Senior Data Engineer

Yesterday

Apply Now
Logo of eSimplicity

eSimplicity

Cybersecurity • Information Technology • Telecommunication • Healthcare IT • Spectrum Management

51 - 200

Description

• eSimplicity is a modern digital services company that works across government • Partnering with our clients to improve the health and lives of millions of Americans • Ensure the security of all Americans—from soldiers and veterans to kids and the elderly • Defend national interests on the battlefield • Design, develop, and maintain scalable data pipelines using Spark, Hive, and Airflow • Develop and deploy data processing workflows on the Databricks platform • Develop API services to facilitate data access and integration • Create interactive data visualizations and reports using AWS QuickSight • Builds required infrastructure for optimal extraction, transformation and loading of data from various data sources using AWS and SQL technologies • Monitor and optimize the performance of data infrastructure and processes • Develop data quality and validation jobs • Assembles large, complex sets of data that meet non-functional and functional business requirements • Write unit and integration tests for all data processing code • Work with DevOps engineers on CI, CD, and IaC • Read specs and translate them into code and design documents • Perform code reviews and develop processes for improving code quality • Improve data availability and timeliness by implementing more frequent refreshes, tiered data storage, and optimizations of existing datasets • Maintain security and privacy for data at rest and while in transit

Requirements

• 7+ years of hands-on software development experience • 4+ years of data pipeline experience using Python, Java and cloud technologies • Bachelor's degree in computer science, Information Systems, Engineering, Business, or other related scientific or technical discipline • Experienced in Spark and Hive for big data processing • Experience building job workflows with the Databricks platform • Strong understanding of AWS products including S3, Redshift, RDS, EMR, AWS Glue, AWS Glue DataBrew, Jupyter Notebooks, Athena, QuickSight, EMR, and Amazon SNS • Familiar with work to build processes that support data transformation, workload management, data structures, dependency and metadata • Experienced in data governance process to ingest (batch, stream), curate, and share data with upstream and downstream data users. • Experienced in data pipeline builder and data wrangler who enjoys optimizing data systems and building them from the ground up. • Demonstrated understanding using software and tools including relational NoSQL and SQL databases including Cassandra and Postgres; workflow management and pipeline tools such as Airflow, Luigi and Azkaban; stream-processing systems like Spark-Streaming and Storm; and object function/object-oriented scripting languages including Scala, C++, Java and Python. • Familiar with DevOps methodologies, including CI/CD pipelines (Github Actions) and IaC (Terraform) • Ability to obtain and maintain a Public Trust; residing in the United States • Experience with Agile methodology, using test-driven development.

Benefits

• Highly competitive salary • Full healthcare benefits

Apply Now

Similar Jobs

2 days ago

Diverzify

1001 - 5000

Transform data into valuable tools for Diverzify's operational stakeholders.

3 days ago

Discord

501 - 1000

Senior Data Engineer at Discord building data pipelines for analytics and modeling.

Built by Lior Neu-ner. I'd love to hear your feedback — Get in touch via DM or lior@remoterocketship.com