MQ Prime is a Virginia based small business who are experts in the world of Cyber Engineering and Software Development. Built on decades of expertise, we support Commercial and Government clients, providing development, design and implementation of cutting-edge solutions. Our personnel maintain support for cyber solutions throughout all of the Government and continue to develop capabilities to fill operational gaps. MQ Prime offers a salary and benefits package that surpasses industry standards while also providing a varied and expanding portfolio of programs at multiple classification levels to enable employee growth. We want you to grow as we do. Come join us! We are looking for a skilled Data Engineer to support a variety of infrastructure, services and analytic deliverables. The primary focus will be to learn existing solutions, and assist in migration of these existing solutions to a recently deployed data lake house environment. After migration, focus will be on developing data-pipeline best practices while helping other departments on-board their data to the new environment. This will be a time of rapid change, and being flexible and adaptable will be the keys to success. Principal Responsibilities: Operate in a cloud-based AWS environment Gain exposure to one of the world's foremost repositories of geospatial data Design, develop, and support scalable, reliable cloud data solutions using open-source and COTS tooling Develop high-quality, resilient data pipelines and business solutions Analyze and interpret data from diverse systems for reporting applications Actively identify opportunities to improve our infrastructure and propose solutions to realize them Collaborate with a team of skilled DevOps Engineers, Data Engineers, and Business Intelligence Developers Minimum Requirements: Bachelor's degree in computer science, geography, or related field. Four additional years of experience may be substituted for a degree Minimum of 5 years of technical experience Minimum of 5 years' experience with SQL including complex queries, SQL tuning, CTEs Minimum of 3 years' experience with Python Demonstrated experience building & orchestrating automated, production-level data pipelines and solutions (ETL/ELT) Experience with file-based data storage, including Parquet or Iceberg Experience with data catalogs (ex. Hive, AWS Glue) General understanding of key AWS services (e.g. EC2, S3, EKS, IAM, lambda) Experience building and/or using data APIs Experience with GitHub Experience with structured, semi-structure, and unstructured data Demonstrated history of exploring and learning new technologies quickly Willing to work on-site in Herndon, VA TS/SCI clearance with CI Poly Preferred qualifications: Experience with software development Experience with geospatial data Experience building data-streaming processes Experience using PostGIS Experience with any of the following: Apache-Hive, Trino, Presto, Starburst, OpenMetadata, Apache-SuperSet, Terraform, dbt, Tableau, Fivetran, Airflow Experience implementing resilient, scalable, and supportable systems in AWS Experience using a wide variety of open-source technologies and cloud services Experience developing multi-step ETLs including DAG creation and scheduling in tools such as Airflow Experience using containerization systems like Docker and Kubernetes Experience with CI/CD pipeline tools such as Jenkins