Posted 2w ago

Senior Data Engineer

@ Steampunk
McLean, Virginia, United States
RemoteFull Time
Responsibilities:Build pipelines, Prepare data, Lead migration
Requirements Summary:Senior data engineering with cloud data warehousing, Python, AWS; strong SQL; experience with big data tools; able to lead migrations and address enterprise data concerns.
Technical Tools Mentioned:Python, AWS, Hadoop, Spark, Kafka, PostgreSQL, Cassandra, Azkaban, Luigi, Airflow, EC2, EMR, Redshift, Solr, Elasticsearch, Lucene, Storm, Spark-Streaming
Save
Mark Applied
Hide Job
Report & Hide
Job Description

We are looking for seasoned Senior Data Engineer to work with our team and our clients to develop enterprise grade data platforms, services, and pipelines. We are looking for more than just a "Senior Data Engineer", but a technologist with excellent communication and customer service skills and a passion for data and problem solving. 

  • Build and manage data pipelines and data models
  • Prepare data for downstream analytics and reporting
  • Support reporting and visualization needs
  • Lead and architect migration of data environments with performance and reliability.
  • Assess and understand the ETL jobs, workflows, BI tools, and reports
  • Address technical inquiries concerning customization, integration, enterprise architecture and general feature / functionality of data products
  • Experience in crafting database / data warehouse solutions in cloud (Preferably AWS. Alternatively Azure, GCP).
  • Key must have skill sets – Python, AWS
  • Support an Agile software development lifecycle
  • Ability to hold a position of Public Trust with the US government
  • Bachelor’s degree and 10 years of experience
    • OR Master's degree and 8 years of experience
  • 5-7 years industry experience coding commercial software and a passion for solving complex problems.  
  • 5-7 years direct experience in Data Engineering with experience in tools such as: 
    • Big data tools: Hadoop, Spark, Kafka, etc. 
    • Relational SQL and NoSQL databases, including Postgres and Cassandra. 
    • Data pipeline and workflow management tools: Azkaban, Luigi, Airflow, etc. 
    • AWS cloud services: EC2, EMR, RDS, Redshift (or Azure equivalents) 
    • Data streaming systems: Storm, Spark-Streaming, etc. 
    • Search tools: Solr, Lucene, Elasticsearch 
    • Object-oriented/object function scripting languages: Python, Java, C++, Scala, etc. 
  • Strong SQL and data modeling experience
  • Experience with data warehousing/lake architectures
  • Hands-on experience with reporting tools (Power BI or similar)
  • Ability to translate business needs into usable data structures