Posted 1d ago

Data Automation Engineer

@ Smart Tech Skills
United States
RemoteFull Time
Responsibilities:Ingest pipelines, Optimize performance, Automate deployment
Requirements Summary:12+ years data engineering and automation; Python production pipelines; Snowflake, dbt; AWS (S3); Kafka/Spark; CI/CD for data pipelines; strong SQL.
Technical Tools Mentioned:Python, Snowflake, dbt, Kafka, Spark, AWS, SQL
Save
Mark Applied
Hide Job
Report & Hide
Job Description
Benefits:
  • Competitive salary
Location
100% Remote

Experience Level
Senior Level (12 or more years of data engineering experience)

Role Overview
The Data Automation Engineer is responsible for designing, building, and optimizing scalable data ingestion and automation pipelines that support both product telemetry and enterprise business data. This is a senior, hands-on role operating in a fast-paced SaaS environment, focused on rapid prototyping, performance optimization, and production-grade pipeline delivery. The role supports analytics, data exploration, and downstream consumption using a modern cloud-based data stack.

Key Responsibilities
Data Ingestion & Pipeline Development
  • Design and build end-to-end data ingestion pipelines from source systems into Snowflake
  • Ingest data from product telemetry and observability platforms
  • Ingest data from enterprise business systems including CRM, ERP, financial, consumption, and renewal platforms
  • Develop ingestion solutions using Python, bulk APIs, and AWS-based object storage
  • Implement streaming and distributed processing using Kafka, Spark, or similar frameworks
Data Processing & Optimization
  • Apply advanced SQL transformations in Snowflake to produce analytics-ready datasets
  • Develop and maintain data transformations using dbt
  • Optimize ingestion performance for large-scale, high-volume datasets
  • Implement incremental loading, change data capture, micro-batching, and API rate-limit handling
  • Ensure pipelines are scalable, efficient, resilient, and production-ready
Automation, DevOps & Reliability
  • Implement CI/CD pipelines and DevOps best practices for data workflows
  • Automate deployment, versioning, and environment management
  • Build observability, monitoring, and alerting into ingestion pipelines
  • Monitor, troubleshoot, and optimize ingestion jobs to ensure data freshness and accuracy
Collaboration & Rapid Prototyping
  • Partner with data analysts, subject matter experts, and platform teams to prototype ingestion solutions
  • Translate evolving business and product requirements into working data pipelines
  • Support data exploration initiatives with fast turnaround and iterative development
Required Qualifications
  • 12 or more years of hands-on experience in data engineering and data automation
  • Strong experience supporting SaaS data platforms
  • Strong expertise in Python for production-grade pipeline development
  • Strong SQL skills with experience supporting analytical workloads
  • Extensive hands-on experience with Snowflake for data ingestion, transformation, and performance optimization
  • Hands-on experience with dbt for data modeling and transformations
  • Experience working with AWS, including S3 and cloud-based data architectures
  • Proven experience with Kafka, Spark, or similar distributed or streaming frameworks
  • Deep understanding of change data capture, incremental loading, micro-batching, and API rate limiting
  • Experience implementing CI/CD and DevOps practices for data pipelines
  • Ability to pass hands-on SQL and Python coding assessments
Preferred Qualifications
  • Experience with open-source ingestion frameworks or connectors
  • Experience working in telemetry-heavy or event-driven architectures
  • Experience supporting high-scale SaaS or data platform environments
  • Consideration may be given to candidates with 8 or more years of experience who demonstrate deep streaming, Snowflake, and Python expertise
Core Skills & Attributes
  • Strong problem-solving and performance optimization mindset
  • Ability to work independently in fast-paced, iterative environments
  • High attention to data quality, reliability, and scalability
  • Strong collaboration and communication skills across technical teams
  • Focus on building durable, production-grade data automation solutions

This is a remote position.