Hybrid Senior Data Engineer – Airflow, PySpark

Posted 4 weeks ago

Apply now

About the role

  • Senior Data Engineer leading data engineering initiatives for automotive client. Building and optimizing data pipelines using Python and PySpark in a hybrid work model.

Responsibilities

  • Design, develop, and optimize scalable data pipelines using Python and PySpark for batch and streaming workloads
  • Build, schedule, and monitor complex workflows using Airflow, ensuring reliability and maintainability
  • Architect and implement CI/CD pipelines for data engineering projects using GitHub, Docker, and cloud-native solutions (GCP preferred)
  • Apply test-driven development (TDD) practices and automate unit/integration tests for data pipelines
  • Implement secure coding best practices and design patterns throughout the development lifecycle
  • Work closely with Data Architects, QA teams, and business stakeholders to translate requirements into technical solutions
  • Create and maintain technical documentation, including process/data flow diagrams and system design artifacts
  • Lead and mentor junior engineers, providing guidance on coding, testing, and deployment best practices
  • Analyze and resolve technical issues across the data stack, including pipeline failures and performance bottlenecks
  • Cross-train team members outside the project team (e.g., operations support) for full knowledge coverage

Requirements

  • 7+ years of Data Engineering experience building production-grade data pipelines using Python and PySpark
  • Experience designing, deploying, and managing Airflow DAGs in enterprise environments
  • Experience maintaining CI/CD pipelines for data engineering workflows, including automated testing and deployment
  • Experience with cloud workflows and containerization, using Docker and cloud platforms (GCP preferred) for data engineering workloads
  • Knowledge and ability to follow twelve-factor design principles
  • Experience and ability to write object-oriented Python code, manage dependencies, and follow industry best practices
  • Proficiency with Git for source code management and collaboration (commits, branching, merging, GitHub/GitLab workflows).
  • Experience working with command lines in Unix/Linux-like environments
  • Solid understanding of SQL for data ingestion and analysis
  • Engineering mindset. Able to write code with an eye for maintainability and testability
  • Collaborative mindset. Comfortable with code reviews, paired programming, and using remote collaboration tools effectively
  • Detroit Labs is not currently able to hire candidates who will reside outside of the United States during their term of employment

Benefits

  • Full medical, dental, vision benefits
  • 401K contribution options
  • Quarterly outings and events
  • Paid holidays and vacation time
  • Parental leave program
  • Monthly budgets for “team fun” bonding events
  • Free lunch for various company meetings and Lunch & Learns
  • Access to our mentorship program and employee resource groups (ERGs)
  • Volunteer opportunities
  • All-company remote-friendly activities
  • Plenty of Detroit Labs swag

Job title

Senior Data Engineer – Airflow, PySpark

Job type

Experience level

Senior

Salary

$160,000 - $180,000 per year

Degree requirement

Bachelor's Degree

Location requirements

Report this job

See something inaccurate? Let us know and we'll update the listing.

Report job