**Architect and Evolve Our Core Data Platform** You will own the technical vision and roadmap for our data platform, steering its evolution on our modern cloud stack and ensuring it meets the demands of a rapidly scaling business.
**Own the Architecture:** Design, implement, and refine a robust data lakehouse architecture (e.g., Medallion) using Databricks and Delta Lake to ensure data reliability and performance.
**Build Scalable Ingestion Frameworks:** Develop and maintain resilient, reusable patterns for ingesting data from a diverse set of sources, including our systems, transactional databases, event streams, and third-party SaaS APIs.
**Define Data Modelling Standards:** Lead the implementation of our core data modelling principles (e.g., Kimball dimensional modelling) to produce curated, intuitive datasets for business intelligence and product analytics.
**Implement Robust Governance:** Use tools like Unity Catalog to establish a comprehensive data governance framework, covering data lineage, fine-grained access controls, and a user-friendly data catalogue.
**Manage Platform Performance and Cost:** Develop and implement strategies for monitoring, optimising, and forecasting our Databricks and cloud expenditure, ensuring the platform is both powerful and cost-effective.
**Champion Engineering Excellence and Best Practice** You will be the driving force for maturing our data operations, embedding a culture of quality, automation, and reliability into everything we do.
**Automate Everything with CI/CD:** Implement and advocate for automated CI/CD pipelines (e.g., using GitHub Actions) for all data assets, including dbt models, infrastructure changes, and Databricks jobs.
**Embed Git-Based Workflows:** Champion a Git-first culture for all data transformation code, establishing clear processes for branching, code reviews, and version control.
**Embed Automated Data Quality: **Implement comprehensive, automated data quality testing at every stage of our pipelines using tools like dbt test, ensuring data is accurate and trustworthy.
**Introduce Data Observability:** Establish thorough monitoring, logging, and alerting for all data pipelines to proactively detect, diagnose, and resolve issues before they impact the business**
Requirements
Mastery of data architecture principles, data modelling frameworks (e.g., dimensional modelling), and a strong understanding of data governance and security best practices.
A strong software engineering mindset, with significant experience implementing CI/CD for data, Git-based workflows, and automated data quality testing.
Exceptional communication and stakeholder management skills, with a proven ability to translate complex technical concepts for non-technical audiences and influence business decisions.
A genuine passion for leadership and mentorship, with a track record of elevating the technical skills of those around you.
**Tech Stack:**
Dbt
Databricks, Unity Catalog
Terraform
AWS: Redshift, Dynamo db, API gateway, Cloud Watch, Lambda, Streaming with Kenisis/Firehose, Glue, Bedrock
Stitch & Fivetran
Languages required include advanced SQL, python
Benefits
Enjoy a flexible remote-first work policy (with a work-from-home stipend to set you up for success!)
Own A piece of Deputy via our Employee Share Ownership Plan (ESOP)
Take paid parental leave to support you and your family
Stay protected with Group Salary Continuance Insurance
Access support through our Employee Assistance Program
Enjoy additional leave days — including study assistance, celebration days and volunteering
Join our global working groups focused on collaboration, belonging and connection
Get creative at our annual Hackathons
Take advantage of our novated leasing for electric vehicles, internet reimbursement and more!
Data Engineer joining Alterric to collaborate on data platform projects and analytics solutions. Working with Azure Cloud technologies to ensure data quality and integrity for informed decision - making.
Data Engineer at Kyndryl transforming raw data into actionable insights using ELK Stack. Responsible for developing, implementing, and maintaining data pipelines and processing workflows.
Senior Data Engineer at Clorox developing cloud - based data solutions. Leading data engineering projects and collaborating with business stakeholders to optimize data flows.
Data Engineer building solutions on AWS for high - performance data processing. Leading initiatives in data architecture and analytics for operational support.
Senior Data Engineer overseeing Databricks platform integrity, optimizing data practices for efficient usage. Leading teams on compliance while mentoring a junior Data Engineer.
Associate Data Engineer contributing to software applications development and maintenance using Python. Collaborating with teams for clean coding and debugging practices in Pune, India.
Data Engineer focusing on development and optimization of data pipelines in an insurance context. Ensuring data integrity and supporting data - driven decision - making processes.
Lead Data Engineer responsible for delivering scalable cloud - based data solutions and managing cross - functional teams. Collaborating with global stakeholders and ensuring high - quality project execution in a fast - paced environment.
Full Stack Data Engineer on a Central Engineering Portfolio Team in Chennai delivering curated data products and collaborating with data engineers and product owners.