Big Data Engineer building scalable data ingestion platforms at Allegro. Work with advanced data science and AI applications in a dynamic and collaborative environment.
Responsibilities
Build a highly scalable and fault-tolerant data ingestion for millions of Allegro customers
Process 5 billion clickstream events every day from all Allegro sites and mobile applications
Engage in projects based on practical applications of data science and AI
Collaborate within a team of experienced engineers organized into various specialized teams
Requirements
Programming in languages such as Scala or Java, Python
Strong understanding of distributed systems, data storage, and processing frameworks like dbt, Spark or Apache Beam
Knowledge of GCP (especially Dataflow and Composer) or other public cloud environments like Azure or AWS
Use good practices such as clean code, code review, TDD, CI/CD
Navigate efficiently within Unix/Linux systems
Positive attitude and team-working skills
Eager for personal development and keeping knowledge up to date
English at B2 level
Benefits
Flexible working hours in an office first model
Annual bonus depending on your annual assessment and the company's results
Well-located offices with fully equipped kitchens and bicycle parking facilities
Excellent working tools including height-adjustable desks and interactive conference rooms
A wide selection of varied benefits in a cafeteria plan
English classes paid for related to the specific nature of your job
Macbook Pro / Air or Dell with Windows depending on preference
High degree of autonomy in terms of organizing your team’s work
Team tourism, training budget, and an internal educational platform
Senior Data Engineer designing and scaling data foundations for AI adoption across Ad Tech. Collaborating with cross - functional teams to deliver robust pipelines for high - profile AI applications.
Specialist in Data Engineering leading pipeline optimization at Inmetrics. Collaborating in innovative data - driven projects within a hybrid work environment.
Data Architect responsible for designing and implementing data architecture at Stefanini. Collaborate with technical teams and stakeholders in a hybrid work environment.
Senior Data Engineer at Reos responsible for scalable ETL pipelines using Microsoft Fabric. Focused on data integration from various sources and data modeling processes.
Junior Data Engineer developing and maintaining data pipelines for AI - powered identity platform at Saviynt. Collaborating with senior engineers, analysts, and BI developers to ensure reliable data for decision - making.
Strategic technical leader architecting data landscape for Sales and Trade at Conagra Brands. Designing scalable solutions and enhancing data integration across enterprise platforms.
Consultant - Data Engineering Specialist supporting public health surveillance data ecosystem. Focused on automated data integrations and ensuring data flows securely across systems.
Data Engineer developing and maintaining data pipelines for SkyShowtime's streaming data ecosystem. Collaborating with teams to facilitate analysis and operationalise data processing systems.
Tech Lead Data Engineering overseeing data engineering, ETL processes, and cloud technologies. Leading project delivery with strong hands - on experience in Informatica, Python, and GCP.
Data Engineer developing and maintaining data infrastructure for healthcare solutions in Portugal. Working with Oracle databases and Pentaho ETL pipelines in a flexible hybrid model.