Data Engineer architecting and building cloud-based systems for Semios Group, an agricultural technology company. Responsibilities include managing data interfaces, scalable infrastructure, and delivering actionable insights.
Responsibilities
Architect and build Cloud based systems to manage and improve the interface between Semios data and its consumers.
Design, develop and maintain scalable infrastructure to process and store data, integrate data driven models and automate manual processes.
Implement highly scalable big data analytics systems in a cloud environment.
Design and build reliable, monitorable and fault-tolerant data systems & data processes.
Create data tools for analytics and data science team members that assist them in building and optimizing our product into an innovative industry leader.
Assemble large, complex data sets that meet functional / non-functional business requirements.
Continuously identify bottlenecks in the data stack and optimize queries and processes for cost and performance.
Write clear documentation of data processes and products.
Requirements
Advanced skills in SQL; how to write elegant queries; written for humans first, machines second.
The ability to thrive both autonomously and in a team environment.
Hands-on experience with provisioning and developing on cloud platforms (familiarity with GCP is a definite plus).
Experience with at least one Data Warehouse (BigQuery, RedShift, Snowflake, On-Prem).
Excellent verbal & written communication skills: a talent to distill complex ideas to different audiences.
An in-depth experience with Big Data. A proven track-record of effective collection, storage, and access.
Proven experience with workflow and scheduling tools (e.g., like Prefect, Airflow, Dagster, Kubeflow, etc.) and version control (Git).
A fluency in Python, Node or other imperative language or ability to learn quickly and with enthusiasm.
Excellent troubleshooting skills to rapidly identify and resolve issues.
**Nice to have:**
Significant exposure to at least one relational database (Postgres, MySQL).
Real world experience with containers (Docker) & container management systems (Kubernetes).
Experience or Interest in working with IoT Cloud and IoT data.
Familiarity with data transformation tools (dbt, SQLMesh, Dataform) and syncing tools (e.g., dlt, Fivetran, Airbyte).
Senior Data Engineer supporting AI - enabled financial compliance initiative with data pipelines and ingestion processes. Collaborating with diverse teams in a mission - critical regulated environment.
Data Architect leading the definition and construction of cloud data architecture for Kyndryl. Participating in significant technological modernization initiatives, focusing on Google Cloud Platform.
Senior Data Engineer driving data intelligence requirements and scalable data solutions for a global consulting firm. Collaborating across functions to enhance Microsoft architecture and analytics capabilities.
Experienced AI Engineer designing and building production - grade agentic AI systems using generative AI and large language models. Collaborating with data engineers, data scientists in a tech - driven company.
Intermediate Data Engineer designing and building data pipelines for travel industry data management. Collaborating across teams to ensure reliable data for analytics and reporting.
Data Engineer managing and organizing datasets for AI models at Walaris, developing AI - driven autonomous systems for defense and security applications.
Data Engineer designing and maintaining data pipelines at Black Semiconductor. Collaborating with process, equipment, and IT teams to support manufacturing analytics and decision - making.
Junior Data Engineer role focusing on Business Intelligence and Big Data at Avanade. Collaborating on data analysis and SQL queries in a supportive learning environment.
GCP Data Engineer designing and developing data processing modules for Ki, an algorithmic insurance carrier. Working closely with multiple teams to optimize data pipelines and reporting.
Data Engineer at Securian Financial optimizing scalable data pipelines for AI and advanced analytics. Collaborating with teams to deliver secure and accessible data solutions.