Design data structures for data ingestion, integration, and analytics layers on a variety of technologies.
Working in a cross-skilled squad with people both from business and technology to build end-to-end features.
Contribute to design patterns for on-premises and cloud
Optimise data flows by building a robust, fault-tolerant data pipeline that cleans, transforms, and aggregates unorganized and messy data into databases or data sources.
Support, run, and optimize the data assets that are built
Requirements & Skills:
MS or BS in Computer Science, Engineering, Mathematics, Statistics, or equivalent practical experience in Data Engineering
A strong background in data processing and software engineering and can build high-quality, scalable data pipelines
Strong Software Engineering experience with proficiency in at least one high-level programming language (Python, Java, Scala)
Experience and proficiency in at least one high-volume data processing environment (Teradata, Cloudera, Trino/Presto, Big Query, or equivalent)
Extensive experience working within cross-functional, agile, or continuous delivery teams
Exposure to CICD/DevOps to deploying the pipelines
Strong ability to translate data insights into practical business recommendations
Understand the fundamentals of distributed computing