Company | Norstella |
Job title | Sr. Data Engineer |
Job location | Remote – India |
Type | Full Time |
Responsibilities:
- Design, develop, and optimize scalable data pipelines and ETL processes.
- Work with AWS services (ECS, Lambda, API Gateway, S3, RDS) to deploy and manage data workflows.
- Develop and implement ETL solutions using Python, PySpark, AWS Glue, and AWS Airflow.
- Utilize Postgres (knowledge of RDS and Aurora beneficial) and advanced SQL for complex data transformations, including joins, subqueries, and CTEs.
- Perform data completeness and quality checks to ensure high data accuracy across all feeds.
- Collaborate with data modelers to implement data models (e.g., Snowflake, Star Schema) and optimize performance for data warehousing.
- Implement and maintain CI/CD pipelines for ETL and data processes using AWS DevOps practices.
Requirements & Skills:
-
Bachelor’s degree in Computer Science, Engineering, or a related field.
- 5+ years of experience in data engineering, specifically with cloud data warehousing and ETL development.
- Strong programming skills in Python and PySpark.
- Proficiency with RDBMS, Advanced SQL (including joins, subqueries, CTEs, and writing complex queries), Preference: Postgres.
- Hands-on experience with AWS services (ECS, Lambda, API Gateway, S3, RDS).
- Familiarity with AWS Glue and Airflow for workflow orchestration.
- Solid understanding of data modelling principles, including Snowflake schema and Star schema.
- Experience with ETL testing for data quality and completeness.
- Proven experience with AWS DevOps CI/CD practices.
- Ability to create Functions/Views/Stored Procs. Experience with complex queries including use of CTEs, table variables, merge and dynamic SQL.
- Experience with complex business logic including applying logic to transforms that falls outside of basic schema formatting conversion.