Design, develop and maintain large scale batch and real-time pipelines with data processing frameworks like Scio, Apache Beam and Google Cloud Platform
Deliver high quality code that is scalable, testable and maintainable
Collaborate with diverse, cross-functional teams to define data requirements and translate them into actionable solutions
Implement best practices for data quality, security and governance
Continuously learn and take on leadership roles that make a real impact
Support and learn from engineers in your domain and across the organization
Requirements & Skills:
You have Data Engineering experience and you know how to work with high-volume, heterogeneous data, preferably with distributed systems such as Hadoop, BigTable, Cassandra, or DynamoDB
You have experience building high-volume data pipelines using Java and/or Scala
You have a strong understanding of data modeling concepts, data access patterns and various data storage technologies
You proactively ensure the adaptability and scalability of data models and pipelines
You value agile methodologies, data-driven decision making, reliability and incremental delivery
You value clear documentation and possess strong data debugging skills