
Company |
Capgemini |
Job title |
Big Data Engineer |
Job location |
Bridgewater NJ, Chicago IL, Atlanta GA |
Type |
Full Time |
Responsibilities:
- Design, develop, and maintain scalable and reliable data pipelines for the extraction, transformation, and loading (ETL) of data from various sources.
- Develop, optimize, and maintain complex data processing jobs, leveraging distributed computing frameworks such as Apache Spark, Flink, or similar.
- Monitor, troubleshoot, and improve the performance of data systems, ensuring minimal latency and optimal throughput for data processing
Requirements & Skills:
- 6- 8 years of experience on Big Data
- Strong hands-on experience as a Hadoop developer /big data engineer.
- Advanced knowledge of the Hadoop ecosystem and its components
- In-depth knowledge of Scala, Spark, and similar frameworks.
- In-depth knowledge of HBase, SOLR, Hive, Gremlin and Pig
- Nice to have: Knowledge/experience on Spark and Flume/NiFi
- Nice to have: Experience on Azure stack as data engineer working on Synapse analytics or ADF or Azure Databricks
