Responsible for building and supporting a GCP-based ecosystem designed for enterprise-wide analysis of structured, semi-structured, and unstructured data.
Support regular requests to move data from one cluster to another
Make sure service levels are maintained, and any interruption is resolved in a timely fashion
Analyze requirements, design AI/ML-based solutions, and integrate those solutions for customer environments.
Closely collaborates with team members to successfully execute development initiatives using Agile practices and principles
Leads efforts to design, develop, deploy, and support software systems
Collaborates with business analysts, project lead, management, and customers on requirements
Participates in the deployment, change, configuration, management, administration, and maintenance of the deployment process and systems
Proven experience in effectively prioritizing workload to meet deadlines and work objectives
Gather requirements, designs, constructs, and delivers solutions with minimal team interaction
Works in an environment with rapidly changing business requirements and priorities
Bring new data sources into GCP, transform, and load to BQ and databases.
Work collaboratively with Data Scientists and business and IT leaders throughout the company to understand Data needs and use cases.
Requirements & Skills:
Bachelor’s degree in computer science, related technical field, or equivalent experience
Master’s degree in computer science or related field preferred. 3+ years of experience in Data Engineering required
1+ year(s) of experience in Healthcare preferred
5+ years of experience in Information Technology required
Good understanding of best practices and standards for GCP Data process design and implementation.
One plus Years of hands-on experience with GCP platform and experience with many of the following components: Cloud Run, GKE, Cloud Functions, Spark Streaming, Kafka, Pub/Sub, Bigtable, Firestore, Cloud SQL, Cloud Spanner, JSON, Avro, Parquet, Python, Java, Terraform, Big Query, Dataflow, Data Fusion, Cloud Composer, DataProc, CI/CD, Cloud Logging, Vertex AI, NLP, GitHub
Ability to multitask and to balance competing priorities.
Ability to define and utilize best practice techniques and to impose order in a fast-changing environment.
Strong problem-solving skills.
Strong verbal, written, and interpersonal skills, including a desire to work within a highly matrixed, team-oriented environment.
Demonstrates an empathetic and growth mindset with a willingness to learn new skills, technologies, and methodologies – Required
Growing knowledge of public cloud best practices and design patterns used in creating, automating, and supporting data pipelines – Required
Ability to assemble large, complex sets of data that meet functional and non-functional product requirements – Required
Helps create and use analytical tools to monitor data pipeline metrics and provide actionable intelligence to increase operational efficiency and valuable data outcomes – Required
Ability to use source control management tools such as Git/GitHub – Required
Ability to use CI/CD automation tools – Required
Understanding of SQL and analytical data warehouses- Required
Understanding of Agile methodologies and how to apply Agile within the team – Required
Proven ability to complete work, make sound decisions, and plan and accomplish goals with direction/guidance from leadership – Required
Builds and nurtures healthy relationships with all colleagues – Required
Stays abreast of public cloud technologies, capabilities, and industry use of public cloud to help guide HCA’s strategy and adoption – Required