This role will involve working alongside the CTO, and a team of engineers, on various business engagements, such as client projects and internal company products.
Some examples of these engagements include client scoping programmes, and PoC/MVP software development.
A significant focus will be on building and optimizing Large Language Model (LLM) inferences and creating robust web services.
This includes developing event-driven and request-response systems to run RAG (Retrieval-Augmented Generation) answer generation pipelines, essential for delivering sophisticated AI-driven solutions.
Requirements & Skills:
Proficient in using Amazon Web Services(AWS) for cloud-based AI solution management and deployment.
Skilled in infrastructure as code for efficient cloud infrastructure setup using Terraform
Capable of orchestrating and managing containerized applications using Kubernetes
Experience with MLOps frameworks like Kubeflow, MLflow, Ray, and Spark is advantageous, facilitating efficient management and orchestration of machine learning workflows and scalable data processing.
Excellent programming skills (10 years+) in Python and/or C/C++ and familiarity with object-oriented programming.
Proficiency in git version control, branching, and code versioning.
Passionate about code quality, adhering to best practices for code quality, performance, testing, monitoring, documentation, CI/CD.
Experience working in an agile framework, defining functional and non-functional requirements and sprint tasks.
Strong communication skills with the ability to communicate with both technical and commercial people.