Lead AI Platform Engineer
Hybrid in Mexico
Cloud Native Development
& 7 others
Mexico
We are seeking a skilled and motivated Lead AI Platform Engineer to join our team.
In this position, you will be responsible for designing and implementing large-scale AI/ML infrastructure to tackle significant challenges in healthcare and drug discovery. This role offers a unique opportunity to build advanced platforms that empower data scientists to develop impactful solutions, driving innovation in global healthcare.
Responsibilities
- Develop and manage infrastructure and platforms to facilitate the deployment and monitoring of machine learning solutions in production environments
- Improve system performance and scalability to support large-scale operations
- Collaborate with data science teams to create and implement AI/ML workflows and environments on AWS
- Work closely with R&D data scientists to operationalize machine learning algorithms, pipelines, and models
- Oversee the entire software engineering lifecycle, from architecture and development to testing and maintenance
- Lead the execution of technology initiatives, from concept to successful implementation
- Upgrade the existing technology stack by integrating the latest advancements in AI and data processing
- Oversee an enterprise-level platform and service, addressing customer requirements and feature requests
- Adopt DevOps practices and modern tools to enhance automation and streamline processes
- Scale MLOps environments to meet production-grade standards
- Ensure systems comply with GxP standards when required
Requirements
- At least 5 years of experience working in AWS cloud environments, with expertise in services like SageMaker, Athena, S3, EC2, RDS, Glue, Lambda, Step Functions, EKS, and ECS
- Proficiency in infrastructure-as-code tools such as Terraform, Ansible, or CloudFormation
- Strong programming skills, particularly in Python, with openness to other programming expertise
- Experience with containerization and microservices architectures, using platforms like Kubernetes or Docker
- In-depth knowledge of Continuous Integration and Continuous Delivery pipelines, using tools like CodePipeline, CodeBuild, or CodeDeploy
- Proven success managing large-scale enterprise platforms and addressing feature requests from end users
- Hands-on experience with DevOps tools and practices, including Docker and Git
- Familiarity with GxP compliance standards
- Strong problem-solving, communication, and analytical skills
Nice to have
- Experience building large-scale data processing systems with technologies such as Hadoop, Spark, or SQL
- Proficiency in data science modeling tools and environments, including R, Python, or Jupyter Notebooks
- Knowledge of multi-cloud platforms, such as AWS, Azure, and GCP
- Experience mentoring and guiding team members or clients in a professional capacity
- Familiarity with SAFe Agile methodologies and frameworks
We offer/Benefits
- International projects with top brands
- Work with global teams of highly skilled, diverse peers
- Healthcare benefits
- Employee financial programs
- Paid time off and sick leave
- Upskilling, reskilling and certification courses
- Unlimited access to the LinkedIn Learning library and 22,000+ courses
- Global career opportunities
- Volunteer and community involvement opportunities
- EPAM Employee Groups
- Award-winning culture recognized by Glassdoor, Newsweek and LinkedIn