We are looking for a Data Software Engineer / MLOps / MLE to join our team and assist in the implementation of multi-country support for the LATAM market.
The ideal candidate will have a robust background in using Python, Spark, and PySpark and have experience working with AWS and Databricks.
Responsibilities
- Design and build scalable data processing pipelines using Python, Spark, and PySpark
- Implement and maintain AWS infrastructure and deploy applications using Databricks
- Develop and manage CI/CD pipelines for machine learning models using Jenkins and other tools
- Utilize Amazon SageMaker for building, training, and deploying machine learning models
- Monitor and optimize machine learning model performance and efficiency
- Ensure data quality and integrity throughout the data lifecycle
- Collaborate with cross-functional teams during brainstorm meetings to develop new use cases
- Create comprehensive unit tests for various machine learning processes from preprocessing to post-processing
- Adapt data science pipelines based on specific project prerequisites
- Communicate effectively to seek clarifications and discuss technical details
Requirements
- Proficiency in Python, Spark, and PySpark
- Experience in AWS and Databricks
- Expertise in MLOps practices and tools like Jenkins
- Skills in Amazon SageMaker for machine learning model management
- Background in CI/CD tools and Terraform for infrastructure as code
- Capability to wrestle new use-cases quickly and comfortably
- Understanding of key concepts in productionized data science pipelines
- Experience in data preprocessing and modeling
- Competency in creating unit tests for multiple stages of machine learning processes
- Clear communication skills
Nice to have
- Familiarity with Apache Airflow
- Knowledge of data quality best practices
- Experience with Databricks and Jenkins
Looking for something else?
Find a vacancy that works for you. Send us your CV to receive a personalized offer.
Find me a job