Senior Data Software Engineer
Colombia
We are seeking a highly skilled Senior Data Software Engineer with expertise in full-stack development and an automation-first approach to engineering within a modern cloud data warehouse stack (BigQuery/Databricks).
This role involves creating scalable, production-grade data infrastructure while working collaboratively with Engineers, Data Analysts, and Data Scientists to deliver real-time insights and support data-driven decision-making for senior leadership. The ideal candidate is a hands-on contributor who excels in code-focused environments, with an emphasis on automation, performance optimization, and adherence to clean code principles.
Responsibilities
- Design and develop high-performance, fault-tolerant data pipelines using Python and SQL with an emphasis on scalability and automation
- Architect end-to-end production-grade data solutions, integrating ingestion, transformation, and model deployment workflows
- Build and maintain real-time streaming pipelines and batch data workflows leveraging BigQuery/Databricks, Apache Airflow, and DBT
- Write clean, modular code with a focus on reusability and automating manual data engineering tasks
- Collaborate with cross-functional teams to translate business requirements into technical solutions, focusing on automation-friendly approaches
- Implement tools for monitoring, logging, and alerting to ensure the reliability and scalability of data pipelines
- Integrate data workflows with broader application development efforts, merging backend and business logic
- Contribute to design discussions around architecture, pipelines, and cloud infrastructure in data engineering projects
Requirements
- BS/MS in Computer Science, Software Engineering, or a related field
- 3+ years of experience in production-grade data engineering, with a focus on full-stack development and automation
- Proficiency in Python, SQL, and data frameworks such as Spark/PySpark for large-scale data processing
- Expertise in modern Cloud Data Warehousing tools such as BigQuery or Databricks, with understanding of cloud-native architectures (AWS/GCP/Azure)
- Hands-on experience with CI/CD pipelines, version control (Git), and testing frameworks
- Competency in containerization and orchestration technologies like Docker and Kubernetes for scalable data applications
- Understanding of workflow orchestration with tools like Apache Airflow and DBT to automate pipelines
- Familiarity with event-driven architectures and streaming technologies (Kafka, Kinesis, or similar)
- Background in Agile, DevOps, or DataOps methodologies, including infrastructure as code (Terraform, Pulumi)
- Strong communication skills in English, meeting at least a B2 proficiency level
Nice to have
- Experience using MySQL and visualization tools such as Looker/Tableau, or working with large-scale analytics platforms like Amplitude, Snowplow, or Segment
- Hands-on cloud DevOps expertise with AWS, GCP, or Azure
- Basic Linux/Unix system administration and shell scripting skills
- Familiarity with machine learning pipelines, MLOps practices, and deploying ML models
- Experience implementing real-time analytics solutions and streaming technologies like Apache Flink or Spark Streaming
Benefits
- International projects with top brands
- Work with global teams of highly skilled, diverse peers
- Healthcare benefits
- Employee financial programs
- Paid time off and sick leave
- Upskilling, reskilling and certification courses
- Unlimited access to the LinkedIn Learning library and 22,000+ courses
- Global career opportunities
- Volunteer and community involvement opportunities
- EPAM Employee Groups
- Award-winning culture recognized by Glassdoor, Newsweek and LinkedIn