Skip To Main Content
backBack to Search

Senior Data Software Engineer

Remote in Colombia, Mexico
Data Software Engineering
& 11 others

We are thrilled to invite a highly skilled Senior Data Software Engineer with expertise in full-stack development and a focus on automation-first engineering within a modern cloud data warehouse stack (BigQuery/Databricks).

This role centers around creating scalable, production-grade data infrastructure while collaborating with Engineers, Data Analysts, and Data Scientists to deliver real-time insights and enable data-driven decision-making for senior leadership.

The ideal candidate thrives as a hands-on contributor in code-driven environments, strongly emphasizing automation, performance optimization, and adherence to clean code principles.

Responsibilities
  • Design and develop high-performance, fault-tolerant data pipelines using Python and SQL, emphasizing scalability and automation
  • Architect end-to-end production-grade data solutions by integrating ingestion, transformation, and model deployment workflows
  • Build and maintain real-time streaming pipelines and batch workflows using BigQuery/Databricks, Apache Airflow, and DBT
  • Write clean, modular code while prioritizing reusability and automating manual data engineering processes
  • Collaborate with cross-functional teams to translate business requirements into technical solutions, prioritizing automation capabilities
  • Implement monitoring, logging, and alerting tools to ensure reliable and scalable data pipelines
  • Integrate data workflows with broader application development efforts by merging backend systems with business logic
  • Contribute to architecture, pipeline, and cloud infrastructure design discussions within data engineering projects
Requirements
  • BS/MS in Computer Science, Software Engineering, or a related field
  • 3+ years of experience in production-grade data engineering with attention to full-stack development and automation
  • Proficiency in Python, SQL, and data frameworks such as Spark/PySpark for large-scale data processing
  • Expertise in modern Cloud Data Warehousing tools such as BigQuery or Databricks, with understanding of cloud-native architectures (AWS/GCP/Azure)
  • Hands-on experience with CI/CD pipelines, version control (Git), and testing frameworks
  • Competency in containerization and orchestration technologies like Docker and Kubernetes for scalable data applications
  • Understanding of workflow orchestration with tools like Apache Airflow and DBT to automate data pipelines
  • Familiarity with event-driven architectures and streaming technologies (Kafka, Kinesis, or similar)
  • Background in Agile, DevOps, or DataOps methodologies, including infrastructure as code (Terraform, Pulumi)
  • Strong communication skills in English, meeting at least a B2 proficiency level
Nice to have
  • Skills in MySQL and visualization tools such as Looker/Tableau, or experience with large-scale analytics platforms like Amplitude, Snowplow, or Segment
  • Showcase of cloud DevOps expertise with AWS, GCP, or Azure
  • Background in basic Linux/Unix system administration and shell scripting
  • Familiarity with machine learning pipelines, MLOps practices, and deploying ML models
  • Understanding of real-time analytics solutions and streaming technologies like Apache Flink or Spark Streaming
Benefits
  • International projects with top brands
  • Work with global teams of highly skilled, diverse peers
  • Healthcare benefits
  • Employee financial programs
  • Paid time off and sick leave
  • Upskilling, reskilling and certification courses
  • Unlimited access to the LinkedIn Learning library and 22,000+ courses
  • Global career opportunities
  • Volunteer and community involvement opportunities
  • EPAM Employee Groups
  • Award-winning culture recognized by Glassdoor, Newsweek and LinkedIn