Skip To Main Content
backBack to Search

Senior Data Software Engineer

Remote in Colombia, Mexico
Data Software Engineering
& 11 others

We are on the lookout for a seasoned Senior Data Software Engineer skilled in full-stack development and dedicated to automation-first engineering within a modern cloud data warehouse stack (BigQuery/Databricks).

This position focuses on building scalable, production-grade data infrastructure while collaborating with Engineers, Data Analysts, and Data Scientists to generate real-time insights and support data-informed decision-making for senior leadership.

The ideal candidate excels as a contributor in code-driven environments, emphasizing automation, performance tuning, and commitment to clean code principles.

Responsibilities
  • Design high-performance, fault-tolerant data pipelines using Python and SQL with a strong focus on scalability and automation
  • Develop end-to-end production-grade data solutions that integrate ingestion, transformation, and model deployment workflows
  • Build real-time streaming pipelines and batch workflows using BigQuery/Databricks, Apache Airflow, and DBT
  • Write clean, modular code to ensure reusability and eliminate manual data engineering bottlenecks
  • Collaborate with cross-functional teams to translate business requirements into technical deliverables centered on automation capabilities
  • Deploy monitoring, logging, and alerting tools for reliable and scalable data operations
  • Merge backend systems with business logic to streamline broader application development efforts
  • Lead architecture, pipeline, and cloud infrastructure design discussions as part of data engineering projects
Requirements
  • Bachelor's or Master's degree in Computer Science, Software Engineering, or a related discipline
  • 3+ years of experience in production-grade data engineering with attention to full-stack development and automation
  • Proficiency in Python, SQL, and data frameworks such as Spark/PySpark for managing large-scale data processing
  • Competency in modern Cloud Data Warehousing tools like BigQuery or Databricks, with understanding of cloud-native architectures (AWS/GCP/Azure)
  • Knowledge of CI/CD pipelines, version control systems such as Git, and testing frameworks
  • Flexibility to use containerization and orchestration tools, including Docker and Kubernetes, for scalable data applications
  • Familiarity with workflow orchestration platforms such as Apache Airflow and DBT to streamline data pipelines
  • Understanding of event-driven architectures and streaming solutions like Kafka or Kinesis
  • Background in Agile, DevOps, or DataOps methodologies, incorporating infrastructure as code tools (Terraform, Pulumi)
  • Strong communication skills in English, meeting at least a B2 proficiency level
Nice to have
  • Competency in MySQL and visualization platforms such as Looker/Tableau, or familiarity with large-scale analytics systems like Amplitude, Snowplow, or Segment
  • Expertise in cloud DevOps technologies across AWS, GCP, or Azure ecosystems
  • Background in Linux/Unix system administration and basic shell scripting capabilities
  • Showcase of machine learning pipelines, MLOps practices, and deploying operational ML models
  • Understanding of real-time analytics tools and experience with streaming technologies such as Apache Flink or Spark Streaming
Benefits
  • International projects with top brands
  • Work with global teams of highly skilled, diverse peers
  • Healthcare benefits
  • Employee financial programs
  • Paid time off and sick leave
  • Upskilling, reskilling and certification courses
  • Unlimited access to the LinkedIn Learning library and 22,000+ courses
  • Global career opportunities
  • Volunteer and community involvement opportunities
  • EPAM Employee Groups
  • Award-winning culture recognized by Glassdoor, Newsweek and LinkedIn