Senior Data Software Engineer
Colombia
We are on the lookout for a seasoned Senior Data Software Engineer skilled in full-stack development and dedicated to automation-first engineering within a modern cloud data warehouse stack (BigQuery/Databricks).
This position focuses on building scalable, production-grade data infrastructure while collaborating with Engineers, Data Analysts, and Data Scientists to generate real-time insights and support data-informed decision-making for senior leadership.
The ideal candidate excels as a contributor in code-driven environments, emphasizing automation, performance tuning, and commitment to clean code principles.
Responsibilities
- Design high-performance, fault-tolerant data pipelines using Python and SQL with a strong focus on scalability and automation
- Develop end-to-end production-grade data solutions that integrate ingestion, transformation, and model deployment workflows
- Build real-time streaming pipelines and batch workflows using BigQuery/Databricks, Apache Airflow, and DBT
- Write clean, modular code to ensure reusability and eliminate manual data engineering bottlenecks
- Collaborate with cross-functional teams to translate business requirements into technical deliverables centered on automation capabilities
- Deploy monitoring, logging, and alerting tools for reliable and scalable data operations
- Merge backend systems with business logic to streamline broader application development efforts
- Lead architecture, pipeline, and cloud infrastructure design discussions as part of data engineering projects
Requirements
- Bachelor's or Master's degree in Computer Science, Software Engineering, or a related discipline
- 3+ years of experience in production-grade data engineering with attention to full-stack development and automation
- Proficiency in Python, SQL, and data frameworks such as Spark/PySpark for managing large-scale data processing
- Competency in modern Cloud Data Warehousing tools like BigQuery or Databricks, with understanding of cloud-native architectures (AWS/GCP/Azure)
- Knowledge of CI/CD pipelines, version control systems such as Git, and testing frameworks
- Flexibility to use containerization and orchestration tools, including Docker and Kubernetes, for scalable data applications
- Familiarity with workflow orchestration platforms such as Apache Airflow and DBT to streamline data pipelines
- Understanding of event-driven architectures and streaming solutions like Kafka or Kinesis
- Background in Agile, DevOps, or DataOps methodologies, incorporating infrastructure as code tools (Terraform, Pulumi)
- Strong communication skills in English, meeting at least a B2 proficiency level
Nice to have
- Competency in MySQL and visualization platforms such as Looker/Tableau, or familiarity with large-scale analytics systems like Amplitude, Snowplow, or Segment
- Expertise in cloud DevOps technologies across AWS, GCP, or Azure ecosystems
- Background in Linux/Unix system administration and basic shell scripting capabilities
- Showcase of machine learning pipelines, MLOps practices, and deploying operational ML models
- Understanding of real-time analytics tools and experience with streaming technologies such as Apache Flink or Spark Streaming
Benefits
- International projects with top brands
- Work with global teams of highly skilled, diverse peers
- Healthcare benefits
- Employee financial programs
- Paid time off and sick leave
- Upskilling, reskilling and certification courses
- Unlimited access to the LinkedIn Learning library and 22,000+ courses
- Global career opportunities
- Volunteer and community involvement opportunities
- EPAM Employee Groups
- Award-winning culture recognized by Glassdoor, Newsweek and LinkedIn