‹ Back

DataOps Engineer

JOB SUMMARY

ArgentinaPosted on 2/12/2026
Apply
Sponsored
SwiftPrep Logo

SwiftPrep

Ace your interview at Mutt Data

Get a tailored interview study plan, cheat-sheet, and find contacts for referrals.

Real interview questions and answers from Glassdoor, Reddit, Blind
Role-specific prep plan and cheatsheet tailored to Mutt Data
Find insiders for referrals
Get Your Prep Plan
Optimize your resume with Teal - AI-powered resume builder and job tracking tools

Job details

🚀 Join Our Data Products and Machine Learning Development Remote Startup. 🚀Muttdata is a dynamic startup committed to crafting innovative systems using cutting-edge Big Data and Machine Learning technologies.

We’re looking for a DataOps Engineer to help take our expertise —and our internal optimization platform— to the next level.

If you consider yourself a data nerd like us, we’d love to connect.

As a key contributor, you'll play a pivotal role in designing, deploying, and maintaining ML models that power our product for marketing campaign budget optimization and al

location
.


You’ll work closely with the product team and business stakeholders, bringing robust systems into production, introducing fresh technical insights, and ensuring our models deliver real business impact.

If you thrive on building scalable ML pipelines, solving complex challenges, and shaping products that drive decision-making at scale, we’re eager to connect with you. 🐶🚀🚀 What We DoLeveraging our expertise, we build modern Machine Learning systems for demand planning and budget forecasting. Developing scalable data infrastructures, we enhance high-level decision-making, tailored to each client. Offering comprehensive Data Engineering and custom AI solutions, we optimize cloud-based systems. Using Generative AI, we help e-commerce platforms and retailers create higher-quality ads, faster. Building deep learning models, we enhance visual recognition and automation for various industries, improving product categorization, quality control, and information retrieval. Developing recommendation models, we personalize user experiences in e-commerce, streaming, and digital platforms, driving engagement and conversions. 🌟 Our PartnershipsAmazon Web ServicesAstronomerDatabricks🌟

Our Values

📊 We are Data Nerds🤗 We are Open Team Players🚀 We Take Ownership🌟 We Have a Positive Mindset🔍 Curious about what we’re up to. Check out our case studies and dive into our blog post to learn more about our culture and the exciting projects we’re working on. 🚀Responsibilities 🤓Orchestration Integration: Build and manage scalable data pipelines using Apache Airflow, dbt, and Airbyte, ensuring seamless data ingestion and movement. Product Development: Work hand in hand with Mixilo’s product team to address real client and internal users' issues, design technical solutions, and help prioritize the roadmap.

Infrastructure as Code (IaC): Use Terraform to provision and manage cloud resources on AWS, maintaining a secure and cost-effective infrastructure. Kubernetes GitOps: Manage containerized applications and services on Kubernetes (EKS), implementing continuous delivery practices to keep Mixilo running smoothly. Enhance DX (Developer Experience): Abstract away complex DAG and dbt logic to reduce manual work for the team and optimize our time to market. Data Reliability: Implement rigorous testing frameworks—specifically leveraging dbt tests—to ensure data quality and catch errors before they impact client recommendations. CI/CD for Data: Maintain and improve our CI/CD pipelines to automate testing, deployment, and infrastructure changes. Required Skills 💻Senior Data Engineering Foundations (Minimum 5+ Years): Strong experience in Python and a deep mastery of SQL and Postgres. Modern Data Stack: Hands-on experience with Airflow, dbt, and Airbyte. Analytical Data Systems: Experience in constructing analytical data systems over Data Lakes (e. g. ,AWS S3, Athena, EMR, Glue, Iceberg/Delta, etc. ). Cloud Mastery: Solid understanding of AWS services (S3, EC2, RDS, IAM, etc. ).

Infrastructure Containers: Proficiency with Terraform, Docker, and Kubernetes. Operations Mindset: A strong understanding of GitOps, CI/CD principles, and a passion for automation. Teamwork: Great capacity for collaborative, async, and written communication.

You take ownership and follow through on your commitments. Nice to have skills 😉Observability: Experience setting up monitoring and alerting systems to ensure the health of data pipelines and infrastructure. Code Hygiene: A sharp sense of code hygiene, including code review, documentation, testing, and CI/CD (Continuous Integration/Continuous Delivery). Stream Processing Knowledge: Experience with stream processing tools like Kafka Streams, Kinesis, or Spark. Python's Scientific Stack: Proficiency in Python's Scientific Stack, including numpy, pandas, jupyter, matplotlib, scikit-learn, and related tools. English Proficiency: Solid command of the English language for writing technical documents, such as Design Documents. 🎁 PerksRemote-first culture – work from anywhere. 🌍AWS, DBT, Google Cloud, Azure Databricks certifications fully covered In-Company English Lessons. Birthday off + an extra vacation week (Mutt Week. 🏖️)Referral bonuses – help us grow the team get rewarded. Maslow: Monthly credits to spend in our

benefits

marketplace. ✈️🏝️ Annual Mutters' Trip – an unforgettable getaway with the team.