See all jobs
Darwoft
Data Engineer
Posted 2 weeks ago

Location: LATAM (Remote supporting US-based teams)
Job Type: Contractor (Full-Time, Remote)
Project: Data Platform Modernization Healthcare Sector
Time Zone: Aligned with GMT-3 (Argentina)
English Level: B2/C1

Get to Know Us

At Darwoft, we build software that drives real change. But were more than just tech were people first. With a remote-first culture and a highly collaborative team spread across LATAM, we partner with global companies to co-create reliable, scalable, and impactful digital products.

Were currently working with a leading US-based healthtech platform, in a major transformation of their data pipeline ecosystem migrating legacy SQL logic into modern, scalable cloud-based infrastructure using DBT, Spark, Argo, and AWS.

Were Looking For a Senior Data Engineer (DBT + Spark + Argo)

In this role, you will be at the core of a strategic data transformation initiative: converting monolithic SQL Server logic into a modular, testable DBT architecture, while integrating Spark for performance and Argo for orchestration. You will work with cutting-edge lakehouse formats like Apache Hudi, Parquet, and Iceberg, and enable real-time analytics through ElasticSearch integration.

If you’re passionate about modern data engineering and want to work in a data-driven, cloud-native, healthcare-focused environment, this is the role for you.

What Youll Be Doing
  • Translate legacy T-SQL logic into modular, scalable DBT models powered by Spark SQL
  • Build reusable and performant data transformation pipelines
  • Develop testing frameworks to ensure data accuracy and integrity in DBT workflows
  • Design and orchestrate workflows using Argo Workflows and CI/CD pipelines with Argo CD
  • Manage mock data and reference datasets (e.g., ICD-10, CPT), ensuring version control and governance
  • Implement efficient storage/query strategies using Apache Hudi, Parquet, and Iceberg
  • Integrate ElasticSearch for analytics by building APIs and pipelines to support indexing and querying
  • Collaborate with DevOps teams to optimize S3 usage, enforce data security, and ensure compliance
  • Work in Agile squads and participate in planning, estimation, and sprint reviews
What You Bring
  • Strong experience with DBT for data modeling, testing, and deployment
  • Hands-on proficiency in Spark SQL, including performance tuning
  • Solid programming skills in Python for automation and data manipulation
  • Familiarity with Jinja templating for building reusable DBT components
  • Practical experience with data lake formats: Apache Hudi, Parquet, Iceberg
  • Expertise in Argo Workflows and CI/CD integration with Argo CD
  • Deep understanding of AWS S3 data storage, performance tuning, and cost optimization
  • Strong command of ElasticSearch for indexing structured/unstructured data
  • Knowledge of ICD-10, CPT, and other healthcare data standards
  • Ability to work cross-functionally in Agile environments
Nice to Have
  • Experience with Docker, Kubernetes, and container orchestration
  • Familiarity with cloud-native data tools: AWS Glue, Databricks, EMR, or GCP equivalents
  • Prior work on CI/CD automation for data engineering workflows
  • Knowledge of data compliance standards: HIPAA, SOC2, etc.
  • Contributions to open-source projects in DBT, Spark, or data engineering frameworks
    Perks & Benefits
    • Contractor agreement with payment in USD
    • 100% remote work
    • Argentinas public holidays
    • English classes
    • Referral program
    • Access to learning platforms

    Explore this and other opportunities at:

About The Company

Status

Accepting Applications

Apply

Similar jobs

New
UHS
Licensed Telehealth Therapist
UHS
Remote Latam's
New
Tractian
Deal Desk Lead
Tractian
Remote Latam's
New
Ocient
Lead Revenue Accountant
Ocient
Remote Latam's