Data Engineer (Hadoop and Spark)

Vori Health

A better approach to back, neck, and orthopedic care – all in one convenient place.

About the Company

Vori Health is a nationally recognized, virtual-first musculoskeletal medical practice focused on delivering holistic and evidence-based care. The organization connects patients with a multidisciplinary care team—including physical medicine physicians, health coaches, and physical therapists—to guide treatment journeys. The mission of Vori Health is to empower individuals to lead healthier lives.

About the Role

The Data Engineer will play a key role in designing and building tools that support seamless care delivery for both consumers and clinicians. This role emphasizes creating robust, system-wide solutions and maintaining technical excellence in support of scalable healthcare technology.

Responsibilities

  • Design and maintain data lakes and data warehouses
  • Build and optimize ETL/ELT pipelines
  • Support integrations involving payment gateways, fraud detection, subscriptions, and healthcare systems
  • Collaborate across clinical, operational, and technology teams to define and implement business requirements
  • Work closely with product managers, developers, QA, and architects
  • Document scalable solution architectures and technical specifications
  • Guide data best practices and enforce coding standards within the team
  • Evaluate new tools and technologies to enhance resilience, scalability, and security
  • Lead research and development efforts and provide innovative recommendations

Required Skills

  • 3–5 years of experience in healthcare-related data engineering, data science, analysis, modeling, or ETL
  • Proficiency with cloud platforms such as AWS, GCP, or Azure
  • Strong skills in SQL, dbt, Python, R, and NoSQL databases
  • Experience using Airflow, dbt, Looker, Jupyter Notebook, and PySpark
  • Familiarity with Agile methodologies
  • Excellent communication, organizational, and problem-solving skills
  • Ability to operate in a dynamic and fast-paced environment

Preferred Qualifications

  • Experience with big data tools such as Spark, Hadoop, Hive, and Yarn
  • Knowledge of streaming data pipelines and tools like Pandas, Scala, Databricks, and Sagemaker
  • Comfortable working independently and collaboratively
  • Passion for building elegant, scalable, business-focused data solutions
  • Enthusiastic about remote work in a team-driven environment

Visit the official website below to access the full details of this vacancy:

Copyright © 2025 hadoop-jobs. All Rights Reserved.