leap
leap1mo ago
New

Senior Data Engineer

United StatesUnited StatesRemotefull-timesenior
Data EngineerData
0 views0 saves0 applied

Quick Summary

Overview

About Leap Leap is one of the fastest-growing benefits solutions and a category-defining pioneer in employer specialty pharmacy. We are reshaping how life-changing therapies are delivered and financed, ensuring patients get the treatment they need while employers finally get a fair deal.

Key Responsibilities

Pipelines and Warehouse Build and own data pipelines and ETL processes for claims ingestion, drug pricing, and CRM sync using BigQuery and Python Design production pipelines for batch and streaming workloads, with a particular focus on high-volume…

Requirements Summary

Required 5+ years of experience with Python, SQL, and dbt, with hands-on expertise in BigQuery, Snowflake, or a comparable cloud data warehouse and proficiency with orchestration tools such as Airflow, Dagster, or Prefect Demonstrated experience…

Technical Tools
airflowawsbigquerydbtgcphubspotkafkapythonsalesforcesnowflakesqletllinux

Leap is one of the fastest-growing benefits solutions and a category-defining pioneer in employer specialty pharmacy. We are reshaping how life-changing therapies are delivered and financed, ensuring patients get the treatment they need while employers finally get a fair deal.

About the Role

~1 min read
  • Build and own data pipelines and ETL processes for claims ingestion, drug pricing, and CRM sync using BigQuery and Python

  • Design production pipelines for batch and streaming workloads, with a particular focus on high-volume claims data and new large-scale data sources on the roadmap

  • Architect warehouse schemas and transformations with clear separation between raw, staging, and modeled layers

  • Maintain data quality and reliability across systems that feed both human users and AI workloads, including row-count checks, schema drift detection, anomaly alerting, and silent upstream change detection

  • Design pipelines to be idempotent and replayable, with raw data always preserved to enable reprocessing when logic changes

  • Track data lineage across the full lifecycle — origin, transformation, and downstream dependencies

  • Validate data at every stage before it reaches a dashboard or AI system

  • Build reporting systems that give sales, clinical, and leadership teams live visibility into business performance

  • Create automated alerting that surfaces meaningful changes in data so the team acts on insights rather than requesting them

  • Build PHI-safe pipelines that support LLM workloads, agent systems, and automation

  • Design a unified data architecture that connects claims, drug pricing, patient records, CRM activity, and clinical workflows into a coherent whole

  • Own ingestion of external data from non-standard formats and sources across a diverse and growing provider base

Requirements

~1 min read
  • 5+ years of experience with Python, SQL, and dbt, with hands-on expertise in BigQuery, Snowflake, or a comparable cloud data warehouse and proficiency with orchestration tools such as Airflow, Dagster, or Prefect

  • Demonstrated experience architecting data platforms, including decisions around batch vs. streaming, incremental vs. full-refresh, and warehouse structure

  • Proven ability to build monitoring, lineage tracking, and governance systems that trace data from source to report

  • Experience using AI tools in day-to-day work and building data infrastructure that AI systems can rely on in production

  • Background as an early employee or founding data engineer responsible for building a data stack from the ground up

Nice to Have

~1 min read
  • Healthcare or HIPAA experience; familiarity with ingestion tools such as Fivetran; CRM integrations (Salesforce, HubSpot); or prior experience building data infrastructure for LLM or AI workloads

  • Experience with streaming frameworks such as Kafka, Pub/Sub, or Flink, or designing systems that handle both batch and real-time data flows

  • Comfort with cloud infrastructure (GCP, AWS) and Linux/sysadmin fundamentals, including VM debugging, log management, and service administration

  • A bias toward simple, cost-effective solutions — defaulting to open-source and applying sound judgment about when managed services justify their cost and lock-in

At Leap, we’re building an outlier company with real impact — and that takes focus, energy, and commitment. If that excites you, we’d love to hear from you.

Leap is an equal opportunity employer and welcomes applicants from all backgrounds. We’re committed to building a team that reflects a diversity of perspectives, experiences, and identities.

Location & Eligibility

Where is the job
United States
Remote within one country
Who can apply
US

Listing Details

Posted
March 20, 2026
First seen
May 7, 2026
Last seen
May 7, 2026

Posting Health

Days active
0
Repost count
0
Trust Level
23%
Scored at
May 7, 2026

Signal breakdown

freshnesssource trustcontent trustemployer trust
Newsletter

Stay ahead of the market

Get the latest job openings, salary trends, and hiring insights delivered to your inbox every week.

A
B
C
D
Join 12,000+ marketers

No spam. Unsubscribe at any time.

leapSenior Data Engineer