As a Data Engineer I, you'll play a hands-on role in building and supporting scalable data pipelines within our cloud-based Lakehouse environment. You'll help bring in healthcare data from various sources, ensuring it's accurate, reliable, and ready to support analytics and reporting needs. You'll partner closely with the Analytics team to deliver data in a clear and actionable way.
Key Highlights
Key Responsibilities
Technical Skills Required
Benefits & Perks
Job Description
Vālenz® Health is the platform to simplify healthcare – the destination for employers, payers, providers and members to reduce costs, improve quality, and elevate the healthcare experience. The Valenz mindset and culture of innovation combine to create a distinctly different approach to an inefficient, uninspired health system. With fully integrated solutions, Valenz engages early and often to execute across the entire patient journey – from care navigation and management to payment integrity, plan performance and provider verification. With a 99% client retention rate, we elevate expectations to a new level of efficiency, effectiveness and transparency where smarter, better, faster healthcare is possible.
About This Opportunity:
As a Data Engineer I, you’ll play a hands-on role in building and supporting scalable data pipelines within our cloud-based Lakehouse environment (Azure Databricks, Delta Lake), leveraging tools like Spark and PySpark. You’ll help bring in healthcare data from a variety of sources, ensuring it’s accurate, reliable, and ready to support analytics and reporting needs across the organization. You’ll also partner closely with the broader Analytics team to make sure data is delivered in a way that’s clear and actionable. Over time, you’ll build expertise in managing large, complex datasets and contribute to evolving our data architecture to support new and emerging data sources as the business grows.
Things You’ll Do Here:
- Create and maintain processes to acquire, validate, and enrich data from various sources
- Support the migration of on-premise data systems (SQL Server) to a cloud-based lakehouse architecture (Azure Databricks, Delta Lake), including data transformation and pipeline re-architecture
- Develop and optimize ETL/ELT pipelines using PySpark and Spark SQL
- Implement Lakehouse + Delta architecture best practices to ensure a standardized and scalable way that we store and process our data, including schema enforcement, ACID transactions, and data versioning
- Orchestrate data pipelines using Databricks Workflows (Jobs) or similar tools
- Implement data quality frameworks, validation checks, and monitoring for pipeline reliability
- Optimize performance and cost of data pipelines
- Collaborate on CI/CD practices for data pipelines, including testing, deployment, and versioning
- Partner with data analysts, data scientists, and business stakeholders to identify new sources of data and estimate feasibility of acquiring specific data sources
- Design and implement data models to support analytics, reporting, and data warehousing use cases
- Take an active role in agile processes
- Perform other duties as assigned
What You’ll Bring to the Team:
- 1+ years of work experience in a data engineering role
- Bachelor’s degree or greater in a quantitative field such as statistics, mathematics, engineering, computer science, finance, or economics or equivalent practical experience
- Hands-on experience with Databricks (Spark, PySpark, Delta Lake) and/or migrating RDBMS systems to a data lakehouse
- Experience working the most common types of healthcare data (medical claims, eligibility, provider network rosters, Rx claims, etc) from a variety of sources
- Strong organizational skills and time management capacity to balance multiple projects with limited supervision
- Ability to build (and re-evaluate) a process from the ground up
- Strong investigative skills with ability to search beyond the initial results
- High attention to detail, with overwhelming desire to test and double-check your own results
- Comfortable working with messy data and ambiguous results
- Hands-on experience with SQL and Python (including PySpark) for distributed data processing
- Experience building and optimizing large-scale distributed data pipelines for both batch and streaming ingestion
- Our data platform is undergoing a transformation from a traditional on-premise architecture to a modern cloud-based lakehouse on Azure. Technologies used:
- Cloud & Modern Data Platform:
- Azure (Blob Storage / Data Lake Storage, Synapse Analytics)
- Databricks (Spark, PySpark, Delta Lake, Databricks Workflows)
- Delta Lake architecture (ACID transactions, schema enforcement, time travel)
- Data Engineering & Development:
- Python (including PySpark), SQL
- Data pipeline orchestration and workflow management
- Version control (Git, Azure DevOps)
- Legacy / Transitional Systems:
- SQL Server (on-premise RDBMS)
- .NET / C#-based data processing applications
- Migration from traditional ETL and relational systems to cloud-based lakehouse architecture
Interested in remote work opportunities in Data Science? Discover Data Science Remote Jobs featuring exclusive positions from top companies that offer flexible work arrangements.
Browse our curated collection of remote jobs across all categories and industries, featuring positions from top companies worldwide.
- Work Environment: You’ll need a quiet workspace that is free from distractions
- Technology: Reliable internet connection—if you can use streaming services, you’re good to go!
- Security: Adherence to company security protocols, including the use of VPNs, secure passwords, and company-approved devices/software
- Location: You must be US based, in a location where you can work effectively and comply with company policies such as HIPAA
Valenz is proud to be recognized by Inc. 5000 as one of America’s fastest-growing private companies. Our team is committed to delivering on our promise to engage early and often for smarter, better, faster healthcare. With this commitment, you’ll find an engaged culture – one that stands strong, vigorous, and healthy in all we do.
Benefits
- Generously subsidized company-sponsored Medical, Dental, and Vision insurance, with access to services through our own products, Healthcare Blue Book and KISx Card
- Spending account options: HSA, FSA, and DCFSA
- 401K with company match and immediate vesting
- Flexible working environment
- Generous Paid Time Off to include vacation, sick leave, and paid holidays
- Employee Assistance Program that includes professional counseling, referrals, and additional services
- Paid maternity and paternity leave
- Pet insurance
- Employee discounts on phone plans, car rentals and computers
- Community giveback opportunities, including paid time off for philanthropic endeavors
Powered by JazzHR
EZsH07uDNv
Similar Jobs
Explore other opportunities that match your interests
AP Professionals
Senior Data Engineer - Property & Casualty Insurance
prismhr