ob Title : Data Engineer (Databricks + Azure)
Client : One of our Consulting Clients (Global Analytics & Digital Transformation Firm)
Location : Columbus, OH (Remote / Hybrid)
Duration : Full-Time
About the Role
We are seeking a highly skilled Data Engineer with deep expertise in Databricks and Azure Cloud to join a decision analytics and data engineering team within one of our global consulting clients. The role involves building, optimizing, and maintaining large-scale data pipelines that fuel enterprise analytics, reporting, and AI-driven insights-primarily supporting clients in the insurance and financial services domains .
Key Responsibilities
Data Pipeline Development & Optimization
- Design, build, and enhance ETL / ELT data pipelines using Azure Data Factory , Databricks (PySpark, SQL, Python) , and related services.
- Develop and manage Delta Live Tables , Autoloader , and Unity Catalog within the Databricks ecosystem for structured, incremental data processing.
- Implement data ingestion, transformation, and validation frameworks that ensure high performance, scalability, and reliability.
- Monitor data pipelines, troubleshoot issues, and ensure optimal system performance and SLA adherence.
Data Modeling & Architecture
Collaborate with business analysts and reporting teams to define logical and physical data models supporting analytical and operational needs.Implement data warehousing and lakehouse solutions using Azure Data Lake and Delta Lake .Optimize data structures for query performance, cost efficiency, and reusability.Data Quality, Governance & Automation
Design and implement robust data quality checks and validation mechanisms to maintain integrity across sources and transformations.Automate repetitive processes using scripts, parameterized pipelines, and reusable frameworks.Conduct periodic audits and compliance checks aligned with governance policies.Contribute to metadata management, documentation, and lineage tracking.Required Skills & Experience
7 12 years of experience in Data Engineering with proven expertise in Databricks and Azure Cloud ecosystems.Strong hands-on experience in PySpark , Python , and SQL for data transformation, validation, and performance tuning.Solid understanding of Delta Lake architecture , ETL / ELT frameworks , and data warehousing principles.Proficiency with Azure services including Data Factory (ADF) , Data Lake (ADLS) , and Databricks Notebooks .Experience with Delta Live Tables , Unity Catalog , and Autoloader for batch and streaming data processing.Strong background in data modeling , performance optimization , and automation scripting .Familiarity with Agile methodologies and DevOps-based deployment practices (Git, CI / CD preferred).Strong analytical, communication, and problem-solving skills to collaborate effectively across diverse teams.Preferred : Exposure to insurance, healthcare, or financial services data ecosystems.Nice to Have
Experience in data migration projects (on-prem to cloud or multi-cloud).Familiarity with Delta Sharing , Databricks SQL Warehouses , or MLflow for advanced use cases.Experience with data cataloging, lineage, or quality frameworks such as Purview, Collibra, or Great Expectations.Exposure to BI / reporting tools like Power BI or Tableau for end-to-end integration understanding.