Data Architect
Chicago, IL – hybrid WFH : days onsite in the loop, days work from home
Summary
We're looking for a Data Architect with a background in Azure, Databricks, Snowflake and Python, which is our primary tech stack for designing, building, and optimizing data pipelines. This role is foundational for our organization to deliver best in class AI models, Machine Learning, Analytics and BI – where we have an industry leading real estate data platform, to power critical business decisions in our space.
What you’ll be doing
- Design, develop, and maintain robust data pipelines using Databricks to process and transform large-scale datasets.
- Write efficient, reusable, and scalable SQL and / or Python code for data ingestion, transformation, and integration.
- Optimize data workflows for performance, reliability, and cost-efficiency in cloud environments.
- Collaborate with data scientists, analysts, and stakeholders to understand data requirements and deliver solutions.
- Implement data governance, security, and compliance best practices within data pipelines.
- Monitor and troubleshoot data pipeline issues, ensuring high availability and data integrity.
- Leverage Databricks for advanced analytics, machine learning workflows, and real-time data processing.
- Integrate data from various sources (APIs, databases, blob, SFTP, streams) into Datalake for centralized storage and querying.
- Mentor junior data engineers and contribute to team knowledge sharing.
- Stay updated on emerging data technologies and recommend improvements to existing systems.
- Performance Tuning and Optimization of all Data Ingestion and Data Integration processes, including the Data Platform and databases
Skills we’re seeking
years of experience with Data Architecture / Data EngineeringMust have strong experience designing, building and optimizing data pipelinesyears of Data Architect experience specificallyMust have strong experience with AzureMust have strong experience with DatabricksMust have strong experience with SnowflakeMust have advanced proficiency with SQLMust have data modeling experience Understanding of data schema modeling (dimensions, measures, slowly changing dimensions)Strong Python experience, or other programming experience (Python highly preferred) for data processing, scripting, and automationBachelor's or Master's Degree in Computer Science or related IT or Data fieldExperience with Azure Durable FunctionsExperience with SparkExperience with DuckDBExperience with Parquet, Delta Lake, and / or Iceberg formatsExperience with ELT and / or Bronze, Silver, Gold transformation layersExperience with data governanceInmon vs Kimball (Star‑Schema vs NF)
Must have excellent communication skillsNice to haves (in this order)
Strong Python experience, or other programming experience (Python highly preferred) for data processing, scripting, and automationBachelor's or Master's Degree in Computer Science or related IT or Data fieldExperience with Azure Durable FunctionsExperience with SparkExperience with Azure Durable FunctionsExperience with DuckDBExperience with Parquet, Delta Lake, and / or Iceberg formatsExperience with ELT and / or Bronze, Silver, Gold transformation layersExperience with data governance#J-18808-Ljbffr