Location Hybrid Remote in Atlanta (3 days onsite / week).
Note : Initial 100% onsite required for the first six months.
Employment Type : Permanent / Direct Hire / Full-time
Salary Up to $160,000 (depending on experience) + bonus
The Role :
We're seeking a highly skilled and hands-on Data Architect to lead the design, implementation, and ongoing evolution of our enterprise-grade data systems. This role is crucial for building scalable, secure, and intelligent data infrastructure that supports core analytics, operational excellence, and future AI initiatives. Success requires a seasoned technologist who can seamlessly integrate cloud-native services with traditional data warehousing to create a modern, unified data platform.
What You'll Do :
- Architecture & Strategy : Lead the design and implementation of modern data platforms, including Data Lakes, Data Warehouses, and Lakehouse architectures, to enable a single source of truth for the enterprise.
- Data Modeling & Integration : Architect unified data models that support both modular monoliths and microservices-based platforms. Design and optimize high-volume, low-latency streaming / batch ETL / ELT pipelines.
- Technical Leadership : Drive the technical execution across the entire data lifecycle. Build and optimize core data processing scripts using Spark and Python.
- Governance & Quality : Define and enforce standards for data governance, metadata management, and data observability across distributed systems. Implement automated data lineage tracking, schema evolution, and data quality monitoring.
- Cloud Infrastructure : Configure and manage cloud-native data services, including core data storage and event ingestion infrastructure.
Required Experience :
Experience : 10+ years of proven experience in enterprise data architecture and engineering.Core Platform Expertise : Strong, hands-on experience with the Azure Data Ecosystem including Azure Data Lake Storage (ADLS), Azure Synapse Analytics (or equivalent cloud DW), and Azure Purview (or equivalent data catalog).Processing : Deep expertise in Databricks (or Apache Spark) for ETL / ELT pipeline implementation, using Delta Lake and SQL Server (or equivalent RDBMS).Coding & Scripting : Strong proficiency in Python, Spark, and advanced SQL.Data Governance : Hands-on experience implementing data lineage tracking and data quality monitoring (e.g., using Great Expectations or dbt).Preferred Skills :
Semantic Technologies : Hands-on experience developing ontology frameworks using OWL, RDF, and SPARQL to enable semantic interoperability.Advanced AI Data : Experience integrating structured / unstructured data into Knowledge Graphs and Vector Databases.Streaming / Telemetry : Experience developing and maintaining semantic telemetry pipelines using services like Azure Event Hubs or Kafka.Emerging Concepts : Exposure to linked data ecosystems, data mesh, or data fabric concepts.