Must Have Skills
Hands-on experience with writing complex SQL queries using :
Joins, Self Joins, Views, Materialized Views, Cursors, Recursive queries
GROUP BY, PARTITION BY functions
SQL performance tuning
Hands-on experience with ETL and Dimensional Data Modelling , including :
Slowly Changing Dimensions (SCD) – Type 1, Type 2, Type 3
Strong understanding of schema types , table types (fact / dimension), and data warehouse design considerations
Proficiency in Python scripting and programming using :
Pandas , PyParsing , Airflow
Modules such as Tableau Server modules , NumPy , Datetime , Apache Airflow-related modules , APIs
Setting up Python scripts on DataLab , scheduling processes , and connecting with DataLake / S3
Data pipeline automation and strong hands-on Python programming skills
Experience with Apache Kafka and Python using client libraries such as :
Confluent’s librdkafka or kafka-python
Building streaming applications , data pipelines , and microservices
Strong understanding of Snowflake Architecture with experience in :
Designing and building data solutions
Architecture and design aspects , performance tuning , time travel , warehouse scaling , clustering , micro-partitioning
Working with SnowSQL and Snowpipe
Good to Have
Experience with Snowflake performance optimization techniques
Experience with Vertica and SingleStore
Lead Experience
Experience interacting directly with business stakeholders and independently leading data projects
Collaborating with offshore teams and owning overall project delivery
Actively participating in discussions with business users to understand requirements
Performing impact analysis and providing suitable data solutions
Senior Data Engineer • Sunnyvale, CA, United States