Job Summary
A company is looking for a Senior Data Engineer to lead the design, development, and optimization of scalable data pipelines.
Key Responsibilities
- Build and optimize distributed data processing jobs using Apache Spark on Databricks
- Design and automate ETL pipelines and integrate data from diverse sources
- Develop dimensional models and ensure efficient querying and transformation using SQL and PySpark
Required Qualifications
- Bachelor's or Master's degree in Computer Science, Engineering, or related field
- 5+ years of experience in data engineering with at least 2 years on Databricks
- Proficiency in Python, Scala, SQL, and Spark
- Hands-on experience with Azure Data Services
- Strong understanding of ETL, data warehousing, and data modeling concepts
Comments