Job Summary
A company is looking for a Senior Data Engineer, Data Lakehouse Infrastructure.
Key Responsibilities
- Architect and scale a high-performance data lakehouse on GCP using various technologies
- Design, build, and optimize distributed query engines for complex analytical workloads
- Implement metadata management and develop robust ETL/ELT pipelines
Required Qualifications
- 5+ years of experience in data or software engineering focused on distributed data systems
- Proven experience building and scaling data platforms on GCP
- Strong command of query engines such as Trino, Presto, Spark, or Snowflake
- Experience with modern table formats like Apache Hudi, Iceberg, or Delta Lake
- Exceptional programming skills in Python and proficiency in SQL or SparkSQL
Comments