Job Summary
A company is looking for a Senior Data Engineer, Data Lakehouse Infrastructure.
Key Responsibilities
- Architect and scale a high-performance data lakehouse on GCP
- Design, build, and optimize distributed query engines for complex analytical workloads
- Develop and orchestrate robust ETL/ELT pipelines using Apache Airflow and GCP-native tools
Required Qualifications
- 5+ years of experience in data or software engineering, focusing on distributed data systems and cloud-native architectures
- Proven experience building and scaling data platforms on GCP
- Strong command of one or more query engines such as Trino, Presto, or Spark
- Experience with modern table formats like Apache Hudi or Iceberg
- Exceptional programming skills in Python and SQL or SparkSQL
Comments