Tesla is seeking a skilled Data Engineer to architect and implement a cutting-edge data platform, focusing on data pipelines, warehousing, and reporting for its energy business, including Industrial, Residential, Supercharger, and Solar products.
Requirements
Strong data engineering background with expertise in at least two programming languages (Python, Scala, Java, or Rust)
Expert-level experience with Apache Spark (PySpark, Spark SQL, Spark Streaming)
Experience with modern data lake technologies (Delta Lake, Apache Iceberg, Apache Hudi)
Experience with containerization and orchestration (Docker, Kubernetes)
Proficiency in SQL and familiarity with database internals and performance concepts
Demonstrated experience working with terabyte-scale data processing
Responsibilities
Design and implement robust, scalable solutions for processing terabyte-scale data
Develop and maintain high-performance data pipelines using Apache Spark, Kafka, and other modern data technologies
Build and optimize batch and streaming data processing systems
Develop aggregate and summary tables for various engineering teams that span across multiple product lines and geographies
Design and develop new systems and tools to enable cross-functional teams to consume and understand data faster
Implement and maintain CI/CD pipelines for data applications
Research and incorporate emerging data infrastructures, tools, and technologies