Sr. Data Engineer
Ford Motor Company.com
Office
Dearborn, MI, United States
Full Time
The Global Data Insight & Analytics organization is looking for a Senior Data Engineer with deep expertise in Google Cloud Platform (GCP) and a proven track record of designing and implementing complex ETL processes to build robust, scalable, and efficient data pipelines . In this role, you will be part of a dynamic, cross-functional team, collaborating closely and consistently with other engineers, business partners, product managers, and designers. You will be involved in frequent and iterative releases. Your focus will be on shaping our data architecture, enabling advanced analytics, and supporting the integration of AI/ML and LLM capabilities.
Key Responsibilities:
- Design, develop, and maintain highly scalable and reliable ETL pipelines to ingest, transform, and load large datasets from various sources into our data ecosystem.
- Build and optimize data models in BigQuery for analytical and operational use cases, ensuring data quality, consistency, and accessibility.
- Leverage GCP services, including Dataproc, for distributed data processing using PySpark.
- Collaborate with data scientists, analysts, and product teams to understand data requirements and translate them into technical solutions.
- Implement data governance, security, and compliance best practices within the data platform.
- Monitor, troubleshoot, and optimize data pipeline performance and reliability.
- Mentor junior engineers and contribute to the continuous improvement of our data engineering practices and standards.
- Stay abreast of emerging technologies and industry trends, particularly in big data, cloud computing, AI/ML, and LLMs, and recommend their adoption where appropriate.
Required Qualifications:
- Bachelor's degree in Computer Science, Engineering, Business Administration or a related quantitative field.
- 5+ years of experience in data engineering, with a strong focus on building and managing large-scale data platforms.
- 3+ Years of experience with Google Cloud Platform (GCP) data services (e.g., BigQuery, Cloud Storage, Dataflow, Pub/Sub, Cloud Functions).
- Demonstrable experience in designing, developing, and optimizing complex ETL processes.
- Extensive hands-on experience with BigQuery for data warehousing, modeling, and query optimization.
- Strong programming skills in Python, with significant experience in PySpark for data manipulation and processing on distributed frameworks like Dataproc.
- Solid understanding of data warehousing concepts, dimensional modeling, and SQL.
- Experience with version control systems (e.g., Git) and CI/CD practices.
- Experience working in a product-driven organization, contributing to data solutions that directly support product development and user needs.
Preferred Qualifications & Experience:
- Hands-on experience with AI/ML workflows, including data preparation for model training, feature engineering, and deploying ML pipelines.
- Familiarity with Large Language Models (LLM) usage, including data considerations for fine-tuning, prompt engineering, or integrating LLM APIs into data applications.
- Experience with other GCP services like Cloud Composer (Apache Airflow), Data Catalog, or Looker.
- Knowledge of streaming data technologies (e.g., Kafka, Pub/Sub).
- Experience with Infrastructure as Code (e.g., Terraform).
Sr. Data Engineer
Office
Dearborn, MI, United States
Full Time
September 24, 2025