Senior ML Ops Engineer (Machine Learning Infrastructure)
Parallel Systems.com
150k - 240k USD/year
Office
Los Angeles, CA
Full Time
Parallel Systems is pioneering autonomous battery-electric rail vehicles designed to transform freight transportation by shifting portions of the $900 billion U.S. trucking industry onto rail. Our innovative technology offers cleaner, safer, and more efficient logistics solutions. Join our dynamic team and help shape a smarter, greener future for global freight.
Senior ML Ops Engineer (Machine Learning Infrastructure)
Parallel Systems is seeking an experienced MLOps/ML Infrastructure Engineer to lead the design and development of the scalable systems that power our autonomy and perception pipelines. As we build the first fully autonomous, battery-electric rail vehicles, you will play a critical role in enabling the ML teams to develop, train, and deploy models efficiently and reliably in both R&D and real-world environments.
This is an opportunity to take full ownership of the ML infrastructure stack, from distributed training environments and experiment tracking to deployment and monitoring at scale. You’ll collaborate closely with world-class engineers in autonomy, robotics, and software, helping shape the core systems that make real-time, safety-critical ML possible. If you're driven by building robust platforms that unlock innovation in AI and robotics, we’d love to work with you.
Responsibilities:
- Design and implement robust MLOps solutions, including automated pipelines for data management, model training, deployment and monitoring.
- Architect, deploy, and manage scalable ML infrastructure for distributed training and inference.
- Collaborate with ML engineers to gather requirements and develop strategies for data management, model development and deployment.
- Build and operate cloud-based systems (e.g., AWS, GCP) optimized for ML workloads in R&D, and production environments.
- Build scalable ML infrastructure to support continuous integration/deployment, experiment management, and governance of models and datasets.
- Support the automation of model evaluation, selection, and deployment workflows.
What Success Looks Like:
- After 30 Days: You have developed a deep understanding of the product goals, existing infrastructure, and stakeholder requirements. You've conducted technical discovery and proposed a preliminary MLOps architecture—evaluating various ML tools, cloud services, and workflow strategies—clearly outlining pros and cons for each option.
- After 60 Days: You’ve delivered a detailed design document that outlines the end-to-end ML pipeline, including data ingestion, model training, deployment, and monitoring. Based on feedback from ML engineers and stakeholders, you’ve iterated on the design and built PoC for the core ML workflow aligned with the approved architecture.
- After 90 Days: You have delivered the core features of the MLOps pipeline and successfully integrated key tools (e.g., MLflow, SageMaker, or Kubeflow). You’ve also initiated the implementation of the remaining features, ensuring the infrastructure supports scalable, repeatable workflows for model experimentation and deployment in both R&D and production environments.
Basic Requirements:
- Bachelor’s or higher degree in Computer Science, Machine Learning, or a relevant engineering discipline.
- 5+ years of experience building large-scale, reliable systems; 2+ years focused on ML infrastructure or MLOps.
- Proven experience architecting and deploying production-grade ML pipelines and platforms.
- Strong knowledge of ML lifecycle: data ingestion, model training, evaluation, packaging, and deployment.
- Hands-on experience with MLOps tools (e.g., MLflow, Kubeflow, SageMaker, Airflow, Metaflow, or similar).
- Deep understanding of CI/CD practices applied to ML workflows.
- Proficiency in Python, Git, and system design with solid software engineering fundamentals.
- Experience with cloud platforms (AWS, GCP, or Azure) and designing ML architectures in those environments.
Preferred Qualifications:
- Experience with deep learning architectures (CNNs, RNNs, Transformers) or computer vision.
- Hands-on experience with distributed training tools (e.g., PyTorch DDP, Horovod, Ray).
- Background in real-time ML systems and batch inference, including CPU/GPU-aware orchestration.
- Previous work in autonomous vehicles, robotics, or other real-time ML-driven systems.
We are committed to providing fair and transparent compensation in accordance with applicable laws. Salary ranges are listed below and reflect the expected range for new hires in this role, based on factors such as skills, experience, qualifications, and location. Final compensation may vary and will be determined during the interview process. The target hiring range for this position is listed below.
Target Salary Range:$150,000—$240,000 USDParallel Systems is an equal opportunity employer committed to diversity in the workplace. All qualified applicants will receive consideration for employment without regard to any discriminatory factor protected by applicable federal, state or local laws. We work to build an inclusive environment in which all people can come to do their best work.
Parallel Systems is committed to the full inclusion of all qualified individuals. As part of this commitment, Parallel Systems will ensure that persons with disabilities are provided reasonable accommodations. If reasonable accommodation is needed to participate in the job application or interview process, to perform essential job functions, and/or to receive other benefits and privileges of employment, please contact your recruiter.
Senior ML Ops Engineer (Machine Learning Infrastructure)
Office
Los Angeles, CA
Full Time
150k - 240k USD/year
October 7, 2025