[Skip To Content]

Senior ML Infrastructure Engineer, Inference Platform

  • 위치
    • Austin, Texas
    • Mountain View, California
    • Sunnyvale, California
    • Warren, Michigan
  • 직무 유형 Full time
  • 게시됨
  • Job Requisition JR-202604629

설명

About the Team:

The ML Inference Platform is part of the AV ML Infrastructure organization. Our team owns the cloud-agnostic, reliable, and cost-efficient platform that powers GM’s AI efforts. We’re proud to serve teams developing autonomous vehicles (L3/L4/L5), as well as other groups building AI-driven products for GM and its customers. We enable rapid innovation and feature development by optimizing for high-priority, ML-centric use cases. Our platform supports the serving of state-of-the-art (SOTA) machine learning models for experimental, online and bulk inference, with a focus on performance, availability, concurrency, and scalability. We’re committed to maximizing GPU utilization across platforms (B200, H100, A100, and more) while maintaining reliability and cost efficiency.

About the Role:

We are seeking a Senior ML Infrastructure engineer to help build and scale robust platforms for ML Inference workflows. In this role, you’ll work closely with ML engineers and researchers to ensure efficient model serving and inference in production, for workflows such as data mining, labeling, model distillation, evaluations, simulations and more. This is a high-impact opportunity to influence the future of AI infrastructure at GM. You will play a key role in shaping the architecture, roadmap and user-experience of a robust ML inference service supporting real-time, batch, and experimental inference needs. The ideal candidate brings experience in designing distributed systems for ML, strong problem-solving skills, and a product mindset focused on platform usability and reliability.

What you’ll be doing:

  • Design and implement core platform backend software components.

  • Collaborate with ML engineers and researchers to understand critical workflows, parse them to platform requirements, and deliver incremental value.

  • Lead technical decision-making on model serving strategies, orchestration, caching, model versioning, and auto-scaling mechanisms for highly optimized use of accelerators.

  • Drive the development of monitoring, observability, and metrics to ensure reliability, performance, and resource optimization of inference services.

  • Proactively research and integrate state-of-the-art model serving frameworks, hardware accelerators, and distributed computing techniques.

  • Lead technical initiatives across GM’s ML ecosystem.

  • Raise the engineering bar through technical leadership, establishing best practices.

  • Contribute to open source projects; represent GM in relevant communities.

Minimum Requirements

  • 5+ years of industry experience, with focus on machine learning systems or high performance backend services.

  • Expertise in either Python, C++ or other relevant coding languages.

  • Expertise in ML inference, model serving frameworks (triton, rayserve, vLLM etc).

  • Strong communication skills and a proven ability to drive cross-functional initiatives.

  • Ability to thrive in a dynamic, multi-tasking environment with ever-evolving priorities.

Preferred Qualifications

  • Deep expertise building zero-to-one ML infrastructure platforms.

  • Experience working with or designing interfaces, apis and clients for ML workflows.

  • Experience with Ray framework, and/or vLLM.

  • Experience with distributed systems, and handling large-scale data processing.

  • Familiarity with telemetry, and other feedback loops to inform product improvements.

  • Familiarity with hardware acceleration (GPUs) and optimizations for inference workloads.

Compensation:  The compensation information is a good faith estimate only. It is based on what a successful applicant might be paid in accordance with applicable state laws. The compensation may not be representative for positions located outside of New York, Colorado, California, or Washington.   

  • The salary range for this role is $155,420 to $205,900. The actual base salary a successful candidate will be offered within this range will vary based on factors relevant to the position.  

  • Bonus Potential: An incentive pay program offers payouts based on company performance, job level, and individual performance. 

  • Benefits: GM offers a variety of health and wellbeing benefit programs. Benefit options include medical, dental, vision, Health Savings Account, Flexible Spending Accounts, retirement savings plan, sickness and accident benefits, life insurance, paid vacation & holidays, tuition assistance programs, employee assistance program, GM vehicle discounts and more.  

Benefits:   

  • Benefits: GM offers a variety of health and wellbeing benefit programs. Benefit options include medical, dental, vision, Health Savings Account, Flexible Spending Accounts, retirement savings plan, sickness and accident benefits, life insurance, paid vacation & holidays, tuition assistance programs, employee assistance program, GM vehicle discounts and more.  

Relocation:  This job may be eligible for relocation benefits.  

Remote/Hybrid: This role is based remotely but if you live within a 50-mile radius of Mountain View, you are expected to report to that location three times a week, at minimum.  

다양성 정보

General Motors는 법적으로 금지된 차별을 배제하는 것은 물론 포용성과 소속감을 진정으로 장려하는 직장이 되기 위해 노력하고 있습니다. 당사는 다양성이 보장되는 환경에서 직원들이 역량을 발휘하고 우리 고객을 위한 더 좋은 제품을 개발할 수 있다고 믿습니다. 따라서 입사에 관심 있는 사람이 있다면 포지션별 주요 업무와 자격을 확인하고 본인이 보유한 기술과 능력에 부합하는 모든 포지션에 적극적으로 지원하기를 장려합니다. 지원자는 채용 과정에서 역할 관련 평가(해당하는 경우) 및/또는 채용 전 스크리닝을 통과해야 합니다.  자세한 정보는 GM 채용 과정 안내를 참고하십시오.

공평한 취업 기회 선언 (미국)

General Motors는 공평한 기회를 제공하는 고용주임을 자부합니다.  자격을 만족하는 지원자는 인종과 피부색, 성별, 성적 지향, 성별 정체성, 국적, 장애, 재향 군인 보호법 적용 여부와 상관없이 채용 후보로서 심사를 받습니다. 

숙소 (미국 및 캐나다)

General Motors는 장애인을 포함한 모든 구직자들에게 취업 기회를 제공합니다. 구직이나 취업 지원에 도움이 되는 합리적인 숙소가 필요한 경우 [email protected]으로 이메일을 보내시거나 800-865-7580으로 전화주십시오. 이메일에, 귀하가 요청하는 특정한 숙소에 대한 설명과 귀하가 지원하는 직무와 채용 요청서 번호를 포함해주세요.