Senior ML Inference Engineer

Only for registered members United States

1 week ago

Default job background

Job summary

We are looking for a Senior Inference System Engineer with 5-10 years of experience to join our team.


Lorem ipsum dolor sit amet
, consectetur adipiscing elit. Nullam tempor vestibulum ex, eget consequat quam pellentesque vel. Etiam congue sed elit nec elementum. Morbi diam metus, rutrum id eleifend ac, porta in lectus. Sed scelerisque a augue et ornare.

Donec lacinia nisi nec odio ultricies imperdiet.
Morbi a dolor dignissim, tristique enim et, semper lacus. Morbi laoreet sollicitudin justo eget eleifend. Donec felis augue, accumsan in dapibus a, mattis sed ligula.

Vestibulum at aliquet erat. Curabitur rhoncus urna vitae quam suscipit
, at pulvinar turpis lacinia. Mauris magna sem, dignissim finibus fermentum ac, placerat at ex. Pellentesque aliquet, lorem pulvinar mollis ornare, orci turpis fermentum urna, non ullamcorper ligula enim a ante. Duis dolor est, consectetur ut sapien lacinia, tempor condimentum purus.
Get full access

Access all high-level positions and get the job of your dreams.



Similar jobs

  • Only for registered members *HQ - San Francisco, CA

    We're hiring an Inference Engineer to advance our mission of building real-time multimodal intelligence. · ...

  • Only for registered members San Francisco Bay Area Remote job

    We're looking for engineers who can bridge the gap between ML research and high-performance inference. · JAX / Equinox / Pallas stack · Rust systems programming with a focus on developer experience · ...

  • Only for registered members San Francisco

    Mirai builds on-device inference layer for AI. · We enable model makers to run AI models directly on edge devices,Our stack spans low-level GPU kernels to high-level model conversion tools. · We're looking for engineers who bridge ML research and high-performance inference,You'll ...

  • Only for registered members San Francisco, CA

    About Us Most AI is frozen in place - it doesn't adapt to the world. We think that's backwards. · ...

  • Only for registered members Palo Alto Full time

    We're seeking an experienced LLM Inference Engineer to optimize our large language model serving infrastructure.The ideal candidate has extensive hands-on experience with state-of-the-art inference optimization techniques and a track record of deploying efficient scalable LLM sys ...

  • Only for registered members North America

    BentoML is a leading inference platform provider that helps AI teams run large language models and other generative AI workloads at scale. · You will improve the speed and efficiency of large language models at the GPU kernel level through the inference engine and across distribu ...

  • Only for registered members San Jose, CA

    We are looking for an AI Inference Engineer with a solid background in speech recognition and model inference. · In this role, you will develop state-of-the-art automatic speech recognition system and ship it to various Zoom products. You will work on the most cutting edge speech ...

  • Only for registered members Cupertino, CA

    About Etched: building AI chips that are hard-coded for individual model architectures. · Contribute to the architecture and design of the Sohu host software stack · ...

  • Only for registered members San Jose Full time $151,800 - $332,200 (USD)

    +Job summary · We are looking for an AI Inference Engineer with a solid background in speech recognition and model inference.Developing state-of-the-art speech services for Zoom products. · Optimizing ASR inference systems for production deployment. · +We are developing speech re ...

  • Only for registered members San Francisco, CA

    We're looking for a software engineer to help us serve OpenAI's multimodal models at scale. · This work is inherently cross-functional: you'll collaborate directly with researchers training these models and with product teams defining new modalities of interaction. · ...

  • Only for registered members San Francisco $200,000 - $350,000 (USD)

    We are looking for an AI Inference engineer to join our growing team. · ...

  • Only for registered members Seattle, WA

    We are looking for a software engineer to join our team responsible for developing and optimizing LLM inference framework. You will be responsible for developing and optimizing LLL inference framework, GPU and CUDA performance optimization to create an industry-leading high-perfo ...

  • Only for registered members San Francisco

    We are looking for an Inference Engineering Manager to lead our AI Inference team. This is a unique opportunity to build and scale the infrastructure that powers Perplexity's products and APIs serving millions of users with state-of-the-art AI capabilities. · ...

  • Only for registered members San Francisco, CA

    We are looking for an AI Inference engineer to join our growing team. · Our current stack is Python, Rust, C++, PyTorch, Triton, CUDA, Kubernetes.Responsibilities · Develop APIs for AI inference that will be used by both internal and external customers · Benchmark and address bot ...

  • Only for registered members San Francisco $300,000 - $385,000 (USD)

    We are looking for an Inference Engineering Manager to lead our AI Inference team. · This is a unique opportunity to build and scale the infrastructure that powers Perplexity's products and APIs, · serving millions of users with state-of-the-art AI capabilities. · ...

  • Only for registered members San Francisco Full time

    We are looking for an Inference Engineering Manager to lead our AI Inference team. This is a unique opportunity to build and scale the infrastructure that powers Perplexity's products and APIs, serving millions of users with state-of-the-art AI capabilities. · ...

  • Only for registered members San Francisco Bay Area

    TheAI Inference Engineer at Quadric will port AI models to Quadric platform; optimize model deployment for efficient inference; profile and benchmark model performance. · Bachelor's or Master's in Computer Science and/or Electric Engineering. · 5+ years of experience in AI/LLM mo ...

  • Only for registered members San Francisco, CA

    We are looking for an Inference Engineering Manager to lead our AI Inference team. This is a unique opportunity to build and scale the infrastructure that powers Perplexity's products and APIs, · serving millions of users with state-of-the-art AI capabilities.Lead and grow a high ...

  • Only for registered members Palo Alto

    We're seeking an experienced LLM Inference Engineer to optimize our large language model (LLM) serving infrastructure. · The ideal candidate has: · Extensive hands-on experience with state-of-the-art inference optimization techniques · A track record of deploying efficient, scala ...

  • Only for registered members Seattle $151,800 - $332,200 (USD)

    We are looking for an AI Inference Engineer with a solid background in speech recognition and model inference. This role will include collaboration with cross-functional teams to deliver high-impact projects from the ground up.About The Team Zoom's AI Speech Team is developing sp ...

  • Only for registered members Seattle $151,800 - $332,200 (USD)

    We are looking for an AI Inference Engineer with a solid background in speech recognition and model inference to develop state-of-the-art automatic speech recognition system and ship it to various Zoom products. · ...