Software Engineer, Inference Deployment - Seattle
2 weeks ago

Job description
, consectetur adipiscing elit. Nullam tempor vestibulum ex, eget consequat quam pellentesque vel. Etiam congue sed elit nec elementum. Morbi diam metus, rutrum id eleifend ac, porta in lectus. Sed scelerisque a augue et ornare.
Donec lacinia nisi nec odio ultricies imperdiet.
Morbi a dolor dignissim, tristique enim et, semper lacus. Morbi laoreet sollicitudin justo eget eleifend. Donec felis augue, accumsan in dapibus a, mattis sed ligula.
Vestibulum at aliquet erat. Curabitur rhoncus urna vitae quam suscipit
, at pulvinar turpis lacinia. Mauris magna sem, dignissim finibus fermentum ac, placerat at ex. Pellentesque aliquet, lorem pulvinar mollis ornare, orci turpis fermentum urna, non ullamcorper ligula enim a ante. Duis dolor est, consectetur ut sapien lacinia, tempor condimentum purus.
Access all high-level positions and get the job of your dreams.
Similar jobs
We're looking for a systems-minded AI Software Engineer to join our core inference platform team. · Architect, extend, and optimize core components of our AI serving platform for throughput, latency, and scalability. · Customize open-source serving frameworks (e.g., vLLM) for pro ...
1 week ago
We are looking for a talented Machine Learning Engineer to play a key role in building our core AI inference platform.You will be responsible for designing and developing critical components, · including ML model deployment, · innovative model optimization pipelines, · and perfor ...
1 week ago
We're looking for a systems-minded AI Software Engineer to join our core inference platform team. · Architect, extend, and optimize core components of our AI serving platform for throughput, latency, and scalability. · Customize open-source serving frameworks (e.g., vLLM) for pro ...
1 week ago
We are looking for a talented Machine Learning Engineer to play a key role in building our core AI inference platform. · ...
1 week ago
We are looking for an AI Inference Engineer with a solid background in speech recognition and model inference. This role will include collaboration with cross-functional teams to deliver high-impact projects from the ground up.About The Team Zoom's AI Speech Team is developing sp ...
2 weeks ago
We are looking for an AI Inference Engineer with a solid background in speech recognition and model inference to develop state-of-the-art automatic speech recognition system and ship it to various Zoom products. · ...
1 month ago
FieldAI is transforming how robots interact with the real world. Our growing ML team in Seattle builds risk-aware, reliable, field-ready AI systems that tackle the hardest problems in robotics and unlock the potential of embodied intelligence. We take a pragmatic approach that go ...
19 hours ago
We are now looking for a Senior System Software Engineer to work on user facing tools for Dynamo Inference Server NVIDIA is hiring software engineers for its GPU-accelerated deep learning software team, and we are a remote friendly work environment. · We are a fast-paced team bui ...
3 weeks ago
We are now looking for a Senior System Software Engineer to work on user facing tools for Dynamo Inference Server NVIDIA is hiring software engineers for its GPU-accelerated deep learning software team and we are a remote friendly work environment. · ...
1 month ago
As a Technical Program Manager for Inference you'll be the critical bridge between our inference systems and the broader organization driving strategic initiatives around inference infrastructure adoption capacity management and deployment processes while ensuring seamless coordi ...
1 month ago
We are seeking a visionary and hands-on Hardware Design Engineer to contribute to the design, definition, · and implementation of our core AI inference engine. · ...
1 week ago
Senior GenAI Algorithms Engineer — Model Optimizations for Inference
Only for registered members
NVIDIA is at the forefront of the generative AI revolution The Algorithmic Model Optimization Team specifically focuses on optimizing generative AI models such as large language models (LLM) and diffusion models for maximal inference efficiency using techniques ranging from quant ...
1 month ago
The Inference Enablement and Acceleration team at Amazon Web Services (AWS) builds AWS Neuron, the software development kit used to accelerate deep learning and GenAI workloads on Amazon's custom machine learning accelerators. · ...
1 month ago
The Annapurna Labs team at Amazon Web Services (AWS) builds AWS Neuron, the software development kit used to accelerate deep learning and GenAI workloads on Amazon's custom machine learning accelerators, Inferentia and Trainium. · The Inference Enablement and Acceleration team fo ...
1 month ago
We combine deep hardware knowledge with ML expertise to push the boundaries of what's possible in AI acceleration.The team works closely with customers on their model enablement, providing direct support and optimization expertise to ensure their machine learning workloads achiev ...
1 month ago
The Annapurna Labs team at Amazon Web Services (AWS) builds AWS Neuron, the software development kit used to accelerate deep learning and GenAI workloads on Amazon's custom machine learning accelerators. · This comprehensive toolkit includes an ML compiler, runtime, and applicati ...
1 month ago
We are looking for a talented Machine Learning Research Scientist to play a key role in designing our core AI inference platform.You will be actively conducting high impact research on novel ML solutions towards the next generation of efficient intelligence platforms. · ...
1 week ago
The Annapurna Labs team at Amazon Web Services (AWS) builds AWS Neuron software development kit used to accelerate deep learning. · Design develop and optimize machine learning models frameworks for deployment on custom ML hardware accelerators. · ...
2 weeks ago
The Cloud Inference team scales and optimizes Claude to serve the massive audiences of developers and enterprise companies across AWS, GCP, Azure, and future cloud service providers (CSPs). We own the end-to-end product of Claude on each cloud platform—from API integration and in ...
3 weeks ago
We're looking for a Lead Engineer Inference Platform to join our team building inference platform for embedding models that power semantic search retrieval and AI native features across MongoDB Atlas. · This role is part of broader Search AI Platform team involves close collabora ...
4 weeks ago