Founding ML inference Engineer - Palo Alto

Only for registered members Palo Alto, United States

1 month ago

Default job background

Job summary

We're a stealth-mode VC-backed AI startup developing the world's fastest most efficient AI models redefining how large language models LLMs are trained deployed and used.

Qualifications

  • BSMSPhD in Computer Science Engineering or equivalent experience.

Lorem ipsum dolor sit amet
, consectetur adipiscing elit. Nullam tempor vestibulum ex, eget consequat quam pellentesque vel. Etiam congue sed elit nec elementum. Morbi diam metus, rutrum id eleifend ac, porta in lectus. Sed scelerisque a augue et ornare.

Donec lacinia nisi nec odio ultricies imperdiet.
Morbi a dolor dignissim, tristique enim et, semper lacus. Morbi laoreet sollicitudin justo eget eleifend. Donec felis augue, accumsan in dapibus a, mattis sed ligula.

Vestibulum at aliquet erat. Curabitur rhoncus urna vitae quam suscipit
, at pulvinar turpis lacinia. Mauris magna sem, dignissim finibus fermentum ac, placerat at ex. Pellentesque aliquet, lorem pulvinar mollis ornare, orci turpis fermentum urna, non ullamcorper ligula enim a ante. Duis dolor est, consectetur ut sapien lacinia, tempor condimentum purus.
Get full access

Access all high-level positions and get the job of your dreams.



Similar jobs

  • Only for registered members Palo Alto Full time

    We're seeking an experienced LLM Inference Engineer to optimize our large language model serving infrastructure.The ideal candidate has extensive hands-on experience with state-of-the-art inference optimization techniques and a track record of deploying efficient scalable LLM sys ...

  • Only for registered members Palo Alto, CA

    We're seeking an experienced LLM Inference Engineer to optimize our large language model (LLM) serving infrastructure. · Design and implement multi-node serving architectures for distributed LLM inference ...

  • Only for registered members Palo Alto, CA

    We're seeking an experienced LLM Inference Engineer to optimize our large language model (LLM) serving infrastructure.The ideal candidate has: · Extensive hands-on experience with state-of-the-art inference optimization techniques · A track record of deploying efficient, scalable ...

  • Only for registered members San Francisco Bay Area Remote job

    We're looking for engineers who can bridge the gap between ML research and high-performance inference. · JAX / Equinox / Pallas stack · Rust systems programming with a focus on developer experience · ...

  • Only for registered members Palo Alto

    We're seeking an experienced LLM Inference Engineer to optimize our large language model (LLM) serving infrastructure. · The ideal candidate has: · Extensive hands-on experience with state-of-the-art inference optimization techniques · A track record of deploying efficient, scala ...

  • Only for registered members Palo Alto $137,000 - $270,000 (USD)

    We're looking for a Lead Engineer, Inference Platform to join our team building the inference platform for embedding models that power semantic search retrieval and AI-native features across MongoDB Atlas. · This role is part of the broader Search and AI Platform team and involve ...

  • Only for registered members San Francisco

    Mirai builds on-device inference layer for AI. · We enable model makers to run AI models directly on edge devices,Our stack spans low-level GPU kernels to high-level model conversion tools. · We're looking for engineers who bridge ML research and high-performance inference,You'll ...

  • Only for registered members Palo Alto, CA

    What To Expect · Our team productionizes and experiments with novel ML models for our custom HW- we train and deploy large neural networks for efficient inference on compute-constrained edge devices (CPU / GPU / AI ASIC). · ...

  • Only for registered members Palo Alto, CA

    We're looking for a Senior Engineer to help build the next-generation inference platform that supports embedding models used for semantic search retrieval and AI-native experiences in MongoDB Atlas. · We'll join the broader Search and AI Platform organization collaborate with ML ...

  • Only for registered members Palo Alto Full time

    We are looking for an RL Systems Engineer to build large-scale reinforcement learning systems for post-training, fine-tuning and alignment. · Build large-scale reinforcement learning systems for post-training, fine-tuning and alignmentScales RL workloads across multi-node cluster ...

  • Only for registered members Palo Alto, CA

    We train and deploy large neural networks for efficient inference on compute-constrained edge devices. · Build robust AI frameworks to lower neural networks to edge devices · Create robust AI infrastructure to train and fine-tune networks for Autopilot and Optimus · ...

  • Only for registered members Palo Alto $132,000 - $390,000 (USD)

    +Job summary · Build robust AI frameworks to lower neural networks to edge devices · +ResponsibilitiesBuild robust AI infrastructure to train and fine-tune networks for Autopilot and Optimus on large GPU clusters ...

  • Only for registered members San Francisco, CA

    About Us Most AI is frozen in place - it doesn't adapt to the world. We think that's backwards. · ...

  • Only for registered members Los Altos, CA

    We are looking for a Platform Engineer to own and improve the systems that support our Cloud Inference products. · 5+ years of relevant experience · Experience maintaining Kubernetes clusters and applications · ...

  • Only for registered members Palo Alto $118,000 - $390,000 (USD)

    We train and deploy large neural networks for efficient inference on compute-constrained edge devices (CPU / GPU / AI ASIC). · ...

  • Only for registered members Cupertino, CA

    About Etched: building AI chips that are hard-coded for individual model architectures. · Contribute to the architecture and design of the Sohu host software stack · ...

  • Only for registered members Los Altos $198,000 - $286,000 (USD)

    We're on a mission to revolutionize AI infrastructure by systematically rebuilding the AI software stack from the ground up.We are looking for a Platform Engineer to own and improve the systems that support our Cloud Inference products. · The estimated base salary range for this ...

  • Only for registered members Mountain View

    We are seeking a dynamic and experienced Product manager to lead our Inference Engine product strategy and execution for GMI Cloud.This role will be instrumental in shaping the future of our cloud offerings with a strong focus on the rapidly evolving AI infrastructure and Maas la ...

  • Only for registered members Mountain View, CA

    We are seeking a dynamic and experienced Product manager to lead our Inference Engine product strategy and execution for GMI Cloud. · Bonus: Experience with Model as a service product · Bonus: Experience with serverless GPU product · ...

  • Only for registered members Mountain View, CA

    We're looking for a BD manager to drive the identification, execution and commercialization of customer POC for our AI infrastructure platform. · Own and Drive Customer POCs Identify, qualify and source high-potential customer POCs aligned with strategic priorities. · ...