Demo

Member of Technical Staff- Inference

Acceler8 Talent
Palo Alto, CA Full Time
POSTED ON 4/5/2025
AVAILABLE BEFORE 5/5/2025

Member of Technical Staff, Research Engineer (Inference) - Palo Alto, CA

Join a team at the forefront of AI innovation, where your expertise in model inference can make a tangible impact. This role is ideal for engineers who thrive in a focused, high-tech environment, solving complex challenges related to large-scale AI deployments. As a Member of Technical Staff, Research Engineer (Inference), you'll play a pivotal role in optimizing and deploying state-of-the-art models for real-world applications.

About the Company

This AI studio, recognized for its groundbreaking work in developing and deploying highly effective language models, is now focused on scaling its technology for enterprise use cases. With a strong foundation in model alignment and fine-tuning, the team is well-funded and equipped with cutting-edge resources, offering a unique environment for those passionate about pushing AI boundaries. Their culture is centered on collaboration, technical excellence, and a pragmatic approach to AI advancements.

About the Role

As a Member of Technical Staff, Research Engineer (Inference), you’ll be involved in optimizing AI models for enterprise deployment, ensuring they perform efficiently under varying conditions. Your work will focus on reducing latency, improving throughput, and maintaining model performance during inference. Engineers in this role should have a deep understanding of the trade-offs in model inference, including balancing hardware constraints with real-time processing demands.

What We Can Offer You :

  • Competitive compensation aligned with your experience and contributions.
  • Unlimited paid time off and flexible parental leave.
  • Comprehensive medical, dental, and vision coverage.
  • Visa sponsorship for qualified hires.
  • Professional growth opportunities through coaching, conferences, and training.

Key Responsibilities :

  • Optimize and deploy large language models (LLMs) for inference across cloud and on-prem environments.
  • Utilize frameworks like ONNX, TensorRT, and TVM to accelerate model performance.
  • Troubleshoot complex issues related to model scaling and performance.
  • Collaborate with cross-functional teams to refine and deploy inference pipelines using PyTorch, Docker, and Kubernetes.
  • Balance competing demands, such as model accuracy and inference speed, in enterprise settings.
  • If you have experience with LLM inference, model optimization tools, and infrastructure management, this role aligns perfectly with your skills.

    If your compensation planning software is too rigid to deploy winning incentive strategies, it’s time to find an adaptable solution. Compensation Planning
    Enhance your organization's compensation strategy with salary data sets that HR and team managers can use to pay your staff right. Surveys & Data Sets

    What is the career path for a Member of Technical Staff- Inference?

    Sign up to receive alerts about other jobs on the Member of Technical Staff- Inference career path by checking the boxes next to the positions that interest you.
    Income Estimation: 
    $36,436 - $44,219
    Income Estimation: 
    $50,145 - $86,059
    Income Estimation: 
    $48,515 - $60,705
    Income Estimation: 
    $103,114 - $138,258
    Income Estimation: 
    $118,163 - $145,996
    Income Estimation: 
    $120,777 - $151,022
    Income Estimation: 
    $129,363 - $167,316
    Income Estimation: 
    $86,891 - $130,303
    Income Estimation: 
    $85,996 - $102,718
    Income Estimation: 
    $111,859 - $131,446
    Income Estimation: 
    $110,457 - $133,106
    Income Estimation: 
    $105,809 - $128,724
    Income Estimation: 
    $122,763 - $145,698
    Income Estimation: 
    $105,809 - $128,724
    Income Estimation: 
    $136,611 - $163,397
    Income Estimation: 
    $135,163 - $163,519
    Income Estimation: 
    $131,953 - $159,624
    Income Estimation: 
    $150,859 - $181,127
    Income Estimation: 
    $129,363 - $167,316
    Income Estimation: 
    $145,845 - $177,256
    Income Estimation: 
    $147,836 - $182,130
    Income Estimation: 
    $154,597 - $194,610
    Income Estimation: 
    $86,891 - $130,303
    View Core, Job Family, and Industry Job Skills and Competency Data for more than 15,000 Job Titles Skills Library

    Job openings at Acceler8 Talent

    Acceler8 Talent
    Hired Organization Address New York, NY Full Time
    Senior Growth Hacker | Hybrid (NYC or Pittsburgh) ???? Are you a data-driven, creative, and technical marketer with a pa...
    Acceler8 Talent
    Hired Organization Address Santa Clara, CA Full Time
    Staff Full-Stack Engineer (AI Product, LLM Series A Startup) We are looking for a Staff Full-Stack Engineer to join an A...
    Acceler8 Talent
    Hired Organization Address Hayward, CA Full Time
    Join Us as a Research Scientist in AI : Shape the Future of Collaborative AI! About Us Our mission is to strengthen the ...
    Acceler8 Talent
    Hired Organization Address Menlo, CA Full Time
    Research Engineer (GPU Kernels) Are you an experienced professional in high-performance computing? If so, we have an exc...

    Not the job you're looking for? Here are some other Member of Technical Staff- Inference jobs in the Palo Alto, CA area that may be a better fit.

    Junior Product Manager

    Inference, Palo Alto, CA

    AI Assistant is available now!

    Feel free to start your new journey!