Deepgram Logo

Deepgram

Research Staff, LLMs

Reposted 11 Hours Ago
In-Office or Remote
Hiring Remotely in USA
150K-250K Annually
Mid level
In-Office or Remote
Hiring Remotely in USA
150K-250K Annually
Mid level
The role involves researching and developing large language models (LLMs) with a focus on transformer architecture, data curation, distributed training, and optimization. Responsibilities include conducting experiments, collaborating with teams, and staying updated on deep learning advancements.
The summary above was generated by AI
Company Overview

Deepgram is the leading platform underpinning the emerging trillion-dollar Voice AI economy, providing real-time APIs for speech-to-text (STT), text-to-speech (TTS), and building production-grade voice agents at scale. More than 200,000 developers and 1,300+ organizations build voice offerings that are ‘Powered by Deepgram’, including Twilio, Cloudflare, Sierra, Decagon, Vapi, Daily, Cresta, Granola, and Jack in the Box. Deepgram’s voice-native foundation models are accessed through cloud APIs or as self-hosted and on-premises software, with unmatched accuracy, low latency, and cost efficiency. Backed by a recent Series C led by leading global investors and strategic partners, Deepgram has processed over 50,000 years of audio and transcribed more than 1 trillion words. There is no organization in the world that understands voice better than Deepgram.

Company Operating Rhythm

At Deepgram, we expect an AI-first mindset—AI use and comfort aren’t optional, they’re core to how we operate, innovate, and measure performance.

Every team member who works at Deepgram is expected to actively use and experiment with advanced AI tools, and even build your own into your everyday work. We measure how effectively AI is applied to deliver results, and consistent, creative use of the latest AI capabilities is key to success here. Candidates should be comfortable adopting new models and modes quickly, integrating AI into their workflows, and continuously pushing the boundaries of what these technologies can do.

Additionally, we move at the pace of AI. Change is rapid, and you can expect your day-to-day work to evolve just as quickly. This may not be the right role if you’re not excited to experiment, adapt, think on your feet, and learn constantly, or if you’re seeking something highly prescriptive with a traditional 9-to-5.

The Opportunity

Voice is the most natural modality for human interaction with machines. However, current sequence modeling paradigms based on jointly scaling model and data cannot deliver voice AI capable of universal human interaction. The challenges are rooted in fundamental data problems posed by audio: real-world audio data is scarce and enormously diverse, spanning a vast space of voices, speaking styles, and acoustic conditions. Even if billions of hours of audio were accessible, its inherent high dimensionality creates computational and storage costs that make training and deployment prohibitively expensive at world scale. We believe that entirely new paradigms for audio AI are needed to overcome these challenges and make voice interaction accessible to everyone.


The Role

Deepgram is currently looking for an experienced researcher to who has worked extensively with Large Language Models (LLMS) and has a deep understanding of transformer architecture to join our Research Staff. As a Member of the Research Staff, this individual should have extensive experience working on the hard technical aspects of LLMs, such as data curation, distributed large-scale training, optimization of transformer architecture, and Reinforcement Learning (RL) training.

The Challenge

We are seeking researchers who:

  • See "unsolved" problems as opportunities to pioneer entirely new approaches

  • Can identify the one critical experiment that will validate or kill an idea in days, not months

  • Have the vision to scale successful proofs-of-concept 100x

  • Are obsessed with using AI to automate and amplify your own impact

If you find yourself energized rather than daunted by these expectations—if you're already thinking about five ideas to try while reading this—you might be the researcher we need. This role demands obsession with the problems, creativity in approach, and relentless drive toward elegant, scalable solutions. The technical challenges are immense, but the potential impact is transformative.

What You'll Do
  • Brainstorming and collaborating with other members of the Research Staff to define new LLM research initiatives

  • Broad surveying of literature, evaluating, classifying, and distilling current methods

  • Designing and carrying out experimental programs for LLMs

  • Driving transformer (LLM) training jobs successfully on distributed compute infrastructure and deploying new models into production

  • Documenting and presenting results and complex technical concepts clearly for a target audience

  • Staying up to date with the latest advances in deep learning and LLMs, with a particular eye towards their implications and applications within our products

You'll Love This Role if You
  • Are passionate about AI and excited about working on state of the art LLM research

  • Have an interest in producing and applying new science to help us develop and deploy large language models

  • Enjoy building from the ground up and love to create new systems.

  • Have strong communication skills and are able to translate complex concepts clearly

  • Are highly analytical and enjoy delving into detailed analyses when necessary


It's Important to Us That You Have
  • 3+ years of experience in applied deep learning research, with a solid understanding toward the applications and implications of different neural network types, architectures, and loss mechanism

  • Proven experience working with large language models (LLMs) - including experience with data curation, distributed large-scale training, optimization of transformer architecture, and RL Learning

  • Strong experience coding in Python and working with Pytorch

  • Experience with various transformer architectures (auto-regressive, sequence-to-sequence.etc)

  • Experience with distributed computing and large-scale data processing

  • Prior experience in conducting experimental programs and using results to optimize models

It Would Be Great if You Had
  • Deep understanding of transformers, causal LMs, and their underlying architecture

  • Understanding of distributed training and distributed inference schemes for LLMs

  • Familiarity with RLHF labeling and training pipelines

  • Up-to-date knowledge of recent LLM techniques and developments

The Challenge

We are seeking researchers who:

  • See "unsolved" problems as opportunities to pioneer entirely new approaches

  • Can identify the one critical experiment that will validate or kill an idea in days, not months

  • Have the vision to scale successful proofs-of-concept 100x

  • Are obsessed with using AI to automate and amplify your own impact

If you find yourself energized rather than daunted by these expectations—if you're already thinking about five ideas to try while reading this—you might be the researcher we need. This role demands obsession with the problems, creativity in approach, and relentless drive toward elegant, scalable solutions. The technical challenges are immense, but the potential impact is transformative.

What You'll Do
  • Brainstorming and collaborating with other members of the Research Staff to define new LLM research initiatives

  • Broad surveying of literature, evaluating, classifying, and distilling current methods

  • Designing and carrying out experimental programs for LLMs

  • Driving transformer (LLM) training jobs successfully on distributed compute infrastructure and deploying new models into production

  • Documenting and presenting results and complex technical concepts clearly for a target audience

  • Staying up to date with the latest advances in deep learning and LLMs, with a particular eye towards their implications and applications within our products

You'll Love This Role if You
  • Are passionate about AI and excited about working on state of the art LLM research

  • Have an interest in producing and applying new science to help us develop and deploy large language models

  • Enjoy building from the ground up and love to create new systems.

  • Have strong communication skills and are able to translate complex concepts clearly

  • Are highly analytical and enjoy delving into detailed analyses when necessary


It's Important to Us That You Have
  • 3+ years of experience in applied deep learning research, with a solid understanding toward the applications and implications of different neural network types, architectures, and loss mechanism

  • Proven experience working with large language models (LLMs) - including experience with data curation, distributed large-scale training, optimization of transformer architecture, and RL Learning

  • Strong experience coding in Python and working with Pytorch

  • Experience with various transformer architectures (auto-regressive, sequence-to-sequence.etc)

  • Experience with distributed computing and large-scale data processing

  • Prior experience in conducting experimental programs and using results to optimize models

It Would Be Great if You Had
  • Deep understanding of transformers, causal LMs, and their underlying architecture

  • Understanding of distributed training and distributed inference schemes for LLMs

  • Familiarity with RLHF labeling and training pipelines

  • Up-to-date knowledge of recent LLM techniques and developments

  • Published papers in Deep Learning Research, particularly related to LLMs and deep neural networks

  • Published papers in Deep Learning Research, particularly related to LLMs and deep neural networks

Benefits & Perks*

Holistic health
  • Medical, dental, vision benefits

  • Annual wellness stipend

  • Mental health support

  • Life, STD, LTD Income Insurance Plans

Work/life blend
  • Unlimited PTO

  • Generous paid parental leave

  • Flexible schedule

  • 12 Paid US company holidays

  • Quarterly personal productivity stipend

  • One-time stipend for home office upgrades

  • 401(k) plan with company match

  • Tax Savings Programs

Continuous learning
  • Learning / Education stipend

  • Participation in talks and conferences

  • Employee Resource Groups

  • AI enablement workshops / sessions

*For candidates outside of the US, we use an Employer of Record model in many countries, which means benefits are administered locally and governed by country-specific regulations. Because of this, benefits will differ by region — in some cases international employees receive benefits US employees do not, and vice versa. As we scale, we will continue to evaluate where we can create more alignment, but a 1:1 global benefits structure is not always legally or operationally possible.

Backed by prominent investors including Y Combinator, Madrona, Tiger Global, Wing VC and NVIDIA, Deepgram has raised over $215M in total funding. If you're looking to work on cutting-edge technology and make a significant impact in the AI industry, we'd love to hear from you!

Deepgram is an equal opportunity employer. We want all voices and perspectives represented in our workforce. We are a curious bunch focused on collaboration and doing the right thing. We put our customers first, grow together and move quickly. We do not discriminate on the basis of race, religion, color, national origin, gender, sexual orientation, gender identity or expression, age, marital status, veteran status, disability status, pregnancy, parental status, genetic information, political affiliation, or any other status protected by the laws or regulations in the locations where we operate.

We are happy to provide accommodations for applicants who need them.

Top Skills

Distributed Computing
Large Language Models
Python
PyTorch
Transformer Architectures

Similar Jobs at Deepgram

11 Hours Ago
Remote
USA
180K-230K Annually
Mid level
180K-230K Annually
Mid level
Artificial Intelligence • Machine Learning • Natural Language Processing • Software • Conversational AI
As Model Evaluation QA Lead, you'll ensure model quality assurance across Deepgram's AI pipeline, designing automated evaluation frameworks, integrating quality gates into release pipelines, and ensuring high standards for model deployments.
Top Skills: ArgoGithub ActionsJenkinsMlflowNumpyPandasPython
11 Hours Ago
Remote
USA
150K-195K Annually
Senior level
150K-195K Annually
Senior level
Artificial Intelligence • Machine Learning • Natural Language Processing • Software • Conversational AI
As a Customer Success Engineer, you'll drive adoption and satisfaction for enterprise customers using Deepgram's voice AI technology, fostering strong relationships and aligning technical solutions with business objectives.
Top Skills: AIAPIsMachine LearningSpeech-To-TextText-To-Speech
11 Hours Ago
Remote
USA
120K-130K Annually
Mid level
120K-130K Annually
Mid level
Artificial Intelligence • Machine Learning • Natural Language Processing • Software • Conversational AI
The Account Executive at Deepgram is responsible for building a sales pipeline, exceeding targets, collaborating with teams, and understanding customer needs to sell the voice AI platform.
Top Skills: APIsSelf-Managed SoftwareVoice Ai

What you need to know about the Los Angeles Tech Scene

Los Angeles is a global leader in entertainment, so it’s no surprise that many of the biggest players in streaming, digital media and game development call the city home. But the city boasts plenty of non-entertainment innovation as well, with tech companies spanning verticals like AI, fintech, e-commerce and biotech. With major universities like Caltech, UCLA, USC and the nearby UC Irvine, the city has a steady supply of top-flight tech and engineering talent — not counting the graduates flocking to Los Angeles from across the world to enjoy its beaches, culture and year-round temperate climate.

Key Facts About Los Angeles Tech

  • Number of Tech Workers: 375,800; 5.5% of overall workforce (2024 CompTIA survey)
  • Major Tech Employers: Snap, Netflix, SpaceX, Disney, Google
  • Key Industries: Artificial intelligence, adtech, media, software, game development
  • Funding Landscape: $11.6 billion in venture capital funding in 2024 (Pitchbook)
  • Notable Investors: Strong Ventures, Fifth Wall, Upfront Ventures, Mucker Capital, Kittyhawk Ventures
  • Research Centers and Universities: California Institute of Technology, UCLA, University of Southern California, UC Irvine, Pepperdine, California Institute for Immunology and Immunotherapy, Center for Quantum Science and Engineering

Sign up now Access later

Create Free Account

Please log in or sign up to report this job.

Create Free Account