×
Register Here to Apply for Jobs or Post Jobs. X

Machine Learning Engineer - Audio Specialist

Job in Austin, Travis County, Texas, 78716, USA
Listing for: Breaker
Full Time position
Listed on 2025-11-22
Job specializations:
  • IT/Tech
    Robotics, Systems Engineer, AI Engineer
  • Engineering
    Robotics, Systems Engineer, AI Engineer
Salary/Wage Range or Industry Benchmark: 125000 - 150000 USD Yearly USD 125000.00 150000.00 YEAR
Job Description & How to Apply Below

Machine Learning Engineer - Audio Specialist

Join to apply for the Machine Learning Engineer - Audio Specialist role at Breaker

Direct message the job poster from Breaker

  • Build audio understanding models (STT, STS) from scratch for voice‑controlled robotic systems
  • Own the entire audio ML pipeline: data collection, training infrastructure, deployment, and optimization
  • Deep audio/signal processing expertise required – this isn't general ML
  • Travel for field testing in real‑world conditions
  • Join an exciting startup backed by globally recognised investors at the bleeding edge of physical AI

About Us

The way humans use robots is broken. Modern warfare demands more robots than we have operators. Every drone, ground vehicle, and maritime system requires dedicated training, manual control, and constant human oversight. One operator per robot. One pilot per mission. This operator bottleneck is the constraint on military capability today.

Breaker's AI agent, Avalon, breaks this constraint. Instead of piloting individual robots, operators command entire teams of autonomous systems across air, ground, and maritime domains, all through natural conversation.

A single operator can now coordinate multiple drones, ground vehicles, and other platforms simultaneously. Instead of flying search patterns on three different screens, you say “survey this area and flag anything unusual” – and a team of robots figures out how to divide the task, coordinate their movements, and report back what matters.

We're not just making robots easier to control. We're fundamentally changing the operator‑to‑robot ratio, turning small teams into force multipliers.

Our software deploys models directly onboard each robot, enabling real‑time, intent‑driven control even in contested environments with limited bandwidth. We're solving problems most AI companies never touch: sub‑second inference on edge hardware with strict latency, power, and connectivity constraints.

We're backed by some of the best global investors and are growing our team across Austin, Texas, and Sydney, Australia. We're a small team of experienced engineers, moving fast on technology that will define how humans and machines work together for decades to come.

Join us if you want to help create the robots we were promised 🤖

About the Role

Voice is the primary human machine interface our customers use with autonomous systems – and you'll own that capability from the ground up. You'll build speech recognition from scratch: building training pipelines, curating datasets, and shipping models that work reliably in real‑world conditions where radio communication, wind noise, and varying audio quality are the norm, not the exception.

This is a rare opportunity to build foundational IP rather than integrate off‑the‑shelf solutions. You'll establish the entire ML infrastructure for the speech stack, from data collection strategies to model deployment on edge hardware. The work spans the full spectrum: running field tests to capture training data, experimenting with state‑of‑the‑art architectures, and optimizing models to run efficiently on compute‑constrained robotic platforms.

The technical challenge is unusually deep for an ML role. You're not just training models – you're making architectural decisions about how audio signals are processed, understanding the tradeoffs between different model families for varying audio quality scenarios, and building the instrumentation to track performance improvements over time. This is the kind of work that could result in patents and define how voice‑controlled robotics systems perform for years to come.

Key Responsibilities

  • Evaluate and implement state‑of‑the‑art architectures, making informed decisions about model selection based on audio quality constraints and deployment requirements
  • Own metrics such as word error rate (WER), establishing baselines and demonstrating measurable improvements over time
  • Build and maintain infrastructure for model training, including experiment tracking, performance monitoring, and version control
  • Design data collection campaigns and field testing protocols to capture representative training data across varying environmental…
To View & Apply for jobs on this site that accept applications from your location or country, tap the button below to make a Search.
(If this job is in fact in your jurisdiction, then you may be using a Proxy or VPN to access this site, and to progress further, you should change your connectivity to another mobile device or PC).
 
 
 
Search for further Jobs Here:
(Try combinations for better Results! Or enter less keywords for broader Results)
Location
Increase/decrease your Search Radius (miles)

Job Posting Language
Employment Category
Education (minimum level)
Filters
Education Level
Experience Level (years)
Posted in last:
Salary