
Senior / Staff Backend Engineer
Hamming, Austin, TX, United States
Location: Remote (North America) or Austin, TX
Employment Type: Full-time (no contractors)
Department: Engineering
Why now
Hamming automates QA for voice AI agents. Everyone is building voice agents. We secure them. In fact, we invented this category. With one click,
thousands of our agents call our customers’ agents
across accents, background noise, and personalities—then we generate
crisp bug reports
and production-grade analytics. Reliability is the moat in voice AI, and that’s our whole job.
We are one of the fastest engineering teams in the world. We prod deploy 4x / day.
I’m looking for someone who can
own reliability and scale
across our LLM-enabled platform, shipping precise, outcome-driven improvements to high-availability systems.
—
Sumanyu (CEO)
Previously: grew Citizen 4× and scaled an AI sales program to $100Ms/yr at Tesla.
What you’ll do
Own core services
in
TypeScript/Node.js
and
Python
that orchestrate
LiveKit ,
Temporal , STT/TTS, and LLM tooling for real-time voice agents.
Scale 1 → N → 100× : take what works today and harden it for 10K parallel calls with
99.99%
uptime. Turn human playbooks into productized systems.
Harden pipelines
for ingestion, evaluation, and analytics so telephony events, recordings, and outcomes propagate reliably across services.
Level-up observability : deepen
OpenTelemetry/SigNoz
and trace-first practices to shrink mean-time-to-truth in prod.
Prototype → test → prod : partner with product to ship new LLM-driven behaviors with clear success metrics, guardrails, and regressions blocked in CI.
Infrastructure readiness : CI/CD, environment automation, incident response playbooks—customer conversations stay online.
You might be a fit if you
Have
senior/staff
experience running distributed backends with
real-time/streaming
constraints.
Are fluent in
TypeScript/Node.js
and comfortable jumping into
Python
for ML/audio jobs.
Know
Temporal
(or similar workflow engines), queues, Redis, and
PostgreSQL .
Have
shipped production LLM apps
and understand prompt/tool design, evals, and guardrail instrumentation.
Operate cloud-native on
AWS
with
Terraform ; k8s doesn’t scare you.
Are a
power user of Cursor/Zed/Devin
and were using code-gen before it was cool.
Have intuition for what current-gen LLMs can/can’t do—and what tomorrow’s models will unlock.
Think independently,
grind with customers , and do whatever it takes—without dropping the quality bar.
Bonus: built 0→1
real-time systems
in Telecom/Networking, Autonomous Vehicles, or HFT; founded something; built
AI voice
apps.
Interesting problems you’ll touch
Voice simulations that feel real : accents, overlapping speech, crosstalk, background noise, barge-ins.
Massive concurrency :
10,000+ parallel calls
with deterministic behavior and graceful degradation.
Temporal-driven orchestration
for long-running, interruptible call flows.
Closed-loop reliability : turn prod failures into auto-generated tests and blocked deploys.
Trace-everything
culture: make “what happened?” a 30-second question, not a war room.
How we work
Outcomes over output : we adjust roadmaps when new data lands.
Demo early and document decisions so context moves fast.
Own incidents : lead the investigation, write crisp notes, land durable fixes.
Direct, candid, respectful
communication keeps remote teammates in lockstep with Austin HQ.
Our stack
App : Next.js, TypeScript, Tailwind
AI : OpenAI, Anthropic, STT/TTS providers
Realtime/Orchestration : LiveKit, Pipecat/Daily, Temporal
Infra/DB : AWS, k8s, PostgreSQL, Redis, Terraform
Observability : OpenTelemetry, SigNoz
Apply
If you want to make
AI voice agents reliable at scale , let’s talk.
Send a short note (links to work > resumes) and tell us about something
reliability-critical
you shipped: what broke, what you fixed, and how you knew it worked.
#J-18808-Ljbffr
Employment Type: Full-time (no contractors)
Department: Engineering
Why now
Hamming automates QA for voice AI agents. Everyone is building voice agents. We secure them. In fact, we invented this category. With one click,
thousands of our agents call our customers’ agents
across accents, background noise, and personalities—then we generate
crisp bug reports
and production-grade analytics. Reliability is the moat in voice AI, and that’s our whole job.
We are one of the fastest engineering teams in the world. We prod deploy 4x / day.
I’m looking for someone who can
own reliability and scale
across our LLM-enabled platform, shipping precise, outcome-driven improvements to high-availability systems.
—
Sumanyu (CEO)
Previously: grew Citizen 4× and scaled an AI sales program to $100Ms/yr at Tesla.
What you’ll do
Own core services
in
TypeScript/Node.js
and
Python
that orchestrate
LiveKit ,
Temporal , STT/TTS, and LLM tooling for real-time voice agents.
Scale 1 → N → 100× : take what works today and harden it for 10K parallel calls with
99.99%
uptime. Turn human playbooks into productized systems.
Harden pipelines
for ingestion, evaluation, and analytics so telephony events, recordings, and outcomes propagate reliably across services.
Level-up observability : deepen
OpenTelemetry/SigNoz
and trace-first practices to shrink mean-time-to-truth in prod.
Prototype → test → prod : partner with product to ship new LLM-driven behaviors with clear success metrics, guardrails, and regressions blocked in CI.
Infrastructure readiness : CI/CD, environment automation, incident response playbooks—customer conversations stay online.
You might be a fit if you
Have
senior/staff
experience running distributed backends with
real-time/streaming
constraints.
Are fluent in
TypeScript/Node.js
and comfortable jumping into
Python
for ML/audio jobs.
Know
Temporal
(or similar workflow engines), queues, Redis, and
PostgreSQL .
Have
shipped production LLM apps
and understand prompt/tool design, evals, and guardrail instrumentation.
Operate cloud-native on
AWS
with
Terraform ; k8s doesn’t scare you.
Are a
power user of Cursor/Zed/Devin
and were using code-gen before it was cool.
Have intuition for what current-gen LLMs can/can’t do—and what tomorrow’s models will unlock.
Think independently,
grind with customers , and do whatever it takes—without dropping the quality bar.
Bonus: built 0→1
real-time systems
in Telecom/Networking, Autonomous Vehicles, or HFT; founded something; built
AI voice
apps.
Interesting problems you’ll touch
Voice simulations that feel real : accents, overlapping speech, crosstalk, background noise, barge-ins.
Massive concurrency :
10,000+ parallel calls
with deterministic behavior and graceful degradation.
Temporal-driven orchestration
for long-running, interruptible call flows.
Closed-loop reliability : turn prod failures into auto-generated tests and blocked deploys.
Trace-everything
culture: make “what happened?” a 30-second question, not a war room.
How we work
Outcomes over output : we adjust roadmaps when new data lands.
Demo early and document decisions so context moves fast.
Own incidents : lead the investigation, write crisp notes, land durable fixes.
Direct, candid, respectful
communication keeps remote teammates in lockstep with Austin HQ.
Our stack
App : Next.js, TypeScript, Tailwind
AI : OpenAI, Anthropic, STT/TTS providers
Realtime/Orchestration : LiveKit, Pipecat/Daily, Temporal
Infra/DB : AWS, k8s, PostgreSQL, Redis, Terraform
Observability : OpenTelemetry, SigNoz
Apply
If you want to make
AI voice agents reliable at scale , let’s talk.
Send a short note (links to work > resumes) and tell us about something
reliability-critical
you shipped: what broke, what you fixed, and how you knew it worked.
#J-18808-Ljbffr