Try our Voice Characters to automate voice agent testing.

    Launch trustworthyAI voice agents in weeks

    Automated voice agent testing, call analytics, trust and safety reports and more.

    Test your AI voice agent against 1000s of simulated users in minutes.

    Backed by
    Y Combinator LogoAI Grant Logo
    Hamming AI (YC S24) - Automated testing for Voice Agents | Product HuntProduct Hunt Daily Badge

    All-in-one experimentation platform for AI voice agents

    AI voice agents are hard to get right. A small change in prompts, function call definitions or model providers can cause large changes in voice agent call quality.

    We're the only end-to-end platform that supports you from development to production.

    Automated Voice Agent Testing

    Teams currently spend hours testing their voice agents by hand. Use our voice characters (see demo) to create 1000s of concurrent phone calls to your voice agents and find bugs.

    Automated voice agent testing is 1000x more efficient than testing your voice agents by hand.

    Featured in our YC launch here.

    Prompt Management

    For B2B teams, each of your customers requires a slightly different prompt.

    You can store, manage, version and keep your prompts synced with voice infra providers from Hamming.

    Featured in our YC launch here.

    Prompt Optimizer & Playground

    Writing prompts by hand is slow and tedius. Use our prompt optimizer (free to try) to automatically generate optimized prompts for your LLM.

    Use our prompt playground to test LLM outputs on a dataset of inputs. Our LLM judges the quality of generated outputs. Save 80% of manual prompt engineering effort.

    Featured in our YC launch here.

    Call Analytics

    Go beyond passive monitoring. We actively track and score how users are using your AI app in production and flag cases using LLM judges that need your attention.

    Easily convert calls and traces into test cases and add them to your golden dataset.

    Featured in our YC launch here.

    Trusted by AI-forward enterprises

    Shelden Shi

    Shelden Shi

    Co-Founder & CTO @ Lilac Labs

    We automate the person taking orders at the drive-thru with voice. Order accuracy is extremely critical for our customers. Getting an order wrong (i.e., missing allergies) means significant financial loss.

    Hamming's automated voice agent testing helps us simulate 1000s of customer calls (dietary restrictions, allergies, large group orders, etc.) and find gaps in our system in minutes. This gives us a huge peace of mind and clarity on where we need to improve.

    Hamming is an essential part of our infra that directly unlocks revenue.

    Yossi Eliyahu
    Yossi Eliyahu
    VP of Engineering @ Fora
    There are a lot of low quality AI apps out there. We care a lot about quality. Hamming helps us launch accurate, robust and resilient AI apps that our users love.
    Chris Chen
    Chris Chen
    PM @ Fora
    Hamming allows me to test new changes to my AI pipeline 100x faster than vibe checking.
    Mark Wai
    Mark Wai
    Co-Founder & CTO @ Inkly
    At Inkly, we're building the modern legal experience for startups using GenAI. Being able to test our system against a dataset of test cases gives us a huge peace of mind and clarity on where we need to improve.
    Conner Swann
    Conner Swann
    Co-Founder @ Intuitive Systems
    The team is tackling a huge pain point for me - running evaluations continuously while I'm fine-tuning custom models.

    Automated voice agent testing and call analytics

    We're experts in supporting companies tackling high-stakes domains where making mistakes leads to high churn or regulatory consequences.

    Our customers use Hamming to automatically test and improve both inbound and outbound AI voice agents.

    Our agents call your agent

    Our voice agents are trained to mimic human conversations, including accents, background noise, and pauses.

    For each call, we automatically score the call quality and provide detailed analytics.

    Icon 01
    Icon 02
    Icon 02
    Icon 03
    Icon 04
    Icon 05
    Icon 06

    In any language

    Our AI voice agents can call your agents in any language, including English, French, German, Hindi, Spanish, Italian, and more.

    🇺🇸

    English

    🇫🇷

    French

    🇩🇪

    German

    🇮🇳

    Hindi

    🇪🇸

    Spanish

    🇮🇹

    Italian

    Built for teams building AI voice agents

    Trustworthy AI voice agents made simple

    Building reliable AI voice agents is a team effort. Hamming is built to support cross-team collaboration.

    ML Engineer
    Mark
    ML Engineer
    I can simulate thousands of voice conversations with different accents, background noise levels, and user scenarios in minutes. The detailed reports help me quickly identify where my voice agent needs improvement.
    Data Scientist
    Julia
    Data Scientist
    I can understand exactly why certain voice conversations score poorly and quickly iterate on our conversation design to improve quality.
    Product Engineer
    Victor
    Product Engineer
    This is like Optimizely for building AI voice agents. I can quickly test new ideas and iterate on my voice agent design.
    DevOps Engineer
    Sarah
    DevOps Engineer
    We catch regressions before they reach users by automatically testing our voice agents.

    Voice Scenarios

    Create diverse voice scenarios with different accents, background noise levels, and user behaviors to test your AI voice agents thoroughly.

    Voice Experiment Tracking

    Track hypotheses and results for each voice agent experiment. Measure improvements in accuracy, latency, and call completion.

    Score Override

    Override AI conversation scores to align with your quality standards. Each override helps train the AI judge to better evaluate voice interactions.

    Call Analytics

    Search and analyze voice conversations to understand why your AI agent responded in a particular way. Identify patterns in successful and failed interactions.

    Sampling

    Run multiple tests of the same scenario to understand performance variations. Identify and fix flaky interactions before they reach users.

    Team Collaboration

    Share voice datasets, test results, and conversation traces with your team. Collaborate to improve voice agent performance across all scenarios.

    FAQs

    Use any LLM, agent and voice infra providers

    We provide platform-agnostic hooks to simulate conversations, evaluate and log your traces.

    Anthropic
    OpenAI
    Google
    Bland
    VoCode
    Retell
    Vapi
    Logo

    Ship reliable AI voice agents with confidence

    We've built mission critical data products at
    • Tesla
    • Microsoft
    • Anduril
    • Square
    • Citizen