Real World Speech Interfaces: What Builders Are Seeing

by Jaim Zuber, Ian Bicking, and Vaish Sagar | at Minnebar20

Are you building with speech, or thinking about it? Let’s compare notes.

Speech interfaces (e.g. speech-to-text, text-to-speech) have become one of the most practical (and widespread) ways AI shows up in real products.

2025 was the year of AI meeting note takers. Transcription apps are everywhere (and seemingly popped up out of nowhere).

2026 is about dictation. WisprFlow and Willow are changing how we interact with computers. Prompting Claude from a keyboard is sooo 2025.

Yet, it feels like we’ve only scratched the surface of what these technologies can do.

Building in the space? Let’s talk shop and make wild guesses on where this is headed.

Jaim Zuber

Apple Platforms Engineering and Leadership. Mobile and desktop.

Hacking on speech-to-text (ASR) systems while he searches for the next gig.

He likes hockey, BBQ, and making noise with a modest array of instruments… sometimes in public.

Links: - jaimzuber.com - BlueSky - Mastodon - GitHub

Ian Bicking

Ian Bicking is a software developer, currently independent. His accidental largest contribution was as author of pip and virtualenv.

He previously worked at Brilliant.org, Meta, and Mozilla.

Links:

Vaish Sagar

I'm an AI engineer, full-stack developer, and two-time founder with a Master's in Computer Science specializing in NLP from Arizona State University. Recognized by the U.S. government with an O-1A for extraordinary abilities in AI, I have five years of experience spanning enterprise data engineering at Oracle, consumer product development, and applied AI research. I've built and shipped two startup platforms from scratch, was a Sequoia Capital Arc Accelerator finalist, and was featured on Fox9 News for consumer tech innovation. What sets my work apart is where it lives: at the intersection of AI and music. I've built AI tools inside DAW environments, applied music theory principles to reduce artifacts in AI-generated audio, and trained a GPT-style transformer from scratch to study musical structures in song lyrics. As a guitarist and vocalist, I bring a musician's ear to my technical work. My current project, "What Does B-Major Sound Like in a Parallel Universe?", is a reflection of that curiosity, of exploring what music could be.


Are you interested in this session?

This will add your name to the list of interested participants. It will help us gauge interest for scheduling purposes.

Interested Participants