Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

Can you describe how it pick different voices? Does it need separate audio channels, or does it recognize different voices on the same audio input?


It separate mic/speaker as 2 channel. So you can reliably get "what you said" vs "what you heard".

For splitting speaker within channel, we need AI model to do that. It is not implemented yet, but I think we'll be in good shape somewhere in September.

Also we have transcript editor that you can easily split segment, assign speakers.




Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: