Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

Most of the Pipecat examples we've been working on are focused on speech-to-speech. The examples guide you through how to do that (or you can give the hosted storytelling example a try: https://storytelling-chatbot.fly.dev/)

We should probably update the example in the README to better represent that, thank you!



Your project is amazing and I'm not trying to take away from what you have accomplished.

But..I looked at the code but didn't see any audio-to-audio service or model. Can you link to an example of that?

I don't mean speech to text to LLM to text to speech. I mean speech-to-speech directly, as in the ML model takes audio as input and outputs audio. As they have now in OpenAI.

I am very familiar with the typical multi-model workflow and have implemented it several times.


That's absolutely amazing, both visually and technically! Do you share any insights of the development process, perhaps some code?


I just realized this is exactly the example provided in the repo which I haven't run yet! Thanks for adding this!




Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: