Toot

Written by hok@lemmy.dbzer0.com on 2024-12-07 at 17:07

I followed their instructions here: speech.fish.audio

I am using the API server to do inference: speech.fish.audio/inference/#http-api-inference

I don’t know about other ways. To be clear, this is not (necessarily) an LLM, it’s just for speech synthesis, so you don’t run it on ollama. That said I think it does technically use Llama under the hood since there are two models, one for encoding text and the other for decoding to audio. Honestly the paper is terrible but it explains the architecture somewhat: arxiv.org/pdf/2411.01156

=> More informations about this toot | View the thread | More toots from hok@lemmy.dbzer0.com

Mentions

=> View PerogiBoi@lemmy.ca profile

Tags

Proxy Information
Original URL
gemini://mastogem.picasoft.net/toot/113612635963742302
Status Code
Success (20)
Meta
text/gemini
Capsule Response Time
230.013498 milliseconds
Gemini-to-HTML Time
0.423562 milliseconds

This content has been proxied by September (3851b).