Skip to content

Conversation

@mpetruc
Copy link

@mpetruc mpetruc commented Oct 7, 2025

  • Added TTS server configuration to TextToSpeech component in Settings page
  • Fixed voice type checking and server config passing for Kokoro voices
  • Ensured both browser and Kokoro voices work in chat messages

Copy link
Owner

@olegshulyakov olegshulyakov left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

@mpetruc There are some changes needed:

  • src/types/provider.ts
    You need to introduce types VoiceProvider and TextToSpeechProvider.
  • src/types/voice.ts
    Create types for OpenAI Text to Speech like src/types/inference.ts does for inference.
  • Create BaseVoiceOpenAIProvider in src/api/providers implementing LLMProvider, VoiceProvider and TextToSpeechProvider. Use can use src/api/providers/BaseOpenAIProvider.ts as reference.
  • Add ability to switch between browser and LLM voice generation in Settings -> Voice.
  • Create src/components/LLMTextToSpeech.tsx with all logic there.
  • src/config/config-default.json should contain ttsBaseUrl and ttsApiKey

I hope it will help https://github.com/openai/openai-node/tree/master/src/resources/audio

@olegshulyakov olegshulyakov linked an issue Oct 12, 2025 that may be closed by this pull request
3 tasks
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

[Feature]: Text-to-Speech using LLM

2 participants