You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
openai-whisper-talk is a sample voice conversation application powered by OpenAI technologies such as Whisper, Completions, Embeddings, and the latest Text-to-Speech. The application is built using Nuxt, a Javascript framework based on Vue.js.
React / Vanilla JS text-to-speech with highlighting the words and sentences that are being spoken using audio files, text-to-speech API, and web speech synthesis API
Code snippets showing how to record I2S audio and store as .wav file on ESP32 with SD card, how to transcribe pre-recorded audio via Deepgram SpeechToText API, how to generate audio from text via TextToSpeech API from OpenAI a/o Google TTS. Triggering ESP32 actions via Voice.
This repository features a Gradio interface designed to leverage the OpenAI Text-To-Speech (TTS) API. The interface lets users create speech from provided text using different models and voice options.
An AI Voice Assistant that can read emails, WhatsApp messages, clipboard data, and captures webcam images and screenshots for contextual understanding.
This project allows you to create short videos based on a given topic. Version 2 (improve transform image and real clip) can be found at: Colab notebook. The reason why version 2 is not in this GitHub is that my laptop cannot afford the Nvidia-smi utility, which is required to install xformers for stable diffusion.
This Next.js application provides a user interface for interacting with OpenAI's audio APIs, including text-to-speech and speech-to-text. Features include history tracking, audio playback, and download functionality.
This Python script processes a video file, generates a compelling description, creates a voiceover script in the style of David Attenborough, and synthesizes the voiceover using OpenAI's Text-to-Speech API.