This tool extracts audio from video files and transcribes it using OpenAI's Whisper model.
- Python 3.11 or later
- FFmpeg installed on your system
- uv (Python package installer)
- CUDA-compatible GPU (optional, for faster transcription)
brew install ffmpegsudo apt update
sudo apt install ffmpegDownload from ffmpeg.org and add ffmpeg into the user PATH
- Download the
.7zor.zipfile. - Extract the downloaded ZIP file to a folder
ffmpeg. - Edit the path variable.
Path->New-> place theffmpeg/bindirectory. - Save and restart the terminal.
Before running the project, please check if ffmpeg is working in you computer or not.
ffmpeg -versionpip install uvClone this repository and navigate to the project directory:
git clone https://github.com/Badhansen/transcribe.git
cd transcribeuv sync
source .venv/bin/activate # On Windows, use `.venv\Scripts\activate`- Place your video file in the project directory
- Update the video filename in
main.pyif needed - Run the transcription:
python main.pyThe transcription will be saved to transcript.txt.
- ffmpeg-python: Python bindings for FFmpeg
- openai-whisper: OpenAI's Whisper speech recognition model
- PyTorch: Required for Whisper model execution