# wingmanAI
**Repository Path**: data_factory/wingmanAI
## Basic Information
- **Project Name**: wingmanAI
- **Description**: No description available
- **Primary Language**: Unknown
- **License**: MIT
- **Default Branch**: main
- **Homepage**: None
- **GVP Project**: No
## Statistics
- **Stars**: 0
- **Forks**: 0
- **Created**: 2024-04-26
- **Last Updated**: 2024-07-20
## Categories & Tags
**Categories**: Uncategorized
**Tags**: None
## README
WingmanAI
WingmanAI is a powerful tool for interacting with real-time transcription of both system and microphone audio. Powered by ChatGPT, this tool lets you interact in real-time with the transcripts as an extensive memory base for the bot, providing a unique communication platform.
## Demo
https://github.com/e-johnstonn/wingmanAI/assets/30129211/6f9f8e09-f43e-47d5-87ae-ac5bc693963d
As you can see, the bot can answer questions about past conversations when you load the transcripts for a designated person.
## Features
- **Real-time Transcription**: WingmanAI can transcribe both system output and microphone input audio, allowing you to view the live transcription in an easy-to-read format.
- **ChatGPT Integration**: You can chat with a ChatGPT powered bot that reads your transcripts in real-time.
- **Efficient Memory Management**: The bot maintains a record of the conversation but in a token-efficient manner, as only the current chunk of transcript is passed to the bot.
- **Save and Load Transcripts**: WingmanAI allows you to save transcripts for future use. You can load them up anytime later, and any query made to the bot will be cross-referenced with a vector database of the saved transcript, providing the bot with a richer context.
- **Append Conversations**: You can keep appending to the saved transcripts, building a vast database over time for the bot to pull from.
## Installation
1. Clone the repository.
2. Install the requirements: ```pip install -r requirements.txt```
3. If you wish to use CUDA for Whisper (which is highly recommended), uninstall (```pip uninstall torch```) torch and run: ```pip3 install torch torchvision torchaudio --index-url https://download.pytorch.org/whl/cu117```
**Note**: This application is currently compatible only with Windows.
## Prerequisites
Ensure you have `ffmpeg` installed on your system.
Have a working OpenAI API key.
Works best using CUDA! CPU transcription is not real-time.
The model currently being used is the "base" model - if your hardware can't run it, change it to "tiny". Language is currently set to English.
## Getting Started
1. Add your OpenAI API key to the `keys.env` file.
2. Run `main.py`.
For any queries or issues, feel free to open a new issue in the repository.
Contributions are always welcomed to improve the project.
## Acknowledgements
This project uses a modified version of SevaSk's "Ecoute" project for the transcriptions - check it out [here](https://github.com/SevaSk/ecoute)!