Skip to content

Navigation Menu

Sign in
Appearance settings

Search code, repositories, users, issues, pull requests...

Provide feedback

We read every piece of feedback, and take your input very seriously.

Saved searches

Use saved searches to filter your results more quickly

Sign up
Appearance settings

codeperfectplus/transmeet

Repository files navigation

πŸŽ™οΈ TransMeet β€” AI-Powered Meeting Summarizer

Turn your meeting recordings into clear, structured minutes using LLMs like Groq Whisper and Google Speech Recognition.


πŸš€ Features

  • βœ… Audio Transcription β€” Automatically convert .wav or .mp3 files into text
  • 🧠 LLM-Powered Summarization β€” Generate concise and structured meeting minutes
  • πŸ” Groq & Google Support β€” Choose between Groq Whisper models or Google Speech API
  • πŸͺ“ Automatic Chunking β€” Splits large files intelligently for smoother transcription
  • βš™οΈ Fully Customizable β€” Pick your preferred transcription and summarization models
  • 🧾 CLI & Python API β€” Use it from the terminal or integrate in your Python workflows
  • πŸ“ Clean Output β€” Saves transcripts and summaries neatly in your desired folder

πŸ“¦ Installation

pip install transmeet

Dependencies

sudo apt-get update && sudo apt-get install -y ffmpeg gcc && sudo apt-get clean && sudo rm -rf /var/lib/apt/lists/*

πŸ” Setup

Set your GROQ API Key/OPENAI API Key in your environment variables.

export GROQ_API_KEY=your_groq_api_key

To make this permanent:

echo 'export GROQ_API_KEY=your_groq_api_key' >> ~/.bashrc

If using OPENAI, set the OPENAI_API_KEY similarly. For Google Speech, no API key is needed; it uses the default model.


πŸ§‘β€πŸ’» How to Use

βœ… Option 1: Import as a Python Module

from transmeet import generate_meeting_transcript_and_minutes
generate_meeting_transcript_and_minutes(
 meeting_audio_file="/path/to/audio.wav",
 output_dir="complete_path_to_output_dir/",
 transcription_client="groq", # or "openai"
 transcription_model="whisper-large-v3-turbo", # change as per your need
 llm_client="groq", # or "openai"
 llm_model="llama-3.3-70b-versatile", # change as per your need
)

This will save two files in your output directory:

  • transcription_<timestamp>.txt
  • meeting_minutes_<timestamp>.md

πŸ”§ Option 2: Use the CLI

πŸ”Ή Basic Usage (Default: GROQ)

transmeet -i /path/to/audio.wav -o output/

πŸ”Έ Advanced Usage

transmeet \
 -i /path/to/audio.wav \
 -o output/ \
 --transcription-client groq \
 --transcription-model whisper-large-v3-turbo \
 --llm-client groq \
 --llm-model llama-3.3-70b-versatile \

πŸ—‚οΈ Output Structure

output/
β”œβ”€β”€ transcriptions/
β”‚ └── transcription_20250510_213038.txt
β”œβ”€β”€ meeting_minutes/
β”‚ └── meeting_minutes_20250510_213041.md

πŸ§ͺ Supported Formats

  • .wav
  • .mp3

βš™οΈ CLI Options

Argument Description
-i, --audio-path Path to the input audio file
-o, --output-dir Output directory (default: output/)
--transcription-client groq or google (default: groq)
--transcription-model e.g., whisper-large-v3-turbo
--llm-client groq or openai (default: groq)
--llm-model e.g., llama-3.3-70b-versatile

πŸ€– LLM Models

  • Groq Whisper: whisper-large, whisper-large-v3-turbo, etc.
  • Google Speech: Model defaults to their API standard
  • LLMs for minutes: llama-3, mixtral, gpt-4, etc. (Groq/OpenAI)

πŸ“‹ Roadmap

  • Add support for multi-language meetings
  • Speaker diarization support
  • Upload directly to Notion or Google Docs
  • Slack/Discord bots

πŸ§‘β€πŸŽ“ Author

Deepak Raj πŸ‘¨β€πŸ’» GitHub β€’ 🌐 LinkedIN


🀝 Contributing

Pull requests are welcome! Found a bug or need a feature? Open an issue or submit a PR.


βš–οΈ License

MIT License

About

LLM based meeting summarization tool

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

AltStyle γ«γ‚ˆγ£γ¦ε€‰ζ›γ•γ‚ŒγŸγƒšγƒΌγ‚Έ (->γ‚ͺγƒͺγ‚ΈγƒŠγƒ«) /