Skip to content
#

multi-model-inference

Here is 1 public repository matching this topic...

Multi-model inference pipeline combining BART and T5 for summarization with Whisper for transcription (via yt-dlp fallback and ffmpeg normalization), plus vector-augmented RAG powered by Groq's Llama-3.3 70b versatile for intelligent Q&A

  • Updated Apr 14, 2026
  • Python

Improve this page

Add a description, image, and links to the multi-model-inference topic page so that developers can more easily learn about it.

Curate this topic

Add this topic to your repo

To associate your repository with the multi-model-inference topic, visit your repo's landing page and select "manage topics."

Learn more