Back to Templates
Execution video: Youtube Link
I built an AI voice-triggered RAG assistant where ElevenLabs’ conversational model acts as the front end and n8n handles the brain....here’s the real breakdown of what’s happening in that workflow:
Webhook (/inf
)
user_question
.Embed User Message (Together API - BAAI/bge-large-en-v1.5)
Search Embeddings (Supabase RPC)
matchembeddings1
to find the top 5 most relevant context chunks from your stored knowledge base.Aggregate
chunk
values into one block of text so the LLM gets full context at once.Basic LLM Chain (LangChain node)
Respond to Webhook
You essentially have:
Voice → Text → Embedding → Vector Search → Context Injection → LLM → Response → Voice