RAG Q&A Agent
Search your Qdrant vector store for relevant context, then answer with an LLM — grounded in your own documents.
RAG Q&A Agent
The retrieval-augmented answer pattern in four nodes: question → search → answer → output. Ground your LLM in your own documents and stop hallucinations at the source.
What this workflow does
- Input receives the user question
- RAG node (search mode) retrieves the top-K most relevant chunks from Qdrant
- LLM node uses the retrieved context to answer accurately
- Output returns the grounded answer
Use cases
- Internal knowledge base chatbot
- Policy and compliance Q&A
- Product documentation assistant
Setup
Configure the RAG node with your Qdrant collection (same one used in the RAG Document Ingest template). The LLM system instruction already references $RagSearch.context — connect your preferred model.
FAQ
Can I combine both RAG templates? Yes — run Ingest once per document, then re-use Q&A for every question.
Which embedding model should I use? Use the same model in both Ingest and Q&A to ensure vector compatibility.
How to import this template
- 1Click Import → Copy JSON on this page.
- 2Open your Heym and navigate to a workflow canvas.
- 3Press Cmd+V / Ctrl+V — nodes appear instantly.
- 4Add your API keys in the node config panels and click Run.
Click a node to select it — same as the Heym editor; the panel shows its settings.
6 nodes · Free & source-available
More workflow templates
Explore related automations — each page links to other templates so you can discover more use cases.
- Batch LLM Status TrackerSend an array through the OpenAI Batch API, branch on live status updates, and collect the final per-item results.
- Gemini Image CreatorGenerate images from a text prompt using Gemini's native image output.
- PDF / DOCX Translation AgentTranslate the full text of any uploaded document using an AI agent.
- Meeting Notes → JSON TasksTurn messy meeting notes into structured JSON tasks with the LLM node's JSON output mode — no image pipeline required.
- Inbox TL;DR SummarizerPaste a long email or thread — one LLM call returns a short TL;DR with next actions.
- RAG Document IngestChunk and embed a document into a Qdrant vector store so it can be retrieved later by the RAG Search node.
- Language Switch RouterDetect the language of incoming text with an LLM and route to the matching branch using the Switch node.
- Telegram FAQ Auto ReplyReply to inbound Telegram questions with an LLM and keep the latest question in a global variable.
- IMAP Support Inbox TriageWatch a shared mailbox, summarize incoming support email, and route urgent messages to Slack.
- Jina Web FetcherFetch clean, LLM-ready text from any URL using the Jina Reader API.
- Cursor Post NotifierMonitor the Cursor blog on a schedule and Slack-notify your team when a new post goes live.
- Claude Blog MonitorMonitor the Anthropic blog on a schedule and Slack-notify your team on new Claude posts.