Turn live chat questions into clear AI replies that follow a strict JSON format. Great for support teams that need fast answers and clean data they can send to other systems. It speeds up replies and keeps the structure consistent across every message.
A chat event starts the flow and passes the text into an LLM chain. The chain uses the Ollama model to produce a JSON object with two fields called Prompt and Response. A mapping step then normalizes the output into a clean object, and a final set node controls what the user sees. If anything fails, an error branch returns a safe fallback so the chat never goes silent. You get reliable structure, simple routing, and outputs ready for storage or follow up tasks.
To set it up, you need n8n and an Ollama server with the llama3.2 model available. Add your Ollama connection in n8n, keep the provided prompt template, and confirm the mapping fields match your needs. Teams can expect faster answers and fewer mistakes, often cutting reply prep from minutes to seconds. Good fits include help desk chat, FAQ assistants, and any tool that needs structured AI responses for logging or downstream processing.