Turn incoming chat messages into fast, helpful replies using a mix of cloud and local AI. Great for help desks or internal teams that need a responsive assistant that remembers recent context.
A chat event starts the flow. Messages can be handled by a simple LLM chain powered by a local Ollama model, or sent to DeepSeek using direct HTTP calls. There is also an AI Agent option with a memory window so the bot can keep track of the last messages in a conversation. The setup shows both JSON and raw body calls to DeepSeek Chat V3 and the Reasoner model, plus a system message that guides tone and role. You can choose between local processing for cost control and cloud calls for higher capacity.
You will need a DeepSeek API key and a running Ollama instance with the deepseek r1 model. Expect faster first replies, fewer repeated questions for agents, and more consistent answers. Useful for website chat, internal IT Q and A, and triage for common tickets. Follow the steps below to connect credentials, set model names, and run a quick end to end test.