Compare answers from several language models in one place. Send one chat message and see how each model writes, how fast it replies, and how easy it is to read. Useful for content, support, and product teams that need clear, simple text.
A new chat message starts the run. The flow pulls the list of loaded models from your local LM server, then runs the same prompt through each model. A system prompt can guide tone and reading level. Start and end times are recorded to measure latency. A code step scores readability, counts words and sentences, and calculates averages. All results can be saved to Google Sheets for side by side review.
Setup needs LM Studio running with your chosen models and a Google Sheet if you want logs. Update the base URL to your local server and tune temperature, top p, and presence penalty to match your test plan. Most teams can cut manual comparisons from an hour to a few minutes. Expect faster evaluations and fewer copy paste steps, so you can pick the best model for FAQs, training notes, release summaries, or help center text.