Lost in Conversation is a code repository designed for benchmarking large language models (LLMs) on multi-turn task completion, enabling the reproduction of experiments from the paper "LLMs Get Lost in Multi-Turn Conversation." It includes tools for simulating conversations across various tasks, a web-based viewer, and instructions for integrating with LLMs. The repository is intended for research purposes and emphasizes careful evaluation and oversight of outputs to ensure accuracy and safety.
llm ✓
simulation ✓
research ✓
tasks ✓
benchmarking ✓