The paper introduces the Chain of Draft (CoD) paradigm, which enables Large Language Models (LLMs) to generate concise intermediate reasoning outputs, mimicking human draft strategies. By focusing on essential information and reducing verbosity, CoD achieves comparable or superior accuracy to Chain-of-Thought prompting while utilizing significantly fewer tokens, thus lowering costs and latency in reasoning tasks.