Large Language Models (LLMs) struggle with computational overhead in complex reasoning tasks due to lengthy Chain-of-Thought (CoT) sequences. TokenSkip, developed by researchers from The Hong Kong Polytechnic University and the University of Science and Technology of China, aims to optimize CoT processing by skipping less
Sort: