Layer-Guided Latent Reasoning: Exploring Targeted Manipulation of Processing Regions in LLMs
Abstract
Layer-Guided Latent Reasoning (LGLR) is a study that explores whether adjusting just a few layers in a transformer model can achieve the benefits of recent reasoning techniques—like Coconut—without the high computational cost those methods usually require. The idea comes from observing where reasoning actually occurs within large language models (LLMs). The method is tested on the GSM8K benchmark, which focuses on arithmetic reasoning. Results show that focusing on specific transformer layers can cut the computation needed at each step by 27–67%, while still keeping accuracy close to that of more complex reasoning methods. This points to a more efficient way for language models to reason.
Permanent Link
http://digital.library.wisc.edu/1793/95538Type
Thesis
Description
Senior Honors Thesis, Department of Computer Sciences, University of Wisconsin-Madison

