Show simple item record

dc.contributor.authorBiswas, Aldrin Kabya
dc.date.accessioned2025-05-25T04:45:54Z
dc.date.available2025-05-25T04:45:54Z
dc.date.issued2025-05
dc.identifier.urihttp://ar.iub.edu.bd/handle/11348/997
dc.description.abstractAdapting large language models (LLMs) to downstream tasks via full fine-tuning is computationally prohibitive. While parameter-efficient fine-tuning (PEFT) methods exist, they often rely on predefined heuristics, incur training overhead for parameter selection, or suffer from poor generalization. This work introduces a novel, training-free layer selection strategy for partial fine-tuning. Our approach leverages the geometric relationships between layer representations by computing the cosine similarity of the [CLS] token embeddings across all layers before fine-tuning begins, using only a single forward pass on a sample of the data. This yields layer-wise importance scores, allowing us to strategically select a small subset of layers for adaptation while freezing the rest. Extensive experiments across 15 diverse NLP tasks, including single-sentence and sentence-pair classifications, demonstrate that our method consistently outperforms various PEFT baselines, including heuristic selections, dynamic/gradient-based methods, and I/O similarity-based selection. Critically, it achieves performance remarkably close (often within 1-2%) to full fine-tuning while drastically reducing trainable parameters by up to 75% and reaching training speedups of 1.5×. Furthermore, the method exhibits superior robustness in cross-domain evaluations compared to baselines and generalizes effectively across different model architectures. By exploiting inherent structural properties of pre-trained models via inter-layer [CLS] token similarity, our approach offers an efficient, effective, and robust paradigm for partial LLM fine-tuning.en_US
dc.language.isoenen_US
dc.publisherIUBen_US
dc.subjectlarge language modelsen_US
dc.subjectparameter-efficient fine-tuningen_US
dc.titleTraining-Free Layer Selection for Parameter-Efficient Fine-Tuning of Language Modelsen_US
dc.typeThesisen_US


Files in this item

Thumbnail

This item appears in the following Collection(s)

Show simple item record


Copyright © 2002-2021  IUB Academic Repository.
Maintained by  Library Information Technology (LIT)
LIT