Abdullah Şamil Güser

Paper Summaries

Title Author Publish Date Description
LoRA: Low-Rank Adaptation of Large Language Models Edward Hu et. al. 2021-10-16 LoRA introduces a resource-efficient approach for adapting large pre-trained language models, such as GPT-3, to specific tasks without the heavy costs of traditional fine-tuning. It maintains model quality, minimizes inference latency, and facilitates quick task-switching.
Parameter-Efficient Fine-Tuning Methods for Pretrained Language Models: A Critical Review and Assessment Lingling Xu et. al. 2023-12-19 This paper critically reviews Parameter Efficient Fine-Tuning (PEFT) methods for large pretrained language models (PLMs), highlighting their benefits in resource-limited settings. It assesses these methods’ performance, efficiency, and memory usage across tasks like natural language understanding, machine translation, and generation.
PaLM 2 Technical Report Rohan Anil et. al. 2023-05-17 We introduce PaLM 2, a new state-of-the-art language model that has better multilingual and reasoning capabilities and is more compute-efficient than its predecessor PaLM.
LIMA: Less Is More for Alignment Chunting Zhou et. al. 2023-05-18 Large language models are trained in two stages: (1) unsupervised pretraining from raw text, to learn general-purpose representations, and (2) large scale instruction tuning and reinforcement learning, to better align to end tasks and user preferences. We measure the relative importance of these two stages.
Training Compute-Optimal Large Language Models Jordan Hoffmann et. al. 2022-03-29 We investigate the optimal model size and number of tokens for training a transformer language model under a given compute budget.