LORD: Low Rank Decomposition Of Monolingual Code LLMs For One-Shot Compression
This paper explores the potential to compress Large Language Models (LLMs) using Low Rank Decomposition (LoRD). The researchers found that ranks for linear layers in these models can be reduced…
Continue reading