This is the Linux app named LLM-Pruner whose latest release can be downloaded as LLM-Prunersourcecode.tar.gz. It can be run online in the free hosting provider OnWorks for workstations.
Download and run online this app named LLM-Pruner with OnWorks for free.
Follow these instructions in order to run this app:
- 1. Downloaded this application in your PC.
- 2. Enter in our file manager https://www.onworks.net/myfiles.php?username=XXXXX with the username that you want.
- 3. Upload this application in such filemanager.
- 4. Start the OnWorks Linux online or Windows online emulator or MACOS online emulator from this website.
- 5. From the OnWorks Linux OS you have just started, goto our file manager https://www.onworks.net/myfiles.php?username=XXXXX with the username that you want.
- 6. Download the application, install it and run it.
SCREENSHOTS
Ad
LLM-Pruner
DESCRIPTION
LLM-Pruner is an open-source framework designed to compress large language models through structured pruning techniques while maintaining their general capabilities. Large language models often require enormous computational resources, making them expensive to deploy and inefficient for many practical applications. LLM-Pruner addresses this issue by identifying and removing non-essential components within transformer architectures, such as redundant attention heads or feed-forward structures. The framework relies on gradient-based analysis to determine which parameters contribute least to model performance, enabling targeted structural pruning rather than simple weight removal. After pruning, the framework applies lightweight fine-tuning methods such as LoRA to recover performance using relatively small datasets and short training times.
Features
- Structured pruning of transformer components such as layers and attention heads
- Gradient-based importance scoring for identifying removable parameters
- Compatibility with multiple LLM architectures including LLaMA and Vicuna
- Lightweight performance recovery using LoRA fine-tuning
- Automated scripts for pruning and model compression workflows
- Reduced memory usage and faster inference for large language models
Programming Language
Python
Categories
This is an application that can also be fetched from https://sourceforge.net/projects/llm-pruner.mirror/. It has been hosted in OnWorks in order to be run online in an easiest way from one of our free Operative Systems.