This is the Linux app named mllm whose latest release can be downloaded as MLLM-V2V2.0.0Releasesourcecode.tar.gz. It can be run online in the free hosting provider OnWorks for workstations.
Download and run online this app named mllm with OnWorks for free.
Follow these instructions in order to run this app:
- 1. Downloaded this application in your PC.
- 2. Enter in our file manager https://www.onworks.net/myfiles.php?username=XXXXX with the username that you want.
- 3. Upload this application in such filemanager.
- 4. Start the OnWorks Linux online or Windows online emulator or MACOS online emulator from this website.
- 5. From the OnWorks Linux OS you have just started, goto our file manager https://www.onworks.net/myfiles.php?username=XXXXX with the username that you want.
- 6. Download the application, install it and run it.
SCREENSHOTS
Ad
mllm
DESCRIPTION
mllm is an open-source inference engine designed to run multimodal large language models efficiently on mobile devices and edge computing environments. The framework focuses on delivering high-performance AI inference in resource-constrained systems such as smartphones, embedded hardware, and lightweight computing platforms. Implemented primarily in C and C++, it is designed to operate with minimal external dependencies while taking advantage of hardware-specific acceleration technologies such as ARM NEON and x86 AVX2 instructions. The system supports multiple optimization techniques including quantization, pruning, and speculative decoding to improve performance while reducing computational overhead. It also provides tools to convert models from popular formats like PyTorch checkpoints into optimized runtime formats that can be executed on supported hardware platforms.
Features
- Lightweight multimodal LLM inference engine optimized for mobile and edge devices
- Support for ARM CPUs, x86 processors, and specialized accelerators such as Qualcomm NPUs
- Model conversion utilities for importing PyTorch and SafeTensors checkpoints
- Advanced optimization techniques including quantization, pruning, and speculative decoding
- Command-line and Android demonstration applications for running local inference
- Support for multimodal models combining text, vision, and image understanding tasks
Programming Language
C++
Categories
This is an application that can also be fetched from https://sourceforge.net/projects/mllm.mirror/. It has been hosted in OnWorks in order to be run online in an easiest way from one of our free Operative Systems.