This is the Linux app named VisualGLM-6B whose latest release can be downloaded as VisualGLM-6Bsourcecode.tar.gz. It can be run online in the free hosting provider OnWorks for workstations.
Download and run online this app named VisualGLM-6B with OnWorks for free.
Follow these instructions in order to run this app:
- 1. Downloaded this application in your PC.
- 2. Enter in our file manager https://www.onworks.net/myfiles.php?username=XXXXX with the username that you want.
- 3. Upload this application in such filemanager.
- 4. Start the OnWorks Linux online or Windows online emulator or MACOS online emulator from this website.
- 5. From the OnWorks Linux OS you have just started, goto our file manager https://www.onworks.net/myfiles.php?username=XXXXX with the username that you want.
- 6. Download the application, install it and run it.
SCREENSHOTS
Ad
VisualGLM-6B
DESCRIPTION
VisualGLM-6B is an open-source multimodal conversational language model developed by ZhipuAI that supports both images and text in Chinese and English. It builds on the ChatGLM-6B backbone, with 6.2 billion language parameters, and incorporates a BLIP2-Qformer visual module to connect vision and language. In total, the model has 7.8 billion parameters. Trained on a large bilingual dataset — including 30 million high-quality Chinese image-text pairs from CogView and 300 million English pairs — VisualGLM-6B is designed for image understanding, description, and question answering. Fine-tuning on long visual QA datasets further aligns the model’s responses with human preferences. The repository provides inference APIs, command-line demos, web demos, and efficient fine-tuning options like LoRA, QLoRA, and P-tuning. It also supports quantization down to INT4, enabling local deployment on consumer GPUs with as little as 6.3 GB VRAM.
Features
- 7.8B parameter multimodal conversational model (6.2B language + vision module)
- Supports Chinese and English image-based dialogue
- Pretrained on 330M bilingual image-text pairs for strong alignment
- Fine-tuning support via LoRA, QLoRA, and P-tuning for domain-specific tasks
- Efficient INT4 quantization allows inference with only 6.3 GB GPU memory
- Provides CLI demos, web demos, and REST API deployment options
Programming Language
Python, Unix Shell
Categories
This is an application that can also be fetched from https://sourceforge.net/projects/visualglm-6b.mirror/. It has been hosted in OnWorks in order to be run online in an easiest way from one of our free Operative Systems.