This is the Linux app named Gemma whose latest release can be downloaded as v3.2.1sourcecode.tar.gz. It can be run online in the free hosting provider OnWorks for workstations.
Download and run online this app named Gemma with OnWorks for free.
Follow these instructions in order to run this app:
- 1. Downloaded this application in your PC.
- 2. Enter in our file manager https://www.onworks.net/myfiles.php?username=XXXXX with the username that you want.
- 3. Upload this application in such filemanager.
- 4. Start the OnWorks Linux online or Windows online emulator or MACOS online emulator from this website.
- 5. From the OnWorks Linux OS you have just started, goto our file manager https://www.onworks.net/myfiles.php?username=XXXXX with the username that you want.
- 6. Download the application, install it and run it.
SCREENSHOTS
Ad
Gemma
DESCRIPTION
Gemma, developed by Google DeepMind, is a family of open-weights large language models (LLMs) built upon the research and technology behind Gemini. This repository provides the official implementation of the Gemma PyPI package, a JAX-based library that enables users to load, interact with, and fine-tune Gemma models. The framework supports both text and multi-modal input, allowing natural language conversations that incorporate visual content such as images. It includes APIs for conversational sampling, parameter management, and integration with fine-tuning methods like LoRA. The Gemma library can operate efficiently on CPUs, GPUs, or TPUs, with recommended configurations depending on model size. Through included tutorials and Colab notebooks, users can explore examples covering sampling, multi-modal interactions, and fine-tuning workflows. By providing accessible open-weight models, Gemma enables researchers and developers to experiment with state-of-the-art LLM architectures.
Features
- JAX-based library for running and fine-tuning Gemma large language models
- Supports multi-turn and multi-modal conversations, including image inputs
- Provides open-weight checkpoints for different model sizes (2B, 7B, etc.)
- Compatible with CPU, GPU, and TPU environments
- Includes examples for sampling, fine-tuning, and LoRA-based adaptation
- Integrates easily through a simple PyPI installation and Python interface
Programming Language
Python
Categories
This is an application that can also be fetched from https://sourceforge.net/projects/gemma.mirror/. It has been hosted in OnWorks in order to be run online in an easiest way from one of our free Operative Systems.