This is the Linux app named FasterTransformer whose latest release can be downloaded as v5.3releasesourcecode.tar.gz. It can be run online in the free hosting provider OnWorks for workstations.
Download and run online this app named FasterTransformer with OnWorks for free.
Follow these instructions in order to run this app:
- 1. Downloaded this application in your PC.
- 2. Enter in our file manager https://www.onworks.net/myfiles.php?username=XXXXX with the username that you want.
- 3. Upload this application in such filemanager.
- 4. Start the OnWorks Linux online or Windows online emulator or MACOS online emulator from this website.
- 5. From the OnWorks Linux OS you have just started, goto our file manager https://www.onworks.net/myfiles.php?username=XXXXX with the username that you want.
- 6. Download the application, install it and run it.
SCREENSHOTS:
FasterTransformer
DESCRIPTION:
FasterTransformer is a high-performance inference library designed to accelerate transformer-based models such as BERT, GPT, and T5 on NVIDIA GPUs. It provides optimized implementations of transformer encoder and decoder layers using CUDA, cuBLAS, and custom kernels to maximize throughput and minimize latency. The library supports multiple deep learning frameworks, including TensorFlow, PyTorch, and Triton, allowing developers to integrate it into existing pipelines without major changes. It includes advanced optimization techniques such as mixed precision, tensor parallelism, and efficient memory management, enabling large models to run across multiple GPUs and nodes. FasterTransformer is particularly focused on inference workloads, where it significantly improves performance compared to standard framework implementations. Although development has transitioned toward TensorRT-LLM, the project remains an important reference for understanding optimized transformer execution.
Features
- Optimized transformer encoder and decoder implementations
- Support for BERT, GPT, T5, and related architectures
- Multi-GPU and multi-node inference with parallelism
- Mixed precision support including FP16 and INT8
- Integration with TensorFlow, PyTorch, and Triton
- High-performance CUDA and cuBLAS-based kernels
Programming Language
C++
Categories
This is an application that can also be fetched from https://sourceforge.net/projects/fastertransformer.mirror/. It has been hosted in OnWorks in order to be run online in an easiest way from one of our free Operative Systems.