This is the Linux app named XNNPACK whose latest release can be downloaded as XNNPACKsourcecode.tar.gz. It can be run online in the free hosting provider OnWorks for workstations.
Download and run online this app named XNNPACK with OnWorks for free.
Follow these instructions in order to run this app:
- 1. Downloaded this application in your PC.
- 2. Enter in our file manager https://www.onworks.net/myfiles.php?username=XXXXX with the username that you want.
- 3. Upload this application in such filemanager.
- 4. Start the OnWorks Linux online or Windows online emulator or MACOS online emulator from this website.
- 5. From the OnWorks Linux OS you have just started, goto our file manager https://www.onworks.net/myfiles.php?username=XXXXX with the username that you want.
- 6. Download the application, install it and run it.
XNNPACK
Ad
DESCRIPTION
XNNPACK is a highly optimized, low-level neural network inference library developed by Google for accelerating deep learning workloads across a variety of hardware architectures, including ARM, x86, WebAssembly, and RISC-V. Rather than serving as a standalone ML framework, XNNPACK provides high-performance computational primitives—such as convolutions, pooling, activation functions, and arithmetic operations—that are integrated into higher-level frameworks like TensorFlow Lite, PyTorch Mobile, ONNX Runtime, TensorFlow.js, and MediaPipe. The library is written in C/C++ and designed for maximum portability, efficiency, and performance, leveraging platform-specific instruction sets (e.g., NEON, AVX, SIMD) for optimized execution. It supports NHWC tensor layouts and allows flexible striding along the channel dimension to efficiently handle channel-split and concatenation operations without additional cost.
Features
- Cross-platform neural network inference backend optimized for ARM, x86, WebAssembly, and RISC-V
- High-performance implementations for 2D convolutions, pooling, activation, and quantization operators
- Supports both FP32 and INT8 inference with per-channel quantization
- Efficient NHWC tensor layout with flexible channel stride
- Integrates seamlessly with frameworks like TensorFlow Lite, TensorFlow.js, PyTorch, ONNX Runtime, and MediaPipe
- Multi-threaded and vectorized operator implementations
Programming Language
Assembly, C, C++, Unix Shell
Categories
This is an application that can also be fetched from https://sourceforge.net/projects/xnnpack.mirror/. It has been hosted in OnWorks in order to be run online in an easiest way from one of our free Operative Systems.