1,119 results for “topic:gpu-computing”
A fast, scalable, high performance Gradient Boosting on Decision Trees library, used for ranking, classification, regression and other machine learning tasks for Python, R, Java, C++. Supports computation on CPU and GPU.
Video stabilization using gyroscope data
High-performance TensorFlow library for quantitative finance.
[ARCHIVED] The C++ parallel algorithms library. See https://github.com/NVIDIA/cccl
The fastest and most memory efficient lattice Boltzmann CFD software, running on all GPUs and CPUs via OpenCL. Free for non-commercial use.
Lingvo
Resource scheduling and cluster management for AI
General purpose GPU compute framework built on Vulkan to support 1000s of cross vendor graphics cards (AMD, Qualcomm, NVIDIA & friends). Blazing fast, mobile-enabled, asynchronous and optimized for advanced GPU data processing usecases. Backed by the Linux Foundation.
A modular and open-ended toolkit for WebGPU, with advanced type inference and the ability to write shaders in TypeScript
CUDA Core Compute Libraries
GPGPU microprocessor architecture
CUDA integration for Python, plus shiny features
Parallel Computing and Scientific Machine Learning (SciML): Methods and Applications (MIT 18.337J/6.338J)
Deep learning in Rust, with shape checked tensors and neural networks
😎 Curated list of awesome things around WebGPU ecosystem.
Compiler for multiple programming models (SYCL, C++ standard parallelism, HIP/CUDA) for CPUs and GPUs from all vendors: The independent, community-driven compiler for C++-based heterogeneous programming models. Lets applications adapt themselves to all the hardware in the system - even at runtime!
Simulation of spiking neural networks (SNNs) using PyTorch.
The write-once-run-anywhere GPGPU library for Rust
TornadoVM: A practical and efficient heterogeneous programming framework for managed languages
An efficient C++20 GPU numerical computing library with Python-like syntax
A fast, ergonomic and portable tensor library in Nim with a deep learning focus for CPU, GPU and embedded devices via OpenMP, Cuda and OpenCL backends
LuxCore source repository
stdgpu: Efficient STL-like Data Structures on the GPU
Fast Clojure Matrix Library
Embedded language for high-performance array computations
Thin, unified, C++-flavored wrappers for the CUDA APIs
Blender Integration for LuxCore
This guide should help fellow researchers and hobbyists to easily automate and accelerate there deep leaning training with their own Kubernetes GPU cluster.
List of Deep Learning Cloud Providers
Performance-Portable Particle-in-Cell Simulations for the Exascale Era :sparkles: