Driss Guessous
drisspg
@pytorch core
Languages
Loading contributions...
Top Repositories
A place to store reusable transformer components of my own creation or found on the interwebs
Cuda extensions for PyTorch
Learnings + Exercises from the PMPP book!
Lean 4 formalizations of proofs from Stephen Abbott's Understanding Analysis textbook
PyTorch Job Queue — dispatch agents to fix PyTorch issues
Repositories
62Rust-based GPU kernel library with PyTorch bindings via cutile-rs + PyO3
No description provided.
No description provided.
No description provided.
PyTorch Job Queue — dispatch agents to fix PyTorch issues
A place to store reusable transformer components of my own creation or found on the interwebs
Fast and memory-efficient exact attention
A tool for working with stacked PRs on github.
Learnings + Exercises from the PMPP book!
No description provided.
Open Source Developer Cloud
Lean 4 formalizations of proofs from Stephen Abbott's Understanding Analysis textbook
A Quirky Assortment of CuTe Kernels
Claude code skills for PT2 triage
Tensors and Dynamic neural networks in Python with strong GPU acceleration
A high-throughput and memory-efficient inference and serving engine for LLMs
No description provided.
Cuda extensions for PyTorch
No description provided.
TORCH_LOGS parser for PT2
FB (Facebook) + GEMM (General Matrix-Matrix Multiplication) - https://code.fb.com/ml-applications/fbgemm/
No description provided.
Helpful tools and examples for working with flex-attention
Tritonbench is a collection of PyTorch custom operators with example inputs to measure their performance.
No description provided.
CUDA Templates for Linear Algebra Subroutines
SGLang is a fast serving framework for large language models and vision language models.
Hackable implementation of state-of-the-art open-source LLMs based on nanoGPT. Supports flash attention, 4-bit and 8-bit quantization, LoRA and LLaMA-Adapter fine-tuning, pre-training. Apache 2.0-licensed.
Development repository for the Triton language and compiler
The simplest, fastest repository for training/finetuning medium-sized GPTs.