KarhouTam
Music ๐ถ Coding ๐จโ๐ป Workout ๐โโ๏ธ Memes ๐บ Exploring ๐
Languages
Top Repositories
Benchmark of federated learning. Dedicated to the community. ๐ค
PyTorch implementation of Layer-wised Model Aggregation for Personalized Federated Learning
Implementation of SCAFFOLD: Stochastic Controlled Averaging for Federated Learning
PyTorch Implementation of Personalized federated learning with theoretical guarantees: A model-agnostic meta-learning approach
Some common CUDA kernel implementations (Not the fastest).
Implementation of FedAvg
Repositories
22Benchmark of federated learning. Dedicated to the community. ๐ค
LLM้ฉฑๅจ็ A/H่กๆบ่ฝๅๆๅจ๏ผๅคๆฐๆฎๆบ่กๆ + ๅฎๆถๆฐ้ป + Gemini ๅณ็ญไปช่กจ็ + ๅคๆธ ้ๆจ้๏ผ้ถๆๆฌ๏ผ็บฏ็ฝๅซ๏ผๅฎๆถ่ฟ่ก
PyTorch implementation of Layer-wised Model Aggregation for Personalized Federated Learning
PyTorch Implementation of Personalized federated learning with theoretical guarantees: A model-agnostic meta-learning approach
Tensors and Dynamic neural networks in Python with strong GPU acceleration
Some common CUDA kernel implementations (Not the fastest).
Implementation of SCAFFOLD: Stochastic Controlled Averaging for Federated Learning
No description provided.
torchcomms: a modern PyTorch communications API
๐LeetCUDA: Modern CUDA Learn Notes with PyTorch for Beginners๐, 200+ CUDA Kernels, Tensor Cores, HGEMM, FA-2 MMA.๐
PyTorch Implementation of Federated Reconstruction: Partially Local Federated Learning
TritonParse: A Compiler Tracer, Visualizer, and Reproducer for Triton Kernels
Implementation of FedAvg
A Python-embedded DSL that makes it easy to write fast, scalable ML kernels with minimal boilerplate.
A tweak to enhance Spotify experience
Material for gpu-mode lectures
SGLang is a fast serving framework for large language models and vision language models.
LLM training in simple, raw C/CUDA
Community maintained hardware plugin for vLLM on Ascend
Implementation of Improving Federated Learning Personalization via Model Agnostic Meta Learning
A high-throughput and memory-efficient inference and serving engine for LLMs
Ray is a unified framework for scaling AI and Python applications. Ray consists of a core distributed runtime and a set of AI Libraries for accelerating ML workloads.