12 results for “topic:gpu-scheduling”
A control plane for concurrent LLM RL on shared GPUs
Fully Autonomous AI Research System with Self-Evolution, built natively on Claude Code
Tensor Fusion is a state-of-the-art GPU virtualization and pooling solution designed to optimize GPU cluster utilization to its fullest potential.
A tool for examining GPU scheduling behavior.
PipelineScheduler optimizes workload distribution between servers and edge devices, setting optimal batch sizes to maximize throughput and minimize latency amid content dynamics and network instability. It also addresses resource contention with spatiotemporal inference scheduling to reduce co-location interference.
The GPU Optimizer for ML Models enhances GPU performance for machine learning. It offers advanced scheduling, real-time monitoring, and efficient resource management through a user-friendly web interface and robust API, integrating big data technologies for seamless data processing and model optimization. @nvidia
Topology-aware Kubernetes scheduler for multi-tenant, heterogeneous clusters
# Fraud-Detection-Service ## The **fraud-detection-service** detects fraudulent orders and user activity. ### Endpoints - `GET /health` — service status - `POST /fraud/check` — check an order for fraud (sample) - `GET /fraud/:orderId` — get fraud status for an order (sample) ## Tracing This service reports telemetry
A distributed CPU/GPU task scheduler for large-scale batch jobs across thousands of machines. Zero dependencies, sub-millisecond latency.
Transparent suspend/resume runtime enabling preemptible GPU workloads via memory snapshotting, UVM paging, and execution state orchestration.
HPC research toolkit infrastructure for interfacing & analyzing LLMs (Kit is composed of: API gateway service, GPU scheduler, model servicer, and web interface)
Design of a GPU Dynamic LLM Inference Task Scheduling Architecture Based on KubeAI