Repos
9
Stars
0
Forks
1
Top Language
Python
Loading contributions...
Repositories
9A Datacenter Scale Distributed Inference Serving Framework
helm charts for deploying models with llm-d
Achieve state of the art inference performance with modern accelerators on Kubernetes
A high-throughput and memory-efficient inference and serving engine for LLMs
Unified Communication X (mailing list - https://elist.ornl.gov/mailman/listinfo/ucx-group)
Time-HD-Lib: A Library for High-Dimensional Time Series Forecasting
Generative AI Examples is a collection of GenAI examples such as ChatQnA, Copilot, which illustrate the pipeline capabilities of the Open Platform for Enterprise AI (OPEA) project.
Extending Hugging Face transformers APIs for Transformer-based models and improve the productivity of inference deployment. With extremely compressed models, the toolkit can greatly improve the inference efficiency on Intel platforms.
No description provided.