309 results for “topic:ai-infrastructure”
Open-source context retrieval layer for AI agents
Local persistent memory store for LLM applications including claude desktop, github copilot, codex, antigravity, etc.
🦞 LLM Token Compression & Reduction Tool — Cut AI agent token costs by up to 97%. 6-layer deterministic context compression for AI agent workspaces. No LLM required. Prompt compression, context window optimization & cost reduction for any LLM pipeline.
Semantica 🧠 — A framework for building semantic layers, context graphs, and decision intelligence systems with explainability and provenance.
Plug-and-play memory for LLMs in 3 lines of code. Add persistent, intelligent, human-like memory and recall to any model in minutes.
Grov automatically captures the context from your private AI sessions and syncs it to a shared team memory. It auto injects relevant memories across developers and future sessions to save tokens and time spent on tasks.
Route inference across LLM providers. Track cost per request.
Stop rewriting integrations. Open-source protocol suite that standardizes LLM, Vector, Graph, and Embedding interfaces across LangChain, LlamaIndex, AutoGen, CrewAI, Semantic Kernel, MCP — and any provider.
Distributed data mesh for real-time access, migration, and replication across diverse databases — built for AI, security, and scale.
A Rust runtime that unifies relational tables, graph relationships, and vector embeddings in a single tensor-based storage layer with distributed consensus and semantic search
NPU powered On-device AI Mobile applications using Melange
AI Infrastructure Engineer Learning Track - Production ML infrastructure curriculum (2-4 years experience)
Stop paying for AI APIs during development. LocalCloud runs everything locally - GPT-level models, databases, all free.
A curated list of awesome tools, frameworks, platforms, and resources for building scalable and efficient AI infrastructure, including distributed training, model serving, MLOps, and deployment.
Predictive memory layer for AI agents. MongoDB + Qdrant + Neo4j with multi-tier caching, custom schema support & GraphQL. 91% Stanford STARK accuracy, <100ms on-device retrieval
CX Linux — AI-powered Linux OS. Natural language system administration for Ubuntu & Debian. The AI layer for Linux infrastructure.
TME: Structured memory engine for LLM agents to plan, rollback, and reason across multi-step tasks.
Zero-code LLM security & observability proxy. Real-time prompt injection detection, PII scanning, and cost control for OpenAI-compatible APIs. Built in Rust.
GPU-aware inference mesh for large-scale AI serving
Production-ready AI infrastructure: RAG with smart reindexing, persistent memory, browser automation, and MCP integration. Stop rebuilding tools for every AI project.
Kubernetes operator for GPU-accelerated LLM inference - air-gapped, edge-native, production-ready
AI SRE tools for RCA, Incident Response, Cost-Saving, Infra management, DevOps and more
ARF is an agentic reliability intelligence platform that separates decision intelligence (OSS) from governed execution (Enterprise), enabling autonomous operations with deterministic safety guarantees.
UniRobot is an embodied intelligent software framework that integrates the robot brain (data, models, model training) with the robot body (perception, model inference, control).
Open-source AI Gateway written in Go, one API for OpenAI, Anthropic, Bedrock, Azure, and 100+ LLMs. Built-in caching, guardrails, retries, and cost optimization. Run as a proxy or embed as a library.
AI Infrastructure Junior Engineer Learning Track - Comprehensive curriculum for entry-level ML infrastructure engineers (0-2 years experience)
A production-grade LLM gateway that abstracts multiple model providers, implements intelligent routing, caching, retries, and observability to deliver reliable, cost-aware LLM access.
Lightfast surfaces every decision your team makes across your tools — searchable, cited, and ready for people and agents.
Orchestrator for shared AI agents and documentation synchronization across multi-repo organizations.
This repository contains a list of various service-specific Azure Landing Zone implementation options.