Efficient LLM inference on Slurm clusters.
-
Updated
Jan 22, 2026 - Python
Efficient LLM inference on Slurm clusters.
A practical, multi-layered JSON repair library for Elixir that intelligently fixes malformed JSON strings commonly produced by LLMs, legacy systems, and data pipelines.
PipelineLLM 是一个系统性的大语言模型(LLM)后训练学习项目,涵盖从监督微调(SFT)到偏好优化(DPO)、强化学习(RLHF/PPO/GRPO)再到持续学习(Continual Learning)的完整技术栈。
A lightweight Bun + Express template that connects to the Testune AI API and streams chat responses in real time using Server-Sent Events (SSE)
Full-featured Elixir client for the Model Context Protocol (MCP) with multi-transport support, resources, prompts, tools, and telemetry.
A Compute-Agnostic, WebSocket-first protocol for AI Agents. The high-performance alternative to MCP. Runs on Serverless or stateful servers with sub-30ms latency.
【非结构化数据pipeline】目标是自动化原始数据—>特定信息提取。first_example:收集任何的文档将其可视化为思维导图(进度1/3)
A production-ready, enterprise-grade Agentic RAG ingestion pipeline built with n8n, Supabase (pgvector), and AI embeddings. Implements event-driven orchestration, hybrid RAG for structured and unstructured data, vector similarity search, and multi-tenant architecture to deliver client-isolated, retrieval-ready knowledge bases.
MindScript Ledger is the temporal memory and pattern-recall layer of the MindScript ecosystem. It stores normalized prompts, logic states, patterns, and threads as a structured ledger, enabling deterministic recall and reconstruction of long-running AI interactions.
Enterprise-grade Sovereign AI Stack optimized for NVIDIA Blackwell (sm_120) & vLLM. Features 256K context window, 5.8k tok/s prefill, and integrated observability via Langfuse.
⚡ Streamline AI agent communication with Agent Socket, a high-performance WebSocket protocol designed for diverse computing environments.
A production-ready, enterprise-grade Agentic RAG ingestion pipeline built with n8n, Supabase (pgvector), and AI embeddings. Implements event-driven orchestration, hybrid RAG for structured and unstructured data, vector similarity search, and multi-tenant architecture to deliver client-isolated, retrieval-ready knowledge bases.
Add a description, image, and links to the llm-infrastructure topic page so that developers can more easily learn about it.
To associate your repository with the llm-infrastructure topic, visit your repo's landing page and select "manage topics."