Highlights
- Pro
Train_Tune 🔢
Unified Efficient Fine-Tuning of 100+ LLMs & VLMs (ACL 2024)
📘 The experiment tracker for foundation model training
Debug, evaluate, and monitor your LLM applications, RAG systems, and agentic workflows with comprehensive tracing, automated evaluations, and production-ready dashboards.
🤗 PEFT: State-of-the-art Parameter-Efficient Fine-Tuning.
Finetune Llama 3.3, DeepSeek-R1 & Reasoning LLMs 2x faster with 70% less memory! 🦥
DeepSpeed is a deep learning optimization library that makes distributed training and inference easy, efficient, and effective.
🚀 A simple way to launch, train, and use PyTorch models on almost any device and distributed configuration, automatic mixed precision (including fp8), and easy-to-configure FSDP and DeepSpeed support
Easily embed, cluster and semantically label text datasets
[NAACL 2025] KnowAgent: Knowledge-Augmented Planning for LLM-Based Agents
LM Studio JSON configuration file format and a collection of example config files.
Train transformer language models with reinforcement learning.
🤖 The free, Open Source alternative to OpenAI, Claude and others. Self-hosted and local-first. Drop-in replacement for OpenAI, running on consumer-grade hardware. No GPU required. Runs gguf, transf…
A connector for SingleStore and Spark
Get up and running with Llama 3.3, DeepSeek-R1, Phi-4, Gemma 2, and other large language models.
OpenUI let's you describe UI using your imagination, then see it rendered live.
Stanford NLP Python library for Representation Finetuning (ReFT)
Stanford NLP Python library for understanding and improving PyTorch models via interventions
Inference-Time Intervention: Eliciting Truthful Answers from a Language Model
Devika is an Agentic AI Software Engineer that can understand high-level human instructions, break them down into steps, research relevant information, and write code to achieve the given objective…
20+ high-performance LLMs with recipes to pretrain, finetune and deploy at scale.
Pretrain, finetune ANY AI model of ANY size on multiple GPUs, TPUs with zero code changes.
Development repository for the Triton language and compiler
A repo for distributed training of language models with Reinforcement Learning via Human Feedback (RLHF)