Stars
SWE-agent takes a GitHub issue and tries to automatically fix it, using GPT-4, or your LM of choice. It can also be employed for offensive cybersecurity or competitive coding challenges. [NeurIPS 2…
Official codebase used to develop Vision Transformer, SigLIP, MLP-Mixer, LiT and more.
DSPy: The framework for programming—not prompting—language models
[ICLR'24 spotlight] An open platform for training, serving, and evaluating large language model for tool learning.
Simple, safe way to store and distribute tensors
LLM powered development for VSCode
RWKV (pronounced RwaKuv) is an RNN with great LLM performance, which can also be directly trained like a GPT transformer (parallelizable). We are at RWKV-7 "Goose". So it's combining the best of RN…
DeepSpeed is a deep learning optimization library that makes distributed training and inference easy, efficient, and effective.
OpenAssistant is a chat-based assistant that understands tasks, can interact with third-party systems, and retrieve information dynamically to do so.
📺 Discover the latest machine learning / AI courses on YouTube.
DAM Data Acquisition for ML Benchmark, as part of the DataPerf benchmark suite, https://dataperf.org/
Official Open Source code for "Masked Autoencoders As Spatiotemporal Learners"
Code repo for the paper BiT Robustly Binarized Multi-distilled Transformer
Hackable and optimized Transformers building blocks, supporting a composable construction.
🧑🏫 60+ Implementations/tutorials of deep learning papers with side-by-side notes 📝; including transformers (original, xl, switch, feedback, vit, ...), optimizers (adam, adabelief, sophia, ...), ga…
Aim 💫 — An easy-to-use & supercharged open-source experiment tracker.
Ekphrasis is a text processing tool, geared towards text from social networks, such as Twitter or Facebook. Ekphrasis performs tokenization, word normalization, word segmentation (for splitting has…
Code for T-Few from "Few-Shot Parameter-Efficient Fine-Tuning is Better and Cheaper than In-Context Learning"
Code of NAACL 2022 "Efficient Hierarchical Domain Adaptation for Pretrained Language Models" paper.
Handout for the tutorial "Creating publication-quality figures with matplotlib"
Solving reinforcement learning tasks which require language and vision
Google Research