Stars
Official repository for MATES: Model-Aware Data Selection for Efficient Pretraining with Data Influence Models [NeurIPS 2024]
[NeurIPS 2024 Main Track] Code for the paper titled "Instruction Tuning With Loss Over Instructions"
This repository is the official implementation of DoFIT (NeurIPS 2024).
Repo for Rho-1: Token-level Data Selection & Selective Pretraining of LLMs.
An Easy-to-use, Scalable and High-performance RLHF Framework (70B+ PPO Full Tuning & Iterative DPO & LoRA & RingAttention & RFT)
Code for the paper "Language Models are Unsupervised Multitask Learners"
I built a gpt-2 style tokenizer that can be trained on any .txt data to generate tokens
TyDi QA contains 200k human-annotated question-answer pairs in 11 Typologically Diverse languages, written without seeing the answer and without the use of translation, and is designed for the trai…
This is an NVIDIA AI Workbench example project that demonstrates an end-to-end model development workflow using Llamafactory.
[ICML 2024] LESS: Selecting Influential Data for Targeted Instruction Tuning
Generates and optimizes Haiku system and user prompts for classification
This repo includes ChatGPT prompt curation to use ChatGPT better.
A method for calculating scaling laws for LLMs from publicly available models
minimal LLM scripts for 24GB VRAM GPUs. training, inference, whatever
MiniCPM3-4B: An edge-side LLM that surpasses GPT-3.5-Turbo.
The Open Cookbook for Top-Tier Code Large Language Model
Reading list of Instruction-tuning. A trend starts from Natrural-Instruction (ACL 2022), FLAN (ICLR 2022) and T0 (ICLR 2022).
DSPy: The framework for programming—not prompting—language models
A framework for prompt tuning using Intent-based Prompt Calibration
[ICML 2024] TrustLLM: Trustworthiness in Large Language Models
This is the code repo for our paper "RAG-DDR: Optimizing Retrieval-Augmented Generation Using Differentiable Data Rewards".
Evaluating the Ripple Effects of Knowledge Editing in Language Models
Demystifying Verbatim Memorization in Large Language Models