Stars
ModelScope-Agent: An agent framework connecting models in ModelScope with the world
Unify Efficient Fine-tuning of RAG Retrieval, including Embedding, ColBERT, ReRanker.
Supercharge Your LLM Application Evaluations 🚀
Basic data mining model, including feature importance display
本仓库是关于大模型面试中常见面试试题和面试经验的整理。这里收集了各类与大模型相关的面试题目,并提供详细的解答和分析。本仓库由上海交大交影社区维护
一键中文数据增强包 ; NLP数据增强、bert数据增强、EDA:pip install nlpcda
An implement of the paper of EDA for Chinese corpus.中文语料的EDA数据增强工具。NLP数据增强。论文阅读笔记。
Aligning pretrained language models with instruction data generated by themselves.
Open-source Self-Instruction Tuning Code LLM
Qwen2.5 is the large language model series developed by Qwen team, Alibaba Cloud.
20+ high-performance LLMs with recipes to pretrain, finetune and deploy at scale.
Run any open-source LLMs, such as Llama, Mistral, as OpenAI compatible API endpoint in the cloud.
Get up and running with Llama 3.3, Phi 4, Gemma 2, and other large language models.
🦜🔗 Build context-aware reasoning applications
Gorilla: Training and Evaluating LLMs for Function Calls (Tool Calls)
Run Mixtral-8x7B models in Colab or consumer desktops
DeepSeek-V2: A Strong, Economical, and Efficient Mixture-of-Experts Language Model
RankLLM is a Python toolkit for reproducible information retrieval research using rerankers, with a focus on listwise reranking.
PyTorch library for cost-effective, fast and easy serving of MoE models.
An implementation of model parallel autoregressive transformers on GPUs, based on the Megatron and DeepSpeed libraries