Lists (22)
Sort Name ascending (A-Z)
ASR
Assistant
bitnet
ctx_sz
Curation
Diffusion
Embedding
Finanace
Framework
Inference
Integrated-AI
LLM
LoRA
Multi-modal
Not transformer
On-Device
Paper Curation
prompt
Quantization
RAG
tools
TTS
Stars
📰 Must-read papers and blogs on LLM based Long Context Modeling 🔥
An open-source RAG-based tool for chatting with your documents.
prime is a framework for efficient, globally distributed training of AI models over the internet.
[TMLR 2024] Efficient Large Language Models: A Survey
A curated list for Efficient Large Language Models
Agent Laboratory is an end-to-end autonomous research workflow meant to assist you as the human researcher toward implementing your research ideas
A self-hostable bookmark-everything app (links, notes and images) with AI-based automatic tagging and full text search
PDF scientific paper translation with preserved formats - 基于 AI 完整保留排版的 PDF 文档全文双语翻译,支持 Google/DeepL/Ollama/OpenAI 等服务,提供 CLI/GUI/Docker/Zotero
A generative world for general-purpose robotics & embodied AI learning.
Python tool for converting files and office documents to Markdown.
Kheish: A multi-role LLM agent for tasks like code auditing, file searching, and more—seamlessly leveraging RAG and extensible modules.
Code for studying the super weight in LLM
arXiv 논문을 자동으로 한국어로 번역하고, LaTeX 소스에서 PDF 파일을 생성하는 도구입니다. OpenAI GPT API를 사용하여 번역을 수행합니다.
A proxy server for multiple ollama instances with Key security
Chrome Extension to Summarize or Chat with Web Pages/Local Documents Using locally running LLMs. Keep all of your data and conversations private. 🔐
Tensor parallelism is all you need. Run LLMs on an AI cluster at home using any device. Distribute the workload, divide RAM usage, and increase inference speed.
Official repository of "SAMURAI: Adapting Segment Anything Model for Zero-Shot Visual Tracking with Motion-Aware Memory"
Swift Package to implement a transformers-like API in Swift
swyxio / bolt.talk
Forked from stackblitz-labs/bolt.diyPrompt, run, edit, and deploy full-stack web applications using any LLM you want!
Best practices for distilling large language models.
MobileLLM Optimizing Sub-billion Parameter Language Models for On-Device Use Cases. In ICML 2024.
A blazing fast inference solution for text embeddings models