-
Meta
-
16:56
(UTC -08:00) - wuyujack.github.io
Highlights
- Pro
Stars
The official repo of Qwen (通义千问) chat & pretrained large language model proposed by Alibaba Cloud.
Lightweight coding agent that runs in your terminal
ModelScope: bring the notion of Model-as-a-Service to life.
Analyzing Hacker News discussions from a decade ago in hindsight with LLMs
An Open-source RL System from ByteDance Seed and Tsinghua AIR
MCP-Bench: Benchmarking Tool-Using LLM Agents with Complex Real-World Tasks via MCP Servers
Lumina-DiMOO - An Open-Sourced Multi-Modal Large Diffusion Language Model
Official PyTorch implementation for "Large Language Diffusion Models"
MMaDA - Open-Sourced Multimodal Large Diffusion Language Models
Official Implementation of "MMaDA-Parallel: Multimodal Large Diffusion Language Models for Thinking-Aware Editing and Generation"
Cosmos-Reason1 models understand the physical common sense and generate appropriate embodied decisions in natural language through long chain-of-thought reasoning processes.
GLM-4.6V/4.5V/4.1V-Thinking: Towards Versatile Multimodal Reasoning with Scalable Reinforcement Learning
Agent framework and applications built upon Qwen>=3.0, featuring Function Calling, MCP, Code Interpreter, RAG, Chrome extension, etc.
An open-source AI agent that lives in your terminal.
🚀🚀 Efficient implementations of Native Sparse Attention
[CoRL 2025] TWIST: Teleoperated Whole-Body Imitation System
Official repository for the Boltz biomolecular interaction models
NEO Series: Native Vision-Language Models from First Principles
The simplest, fastest repository for training/finetuning small-sized VLMs.
Post-training with Tinker
Qwen3-omni is a natively end-to-end, omni-modal LLM developed by the Qwen team at Alibaba Cloud, capable of understanding text, audio, images, and video, as well as generating speech in real time.
[NeurIPS 2025]《SD-VLM: Spatial Measuring and Understanding with Depth-encoded Vision Language Models》

