-
Millennium Science School
- Beijing, China
-
23:26
(UTC +08:00) - @llamafactory_ai
- https://huggingface.co/hiyouga
Lists (2)
Sort Name ascending (A-Z)
Starred repositories
An open-source AI agent that brings the power of Gemini directly into your terminal.
qwen-code is a coding agent that lives in digital world.
Text-audio foundation model from Boson AI
[ICLR 2025] Repository for Show-o series, One Single Transformer to Unify Multimodal Understanding and Generation.
Kimi K2 is the large language model series developed by Moonshot AI team
Skywork-R1V is an advanced multimodal AI model series developed by Skywork AI (Kunlun Inc.), specializing in vision-language reasoning.
What are the principles we can use to build LLM-powered software that is actually good enough to put in the hands of production customers?
GraphGen: Enhancing Supervised Fine-Tuning for LLMs with Knowledge-Driven Synthetic Data Generation
Fused Qwen3 MoE layer for faster training, compatible with HF Transformers, LoRA, 4-bit quant, Unsloth
⚙️🗑️ A GitHub Action to free disk space on an Ubuntu runner.
GLM-4.1V-Thinking: Towards Versatile Multimodal Reasoning with Scalable Reinforcement Learning.
slime is a LLM post-training framework aiming for RL Scaling.
Bridge Megatron-Core to Hugging Face/Reinforcement Learning
The official Python SDK for Model Context Protocol servers and clients
Visual Planning: Let's Think Only with Images
patches for huggingface transformers to save memory
🚀 Efficient implementations of state-of-the-art linear attention models
🚀ReVisual-R1 is a 7B open-source multimodal language model that follows a three-stage curriculum—cold-start pre-training, multimodal reinforcement learning, and text-only reinforcement learning—to …
Reverse Engineering Gemma 3n: Google's New Edge-Optimized Language Model
Get started with building Fullstack Agents using Gemini 2.5 and LangGraph
🚀 The fast, Pythonic way to build MCP servers and clients
[ICCV 2025] Scaling Inference-Time Optimization for Text-to-Image Diffusion Models via Reflection Tuning
Production ready LLM model compression/quantization toolkit with hw accelerated inference support for both cpu/gpu via HF, vLLM, and SGLang.