-
Fudan University
- Shanghai, China
- bug-orz.github.io
Lists (1)
Sort Name ascending (A-Z)
Stars
🚀 「大模型」1小时从0训练26M参数的视觉多模态VLM!🌏 Train a 26M-parameter VLM from scratch in just 1 hours!
《大模型白盒子构建指南》:一个全手搓的Tiny-Universe
a toolkit on knowledge distillation for large language models
[CVPR2025] Official implementation of "Decouple-Then-Merge: Finetune Diffusion Models as Multi-Task Learning"
[ACL2025 main] Official implementation of "LED-Merging: Mitigating Safety-Utility Conflicts in Model Merging with Location-Election-Disjoint"
Generating sets of formulaic alpha (predictive) stock factors via reinforcement learning.
Do Large Language Models Know What They Don’t Know?
[ACL'24] Selective Reflection-Tuning: Student-Selected Data Recycling for LLM Instruction-Tuning
An Easy-to-use, Scalable and High-performance RLHF Framework based on Ray (PPO & GRPO & REINFORCE++ & vLLM & Ray & Dynamic Sampling & Async Agentic RL)
MNBVC(Massive Never-ending BT Vast Chinese corpus)超大规模中文语料集。对标chatGPT训练的40T数据。MNBVC数据集不但包括主流文化,也包括各个小众文化甚至火星文的数据。MNBVC数据集包括新闻、作文、小说、书籍、杂志、论文、台词、帖子、wiki、古诗、歌词、商品介绍、笑话、糗事、聊天记录等一切形式的纯文本中文数据。
Official code for "MAmmoTH2: Scaling Instructions from the Web" [NeurIPS 2024]
A comprehensive library for implementing LLMs, including a unified training pipeline and comprehensive model evaluation.
中文对话0.2B小模型(ChatLM-Chinese-0.2B),开源所有数据集来源、数据清洗、tokenizer训练、模型预训练、SFT指令微调、RLHF优化等流程的全部代码。支持下游任务sft微调,给出三元组信息抽取微调示例。
用于从头预训练+SFT一个小参数量的中文LLaMa2的仓库;24G单卡即可运行得到一个具备简单中文问答能力的chat-llama2.
SuperCLUE: 中文通用大模型综合性基准 | A Benchmark for Foundation Models in Chinese
A framework for few-shot evaluation of language models.
从0到1构建一个MiniLLM (pretrain+sft+dpo实践中)
Qwen3 is the large language model series developed by Qwen team, Alibaba Cloud.
🚀 大语言模型高效转发服务 · An efficient forwarding service designed for LLMs. · OpenAI API Reverse Proxy
Unified Efficient Fine-Tuning of 100+ LLMs & VLMs (ACL 2024)
The official repo of Qwen (通义千问) chat & pretrained large language model proposed by Alibaba Cloud.
The repository for paper <Evaluating Open-QA Evaluation>
The repository for the survey paper <<Survey on Large Language Models Factuality: Knowledge, Retrieval and Domain-Specificity>>
Use ChatGPT to summarize the arXiv papers. 全流程加速科研,利用chatgpt进行论文全文总结+专业翻译+润色+审稿+审稿回复
Large-scale Self-supervised Pre-training Across Tasks, Languages, and Modalities
Self-study on Larry Wasserman's "All of Statistics"