-
12:43
(UTC -05:00)
Highlights
- Pro
Stars
A lightweight LMM-based Document Parsing Model
Python tool for converting files and office documents to Markdown.
FrogAi / FrogPilot
Forked from commaai/openpilotA monthly updated and 100% open-sourced fork of openpilot with clean commits dedicated to serve the openpilot community! FrogPilot is shaped by user and developer contributions, emphasizing collabo…
[COLM'25] DeepRetrieval - 🔥 Training Search Agent with Retrieval Outcomes via Reinforcement Learning
Multi-Joint dynamics with Contact. A general purpose physics simulator.
Official inference framework for 1-bit LLMs
Python Finite State Machines made easy.
Convert PDF to markdown + JSON quickly with high accuracy
aider is AI pair programming in your terminal
📃 A better UX for chat, writing content, and coding with LLMs.
Aruco marker detection with intel realsense camera D435 with less delay on 2DOF pose estimation
OctoTools: An agentic framework with extensible tools for complex reasoning
Igus Rebel controller with ROS2 and MoveIt2: hardware interfaces and commander demos
Official implementation of "ASAP: Aligning Simulation and Real-World Physics for Learning Agile Humanoid Whole-Body Skills"
Official code implementation of General OCR Theory: Towards OCR-2.0 via a Unified End-to-end Model
[ACM'MM 2024 Oral] Official code for "OneChart: Purify the Chart Structural Extraction via One Auxiliary Token"
[ECCV 2024] Official code implementation of Vary: Scaling Up the Vision Vocabulary of Large Vision Language Models.
Fully open reproduction of DeepSeek-R1
ExtractThinker is a Document Intelligence library for LLMs, offering ORM-style interaction for flexible and powerful document workflows.
openpilot is an operating system for robotics. Currently, it upgrades the driver assistance system on 300+ supported cars.
💡 All-in-one open-source AI framework for semantic search, LLM orchestration and language model workflows
An AI memory layer with short- and long-term storage, semantic clustering, and optional memory decay for context-aware applications.
A fast inference library for running LLMs locally on modern consumer-class GPUs
We write your reusable computer vision tools. 💜
Investment Research for Everyone, Everywhere.
MS-Agent: Lightweight Framework for Empowering Agents with Autonomous Exploration
[ICLR 2023] ReAct: Synergizing Reasoning and Acting in Language Models
Accelerate local LLM inference and finetuning (LLaMA, Mistral, ChatGLM, Qwen, DeepSeek, Mixtral, Gemma, Phi, MiniCPM, Qwen-VL, MiniCPM-V, etc.) on Intel XPU (e.g., local PC with iGPU and NPU, discr…