Lang ZH

SMALL SIZE, SUPER POWER

Edge Model for Everyone, Everyday, Everywhere
MiniCPM InsidePhones
MiniCPM InsideAIPC
MiniCPM InsideIntelligent Cabins
MiniCPM InsideEmbodied Robots
MiniCPM InsideWearable Devices

Put ChatGPT, GPT-4V Level LLMs on Your Phone, Pad and PC

Learn More

The 'MiniCPM' edge model series is a world-leading, lightweight, and high-performance LLM. Since its release in February 2024, it has been widely tested and acclaimed by the global open-source community for its "achieving more with less" efficiency and outstanding on-device performance. It has repeatedly topped GitHub and Hugging Face trending charts, becoming one of the most popular LLMs on Hugging Face in 2024. The 'MiniCPM' has partnered with industry benchmark leaders, emerging as an indispensable player in driving innovation across sectors such as AIPC, AI phones, intelligent cabins, and embodied robots.

High Efficiency, Low Cost, Achieving More with LessFundation Model MiniCPM
4B2.4B1.2B
The On-Device ChatGPT Moment
4B 2.4B 1.2B
githubhuggingface
Unbelievably Strong for 4B size edge Model on your device!
ChatGPT-level Basic Performance Surpassing GPT-3.5, Qwen2-7B, GLM4-9B


New Architecture, New Benchmark of LLM Knowledge Density

Light! Fast! On-Device Friendly
Only 2GB of memory after quantization
Versatile and Sharp as a Swiss Army KnifeSurpassing Kimi! Infinite Long Text
32, 128, 256, 512K... Unlimited Context Expansion

GPT-4o-level Function Calling
Surpassing GPT-3.5, GLM4-9B, Close to GPT-4o

Superior RAG External Attachment Set Number One in Chinese Retrieval, Results Generation Surpassing Llama3-8B
Learn More
View the detailed features of each version
GPT-4o level Omni Model runs on deviceMultimodal Model MiniCPM-V
8B Full-Modal8B Live Video8B2.8B
The On-Device GPT-4o New Era
8B Full-Modal 8B Live Video 8B 2.8B
githubhuggingface
Edge-Side GPT-4oReal-time streaming, end-to-end Full-modal, all SOTA The best edge visual general model The best audio general model
Continuous watching, real videos Not just a single frame-based model Real-time listening, truly smooth Hear clearly, understand distinctly Natural speaking, emotional engagement Real-time interruptions without confusionFull Capability, End-to-EndHigh performance, low latency More natural, more coherent Context understanding Interruptible at any time Noise resistance Easy deployment and maintenanceLearn More
View the detailed features of each version
Compare the functionalities of various versions

Global Partner

amd
amd
amd
amd
amd
amd
amd
amd
amd
amd
amd
amd
amd
amd
amd
amd
amd
amd
amd
amd
amd
amd
amd
Technical Blog

Efficiency FirstWe believe the best model is the one with superior power, faster speed and lower costEfficiency comes from mastering the science of large language models (LLMs), with knowledge density as the key principle. As knowledge density grows, it becomes a core competitive advantage, unlocking vast potential for edge intelligence and applications.
Modelbest LawMoore’s Law
The capability density of LLMs increases exponentially over time. Since 2023, the maximum capability density of LLMs doubles approximately every 3.3 months.Capability density: The ratio of effective parameter size to actual parameter size. Effective parameter size refers to the minimum number of parameters required for the reference model (e.g., MiniCPM) to achieve performance equivalent to the given target model.
News

A G I  F O R  L I V E S