MiniCPM-2B
MiniMind| Launched | |
| Pricing Model | Free |
| Starting Price | |
| Tech used | |
| Tag | Language Learning |
| Launched | |
| Pricing Model | Free |
| Starting Price | |
| Tech used | |
| Tag | Mlops,Developer Tools,Software Development |
| Global Rank | |
| Country | |
| Month Visit |
| Global Rank | |
| Country | |
| Month Visit |
Estimated traffic data from Similarweb
MiniCPM-Llama3-V 2.5 - With a total of 8B parameters, the model surpasses proprietary models such as GPT-4V-1106, Gemini Pro, Qwen-VL-Max and Claude 3 in overall performance.
MiniCPM3-4B - MiniCPM3-4B is the 3rd generation of MiniCPM series. The overall performance of MiniCPM3-4B surpasses Phi-3.5-mini-Instruct and GPT-3.5-Turbo-0125, being comparable with many recent 7B~9B models.
SmolLM - SmolLM is a series of state-of-the-art small language models available in three sizes: 135M, 360M, and 1.7B parameters.
OpenBMB - OpenBMB: Building a large-scale pre-trained language model center and tools to accelerate training, tuning, and inference of big models with over 10 billion parameters. Join our open-source community and bring big models to everyone.