Minicpm: Unveiling the potential of small language models with scalable training strategies
The burgeoning interest in developing Large Language Models (LLMs) with up to trillion
parameters has been met with concerns regarding resource efficiency and practical
expense, particularly given the immense cost of experimentation. This scenario underscores
the importance of exploring the potential of Small Language Models (SLMs) as a resource-
efficient alternative. In this context, we introduce MiniCPM, specifically the 1.2 B and 2.4 B
non-embedding parameter variants, not only excel in their respective categories but also …
parameters has been met with concerns regarding resource efficiency and practical
expense, particularly given the immense cost of experimentation. This scenario underscores
the importance of exploring the potential of Small Language Models (SLMs) as a resource-
efficient alternative. In this context, we introduce MiniCPM, specifically the 1.2 B and 2.4 B
non-embedding parameter variants, not only excel in their respective categories but also …
[CITATION][C] Minicpm: Unveiling the potential of small language models with scalable training strategies. 2024
S Hu, Y Tu, X Han, C He, G Cui, X Long, Z Zheng… - URL https://doi. org/10.48550/arXiv
Showing the best results for this search. See all results