PolyLM

6 comments
PolyLM is a multilingual large language model designed to address the gaps and limitations in curren0
Visit website

What is PolyLM?

PolyLM is a polyglot large language model, which is aimed to address the following blanks and limitations in current LLM research, offering a comprehensive and innovative solution to advance this field.


Covering 18 of the most commonly spoken languages. PolyLM is proficient in the major non-English languages spoken worldwide, such as Spanish, Russian, Arabic, Japanese, Korean, Thai, Indonesian, and Chinese etc. It is a perfect complement to the existing open-source models, including: (1) LLaMA, in which English is predominant among the whole dataset. (2) BLOOM, fails to address languages spoken by significant populations, such as Japanese, Korean and Thai.

Better multingual instruction-following capability. We propose MULTIALPACA to complement ALPACA and CHINESEALPACA, making LLMs better follow multilingual instructions, particularly those coming from non-native English speakers.

Strong performance. In comparison with popular multilingual LLMs of the similar model size, PolyLM demonstrates remarkable performance on various tasks, including QA, understanding, and generation.


More information on PolyLM

Launched
2023
Pricing Model
Free
Starting Price
Global Rank
Country
Month Visit
<5k
Tech used
PolyLM was manually vetted by our editorial team and was first featured on September 4th 2024.
Aitoolnet Featured banner

PolyLM Alternatives

Load more Alternatives
  1. Enhance language models, improve performance, and get accurate results. WizardLM is the ultimate tool for coding, math, and NLP tasks.

  2. A high-throughput and memory-efficient inference and serving engine for LLMs

  3. MiniCPM is an End-Side LLM developed by ModelBest Inc. and TsinghuaNLP, with only 2.4B parameters excluding embeddings (2.7B in total).

  4. Discover PaLM 2, Google's advanced language model for reasoning, translation, and coding tasks. Built with responsible AI practices, PaLM 2 excels in multilingual collaboration and specialized code generation.

  5. GLM-130B: An Open Bilingual Pre-Trained Model (ICLR 2023)