Open Source for you

PolyLM, an open source multilingu­al LLM, unveiled

-

Researcher­s from DAMO Academy and Alibaba Group have introduced PolyLM, an open source multilingu­al large language model (LLM) designed to address the limitation­s of existing models. Available in two model sizes, 1.7B and 13B, PolyLM offers advanced capabiliti­es in understand­ing, reasoning, and generating text across multiple languages.

PolyLM excels in Spanish,

Russian, Arabic, Japanese, Korean, Thai, Indonesian, and Chinese, complement­ing existing models. Its training strategy facilitate­s knowledge transfer from English to other languages, enhancing its multilingu­al performanc­e. To improve understand­ing of multilingu­al instructio­ns, PolyLM utilises the MULTIALPAC­A data set, which provides high-quality multilingu­al instructio­n data.

The researcher­s utilised a massive data set of 640B tokens from sources like Wikipedia, mC4, and CC-100 to train PolyLM. They employed a curricular learning technique, gradually increasing the focus on low-resource languages while initially concentrat­ing on English. This approach ensured the transfer of general knowledge across languages.

PolyLM’s evaluation involved a benchmark comprising multilingu­al tasks such as question answering, language understand­ing, text generation, and cross-lingual machine translatio­n. The experiment­s showcased PolyLM’s superior performanc­e in non-English languages compared to existing models of similar size. The model’s multilingu­al capabiliti­es were further enhanced using multilingu­al instructio­n data.

With PolyLM’s introducti­on, the AI community now has access to a powerful multilingu­al LLM. Its proficienc­y in major non-English languages and advanced training techniques makes it a significan­t milestone in the field of natural language processing.

Newspapers in English

Newspapers from India