RakutenAI-7B: Extending Large Language Models for Japanese

Rakuten Group, null, Levine, Aaron, Huang, Connie, Wang, Chenguang, Batista, Eduardo, Szymanska, Ewa, Ding, Hongyi, Chou, Hou Wei, Pessiot, Jean-François, Effendi, Johanes, Chiu, Justin, Ohlhus, Kai Torben, Chopra, Karan, Shinzato, Keiji, Murakami, Koji, Xiong, Lee, Chen, Lei, Kubota, Maki, Tkachenko, Maksim, Lee, Miroku, Takahashi, Naoki, Jwalapuram, Prathyusha, Tatsushima, Ryutaro, Jain, Saurabh, Yadav, Sunil Kumar, Cai, Ting, Chen, Wei-Te, Xia, Yandi, Nakayama, Yuki, Higashiyama, Yutaka

arXiv.org Artificial Intelligence 

We introduce RakutenAI-7B, a suite of Japanese-oriented large language models that achieve the best performance on the Japanese LM Harness benchmarks among the open 7B models. Along with the foundation model, we release instruction- and chat-tuned models, RakutenAI-7B-instruct and RakutenAI-7B-chat respectively, under the Apache 2.0 license.

Duplicate Docs Excel Report

Title
None found

Similar Docs  Excel Report  more

TitleSimilaritySource
None found