Stability AI Beats SoftBank in Releasing Japanese Language Model
Stability AI today released its first Japanese language model (LM), Japanese StableLM Alpha.
The 7 billion-parameter general-purpose language model is currently the only publicly available LM for Japanese speakers.
With this release, Stability has beaten SoftBank to the punch as the latter announced last week that it will be developing homegrown large language models for the Japanese market.
Stability AI, the generative AI company behind Stable Diffusion, today announced the release of its first Japanese Language Model (LM) named Japanese StableLM Alpha, accessible via Hugging Face.
The company claims that the 7 billion-parameter general-purpose language model is currently the only best-performing publicly available LM for Japanese speakers, according to a benchmark suite against four sets of other Japanese LMs.
A commercially available model, the Japanese StableLM Base Alpha 7B, will be released under the Apache License 2.0. The model is trained on 750 billion tokens of Japanese and English text using large scale data sourced from the web.
In addition to open datasets, training data includes datasets created by Stability AI’s Japanese community, in cooperation with the Japanese team of the EleutherAI Polyglot project. Stability AI used an extension of EleutherAI’s GPT-NeoX software to train the Japanese StableLM Base Alpha 7B model.
Another model, the Japanese StableLM Instruct Alpha 7B, is created solely for research purposes and released exclusively for research use. “This model is additionally tuned to follow user instructions, and trained with Supervised Fine-tuning (SFT) using multiple open datasets,” Stability AI tweeted.
Both models were tested using EleutherAI’s Language Model Evaluation Harness on tasks like sentence classification, sentence pair classification, question answering, and sentence summarization, with an average score 54.71%. Stability AI claims that this score puts its Japanese StableLM Instruct Alpha 7B far ahead of other Japanese models.
“We are proud of our first big step towards contributing to the Japanese generative AI ecosystem,” said Meng Lee, Project Lead of Japanese StableLM.
”We look forward to continuing to create models across several modalities, built specifically to reflect Japanese culture, language and aesthetics”.
With the release of its Japanese LM, Stability AI has beaten SoftBank to the punch of releasing language models for the Japanese market. Last Friday, SoftBank announced that it has launched a new company to research and develop homegrown Large Language Models (LLM) for the Japanese market.
Furthermore, SoftBank is allocating around 20 billion JPY (more than $140 million) to its generative AI computing platform, set to launch in the fall of this year. It’s a waiting game to determine whose Japanese Language Model will emerge triumphant in the long run.
Any data, text, or other content on this page is provided as general market information and not as investment advice. Past performance is not necessarily an indicator of future results.
The Trust Project is a worldwide group of news organizations working to establish transparency standards.