PLaMo-100B is a 100B model pre-trained on English and Japanese open datasets, developed by Preferred Elements, Inc.
PLaMo-100B is released under both Commercial and Non-Commercial Licenses. Please check the
LICENSE
for
non-commercial
use, both Japanese version and English version of the license are available. For
commercial
use, please contact us via this
form
(Japanese Only).
NOTE
: This model has
NOT
been instruction-tuned for chat dialog or other downstream tasks. We provide instruction-tuned version of PLaMo-100B model via our API and solution packages. Please check our official PLaMo
website
(Japanese only) for details.
Usage
Requirements
numpy
sentencepiece
torch
transformers
Use a pipeline as a high-level helper
import transformers
pipeline = transformers.pipeline("text-generation", model="pfnet/plamo-100b", trust_remote_code=True)
print(pipeline("The future of artificial intelligence technology is ", max_new_tokens=32))
We trained PLaMo-100B in two phases, phase 1 with 1.5T tokens and phase 2 with 0.5T tokens.
The percentage of datasets in each phase is shown in the following table.
1.5T (phase 1)
0.5T (phase 2)
RefinedWeb (English)
42%
17%
Other English Dataset
28%
33%
Proprietary CommonCrawl-JP
18%
46%
Other Japanese Dataset
12%
4%
Tokenizer
PLaMo-100B uses sentencepiece tokenizer which is trained on a subset of the datasets for model pre-training.
PLaMo-100B is a new technology that carries risks with use. Testing conducted to date has been in English and Japanese, and has not covered, nor could it cover all scenarios. For these reasons, as with all LLMs, PLaMo-100B’s potential outputs cannot be predicted in advance, and the model may in some instances produce inaccurate, biased or other objectionable responses to user prompts. Therefore, before deploying any applications of PLaMo-100B, developers should perform safety testing and tuning tailored to their specific applications of the model.
How to cite
@article{plamo100b,
author = {Kenshin Abe, Kaizaburo Chubachi, Yasuhiro Fujita, Yuta Hirokawa, Kentaro Imajo, Toshiki Kataoka, Hiroyoshi Komatsu, Hiroaki Mikami, Tsuguo Mogami, Shogo Murai, Kosuke Nakago, Daisuke Nishino, Toru Ogawa, Daisuke Okanohara, Yoshihiko Ozaki, Shotaro Sano, Shuji Suzuki, Tianqi Xu, Toshihiko Yanase (Preferred Elements, Inc.)},
title = {PLaMo-100B: A Ground-Up Language Model Designed for Japanese Proficiency},
year = {2024},
url = {https://arxiv.org/abs/2410.07563},
journal = {arXiv}
}
Acknowledgement
This model is trained under the project, “Research and Development Project of the Enhanced Infrastructures for Post 5G Information and Communication System” (JPNP 20017), subsidized by the New Energy and Industrial Technology Development Organization (NEDO).
Runs of pfnet plamo-100b on huggingface.co
83
Total runs
0
24-hour runs
-6
3-day runs
-21
7-day runs
-30
30-day runs
More Information About plamo-100b huggingface.co Model
plamo-100b huggingface.co is an AI model on huggingface.co that provides plamo-100b's model effect (), which can be used instantly with this pfnet plamo-100b model. huggingface.co supports a free trial of the plamo-100b model, and also provides paid use of the plamo-100b. Support call plamo-100b model through api, including Node.js, Python, http.
plamo-100b huggingface.co is an online trial and call api platform, which integrates plamo-100b's modeling effects, including api services, and provides a free online trial of plamo-100b, you can try plamo-100b online for free by clicking the link below.
pfnet plamo-100b online free url in huggingface.co:
plamo-100b is an open source model from GitHub that offers a free installation service, and any user can find plamo-100b on GitHub to install. At the same time, huggingface.co provides the effect of plamo-100b install, users can directly use plamo-100b installed effect in huggingface.co for debugging and trial. It also supports api for free installation.