pfnet / plamo-100b

huggingface.co
Total runs: 83
24-hour runs: 0
7-day runs: -21
30-day runs: -30
Model's Last Updated: October 15 2024
text-generation

Introduction of plamo-100b

Model Details of plamo-100b

PLaMo-100B

Model Description

PLaMo-100B is a 100B model pre-trained on English and Japanese open datasets, developed by Preferred Elements, Inc. PLaMo-100B is released under both Commercial and Non-Commercial Licenses. Please check the LICENSE for non-commercial use, both Japanese version and English version of the license are available. For commercial use, please contact us via this form (Japanese Only).

NOTE : This model has NOT been instruction-tuned for chat dialog or other downstream tasks. We provide instruction-tuned version of PLaMo-100B model via our API and solution packages. Please check our official PLaMo website (Japanese only) for details.

Usage
Requirements
  • numpy
  • sentencepiece
  • torch
  • transformers
Use a pipeline as a high-level helper
import transformers
pipeline = transformers.pipeline("text-generation", model="pfnet/plamo-100b", trust_remote_code=True)
print(pipeline("The future of artificial intelligence technology is ", max_new_tokens=32))
Load model directly
from transformers import AutoTokenizer, AutoModelForCausalLM
tokenizer = AutoTokenizer.from_pretrained("pfnet/plamo-100b", trust_remote_code=True)
model = AutoModelForCausalLM.from_pretrained("pfnet/plamo-100b", trust_remote_code=True)
text = "これからの人工知能技術は"
input_ids = tokenizer(text, return_tensors="pt").input_ids
generated_tokens = model.generate(
    inputs=input_ids,
    max_new_tokens=32,
    do_sample=True,
    top_k=50,
    top_p=0.95,
    temperature=1.0,
)[0]
generated_text = tokenizer.decode(generated_tokens)
print(generated_text)
Model Details
  • Model size: 100B
  • Trained tokens: 2T tokens (English: 1.3T tokens, Japanese: 0.7T tokens)
  • Developed by: Preferred Elements, Inc
  • Model type: Causal decoder-only
  • Language(s): English, Japanese
  • License: Commercial and Non-Commercial
Training Dataset

We trained PLaMo-100B in two phases, phase 1 with 1.5T tokens and phase 2 with 0.5T tokens. The percentage of datasets in each phase is shown in the following table.

1.5T (phase 1) 0.5T (phase 2)
RefinedWeb (English) 42% 17%
Other English Dataset 28% 33%
Proprietary CommonCrawl-JP 18% 46%
Other Japanese Dataset 12% 4%
Tokenizer

PLaMo-100B uses sentencepiece tokenizer which is trained on a subset of the datasets for model pre-training.

Tech Blog

https://tech.preferred.jp/ja/blog/plamo-100b/

Bias, Risks, and Limitations

PLaMo-100B is a new technology that carries risks with use. Testing conducted to date has been in English and Japanese, and has not covered, nor could it cover all scenarios. For these reasons, as with all LLMs, PLaMo-100B’s potential outputs cannot be predicted in advance, and the model may in some instances produce inaccurate, biased or other objectionable responses to user prompts. Therefore, before deploying any applications of PLaMo-100B, developers should perform safety testing and tuning tailored to their specific applications of the model.

How to cite
@article{plamo100b,
    author    = {Kenshin Abe, Kaizaburo Chubachi, Yasuhiro Fujita, Yuta Hirokawa, Kentaro Imajo, Toshiki Kataoka, Hiroyoshi Komatsu, Hiroaki Mikami, Tsuguo Mogami, Shogo Murai, Kosuke Nakago, Daisuke Nishino, Toru Ogawa, Daisuke Okanohara, Yoshihiko Ozaki, Shotaro Sano, Shuji Suzuki, Tianqi Xu, Toshihiko Yanase (Preferred Elements, Inc.)},
    title     = {PLaMo-100B: A Ground-Up Language Model Designed for Japanese Proficiency},
    year      = {2024},
    url       = {https://arxiv.org/abs/2410.07563},
    journal   = {arXiv}
}
Acknowledgement

This model is trained under the project, “Research and Development Project of the Enhanced Infrastructures for Post 5G Information and Communication System” (JPNP 20017), subsidized by the New Energy and Industrial Technology Development Organization (NEDO).

Runs of pfnet plamo-100b on huggingface.co

83
Total runs
0
24-hour runs
-6
3-day runs
-21
7-day runs
-30
30-day runs

More Information About plamo-100b huggingface.co Model

plamo-100b huggingface.co

plamo-100b huggingface.co is an AI model on huggingface.co that provides plamo-100b's model effect (), which can be used instantly with this pfnet plamo-100b model. huggingface.co supports a free trial of the plamo-100b model, and also provides paid use of the plamo-100b. Support call plamo-100b model through api, including Node.js, Python, http.

plamo-100b huggingface.co Url

https://huggingface.co/pfnet/plamo-100b

pfnet plamo-100b online free

plamo-100b huggingface.co is an online trial and call api platform, which integrates plamo-100b's modeling effects, including api services, and provides a free online trial of plamo-100b, you can try plamo-100b online for free by clicking the link below.

pfnet plamo-100b online free url in huggingface.co:

https://huggingface.co/pfnet/plamo-100b

plamo-100b install

plamo-100b is an open source model from GitHub that offers a free installation service, and any user can find plamo-100b on GitHub to install. At the same time, huggingface.co provides the effect of plamo-100b install, users can directly use plamo-100b installed effect in huggingface.co for debugging and trial. It also supports api for free installation.

plamo-100b install url in huggingface.co:

https://huggingface.co/pfnet/plamo-100b

Url of plamo-100b

plamo-100b huggingface.co Url

Provider of plamo-100b huggingface.co

pfnet
ORGANIZATIONS

Other API from pfnet

huggingface.co

Total runs: 54.3K
Run Growth: 4.2K
Growth Rate: 7.71%
Updated:November 07 2025
huggingface.co

Total runs: 152
Run Growth: 27
Growth Rate: 17.76%
Updated:October 10 2023
huggingface.co

Total runs: 22
Run Growth: -58
Growth Rate: -263.64%
Updated:October 31 2025
huggingface.co

Total runs: 0
Run Growth: 0
Growth Rate: 0.00%
Updated:December 19 2024