We used the same dataset as
eques/jpharmatron
for training our JpharmaBERT, which consists of:
Japanese text data (2B tokens) collected from pharmaceutical documents such as academic papers and package inserts
English data (8B tokens) obtained from PubMed abstracts
Pharmaceutical-related data (1.2B tokens) extracted from the multilingual CC100 dataset
After removing duplicate entries across these sources, the final dataset contains approximately 9 billion tokens.
(For details, please refer to our paper about Jpharmatron:
link
)
Training Hyperparameters
The model was continually pre-trained with the following settings:
More Information About jpharma-bert-base huggingface.co Model
jpharma-bert-base huggingface.co
jpharma-bert-base huggingface.co is an AI model on huggingface.co that provides jpharma-bert-base's model effect (), which can be used instantly with this EQUES jpharma-bert-base model. huggingface.co supports a free trial of the jpharma-bert-base model, and also provides paid use of the jpharma-bert-base. Support call jpharma-bert-base model through api, including Node.js, Python, http.
jpharma-bert-base huggingface.co is an online trial and call api platform, which integrates jpharma-bert-base's modeling effects, including api services, and provides a free online trial of jpharma-bert-base, you can try jpharma-bert-base online for free by clicking the link below.
EQUES jpharma-bert-base online free url in huggingface.co:
jpharma-bert-base is an open source model from GitHub that offers a free installation service, and any user can find jpharma-bert-base on GitHub to install. At the same time, huggingface.co provides the effect of jpharma-bert-base install, users can directly use jpharma-bert-base installed effect in huggingface.co for debugging and trial. It also supports api for free installation.