A model trained on a synthesis dataset of over
120 million
entries, this dataset having been generated through the application of state-of-the-art language models utilizing large context windows, alongside methodologies akin to retrieval-augmented generation and knowledge graph integration, where the data synthesis is conducted within clusters derived from a curated pretraining corpus of 20 billion tokens, with subsequent validation performed by the model itself.
Despite the absence of thorough alignment with human preferences, the model is under no obligation to cater to poorly constructed prompts or the clichés often found in conventional benchmarks. Bonus: Included is an implementation of a
Vision Language Model
that has undergone Locked-Image Tuning.
Supported Input Modalities
: text, image
Context Window:
1M tokens
Model Parameters:
LLM - 9B (initialized from THUDM/glm-4-9b-chat-1m); Optional ViT - 5B
Cautionary Notes:
It is strongly recommended to utilize a standardized implementation for inference
, such as Hugging Face Transformers, to avoid the significant performance degradation that might occur when using accelerated kernels like vllm or lmdeploy - not to mention the potentially catastrophic effects of model quantization.
As of now, these accelerated inference implementations are known to severely compromise effective
vision inference, though they have a less pronounced impact on pure text performance.
Inference Parameters:
Our observations suggest that, if one desires to achieve results with fewer hallucinations, it is advisable to employ sampling with top_p=0.8 followed by a temperature setting of 0.3, or alternatively, to use pure temperature sampling with a setting of 0.2.
In general, a lower temperature is required compared to similar models
, which we tentatively attribute to overfitting on the vast dataset.
Regarding Formatting:
We strongly recommend you double-check your input to ensure: 1. The system prompt is not empty. Even something as simple as "You are a helpful assistant." is expected. 2. Each role's content ends with a newline character ('\n') before being concatenated with the <|role|> tag. 3. There is always a newline character after the <|role|> tag. This will help ensure proper parsing and processing of your input.
Regarding
Benchmark Scores
:
Generally, you shouldn't worry too much about them, as people can always train specifically to achieve good results. We mainly use them as a smoke test, a quick check to ensure no major regressions have occurred. In fact, if you actually read through the benchmark questions themselves, you'll often find yourself chuckling at how inane, low-quality, or even downright silly they are.
Regarding training:
The final released version was trained using a merge of multiple candidate models in an attempt to improve performance. However, we were unable to conclusively determine whether this was effective. Excluding candidate versions, an efficient naive fine-tuning should be achievable within one day on 16 nodes of 8*A100-80G. Based on this, we estimate the carbon emissions to be 700 kg CO2 eq.
Disclaimer:
Please note that the model was trained on unfiltered internet data. Since we do not have the capacity to vet all of it, there may be a substantial amount of objectionable content, pornography, violence, and offensive language present that we are unable to remove. Therefore, you will still need to complete your own checks on the model's safety and filter keywords in the output. Due to computational resource constraints, we are presently unable to implement RLHF for the model's ethics and safety, nor training on SFT samples that refuse to answer certain questions for restrictive fine-tuning.
Seeking Unconditional Sponsorship:
We are actively training larger parameter models and scaling up data synthesis, and are seeking substantial compute resources and generous
unconditional
grants. While this is for the purpose of commercial exploration and technology selection, we are currently under no immediate pressure to generate profit and remain committed to sharing more with the open-source community.
miniG huggingface.co is an AI model on huggingface.co that provides miniG's model effect (), which can be used instantly with this CausalLM miniG model. huggingface.co supports a free trial of the miniG model, and also provides paid use of the miniG. Support call miniG model through api, including Node.js, Python, http.
miniG huggingface.co is an online trial and call api platform, which integrates miniG's modeling effects, including api services, and provides a free online trial of miniG, you can try miniG online for free by clicking the link below.
miniG is an open source model from GitHub that offers a free installation service, and any user can find miniG on GitHub to install. At the same time, huggingface.co provides the effect of miniG install, users can directly use miniG installed effect in huggingface.co for debugging and trial. It also supports api for free installation.