These models are trained on 🤗
SKGInstruct Dataset
, an instruction-tuning dataset containing mixture of 19 SKG tasks combined with 🤗
SlimOrca
. Check out the dataset card for more details.
Training Procedure
The models are fine-tuned with CodeLlama-Instruct-hf models as base models. Each model is trained for 3 epochs, and the best checkpoint is selected.
For this 7B model, the prompt format (different from 13B, 34B) is
[INST] <<SYS>>
You are an AI assistant that specializes in analyzing and reasoning over structured information. You will be given a task, optionally with some structured knowledge input. Your answer must strictly adhere to the output format, if specified.
<</SYS>>
{instruction} [/INST]
To see concrete examples of this linearization, you can directly reference the 🤗
SKGInstruct Dataset
(coming soon).
We will provide code for linearizing this data shortly.
[INST] <<SYS>>
You are an AI assistant that specializes in analyzing and reasoning over structured information. You will be given a task, optionally with some structured knowledge input. Your answer must strictly adhere to the output format, if specified.
<</SYS>>
Use the information in the following table to solve the problem, choose between the choices if they are provided. table:
col : day | kilometers row 1 : tuesday | 0 row 2 : wednesday | 0 row 3 : thursday | 4 row 4 : friday | 0 row 5 : saturday | 0
question:
Allie kept track of how many kilometers she walked during the past 5 days. What is the range of the numbers? [/INST]
Intended Uses
These models are trained for research purposes. They are designed to be proficient in interpreting linearized structured input. Downstream uses can potentially include various applications requiring the interpretation of structured data.
Limitations
While we've tried to build an SKG-specialized model capable of generalizing, we have shown that this is a challenging domain, and it may lack performance characteristics that allow it to be directly used in chat or other applications.
Citation
If you use the models, data, or code from this project, please cite the original paper:
@misc{zhuang2024structlm,
title={StructLM: Towards Building Generalist Models for Structured Knowledge Grounding},
author={Alex Zhuang and Ge Zhang and Tianyu Zheng and Xinrun Du and Junjie Wang and Weiming Ren and Stephen W. Huang and Jie Fu and Xiang Yue and Wenhu Chen},
year={2024},
eprint={2402.16671},
archivePrefix={arXiv},
primaryClass={cs.CL}
}
Runs of TIGER-Lab StructLM-7B on huggingface.co
29
Total runs
0
24-hour runs
-1
3-day runs
0
7-day runs
9
30-day runs
More Information About StructLM-7B huggingface.co Model
StructLM-7B huggingface.co is an AI model on huggingface.co that provides StructLM-7B's model effect (), which can be used instantly with this TIGER-Lab StructLM-7B model. huggingface.co supports a free trial of the StructLM-7B model, and also provides paid use of the StructLM-7B. Support call StructLM-7B model through api, including Node.js, Python, http.
StructLM-7B huggingface.co is an online trial and call api platform, which integrates StructLM-7B's modeling effects, including api services, and provides a free online trial of StructLM-7B, you can try StructLM-7B online for free by clicking the link below.
TIGER-Lab StructLM-7B online free url in huggingface.co:
StructLM-7B is an open source model from GitHub that offers a free installation service, and any user can find StructLM-7B on GitHub to install. At the same time, huggingface.co provides the effect of StructLM-7B install, users can directly use StructLM-7B installed effect in huggingface.co for debugging and trial. It also supports api for free installation.