(1) To comprehensively and rigorously benchmark the ability of the off-the-shelf MLLMs in chart domain, we construct an evaluation set covering multi-modal (image, code, csv, text description), multi-task, multi-disciplinary, high-quality chart data, and evaluate the performance of mainstream MLLMs.
(2) We develop ChartVLM, offering a new perspective on handling the multi-modal tasks that strongly depend on interpretable patterns such as reasoning tasks in the field of chart or geometric images. To augment the model’s interpretability in cognition tasks in chart domain, ChartVLM incorporates the intermediate chart representations such as CSV data, chart title, chart type, etc.
We collected 48K multi-modal chart data covering
22 topics
,
18 chart types
, and
7 tasks
. Each chart data within this dataset includes four modalities: image, CSV, python code, and text description.
ChartX下载(ChartX Download)
Data Download
Please download the official
ChartX Evaluation Set
dataset and organize the downloaded files as follows:
(1)
To enhance the interpretability of the chart model in cognition tasks (e.g. answer questions based on chart image), ChartVLM first performs the base perception task (e.g. structural extraction from the given chart image to a predicted CSV data), and then, finishes other cognition tasks (e.g. chart redrawing, description, summary, and QA) based on the extracted structural data.
(2)
To choose the task that users expect to perform according to the prompts they used, the instruction adapter is designed, which can cover a variety of user instructions as illustrated in this figure
from tools.ChartVLM import infer_ChartVLM
if __name__ == '__main__':
model = '${PATH_TO_PRETRAINED_MODEL}/ChartVLM/base/' #${PATH_TO_PRETRAINED_MODEL}
image = './base_decoder/train/data/test.png'
text = 'who has the largest value?'
output = infer_ChartVLM(image, text, model)
print(output)
Runs of U4R ChartVLM-base on huggingface.co
32
Total runs
0
24-hour runs
0
3-day runs
0
7-day runs
0
30-day runs
More Information About ChartVLM-base huggingface.co Model
ChartVLM-base huggingface.co
ChartVLM-base huggingface.co is an AI model on huggingface.co that provides ChartVLM-base's model effect (), which can be used instantly with this U4R ChartVLM-base model. huggingface.co supports a free trial of the ChartVLM-base model, and also provides paid use of the ChartVLM-base. Support call ChartVLM-base model through api, including Node.js, Python, http.
ChartVLM-base huggingface.co is an online trial and call api platform, which integrates ChartVLM-base's modeling effects, including api services, and provides a free online trial of ChartVLM-base, you can try ChartVLM-base online for free by clicking the link below.
U4R ChartVLM-base online free url in huggingface.co:
ChartVLM-base is an open source model from GitHub that offers a free installation service, and any user can find ChartVLM-base on GitHub to install. At the same time, huggingface.co provides the effect of ChartVLM-base install, users can directly use ChartVLM-base installed effect in huggingface.co for debugging and trial. It also supports api for free installation.