Babelscape / rebel-large

huggingface.co
Total runs: 404.5K
24-hour runs: 57.6K
7-day runs: 308.8K
30-day runs: 82.3K
Model's Last Updated: June 20 2023
text-generation

Introduction of rebel-large

Model Details of rebel-large

PWC PWC PWC PWC PWC

Multilingual update! Check mREBEL , a multilingual version covering more relation types, languages and including entity types.

REBEL hf-rebel : Relation Extraction By End-to-end Language generation

This is the model card for the Findings of EMNLP 2021 paper REBEL: Relation Extraction By End-to-end Language generation . We present a new linearization approach and a reframing of Relation Extraction as a seq2seq task. The paper can be found here . If you use the code, please reference this work in your paper:

@inproceedings{huguet-cabot-navigli-2021-rebel-relation,
    title = "{REBEL}: Relation Extraction By End-to-end Language generation",
    author = "Huguet Cabot, Pere-Llu{\'\i}s  and
      Navigli, Roberto",
    booktitle = "Findings of the Association for Computational Linguistics: EMNLP 2021",
    month = nov,
    year = "2021",
    address = "Punta Cana, Dominican Republic",
    publisher = "Association for Computational Linguistics",
    url = "https://aclanthology.org/2021.findings-emnlp.204",
    pages = "2370--2381",
    abstract = "Extracting relation triplets from raw text is a crucial task in Information Extraction, enabling multiple applications such as populating or validating knowledge bases, factchecking, and other downstream tasks. However, it usually involves multiple-step pipelines that propagate errors or are limited to a small number of relation types. To overcome these issues, we propose the use of autoregressive seq2seq models. Such models have previously been shown to perform well not only in language generation, but also in NLU tasks such as Entity Linking, thanks to their framing as seq2seq tasks. In this paper, we show how Relation Extraction can be simplified by expressing triplets as a sequence of text and we present REBEL, a seq2seq model based on BART that performs end-to-end relation extraction for more than 200 different relation types. We show our model{'}s flexibility by fine-tuning it on an array of Relation Extraction and Relation Classification benchmarks, with it attaining state-of-the-art performance in most of them.",
}

The original repository for the paper can be found here

Be aware that the inference widget at the right does not output special tokens, which are necessary to distinguish the subject, object and relation types. For a demo of REBEL and its pre-training dataset check the Spaces demo .

Pipeline usage
from transformers import pipeline

triplet_extractor = pipeline('text2text-generation', model='Babelscape/rebel-large', tokenizer='Babelscape/rebel-large')
# We need to use the tokenizer manually since we need special tokens.
extracted_text = triplet_extractor.tokenizer.batch_decode([triplet_extractor("Punta Cana is a resort town in the municipality of Higuey, in La Altagracia Province, the eastern most province of the Dominican Republic", return_tensors=True, return_text=False)[0]["generated_token_ids"]])
print(extracted_text[0])
# Function to parse the generated text and extract the triplets
def extract_triplets(text):
    triplets = []
    relation, subject, relation, object_ = '', '', '', ''
    text = text.strip()
    current = 'x'
    for token in text.replace("<s>", "").replace("<pad>", "").replace("</s>", "").split():
        if token == "<triplet>":
            current = 't'
            if relation != '':
                triplets.append({'head': subject.strip(), 'type': relation.strip(),'tail': object_.strip()})
                relation = ''
            subject = ''
        elif token == "<subj>":
            current = 's'
            if relation != '':
                triplets.append({'head': subject.strip(), 'type': relation.strip(),'tail': object_.strip()})
            object_ = ''
        elif token == "<obj>":
            current = 'o'
            relation = ''
        else:
            if current == 't':
                subject += ' ' + token
            elif current == 's':
                object_ += ' ' + token
            elif current == 'o':
                relation += ' ' + token
    if subject != '' and relation != '' and object_ != '':
        triplets.append({'head': subject.strip(), 'type': relation.strip(),'tail': object_.strip()})
    return triplets
extracted_triplets = extract_triplets(extracted_text[0])
print(extracted_triplets)
Model and Tokenizer using transformers
from transformers import AutoModelForSeq2SeqLM, AutoTokenizer

def extract_triplets(text):
    triplets = []
    relation, subject, relation, object_ = '', '', '', ''
    text = text.strip()
    current = 'x'
    for token in text.replace("<s>", "").replace("<pad>", "").replace("</s>", "").split():
        if token == "<triplet>":
            current = 't'
            if relation != '':
                triplets.append({'head': subject.strip(), 'type': relation.strip(),'tail': object_.strip()})
                relation = ''
            subject = ''
        elif token == "<subj>":
            current = 's'
            if relation != '':
                triplets.append({'head': subject.strip(), 'type': relation.strip(),'tail': object_.strip()})
            object_ = ''
        elif token == "<obj>":
            current = 'o'
            relation = ''
        else:
            if current == 't':
                subject += ' ' + token
            elif current == 's':
                object_ += ' ' + token
            elif current == 'o':
                relation += ' ' + token
    if subject != '' and relation != '' and object_ != '':
        triplets.append({'head': subject.strip(), 'type': relation.strip(),'tail': object_.strip()})
    return triplets

# Load model and tokenizer
tokenizer = AutoTokenizer.from_pretrained("Babelscape/rebel-large")
model = AutoModelForSeq2SeqLM.from_pretrained("Babelscape/rebel-large")
gen_kwargs = {
    "max_length": 256,
    "length_penalty": 0,
    "num_beams": 3,
    "num_return_sequences": 3,
}

# Text to extract triplets from
text = 'Punta Cana is a resort town in the municipality of Higüey, in La Altagracia Province, the easternmost province of the Dominican Republic.'

# Tokenizer text
model_inputs = tokenizer(text, max_length=256, padding=True, truncation=True, return_tensors = 'pt')

# Generate
generated_tokens = model.generate(
    model_inputs["input_ids"].to(model.device),
    attention_mask=model_inputs["attention_mask"].to(model.device),
    **gen_kwargs,
)

# Extract text
decoded_preds = tokenizer.batch_decode(generated_tokens, skip_special_tokens=False)

# Extract triplets
for idx, sentence in enumerate(decoded_preds):
    print(f'Prediction triplets sentence {idx}')
    print(extract_triplets(sentence))

Runs of Babelscape rebel-large on huggingface.co

404.5K
Total runs
57.6K
24-hour runs
150.5K
3-day runs
308.8K
7-day runs
82.3K
30-day runs

More Information About rebel-large huggingface.co Model

More rebel-large license Visit here:

https://choosealicense.com/licenses/cc-by-nc-sa-4.0

rebel-large huggingface.co

rebel-large huggingface.co is an AI model on huggingface.co that provides rebel-large's model effect (), which can be used instantly with this Babelscape rebel-large model. huggingface.co supports a free trial of the rebel-large model, and also provides paid use of the rebel-large. Support call rebel-large model through api, including Node.js, Python, http.

Babelscape rebel-large online free

rebel-large huggingface.co is an online trial and call api platform, which integrates rebel-large's modeling effects, including api services, and provides a free online trial of rebel-large, you can try rebel-large online for free by clicking the link below.

Babelscape rebel-large online free url in huggingface.co:

https://huggingface.co/Babelscape/rebel-large

rebel-large install

rebel-large is an open source model from GitHub that offers a free installation service, and any user can find rebel-large on GitHub to install. At the same time, huggingface.co provides the effect of rebel-large install, users can directly use rebel-large installed effect in huggingface.co for debugging and trial. It also supports api for free installation.

rebel-large install url in huggingface.co:

https://huggingface.co/Babelscape/rebel-large

Url of rebel-large

Provider of rebel-large huggingface.co

Babelscape
ORGANIZATIONS

Other API from Babelscape

huggingface.co

Total runs: 139
Run Growth: 129
Growth Rate: 92.81%
Updated:June 18 2024
huggingface.co

Total runs: 61
Run Growth: 22
Growth Rate: 33.85%
Updated:December 13 2024
huggingface.co

Total runs: 0
Run Growth: 0
Growth Rate: 0.00%
Updated:August 29 2024