yasserrmd / LexEmbed-Contracts

huggingface.co
Total runs: 6
24-hour runs: 0
7-day runs: 1
30-day runs: 3
Model's Last Updated: August 28 2025
sentence-similarity

Introduction of LexEmbed-Contracts

Model Details of LexEmbed-Contracts

SentenceTransformer

This is a sentence-transformers model trained. It maps sentences & paragraphs to a 1024-dimensional dense vector space and can be used for semantic textual similarity, semantic search, paraphrase mining, text classification, clustering, and more.

Model Details
Model Description
  • Model Type: Sentence Transformer
  • Maximum Sequence Length: 512 tokens
  • Output Dimensionality: 1024 dimensions
  • Similarity Function: Cosine Similarity
Model Sources
Full Model Architecture
SentenceTransformer(
  (0): Transformer({'max_seq_length': 512, 'do_lower_case': False, 'architecture': 'XLMRobertaModel'})
  (1): Pooling({'word_embedding_dimension': 1024, 'pooling_mode_cls_token': False, 'pooling_mode_mean_tokens': True, 'pooling_mode_max_tokens': False, 'pooling_mode_mean_sqrt_len_tokens': False, 'pooling_mode_weightedmean_tokens': False, 'pooling_mode_lasttoken': False, 'include_prompt': True})
)
Usage
Direct Usage (Sentence Transformers)

First install the Sentence Transformers library:

pip install -U sentence-transformers

Then you can load this model and run inference.

from sentence_transformers import SentenceTransformer

# Download from the 🤗 Hub
model = SentenceTransformer("sentence_transformers_model_id")
# Run inference
sentences = [
    'License Grant',
    "In the event FCE notifies ExxonMobil that it has formally decided not to pursue Generation 2 Technology for Power Applications, then upon ExxonMobil's written request, FCE agrees to negotiate a grant to ExxonMobil and its Affiliates, under commercially reasonable terms to be determined in good faith, a worldwide, royalty-bearing (with the royalty to be negotiated), non-exclusive, sub-licensable right and license to practice FCE Background Information and FCE Background Patents for Generation 2 Technology in any application outside of Carbon Capture Applications and Hydrogen Applications.",
    'Aucta should continue to receive 15% of Net Sales Royalty for as long as ETON is selling the Product(s) in the Territory, unless otherwise agreed to under this Agreement.',
]
embeddings = model.encode(sentences)
print(embeddings.shape)
# [3, 1024]

# Get the similarity scores for the embeddings
similarities = model.similarity(embeddings, embeddings)
print(similarities)
# tensor([[1.0000, 0.7920, 0.3253],
#         [0.7920, 1.0000, 0.4614],
#         [0.3253, 0.4614, 1.0000]])
Training Details
Training Dataset
Unnamed Dataset
  • Size: 16,129 training samples
  • Columns: sentence_0 , sentence_1 , and label
  • Approximate statistics based on the first 1000 samples:
    sentence_0 sentence_1 label
    type string string float
    details
    • min: 3 tokens
    • mean: 54.18 tokens
    • max: 512 tokens
    • min: 3 tokens
    • mean: 95.75 tokens
    • max: 512 tokens
    • min: 1.0
    • mean: 1.0
    • max: 1.0
  • Samples:
    sentence_0 sentence_1 label
    Parties STARTEC GLOBAL COMMUNICATIONS CORPORATION 1.0
    The proceeds of the Revolving Loans and the Swingline Loans, and the Letters of Credit, shall be used for general corporate purposes, including, but not limited to, repayment of any Indebtedness and to backstop the issuance of commercial paper. Use the proceeds of the Loans and the Letters of Credit only as contemplated in Section  3.12 . The Borrower will not request any Borrowing, and the Borrower shall not use, and shall procure that its Subsidiaries and its or their respective directors, officers, employees and agents shall not use, the proceeds of any Borrowing (a) in furtherance of an offer, payment, promise to pay, or authorization of the payment or giving of money, or anything else of value, to any Person in violation of any Anti-Corruption Laws in any material respect, (b) for the purpose of funding, financing or facilitating any unauthorized activities, business or transaction of or with any Sanctioned Person, or in any Sanctioned Country, or (c) knowingly in any manner that would result in the violation of any Sanctions Laws applicable to any party hereto. 1.0
    Governing Law state. 1.0
  • Loss: MultipleNegativesRankingLoss with these parameters:
    {
        "scale": 20.0,
        "similarity_fct": "cos_sim",
        "gather_across_devices": false
    }
    
Training Hyperparameters
Non-Default Hyperparameters
  • per_device_train_batch_size : 2
  • per_device_eval_batch_size : 2
  • num_train_epochs : 1
  • fp16 : True
  • multi_dataset_batch_sampler : round_robin
All Hyperparameters
Click to expand
  • overwrite_output_dir : False
  • do_predict : False
  • eval_strategy : no
  • prediction_loss_only : True
  • per_device_train_batch_size : 2
  • per_device_eval_batch_size : 2
  • per_gpu_train_batch_size : None
  • per_gpu_eval_batch_size : None
  • gradient_accumulation_steps : 1
  • eval_accumulation_steps : None
  • torch_empty_cache_steps : None
  • learning_rate : 5e-05
  • weight_decay : 0.0
  • adam_beta1 : 0.9
  • adam_beta2 : 0.999
  • adam_epsilon : 1e-08
  • max_grad_norm : 1
  • num_train_epochs : 1
  • max_steps : -1
  • lr_scheduler_type : linear
  • lr_scheduler_kwargs : {}
  • warmup_ratio : 0.0
  • warmup_steps : 0
  • log_level : passive
  • log_level_replica : warning
  • log_on_each_node : True
  • logging_nan_inf_filter : True
  • save_safetensors : True
  • save_on_each_node : False
  • save_only_model : False
  • restore_callback_states_from_checkpoint : False
  • no_cuda : False
  • use_cpu : False
  • use_mps_device : False
  • seed : 42
  • data_seed : None
  • jit_mode_eval : False
  • use_ipex : False
  • bf16 : False
  • fp16 : True
  • fp16_opt_level : O1
  • half_precision_backend : auto
  • bf16_full_eval : False
  • fp16_full_eval : False
  • tf32 : None
  • local_rank : 0
  • ddp_backend : None
  • tpu_num_cores : None
  • tpu_metrics_debug : False
  • debug : []
  • dataloader_drop_last : False
  • dataloader_num_workers : 0
  • dataloader_prefetch_factor : None
  • past_index : -1
  • disable_tqdm : False
  • remove_unused_columns : True
  • label_names : None
  • load_best_model_at_end : False
  • ignore_data_skip : False
  • fsdp : []
  • fsdp_min_num_params : 0
  • fsdp_config : {'min_num_params': 0, 'xla': False, 'xla_fsdp_v2': False, 'xla_fsdp_grad_ckpt': False}
  • fsdp_transformer_layer_cls_to_wrap : None
  • accelerator_config : {'split_batches': False, 'dispatch_batches': None, 'even_batches': True, 'use_seedable_sampler': True, 'non_blocking': False, 'gradient_accumulation_kwargs': None}
  • deepspeed : None
  • label_smoothing_factor : 0.0
  • optim : adamw_torch_fused
  • optim_args : None
  • adafactor : False
  • group_by_length : False
  • length_column_name : length
  • ddp_find_unused_parameters : None
  • ddp_bucket_cap_mb : None
  • ddp_broadcast_buffers : False
  • dataloader_pin_memory : True
  • dataloader_persistent_workers : False
  • skip_memory_metrics : True
  • use_legacy_prediction_loop : False
  • push_to_hub : False
  • resume_from_checkpoint : None
  • hub_model_id : None
  • hub_strategy : every_save
  • hub_private_repo : None
  • hub_always_push : False
  • hub_revision : None
  • gradient_checkpointing : False
  • gradient_checkpointing_kwargs : None
  • include_inputs_for_metrics : False
  • include_for_metrics : []
  • eval_do_concat_batches : True
  • fp16_backend : auto
  • push_to_hub_model_id : None
  • push_to_hub_organization : None
  • mp_parameters :
  • auto_find_batch_size : False
  • full_determinism : False
  • torchdynamo : None
  • ray_scope : last
  • ddp_timeout : 1800
  • torch_compile : False
  • torch_compile_backend : None
  • torch_compile_mode : None
  • include_tokens_per_second : False
  • include_num_input_tokens_seen : False
  • neftune_noise_alpha : None
  • optim_target_modules : None
  • batch_eval_metrics : False
  • eval_on_start : False
  • use_liger_kernel : False
  • liger_kernel_config : None
  • eval_use_gather_object : False
  • average_tokens_across_devices : False
  • prompts : None
  • batch_sampler : batch_sampler
  • multi_dataset_batch_sampler : round_robin
  • router_mapping : {}
  • learning_rate_mapping : {}
Training Logs
Epoch Step Training Loss
0.0620 500 0.62
0.1240 1000 0.3153
0.1860 1500 0.2382
Framework Versions
  • Python: 3.12.11
  • Sentence Transformers: 5.1.0
  • Transformers: 4.55.4
  • PyTorch: 2.8.0+cu126
  • Accelerate: 1.10.1
  • Datasets: 4.0.0
  • Tokenizers: 0.21.4
Citation
BibTeX
Sentence Transformers
@inproceedings{reimers-2019-sentence-bert,
    title = "Sentence-BERT: Sentence Embeddings using Siamese BERT-Networks",
    author = "Reimers, Nils and Gurevych, Iryna",
    booktitle = "Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing",
    month = "11",
    year = "2019",
    publisher = "Association for Computational Linguistics",
    url = "https://arxiv.org/abs/1908.10084",
}
MultipleNegativesRankingLoss
@misc{henderson2017efficient,
    title={Efficient Natural Language Response Suggestion for Smart Reply},
    author={Matthew Henderson and Rami Al-Rfou and Brian Strope and Yun-hsuan Sung and Laszlo Lukacs and Ruiqi Guo and Sanjiv Kumar and Balint Miklos and Ray Kurzweil},
    year={2017},
    eprint={1705.00652},
    archivePrefix={arXiv},
    primaryClass={cs.CL}
}

Runs of yasserrmd LexEmbed-Contracts on huggingface.co

6
Total runs
0
24-hour runs
0
3-day runs
1
7-day runs
3
30-day runs

More Information About LexEmbed-Contracts huggingface.co Model

LexEmbed-Contracts huggingface.co

LexEmbed-Contracts huggingface.co is an AI model on huggingface.co that provides LexEmbed-Contracts's model effect (), which can be used instantly with this yasserrmd LexEmbed-Contracts model. huggingface.co supports a free trial of the LexEmbed-Contracts model, and also provides paid use of the LexEmbed-Contracts. Support call LexEmbed-Contracts model through api, including Node.js, Python, http.

LexEmbed-Contracts huggingface.co Url

https://huggingface.co/yasserrmd/LexEmbed-Contracts

yasserrmd LexEmbed-Contracts online free

LexEmbed-Contracts huggingface.co is an online trial and call api platform, which integrates LexEmbed-Contracts's modeling effects, including api services, and provides a free online trial of LexEmbed-Contracts, you can try LexEmbed-Contracts online for free by clicking the link below.

yasserrmd LexEmbed-Contracts online free url in huggingface.co:

https://huggingface.co/yasserrmd/LexEmbed-Contracts

LexEmbed-Contracts install

LexEmbed-Contracts is an open source model from GitHub that offers a free installation service, and any user can find LexEmbed-Contracts on GitHub to install. At the same time, huggingface.co provides the effect of LexEmbed-Contracts install, users can directly use LexEmbed-Contracts installed effect in huggingface.co for debugging and trial. It also supports api for free installation.

LexEmbed-Contracts install url in huggingface.co:

https://huggingface.co/yasserrmd/LexEmbed-Contracts

Url of LexEmbed-Contracts

LexEmbed-Contracts huggingface.co Url

Provider of LexEmbed-Contracts huggingface.co

yasserrmd
ORGANIZATIONS

Other API from yasserrmd

huggingface.co

Total runs: 28
Run Growth: 3
Growth Rate: 10.71%
Updated:October 19 2025