rasyosef / SPLADE-BERT-Tiny

huggingface.co
Total runs: 15
24-hour runs: 0
7-day runs: 0
30-day runs: 0
Model's Last Updated: July 08 2025
feature-extraction

Introduction of SPLADE-BERT-Tiny

Model Details of SPLADE-BERT-Tiny

SPLADE-BERT-Tiny

This is a SPLADE Sparse Encoder model finetuned from prajjwal1/bert-tiny using the sentence-transformers library. It maps sentences & paragraphs to a 30522-dimensional sparse vector space and can be used for semantic search and sparse retrieval.

Model Details
Model Description
  • Model Type: SPLADE Sparse Encoder
  • Base model: prajjwal1/bert-tiny
  • Maximum Sequence Length: 512 tokens
  • Output Dimensionality: 30522 dimensions
  • Similarity Function: Dot Product
  • Language: en
  • License: mit
Model Sources
Full Model Architecture
SparseEncoder(
  (0): MLMTransformer({'max_seq_length': 512, 'do_lower_case': False, 'architecture': 'BertForMaskedLM'})
  (1): SpladePooling({'pooling_strategy': 'max', 'activation_function': 'relu', 'word_embedding_dimension': 30522})
)
Usage
Direct Usage (Sentence Transformers)

First install the Sentence Transformers library:

pip install -U sentence-transformers

Then you can load this model and run inference.

from sentence_transformers import SparseEncoder

# Download from the 🤗 Hub
model = SparseEncoder("rasyosef/SPLADE-BERT-Tiny")
# Run inference
queries = [
    "benefits of health care act",
]
documents = [
    'Every health insurance plan sold in the Marketplace will offer 10 essential health benefits. These essential health benefits include at least the following items and services: 1  Outpatient care—the kind you get without being admitted to a hospital.  Trips to the emergency room.',
    'Essential Health Benefits. A set of 10 categories of services health insurance plans must cover under the Affordable Care Act. These include doctors’ services, inpatient and outpatient hospital care, prescription drug coverage, pregnancy and childbirth, mental health services, and more. Some plans cover more services.',
    'Find Baker Hughes in Minot, ND 58701-5870 on Yellowbook. Get contact details or leave a review about this business.',
]
query_embeddings = model.encode_query(queries)
document_embeddings = model.encode_document(documents)
print(query_embeddings.shape, document_embeddings.shape)
# [1, 30522] [3, 30522]

# Get the similarity scores for the embeddings
similarities = model.similarity(query_embeddings, document_embeddings)
print(similarities)
# tensor([[13.6736, 18.2270,  0.0508]])
Evaluation
Metrics
Sparse Information Retrieval
Metric Value
dot_accuracy@1 0.7258
dot_accuracy@3 0.8584
dot_accuracy@5 0.8986
dot_accuracy@10 0.9372
dot_precision@1 0.7258
dot_precision@3 0.2861
dot_precision@5 0.1797
dot_precision@10 0.0937
dot_recall@1 0.7258
dot_recall@3 0.8584
dot_recall@5 0.8986
dot_recall@10 0.9372
dot_ndcg@10 0.8336
dot_mrr@10 0.8001
dot_map@100 0.8026
query_active_dims 34.0416
query_sparsity_ratio 0.9989
corpus_active_dims 192.5474
corpus_sparsity_ratio 0.9937
Training Details
Training Dataset
Unnamed Dataset
  • Size: 100,000 training samples
  • Columns: query , positive , and negative
  • Approximate statistics based on the first 1000 samples:
    query positive negative
    type string string string
    details
    • min: 4 tokens
    • mean: 8.07 tokens
    • max: 21 tokens
    • min: 21 tokens
    • mean: 81.21 tokens
    • max: 236 tokens
    • min: 20 tokens
    • mean: 79.54 tokens
    • max: 187 tokens
  • Samples:
    query positive negative
    can a spouse make you move out If you and your spouse purchased a home together or the house was bought during the course of the marriage, you likely cannot make a spouse move out during divorce.However, if your spouse is engaging in abusive behavior towards you or towards your children, then you could seek a protective order or an injunction.his would be the case if you purchased the home before marriage; if the home is in your name only and if your spouse has made no contributions to the home during the time you were husband and wife. Many courts will not award a final restraining order unless there’s been actual physical violence. If you get a temporary restraining order because your spouse threatened violence, it may not be enough to keep him out of the home permanently, pending your divorce.tep 1. Talk it out. Your spouse might not want to be in the house with you any more than you want to live there with him, but he may have some viable concerns about moving out. Try to come up with a plan to address those concerns.
    can i get pregnant right before period simran47. It is difficult to get pregnant right before your period, since ovulation takes place around 2 weeks before the next period. This is true among women who have regular menstrual cycles and ovulation can be predicted to occur around mid-cycle. However, some women have irregular cycles. If your ovulation takes place later than expected (mid-cycle), say on the 24th day of the cycle instead of around 14th day, and fertilization and implantation take place, then pregnancy can occur right before your expected period. The question about whether you can get pregnant on your period or not is probably the most common of the three situations noted above. The answer is yes and no. The probability is that you would not get pregnant having sex during your period.
    average training mileage rates Beginning on Jan. 1, 2014, the standard mileage rates for the use of a car (also vans, pickups or panel trucks) will be: 1 56 cents per mile for business miles driven. 23.5 cents per mile driven for medical or moving purposes. Schneider Increases Mileage Rates. Trucker to Trucker is the major online resource for sourcing trucks for sale and for selling all trucking equipment, including rigs and semi trailers. Schneider National is a name known to all truckers and operators and has been in business for over three-quarters of a century.
  • Loss: SpladeLoss with these parameters:
    {
        "loss": "SparseMultipleNegativesRankingLoss(scale=1.0, similarity_fct='dot_score')",
        "document_regularizer_weight": 0.001,
        "query_regularizer_weight": 0.002
    }
    
Training Hyperparameters
Non-Default Hyperparameters
  • eval_strategy : epoch
  • per_device_train_batch_size : 32
  • per_device_eval_batch_size : 32
  • learning_rate : 4e-05
  • num_train_epochs : 4
  • lr_scheduler_type : cosine
  • warmup_ratio : 0.025
  • fp16 : True
  • optim : adamw_torch_fused
  • batch_sampler : no_duplicates
All Hyperparameters
Click to expand
  • overwrite_output_dir : False
  • do_predict : False
  • eval_strategy : epoch
  • prediction_loss_only : True
  • per_device_train_batch_size : 32
  • per_device_eval_batch_size : 32
  • per_gpu_train_batch_size : None
  • per_gpu_eval_batch_size : None
  • gradient_accumulation_steps : 1
  • eval_accumulation_steps : None
  • torch_empty_cache_steps : None
  • learning_rate : 4e-05
  • weight_decay : 0.0
  • adam_beta1 : 0.9
  • adam_beta2 : 0.999
  • adam_epsilon : 1e-08
  • max_grad_norm : 1.0
  • num_train_epochs : 4
  • max_steps : -1
  • lr_scheduler_type : cosine
  • lr_scheduler_kwargs : {}
  • warmup_ratio : 0.025
  • warmup_steps : 0
  • log_level : passive
  • log_level_replica : warning
  • log_on_each_node : True
  • logging_nan_inf_filter : True
  • save_safetensors : True
  • save_on_each_node : False
  • save_only_model : False
  • restore_callback_states_from_checkpoint : False
  • no_cuda : False
  • use_cpu : False
  • use_mps_device : False
  • seed : 42
  • data_seed : None
  • jit_mode_eval : False
  • use_ipex : False
  • bf16 : False
  • fp16 : True
  • fp16_opt_level : O1
  • half_precision_backend : auto
  • bf16_full_eval : False
  • fp16_full_eval : False
  • tf32 : None
  • local_rank : 0
  • ddp_backend : None
  • tpu_num_cores : None
  • tpu_metrics_debug : False
  • debug : []
  • dataloader_drop_last : False
  • dataloader_num_workers : 0
  • dataloader_prefetch_factor : None
  • past_index : -1
  • disable_tqdm : False
  • remove_unused_columns : True
  • label_names : None
  • load_best_model_at_end : False
  • ignore_data_skip : False
  • fsdp : []
  • fsdp_min_num_params : 0
  • fsdp_config : {'min_num_params': 0, 'xla': False, 'xla_fsdp_v2': False, 'xla_fsdp_grad_ckpt': False}
  • fsdp_transformer_layer_cls_to_wrap : None
  • accelerator_config : {'split_batches': False, 'dispatch_batches': None, 'even_batches': True, 'use_seedable_sampler': True, 'non_blocking': False, 'gradient_accumulation_kwargs': None}
  • deepspeed : None
  • label_smoothing_factor : 0.0
  • optim : adamw_torch_fused
  • optim_args : None
  • adafactor : False
  • group_by_length : False
  • length_column_name : length
  • ddp_find_unused_parameters : None
  • ddp_bucket_cap_mb : None
  • ddp_broadcast_buffers : False
  • dataloader_pin_memory : True
  • dataloader_persistent_workers : False
  • skip_memory_metrics : True
  • use_legacy_prediction_loop : False
  • push_to_hub : False
  • resume_from_checkpoint : None
  • hub_model_id : None
  • hub_strategy : every_save
  • hub_private_repo : None
  • hub_always_push : False
  • hub_revision : None
  • gradient_checkpointing : False
  • gradient_checkpointing_kwargs : None
  • include_inputs_for_metrics : False
  • include_for_metrics : []
  • eval_do_concat_batches : True
  • fp16_backend : auto
  • push_to_hub_model_id : None
  • push_to_hub_organization : None
  • mp_parameters :
  • auto_find_batch_size : False
  • full_determinism : False
  • torchdynamo : None
  • ray_scope : last
  • ddp_timeout : 1800
  • torch_compile : False
  • torch_compile_backend : None
  • torch_compile_mode : None
  • include_tokens_per_second : False
  • include_num_input_tokens_seen : False
  • neftune_noise_alpha : None
  • optim_target_modules : None
  • batch_eval_metrics : False
  • eval_on_start : False
  • use_liger_kernel : False
  • liger_kernel_config : None
  • eval_use_gather_object : False
  • average_tokens_across_devices : False
  • prompts : None
  • batch_sampler : no_duplicates
  • multi_dataset_batch_sampler : proportional
  • router_mapping : {}
  • learning_rate_mapping : {}
Training Logs
Epoch Step Training Loss dot_ndcg@10
1.0 3125 22.6481 0.7960
2.0 6250 0.6743 0.8296
3.0 9375 0.5636 0.8333
4.0 12500 0.5257 0.8336
-1 -1 - 0.8336
Framework Versions
  • Python: 3.11.13
  • Sentence Transformers: 5.0.0
  • Transformers: 4.53.0
  • PyTorch: 2.6.0+cu124
  • Accelerate: 1.8.1
  • Datasets: 3.6.0
  • Tokenizers: 0.21.2
Citation
BibTeX
Sentence Transformers
@inproceedings{reimers-2019-sentence-bert,
    title = "Sentence-BERT: Sentence Embeddings using Siamese BERT-Networks",
    author = "Reimers, Nils and Gurevych, Iryna",
    booktitle = "Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing",
    month = "11",
    year = "2019",
    publisher = "Association for Computational Linguistics",
    url = "https://arxiv.org/abs/1908.10084",
}
SpladeLoss
@misc{formal2022distillationhardnegativesampling,
      title={From Distillation to Hard Negative Sampling: Making Sparse Neural IR Models More Effective},
      author={Thibault Formal and Carlos Lassance and Benjamin Piwowarski and Stéphane Clinchant},
      year={2022},
      eprint={2205.04733},
      archivePrefix={arXiv},
      primaryClass={cs.IR},
      url={https://arxiv.org/abs/2205.04733},
}
SparseMultipleNegativesRankingLoss
@misc{henderson2017efficient,
    title={Efficient Natural Language Response Suggestion for Smart Reply},
    author={Matthew Henderson and Rami Al-Rfou and Brian Strope and Yun-hsuan Sung and Laszlo Lukacs and Ruiqi Guo and Sanjiv Kumar and Balint Miklos and Ray Kurzweil},
    year={2017},
    eprint={1705.00652},
    archivePrefix={arXiv},
    primaryClass={cs.CL}
}
FlopsLoss
@article{paria2020minimizing,
    title={Minimizing flops to learn efficient sparse representations},
    author={Paria, Biswajit and Yeh, Chih-Kuan and Yen, Ian EH and Xu, Ning and Ravikumar, Pradeep and P{'o}czos, Barnab{'a}s},
    journal={arXiv preprint arXiv:2004.05665},
    year={2020}
}

Runs of rasyosef SPLADE-BERT-Tiny on huggingface.co

15
Total runs
0
24-hour runs
0
3-day runs
0
7-day runs
0
30-day runs

More Information About SPLADE-BERT-Tiny huggingface.co Model

More SPLADE-BERT-Tiny license Visit here:

https://choosealicense.com/licenses/mit

SPLADE-BERT-Tiny huggingface.co

SPLADE-BERT-Tiny huggingface.co is an AI model on huggingface.co that provides SPLADE-BERT-Tiny's model effect (), which can be used instantly with this rasyosef SPLADE-BERT-Tiny model. huggingface.co supports a free trial of the SPLADE-BERT-Tiny model, and also provides paid use of the SPLADE-BERT-Tiny. Support call SPLADE-BERT-Tiny model through api, including Node.js, Python, http.

SPLADE-BERT-Tiny huggingface.co Url

https://huggingface.co/rasyosef/SPLADE-BERT-Tiny

rasyosef SPLADE-BERT-Tiny online free

SPLADE-BERT-Tiny huggingface.co is an online trial and call api platform, which integrates SPLADE-BERT-Tiny's modeling effects, including api services, and provides a free online trial of SPLADE-BERT-Tiny, you can try SPLADE-BERT-Tiny online for free by clicking the link below.

rasyosef SPLADE-BERT-Tiny online free url in huggingface.co:

https://huggingface.co/rasyosef/SPLADE-BERT-Tiny

SPLADE-BERT-Tiny install

SPLADE-BERT-Tiny is an open source model from GitHub that offers a free installation service, and any user can find SPLADE-BERT-Tiny on GitHub to install. At the same time, huggingface.co provides the effect of SPLADE-BERT-Tiny install, users can directly use SPLADE-BERT-Tiny installed effect in huggingface.co for debugging and trial. It also supports api for free installation.

SPLADE-BERT-Tiny install url in huggingface.co:

https://huggingface.co/rasyosef/SPLADE-BERT-Tiny

Url of SPLADE-BERT-Tiny

SPLADE-BERT-Tiny huggingface.co Url

Provider of SPLADE-BERT-Tiny huggingface.co

rasyosef
ORGANIZATIONS

Other API from rasyosef

huggingface.co

Total runs: 5.1K
Run Growth: 794
Growth Rate: 15.69%
Updated:October 07 2025
huggingface.co

Total runs: 659
Run Growth: 598
Growth Rate: 90.74%
Updated:October 07 2025
huggingface.co

Total runs: 0
Run Growth: -1
Growth Rate: 0.00%
Updated:September 16 2024