michaelfeil / jina-embeddings-v2-base-code

huggingface.co
Total runs: 14
24-hour runs: 0
7-day runs: -1
30-day runs: 3
Model's Last Updated: May 12 2024
feature-extraction

Introduction of jina-embeddings-v2-base-code

Model Details of jina-embeddings-v2-base-code



Finetuner logo: Finetuner helps you to create experiments in order to improve embeddings on search tasks. It accompanies you to deliver the last mile of performance-tuning for neural search applications.

The text embedding set trained by Jina AI .

Quick Start

The easiest way to starting using jina-embeddings-v2-base-code is to use Jina AI's Embedding API .

Intended Usage & Model Info

jina-embeddings-v2-base-code is an multilingual embedding model speaks English and 30 widely used programming languages . Same as other jina-embeddings-v2 series, it supports 8192 sequence length.

jina-embeddings-v2-base-code is based on a Bert architecture (JinaBert) that supports the symmetric bidirectional variant of ALiBi to allow longer sequence length. The backbone jina-bert-v2-base-code is pretrained on the github-code dataset. The model is further trained on Jina AI's collection of more than 150 millions of coding question answer and docstring source code pairs. These pairs were obtained from various domains and were carefully selected through a thorough cleaning process.

The embedding model was trained using 512 sequence length, but extrapolates to 8k sequence length (or even longer) thanks to ALiBi. This makes our model useful for a range of use cases, especially when processing long documents is needed, including technical question answering and code search.

This model has 161 million parameters, which enables fast and memory efficient inference, while delivering impressive performance. Additionally, we provide the following embedding models:

Supported (Programming) Languages

  • English
  • Assembly
  • Batchfile
  • C
  • C#
  • C++
  • CMake
  • CSS
  • Dockerfile
  • FORTRAN
  • GO
  • Haskell
  • HTML
  • Java
  • JavaScript
  • Julia
  • Lua
  • Makefile
  • Markdown
  • PHP
  • Perl
  • PowerShell
  • Python
  • Ruby
  • Rust
  • SQL
  • Scala
  • Shell
  • TypeScript
  • TeX
  • Visual Basic

Data & Parameters

Jina Embeddings V2 technical report

Usage

Please apply mean pooling when integrating the model.

Why mean pooling?

mean poooling takes all token embeddings from model output and averaging them at sentence/paragraph level. It has been proved to be the most effective way to produce high-quality sentence embeddings. We offer an encode function to deal with this.

However, if you would like to do it without using the default encode function:

import torch
import torch.nn.functional as F
from transformers import AutoTokenizer, AutoModel

def mean_pooling(model_output, attention_mask):
    token_embeddings = model_output[0]
    input_mask_expanded = attention_mask.unsqueeze(-1).expand(token_embeddings.size()).float()
    return torch.sum(token_embeddings * input_mask_expanded, 1) / torch.clamp(input_mask_expanded.sum(1), min=1e-9)

sentences = [
    'How do I access the index while iterating over a sequence with a for loop?',
    '# Use the built-in enumerator\nfor idx, x in enumerate(xs):\n    print(idx, x)',
]

tokenizer = AutoTokenizer.from_pretrained('jinaai/jina-embeddings-v2-base-code')
model = AutoModel.from_pretrained('jinaai/jina-embeddings-v2-base-code', trust_remote_code=True)

encoded_input = tokenizer(sentences, padding=True, truncation=True, return_tensors='pt')

with torch.no_grad():
    model_output = model(**encoded_input)

embeddings = mean_pooling(model_output, encoded_input['attention_mask'])
embeddings = F.normalize(embeddings, p=2, dim=1)

You can use Jina Embedding models directly from transformers package:

!pip install transformers
from transformers import AutoModel
from numpy.linalg import norm

cos_sim = lambda a,b: (a @ b.T) / (norm(a)*norm(b))
model = AutoModel.from_pretrained('jinaai/jina-embeddings-v2-base-code', trust_remote_code=True)
embeddings = model.encode(
    [
        'How do I access the index while iterating over a sequence with a for loop?',
        '# Use the built-in enumerator\nfor idx, x in enumerate(xs):\n    print(idx, x)',
    ]
)
print(cos_sim(embeddings[0], embeddings[1]))
>>> tensor([[0.7282]])

If you only want to handle shorter sequence, such as 2k, pass the max_length parameter to the encode function:

embeddings = model.encode(
    ['Very long ... code'],
    max_length=2048
)

Using the its latest release (v2.3.0) sentence-transformers also supports Jina embeddings (Please make sure that you are logged into huggingface as well):

!pip install -U sentence-transformers
from sentence_transformers import SentenceTransformer
from sentence_transformers.util import cos_sim

model = SentenceTransformer(
    "jinaai/jina-embeddings-v2-base-code",
    trust_remote_code=True
)

# control your input sequence length up to 8192
model.max_seq_length = 1024

embeddings = model.encode([
    'How do I access the index while iterating over a sequence with a for loop?',
    '# Use the built-in enumerator\nfor idx, x in enumerate(xs):\n    print(idx, x)',
])
print(cos_sim(embeddings[0], embeddings[1]))

You can also use the Transformers.js library to compute embeddings in JavaScript.

// npm i @xenova/transformers
import { pipeline, cos_sim } from '@xenova/transformers';

const extractor = await pipeline('feature-extraction', 'jinaai/jina-embeddings-v2-base-code', {
    quantized: false, // Comment out this line to use the 8-bit quantized version
});

const texts = [
    'How do I access the index while iterating over a sequence with a for loop?',
    '# Use the built-in enumerator\nfor idx, x in enumerate(xs):\n    print(idx, x)',
]
const embeddings = await extractor(texts, { pooling: 'mean' });

const score = cos_sim(embeddings[0].data, embeddings[1].data);
console.log(score);
// 0.7281748759529421
Plans
  1. Bilingual embedding models supporting more European & Asian languages, including Spanish, French, Italian and Japanese.
  2. Multimodal embedding models enable Multimodal RAG applications.
  3. High-performt rerankers.
Contact

Join our Discord community and chat with other community members about ideas.

Runs of michaelfeil jina-embeddings-v2-base-code on huggingface.co

14
Total runs
0
24-hour runs
-1
3-day runs
-1
7-day runs
3
30-day runs

More Information About jina-embeddings-v2-base-code huggingface.co Model

More jina-embeddings-v2-base-code license Visit here:

https://choosealicense.com/licenses/apache-2.0

jina-embeddings-v2-base-code huggingface.co

jina-embeddings-v2-base-code huggingface.co is an AI model on huggingface.co that provides jina-embeddings-v2-base-code's model effect (), which can be used instantly with this michaelfeil jina-embeddings-v2-base-code model. huggingface.co supports a free trial of the jina-embeddings-v2-base-code model, and also provides paid use of the jina-embeddings-v2-base-code. Support call jina-embeddings-v2-base-code model through api, including Node.js, Python, http.

jina-embeddings-v2-base-code huggingface.co Url

https://huggingface.co/michaelfeil/jina-embeddings-v2-base-code

michaelfeil jina-embeddings-v2-base-code online free

jina-embeddings-v2-base-code huggingface.co is an online trial and call api platform, which integrates jina-embeddings-v2-base-code's modeling effects, including api services, and provides a free online trial of jina-embeddings-v2-base-code, you can try jina-embeddings-v2-base-code online for free by clicking the link below.

michaelfeil jina-embeddings-v2-base-code online free url in huggingface.co:

https://huggingface.co/michaelfeil/jina-embeddings-v2-base-code

jina-embeddings-v2-base-code install

jina-embeddings-v2-base-code is an open source model from GitHub that offers a free installation service, and any user can find jina-embeddings-v2-base-code on GitHub to install. At the same time, huggingface.co provides the effect of jina-embeddings-v2-base-code install, users can directly use jina-embeddings-v2-base-code installed effect in huggingface.co for debugging and trial. It also supports api for free installation.

jina-embeddings-v2-base-code install url in huggingface.co:

https://huggingface.co/michaelfeil/jina-embeddings-v2-base-code

Url of jina-embeddings-v2-base-code

jina-embeddings-v2-base-code huggingface.co Url

Provider of jina-embeddings-v2-base-code huggingface.co

michaelfeil
ORGANIZATIONS

Other API from michaelfeil