Our appreciation for the sponsors of DolphinVision:
TensorWave
- provided 8x mi300x node used for training, evaluations, and inference
DolphinVision is a multimodal model. It is uncensored, and capable to reason and comment regarding images that other popular models would object to.
import torch
import transformers
from transformers import AutoModelForCausalLM, AutoTokenizer
from PIL import Image
import warnings
# disable some warnings
transformers.logging.set_verbosity_error()
transformers.logging.disable_progress_bar()
warnings.filterwarnings('ignore')
# set device
torch.set_default_device('cuda') # or 'cpu'
model_name = 'cognitivecomputations/dolphin-vision-7b'# create model
model = AutoModelForCausalLM.from_pretrained(
model_name,
torch_dtype=torch.float16,
device_map='auto',
trust_remote_code=True)
tokenizer = AutoTokenizer.from_pretrained(
model_name,
trust_remote_code=True)
# text prompt
prompt = 'Describe this image in detail'
messages = [
{"role": "user", "content": f'<image>\n{prompt}'}
]
text = tokenizer.apply_chat_template(
messages,
tokenize=False,
add_generation_prompt=True
)
print(text)
text_chunks = [tokenizer(chunk).input_ids for chunk in text.split('<image>')]
input_ids = torch.tensor(text_chunks[0] + [-200] + text_chunks[1], dtype=torch.long).unsqueeze(0)
# image, sample images can be found in images folder
image = Image.open('/path/to/image.png')
image_tensor = model.process_images([image], model.config).to(dtype=model.dtype)
# generate
output_ids = model.generate(
input_ids,
images=image_tensor,
max_new_tokens=2048,
use_cache=True)[0]
print(tokenizer.decode(output_ids[input_ids.shape[1]:], skip_special_tokens=True).strip())
Examples
<|im_start|>user
: do OCR on all the text
<|im_start|>assistant
: The text in the image reads as follows:
- STOP USING DOCKER
- Applications were not meant to be trapped in virtual jars like digital fireflies
- Years of software development, yet no one questioned why we were putting programs in imaginary boats
- Let's containerize our monolith. I'll just quickly spin up 17 microservices. Statements dreamt up by the utterly deranged
- Observe the whale-sized deception they fooled us with all this time
- These are real images used to explain Docker
- "Yay Matey, I'd like to run my website in a virtual sea"
- They have played us for absolute fools
This text is a satirical commentary on the use of Docker in software development, using humor and exaggeration to question the practice.
Benchmarks
TBD
Runs of SaisExperiments DV-Experiment on huggingface.co
14
Total runs
0
24-hour runs
0
3-day runs
0
7-day runs
0
30-day runs
More Information About DV-Experiment huggingface.co Model
DV-Experiment huggingface.co is an AI model on huggingface.co that provides DV-Experiment's model effect (), which can be used instantly with this SaisExperiments DV-Experiment model. huggingface.co supports a free trial of the DV-Experiment model, and also provides paid use of the DV-Experiment. Support call DV-Experiment model through api, including Node.js, Python, http.
DV-Experiment huggingface.co is an online trial and call api platform, which integrates DV-Experiment's modeling effects, including api services, and provides a free online trial of DV-Experiment, you can try DV-Experiment online for free by clicking the link below.
SaisExperiments DV-Experiment online free url in huggingface.co:
DV-Experiment is an open source model from GitHub that offers a free installation service, and any user can find DV-Experiment on GitHub to install. At the same time, huggingface.co provides the effect of DV-Experiment install, users can directly use DV-Experiment installed effect in huggingface.co for debugging and trial. It also supports api for free installation.