Skip to content

v1.20.0: multi-modal and OpenCLIP models support, transformers v4.45

Latest
Compare
Choose a tag to compare
@echarlaix echarlaix released this 10 Oct 17:01
· 16 commits to main since this release

OpenVINO

Multi-modal models support

Adding OVModelForVisionCausalLM by @eaidova in #883

OpenCLIP models support

Adding OpenCLIP models support by @sbalandi in #857

from optimum.intel import OVModelCLIPVisual, OVModelCLIPText

visual_model = OVModelCLIPVisual.from_pretrained(model_name_or_path)
text_model  = OVModelCLIPText.from_pretrained(model_name_or_path)
image = processor(image).unsqueeze(0)
text = tokenizer(["a diagram", "a dog", "a cat"])
image_features = visual_model(image).image_features
text_features = text_model(text).text_features

Diffusion pipeline

Adding OVDiffusionPipeline to simplify diffusers model loading by @IlyasMoutawwakil in #889

  model_id = "stabilityai/stable-diffusion-xl-base-1.0"
- pipeline = OVStableDiffusionXLPipeline.from_pretrained(model_id)
+ pipeline = OVDiffusionPipeline.from_pretrained(model_id)
  image = pipeline("sailing ship in storm by Leonardo da Vinci").images[0]

NNCF GPTQ support

GPTQ support by @nikita-savelyevv in #912

Transformers v4.45

Transformers v4.45 support by @echarlaix in #902

Subfolder

Remove the restriction for the model's config to be in the model's subfolder by @tomaarsen in #933

New Contributors