mirror of
https://github.com/huggingface/transformers.git
synced 2025-07-05 22:00:09 +06:00

* toctree * not-doctested.txt * collapse sections * feedback * update * rewrite get started sections * fixes * fix * loading models * fix * customize models * share * fix link * contribute part 1 * contribute pt 2 * fix toctree * tokenization pt 1 * Add new model (#32615) * v1 - working version * fix * fix * fix * fix * rename to correct name * fix title * fixup * rename files * fix * add copied from on tests * rename to `FalconMamba` everywhere and fix bugs * fix quantization + accelerate * fix copies * add `torch.compile` support * fix tests * fix tests and add slow tests * copies on config * merge the latest changes * fix tests * add few lines about instruct * Apply suggestions from code review Co-authored-by: Arthur <48595927+ArthurZucker@users.noreply.github.com> * fix * fix tests --------- Co-authored-by: Arthur <48595927+ArthurZucker@users.noreply.github.com> * "to be not" -> "not to be" (#32636) * "to be not" -> "not to be" * Update sam.md * Update trainer.py * Update modeling_utils.py * Update test_modeling_utils.py * Update test_modeling_utils.py * fix hfoption tag * tokenization pt. 2 * image processor * fix toctree * backbones * feature extractor * fix file name * processor * update not-doctested * update * make style * fix toctree * revision * make fixup * fix toctree * fix * make style * fix hfoption tag * pipeline * pipeline gradio * pipeline web server * add pipeline * fix toctree * not-doctested * prompting * llm optims * fix toctree * fixes * cache * text generation * fix * chat pipeline * chat stuff * xla * torch.compile * cpu inference * toctree * gpu inference * agents and tools * gguf/tiktoken * finetune * toctree * trainer * trainer pt 2 * optims * optimizers * accelerate * parallelism * fsdp * update * distributed cpu * hardware training * gpu training * gpu training 2 * peft * distrib debug * deepspeed 1 * deepspeed 2 * chat toctree * quant pt 1 * quant pt 2 * fix toctree * fix * fix * quant pt 3 * quant pt 4 * serialization * torchscript * scripts * tpu * review * model addition timeline * modular * more reviews * reviews * fix toctree * reviews reviews * continue reviews * more reviews * modular transformers * more review * zamba2 * fix * all frameworks * pytorch * supported model frameworks * flashattention * rm check_table * not-doctested.txt * rm check_support_list.py * feedback * updates/feedback * review * feedback * fix * update * feedback * updates * update --------- Co-authored-by: Younes Belkada <49240599+younesbelkada@users.noreply.github.com> Co-authored-by: Arthur <48595927+ArthurZucker@users.noreply.github.com> Co-authored-by: Quentin Gallouédec <45557362+qgallouedec@users.noreply.github.com>
142 lines
5.8 KiB
Markdown
142 lines
5.8 KiB
Markdown
<!--Copyright 2024 The HuggingFace Team. All rights reserved.
|
|
|
|
Licensed under the Apache License, Version 2.0 (the "License"); you may not use this file except in compliance with
|
|
the License. You may obtain a copy of the License at
|
|
|
|
http://www.apache.org/licenses/LICENSE-2.0
|
|
|
|
Unless required by applicable law or agreed to in writing, software distributed under the License is distributed on
|
|
an "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the License for the
|
|
specific language governing permissions and limitations under the License.
|
|
|
|
⚠️ Note that this file is in Markdown but contain specific syntax for our doc-builder (similar to MDX) that may not be
|
|
rendered properly in your Markdown viewer.
|
|
|
|
-->
|
|
|
|
# Mllama
|
|
|
|
<div class="flex flex-wrap space-x-1">
|
|
<img alt="PyTorch" src="https://img.shields.io/badge/PyTorch-DE3412?style=flat&logo=pytorch&logoColor=white">
|
|
</div>
|
|
|
|
## Overview
|
|
|
|
The Llama 3.2-Vision collection of multimodal large language models (LLMs) is a collection of pretrained and instruction-tuned image reasoning generative models in 11B and 90B sizes (text \+ images in / text out). The Llama 3.2-Vision instruction-tuned models are optimized for visual recognition, image reasoning, captioning, and answering general questions about an image.
|
|
|
|
**Model Architecture:** Llama 3.2-Vision is built on top of Llama 3.1 text-only model, which is an auto-regressive language model that uses an optimized transformer architecture. The tuned versions use supervised fine-tuning (SFT) and reinforcement learning with human feedback (RLHF) to align with human preferences for helpfulness and safety. To support image recognition tasks, the Llama 3.2-Vision model uses a separately trained vision adapter that integrates with the pre-trained Llama 3.1 language model. The adapter consists of a series of cross-attention layers that feed image encoder representations into the core LLM.
|
|
|
|
## Usage Tips
|
|
|
|
- For image+text and text inputs use `MllamaForConditionalGeneration`.
|
|
- For text-only inputs use `MllamaForCausalLM` for generation to avoid loading vision tower.
|
|
- Each sample can contain multiple images, and the number of images can vary between samples. The processor will pad the inputs to the maximum number of images across samples and to a maximum number of tiles within each image.
|
|
- The text passed to the processor should have the `"<|image|>"` tokens where the images should be inserted.
|
|
- The processor has its own `apply_chat_template` method to convert chat messages to text that can then be passed as text to the processor. If you're using `transformers>=4.49.0`, you can also get a vectorized output from `apply_chat_template`. See the **Usage Examples** below for more details on how to use it.
|
|
|
|
|
|
|
|
<Tip warning={true}>
|
|
|
|
Mllama has an extra token used as a placeholder for image positions in the text. It means that input ids and an input embedding layer will have an extra token. But since the weights for input and output embeddings are not tied, the `lm_head` layer has one less token and will fail if you want to calculate loss on image tokens or apply some logit processors. In case you are training, make sure to mask out special `"<|image|>"` tokens in the `labels` as the model should not be trained on predicting them.
|
|
|
|
Otherwise if you see CUDA-side index erros when generating, use the below code to expand the `lm_head` by one more token.
|
|
|
|
|
|
```python
|
|
old_embeddings = model.get_output_embeddings()
|
|
|
|
num_tokens = model.vocab_size + 1
|
|
resized_embeddings = model._get_resized_lm_head(old_embeddings, new_num_tokens=num_tokens, mean_resizing=True)
|
|
resized_embeddings.requires_grad_(old_embeddings.weight.requires_grad)
|
|
model.set_output_embeddings(resized_embeddings)
|
|
```
|
|
</Tip>
|
|
|
|
|
|
## Usage Example
|
|
|
|
#### Instruct model
|
|
```python
|
|
import torch
|
|
from transformers import MllamaForConditionalGeneration, AutoProcessor
|
|
|
|
model_id = "meta-llama/Llama-3.2-11B-Vision-Instruct"
|
|
model = MllamaForConditionalGeneration.from_pretrained(model_id, device_map="auto", torch_dtype=torch.bfloat16)
|
|
processor = AutoProcessor.from_pretrained(model_id)
|
|
|
|
messages = [
|
|
[
|
|
{
|
|
"role": "user",
|
|
"content": [
|
|
{"type": "image", "url": "https://llava-vl.github.io/static/images/view.jpg"},
|
|
{"type": "text", "text": "What does the image show?"}
|
|
]
|
|
}
|
|
],
|
|
]
|
|
inputs = processor.apply_chat_template(messages, add_generation_prompt=True, tokenize=True, return_dict=True, return_tensors="pt").to(model.device)
|
|
output = model.generate(**inputs, max_new_tokens=25)
|
|
print(processor.decode(output[0]))
|
|
```
|
|
|
|
#### Base model
|
|
```python
|
|
import requests
|
|
import torch
|
|
from PIL import Image
|
|
from transformers import MllamaForConditionalGeneration, AutoProcessor
|
|
|
|
model_id = "meta-llama/Llama-3.2-11B-Vision"
|
|
model = MllamaForConditionalGeneration.from_pretrained(model_id, device_map="auto", torch_dtype=torch.bfloat16)
|
|
processor = AutoProcessor.from_pretrained(model_id)
|
|
|
|
prompt = "<|image|>If I had to write a haiku for this one"
|
|
url = "https://llava-vl.github.io/static/images/view.jpg"
|
|
raw_image = Image.open(requests.get(url, stream=True).raw)
|
|
|
|
inputs = processor(text=prompt, images=raw_image, return_tensors="pt").to(model.device)
|
|
output = model.generate(**inputs, do_sample=False, max_new_tokens=25)
|
|
print(processor.decode(output[0], skip_special_tokens=True))
|
|
```
|
|
|
|
|
|
## MllamaConfig
|
|
|
|
[[autodoc]] MllamaConfig
|
|
|
|
## MllamaProcessor
|
|
|
|
[[autodoc]] MllamaProcessor
|
|
|
|
|
|
## MllamaImageProcessor
|
|
|
|
[[autodoc]] MllamaImageProcessor
|
|
|
|
## MllamaForConditionalGeneration
|
|
|
|
[[autodoc]] MllamaForConditionalGeneration
|
|
- forward
|
|
|
|
## MllamaForCausalLM
|
|
|
|
[[autodoc]] MllamaForCausalLM
|
|
- forward
|
|
|
|
## MllamaTextModel
|
|
|
|
[[autodoc]] MllamaTextModel
|
|
- forward
|
|
|
|
## MllamaForCausalLM
|
|
|
|
[[autodoc]] MllamaForCausalLM
|
|
- forward
|
|
|
|
## MllamaVisionModel
|
|
|
|
[[autodoc]] MllamaVisionModel
|
|
- forward
|