Image Feature Extraction
Transformers
Safetensors
feature-extraction
custom_code

flash attention 2

#3
by Itaykatzir - opened

Does it support flash attention 2?
Can you give an example of usage?

NVIDIA org

We use torch.scaled_dot_product_attention, which under the hood will use FA2 if the input conditions are met.

Sign up or log in to comment