Update README.md
Browse files
README.md
CHANGED
@@ -6,6 +6,8 @@ tags:
|
|
6 |
library_name: transformers
|
7 |
datasets:
|
8 |
- bkprocovid19/diving48
|
|
|
|
|
9 |
---
|
10 |
|
11 |
# V-JEPA 2
|
@@ -46,7 +48,7 @@ model = AutoModelForVideoClassification.from_pretrained(hf_repo).to(device)
|
|
46 |
processor = AutoVideoProcessor.from_pretrained(hf_repo)
|
47 |
|
48 |
# To load a video, sample the number of frames according to the model.
|
49 |
-
video_url = "https://huggingface.co/
|
50 |
vr = VideoDecoder(video_url)
|
51 |
frame_idx = np.arange(0, model.config.frames_per_clip, 8) # you can define more complex sampling strategy
|
52 |
video = vr.get_frames_at(indices=frame_idx).data # frames x channels x height x width
|
|
|
6 |
library_name: transformers
|
7 |
datasets:
|
8 |
- bkprocovid19/diving48
|
9 |
+
base_model:
|
10 |
+
- facebook/vjepa2-vitl-fpc64-256
|
11 |
---
|
12 |
|
13 |
# V-JEPA 2
|
|
|
48 |
processor = AutoVideoProcessor.from_pretrained(hf_repo)
|
49 |
|
50 |
# To load a video, sample the number of frames according to the model.
|
51 |
+
video_url = "https://huggingface.co/facebook/vjepa2-vitl-fpc32-256-diving48/resolve/main/sample/diving.mp4"
|
52 |
vr = VideoDecoder(video_url)
|
53 |
frame_idx = np.arange(0, model.config.frames_per_clip, 8) # you can define more complex sampling strategy
|
54 |
video = vr.get_frames_at(indices=frame_idx).data # frames x channels x height x width
|