-
Prompt Cache: Modular Attention Reuse for Low-Latency Inference
Paper • 2311.04934 • Published • 34 -
I2VEdit: First-Frame-Guided Video Editing via Image-to-Video Diffusion Models
Paper • 2405.16537 • Published • 18 -
NV-Embed: Improved Techniques for Training LLMs as Generalist Embedding Models
Paper • 2405.17428 • Published • 20
Pedro Batista
pedrovhb
·
AI & ML interests
None yet
Recent Activity
liked
a Space
29 days ago
nvidia/PartPacker
liked
a model
about 1 month ago
nanonets/Nanonets-OCR-s
liked
a Space
4 months ago
nikigoli/countgd
Organizations
None yet