Ksenia Se

Kseniase

AI & ML interests

None yet

Recent Activity

replied to their post 3 days ago
12 Types of JEPA JEPA, or Joint Embedding Predictive Architecture, is an approach to building AI models introduced by Yann LeCun. It differs from transformers by predicting the representation of a missing or future part of the input, rather than the next token or pixel. This encourages conceptual understanding, not just low-level pattern matching. So JEPA allows teaching AI to reason abstractly. Here are 12 types of JEPA you should know about: 1. I-JEPA -> https://huggingface.co/papers/2301.08243 A non-generative, self-supervised learning framework designed for processing images. It works by masking parts of the images and then trying to predict those masked parts 2. MC-JEPA -> https://huggingface.co/papers/2307.12698 Simultaneously interprets video data - dynamic elements (motion) and static details (content) - using a shared encoder 3. V-JEPA -> https://huggingface.co/papers/2404.08471 Presents vision models trained by predicting future video features, without pretrained image encoders, text, negative sampling, or reconstruction 4. UI-JEPA -> https://huggingface.co/papers/2409.04081 Masks unlabeled UI sequences to learn abstract embeddings, then adds a fine-tuned LLM decoder for intent prediction. 5. Audio-based JEPA (A-JEPA) -> https://huggingface.co/papers/2311.15830 Masks spectrogram patches with a curriculum, encodes them, and predicts hidden representations. 6. S-JEPA -> https://huggingface.co/papers/2403.11772 Signal-JEPA is used in EEG analysis. It adds a spatial block-masking scheme and three lightweight downstream classifiers 7. TI-JEPA -> https://huggingface.co/papers/2503.06380 Text-Image JEPA uses self-supervised, energy-based pre-training to map text and images into a shared embedding space, improving cross-modal transfer to downstream tasks Find more types below 👇 Also, explore the basics of JEPA in our article: https://www.turingpost.com/p/jepa If you liked it, subscribe to the Turing Post: https://www.turingpost.com/subscribe
posted an update 3 days ago
12 Types of JEPA JEPA, or Joint Embedding Predictive Architecture, is an approach to building AI models introduced by Yann LeCun. It differs from transformers by predicting the representation of a missing or future part of the input, rather than the next token or pixel. This encourages conceptual understanding, not just low-level pattern matching. So JEPA allows teaching AI to reason abstractly. Here are 12 types of JEPA you should know about: 1. I-JEPA -> https://huggingface.co/papers/2301.08243 A non-generative, self-supervised learning framework designed for processing images. It works by masking parts of the images and then trying to predict those masked parts 2. MC-JEPA -> https://huggingface.co/papers/2307.12698 Simultaneously interprets video data - dynamic elements (motion) and static details (content) - using a shared encoder 3. V-JEPA -> https://huggingface.co/papers/2404.08471 Presents vision models trained by predicting future video features, without pretrained image encoders, text, negative sampling, or reconstruction 4. UI-JEPA -> https://huggingface.co/papers/2409.04081 Masks unlabeled UI sequences to learn abstract embeddings, then adds a fine-tuned LLM decoder for intent prediction. 5. Audio-based JEPA (A-JEPA) -> https://huggingface.co/papers/2311.15830 Masks spectrogram patches with a curriculum, encodes them, and predicts hidden representations. 6. S-JEPA -> https://huggingface.co/papers/2403.11772 Signal-JEPA is used in EEG analysis. It adds a spatial block-masking scheme and three lightweight downstream classifiers 7. TI-JEPA -> https://huggingface.co/papers/2503.06380 Text-Image JEPA uses self-supervised, energy-based pre-training to map text and images into a shared embedding space, improving cross-modal transfer to downstream tasks Find more types below 👇 Also, explore the basics of JEPA in our article: https://www.turingpost.com/p/jepa If you liked it, subscribe to the Turing Post: https://www.turingpost.com/subscribe
replied to their post 10 days ago
7 Free resources to master Multi-Agent Systems (MAS) Collective intelligence is the future of AI. Sometimes, a single agent isn't enough — a team of simpler, specialized agents working together to solve a task can be a much better option. Building Multi-Agent Systems (MAS) isn’t easy, that's why today we’re offering you a list of sources that may help you master MAS: 1. CrewAI tutorials -> https://docs.crewai.com/introduction#ready-to-start-building%3F At the end of the page you'll find a guide on how to build a crew of agents that research and analyze a topic, and create a report. Also, there are useful guides on how to build a single CrewAI agent and a workflow 2. Building with CAMEL multi-agent framework -> https://github.com/camel-ai/camel Offers guides, cookbooks and other useful information to build even million agent societies, explore and work with MAS 3. Lang Chain multi-agent tutorial -> https://langchain-ai.github.io/langgraph/agents/multi-agent/ Explains how to make agents communicate via handoffs pattern on the example of 2 multi-agent architectures - supervisor and swarm 4. "Multiagent Systems: Algorithmic, Game-Theoretic, and Logical Foundations" by Yoav Shoham and Kevin Leyton-Brown -> https://www.masfoundations.org/download.html This book explains learning between agents, how multiple agents solve shared problems and communicate with focus on theory, practical examples and algorithms, diving into the game theory and logical approaches Also, check out The Turing Post article about MAS -> https://www.turingpost.com/p/mas Our article can be a good starting guide for you to explore what MAS is, its components, architectures, types, top recent developments and current trends More resources in the comments 👇 If you liked it, also subscribe to the Turing Post: https://www.turingpost.com/subscribe
View all activity

Organizations

Turing Post's profile picture Journalists on Hugging Face's profile picture Social Post Explorers's profile picture Hugging Face Discord Community's profile picture Sandbox's profile picture

Posts 24

view post
Post
4014
12 Types of JEPA

JEPA, or Joint Embedding Predictive Architecture, is an approach to building AI models introduced by Yann LeCun. It differs from transformers by predicting the representation of a missing or future part of the input, rather than the next token or pixel. This encourages conceptual understanding, not just low-level pattern matching. So JEPA allows teaching AI to reason abstractly.

Here are 12 types of JEPA you should know about:

1. I-JEPA -> Self-Supervised Learning from Images with a Joint-Embedding Predictive Architecture (2301.08243)
A non-generative, self-supervised learning framework designed for processing images. It works by masking parts of the images and then trying to predict those masked parts

2. MC-JEPA -> MC-JEPA: A Joint-Embedding Predictive Architecture for Self-Supervised Learning of Motion and Content Features (2307.12698)
Simultaneously interprets video data - dynamic elements (motion) and static details (content) - using a shared encoder

3. V-JEPA -> Revisiting Feature Prediction for Learning Visual Representations from Video (2404.08471)
Presents vision models trained by predicting future video features, without pretrained image encoders, text, negative sampling, or reconstruction

4. UI-JEPA -> UI-JEPA: Towards Active Perception of User Intent through Onscreen User Activity (2409.04081)
Masks unlabeled UI sequences to learn abstract embeddings, then adds a fine-tuned LLM decoder for intent prediction.

5. Audio-based JEPA (A-JEPA) -> A-JEPA: Joint-Embedding Predictive Architecture Can Listen (2311.15830)
Masks spectrogram patches with a curriculum, encodes them, and predicts hidden representations.

6. S-JEPA -> S-JEPA: towards seamless cross-dataset transfer through dynamic spatial attention (2403.11772)
Signal-JEPA is used in EEG analysis. It adds a spatial block-masking scheme and three lightweight downstream classifiers

7. TI-JEPA -> TI-JEPA: An Innovative Energy-based Joint Embedding Strategy for Text-Image Multimodal Systems (2503.06380)
Text-Image JEPA uses self-supervised, energy-based pre-training to map text and images into a shared embedding space, improving cross-modal transfer to downstream tasks

Find more types below 👇

Also, explore the basics of JEPA in our article: https://www.turingpost.com/p/jepa

If you liked it, subscribe to the Turing Post: https://www.turingpost.com/subscribe

Articles 41

Article
12

🦸🏻#17: What is A2A and why is it – still! – underappreciated?

models 0

None public yet

datasets 0

None public yet