Semantic-guided LoRA (SG-LoRA) → https://huggingface.co/papers/2509.10535
Generates task-specific LoRA parameters from semantic preferences, enabling zero-shot and privacy-preserving adaptation to new tasksPHLoRA (Post-hoc LoRA) → https://huggingface.co/papers/2509.10971
Extracts LoRA adapters after fine-rank fine-tuning by low-rank factoring the weight differencesLoRA-Gen → https://huggingface.co/papers/2506.11638
Generates LoRA parameters from a large cloud model for small edge models, merging them for efficient task specialization and faster inference.DP-FedLoRA → https://huggingface.co/papers/2509.09097
Adds DP noise to LoRA matrices in federated on-device fine-tuning for privacy preservation
Ksenia Se
Kseniase
AI & ML interests
None yet
Recent Activity
replied to
their
post
6 days ago
10 awesome advanced LoRA approaches
Low-Rank Adaptation (LoRA) is the go-to method for efficient model fine-tuning that adds small low-rank matrices instead of retraining full models. The field isn’t standing still – new LoRA variants push the limits of efficiency, generalization, and personalization. So we’re sharing 10 of the latest LoRA approaches you should know about:
1. Mixture-of-LoRA-experts → https://huggingface.co/papers/2509.13878
Adds multiple low-rank adapters (LoRA) into a model’s layers, and a routing mechanism activates the most suitable ones for each input. This lets the model adapt better to new unseen conditions
2. Amortized Bayesian Meta-Learning for LoRA (ABMLL) → https://huggingface.co/papers/2508.14285
Balances global and task-specific parameters within a Bayesian framework to improve uncertainty calibration and generalization to new tasks without high memory or compute costs
3. AutoLoRA → https://huggingface.co/papers/2508.02107
Automatically retrieves and dynamically aggregates public LoRAs for stronger T2I generation
4. aLoRA (Activated LoRA) → https://huggingface.co/papers/2504.12397
Only applies LoRA after invocation, letting the model reuse the base model’s KV cache instead of recomputing the full turn’s KV cache. Efficient in multi-turn conversations
5. LiLoRA (LoRA in LoRA) → https://huggingface.co/papers/2508.06202
Shares the LoRA matrix A across tasks and additionally low-rank-decomposes matrix B to cut parameters in continual vision-text MLLMs
6. Sensitivity-LoRA → https://huggingface.co/papers/2509.09119
Dynamically assigns ranks to weight matrices based on their sensitivity, measured using second-order derivatives
Read further below ↓
Also, subscribe to the Turing Post: https://www.turingpost.com/subscribe
posted
an
update
6 days ago
10 awesome advanced LoRA approaches
Low-Rank Adaptation (LoRA) is the go-to method for efficient model fine-tuning that adds small low-rank matrices instead of retraining full models. The field isn’t standing still – new LoRA variants push the limits of efficiency, generalization, and personalization. So we’re sharing 10 of the latest LoRA approaches you should know about:
1. Mixture-of-LoRA-experts → https://huggingface.co/papers/2509.13878
Adds multiple low-rank adapters (LoRA) into a model’s layers, and a routing mechanism activates the most suitable ones for each input. This lets the model adapt better to new unseen conditions
2. Amortized Bayesian Meta-Learning for LoRA (ABMLL) → https://huggingface.co/papers/2508.14285
Balances global and task-specific parameters within a Bayesian framework to improve uncertainty calibration and generalization to new tasks without high memory or compute costs
3. AutoLoRA → https://huggingface.co/papers/2508.02107
Automatically retrieves and dynamically aggregates public LoRAs for stronger T2I generation
4. aLoRA (Activated LoRA) → https://huggingface.co/papers/2504.12397
Only applies LoRA after invocation, letting the model reuse the base model’s KV cache instead of recomputing the full turn’s KV cache. Efficient in multi-turn conversations
5. LiLoRA (LoRA in LoRA) → https://huggingface.co/papers/2508.06202
Shares the LoRA matrix A across tasks and additionally low-rank-decomposes matrix B to cut parameters in continual vision-text MLLMs
6. Sensitivity-LoRA → https://huggingface.co/papers/2509.09119
Dynamically assigns ranks to weight matrices based on their sensitivity, measured using second-order derivatives
Read further below ↓
Also, subscribe to the Turing Post: https://www.turingpost.com/subscribe
replied to
their
post
13 days ago
6 Recent & free sources to master Reinforcement Learning
Almost every week new research and resources on RL come out. Knowledge needs to be constantly refreshed and updated with the latest trends. So today, we’re sharing 6 free sources to help you stay on track with RL:
1. A Survey of Continual Reinforcement Learning → https://arxiv.org/abs/2506.21872
Covers continual RL (CRL): how agents can keep learning and adapt to new tasks without forgetting past ones. It analyses methods, benchmarks, evaluation metrics &challenges
2. The Deep Reinforcement Learning course by Hugging Face → https://huggingface.co/learn/deep-rl-course/unit0/introduction
This is a popular free course, regularly updated. Includes community interaction, exercises, leaderboards, etc.
3. Reinforcement Learning Specialization (Coursera, University of Alberta) → https://www.coursera.org/specializations/reinforcement-learning
A 4-course series introducing foundational RL, implementing different algorithms, culminating in a capstone. It's a great structured path
4. A Technical Survey of Reinforcement Learning Techniques for LLMs → https://huggingface.co/papers/2507.04136
Looks at how RL is being used for/with LLMs for alignment, reasoning, preference signals, etc. Covers methods like RLHF, RLAIF, DPO, PPO, GRPO & applications from code gen to tool use
5. A Survey of Reinforcement Learning for Software Engineering → https://arxiv.org/abs/2507.12483
Good if you're interested in RL-applied domains. Examines how RL is used in software engineering tasks: maintenance, development, evaluation. Covering 115 papers since DRL introduction, it summarizes trends, gaps & challenges
6. A Survey of Reinforcement Learning for LRMs → https://arxiv.org/abs/2509.08827
Tracks the way from LLMs to LRMs via RL. Covers reward design, policy optimization, use cases and future approaches like continual, memory, model-based RL and more
If you liked this, subscribe to The Turing Post https://www.turingpost.com/subscribe