Ksenia Se

Kseniase

AI & ML interests

None yet

Recent Activity

replied to their post 5 days ago
10 awesome advanced LoRA approaches Low-Rank Adaptation (LoRA) is the go-to method for efficient model fine-tuning that adds small low-rank matrices instead of retraining full models. The field isn’t standing still – new LoRA variants push the limits of efficiency, generalization, and personalization. So we’re sharing 10 of the latest LoRA approaches you should know about: 1. Mixture-of-LoRA-experts → https://huggingface.co/papers/2509.13878 Adds multiple low-rank adapters (LoRA) into a model’s layers, and a routing mechanism activates the most suitable ones for each input. This lets the model adapt better to new unseen conditions 2. Amortized Bayesian Meta-Learning for LoRA (ABMLL) → https://huggingface.co/papers/2508.14285 Balances global and task-specific parameters within a Bayesian framework to improve uncertainty calibration and generalization to new tasks without high memory or compute costs 3. AutoLoRA → https://huggingface.co/papers/2508.02107 Automatically retrieves and dynamically aggregates public LoRAs for stronger T2I generation 4. aLoRA (Activated LoRA) → https://huggingface.co/papers/2504.12397 Only applies LoRA after invocation, letting the model reuse the base model’s KV cache instead of recomputing the full turn’s KV cache. Efficient in multi-turn conversations 5. LiLoRA (LoRA in LoRA) → https://huggingface.co/papers/2508.06202 Shares the LoRA matrix A across tasks and additionally low-rank-decomposes matrix B to cut parameters in continual vision-text MLLMs 6. Sensitivity-LoRA → https://huggingface.co/papers/2509.09119 Dynamically assigns ranks to weight matrices based on their sensitivity, measured using second-order derivatives Read further below ↓ Also, subscribe to the Turing Post: https://www.turingpost.com/subscribe
posted an update 5 days ago
10 awesome advanced LoRA approaches Low-Rank Adaptation (LoRA) is the go-to method for efficient model fine-tuning that adds small low-rank matrices instead of retraining full models. The field isn’t standing still – new LoRA variants push the limits of efficiency, generalization, and personalization. So we’re sharing 10 of the latest LoRA approaches you should know about: 1. Mixture-of-LoRA-experts → https://huggingface.co/papers/2509.13878 Adds multiple low-rank adapters (LoRA) into a model’s layers, and a routing mechanism activates the most suitable ones for each input. This lets the model adapt better to new unseen conditions 2. Amortized Bayesian Meta-Learning for LoRA (ABMLL) → https://huggingface.co/papers/2508.14285 Balances global and task-specific parameters within a Bayesian framework to improve uncertainty calibration and generalization to new tasks without high memory or compute costs 3. AutoLoRA → https://huggingface.co/papers/2508.02107 Automatically retrieves and dynamically aggregates public LoRAs for stronger T2I generation 4. aLoRA (Activated LoRA) → https://huggingface.co/papers/2504.12397 Only applies LoRA after invocation, letting the model reuse the base model’s KV cache instead of recomputing the full turn’s KV cache. Efficient in multi-turn conversations 5. LiLoRA (LoRA in LoRA) → https://huggingface.co/papers/2508.06202 Shares the LoRA matrix A across tasks and additionally low-rank-decomposes matrix B to cut parameters in continual vision-text MLLMs 6. Sensitivity-LoRA → https://huggingface.co/papers/2509.09119 Dynamically assigns ranks to weight matrices based on their sensitivity, measured using second-order derivatives Read further below ↓ Also, subscribe to the Turing Post: https://www.turingpost.com/subscribe
replied to their post 12 days ago
6 Recent & free sources to master Reinforcement Learning Almost every week new research and resources on RL come out. Knowledge needs to be constantly refreshed and updated with the latest trends. So today, we’re sharing 6 free sources to help you stay on track with RL: 1. A Survey of Continual Reinforcement Learning → https://arxiv.org/abs/2506.21872 Covers continual RL (CRL): how agents can keep learning and adapt to new tasks without forgetting past ones. It analyses methods, benchmarks, evaluation metrics &challenges 2. The Deep Reinforcement Learning course by Hugging Face → https://huggingface.co/learn/deep-rl-course/unit0/introduction This is a popular free course, regularly updated. Includes community interaction, exercises, leaderboards, etc. 3. Reinforcement Learning Specialization (Coursera, University of Alberta) → https://www.coursera.org/specializations/reinforcement-learning A 4-course series introducing foundational RL, implementing different algorithms, culminating in a capstone. It's a great structured path 4. A Technical Survey of Reinforcement Learning Techniques for LLMs → https://huggingface.co/papers/2507.04136 Looks at how RL is being used for/with LLMs for alignment, reasoning, preference signals, etc. Covers methods like RLHF, RLAIF, DPO, PPO, GRPO & applications from code gen to tool use 5. A Survey of Reinforcement Learning for Software Engineering → https://arxiv.org/abs/2507.12483 Good if you're interested in RL-applied domains. Examines how RL is used in software engineering tasks: maintenance, development, evaluation. Covering 115 papers since DRL introduction, it summarizes trends, gaps & challenges 6. A Survey of Reinforcement Learning for LRMs → https://arxiv.org/abs/2509.08827 Tracks the way from LLMs to LRMs via RL. Covers reward design, policy optimization, use cases and future approaches like continual, memory, model-based RL and more If you liked this, subscribe to The Turing Post https://www.turingpost.com/subscribe
View all activity

Organizations

Turing Post's profile picture Journalists on Hugging Face's profile picture Social Post Explorers's profile picture Hugging Face Discord Community's profile picture Sandbox's profile picture