Denys's picture
1 2

Denys

JoniJOST
ยท

AI & ML interests

None yet

Recent Activity

Organizations

None yet

JoniJOST's activity

reacted to m-ric's post with ๐Ÿ”ฅ 13 days ago
view post
Post
3904
๐—ง๐—ต๐—ฒ ๐—›๐˜‚๐—ฏ ๐˜„๐—ฒ๐—น๐—ฐ๐—ผ๐—บ๐—ฒ๐˜€ ๐—ฒ๐˜…๐˜๐—ฒ๐—ฟ๐—ป๐—ฎ๐—น ๐—ถ๐—ป๐—ณ๐—ฒ๐—ฟ๐—ฒ๐—ป๐—ฐ๐—ฒ ๐—ฝ๐—ฟ๐—ผ๐˜ƒ๐—ถ๐—ฑ๐—ฒ๐—ฟ๐˜€!

โœ… Hosting our own inference was not enough: now the Hub 4 new inference providers: fal, Replicate, SambaNova Systems, & Together AI.

Check model cards on the Hub: you can now, in 1 click, use inference from various providers (cf video demo)

Their inference can also be used through our Inference API client. There, you can use either your custom provider key, or your HF token, then billing will be handled directly on your HF account, as a way to centralize all expenses.

๐Ÿ’ธ Also, PRO users get 2$ inference credits per month!

Read more in the announcement ๐Ÿ‘‰ https://huggingface.co/blog/inference-providers
  • 1 reply
ยท
reacted to cfahlgren1's post with ๐Ÿค 13 days ago
view post
Post
1906
If you haven't seen yet, we just released Inference Providers ๐Ÿ”€

> 4 new serverless inference providers on the Hub ๐Ÿคฏ
> Use your HF API key or personal key with all providers ๐Ÿ”‘
> Chat with Deepseek R1, V3, and more on HF Hub ๐Ÿ‹
> We support Sambanova, TogetherAI, Replicate, and Fal.ai ๐Ÿ’ช

Best of all, we don't charge any markup on top of the provider ๐Ÿซฐ Have you tried it out yet? HF Pro accounts get $2 of free usage for the provider inference.
reacted to dylanebert's post with ๐Ÿค 13 days ago
reacted to prithivMLmods's post with ๐Ÿค 13 days ago
view post
Post
5099
Deepswipe by
.
.
.
. Deepseek๐Ÿฌ๐Ÿ—ฟ






Everything is now in recovery. ๐Ÿ“‰๐Ÿ“ˆ
ยท
reacted to singhsidhukuldeep's post with ๐Ÿค 13 days ago
view post
Post
1663
Groundbreaking Research Alert: Can Large Language Models Really Understand Personal Preferences?

A fascinating new study from researchers at University of Notre Dame, Xi'an Jiaotong University, and Universitรฉ de Montrรฉal introduces PERRECBENCH - a novel benchmark for evaluating how well Large Language Models (LLMs) understand user preferences in recommendation systems.

Key Technical Insights:
- The benchmark eliminates user rating bias and item quality factors by using relative ratings and grouped ranking approaches
- Implements three distinct ranking methods: pointwise rating prediction, pairwise comparison, and listwise ranking
- Evaluates 19 state-of-the-art LLMs including Claude-3.5, GPT-4, Llama-3, Mistral, and Qwen models
- Uses Kendall's tau correlation to measure ranking accuracy
- Incorporates BM25 retriever with configurable history items (k=4 by default)

Notable Findings:
- Current LLMs struggle with true personalization, achieving only moderate correlation scores
- Larger models don't always perform better - challenging conventional scaling laws
- Pairwise and listwise ranking methods outperform pointwise approaches
- Open-source models like Mistral-123B and Llama-3-405B compete well with proprietary models
- Weight merging strategy shows promise for improving personalization capabilities

The research reveals that while LLMs excel at many tasks, they still face significant challenges in understanding individual user preferences. This work opens new avenues for improving personalized recommendation systems and highlights the importance of developing better evaluation methods.

A must-read for anyone interested in LLMs, recommender systems, or personalization technology. The team has made their benchmark and code publicly available for further research.
reacted to mkurman's post with ๐Ÿค 13 days ago
view post
Post
2755
Ok, my 14B DeepSeek R1 merge with Qwen2.5 1M is really hot right nowโ€”it's got 2.6k downloads! It's sitting pretty as the top trending model on the third page. ๐Ÿ”ฅ

Check it out if you haven't already!
mkurman/Qwen2.5-14B-DeepSeek-R1-1M
ยท
reacted to csabakecskemeti's post with ๐Ÿค 13 days ago