Join the conversation

Join the community of Machine Learners and AI enthusiasts.

Sign Up

All HF Hub posts

ginipickย 
posted an update about 22 hours ago
view post
Post
1040
๐Ÿค– AI Academic Paper Generator: Your Research Partner ๐ŸŽ“

Hello, researchers! Today I'm introducing my AI Academic Paper Generation System. This application is built with Streamlit and provides AI agents to assist with every stage of the academic research process.

ginipick/AgentX-Papers

โœจ Key Features

๐Ÿ“š Literature Research: AI reviews and summarizes relevant research
๐Ÿ“ Paper Outline: Generates a well-structured paper outline
โœ๏ธ Draft Writing: Creates a paper draft based on your research topic
๐Ÿ”— Citation Generation: Automatically generates academic citations
๐Ÿ–‹๏ธ Editing & Polishing: Checks grammar, context, and logical flow
๐ŸŒ Multilingual Support: Interface available in English and Korean

๐Ÿš€ How to Use

Enter basic information like research topic, paper title, and deadline
AI agents generate everything from literature review to final paper
Download your completed paper or consult with the chatbot for further assistance

๐Ÿ’ก What Makes It Special
This tool integrates all stages of academic research. Going beyond simple text generation, it mimics the actual research process to produce higher quality papers.
Visualization features and social media sharing options will be added in the next update! ๐Ÿ’ช

#AIResearch #AcademicWriting #ResearchAssistant #ArtificialIntelligence
openfreeย 
posted an update 1 day ago
view post
Post
2890
๐Ÿง  ThinkFlow: The Revolutionary Platform That Gives LLMs the Power to Think ๐Ÿš€

Hello AI community! We're excited to introduce you to ThinkFlow, an innovative service that transforms how language models solve problems. ๐ŸŽ‰
VIDraft/ThinkFlow-llama

โœจ What is ThinkFlow?
ThinkFlow is a groundbreaking platform that automatically applies step-by-step reasoning capabilities to existing LLM models without any modifications. It makes complex problem-solving transparent, allowing you to witness the model's thought process in real-time.

๐Ÿ” Key Features

Reasoning Without Model Modifications: Add step-by-step reasoning while utilizing existing LLMs as they are โš™๏ธ
Visualized Thinking Process: See exactly how the model analyzes and solves problems ๐Ÿ‘๏ธ
Before & After Comparison: Compare standard responses with reasoning-enhanced outputs in real-time ๐Ÿ“Š
Improved Accuracy: Deliver more accurate solutions for complex math and logic problems ๐Ÿ“ˆ
Educational Value: Teach students systematic approaches to problem-solving ๐Ÿ‘จโ€๐Ÿซ
User-Friendly Interface: Intuitive and easy-to-use UI for seamless experience ๐Ÿ–ฅ๏ธ

๐Ÿ’ก What Problems Can It Solve?
ThinkFlow is particularly effective for various domains including:

Complex mathematical problems ๐Ÿงฎ
Logic puzzles ๐Ÿงฉ
Questions requiring multi-step reasoning ๐Ÿค”
Scientific analysis challenges ๐Ÿ”ฌ
Complex decision-making processes ๐Ÿ“

๐Ÿ‘จโ€๐Ÿ’ป Technical Details
ThinkFlow is built on the meta-llama/Llama-3.1-8B-Instruct model and uses carefully designed prompt chains to guide the model through step-by-step thinking. Each reasoning step builds upon the results of previous steps, culminating in a comprehensive final answer.

๐Ÿ’ฌ Join Our Community!
If you have questions or suggestions about ThinkFlow, join our Discord community: https://discord.gg/openfreeai
Let's build better AI reasoning experiences together! ๐Ÿ’ช

#AI #LLM #ReasoningAI #ThinkFlow #HuggingFace #OpenSource #AIEducation
ยท
seawolf2357ย 
posted an update about 22 hours ago
view post
Post
1647
๐Ÿ“š Papers Leaderboard - See the Latest AI Research Trends at a Glance! โœจ

Hello, AI research community! Today I'm introducing a new tool for exploring research papers. Papers Leaderboard is an open-source dashboard that makes it easy to find and filter the latest AI research papers.

Heartsync/Papers-Leaderboard

๐ŸŒŸ Key Features

Date Filtering: View only papers published within a specific timeframe (from May 5, 2023 to present)
Title Search: Quickly find papers containing your keywords of interest
Abstract Search: Explore paper content more deeply by searching for keywords within abstracts
Automatic Updates: The database is updated with the latest papers every hour

๐Ÿ’ก How to Use It?

Select a start date and end date
Enter keywords you want to find in titles or abstracts
Adjust the maximum number of search results for abstract searches
Results are displayed neatly in table format
aiqtechย 
posted an update about 22 hours ago
view post
Post
1273
๐ŸŒ AI Token Visualization Tool with Perfect Multilingual Support

Hello! Today I'm introducing my Token Visualization Tool with comprehensive multilingual support. This web-based application allows you to see how various Large Language Models (LLMs) tokenize text.

aiqtech/LLM-Token-Visual

โœจ Key Features

๐Ÿค– Multiple LLM Tokenizers: Support for Llama 4, Mistral, Gemma, Deepseek, QWQ, BERT, and more
๐Ÿ”„ Custom Model Support: Use any tokenizer available on HuggingFace
๐Ÿ“Š Detailed Token Statistics: Analyze total tokens, unique tokens, compression ratio, and more
๐ŸŒˆ Visual Token Representation: Each token assigned a unique color for visual distinction
๐Ÿ“‚ File Analysis Support: Upload and analyze large files

๐ŸŒ Powerful Multilingual Support
The most significant advantage of this tool is its perfect support for all languages:

๐Ÿ“ Asian languages including Korean, Chinese, and Japanese fully supported
๐Ÿ”ค RTL (right-to-left) languages like Arabic and Hebrew supported
๐Ÿˆบ Special characters and emoji tokenization visualization
๐Ÿงฉ Compare tokenization differences between languages
๐Ÿ’ฌ Mixed multilingual text processing analysis

๐Ÿš€ How It Works

Select your desired tokenizer model (predefined or HuggingFace model ID)
Input multilingual text or upload a file for analysis
Click 'Analyze Text' to see the tokenized results
Visually understand how the model breaks down various languages with color-coded tokens

๐Ÿ’ก Benefits of Multilingual Processing
Understanding multilingual text tokenization patterns helps you:

Optimize prompts that mix multiple languages
Compare token efficiency across languages (e.g., English vs. Korean vs. Chinese token usage)
Predict token usage for internationalization (i18n) applications
Optimize costs for multilingual AI services

๐Ÿ› ๏ธ Technology Stack

Backend: Flask (Python)
Frontend: HTML, CSS, JavaScript (jQuery)
Tokenizers: ๐Ÿค— Transformers library
Kseniaseย 
posted an update about 15 hours ago
view post
Post
1391
11 new types of RAG

RAG is evolving fast, keeping pace with cutting-edge AI trends. Today it becomes more agentic and smarter at navigating complex structures like hypergraphs.

Here are 11 latest RAG types:

1. InstructRAG -> InstructRAG: Leveraging Retrieval-Augmented Generation on Instruction Graphs for LLM-Based Task Planning (2504.13032)
Combines RAG with a multi-agent framework, using a graph-based structure, an RL agent to expand task coverage, and a meta-learning agent for better generalization

2. CoRAG (Collaborative RAG) -> CoRAG: Collaborative Retrieval-Augmented Generation (2504.01883)
A collaborative framework that extends RAG to settings where clients train a shared model using a joint passage store

3. ReaRAG -> ReaRAG: Knowledge-guided Reasoning Enhances Factuality of Large Reasoning Models with Iterative Retrieval Augmented Generation (2503.21729)
It uses a Thought-Action-Observation loop to decide at each step whether to retrieve information or finalize an answer, reducing unnecessary reasoning and errors

4. MCTS-RAG -> MCTS-RAG: Enhancing Retrieval-Augmented Generation with Monte Carlo Tree Search (2503.20757)
Combines RAG with Monte Carlo Tree Search (MCTS) to help small LMs handle complex, knowledge-heavy tasks

5. Typed-RAG - > Typed-RAG: Type-aware Multi-Aspect Decomposition for Non-Factoid Question Answering (2503.15879)
Improves answers on open-ended questions by identifying question types (a debate, personal experience, or comparison) and breaking it down into simpler parts

6. MADAM-RAG -> Retrieval-Augmented Generation with Conflicting Evidence (2504.13079)
A multi-agent system where models debate answers over multiple rounds and an aggregator filters noise and misinformation

7. HM-RAG -> HM-RAG: Hierarchical Multi-Agent Multimodal Retrieval Augmented Generation (2504.12330)
A hierarchical multi-agent RAG framework that uses 3 agents: one to split queries, one to retrieve across multiple data types (text, graphs and web), and one to merge and refine answers

8. CDF-RAG -> CDF-RAG: Causal Dynamic Feedback for Adaptive Retrieval-Augmented Generation (2504.12560)
Works with causal graphs and enables multi-hop causal reasoning, refining queries. It validates responses against causal pathways

To explore what is Causal AI, read our article: https://www.turingpost.com/p/causalai

Subscribe to the Turing Post: https://www.turingpost.com/subscribe

Read further ๐Ÿ‘‡
  • 1 reply
ยท
openfreeย 
posted an update about 2 hours ago
view post
Post
207
๐Ÿ“Š Papers Impact: Instant AI Grading for Your Research Papers! ๐Ÿš€

๐ŸŒŸ Introduction
Hello, AI research community! ๐ŸŽ‰
Introducing Papers Impact - the revolutionary AI tool that automatically grades and predicts the potential impact of research papers! ๐Ÿง ๐Ÿ’ก

VIDraft/PapersImpact

โœจ Key Feature: Instant Paper Grading
The core functionality is brilliantly simple: Just enter an arXiv paper ID or URL, and our AI instantly analyzes and grades the paper's potential academic impact! No need to read through the entire paper yourself - our system automatically evaluates the title and abstract to generate a normalized impact score between 0 and 1.
๐ŸŽฏ How It Works

Enter Paper ID or URL: Simply paste an arXiv ID (e.g., "2504.11651") or full URL
Automatic Fetching: The system retrieves the paper's title and abstract
AI Analysis: Our advanced LLaMA-based transformer model analyzes the content
Instant Grading: Receive an impact score and corresponding letter grade in seconds!

๐Ÿ’ก Who Can Benefit?

๐Ÿ”ฌ Researchers: Pre-assess your paper before submission
๐Ÿ“š Students: Quickly gauge the quality of papers for literature reviews
๐Ÿซ Educators: Objectively evaluate student research
๐Ÿ“Š Research Managers: Prioritize which papers to read in depth
๐Ÿงฉ Journal Editors: Get an AI second opinion on submissions

๐Ÿš€ Technical Details
Our model is trained on an extensive dataset of published papers in CS.CV, CS.CL, and CS.AI fields, using NDCG optimization with Sigmoid activation and MSE loss. It's been rigorously cross-validated against historical citation data to ensure accurate impact predictions.
merterbakย 
posted an update 2 days ago
view post
Post
1693
Hereโ€™s a cool paper I found: โ€œMassive Image Embedding Benchmark (MIEB).โ€ It is a new tool to test how good image embedding models are. It has 130 different tasks grouped into 8 categories, like image search, classification, clustering similar images, answering questions based on images, and understanding documents. It even covers 38 different languages.

The authors tested 50 models and found that no single model was best at everything. Some models were great at recognizing text inside images but struggled to handle complicated tasks like matching images and text that appear together.

Paper: https://arxiv.org/pdf/2504.10471v1
Code: https://github.com/embeddings-benchmark/mteb
  • 2 replies
ยท
zhiminyย 
posted an update 1 day ago
view post
Post
991
# ๐Ÿš€ SE Arena: Evaluating Foundation Models for Software Engineering

**SE Arena** is the first open-source platform for evaluating foundation models in real-world software engineering workflows.

## What makes it unique?

- **RepoChat**: Automatically injects repository context (issues, commits, PRs) into conversations for more realistic evaluations
- **Multi-round interactions**: Tests models through iterative workflows, not just single prompts
- **Novel metrics**: Includes a "consistency score" that measures model determinism through self-play matches

Try it now: SE-Arena/Software-Engineering-Arena

## Why it matters

Traditional evaluation frameworks don't capture how developers actually use models in their daily work. SE Arena creates a testing environment that mirrors real engineering workflows, helping you choose the right model for your specific software development needs.

From debugging to requirement refinement, see which models truly excel at software engineering tasks!
nyuuzyouย 
posted an update 2 days ago
view post
Post
1515
๐Ÿฆ… SmolLM2-Eagle Collection - nyuuzyou/smollm2-eagle-680263bf97f0c7e6bbe4936b

Collection of fine-tuned bilingual language models featuring:
- Models in three parameter sizes: 135M, 360M, and 1.7B based on HuggingFaceTB's SmolLM2 models
- Both standard and GGUF formats for flexible deployment in llama.cpp and Ollama
- Fine-tuned on nyuuzyou/EagleSFT dataset (536,231 Russian-English QA pairs derived from 739k+ real user queries)
- Experimental Russian language capabilities while maintaining English performance
- Limited Russian capabilities due to SFT-only approach without Russian pre-training
- Environmental impact: ~19.75 kg CO2eq

This collection provides compact models for research on bilingual language capabilities, resource-constrained environments, and educational applications. Not recommended for production use due to experimental nature and inherent limitations. Available under Apache 2.0 license.
  • 1 reply
ยท
JLouisBizย 
posted an update 1 day ago
view post
Post
984
Back to LLM integration.

ClickDefine.sh -- quickly define or explain anything within your whole desktop environment

You only need to run the model locally, maybe with the **llama.cpp** or **ollama**

- https://github.com/ggml-org/llama.cpp
- https://ollama.com/download

And you get universal explaining tool that works anywhere on your X Org Desktop (on operating systems which are usually Fully Free Software like Debian GNU/Linux)

ClickDefine - Interactive Text Processor Script for Iterative LLM Query Handling:
https://hyperscope.link/9/6/0/9/8/ClickDefine-Interactive-Text-Processor-Script-for-Iterative-LLM-Query-Handling-96098.html

Watch the demonstration here: https://www.youtube.com/watch?v=mQxCYAiReu0&t=2s