File size: 7,546 Bytes
02afdc0 aa56335 02afdc0 37dd96b 02afdc0 37dd96b 02afdc0 37dd96b 02afdc0 37dd96b 02afdc0 aa56335 02afdc0 aa56335 02afdc0 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135 136 137 138 139 140 141 142 143 144 145 146 147 148 149 150 151 152 153 154 |
---
language:
- en
license: apache-2.0
task_categories:
- question-answering
tags:
- agent
- benchmark
- tool-use
- korean
---
<p align="center">
<img src="banner.png" />
</p>
# **๐ฐ๐ท Ko-AgentBench v1**
**"Korean Agent Benchmark Project"**
**English | [ํ๊ตญ์ด](README.md)**
As AI agents become more sophisticated, it has become crucial to precisely measure their performance under conditions similar to real-world environments. However, most benchmarks are designed based on English-speaking environments, which limits their ability to reflect Korea's unique usage contexts.
To address this issue, we have developed a high-quality agent benchmark specialized for the Korean real-world usage environment.
# Ko-AgentBench Key Features โจ
**1. Step-by-step Task Design**
We have comprehensively analyzed agent capabilities across 7 levels, from simple tool calls to long-term contextual abilities and robustness handling capabilities.
**2. 18 Korean-specific APIs and High-quality Scenarios Tailored to Real-life Environments**
Based on APIs from Korean real-world usage environments such as Naver, Maps, Kakao, and websites, we have implemented realistic problem-solving scenarios closely related to domestic users' daily lives, such as 'appointment booking' and 'blog review search'.
**3. Cache-based Iterative Evaluation and Robustness Testing**
We solve chronic problems of existing benchmarks, such as 'information attribute inconsistency changes'.
By improving failed API responses, we ensure benchmark consistency and reliability.
By evaluating error recognition/response capabilities (strategies) in intentional error situations, we select models that operate stably even in real-world environments.
**4. Step-specific Precision Metrics**
We evaluate the necessity/requirements of problem-solving step by step, including tool selection, parameter configuration, and data flow. Through this, we quantitatively identify the strengths and weaknesses of models.
## **Data Loading**
```bash
from datasets import load_dataset
# Load specific level
dataset = load_dataset("Hugging-Face-KREW/Ko-AgentBench", data_files="L1.json")
# Or load all levels
dataset = load_dataset("Hugging-Face-KREW/Ko-AgentBench", data_files="*.json")
```
# Dataset Overview
- Define task classification system for agent benchmark design
- Design to evaluate agent's tool calling capabilities in a step-by-step manner
## Dataset Scope
- Evaluation Target: Open-weight sLLM (supports tool calling), Commercial APIs
- Evaluation Scope: Agent tool calling performance in single-turn and multi-turn conversation situations
- Applied APIs: 18 Korean-specific open APIs
# Task Levels
## Single-Turn
**L1. Single Tool Call**
- Goal: Verify the most basic API calling capability
- Description: Check if the given tool can be executed with correct parameters
- Feature: Evaluate "accuracy only" by performing requests with specified API names or natural language requests as-is
- Example: "Search for 'Rapid Current' using Naver Book API and tell me the price."
- Example: "Tell me the price of the 'Rapid Current' book"
**L2. Tool Selection**
- Goal: Verify the ability to select the optimal API among multiple candidate tools
- Description: Users make requests in natural language, and the model must select the most suitable tool from the given tool list
- Feature: Evaluate accurate tool mapping with input natural language
- Example: "Check the price of the 'All Back English Middle 2-1 Cheonjae (Kim)' book."
- Candidate tools: `hotel_booking_api`, `aladin_books_api`
- Candidate tools must have no mutual correlation.
**L3. Sequential Tool Reasoning**
- Goal: Verify planning and execution capabilities through multi-step reasoning
- Description: Check if a correct pipeline can be constructed by connecting the results of one tool as input to another tool
- Feature: Evaluate "planned chain-of-tools" rather than simple calls
- Example: "Tell me when the Instax11 I bought from 11st Amazon will be delivered"
- Candidate tools: `11st_order_api`, `customs_api`, `cj_delivery_api`
- Tools must be callable sequentially (11st delivery number inquiry โ customs clearance โ courier company)
**L4. Parallel Tool Reasoning**
- Goal: Collect information in parallel and derive conclusions by synthesizing it
- Description: Simultaneously call multiple independent tools, compare and analyze results, then produce final answers
- Feature: Evaluate multi-source aggregation (information synthesis and comparison ability)
- Example: "Check the stock of the 'Hanroro Grapefruit Apricot Club' book."
- Candidate tools: `kyobo_books_api`, `aladin_books_api`
- Expected answer: There are 12 books at Kyobo Book Centre and 18 books at Aladin, totaling 30 books.
- At this time, candidate tools must handle the same function in parallel.
**L5. Error Handling and Robustness**
- Goal: Verify coping ability in error situations
- Description: Evaluate how various failure modes are handled, not just "failed"
- **Sub-items:**
- A. Request for additional questions
- Guide users to make clearer requests when information is insufficient
- B. Hallucination prevention
- Prohibit calling non-existent APIs
- Prohibit "pretending to succeed" answers when failed
- C. Fallback maneuvers
- Whether alternative APIs with the same function can be utilized when specific API errors occur
- Example: "When Naver Movie API call fails โ Report 'API call failed' or call Kakao Movie API as alternative"
## Multi-Turn
**L6. Efficient Tool Utilization**
- Goal: Verify the ability to efficiently reuse previous tool results
- Description: While recalling APIs in all situations is accurate, it's inefficient in terms of cost and delay. Conversely, unconditionally reusing old information also causes accuracy problems.
- Feature: Evaluate whether reasonable choices can be made between "recall vs reuse"
- Example:
- User: "Compare Coupang and Naver prices." โ Result: Coupang 80, Naver 85
- User: "What was the Naver price?"
- Correct answer: 85 (utilize past information, avoid unnecessary recalls)
- Wrong answer: Call API again or "I don't know"
**L7. Long-Context Reasoning**
- Goal: Verify the ability to maintain long-term context in multi-turn conversations
- Description: Remember information from several turns ago and correctly perform tool calling by connecting it with new questions
- Example:
- User's first question: "I'm going to travel to Jeju Island."
- Later: "How's the weather?" โ Call weather API using Jeju Island context
- (Additional turn) "If it rains, find places where I can buy an umbrella." โ Utilize all previous Jeju Island + weather context
## Links
You can check more detailed information about Ko-AgentBench.
- ๐ [Live Leaderboard](https://huggingface.co/spaces/huggingface-KREW/Ko-AgentBench)
- ๐ [Dataset](https://huggingface.co/datasets/huggingface-KREW/Ko-AgentBench)
- ๐ [Github](https://github.com/Hugging-Face-KREW/Ko-AgentBench)
## Contact
If you have any questions about the dataset and benchmark, please contact us!
Hugging Face KREW is a Korean non-profit research organization that strives to deeply understand artificial intelligence through Hugging Face and contribute to open source.
- โ๐ป Blog: [KREW-blog](https://hugging-face-krew.github.io/)
- ๐ฆ HuggingFace Community: [@huggingface-KREW](https://huggingface.co/huggingface-KREW)
- ๐ผ LinkedIn: [Hugging Face KREW](https://www.linkedin.com/company/hugging-face-krew/)
|