File size: 6,652 Bytes
4dcba74 2dba94f 579db59 5d1208d 2dba94f 4dcba74 9da9431 ad604a4 9da9431 1685572 2dba94f dbd7a03 3489875 2dba94f 579db59 5d1208d 808487a 3621285 5d1208d 2dba94f 1685572 2dba94f 1f3e8c7 1685572 2dba94f 579db59 1685572 579db59 2dba94f 1685572 2dba94f 302f5c8 2dba94f 4dcba74 2dba94f |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 |
TITLE = """<h1 align="center" id="space-title">π Online-Mind2Web Leaderboard</h1>"""
LINKS = """
<div align="center">
<a href="https://arxiv.org/abs/2504.01382">Paper</a> |
<a href="https://tiancixue.notion.site/An-Illusion-of-Progress-Assessing-the-Current-State-of-Web-Agents-1ac6cd2b9aac80719cd6f68374aaf4b4?pvs=4">Blog</a> |
<a href="https://github.com/OSU-NLP-Group/Online-Mind2Web">Code</a> |
<a href="https://huggingface.co/datasets/osunlp/Online-Mind2Web">Data</a>
</div>
"""
INTRODUCTION_TEXT = """
Online-Mind2Web is a benchmark designed to evaluate the real-world performance of web agents on live websites, featuring 300 tasks across 136 popular sites in diverse domains.
Based on the number of steps required by human annotators, tasks are divided into three difficulty levels: Easy (1β5 steps), Medium (6β10 steps), and Hard (11+ steps).
"""
LEADERBOARD_TEXT = """
### Leaderboard
Our goal is to conduct a rigorous assessment of the current state of web agents. We maintain two leaderboardsβone for automatic evaluation and another for human evaluation.
Please click "Submission Guideline" for details.
"""
SUBMISSION_TEXT = """
## Submissions
Participants are invited to submit your agent's trajectory to test. The submissions will be evaluated based on our auto-eval.
### Format of submission
Submissions must include a sequence of images (i.e., screenshots in the trajectory) and a result.json file for each task. The JSON file should contain the fields: "Task", "Task_id", and "action_history". You can refer to an example of the submission files.
"""
EVALUATION_DETAILS = """
In certain scenarios, testing on the full Online-Mind2Web dataset may not be feasible due to cost, privacy, or legal constraints. To facilitate fair and apple-to-apple comparisons, we release both our human evaluation labels and auto-eval details.
- **Human Evaluation**: Task-level human evaluation labels are provided in the [file](https://github.com/OSU-NLP-Group/Online-Mind2Web/tree/main/data/human_label.json).
- **Auto-Evaluation**: The results of WebJudge are available in the [folder](https://github.com/OSU-NLP-Group/Online-Mind2Web/tree/main/data/evaluation_results)."""
CITATION_BUTTON_LABEL = "Copy the following snippet to cite these results. Note: Online-Mind2Web is derived from the original Mind2Web dataset. We kindly ask that you cite both the original and this work when using or referencing the data."
CITATION_BUTTON_TEXT = r"""
@article{xue2025illusionprogressassessingcurrent,
title={An Illusion of Progress? Assessing the Current State of Web Agents},
author={Tianci Xue and Weijian Qi and Tianneng Shi and Chan Hee Song and Boyu Gou and Dawn Song and Huan Sun and Yu Su},
year={2025},
eprint={2504.01382},
archivePrefix={arXiv},
primaryClass={cs.AI},
url={https://arxiv.org/abs/2504.01382},
}
@inproceedings{deng2023mind2web,
author = {Deng, Xiang and Gu, Yu and Zheng, Boyuan and Chen, Shijie and Stevens, Sam and Wang, Boshi and Sun, Huan and Su, Yu},
booktitle = {Advances in Neural Information Processing Systems},
editor = {A. Oh and T. Naumann and A. Globerson and K. Saenko and M. Hardt and S. Levine},
pages = {28091--28114},
publisher = {Curran Associates, Inc.},
title = {Mind2Web: Towards a Generalist Agent for the Web},
url = {https://proceedings.neurips.cc/paper_files/paper/2023/file/5950bf290a1570ea401bf98882128160-Paper-Datasets_and_Benchmarks.pdf},
volume = {36},
year = {2023}
}
"""
SUBMIT_INTRODUCTION = """
You should use the script provided in our GitHub repository to obtain automatic evaluation results on your own and submit them along with all trajectories.
To ensure the authenticity and reliability of the reported results, we will also conduct a verification of auto-eval results.
If you have conducted your own human evaluation, please also attach your human-eval results. We will spot-check these before adding them to the human-eval table.
## Important Notes for Reliable Evaluation:
- To enable fair comparisons, please ensure that each task starts from the specified website in our benchmark. Starting from Google Search or alternative websites can lead agents to use different websites to solve the task, resulting in varying difficulty levels and potentially skewed evaluation results.
- The action history should contain only the actions taken by the agent to complete the task (e.g., clicking elements and Typing text). Please avoid including the final response, as it may contain hallucinated content, leading to a high rate of false positives.
- WebJudge powered by o4-mini demonstrates a higher alignment with human judgment, achieving an average agreement rate of 85.7% and maintaining a narrow success rate gap of just 3.8%. Therefore, please use o4-mini as the backbone for automatic evaluation.
## β Please submit the trajectory file with the following format:
The result of each task is stored in a folder named as its `task_id`, containing:
- `trajectory/`: Stores screenshots of each step.
- `result.json`: Task metadata and action history.
Here is an [example](https://github.com/OSU-NLP-Group/Online-Mind2Web/tree/main/data/example/fb7b4f784cfde003e2548fdf4e8d6b4f) of the format.
**Structure:**
```
main_directory/
βββ task_id/
βββ result.json
βββ trajectory/
βββ 0_screenshot.png
βββ 1_screenshot.png
βββ ...
```
**`result.json` format:**
```json
{
"task_id": 123,
"task": "abc",
"action_history": ["abc", "xyz", "..."]
}
```
**`human_result.json` format:**
```json
[
{
"task_id": 123,
"task": "abc",
"human_label": 0 or 1 (failure or success)
},
{
"task_id": 456,
"task": "def",
"human_label": 0 or 1 (failure or success)
},
]
```
Please email your agent's name, model family, and organization to xue.681@osu.edu, and include the trajectory directory and auto-eval results file as attachments (optional: human evaluation results).
"""
DATA_DATASET = """## More Statistics for Online-Mind2Web Benchmark
"""
def format_error(msg):
return f"<p style='color: red; font-size: 20px; text-align: center;'>{msg}</p>"
def format_warning(msg):
return f"<p style='color: orange; font-size: 20px; text-align: center;'>{msg}</p>"
def format_log(msg):
return f"<p style='color: green; font-size: 20px; text-align: center;'>{msg}</p>"
def model_hyperlink(link, model_name):
return f'<a target="_blank" href="{link}" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">{model_name}</a>'
|