Dataset Viewer
Duplicate
The dataset viewer is not available for this split.
Cannot extract the features (columns) for the split 'train' of the config 'default' of the dataset.
Error code:   FeaturesError
Exception:    ArrowTypeError
Message:      ("Expected bytes, got a 'int' object", 'Conversion failed for column metadata with type object')
Traceback:    Traceback (most recent call last):
                File "/usr/local/lib/python3.12/site-packages/datasets/packaged_modules/json/json.py", line 242, in _generate_tables
                  pa_table = paj.read_json(
                             ^^^^^^^^^^^^^^
                File "pyarrow/_json.pyx", line 342, in pyarrow._json.read_json
                File "pyarrow/error.pxi", line 155, in pyarrow.lib.pyarrow_internal_check_status
                File "pyarrow/error.pxi", line 92, in pyarrow.lib.check_status
              pyarrow.lib.ArrowInvalid: JSON parse error: Missing a name for object member. in row 0
              
              During handling of the above exception, another exception occurred:
              
              Traceback (most recent call last):
                File "/src/services/worker/src/worker/job_runners/split/first_rows.py", line 243, in compute_first_rows_from_streaming_response
                  iterable_dataset = iterable_dataset._resolve_features()
                                     ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
                File "/usr/local/lib/python3.12/site-packages/datasets/iterable_dataset.py", line 4195, in _resolve_features
                  features = _infer_features_from_batch(self.with_format(None)._head())
                                                        ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
                File "/usr/local/lib/python3.12/site-packages/datasets/iterable_dataset.py", line 2533, in _head
                  return next(iter(self.iter(batch_size=n)))
                         ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
                File "/usr/local/lib/python3.12/site-packages/datasets/iterable_dataset.py", line 2711, in iter
                  for key, pa_table in ex_iterable.iter_arrow():
                                       ^^^^^^^^^^^^^^^^^^^^^^^^
                File "/usr/local/lib/python3.12/site-packages/datasets/iterable_dataset.py", line 2249, in _iter_arrow
                  yield from self.ex_iterable._iter_arrow()
                File "/usr/local/lib/python3.12/site-packages/datasets/iterable_dataset.py", line 494, in _iter_arrow
                  for key, pa_table in iterator:
                                       ^^^^^^^^
                File "/usr/local/lib/python3.12/site-packages/datasets/iterable_dataset.py", line 384, in _iter_arrow
                  for key, pa_table in self.generate_tables_fn(**gen_kwags):
                                       ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
                File "/usr/local/lib/python3.12/site-packages/datasets/packaged_modules/json/json.py", line 283, in _generate_tables
                  pa_table = pa.Table.from_pandas(df, preserve_index=False)
                             ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
                File "pyarrow/table.pxi", line 4795, in pyarrow.lib.Table.from_pandas
                File "/usr/local/lib/python3.12/site-packages/pyarrow/pandas_compat.py", line 637, in dataframe_to_arrays
                  arrays = [convert_column(c, f)
                            ^^^^^^^^^^^^^^^^^^^^
                File "/usr/local/lib/python3.12/site-packages/pyarrow/pandas_compat.py", line 625, in convert_column
                  raise e
                File "/usr/local/lib/python3.12/site-packages/pyarrow/pandas_compat.py", line 619, in convert_column
                  result = pa.array(col, type=type_, from_pandas=True, safe=safe)
                           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
                File "pyarrow/array.pxi", line 365, in pyarrow.lib.array
                File "pyarrow/array.pxi", line 91, in pyarrow.lib._ndarray_to_array
                File "pyarrow/error.pxi", line 92, in pyarrow.lib.check_status
              pyarrow.lib.ArrowTypeError: ("Expected bytes, got a 'int' object", 'Conversion failed for column metadata with type object')

Need help to make the dataset viewer work? Make sure to review how to configure the dataset viewer, and open a discussion for direct support.

AutoResearchBench

This repository hosts the obfuscated benchmark bundle for AutoResearchBench.

The published file uses public reversible obfuscation for benchmark release. It lowers casual web exposure, but it is not strong access control.

Please do not repost decrypted questions or answers in plain text or images online.

Download

export HF_TOKEN=your_hf_token  # required if this dataset repo is private
curl -L   -H "Authorization: Bearer ${HF_TOKEN}"   -o AutoResearchBench.jsonl.obf.json   https://huggingface.co/datasets/Lk123/AutoResearchBench/resolve/main/AutoResearchBench.jsonl.obf.json

Decrypt

Use decrypt_benchmark.py from the benchmark code repository:

python3 decrypt_benchmark.py   --input-file AutoResearchBench.jsonl.obf.json   --output-file AutoResearchBench.jsonl

Run inference and evaluation on the decrypted AutoResearchBench.jsonl, not on the .obf.json bundle directly.

❤️ Citing Us

If you find this repository or our work useful, please consider giving a star ⭐ and or citing our work, which would be greatly appreciated:

@misc{xiong2026autoresearchbenchbenchmarkingaiagents,
      title={AutoResearchBench: Benchmarking AI Agents on Complex Scientific Literature Discovery}, 
      author={Lei Xiong and Kun Luo and Ziyi Xia and Wenbo Zhang and Jin-Ge Yao and Zheng Liu and Jingying Shao and Jianlyu Chen and Hongjin Qian and Xi Yang and Qian Yu and Hao Li and Chen Yue and Xiaan Du and Yuyang Wang and Yesheng Liu and Haiyu Xu and Zhicheng Dou},
      year={2026},
      eprint={2604.25256},
      archivePrefix={arXiv},
      primaryClass={cs.AI},
      url={https://arxiv.org/abs/2604.25256}, 
}
Downloads last month
83

Paper for Lk123/AutoResearchBench