Datasets:

Modalities:
Image
ArXiv:
Dataset Preview
Duplicate
The full dataset viewer is not available (click to read why). Only showing a preview of the rows.
The dataset generation failed
Error code:   DatasetGenerationError
Exception:    ArrowCapacityError
Message:      array cannot contain more than 2147483646 bytes, have 5368709120
Traceback:    Traceback (most recent call last):
                File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/builder.py", line 1586, in _prepare_split_single
                  writer.write(example, key)
                File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/arrow_writer.py", line 553, in write
                  self.write_examples_on_file()
                File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/arrow_writer.py", line 511, in write_examples_on_file
                  self.write_batch(batch_examples=batch_examples)
                File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/arrow_writer.py", line 631, in write_batch
                  self.write_table(pa_table, writer_batch_size)
                File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/arrow_writer.py", line 646, in write_table
                  pa_table = embed_table_storage(pa_table)
                File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/table.py", line 2248, in embed_table_storage
                  arrays = [
                File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/table.py", line 2249, in <listcomp>
                  embed_array_storage(table[name], feature, token_per_repo_id=token_per_repo_id)
                File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/table.py", line 1795, in wrapper
                  return pa.chunked_array([func(chunk, *args, **kwargs) for chunk in array.chunks])
                File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/table.py", line 1795, in <listcomp>
                  return pa.chunked_array([func(chunk, *args, **kwargs) for chunk in array.chunks])
                File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/table.py", line 2124, in embed_array_storage
                  return feature.embed_storage(array, token_per_repo_id=token_per_repo_id)
                File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/features/image.py", line 281, in embed_storage
                  bytes_array = pa.array(
                File "pyarrow/array.pxi", line 343, in pyarrow.lib.array
                File "pyarrow/array.pxi", line 42, in pyarrow.lib._sequence_to_array
                File "pyarrow/error.pxi", line 154, in pyarrow.lib.pyarrow_internal_check_status
                File "pyarrow/error.pxi", line 91, in pyarrow.lib.check_status
              pyarrow.lib.ArrowCapacityError: array cannot contain more than 2147483646 bytes, have 5368709120
              
              During handling of the above exception, another exception occurred:
              
              Traceback (most recent call last):
                File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/builder.py", line 1595, in _prepare_split_single
                  num_examples, num_bytes = writer.finalize()
                File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/arrow_writer.py", line 658, in finalize
                  self.write_examples_on_file()
                File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/arrow_writer.py", line 511, in write_examples_on_file
                  self.write_batch(batch_examples=batch_examples)
                File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/arrow_writer.py", line 631, in write_batch
                  self.write_table(pa_table, writer_batch_size)
                File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/arrow_writer.py", line 646, in write_table
                  pa_table = embed_table_storage(pa_table)
                File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/table.py", line 2248, in embed_table_storage
                  arrays = [
                File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/table.py", line 2249, in <listcomp>
                  embed_array_storage(table[name], feature, token_per_repo_id=token_per_repo_id)
                File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/table.py", line 1795, in wrapper
                  return pa.chunked_array([func(chunk, *args, **kwargs) for chunk in array.chunks])
                File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/table.py", line 1795, in <listcomp>
                  return pa.chunked_array([func(chunk, *args, **kwargs) for chunk in array.chunks])
                File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/table.py", line 2124, in embed_array_storage
                  return feature.embed_storage(array, token_per_repo_id=token_per_repo_id)
                File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/features/image.py", line 281, in embed_storage
                  bytes_array = pa.array(
                File "pyarrow/array.pxi", line 343, in pyarrow.lib.array
                File "pyarrow/array.pxi", line 42, in pyarrow.lib._sequence_to_array
                File "pyarrow/error.pxi", line 154, in pyarrow.lib.pyarrow_internal_check_status
                File "pyarrow/error.pxi", line 91, in pyarrow.lib.check_status
              pyarrow.lib.ArrowCapacityError: array cannot contain more than 2147483646 bytes, have 5368709120
              
              The above exception was the direct cause of the following exception:
              
              Traceback (most recent call last):
                File "/src/services/worker/src/worker/job_runners/config/parquet_and_info.py", line 1451, in compute_config_parquet_and_info_response
                  parquet_operations, partial, estimated_dataset_info = stream_convert_to_parquet(
                File "/src/services/worker/src/worker/job_runners/config/parquet_and_info.py", line 994, in stream_convert_to_parquet
                  builder._prepare_split(
                File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/builder.py", line 1447, in _prepare_split
                  for job_id, done, content in self._prepare_split_single(
                File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/builder.py", line 1604, in _prepare_split_single
                  raise DatasetGenerationError("An error occurred while generating the dataset") from e
              datasets.exceptions.DatasetGenerationError: An error occurred while generating the dataset

Need help to make the dataset viewer work? Make sure to review how to configure the dataset viewer, and open a discussion for direct support.

id
string
task
string
selected_step
int64
conversations
list
image
list
/mnt/hpfs/baaiei/jyShi/rt_frames_success/rtx_frames_success_42/62_robo_set#episode_1570
Discriminative_Affordance_(Negative)_Task
3
[ { "from": "human", "value": "<image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image>\nIs <reach for the small brown cylindri...
[ "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_1570/frame_0.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_1570/frame_1.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_1570/frame_2.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_1570/fram...
/mnt/hpfs/baaiei/jyShi/rt_frames_success/rtx_frames_success_42/62_robo_set#episode_1570
Discriminative_Affordance_(Negative)_Task
0
[ { "from": "human", "value": "<image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image>\nIs this <place the can on a different spot on the table> you're focused on right now?" }, { "from": "gpt", "value": "no" } ]
[ "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_1570/frame_0.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_1570/frame_1.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_1570/frame_2.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_1570/fram...
/mnt/hpfs/baaiei/jyShi/rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15696
Discriminative_Affordance_(Negative)_Task
5
[ { "from": "human", "value": "<image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image>\nAre you currently engaged in <place ut...
[ "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15696/frame_0.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15696/frame_1.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15696/frame_2.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15696/...
/mnt/hpfs/baaiei/jyShi/rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15696
Discriminative_Affordance_(Negative)_Task
2
[ { "from": "human", "value": "<image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image>\nIs this <move the knife to the left of the pot> you're focused on right now?" }, { "from": "gpt", "value": "no" } ]
[ "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15696/frame_0.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15696/frame_1.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15696/frame_2.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15696/...
/mnt/hpfs/baaiei/jyShi/rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15693
Discriminative_Affordance_(Negative)_Task
1
[ { "from": "human", "value": "<image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image>\nAre you currently engaged in <move the packets over the box>" }, { "from": "gpt", "value": "no" } ]
[ "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15693/frame_0.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15693/frame_1.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15693/frame_2.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15693/...
/mnt/hpfs/baaiei/jyShi/rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15693
Discriminative_Affordance_(Negative)_Task
0
[ { "from": "human", "value": "<image> <image> <image> <image> <image> <image> <image> <image>\nIs <move the chips bag towards the sink> what you're working on at the moment?" }, { "from": "gpt", "value": "no" } ]
[ "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15693/frame_0.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15693/frame_1.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15693/frame_2.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15693/...
/mnt/hpfs/baaiei/jyShi/rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15695
Discriminative_Affordance_(Negative)_Task
0
[ { "from": "human", "value": "<image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image>\nIs this <lift the object from the toolbox> you're handling at present?" }, { "from": "gpt", "value": "no" } ]
[ "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15695/frame_0.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15695/frame_1.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15695/frame_2.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15695/...
/mnt/hpfs/baaiei/jyShi/rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15695
Discriminative_Affordance_(Negative)_Task
1
[ { "from": "human", "value": "<image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image>\nIs <reach for the oven door handle> what you're working on at the moment?" }, { "from": "gpt", "value": "no" } ]
[ "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15695/frame_0.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15695/frame_1.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15695/frame_2.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15695/...
/mnt/hpfs/baaiei/jyShi/rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15698
Discriminative_Affordance_(Negative)_Task
2
[ { "from": "human", "value": "<image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image>\nAre you doing <reach for the green spatula on the cloth> at this very moment?" }, { "from": "gpt", "value"...
[ "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15698/frame_0.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15698/frame_1.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15698/frame_2.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15698/...
/mnt/hpfs/baaiei/jyShi/rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15698
Discriminative_Affordance_(Negative)_Task
3
[ { "from": "human", "value": "<image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image>\nAre you currently engaged in <place the towel on the right sid...
[ "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15698/frame_0.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15698/frame_1.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15698/frame_2.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15698/...
/mnt/hpfs/baaiei/jyShi/rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15694
Discriminative_Affordance_(Negative)_Task
1
[ { "from": "human", "value": "<image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image>\nIs this <place the wooden spoon down> you're handling at present?" }, { "from": "gpt", "value": "no" } ]
[ "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15694/frame_0.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15694/frame_1.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15694/frame_2.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15694/...
/mnt/hpfs/baaiei/jyShi/rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15694
Discriminative_Affordance_(Negative)_Task
0
[ { "from": "human", "value": "<image> <image> <image> <image> <image> <image> <image> <image>\nIs this <place the clothes onto the box> you're focused on right now?" }, { "from": "gpt", "value": "no" } ]
[ "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15694/frame_0.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15694/frame_1.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15694/frame_2.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15694/...
/mnt/hpfs/baaiei/jyShi/rt_frames_success/rtx_frames_success_42/62_robo_set#episode_157
Discriminative_Affordance_(Negative)_Task
0
[ { "from": "human", "value": "<image> <image> <image> <image> <image> <image> <image> <image>\nAre you currently engaged in <insert the green object into the pink base>" }, { "from": "gpt", "value": "no" } ]
[ "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_157/frame_0.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_157/frame_1.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_157/frame_2.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_157/frame_3....
/mnt/hpfs/baaiei/jyShi/rt_frames_success/rtx_frames_success_42/62_robo_set#episode_157
Discriminative_Affordance_(Negative)_Task
1
[ { "from": "human", "value": "<image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image>\nIs this <grab the green chilly> you're handling at present?" }, { "from": "gpt", "value": "no" } ]
[ "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_157/frame_0.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_157/frame_1.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_157/frame_2.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_157/frame_3....
/mnt/hpfs/baaiei/jyShi/rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15699
Discriminative_Affordance_(Negative)_Task
1
[ { "from": "human", "value": "<image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image>\nAre you currently engaged in <move towards the bowl with the sponge>" }, { "from": "gpt", "value": "no" } ]
[ "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15699/frame_0.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15699/frame_1.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15699/frame_2.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15699/...
/mnt/hpfs/baaiei/jyShi/rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15699
Discriminative_Affordance_(Negative)_Task
2
[ { "from": "human", "value": "<image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image>\nIs <move towards the target location w...
[ "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15699/frame_0.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15699/frame_1.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15699/frame_2.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15699/...
/mnt/hpfs/baaiei/jyShi/rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15697
Discriminative_Affordance_(Negative)_Task
5
[ { "from": "human", "value": "<image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image>\nAre you currently engaged in <place the wooden...
[ "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15697/frame_0.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15697/frame_1.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15697/frame_2.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15697/...
/mnt/hpfs/baaiei/jyShi/rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15697
Discriminative_Affordance_(Negative)_Task
3
[ { "from": "human", "value": "<image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image>\nIs this <lower to the surface again> you're focused on right now?" }, { "from": "gpt", "value": "no" } ]
[ "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15697/frame_0.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15697/frame_1.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15697/frame_2.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15697/...
/mnt/hpfs/baaiei/jyShi/rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15700
Discriminative_Affordance_(Negative)_Task
4
[ { "from": "human", "value": "<image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image>\nIs this <move the steak towards the bowl> you're handling at p...
[ "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15700/frame_0.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15700/frame_1.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15700/frame_2.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15700/...
/mnt/hpfs/baaiei/jyShi/rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15700
Discriminative_Affordance_(Negative)_Task
0
[ { "from": "human", "value": "<image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image>\nAre you doing <move towards the metal pot> at this very moment?" }, { "from": "gpt", "value": "no" } ]
[ "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15700/frame_0.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15700/frame_1.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15700/frame_2.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15700/...
/mnt/hpfs/baaiei/jyShi/rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15703
Discriminative_Affordance_(Negative)_Task
1
[ { "from": "human", "value": "<image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image>\nAre you doing <place the lemon in the sink> at this very moment?" }, { "from": "gpt", "value": "no" } ]
[ "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15703/frame_0.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15703/frame_1.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15703/frame_2.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15703/...
/mnt/hpfs/baaiei/jyShi/rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15703
Discriminative_Affordance_(Negative)_Task
3
[ { "from": "human", "value": "<image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image>\nIs this <move the bowl to the left of the eggplant> you're focused on right now?" }, { "from": "gpt", ...
[ "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15703/frame_0.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15703/frame_1.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15703/frame_2.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15703/...
/mnt/hpfs/baaiei/jyShi/rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15709
Discriminative_Affordance_(Negative)_Task
1
[ { "from": "human", "value": "<image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image>\nAre you doing <move the chairs towards...
[ "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15709/frame_0.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15709/frame_1.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15709/frame_2.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15709/...
/mnt/hpfs/baaiei/jyShi/rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15709
Discriminative_Affordance_(Negative)_Task
0
[ { "from": "human", "value": "<image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image>\nIs this <position the knife above the daikon radish> you're handling at present?" }, { "from": "gpt", "value": "no" } ]
[ "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15709/frame_0.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15709/frame_1.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15709/frame_2.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15709/...
/mnt/hpfs/baaiei/jyShi/rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15702
Discriminative_Affordance_(Negative)_Task
0
[ { "from": "human", "value": "<image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image>\nAre you currently engaged in <place the slice down on the table>" }, { "from": "gpt", "value": "no" } ]
[ "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15702/frame_0.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15702/frame_1.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15702/frame_2.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15702/...
/mnt/hpfs/baaiei/jyShi/rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15702
Discriminative_Affordance_(Negative)_Task
1
[ { "from": "human", "value": "<image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image>\nAre you doing <move the inverted cup towards the target cup> at this very moment?" }, { "from": "gpt", "value": "no" } ]
[ "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15702/frame_0.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15702/frame_1.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15702/frame_2.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15702/...
/mnt/hpfs/baaiei/jyShi/rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15708
Discriminative_Affordance_(Negative)_Task
2
[ { "from": "human", "value": "<image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image>\nIs this <move towards the coffee cup with the elastic band> you're handling at present?" }, { "from": "gpt", "valu...
[ "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15708/frame_0.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15708/frame_1.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15708/frame_2.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15708/...
/mnt/hpfs/baaiei/jyShi/rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15708
Discriminative_Affordance_(Negative)_Task
4
[ { "from": "human", "value": "<image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image>\nIs this <throw the second piece of pap...
[ "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15708/frame_0.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15708/frame_1.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15708/frame_2.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15708/...
/mnt/hpfs/baaiei/jyShi/rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15704
Discriminative_Affordance_(Negative)_Task
0
[ { "from": "human", "value": "<image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image>\nIs <place the fork on the right side of the counter> what you're working on at the moment?" }, { "from": "gpt", "value": "no" } ...
[ "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15704/frame_0.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15704/frame_1.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15704/frame_2.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15704/...
/mnt/hpfs/baaiei/jyShi/rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15704
Discriminative_Affordance_(Negative)_Task
2
[ { "from": "human", "value": "<image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image>\nIs this <lift the pipette with the tip> you're handling at present?" ...
[ "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15704/frame_0.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15704/frame_1.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15704/frame_2.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15704/...
/mnt/hpfs/baaiei/jyShi/rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15707
Discriminative_Affordance_(Negative)_Task
3
[ { "from": "human", "value": "<image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image>\nAre you currently engaged in <lower th...
[ "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15707/frame_0.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15707/frame_1.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15707/frame_2.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15707/...
/mnt/hpfs/baaiei/jyShi/rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15707
Discriminative_Affordance_(Negative)_Task
1
[ { "from": "human", "value": "<image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image>\nIs this <lift the water bottle> you're focused on right now?" }, { "from": "gpt", "value": "no" } ]
[ "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15707/frame_0.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15707/frame_1.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15707/frame_2.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15707/...
/mnt/hpfs/baaiei/jyShi/rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15705
Discriminative_Affordance_(Negative)_Task
3
[ { "from": "human", "value": "<image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image>\nIs this <pick up a brown play-doh cont...
[ "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15705/frame_0.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15705/frame_1.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15705/frame_2.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15705/...
/mnt/hpfs/baaiei/jyShi/rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15705
Discriminative_Affordance_(Negative)_Task
0
[ { "from": "human", "value": "<image> <image> <image> <image> <image> <image> <image>\nAre you currently engaged in <move the pickle to the right of the yellow towel>" }, { "from": "gpt", "value": "no" } ]
[ "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15705/frame_0.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15705/frame_1.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15705/frame_2.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15705/...
/mnt/hpfs/baaiei/jyShi/rt_frames_success/rtx_frames_success_42/62_robo_set#episode_1571
Discriminative_Affordance_(Negative)_Task
1
[ { "from": "human", "value": "<image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image>\nIs this <move towards the garbage can> you're focused on right...
[ "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_1571/frame_0.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_1571/frame_1.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_1571/frame_2.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_1571/fram...
/mnt/hpfs/baaiei/jyShi/rt_frames_success/rtx_frames_success_42/62_robo_set#episode_1571
Discriminative_Affordance_(Negative)_Task
2
[ { "from": "human", "value": "<image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image>\nAre you currently engaged in <move the...
[ "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_1571/frame_0.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_1571/frame_1.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_1571/frame_2.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_1571/fram...
/mnt/hpfs/baaiei/jyShi/rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15701
Discriminative_Affordance_(Negative)_Task
0
[ { "from": "human", "value": "<image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image>\nIs <place the yellow block on the board> what you're working on at the moment?" }, { "from": "gpt", "value": "no" } ]
[ "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15701/frame_0.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15701/frame_1.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15701/frame_2.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15701/...
/mnt/hpfs/baaiei/jyShi/rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15701
Discriminative_Affordance_(Negative)_Task
1
[ { "from": "human", "value": "<image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image>\nIs this <grasp the pear-shaped item> you're focused on right now?" }, { "from": "gpt", "value": "no" } ]
[ "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15701/frame_0.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15701/frame_1.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15701/frame_2.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15701/...
/mnt/hpfs/baaiei/jyShi/rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15706
Discriminative_Affordance_(Negative)_Task
2
[ { "from": "human", "value": "<image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image>\nAre you doing <position above the sponge> at this very moment?" }, { "from": "gpt", "value": "no" } ]
[ "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15706/frame_0.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15706/frame_1.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15706/frame_2.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15706/...
/mnt/hpfs/baaiei/jyShi/rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15706
Discriminative_Affordance_(Negative)_Task
4
[ { "from": "human", "value": "<image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image>\nIs this <push the right switch downwar...
[ "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15706/frame_0.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15706/frame_1.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15706/frame_2.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15706/...
/mnt/hpfs/baaiei/jyShi/rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15712
Discriminative_Affordance_(Negative)_Task
1
[ { "from": "human", "value": "<image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image>\nIs <place the toothbrush inside the dark blue toiletry bag> what you're working on at the moment?" }...
[ "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15712/frame_0.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15712/frame_1.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15712/frame_2.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15712/...
/mnt/hpfs/baaiei/jyShi/rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15712
Discriminative_Affordance_(Negative)_Task
2
[ { "from": "human", "value": "<image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image>\nIs this <pick up a wrench> you're focused on right now?" }, { ...
[ "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15712/frame_0.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15712/frame_1.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15712/frame_2.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15712/...
/mnt/hpfs/baaiei/jyShi/rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15715
Discriminative_Affordance_(Negative)_Task
2
[ { "from": "human", "value": "<image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image>\nIs this <move the green block towards ...
[ "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15715/frame_0.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15715/frame_1.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15715/frame_2.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15715/...
/mnt/hpfs/baaiei/jyShi/rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15715
Discriminative_Affordance_(Negative)_Task
1
[ { "from": "human", "value": "<image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image>\nIs this <put the plate with the tape back on the table> you're handling at present?" }, { "from": "gpt", "value": "no" } ]
[ "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15715/frame_0.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15715/frame_1.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15715/frame_2.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15715/...
/mnt/hpfs/baaiei/jyShi/rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15713
Discriminative_Affordance_(Negative)_Task
2
[ { "from": "human", "value": "<image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image>\nIs <place the object in the upper segment of the box> what you're working on at the moment?" }, { "from": "gpt", "value": ...
[ "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15713/frame_0.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15713/frame_1.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15713/frame_2.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15713/...
/mnt/hpfs/baaiei/jyShi/rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15713
Discriminative_Affordance_(Negative)_Task
4
[ { "from": "human", "value": "<image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image>\nIs this <move the green vegetable towa...
[ "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15713/frame_0.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15713/frame_1.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15713/frame_2.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15713/...
/mnt/hpfs/baaiei/jyShi/rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15716
Discriminative_Affordance_(Negative)_Task
0
[ { "from": "human", "value": "<image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image>\nAre you currently engaged in <place the red cup into the black bowl>" }, { "from": "gpt", "value": "no" } ]
[ "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15716/frame_0.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15716/frame_1.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15716/frame_2.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15716/...
/mnt/hpfs/baaiei/jyShi/rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15716
Discriminative_Affordance_(Negative)_Task
1
[ { "from": "human", "value": "<image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image>\nIs this <position the sock over the ch...
[ "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15716/frame_0.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15716/frame_1.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15716/frame_2.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15716/...
/mnt/hpfs/baaiei/jyShi/rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15717
Discriminative_Affordance_(Negative)_Task
0
[ { "from": "human", "value": "<image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image>\nAre you doing <move the jacket towards the center of the bed> at this very moment?" }, { "from": "gpt", "value": "no" } ]
[ "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15717/frame_0.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15717/frame_1.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15717/frame_2.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15717/...
/mnt/hpfs/baaiei/jyShi/rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15717
Discriminative_Affordance_(Negative)_Task
1
[ { "from": "human", "value": "<image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image>\nAre you currently engaged in <move the...
[ "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15717/frame_0.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15717/frame_1.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15717/frame_2.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15717/...
/mnt/hpfs/baaiei/jyShi/rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15711
Discriminative_Affordance_(Negative)_Task
4
[ { "from": "human", "value": "<image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image>\nIs this <approach the small gray can> ...
[ "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15711/frame_0.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15711/frame_1.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15711/frame_2.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15711/...
/mnt/hpfs/baaiei/jyShi/rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15711
Discriminative_Affordance_(Negative)_Task
0
[ { "from": "human", "value": "<image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image>\nAre you doing <move the pawn above the chessboard> at this very moment?" }, { "from": "gpt", "value": "no" } ]
[ "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15711/frame_0.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15711/frame_1.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15711/frame_2.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15711/...
/mnt/hpfs/baaiei/jyShi/rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15714
Discriminative_Affordance_(Negative)_Task
0
[ { "from": "human", "value": "<image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image>\nAre you currently engaged in <move the blue object away from the dowel>" }, { "from": "gpt", "value": "no" } ]
[ "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15714/frame_0.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15714/frame_1.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15714/frame_2.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15714/...
/mnt/hpfs/baaiei/jyShi/rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15714
Discriminative_Affordance_(Negative)_Task
1
[ { "from": "human", "value": "<image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image>\nIs this <move towards the object on th...
[ "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15714/frame_0.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15714/frame_1.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15714/frame_2.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15714/...
/mnt/hpfs/baaiei/jyShi/rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15719
Discriminative_Affordance_(Negative)_Task
3
[ { "from": "human", "value": "<image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image>\nAre you doing <move towards the coca-cola can> at this very moment?" ...
[ "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15719/frame_0.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15719/frame_1.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15719/frame_2.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15719/...
/mnt/hpfs/baaiei/jyShi/rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15719
Discriminative_Affordance_(Negative)_Task
1
[ { "from": "human", "value": "<image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image>\nIs this <grab the clamp> you're focused on right now?" }, { "from": "gpt", "value": "no" } ]
[ "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15719/frame_0.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15719/frame_1.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15719/frame_2.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15719/...
/mnt/hpfs/baaiei/jyShi/rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15710
Discriminative_Affordance_(Negative)_Task
3
[ { "from": "human", "value": "<image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image>\nIs <move towards the checkered box> what you're working on at the moment?" }...
[ "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15710/frame_0.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15710/frame_1.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15710/frame_2.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15710/...
/mnt/hpfs/baaiei/jyShi/rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15710
Discriminative_Affordance_(Negative)_Task
2
[ { "from": "human", "value": "<image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image>\nIs this <lay down the folded towel> you're focused on right now?" }, { "from": "gpt", "value": "no...
[ "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15710/frame_0.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15710/frame_1.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15710/frame_2.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15710/...
/mnt/hpfs/baaiei/jyShi/rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15718
Discriminative_Affordance_(Negative)_Task
0
[ { "from": "human", "value": "<image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image>\nAre you currently engaged in <move towards the specified point on the go board>" }, { "from": "gpt", "value": "no" } ]
[ "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15718/frame_0.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15718/frame_1.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15718/frame_2.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15718/...
/mnt/hpfs/baaiei/jyShi/rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15718
Discriminative_Affordance_(Negative)_Task
1
[ { "from": "human", "value": "<image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image>\nIs this <tip over the metal pot> you're handling at present?" }, { "from": "gpt", "value": "no" } ]
[ "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15718/frame_0.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15718/frame_1.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15718/frame_2.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15718/...
/mnt/hpfs/baaiei/jyShi/rt_frames_success/rtx_frames_success_42/62_robo_set#episode_1572
Discriminative_Affordance_(Negative)_Task
1
[ { "from": "human", "value": "<image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image>\nIs this <place the pot into the rack> you're handling at present?" }, { "from": "gpt", "value": "no" } ]
[ "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_1572/frame_0.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_1572/frame_1.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_1572/frame_2.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_1572/fram...
/mnt/hpfs/baaiei/jyShi/rt_frames_success/rtx_frames_success_42/62_robo_set#episode_1572
Discriminative_Affordance_(Negative)_Task
2
[ { "from": "human", "value": "<image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image>\nIs <move the pot to the left of the brush> what you're working on at the moment?" }, { "from": "gpt", "val...
[ "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_1572/frame_0.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_1572/frame_1.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_1572/frame_2.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_1572/fram...
/mnt/hpfs/baaiei/jyShi/rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15726
Discriminative_Affordance_(Negative)_Task
0
[ { "from": "human", "value": "<image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image>\nIs <place the orange object on the orange cloth> what you're working on at the moment?" }, { "from": "gpt", "value": "no" } ]
[ "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15726/frame_0.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15726/frame_1.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15726/frame_2.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15726/...
/mnt/hpfs/baaiei/jyShi/rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15726
Discriminative_Affordance_(Negative)_Task
2
[ { "from": "human", "value": "<image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image>\nIs this <place the inverted white bottle on top of the clear bottle> you're handling at present?" }, { "from":...
[ "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15726/frame_0.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15726/frame_1.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15726/frame_2.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15726/...
/mnt/hpfs/baaiei/jyShi/rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15728
Discriminative_Affordance_(Negative)_Task
3
[ { "from": "human", "value": "<image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image>\nIs <lift and reposition above the table> what you're working on at the...
[ "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15728/frame_0.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15728/frame_1.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15728/frame_2.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15728/...
/mnt/hpfs/baaiei/jyShi/rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15728
Discriminative_Affordance_(Negative)_Task
0
[ { "from": "human", "value": "<image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image>\nIs this <place the plate near the towel> you're focused on right now?" }, { "from": "gpt", "value": "no" } ]
[ "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15728/frame_0.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15728/frame_1.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15728/frame_2.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15728/...
/mnt/hpfs/baaiei/jyShi/rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15724
Discriminative_Affordance_(Negative)_Task
2
[ { "from": "human", "value": "<image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image>\nAre you doing <drop the potato into the orange bucket> at this very moment?" }, { "from": "gpt", "value": ...
[ "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15724/frame_0.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15724/frame_1.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15724/frame_2.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15724/...
/mnt/hpfs/baaiei/jyShi/rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15724
Discriminative_Affordance_(Negative)_Task
4
[ { "from": "human", "value": "<image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image>\nIs <approach the orange basket> what y...
[ "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15724/frame_0.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15724/frame_1.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15724/frame_2.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15724/...
/mnt/hpfs/baaiei/jyShi/rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15727
Discriminative_Affordance_(Negative)_Task
1
[ { "from": "human", "value": "<image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image>\nAre you doing <wipe the book cover> at this very moment?" }, { "from": "gpt", "value": "no" } ]
[ "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15727/frame_0.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15727/frame_1.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15727/frame_2.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15727/...
/mnt/hpfs/baaiei/jyShi/rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15727
Discriminative_Affordance_(Negative)_Task
3
[ { "from": "human", "value": "<image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image>\nIs <retract from the window> what you're working on at the moment?" ...
[ "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15727/frame_0.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15727/frame_1.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15727/frame_2.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15727/...
/mnt/hpfs/baaiei/jyShi/rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15725
Discriminative_Affordance_(Negative)_Task
0
[ { "from": "human", "value": "<image> <image> <image> <image> <image> <image> <image> <image>\nIs this <move the banana over the table> you're handling at present?" }, { "from": "gpt", "value": "no" } ]
[ "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15725/frame_0.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15725/frame_1.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15725/frame_2.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15725/...
/mnt/hpfs/baaiei/jyShi/rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15725
Discriminative_Affordance_(Negative)_Task
3
[ { "from": "human", "value": "<image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image>\nAre you doing <reach for the napkin in...
[ "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15725/frame_0.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15725/frame_1.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15725/frame_2.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15725/...
/mnt/hpfs/baaiei/jyShi/rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15723
Discriminative_Affordance_(Negative)_Task
1
[ { "from": "human", "value": "<image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image>\nAre you doing <place the fork on the cutlery tray> at this very moment?" }, { "from": "gpt", "value": "no" } ]
[ "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15723/frame_0.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15723/frame_1.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15723/frame_2.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15723/...
/mnt/hpfs/baaiei/jyShi/rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15723
Discriminative_Affordance_(Negative)_Task
0
[ { "from": "human", "value": "<image> <image> <image> <image> <image> <image> <image> <image> <image> <image>\nIs this <place the ring on top of the ball> you're focused on right now?" }, { "from": "gpt", "value": "no" } ]
[ "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15723/frame_0.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15723/frame_1.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15723/frame_2.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15723/...
/mnt/hpfs/baaiei/jyShi/rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15722
Discriminative_Affordance_(Negative)_Task
0
[ { "from": "human", "value": "<image> <image> <image> <image> <image> <image> <image> <image>\nIs this <lift the waffle maker lid> you're handling at present?" }, { "from": "gpt", "value": "no" } ]
[ "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15722/frame_0.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15722/frame_1.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15722/frame_2.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15722/...
/mnt/hpfs/baaiei/jyShi/rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15722
Discriminative_Affordance_(Negative)_Task
2
[ { "from": "human", "value": "<image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image>\nIs <move the water bottle towards the ...
[ "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15722/frame_0.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15722/frame_1.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15722/frame_2.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15722/...
/mnt/hpfs/baaiei/jyShi/rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15721
Discriminative_Affordance_(Negative)_Task
1
[ { "from": "human", "value": "<image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image>\nIs this <lift the lid from the sink> you're focused on right now?" }...
[ "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15721/frame_0.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15721/frame_1.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15721/frame_2.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15721/...
/mnt/hpfs/baaiei/jyShi/rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15721
Discriminative_Affordance_(Negative)_Task
0
[ { "from": "human", "value": "<image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image>\nAre you currently engaged in <pick up the golden component>" }, { "from": "gpt", "value": "no" } ]
[ "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15721/frame_0.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15721/frame_1.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15721/frame_2.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15721/...
/mnt/hpfs/baaiei/jyShi/rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15720
Discriminative_Affordance_(Negative)_Task
0
[ { "from": "human", "value": "<image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image>\nIs this <lift and move the eyeglass case to the right> you're handling at present?" }, { "from": "gpt", "value": "no" } ]
[ "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15720/frame_0.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15720/frame_1.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15720/frame_2.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15720/...
/mnt/hpfs/baaiei/jyShi/rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15720
Discriminative_Affordance_(Negative)_Task
1
[ { "from": "human", "value": "<image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image>\nIs this <position the square-circle object over the corresponding slot> you're focused on right now?" }, { "from": "gp...
[ "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15720/frame_0.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15720/frame_1.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15720/frame_2.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15720/...
/mnt/hpfs/baaiei/jyShi/rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15736
Discriminative_Affordance_(Negative)_Task
2
[ { "from": "human", "value": "<image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image>\nAre you doing <reach towards the dishwasher> at this very moment?" }, { "from": "gpt", "value": "no" } ]
[ "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15736/frame_0.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15736/frame_1.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15736/frame_2.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15736/...
/mnt/hpfs/baaiei/jyShi/rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15736
Discriminative_Affordance_(Negative)_Task
0
[ { "from": "human", "value": "<image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image>\nIs this <release the bottle into the box> you're handling at present?" }, { "from": "gpt", "value": "no" } ]
[ "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15736/frame_0.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15736/frame_1.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15736/frame_2.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15736/...
/mnt/hpfs/baaiei/jyShi/rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15733
Discriminative_Affordance_(Negative)_Task
2
[ { "from": "human", "value": "<image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image>\nIs this <reach for the yellow book> you're handling at present?" }, { "from": "gpt", "value": "no" } ]
[ "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15733/frame_0.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15733/frame_1.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15733/frame_2.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15733/...
/mnt/hpfs/baaiei/jyShi/rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15733
Discriminative_Affordance_(Negative)_Task
0
[ { "from": "human", "value": "<image> <image> <image> <image> <image> <image> <image>\nAre you doing <move the pan between the knife and the towel> at this very moment?" }, { "from": "gpt", "value": "no" } ]
[ "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15733/frame_0.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15733/frame_1.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15733/frame_2.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15733/...
/mnt/hpfs/baaiei/jyShi/rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15732
Discriminative_Affordance_(Negative)_Task
0
[ { "from": "human", "value": "<image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image>\nAre you currently engaged in <lift the hollow cube>" }, { "from": "gpt", "value": "no" } ]
[ "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15732/frame_0.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15732/frame_1.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15732/frame_2.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15732/...
/mnt/hpfs/baaiei/jyShi/rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15732
Discriminative_Affordance_(Negative)_Task
3
[ { "from": "human", "value": "<image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image>\nIs this <release the cup holder> you'r...
[ "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15732/frame_0.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15732/frame_1.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15732/frame_2.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15732/...
/mnt/hpfs/baaiei/jyShi/rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15735
Discriminative_Affordance_(Negative)_Task
3
[ { "from": "human", "value": "<image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image>\nAre you doing <move the orange object towards the cabinet> at this ver...
[ "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15735/frame_0.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15735/frame_1.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15735/frame_2.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15735/...
/mnt/hpfs/baaiei/jyShi/rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15735
Discriminative_Affordance_(Negative)_Task
1
[ { "from": "human", "value": "<image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image>\nAre you currently engaged in <move the spoon towards the rag>" }, { "from": "gpt", "value": "no" } ]
[ "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15735/frame_0.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15735/frame_1.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15735/frame_2.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15735/...
/mnt/hpfs/baaiei/jyShi/rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15731
Discriminative_Affordance_(Negative)_Task
3
[ { "from": "human", "value": "<image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image>\nIs <reach for the gray cup inside the ...
[ "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15731/frame_0.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15731/frame_1.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15731/frame_2.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15731/...
/mnt/hpfs/baaiei/jyShi/rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15731
Discriminative_Affordance_(Negative)_Task
0
[ { "from": "human", "value": "<image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image>\nIs this <insert the coffee pod into the machine> you're handling at present?" }, { "from": "gpt", "value": "no" } ]
[ "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15731/frame_0.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15731/frame_1.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15731/frame_2.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15731/...
/mnt/hpfs/baaiei/jyShi/rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15737
Discriminative_Affordance_(Negative)_Task
0
[ { "from": "human", "value": "<image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image>\nIs this <move the triangle towards the board> you're handling at present?" }, { "from": "gpt", "value": "no" } ]
[ "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15737/frame_0.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15737/frame_1.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15737/frame_2.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15737/...
/mnt/hpfs/baaiei/jyShi/rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15737
Discriminative_Affordance_(Negative)_Task
1
[ { "from": "human", "value": "<image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image>\nAre you currently engaged in <reach fo...
[ "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15737/frame_0.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15737/frame_1.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15737/frame_2.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15737/...
/mnt/hpfs/baaiei/jyShi/rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15729
Discriminative_Affordance_(Negative)_Task
1
[ { "from": "human", "value": "<image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image> <image>\nIs <move an item backwards> what you're working on at the moment?" }, { "from": "gpt",...
[ "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15729/frame_0.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15729/frame_1.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15729/frame_2.png", "rt_frames_success/rtx_frames_success_42/62_robo_set#episode_15729/...
End of preview.

ShareRobot Dataset

ShareRobot, a high-quality heterogeneous dataset that labels multi-dimensional information, including task planning, object affordance, and end-effector trajectory, effectively enhancing various robotic capabilities.

Overview of ShareRobot Dataset

ee709e8b-6f05-428d-abff-2578914aeb0d For planning, we have 51,403 episodes and each with 30 frames. In the process of data generation, we design 5 different templates for each of the 10 question types in RoboVQA [1]. In the process of data generation, we randomly select 2 templates of each question type to generate question-answer pairs for every instance. This process transforms 51,403 instances into 1,027,990 question-answer pairs, with annotators monitoring data generation to maintain the dataset’s integrity.

For Affordance, we have 6,522 images and each with affordance areas aligned with an instruction.

For Trajectory, we have 6,870 images and each with at least 3 {x, y} coordinates aligned with an instruction.

Dataset Sources

a608d080-665a-4ab1-bd8f-d5bd121454da

ShareRobot dataset contains 23 original datasets from Open X-Embodiment dataset [2], 12 embodiments and 107 types of atomic tasks.

Raw Dataset for Planning

Raw Dataset Number of Raws
nyu_door_opening_surprising_effectiveness 421
bridge 15738
dlr_edan_shared_control_converted_externally_to_rlds 63
utokyo_xarm_pick_and_place_converted_externally_to_rlds 92
cmu_stretch 10
asu_table_top_converted_externally_to_rlds 109
dlr_sara_pour_converted_externally_to_rlds 51
utokyo_xarm_bimanual_converted_externally_to_rlds 27
robo_set 18164
dobbe 5200
berkeley_autolab_ur5 882
qut_dexterous_manpulation 192
aloha_mobile 264
dlr_sara_grid_clamp_converted_externally_to_rlds 40
ucsd_pick_and_place_dataset_converted_externally_to_rlds 569
ucsd_kitchen_dataset_converted_externally_to_rlds 39
jaco_play 956
utokyo_pr2_opening_fridge_converted_externally_to_rlds 64
conq_hose_manipulation 56
fmb 7836
plex_robosuite 398
utokyo_pr2_tabletop_manipulation_converted_externally_to_rlds 189
viola 44

Raw Dataset for Affordance

Raw Dataset Number of Raws
utokyo_pr2_tabletop_manipulation_converted_externally_to_rlds 24
utokyo_xarm_pick_and_place_converted_externally_to_rlds 23
ucsd_kitchen_dataset_converted_externally_to_rlds 10
ucsd_pick_and_place_dataset_converted_externally_to_rlds 112
nyu_door_opening_surprising_effectiveness 85
jaco_play 171
bridge 2610
utokyo_pr2_opening_fridge_converted_externally_to_rlds 12
asu_table_top_converted_externally_to_rlds 24
viola 1
berkeley_autolab_ur5 122
aloha_mobile 23
conq_hose_manipulation 1
dobbe 717
fmb 561
plex_robosuite 13
qut_dexterous_manpulation 16
robo_set 1979
dlr_edan_shared_control_converted_externally_to_rlds 18
Summary 6522

Raw Dataset for Trajectory

Raw Dataset Number of Raws
utokyo_pr2_tabletop_manipulation_converted_externally_to_rlds 35
utokyo_xarm_pick_and_place_converted_externally_to_rlds 36
ucsd_kitchen_dataset_converted_externally_to_rlds 19
dlr_sara_grid_clamp_converted_externally_to_rlds 1
ucsd_pick_and_place_dataset_converted_externally_to_rlds 109
nyu_door_opening_surprising_effectiveness 74
jaco_play 175
utokyo_xarm_bimanual_converted_externally_to_rlds 7
bridge 2986
utokyo_pr2_opening_fridge_converted_externally_to_rlds 12
asu_table_top_converted_externally_to_rlds 22
berkeley_autolab_ur5 164
dobbe 759
fmb 48
qut_dexterous_manpulation 29
robo_set 2374
dlr_sara_pour_converted_externally_to_rlds 3
dlr_edan_shared_control_converted_externally_to_rlds 17
Summary 6870

Data Format

Planning

data-demo

{
 "id"{
        "id": 0,
        "task": "Future_Prediction_Task",
        "selected_step": 3,
        "conversations": [
            {
                "from": "human",
                "value": "<image 0-25> After <move the grasped banana towards the mug>, what's the most probable next event?"
            },
            {
                "from": "gpt",
                "value": "<place the banana into the mug>"
            }
        ],
        "image": [
            "/path/to/image_0-25"
        ]
    }        
}

     

Affordance

{

        "id": 2486,
        "meta_data": {
            "original_dataset": "bridge",
            "original_width": 640,
            "original_height": 480
        },
        "instruction": "place the red fork to the left of the left burner",
        "affordance": {
            "x": 352.87425387858815,
            "y": 186.47871614766484,
            "width": 19.296008229513156,
            "height": 14.472006172134865
    }

Visualize Code

import json
import os
import cv2
import numpy as np

img_dir = '/path/to/your/original/images/dir'
affordance_json = '/path/to/your/affordances/json'
output_img_dir = '/path/to/your/visualized/images/dir'

with open(affordance_json, 'r') as f:
    data = json.load(f)
    for item in data:
        filepath = os.path.join(img_dir, item['id'])

        image = cv2.imread(filepath)
        color = (255, 0, 0)
        thickness = 2

        x_min,y_min = item['affordance']['x'], item['affordance']['y']
        x_max,y_max = item['affordance']['x']+item['affordance']['width'], item['affordance']['y']+item['affordance']['height']

        # 定义矩形的四个顶点坐标
        pts = np.array([
            [x_min, y_min],  # 左上角
            [x_max, y_min],  # 右上角
            [x_max, y_max],  # 右下角
            [x_min, y_max]   # 左下角
        ], dtype=np.float32)

        # 绘制矩形框
        cv2.polylines(image, [pts.astype(int)], isClosed=True, color=color, thickness=thickness)

        # 获取相对路径并拼接目标路径
        relative_path = os.path.relpath(filepath, img_dir)  # 获取相对于 img_dir 的相对路径
        output_img_path = os.path.join(output_img_dir, relative_path)  # 拼接目标路径

        # 创建目标文件夹
        output_directory = os.path.dirname(output_img_path)
        if not os.path.exists(output_directory):
            os.makedirs(output_directory)

        # 打印调试信息
        print(f"Input filepath: {filepath}")
        print(f"Output image path: {output_img_path}")
        print(f"Output directory: {output_directory}")

        # 保存图像
        cv2.imwrite(output_img_path, image)

Trajectory

{
        "id": 456,
        "meta_data": {
            "original_dataset": "bridge",
            "original_width": 640,
            "original_height": 480
        },
        "instruction": "reach for the carrot",
        "points": [
            [
                265.45454545454544,
                120.0
            ],
            [
                275.1515151515152,
                162.42424242424244
            ],
            [
                280.0,
                213.33333333333331
            ],
            [
                280.0,
                259.3939393939394
            ]
        ]
    },

Visualize Code

import json
import os
from PIL import Image, ImageDraw

trajectory_final = '/path/to/your/trajectory_json'
img_dir = '/path/to/your/original/images/dir'
output_img_dir = '/path/to/your/visualzed/images/dir'

with open(trajectory_final, 'r') as f:
    data = json.load(f)
    for item in data:
        filepath = os.path.join(img_dir, item['id'])
        points = item['points']

        image = Image.open(filepath).convert("RGB")  # 确保图像是 RGB 模式
        draw = ImageDraw.Draw(image)  # 创建绘图对象
        # 定义颜色和线宽
        color = (255, 0, 0)  # 红色 (RGB 格式)
        thickness = 2


        scaled_points = [
                (point[0], point[1])
                for point in points
            ]
        # 按照顺序连接相邻的点
        for i in range(len(scaled_points) - 1):
            draw.line([scaled_points[i], scaled_points[i + 1]], fill=color, width=thickness)

        # 获取相对路径并拼接目标路径
        relative_path = os.path.relpath(filepath, img_dir)
        output_img_path = os.path.join(output_img_dir, relative_path)

        # 创建目标文件夹
        output_directory = os.path.dirname(output_img_path)
        if not os.path.exists(output_directory):
            os.makedirs(output_directory)

        # 打印调试信息
        print(f"Input filepath: {filepath}")
        print(f"Output image path: {output_img_path}")
        print(f"Output directory: {output_directory}")

        # 保存图像
        image.save(output_img_path)

Evaluation

Reference

[1] Pierre Sermanet, Tianli Ding, Jeffrey Zhao, Fei Xia, Debidatta Dwibedi, Keerthana Gopalakrishnan, Christine Chan,Gabriel Dulac-Arnold, Sharath Maddineni, Nikhil J Joshi,et al. Robovqa: Multimodal long-horizon reasoning forrobotics. In ICRA, pages 645–652, 2024.

[2] Abby O’Neill, Abdul Rehman, Abhinav Gupta, AbhiramMaddukuri, Abhishek Gupta, Abhishek Padalkar, AbrahamLee, Acorn Pooley, Agrim Gupta, Ajay Mandlekar, et al.Open x-embodiment: Robotic learning datasets and rt-xmodels. arXiv preprint arXiv:2310.08864, 2023.

Citation

@article{ji2025robobrain,
  title={RoboBrain: A Unified Brain Model for Robotic Manipulation from Abstract to Concrete},
  author={Ji, Yuheng and Tan, Huajie and Shi, Jiayu and Hao, Xiaoshuai and Zhang, Yuan and Zhang, Hengyuan and Wang, Pengwei and Zhao, Mengdi and Mu, Yao and An, Pengju and others},
  journal={arXiv preprint arXiv:2502.21257},
  year={2025}
}
Downloads last month
3,552

Models trained or fine-tuned on BAAI/ShareRobot

Collection including BAAI/ShareRobot

Papers for BAAI/ShareRobot