FreeAskWorld: An Interactive and Closed-Loop Simulator for Human-Centric Embodied AI Article Swipe
YOU?
·
· 2025
· Open Access
·
· DOI: https://doi.org/10.48550/arxiv.2511.13524
As embodied intelligence emerges as a core frontier in artificial intelligence research, simulation platforms must evolve beyond low-level physical interactions to capture complex, human-centered social behaviors. We introduce FreeAskWorld, an interactive simulation framework that integrates large language models (LLMs) for high-level behavior planning and semantically grounded interaction, informed by theories of intention and social cognition. Our framework supports scalable, realistic human-agent simulations and includes a modular data generation pipeline tailored for diverse embodied tasks.To validate the framework, we extend the classic Vision-and-Language Navigation (VLN) task into a interaction enriched Direction Inquiry setting, wherein agents can actively seek and interpret navigational guidance. We present and publicly release FreeAskWorld, a large-scale benchmark dataset comprising reconstructed environments, six diverse task types, 16 core object categories, 63,429 annotated sample frames, and more than 17 hours of interaction data to support training and evaluation of embodied AI systems. We benchmark VLN models, and human participants under both open-loop and closed-loop settings. Experimental results demonstrate that models fine-tuned on FreeAskWorld outperform their original counterparts, achieving enhanced semantic understanding and interaction competency. These findings underscore the efficacy of socially grounded simulation frameworks in advancing embodied AI systems toward sophisticated high-level planning and more naturalistic human-agent interaction. Importantly, our work underscores that interaction itself serves as an additional information modality.
Related Topics
- Type
- preprint
- Landing Page
- http://arxiv.org/abs/2511.13524
- https://arxiv.org/pdf/2511.13524
- OA Status
- green
- OpenAlex ID
- https://openalex.org/W4416656756
Raw OpenAlex JSON
- OpenAlex ID
-
https://openalex.org/W4416656756Canonical identifier for this work in OpenAlex
- DOI
-
https://doi.org/10.48550/arxiv.2511.13524Digital Object Identifier
- Title
-
FreeAskWorld: An Interactive and Closed-Loop Simulator for Human-Centric Embodied AIWork title
- Type
-
preprintOpenAlex work type
- Publication year
-
2025Year of publication
- Publication date
-
2025-11-17Full publication date if available
- Authors
-
Yafeng Pan, Xinning He, Jun Yang, Xinyu Yin, Han Wang, Xiaoji Zheng, Chao Gao, Jiangtao GongList of authors in order
- Landing page
-
https://arxiv.org/abs/2511.13524Publisher landing page
- PDF URL
-
https://arxiv.org/pdf/2511.13524Direct link to full text PDF
- Open access
-
YesWhether a free full text is available
- OA status
-
greenOpen access status per OpenAlex
- OA URL
-
https://arxiv.org/pdf/2511.13524Direct OA link when available
- Cited by
-
0Total citation count in OpenAlex
Full payload
| id | https://openalex.org/W4416656756 |
|---|---|
| doi | https://doi.org/10.48550/arxiv.2511.13524 |
| ids.doi | https://doi.org/10.48550/arxiv.2511.13524 |
| ids.openalex | https://openalex.org/W4416656756 |
| fwci | |
| type | preprint |
| title | FreeAskWorld: An Interactive and Closed-Loop Simulator for Human-Centric Embodied AI |
| biblio.issue | |
| biblio.volume | |
| biblio.last_page | |
| biblio.first_page | |
| is_xpac | False |
| apc_list | |
| apc_paid | |
| language | |
| locations[0].id | pmh:oai:arXiv.org:2511.13524 |
| locations[0].is_oa | True |
| locations[0].source.id | https://openalex.org/S4306400194 |
| locations[0].source.issn | |
| locations[0].source.type | repository |
| locations[0].source.is_oa | True |
| locations[0].source.issn_l | |
| locations[0].source.is_core | False |
| locations[0].source.is_in_doaj | False |
| locations[0].source.display_name | arXiv (Cornell University) |
| locations[0].source.host_organization | https://openalex.org/I205783295 |
| locations[0].source.host_organization_name | Cornell University |
| locations[0].source.host_organization_lineage | https://openalex.org/I205783295 |
| locations[0].license | |
| locations[0].pdf_url | https://arxiv.org/pdf/2511.13524 |
| locations[0].version | submittedVersion |
| locations[0].raw_type | text |
| locations[0].license_id | |
| locations[0].is_accepted | False |
| locations[0].is_published | False |
| locations[0].raw_source_name | |
| locations[0].landing_page_url | http://arxiv.org/abs/2511.13524 |
| locations[1].id | doi:10.48550/arxiv.2511.13524 |
| locations[1].is_oa | True |
| locations[1].source.id | https://openalex.org/S4306400194 |
| locations[1].source.issn | |
| locations[1].source.type | repository |
| locations[1].source.is_oa | True |
| locations[1].source.issn_l | |
| locations[1].source.is_core | False |
| locations[1].source.is_in_doaj | False |
| locations[1].source.display_name | arXiv (Cornell University) |
| locations[1].source.host_organization | https://openalex.org/I205783295 |
| locations[1].source.host_organization_name | Cornell University |
| locations[1].source.host_organization_lineage | https://openalex.org/I205783295 |
| locations[1].license | |
| locations[1].pdf_url | |
| locations[1].version | |
| locations[1].raw_type | article |
| locations[1].license_id | |
| locations[1].is_accepted | False |
| locations[1].is_published | |
| locations[1].raw_source_name | |
| locations[1].landing_page_url | https://doi.org/10.48550/arxiv.2511.13524 |
| indexed_in | arxiv, datacite |
| authorships[0].author.id | https://openalex.org/A5112300402 |
| authorships[0].author.orcid | |
| authorships[0].author.display_name | Yafeng Pan |
| authorships[0].author_position | middle |
| authorships[0].raw_author_name | Pan, Yizhou |
| authorships[0].is_corresponding | False |
| authorships[1].author.id | https://openalex.org/A5057468874 |
| authorships[1].author.orcid | |
| authorships[1].author.display_name | Xinning He |
| authorships[1].author_position | middle |
| authorships[1].raw_author_name | He, Xinning |
| authorships[1].is_corresponding | False |
| authorships[2].author.id | https://openalex.org/A5100606329 |
| authorships[2].author.orcid | https://orcid.org/0000-0003-4060-3482 |
| authorships[2].author.display_name | Jun Yang |
| authorships[2].author_position | middle |
| authorships[2].raw_author_name | Yang, Jihaoyu |
| authorships[2].is_corresponding | False |
| authorships[3].author.id | https://openalex.org/A5111113606 |
| authorships[3].author.orcid | |
| authorships[3].author.display_name | Xinyu Yin |
| authorships[3].author_position | middle |
| authorships[3].raw_author_name | Yin, Xinyu |
| authorships[3].is_corresponding | False |
| authorships[4].author.id | https://openalex.org/A5100452675 |
| authorships[4].author.orcid | https://orcid.org/0000-0002-1320-0947 |
| authorships[4].author.display_name | Han Wang |
| authorships[4].author_position | middle |
| authorships[4].raw_author_name | Wang, Han |
| authorships[4].is_corresponding | False |
| authorships[5].author.id | https://openalex.org/A5100652158 |
| authorships[5].author.orcid | https://orcid.org/0009-0007-0868-8216 |
| authorships[5].author.display_name | Xiaoji Zheng |
| authorships[5].author_position | middle |
| authorships[5].raw_author_name | Zheng, Xiaoji |
| authorships[5].is_corresponding | False |
| authorships[6].author.id | https://openalex.org/A5112258304 |
| authorships[6].author.orcid | https://orcid.org/0000-0003-3024-9117 |
| authorships[6].author.display_name | Chao Gao |
| authorships[6].author_position | last |
| authorships[6].raw_author_name | Gao, Chao |
| authorships[6].is_corresponding | False |
| authorships[7].author.id | https://openalex.org/A5089973290 |
| authorships[7].author.orcid | https://orcid.org/0000-0002-4310-1894 |
| authorships[7].author.display_name | Jiangtao Gong |
| authorships[7].author_position | middle |
| authorships[7].raw_author_name | Gong, Jiangtao |
| authorships[7].is_corresponding | False |
| has_content.pdf | False |
| has_content.grobid_xml | False |
| is_paratext | False |
| open_access.is_oa | True |
| open_access.oa_url | https://arxiv.org/pdf/2511.13524 |
| open_access.oa_status | green |
| open_access.any_repository_has_fulltext | False |
| created_date | 2025-11-19T00:00:00 |
| display_name | FreeAskWorld: An Interactive and Closed-Loop Simulator for Human-Centric Embodied AI |
| has_fulltext | False |
| is_retracted | False |
| updated_date | 2025-11-28T19:51:09.224214 |
| primary_topic | |
| cited_by_count | 0 |
| locations_count | 2 |
| best_oa_location.id | pmh:oai:arXiv.org:2511.13524 |
| best_oa_location.is_oa | True |
| best_oa_location.source.id | https://openalex.org/S4306400194 |
| best_oa_location.source.issn | |
| best_oa_location.source.type | repository |
| best_oa_location.source.is_oa | True |
| best_oa_location.source.issn_l | |
| best_oa_location.source.is_core | False |
| best_oa_location.source.is_in_doaj | False |
| best_oa_location.source.display_name | arXiv (Cornell University) |
| best_oa_location.source.host_organization | https://openalex.org/I205783295 |
| best_oa_location.source.host_organization_name | Cornell University |
| best_oa_location.source.host_organization_lineage | https://openalex.org/I205783295 |
| best_oa_location.license | |
| best_oa_location.pdf_url | https://arxiv.org/pdf/2511.13524 |
| best_oa_location.version | submittedVersion |
| best_oa_location.raw_type | text |
| best_oa_location.license_id | |
| best_oa_location.is_accepted | False |
| best_oa_location.is_published | False |
| best_oa_location.raw_source_name | |
| best_oa_location.landing_page_url | http://arxiv.org/abs/2511.13524 |
| primary_location.id | pmh:oai:arXiv.org:2511.13524 |
| primary_location.is_oa | True |
| primary_location.source.id | https://openalex.org/S4306400194 |
| primary_location.source.issn | |
| primary_location.source.type | repository |
| primary_location.source.is_oa | True |
| primary_location.source.issn_l | |
| primary_location.source.is_core | False |
| primary_location.source.is_in_doaj | False |
| primary_location.source.display_name | arXiv (Cornell University) |
| primary_location.source.host_organization | https://openalex.org/I205783295 |
| primary_location.source.host_organization_name | Cornell University |
| primary_location.source.host_organization_lineage | https://openalex.org/I205783295 |
| primary_location.license | |
| primary_location.pdf_url | https://arxiv.org/pdf/2511.13524 |
| primary_location.version | submittedVersion |
| primary_location.raw_type | text |
| primary_location.license_id | |
| primary_location.is_accepted | False |
| primary_location.is_published | False |
| primary_location.raw_source_name | |
| primary_location.landing_page_url | http://arxiv.org/abs/2511.13524 |
| publication_date | 2025-11-17 |
| publication_year | 2025 |
| referenced_works_count | 0 |
| abstract_inverted_index.a | 5, 64, 86, 107 |
| abstract_inverted_index.16 | 118 |
| abstract_inverted_index.17 | 129 |
| abstract_inverted_index.AI | 141, 188 |
| abstract_inverted_index.As | 0 |
| abstract_inverted_index.We | 26, 101, 143 |
| abstract_inverted_index.an | 29, 208 |
| abstract_inverted_index.as | 4, 207 |
| abstract_inverted_index.by | 48 |
| abstract_inverted_index.in | 8, 185 |
| abstract_inverted_index.of | 50, 131, 139, 180 |
| abstract_inverted_index.on | 162 |
| abstract_inverted_index.to | 20, 134 |
| abstract_inverted_index.we | 77 |
| abstract_inverted_index.Our | 55 |
| abstract_inverted_index.VLN | 145 |
| abstract_inverted_index.and | 43, 52, 62, 97, 103, 126, 137, 147, 153, 172, 194 |
| abstract_inverted_index.can | 94 |
| abstract_inverted_index.for | 39, 70 |
| abstract_inverted_index.our | 200 |
| abstract_inverted_index.six | 114 |
| abstract_inverted_index.the | 75, 79, 178 |
| abstract_inverted_index.both | 151 |
| abstract_inverted_index.core | 6, 119 |
| abstract_inverted_index.data | 66, 133 |
| abstract_inverted_index.into | 85 |
| abstract_inverted_index.more | 127, 195 |
| abstract_inverted_index.must | 14 |
| abstract_inverted_index.seek | 96 |
| abstract_inverted_index.task | 84, 116 |
| abstract_inverted_index.than | 128 |
| abstract_inverted_index.that | 33, 159, 203 |
| abstract_inverted_index.work | 201 |
| abstract_inverted_index.(VLN) | 83 |
| abstract_inverted_index.These | 175 |
| abstract_inverted_index.hours | 130 |
| abstract_inverted_index.human | 148 |
| abstract_inverted_index.large | 35 |
| abstract_inverted_index.their | 165 |
| abstract_inverted_index.under | 150 |
| abstract_inverted_index.(LLMs) | 38 |
| abstract_inverted_index.63,429 | 122 |
| abstract_inverted_index.agents | 93 |
| abstract_inverted_index.beyond | 16 |
| abstract_inverted_index.evolve | 15 |
| abstract_inverted_index.extend | 78 |
| abstract_inverted_index.itself | 205 |
| abstract_inverted_index.models | 37, 160 |
| abstract_inverted_index.object | 120 |
| abstract_inverted_index.sample | 124 |
| abstract_inverted_index.serves | 206 |
| abstract_inverted_index.social | 24, 53 |
| abstract_inverted_index.toward | 190 |
| abstract_inverted_index.types, | 117 |
| abstract_inverted_index.Inquiry | 90 |
| abstract_inverted_index.capture | 21 |
| abstract_inverted_index.classic | 80 |
| abstract_inverted_index.dataset | 110 |
| abstract_inverted_index.diverse | 71, 115 |
| abstract_inverted_index.emerges | 3 |
| abstract_inverted_index.frames, | 125 |
| abstract_inverted_index.models, | 146 |
| abstract_inverted_index.modular | 65 |
| abstract_inverted_index.present | 102 |
| abstract_inverted_index.release | 105 |
| abstract_inverted_index.results | 157 |
| abstract_inverted_index.support | 135 |
| abstract_inverted_index.systems | 189 |
| abstract_inverted_index.wherein | 92 |
| abstract_inverted_index.actively | 95 |
| abstract_inverted_index.behavior | 41 |
| abstract_inverted_index.complex, | 22 |
| abstract_inverted_index.efficacy | 179 |
| abstract_inverted_index.embodied | 1, 72, 140, 187 |
| abstract_inverted_index.enhanced | 169 |
| abstract_inverted_index.enriched | 88 |
| abstract_inverted_index.findings | 176 |
| abstract_inverted_index.frontier | 7 |
| abstract_inverted_index.grounded | 45, 182 |
| abstract_inverted_index.includes | 63 |
| abstract_inverted_index.informed | 47 |
| abstract_inverted_index.language | 36 |
| abstract_inverted_index.original | 166 |
| abstract_inverted_index.physical | 18 |
| abstract_inverted_index.pipeline | 68 |
| abstract_inverted_index.planning | 42, 193 |
| abstract_inverted_index.publicly | 104 |
| abstract_inverted_index.semantic | 170 |
| abstract_inverted_index.setting, | 91 |
| abstract_inverted_index.socially | 181 |
| abstract_inverted_index.supports | 57 |
| abstract_inverted_index.systems. | 142 |
| abstract_inverted_index.tailored | 69 |
| abstract_inverted_index.tasks.To | 73 |
| abstract_inverted_index.theories | 49 |
| abstract_inverted_index.training | 136 |
| abstract_inverted_index.validate | 74 |
| abstract_inverted_index.Direction | 89 |
| abstract_inverted_index.achieving | 168 |
| abstract_inverted_index.advancing | 186 |
| abstract_inverted_index.annotated | 123 |
| abstract_inverted_index.benchmark | 109, 144 |
| abstract_inverted_index.framework | 32, 56 |
| abstract_inverted_index.guidance. | 100 |
| abstract_inverted_index.intention | 51 |
| abstract_inverted_index.interpret | 98 |
| abstract_inverted_index.introduce | 27 |
| abstract_inverted_index.low-level | 17 |
| abstract_inverted_index.modality. | 211 |
| abstract_inverted_index.open-loop | 152 |
| abstract_inverted_index.platforms | 13 |
| abstract_inverted_index.realistic | 59 |
| abstract_inverted_index.research, | 11 |
| abstract_inverted_index.scalable, | 58 |
| abstract_inverted_index.settings. | 155 |
| abstract_inverted_index.Navigation | 82 |
| abstract_inverted_index.additional | 209 |
| abstract_inverted_index.artificial | 9 |
| abstract_inverted_index.behaviors. | 25 |
| abstract_inverted_index.cognition. | 54 |
| abstract_inverted_index.comprising | 111 |
| abstract_inverted_index.evaluation | 138 |
| abstract_inverted_index.fine-tuned | 161 |
| abstract_inverted_index.framework, | 76 |
| abstract_inverted_index.frameworks | 184 |
| abstract_inverted_index.generation | 67 |
| abstract_inverted_index.high-level | 40, 192 |
| abstract_inverted_index.integrates | 34 |
| abstract_inverted_index.outperform | 164 |
| abstract_inverted_index.simulation | 12, 31, 183 |
| abstract_inverted_index.underscore | 177 |
| abstract_inverted_index.categories, | 121 |
| abstract_inverted_index.closed-loop | 154 |
| abstract_inverted_index.competency. | 174 |
| abstract_inverted_index.demonstrate | 158 |
| abstract_inverted_index.human-agent | 60, 197 |
| abstract_inverted_index.information | 210 |
| abstract_inverted_index.interaction | 87, 132, 173, 204 |
| abstract_inverted_index.interactive | 30 |
| abstract_inverted_index.large-scale | 108 |
| abstract_inverted_index.simulations | 61 |
| abstract_inverted_index.underscores | 202 |
| abstract_inverted_index.Experimental | 156 |
| abstract_inverted_index.FreeAskWorld | 163 |
| abstract_inverted_index.Importantly, | 199 |
| abstract_inverted_index.intelligence | 2, 10 |
| abstract_inverted_index.interaction, | 46 |
| abstract_inverted_index.interaction. | 198 |
| abstract_inverted_index.interactions | 19 |
| abstract_inverted_index.naturalistic | 196 |
| abstract_inverted_index.navigational | 99 |
| abstract_inverted_index.participants | 149 |
| abstract_inverted_index.semantically | 44 |
| abstract_inverted_index.FreeAskWorld, | 28, 106 |
| abstract_inverted_index.counterparts, | 167 |
| abstract_inverted_index.environments, | 113 |
| abstract_inverted_index.reconstructed | 112 |
| abstract_inverted_index.sophisticated | 191 |
| abstract_inverted_index.understanding | 171 |
| abstract_inverted_index.human-centered | 23 |
| abstract_inverted_index.Vision-and-Language | 81 |
| cited_by_percentile_year | |
| countries_distinct_count | 0 |
| institutions_distinct_count | 8 |
| citation_normalized_percentile |