Reanimating Images using Neural Representations of Dynamic Stimuli Article Swipe
YOU?
·
· 2024
· Open Access
·
· DOI: https://doi.org/10.48550/arxiv.2406.02659
While computer vision models have made incredible strides in static image recognition, they still do not match human performance in tasks that require the understanding of complex, dynamic motion. This is notably true for real-world scenarios where embodied agents face complex and motion-rich environments. Our approach, BrainNRDS (Brain-Neural Representations of Dynamic Stimuli), leverages state-of-the-art video diffusion models to decouple static image representation from motion generation, enabling us to utilize fMRI brain activity for a deeper understanding of human responses to dynamic visual stimuli. Conversely, we also demonstrate that information about the brain's representation of motion can enhance the prediction of optical flow in artificial systems. Our novel approach leads to four main findings: (1) Visual motion, represented as fine-grained, object-level resolution optical flow, can be decoded from brain activity generated by participants viewing video stimuli; (2) Video encoders outperform image-based models in predicting video-driven brain activity; (3) Brain-decoded motion signals enable realistic video reanimation based only on the initial frame of the video; and (4) We extend prior work to achieve full video decoding from video-driven brain activity. BrainNRDS advances our understanding of how the brain represents spatial and temporal information in dynamic visual scenes. Our findings demonstrate the potential of combining brain imaging with video diffusion models for developing more robust and biologically-inspired computer vision systems. We show additional decoding and encoding examples on this site: https://brain-nrds.github.io/.
Related Topics
- Type
- preprint
- Language
- en
- Landing Page
- http://arxiv.org/abs/2406.02659
- https://arxiv.org/pdf/2406.02659
- OA Status
- green
- Related Works
- 10
- OpenAlex ID
- https://openalex.org/W4399432438
Raw OpenAlex JSON
- OpenAlex ID
-
https://openalex.org/W4399432438Canonical identifier for this work in OpenAlex
- DOI
-
https://doi.org/10.48550/arxiv.2406.02659Digital Object Identifier
- Title
-
Reanimating Images using Neural Representations of Dynamic StimuliWork title
- Type
-
preprintOpenAlex work type
- Language
-
enPrimary language
- Publication year
-
2024Year of publication
- Publication date
-
2024-06-04Full publication date if available
- Authors
-
Jacob Yeung, Andrew F. Luo, Gabriel Sarch, Margaret M. Henderson, Deva Ramanan, Michael J. TarrList of authors in order
- Landing page
-
https://arxiv.org/abs/2406.02659Publisher landing page
- PDF URL
-
https://arxiv.org/pdf/2406.02659Direct link to full text PDF
- Open access
-
YesWhether a free full text is available
- OA status
-
greenOpen access status per OpenAlex
- OA URL
-
https://arxiv.org/pdf/2406.02659Direct OA link when available
- Concepts
-
Computer science, Neuroscience, Cognitive psychology, Artificial intelligence, Psychology, Cognitive scienceTop concepts (fields/topics) attached by OpenAlex
- Cited by
-
0Total citation count in OpenAlex
- Related works (count)
-
10Other works algorithmically related by OpenAlex
Full payload
| id | https://openalex.org/W4399432438 |
|---|---|
| doi | https://doi.org/10.48550/arxiv.2406.02659 |
| ids.doi | https://doi.org/10.48550/arxiv.2406.02659 |
| ids.openalex | https://openalex.org/W4399432438 |
| fwci | 0.0 |
| type | preprint |
| title | Reanimating Images using Neural Representations of Dynamic Stimuli |
| biblio.issue | |
| biblio.volume | |
| biblio.last_page | |
| biblio.first_page | |
| topics[0].id | https://openalex.org/T10427 |
| topics[0].field.id | https://openalex.org/fields/28 |
| topics[0].field.display_name | Neuroscience |
| topics[0].score | 0.7642999887466431 |
| topics[0].domain.id | https://openalex.org/domains/1 |
| topics[0].domain.display_name | Life Sciences |
| topics[0].subfield.id | https://openalex.org/subfields/2805 |
| topics[0].subfield.display_name | Cognitive Neuroscience |
| topics[0].display_name | Visual perception and processing mechanisms |
| is_xpac | False |
| apc_list | |
| apc_paid | |
| concepts[0].id | https://openalex.org/C41008148 |
| concepts[0].level | 0 |
| concepts[0].score | 0.5148707628250122 |
| concepts[0].wikidata | https://www.wikidata.org/wiki/Q21198 |
| concepts[0].display_name | Computer science |
| concepts[1].id | https://openalex.org/C169760540 |
| concepts[1].level | 1 |
| concepts[1].score | 0.46655815839767456 |
| concepts[1].wikidata | https://www.wikidata.org/wiki/Q207011 |
| concepts[1].display_name | Neuroscience |
| concepts[2].id | https://openalex.org/C180747234 |
| concepts[2].level | 1 |
| concepts[2].score | 0.4102746248245239 |
| concepts[2].wikidata | https://www.wikidata.org/wiki/Q23373 |
| concepts[2].display_name | Cognitive psychology |
| concepts[3].id | https://openalex.org/C154945302 |
| concepts[3].level | 1 |
| concepts[3].score | 0.3968794345855713 |
| concepts[3].wikidata | https://www.wikidata.org/wiki/Q11660 |
| concepts[3].display_name | Artificial intelligence |
| concepts[4].id | https://openalex.org/C15744967 |
| concepts[4].level | 0 |
| concepts[4].score | 0.3755807876586914 |
| concepts[4].wikidata | https://www.wikidata.org/wiki/Q9418 |
| concepts[4].display_name | Psychology |
| concepts[5].id | https://openalex.org/C188147891 |
| concepts[5].level | 1 |
| concepts[5].score | 0.3549012839794159 |
| concepts[5].wikidata | https://www.wikidata.org/wiki/Q147638 |
| concepts[5].display_name | Cognitive science |
| keywords[0].id | https://openalex.org/keywords/computer-science |
| keywords[0].score | 0.5148707628250122 |
| keywords[0].display_name | Computer science |
| keywords[1].id | https://openalex.org/keywords/neuroscience |
| keywords[1].score | 0.46655815839767456 |
| keywords[1].display_name | Neuroscience |
| keywords[2].id | https://openalex.org/keywords/cognitive-psychology |
| keywords[2].score | 0.4102746248245239 |
| keywords[2].display_name | Cognitive psychology |
| keywords[3].id | https://openalex.org/keywords/artificial-intelligence |
| keywords[3].score | 0.3968794345855713 |
| keywords[3].display_name | Artificial intelligence |
| keywords[4].id | https://openalex.org/keywords/psychology |
| keywords[4].score | 0.3755807876586914 |
| keywords[4].display_name | Psychology |
| keywords[5].id | https://openalex.org/keywords/cognitive-science |
| keywords[5].score | 0.3549012839794159 |
| keywords[5].display_name | Cognitive science |
| language | en |
| locations[0].id | pmh:oai:arXiv.org:2406.02659 |
| locations[0].is_oa | True |
| locations[0].source.id | https://openalex.org/S4306400194 |
| locations[0].source.issn | |
| locations[0].source.type | repository |
| locations[0].source.is_oa | True |
| locations[0].source.issn_l | |
| locations[0].source.is_core | False |
| locations[0].source.is_in_doaj | False |
| locations[0].source.display_name | arXiv (Cornell University) |
| locations[0].source.host_organization | https://openalex.org/I205783295 |
| locations[0].source.host_organization_name | Cornell University |
| locations[0].source.host_organization_lineage | https://openalex.org/I205783295 |
| locations[0].license | |
| locations[0].pdf_url | https://arxiv.org/pdf/2406.02659 |
| locations[0].version | submittedVersion |
| locations[0].raw_type | text |
| locations[0].license_id | |
| locations[0].is_accepted | False |
| locations[0].is_published | False |
| locations[0].raw_source_name | |
| locations[0].landing_page_url | http://arxiv.org/abs/2406.02659 |
| locations[1].id | doi:10.48550/arxiv.2406.02659 |
| locations[1].is_oa | True |
| locations[1].source.id | https://openalex.org/S4306400194 |
| locations[1].source.issn | |
| locations[1].source.type | repository |
| locations[1].source.is_oa | True |
| locations[1].source.issn_l | |
| locations[1].source.is_core | False |
| locations[1].source.is_in_doaj | False |
| locations[1].source.display_name | arXiv (Cornell University) |
| locations[1].source.host_organization | https://openalex.org/I205783295 |
| locations[1].source.host_organization_name | Cornell University |
| locations[1].source.host_organization_lineage | https://openalex.org/I205783295 |
| locations[1].license | cc-by |
| locations[1].pdf_url | |
| locations[1].version | |
| locations[1].raw_type | article-journal |
| locations[1].license_id | https://openalex.org/licenses/cc-by |
| locations[1].is_accepted | False |
| locations[1].is_published | |
| locations[1].raw_source_name | |
| locations[1].landing_page_url | https://doi.org/10.48550/arxiv.2406.02659 |
| indexed_in | arxiv, datacite |
| authorships[0].author.id | https://openalex.org/A5014857783 |
| authorships[0].author.orcid | https://orcid.org/0000-0002-7993-6666 |
| authorships[0].author.display_name | Jacob Yeung |
| authorships[0].author_position | first |
| authorships[0].raw_author_name | Yeung, Jacob |
| authorships[0].is_corresponding | False |
| authorships[1].author.id | https://openalex.org/A5113261108 |
| authorships[1].author.orcid | |
| authorships[1].author.display_name | Andrew F. Luo |
| authorships[1].author_position | middle |
| authorships[1].raw_author_name | Luo, Andrew F. |
| authorships[1].is_corresponding | False |
| authorships[2].author.id | https://openalex.org/A5035580849 |
| authorships[2].author.orcid | https://orcid.org/0000-0002-4396-7612 |
| authorships[2].author.display_name | Gabriel Sarch |
| authorships[2].author_position | middle |
| authorships[2].raw_author_name | Sarch, Gabriel |
| authorships[2].is_corresponding | False |
| authorships[3].author.id | https://openalex.org/A5070257872 |
| authorships[3].author.orcid | https://orcid.org/0000-0001-9375-6680 |
| authorships[3].author.display_name | Margaret M. Henderson |
| authorships[3].author_position | middle |
| authorships[3].raw_author_name | Henderson, Margaret M. |
| authorships[3].is_corresponding | False |
| authorships[4].author.id | https://openalex.org/A5004353237 |
| authorships[4].author.orcid | https://orcid.org/0009-0008-9180-8983 |
| authorships[4].author.display_name | Deva Ramanan |
| authorships[4].author_position | middle |
| authorships[4].raw_author_name | Ramanan, Deva |
| authorships[4].is_corresponding | False |
| authorships[5].author.id | https://openalex.org/A5008604805 |
| authorships[5].author.orcid | https://orcid.org/0000-0003-4724-1744 |
| authorships[5].author.display_name | Michael J. Tarr |
| authorships[5].author_position | last |
| authorships[5].raw_author_name | Tarr, Michael J. |
| authorships[5].is_corresponding | False |
| has_content.pdf | False |
| has_content.grobid_xml | False |
| is_paratext | False |
| open_access.is_oa | True |
| open_access.oa_url | https://arxiv.org/pdf/2406.02659 |
| open_access.oa_status | green |
| open_access.any_repository_has_fulltext | False |
| created_date | 2024-06-08T00:00:00 |
| display_name | Reanimating Images using Neural Representations of Dynamic Stimuli |
| has_fulltext | False |
| is_retracted | False |
| updated_date | 2025-11-06T06:51:31.235846 |
| primary_topic.id | https://openalex.org/T10427 |
| primary_topic.field.id | https://openalex.org/fields/28 |
| primary_topic.field.display_name | Neuroscience |
| primary_topic.score | 0.7642999887466431 |
| primary_topic.domain.id | https://openalex.org/domains/1 |
| primary_topic.domain.display_name | Life Sciences |
| primary_topic.subfield.id | https://openalex.org/subfields/2805 |
| primary_topic.subfield.display_name | Cognitive Neuroscience |
| primary_topic.display_name | Visual perception and processing mechanisms |
| related_works | https://openalex.org/W4391375266, https://openalex.org/W2748952813, https://openalex.org/W2931662336, https://openalex.org/W2077865380, https://openalex.org/W2765597752, https://openalex.org/W2134894512, https://openalex.org/W2083375246, https://openalex.org/W2067108088, https://openalex.org/W2085372204, https://openalex.org/W4391301621 |
| cited_by_count | 0 |
| locations_count | 2 |
| best_oa_location.id | pmh:oai:arXiv.org:2406.02659 |
| best_oa_location.is_oa | True |
| best_oa_location.source.id | https://openalex.org/S4306400194 |
| best_oa_location.source.issn | |
| best_oa_location.source.type | repository |
| best_oa_location.source.is_oa | True |
| best_oa_location.source.issn_l | |
| best_oa_location.source.is_core | False |
| best_oa_location.source.is_in_doaj | False |
| best_oa_location.source.display_name | arXiv (Cornell University) |
| best_oa_location.source.host_organization | https://openalex.org/I205783295 |
| best_oa_location.source.host_organization_name | Cornell University |
| best_oa_location.source.host_organization_lineage | https://openalex.org/I205783295 |
| best_oa_location.license | |
| best_oa_location.pdf_url | https://arxiv.org/pdf/2406.02659 |
| best_oa_location.version | submittedVersion |
| best_oa_location.raw_type | text |
| best_oa_location.license_id | |
| best_oa_location.is_accepted | False |
| best_oa_location.is_published | False |
| best_oa_location.raw_source_name | |
| best_oa_location.landing_page_url | http://arxiv.org/abs/2406.02659 |
| primary_location.id | pmh:oai:arXiv.org:2406.02659 |
| primary_location.is_oa | True |
| primary_location.source.id | https://openalex.org/S4306400194 |
| primary_location.source.issn | |
| primary_location.source.type | repository |
| primary_location.source.is_oa | True |
| primary_location.source.issn_l | |
| primary_location.source.is_core | False |
| primary_location.source.is_in_doaj | False |
| primary_location.source.display_name | arXiv (Cornell University) |
| primary_location.source.host_organization | https://openalex.org/I205783295 |
| primary_location.source.host_organization_name | Cornell University |
| primary_location.source.host_organization_lineage | https://openalex.org/I205783295 |
| primary_location.license | |
| primary_location.pdf_url | https://arxiv.org/pdf/2406.02659 |
| primary_location.version | submittedVersion |
| primary_location.raw_type | text |
| primary_location.license_id | |
| primary_location.is_accepted | False |
| primary_location.is_published | False |
| primary_location.raw_source_name | |
| primary_location.landing_page_url | http://arxiv.org/abs/2406.02659 |
| publication_date | 2024-06-04 |
| publication_year | 2024 |
| referenced_works_count | 0 |
| abstract_inverted_index.a | 73 |
| abstract_inverted_index.We | 165, 217 |
| abstract_inverted_index.as | 117 |
| abstract_inverted_index.be | 124 |
| abstract_inverted_index.by | 130 |
| abstract_inverted_index.do | 14 |
| abstract_inverted_index.in | 8, 19, 102, 141, 191 |
| abstract_inverted_index.is | 30 |
| abstract_inverted_index.of | 25, 49, 76, 93, 99, 160, 182, 200 |
| abstract_inverted_index.on | 156, 224 |
| abstract_inverted_index.to | 57, 67, 79, 109, 169 |
| abstract_inverted_index.us | 66 |
| abstract_inverted_index.we | 84 |
| abstract_inverted_index.(1) | 113 |
| abstract_inverted_index.(2) | 135 |
| abstract_inverted_index.(3) | 146 |
| abstract_inverted_index.(4) | 164 |
| abstract_inverted_index.Our | 44, 105, 195 |
| abstract_inverted_index.and | 41, 163, 188, 212, 221 |
| abstract_inverted_index.can | 95, 123 |
| abstract_inverted_index.for | 33, 72, 208 |
| abstract_inverted_index.how | 183 |
| abstract_inverted_index.not | 15 |
| abstract_inverted_index.our | 180 |
| abstract_inverted_index.the | 23, 90, 97, 157, 161, 184, 198 |
| abstract_inverted_index.This | 29 |
| abstract_inverted_index.also | 85 |
| abstract_inverted_index.fMRI | 69 |
| abstract_inverted_index.face | 39 |
| abstract_inverted_index.flow | 101 |
| abstract_inverted_index.four | 110 |
| abstract_inverted_index.from | 62, 126, 174 |
| abstract_inverted_index.full | 171 |
| abstract_inverted_index.have | 4 |
| abstract_inverted_index.made | 5 |
| abstract_inverted_index.main | 111 |
| abstract_inverted_index.more | 210 |
| abstract_inverted_index.only | 155 |
| abstract_inverted_index.show | 218 |
| abstract_inverted_index.that | 21, 87 |
| abstract_inverted_index.they | 12 |
| abstract_inverted_index.this | 225 |
| abstract_inverted_index.true | 32 |
| abstract_inverted_index.with | 204 |
| abstract_inverted_index.work | 168 |
| abstract_inverted_index.Video | 136 |
| abstract_inverted_index.While | 0 |
| abstract_inverted_index.about | 89 |
| abstract_inverted_index.based | 154 |
| abstract_inverted_index.brain | 70, 127, 144, 176, 185, 202 |
| abstract_inverted_index.flow, | 122 |
| abstract_inverted_index.frame | 159 |
| abstract_inverted_index.human | 17, 77 |
| abstract_inverted_index.image | 10, 60 |
| abstract_inverted_index.leads | 108 |
| abstract_inverted_index.match | 16 |
| abstract_inverted_index.novel | 106 |
| abstract_inverted_index.prior | 167 |
| abstract_inverted_index.site: | 226 |
| abstract_inverted_index.still | 13 |
| abstract_inverted_index.tasks | 20 |
| abstract_inverted_index.video | 54, 133, 152, 172, 205 |
| abstract_inverted_index.where | 36 |
| abstract_inverted_index.Visual | 114 |
| abstract_inverted_index.agents | 38 |
| abstract_inverted_index.deeper | 74 |
| abstract_inverted_index.enable | 150 |
| abstract_inverted_index.extend | 166 |
| abstract_inverted_index.models | 3, 56, 140, 207 |
| abstract_inverted_index.motion | 63, 94, 148 |
| abstract_inverted_index.robust | 211 |
| abstract_inverted_index.static | 9, 59 |
| abstract_inverted_index.video; | 162 |
| abstract_inverted_index.vision | 2, 215 |
| abstract_inverted_index.visual | 81, 193 |
| abstract_inverted_index.Dynamic | 50 |
| abstract_inverted_index.achieve | 170 |
| abstract_inverted_index.brain's | 91 |
| abstract_inverted_index.complex | 40 |
| abstract_inverted_index.decoded | 125 |
| abstract_inverted_index.dynamic | 27, 80, 192 |
| abstract_inverted_index.enhance | 96 |
| abstract_inverted_index.imaging | 203 |
| abstract_inverted_index.initial | 158 |
| abstract_inverted_index.motion, | 115 |
| abstract_inverted_index.motion. | 28 |
| abstract_inverted_index.notably | 31 |
| abstract_inverted_index.optical | 100, 121 |
| abstract_inverted_index.require | 22 |
| abstract_inverted_index.scenes. | 194 |
| abstract_inverted_index.signals | 149 |
| abstract_inverted_index.spatial | 187 |
| abstract_inverted_index.strides | 7 |
| abstract_inverted_index.utilize | 68 |
| abstract_inverted_index.viewing | 132 |
| abstract_inverted_index.activity | 71, 128 |
| abstract_inverted_index.advances | 179 |
| abstract_inverted_index.approach | 107 |
| abstract_inverted_index.complex, | 26 |
| abstract_inverted_index.computer | 1, 214 |
| abstract_inverted_index.decoding | 173, 220 |
| abstract_inverted_index.decouple | 58 |
| abstract_inverted_index.embodied | 37 |
| abstract_inverted_index.enabling | 65 |
| abstract_inverted_index.encoders | 137 |
| abstract_inverted_index.encoding | 222 |
| abstract_inverted_index.examples | 223 |
| abstract_inverted_index.findings | 196 |
| abstract_inverted_index.stimuli. | 82 |
| abstract_inverted_index.stimuli; | 134 |
| abstract_inverted_index.systems. | 104, 216 |
| abstract_inverted_index.temporal | 189 |
| abstract_inverted_index.BrainNRDS | 46, 178 |
| abstract_inverted_index.Stimuli), | 51 |
| abstract_inverted_index.activity. | 177 |
| abstract_inverted_index.activity; | 145 |
| abstract_inverted_index.approach, | 45 |
| abstract_inverted_index.combining | 201 |
| abstract_inverted_index.diffusion | 55, 206 |
| abstract_inverted_index.findings: | 112 |
| abstract_inverted_index.generated | 129 |
| abstract_inverted_index.leverages | 52 |
| abstract_inverted_index.potential | 199 |
| abstract_inverted_index.realistic | 151 |
| abstract_inverted_index.responses | 78 |
| abstract_inverted_index.scenarios | 35 |
| abstract_inverted_index.additional | 219 |
| abstract_inverted_index.artificial | 103 |
| abstract_inverted_index.developing | 209 |
| abstract_inverted_index.incredible | 6 |
| abstract_inverted_index.outperform | 138 |
| abstract_inverted_index.predicting | 142 |
| abstract_inverted_index.prediction | 98 |
| abstract_inverted_index.real-world | 34 |
| abstract_inverted_index.represents | 186 |
| abstract_inverted_index.resolution | 120 |
| abstract_inverted_index.Conversely, | 83 |
| abstract_inverted_index.demonstrate | 86, 197 |
| abstract_inverted_index.generation, | 64 |
| abstract_inverted_index.image-based | 139 |
| abstract_inverted_index.information | 88, 190 |
| abstract_inverted_index.motion-rich | 42 |
| abstract_inverted_index.performance | 18 |
| abstract_inverted_index.reanimation | 153 |
| abstract_inverted_index.represented | 116 |
| abstract_inverted_index.object-level | 119 |
| abstract_inverted_index.participants | 131 |
| abstract_inverted_index.recognition, | 11 |
| abstract_inverted_index.video-driven | 143, 175 |
| abstract_inverted_index.(Brain-Neural | 47 |
| abstract_inverted_index.Brain-decoded | 147 |
| abstract_inverted_index.environments. | 43 |
| abstract_inverted_index.fine-grained, | 118 |
| abstract_inverted_index.understanding | 24, 75, 181 |
| abstract_inverted_index.representation | 61, 92 |
| abstract_inverted_index.Representations | 48 |
| abstract_inverted_index.state-of-the-art | 53 |
| abstract_inverted_index.biologically-inspired | 213 |
| abstract_inverted_index.https://brain-nrds.github.io/. | 227 |
| cited_by_percentile_year | |
| countries_distinct_count | 0 |
| institutions_distinct_count | 6 |
| citation_normalized_percentile.value | 0.1054074 |
| citation_normalized_percentile.is_in_top_1_percent | False |
| citation_normalized_percentile.is_in_top_10_percent | False |