A year and a half ago, we published a great feature on the current state of the quest to read the human mind. It included some then in-progress work from Jack Gallant, a neuroscientist at U.C. Berkeley, in which Gallant was attempting to reconstruct a video by reading the brain scans of someone who watched that video--essentially pulling experiences directly from someone's brain. Now, Gallant and his team have published a paper on the subject in the journal Current Biology.
This is the first taste we've gotten of what the study actually produces. Here's a video of the reconstruction in action:
The reconstruction (on the right, obviously) was, according to Gallant, "obtained using only each subject's brain activity and a library of 18 million seconds of random YouTube video that did not include the movies used as stimuli. Brain activity was sampled every one second, and each one-second section of the viewed movie was reconstructed separately."
by "environment clean generations"