Brain-reading

From HandWiki
Revision as of 21:43, 6 February 2024 by NBrush (talk | contribs) (add)
(diff) ← Older revision | Latest revision (diff) | Newer revision → (diff)
Short description: Use of fMRI to decode brain stimuli

Brain-reading or thought identification uses the responses of multiple voxels in the brain evoked by stimulus then detected by fMRI in order to decode the original stimulus. Advances in research have made this possible by using human neuroimaging to decode a person's conscious experience based on non-invasive measurements of an individual's brain activity.[1] Brain reading studies differ in the type of decoding (i.e. classification, identification and reconstruction) employed, the target (i.e. decoding visual patterns, auditory patterns, cognitive states), and the decoding algorithms (linear classification, nonlinear classification, direct reconstruction, Bayesian reconstruction, etc.) employed.

In 2007, Professor of neuropsychology Barbara Sahakian qualified, "A lot of neuroscientists in the field are very cautious and say we can't talk about reading individuals' minds, and right now that is very true, but we're moving ahead so rapidly, it's not going to be that long before we will be able to tell whether someone's making up a story, or whether someone intended to do a crime with a certain degree of certainty."[2]


Applications

Natural images

Identification of complex natural images is possible using voxels from early and anterior visual cortex areas forward of them (visual areas V3A, V3B, V4, and the lateral occipital) together with Bayesian inference. This brain reading approach uses three components:[3] a structural encoding model that characterizes responses in early visual areas; a semantic encoding model that characterizes responses in anterior visual areas; and a Bayesian prior that describes the distribution of structural and semantic scene statistics.[3]

Experimentally the procedure is for subjects to view 1750 black and white natural images that are correlated with voxel activation in their brains. Then subjects viewed another 120 novel target images, and information from the earlier scans is used reconstruct them. Natural images used include pictures of a seaside cafe and harbor, performers on a stage, and dense foliage.[3]

In 2008 IBM applied for a patent on how to extract mental images of human faces from the human brain. It uses a feedback loop based on brain measurements of the fusiform gyrus area in the brain which activates proportionate with degree of facial recognition.[4]

In 2011, a team led by Shinji Nishimoto used only brain recordings to partially reconstruct what volunteers were seeing. The researchers applied a new model, about how moving object information is processed in human brains, while volunteers watched clips from several videos. An algorithm searched through thousands of hours of external YouTube video footage (none of the videos were the same as the ones the volunteers watched) to select the clips that were most similar.[5][6] The authors have uploaded demos comparing the watched and the computer-estimated videos.[7][8]

In 2017 a face perception study in monkeys reported the reconstruction of human faces by analyzing electrical activity from 205 neurons.[9][10]

In 2023 image reconstruction was reported utilizing Stable Diffusion on human brain activity obtained via fMRI.[11][12]

Lie detector

Brain-reading has been suggested as an alternative to polygraph machines as a form of lie detection.[13] Another alternative to polygraph machines is blood oxygenated level dependent functional MRI technology (BOLD fMRI). This technique involves the interpretation of the local change in the concentration of oxygenated hemoglobin in the brain, although the relationship between this blood flow and neural activity is not yet completely understood.[13] Another technique to find concealed information is brain fingerprinting, which uses EEG to ascertain if a person has a specific memory or information by identifying P300 event related potentials.[14]

A number of concerns have been raised about the accuracy and ethical implications of brain-reading for this purpose. Laboratory studies have found rates of accuracy of up to 85%; however, there are concerns about what this means for false positive results among non-criminal populations: "If the prevalence of "prevaricators" in the group being examined is low, the test will yield far more false-positive than true-positive results; about one person in five will be incorrectly identified by the test."[13] Ethical problems involved in the use of brain-reading as lie detection include misapplications due to adoption of the technology before its reliability and validity can be properly assessed and due to misunderstanding of the technology, and privacy concerns due to unprecedented access to individual's private thoughts.[13] However, it has been noted that the use of polygraph lie detection carries similar concerns about the reliability of the results[13] and violation of privacy.[15]

Human–machine interfaces

The Emotiv Epoc is one way that users can give commands to devices using only thoughts.

Brain-reading has also been proposed as a method of improving human–machine interfaces, by the use of EEG to detect relevant brain states of a human.[16] In recent years, there has been a rapid increase in patents for technology involved in reading brainwaves, rising from fewer than 400 from 2009–2012 to 1600 in 2014.[17] These include proposed ways to control video games via brain waves and "neuro-marketing" to determine someone's thoughts about a new product or advertisement.

Emotiv Systems, an Australian electronics company, has demonstrated a headset that can be trained to recognize a user's thought patterns for different commands. Tan Le demonstrated the headset's ability to manipulate virtual objects on screen, and discussed various future applications for such brain-computer interface devices, from powering wheel chairs to replacing the mouse and keyboard.[18]

Detecting attention

It is possible to track which of two forms of rivalrous binocular illusions a person was subjectively experiencing from fMRI signals.[19]

When humans think of an object, such as a screwdriver, many different areas of the brain activate. Marcel Just and his colleague, Tom Mitchell, have used fMRI brain scans to teach a computer to identify the various parts of the brain associated with specific thoughts.[20] This technology also yielded a discovery: similar thoughts in different human brains are surprisingly similar neurologically. To illustrate this, Just and Mitchell used their computer to predict, based on nothing but fMRI data, which of several images a volunteer was thinking about. The computer was 100% accurate, but so far the machine is only distinguishing between 10 images.[20]

Detecting thoughts

The category of event which a person freely recalls can be identified from fMRI before they say what they remembered.[21]

December 16, 2015, a study conducted by Toshimasa Yamazaki at Kyushu Institute of Technology found that during a rock-paper-scissors game a computer was able to determine the choice made by the subjects before they moved their hand. An EEG was used to measure activity in the Broca's area to see the words two seconds before the words were uttered.[22][23][24]

In 2023, the University of Texas in Austin trained a non-invasive brain decoder to translate volunteers' brainwaves into the GPT-1 language model. After lengthy training on each individual volunteer, the decoder usually failed to reconstruct the exact words, but could nevertheless reconstruct meanings close enough that the decoder could, most of the time, identify what timestamp of a given book the subject was listening to.[25][26]

Detecting language

Statistical analysis of EEG brainwaves has been claimed to allow the recognition of phonemes,[27] and (in 1999) at a 60% to 75% level color and visual shape words.[28]

On 31 January 2012 Brian Pasley and colleagues of University of California Berkeley published their paper in PLoS Biology wherein subjects' internal neural processing of auditory information was decoded and reconstructed as sound on computer by gathering and analyzing electrical signals directly from subjects' brains.[29] The research team conducted their studies on the superior temporal gyrus, a region of the brain that is involved in higher order neural processing to make semantic sense from auditory information.[30] The research team used a computer model to analyze various parts of the brain that might be involved in neural firing while processing auditory signals. Using the computational model, scientists were able to identify the brain activity involved in processing auditory information when subjects were presented with recording of individual words.[31] Later, the computer model of auditory information processing was used to reconstruct some of the words back into sound based on the neural processing of the subjects. However the reconstructed sounds were not of good quality and could be recognized only when the audio wave patterns of the reconstructed sound were visually matched with the audio wave patterns of the original sound that was presented to the subjects.[31] However this research marks a direction towards more precise identification of neural activity in cognition.

Predicting intentions

Some researchers in 2008 were able to predict, with 60% accuracy, whether a subject was going to push a button with their left or right hand. This is notable, not just because the accuracy is better than chance, but also because the scientists were able to make these predictions up to 10 seconds before the subject acted – well before the subject felt they had decided.[32] This data is even more striking in light of other research suggesting that the decision to move, and possibly the ability to cancel that movement at the last second,[33] may be the results of unconscious processing.[34]

John Dylan-Haynes has also demonstrated that fMRI can be used to identify whether a volunteer is about to add or subtract two numbers in their head.[20]

Predictive processing in the brain

Neural decoding techniques have been used to test theories about the predictive brain, and to investigate how top-down predictions affect brain areas such as the visual cortex. Studies using fMRI decoding techniques have found that predictable sensory events[35] and the expected consequences of our actions[36] are better decoded in visual brain areas, suggesting that prediction 'sharpens' representations in line with expectations.

Virtual environments

It has also been shown that brain-reading can be achieved in a complex virtual environment.[37]

Emotions

Just and Mitchell also claim they are beginning to be able to identify kindness, hypocrisy, and love in the brain.[20]

Security

In 2013 a project led by University of California Berkeley professor John Chuang published findings on the feasibility of brainwave-based computer authentication as a substitute for passwords. Improvements in the use of biometrics for computer authentication has continually improved since the 1980s, but this research team was looking for a method faster and less intrusive than today's retina scans, fingerprinting, and voice recognition. The technology chosen to improve security measures is an electroencephalogram (EEG), or brainwave measurer, to improve passwords into "pass thoughts." Using this method Chuang and his team were able to customize tasks and their authentication thresholds to the point where they were able to reduce error rates under 1%, significantly better than other recent methods. In order to better attract users to this new form of security the team is still researching mental tasks that are enjoyable for the user to perform while having their brainwaves identified. In the future this method could be as cheap, accessible, and straightforward as thought itself.[38]

John-Dylan Haynes states that fMRI can also be used to identify recognition in the brain. He provides the example of a criminal being interrogated about whether he recognizes the scene of the crime or murder weapons.[20]

Methods of analysis

Classification

In classification, a pattern of activity across multiple voxels is used to determine the particular class from which the stimulus was drawn.[39] Many studies have classified visual stimuli, but this approach has also been used to classify cognitive states.[citation needed]

Reconstruction

In reconstruction brain reading the aim is to create a literal picture of the image that was presented. Early studies used voxels from early visual cortex areas (V1, V2, and V3) to reconstruct geometric stimuli made up of flickering checkerboard patterns.[40][41]

EEG

EEG has also been used to identify recognition of specific information or memories by the P300 event related potential, which has been dubbed 'brain fingerprinting'.[42]

Accuracy

Brain-reading accuracy is increasing steadily as the quality of the data and the complexity of the decoding algorithms improve. In one recent experiment it was possible to identify which single image was being seen from a set of 120.[43] In another it was possible to correctly identify 90% of the time which of two categories the stimulus came and the specific semantic category (out of 23) of the target image 40% of the time.[3]

Limitations

It has been noted that so far brain-reading is limited. "In practice, exact reconstructions are impossible to achieve by any reconstruction algorithm on the basis of brain activity signals acquired by fMRI. This is because all reconstructions will inevitably be limited by inaccuracies in the encoding models and noise in the measured signals. Our results[who?] demonstrate that the natural image prior is a powerful (if unconventional) tool for mitigating the effects of these fundamental limitations. A natural image prior with only six million images is sufficient to produce reconstructions that are structurally and semantically similar to a target image."[3]

Ethical issues

With brain scanning technology becoming increasingly accurate, experts predict important debates over how and when it should be used. One potential area of application is criminal law. Haynes states that simply refusing to use brain scans on suspects also prevents the wrongly accused from proving their innocence.[2] US scholars generally believe that involuntary brain reading, and involuntary polygraph tests, would violate the 5th Amendment's right to not self incriminate.[44][45] One perspective is to consider whether brain imaging is like testimony, or instead like DNA, blood, or semen. Paul Root Wolpe, director of the Center for Ethics at Emory University in Atlanta predicts that this question will be decided by a Supreme Court case.[46]

In other countries outside the United States, thought identification has already been used in criminal law. In 2008 an Indian woman was convicted of murder after an EEG of her brain allegedly revealed that she was familiar with the circumstances surrounding the poisoning of her ex-fiancé.[46] Some neuroscientists and legal scholars doubt the validity of using thought identification as a whole for anything past research on the nature of deception and the brain.[47]

The Economist cautioned people to be "afraid" of the future impact, and some ethicists argue that privacy laws should protect private thoughts. Legal scholar Hank Greely argues that the court systems could benefit from such technology, and neuroethicist Julian Savulescu states that brain data is not fundamentally different from other types of evidence.[48] In Nature, journalist Liam Drew writes about emerging projects to attach brain-reading devices to speech synthesizers or other output devices for the benefit of tetraplegics. Such devices could create concerns of accidentally broadcasting the patient's "inner thoughts" rather than merely conscious speech.[49]

History

MRI scanner that could be used for Thought Identification

Psychologist John-Dylan Haynes experienced breakthroughs in brain imaging research in 2006 by using fMRI. This research included new findings on visual object recognition, tracking dynamic mental processes, lie detecting, and decoding unconscious processing. The combination of these four discoveries revealed such a significant amount of information about an individual's thoughts that Haynes termed it "brain reading".[1]

The fMRI has allowed research to expand by significant amounts because it can track the activity in an individual's brain by measuring the brain's blood flow. It is currently thought to be the best method for measuring brain activity, which is why it has been used in multiple research experiments in order to improve the understanding of how doctors and psychologists can identify thoughts.[50]

In a 2020 study, AI using implanted electrodes could correctly transcribe a sentence read aloud from a fifty-sentence test set 97% of the time, given 40 minutes of training data per participant.[51]

Future research

Experts are unsure of how far thought identification can expand, but Marcel Just believed in 2014 that in 3–5 years there will be a machine that is able to read complex thoughts such as 'I hate so-and-so'.[46]

Donald Marks, founder and chief science officer of MMT, is working on playing back thoughts individuals have after they have already been recorded.[52]

Researchers at the University of California Berkeley have already been successful in forming, erasing, and reactivating memories in rats. Marks says they are working on applying the same techniques to humans. This discovery could be monumental for war veterans who suffer from PTSD.[52]

Further research is also being done in analyzing brain activity during video games to detect criminals, neuromarketing, and using brain scans in government security checks.[46][50]

In popular culture

Episode Black Hole of American medical drama House, which aired on March 15, 2010, featured an experimental "cognitive imaging" device that supposedly allowed seeing into a patient's subconscious mind. The patient was first put in a preparation phase of six hours while watching video clips, attached to a neuroimaging device looking like electroencephalography or Functional near-infrared spectroscopy, to train the neuroimaging classifier. Then the patient was put under twilight anesthesia, and the same device was used to try to infer what was going through the patient's mind. The fictional episode somewhat anticipated the study by Nishimoto et al. published the following year, in which fMRI was used instead.[5][6][7][8] In the movie Dumb and Dumber To, one scene shows a brain reader. In the Henry Danger episode, "Dream Busters," a machine showed Henry's dream.

See also

References

  1. 1.0 1.1 Haynes, John-Dylan; Geraint, Rees (1 July 2006). "Decoding mental states from brain activity in humans". Nature Reviews Neuroscience 7 (7): 523–534. doi:10.1038/nrn1931. PMID 16791142. http://www.nature.com/nrn/journal/v7/n7/full/nrn1931.html. Retrieved 8 December 2014. 
  2. 2.0 2.1 Sample, Ian (9 February 2007). "The brain scan that can read people's intentions". The Guardian. https://www.theguardian.com/science/2007/feb/09/neuroscience.ethicsofscience. 
  3. 3.0 3.1 3.2 3.3 3.4 Naselaris, Thomas; Prenger, Ryan J.; Kay, Kendrick N.; Oliver, Michael; Gallant, Jack L. (2009). "Bayesian Reconstruction of Natural Images from Human Brain Activity". Neuron 63 (6): 902–15. doi:10.1016/j.neuron.2009.09.006. PMID 19778517. 
  4. "IBM Patent Application US 2010/0049076: Retrieving mental images of faces from the human brain". February 25, 2010. http://www.freepatentsonline.com/20100049076.pdf. 
  5. 5.0 5.1 Nishimoto, Shinji; Vu, An T.; Naselaris, Thomas; Benjamini, Yuval; Yu, Bin; Gallant, Jack L. (2011), "Reconstructing Visual Experiences from Brain Activity Evoked by Natural Movies", Current Biology 21 (19): 1641–1646, doi:10.1016/j.cub.2011.08.031, PMID 21945275 
  6. 6.0 6.1 "Breakthrough Could Enable Others to Watch Your Dreams and Memories [Video], Philip Yam". https://blogs.scientificamerican.com/observations/breakthrough-could-enable-others-to-watch-your-dreams-and-memories-video/. 
  7. 7.0 7.1 Nishimoto et al. 2011 uploaded video 1 Movie reconstruction from human brain activity on Youtube
  8. 8.0 8.1 Nishimoto et al. 2011 uploaded video 2 Movie reconstructions from human brain activity: 3 subjects, "Nishimoto.etal.2011.3Subjects.mpeg" on Youtube
  9. Chang, Le; Tsao, Doris Y. (2017-06-01). "The Code for Facial Identity in the Primate Brain" (in English). Cell 169 (6): 1013–1028.e14. doi:10.1016/j.cell.2017.05.011. ISSN 0092-8674. PMID 28575666. 
  10. Devlin, Hannah; Correspondent, Science (2017-06-01). "Scientists discover how the brain recognises faces – by reading monkeys' minds" (in en-GB). The Guardian. ISSN 0261-3077. https://www.theguardian.com/science/2017/jun/01/scientists-discover-how-the-brain-recognises-faces-by-reading-monkeys-minds. 
  11. Takagi, Yu; Nishimoto, Shinji (2022-12-01) (in en). High-resolution image reconstruction with latent diffusion models from human brain activity. pp. 2022.11.18.517004. doi:10.1101/2022.11.18.517004. https://www.biorxiv.org/content/10.1101/2022.11.18.517004v2. 
  12. "AI re-creates what people see by reading their brain scans" (in en). https://www.science.org/content/article/ai-re-creates-what-people-see-reading-their-brain-scans. 
  13. 13.0 13.1 13.2 13.3 13.4 Wolpe, P. R.; Foster, K. R.; Langleben, D. D. (2005). "Emerging neurotechnologies for lie-detection: promises and perils". The American Journal of Bioethics 5 (2): 39–49. doi:10.1080/15265160590923367. PMID 16036700. 
  14. Farwell, Lawrence A.; Richardson, Drew C.; Richardson, Graham M. (5 December 2012). "Brain fingerprinting field studies comparing P300-MERMER and P300 brainwave responses in the detection of concealed information". Cognitive Neurodynamics 7 (4): 263–299. doi:10.1007/s11571-012-9230-0. PMID 23869200. 
  15. Arstila, V.; Scott, F. (2011). "Brain Reading and Mental Privacy". TRAMES: A Journal of the Humanities & Social Sciences 15 (2): 204–212. doi:10.3176/tr.2011.2.08. http://www.kirj.ee/public/trames_pdf/2011/issue_2/Trames-2011-2-204-212.pdf. 
  16. Kirchner, E. A.; Kim, S. K.; Straube, S.; Seeland, A.; Wöhrle, H.; Krell, M. M.; Tabie, M.; Fahle, M. (2013). "On the Applicability of Brain Reading for Predictive Human–Machine Interfaces in Robotics". PLOS ONE 8 (12): e81732. doi:10.1371/journal.pone.0081732. PMID 24358125. 
  17. "Surge in U.S. 'brain-reading' patents". BBC News. May 7, 2015. https://www.bbc.com/news/technology-32623063. 
  18. Le, Tan (2010). "A headset that reads your brainwaves". TEDGlobal. https://www.ted.com/talks/tan_le_a_headset_that_reads_your_brainwaves?language=en. 
  19. Haynes, J; Rees, G (2005). "Predicting the Stream of Consciousness from Activity in Human Visual Cortex". Current Biology 15 (14): 1301–7. doi:10.1016/j.cub.2005.06.026. PMID 16051174. 
  20. 20.0 20.1 20.2 20.3 20.4 "'60 Minutes' video: Tech that reads your mind". CNET News. January 5, 2009. http://news.cnet.com/8301-11386_3-10131643-76.html. 
  21. Polyn, S. M.; Natu, VS; Cohen, JD; Norman, KA (2005). "Category-Specific Cortical Activity Precedes Retrieval During Memory Search". Science 310 (5756): 1963–6. doi:10.1126/science.1117645. PMID 16373577. 
  22. Ito, Takashi; Yamaguchi, Hiromi; Yamaguchi, Ayaka; Yamazaki, Toshimasa; Fukuzumi, Shin'Ichi; Yamanoi, Takahiro (2015-12-16). "Silent Speech BCI - An investigation for practical problems". IEICE Technical Report (IEICE Technical Committee) 114 (514): 81–84. https://www.ieice.org/ken/paper/20150317SBYj/eng/. Retrieved 2016-01-17. 
  23. Danigelis, Alyssa (2016-01-07). "Mind-Reading Computer Knows What You're About to Say". Discovery News. http://news.discovery.com/tech/gear-and-gadgets/mind-reading-computer-knows-what-youre-about-to-say-160107.htm. 
  24. "頭の中の言葉 解読 障害者と意思疎通、ロボット操作も 九工大・山崎教授ら" (in ja). Nishinippon Shimbun. 2016-01-04. http://www.nishinippon.co.jp/nnp/national/article/216215. 
  25. O'Sullivan, Donie (23 May 2023). "How the technology behind ChatGPT could make mind-reading a reality" (in en). CNN. https://www.cnn.com/2023/05/23/tech/chatgpt-mind-reading/index.html. 
  26. Tang, Jerry; LeBel, Amanda; Jain, Shailee; Huth, Alexander G. (May 2023). "Semantic reconstruction of continuous language from non-invasive brain recordings". Nature Neuroscience 26 (5): 858–866. doi:10.1038/s41593-023-01304-9. PMID 37127759. https://www.nature.com/articles/s41593-023-01304-9. 
  27. Suppes, Patrick; Perreau-Guimaraes, Marcos; Wong, Dik Kin (2009). "Partial Orders of Similarity Differences Invariant Between EEG-Recorded Brain and Perceptual Representations of Language". Neural Computation 21 (11): 3228–69. doi:10.1162/neco.2009.04-08-764. PMID 19686069. 
  28. Suppes, Patrick; Han, Bing; Epelboim, Julie; Lu, Zhong-Lin (1999). "Invariance of brain-wave representations of simple visual images and their names". Proceedings of the National Academy of Sciences of the United States of America 96 (25): 14658–63. doi:10.1073/pnas.96.25.14658. PMID 10588761. 
  29. Pasley, BN et al. (2012). "Reconstructing Speech from Human Auditory Cortex". PLOS Biol 10 (1): e1001251. doi:10.1371/journal.pbio.1001251. PMID 22303281. 
  30. Palmer, Jason (31 January 2012). "Science decodes 'internal voices'". BBC News. https://www.bbc.co.uk/news/science-environment-16811042. 
  31. 31.0 31.1 "Secrets of the inner voice unlocked". NHS. 1 Feb 2012. Archived from the original on 2012-02-03. https://web.archive.org/web/20120203022143/http://www.nhs.uk/news/2012/02February/Pages/mind-reading-telephathy-inner-voice.aspx. 
  32. Soon, C.; Brass, M.; Heinze, H.; Haynes, J. (2008). "Unconscious determinants of free decisions in the human brain". Nature Neuroscience 11 (5): 543–545. doi:10.1038/nn.2112. PMID 18408715. 
  33. Kühn, S.; Brass, M. (2009). "Retrospective construction of the judgment of free choice". Consciousness and Cognition 18 (1): 12–21. doi:10.1016/j.concog.2008.09.007. PMID 18952468. 
  34. Matsuhashi, M.; Hallett, M. (2008). "The timing of the conscious intention to move". European Journal of Neuroscience 28 (11): 2344–2351. doi:10.1111/j.1460-9568.2008.06525.x. PMID 19046374. 
  35. Kok, Peter; Jehee, Janneke; de Lange, Floris (2012). "Less Is More: Expectation Sharpens Representations in the Primary Visual Cortex". Neuron 75 (2): 265–270. doi:10.1016/j.neuron.2012.04.034. ISSN 0896-6273. PMID 22841311. 
  36. Yon, Daniel; Gilbert, Sam J.; de Lange, Floris P.; Press, Clare (2018). "Action sharpens sensory representations of expected outcomes" (in en). Nature Communications 9 (1): 4288. doi:10.1038/s41467-018-06752-7. ISSN 2041-1723. PMID 30327503. 
  37. Chu, Carlton; Ni, Yizhao; Tan, Geoffrey; Saunders, Craig J.; Ashburner, John (2010). "Kernel regression for fMRI pattern prediction". NeuroImage 56 (2): 662–673. doi:10.1016/j.neuroimage.2010.03.058. PMID 20348000. 
  38. "New Research: Computers That Can Identify You by Your Thoughts". UC Berkeley. April 3, 2013. http://www.ischool.berkeley.edu/newsandevents/news/20130403brainwaveauthentication. 
  39. Kamitani, Yukiyasu; Tong, Frank (2005). "Decoding the visual and subjective contents of the human brain". Nature Neuroscience 8 (5): 679–85. doi:10.1038/nn1444. PMID 15852014. 
  40. Miyawaki, Y; Uchida, H; Yamashita, O; Sato, M; Morito, Y; Tanabe, H; Sadato, N; Kamitani, Y (2008). "Visual Image Reconstruction from Human Brain Activity using a Combination of Multiscale Local Image Decoders". Neuron 60 (5): 915–29. doi:10.1016/j.neuron.2008.11.004. PMID 19081384. 
  41. Thirion, Bertrand; Duchesnay, Edouard; Hubbard, Edward; Dubois, Jessica; Poline, Jean-Baptiste; Lebihan, Denis; Dehaene, Stanislas (2006). "Inverse retinotopy: Inferring the visual content of images from brain activation patterns". NeuroImage 33 (4): 1104–16. doi:10.1016/j.neuroimage.2006.06.062. PMID 17029988. 
  42. Farwell, Lawrence A.; Richardson, Drew C.; Richardson, Graham M. (2012-12-05). "Brain fingerprinting field studies comparing P300-MERMER and P300 brainwave responses in the detection of concealed information". Cognitive Neurodynamics (Springer) 7 (4): 263–299. doi:10.1007/s11571-012-9230-0. ISSN 1871-4080. PMID 23869200. 
  43. Kay, Kendrick N.; Naselaris, Thomas; Prenger, Ryan J.; Gallant, Jack L. (2008). "Identifying natural images from human brain activity". Nature 452 (7185): 352–5. doi:10.1038/nature06713. PMID 18322462. Bibcode2008Natur.452..352K. 
  44. Allen, Ronald J.; Mace, M. Kristin (Winter 2004). "The Self-Incrimination Clause Explained and Its Future Predicted". The Journal of Criminal Law and Criminology (Elsevier) 94 (2). doi:10.2139/ssrn.480143. https://scholarlycommons.law.northwestern.edu/jclc/vol94/iss2/1. 
  45. Brennan-Marquez, Kiel (2012). "A modest defense of mind reading". Yale Journal of Law & Technology. Yale University. p. 214. https://yjolt.org/modest-defense-mind-reading. "Ronald Allen and Kristen Mace discern 'universal agreement' that the (Mind Reader Machine) is unacceptable." 
  46. 46.0 46.1 46.2 46.3 "How Technology May Soon "Read" Your Mind". December 31, 2008. http://www.cbsnews.com/news/how-technology-may-soon-read-your-mind/. 
  47. Stix, Gary (August 1, 2008). "Can fMRI Really Tell If You're Lying?". http://www.scientificamerican.com/article/new-lie-detector/. 
  48. Smith, Kerri (24 October 2013). "Brain decoding: Reading minds" (in en). Nature News 502 (7472): 428–430. doi:10.1038/502428a. PMID 24153277. 
  49. Drew, Liam (24 July 2019). "The ethics of brain–computer interfaces" (in en). Nature 571 (7766): S19–S21. doi:10.1038/d41586-019-02214-2. PMID 31341310. 
  50. 50.0 50.1 Saenz, Aaron (17 March 2010). "fMRI Reads the Images in Your Brain – We Know What You're Looking At". Singularity University. http://singularityhub.com/2010/03/17/fmri-reads-the-images-in-your-brain-we-know-what-youre-looking-at-video/. 
  51. Davis, Nicola (30 March 2020). "Scientists develop AI that can turn brain activity into text" (in en). The Guardian. https://www.theguardian.com/science/2020/mar/30/scientists-develop-ai-that-can-turn-brain-activity-into-text. 
  52. 52.0 52.1 Cuthbertson, Anthony (2014-08-29). "Mind Reader: Meet The Man Who Records and Stores Your Thoughts, Dreams and Memories". http://www.ibtimes.co.uk/mind-reader-meet-man-who-records-stores-your-thoughts-dreams-memories-1463094. 

External links