作者Morais, Michael J
ProQuest Information and Learning Co
Princeton University. Neuroscience
書名Approximate Bayesian Methods for Optimal Neural Coding and Decision-Making
出版項2021
說明1 online resource (139 pages)
文字text
無媒介computer
成冊online resource
附註Source: Dissertations Abstracts International, Volume: 82-12, Section: B
Advisor: Pillow, Jonathan W
Thesis (Ph.D.)--Princeton University, 2021
Includes bibliographical references
One fundamental goal of theoretical neuroscience is to understand the normative principles governing the functional organization of neural circuits, and, in turn, to what extent they can be considered optimal. Calling neural representations of information in the brain ̀̀optimal'' implies a multifarious equilibrium that balances robustness against flexibility, completeness against relevance, and so on, but it need only imply a solution to some optimization program. The exact forms of these programs varies with the modeling goals, neural circuits, tasks, or even animals under investigation. With this dissertation, we explore how we can define neural codes as optimal when they generate optimal behavior -- an easy principle to state, but a hard one to implement. Such a principle would bridge a gap between classical hypotheses of optimal neural coding, efficient coding and the Bayesian brain, with a common unified theory.In the first study, we analyzed neural population activity in V1 while monkeys performed a visual detection task, and found that a majority of the total choice-related variability is already present in V1 population activity. Such a prominent contribution of non-stimulus activity in classically sensory regions cannot be incorporated into existing models of neural coding, and demands models that can jointly optimize coding and decision-making within a single neural population.In the second study, we derived power-law efficient codes, a natural generalization of classical efficient codes, and show they are sufficient to replicate and explain a diverse set of psychophysical results. This broader family can maximize mutual information or minimize error of perceptual decisions, suggesting that psychophysical phenomena used to validate normative models could be more general features of perceptual systems than previously appreciated.In the third study, we translated the problem of joint model learning and decision-making into Bayesian machine learning, and extended a family of methods for decision-aware approximate inference to include a novel algorithm that we called loss-calibrated expectation propagation. How this problem can be solved by a non-biophysical system could be a constructive reference point for future studies into joint coding and decision-making, and the normative principles that drive decision-related variability in optimal sensory neural codes
Electronic reproduction. Ann Arbor, Mich. : ProQuest, 2021
Mode of access: World Wide Web
主題Neurosciences
Statistics
Logic
Approximate inference
Bayesian statistics
Decision-making
Efficient coding
Neural coding
Perception
Electronic books.
0317
0463
0395
ISBN/ISSN9798515256630
QRCode
相關連結: click for full text (PQDT) (網址狀態查詢中....)
館藏地 索書號 條碼 處理狀態  

Go to Top