John K. Kruschke
{{Short description|American psychologist}}
{{Infobox scientist
| honorific_prefix =
| name = John K. Kruschke
| honorific_suffix =
| native_name =
| native_name_lang =
| image =
| image_size =
| image_upright =
| alt =
| caption =
| birth_name =
| birth_date =
| birth_place =
| death_date =
| death_place =
| death_cause =
| resting_place =
| resting_place_coordinates =
| other_names =
| siglum =
| pronounce =
| citizenship =
| nationality =
| fields = {{hlist|Psychology|Statistics}}
| workplaces = Indiana University Bloomington
| patrons =
| education =
| alma_mater = University of California at Berkeley
| thesis_title = A connectionist model of category learning
| thesis_url = https://search.library.berkeley.edu/permalink/01UCS_BER/1thfj9n/alma991077511199706532
| thesis_year = 1990
| doctoral_advisors = Stephen E. Palmer
Robert Nosofsky
| doctoral_students =
| known_for = {{hlist|connectionism|Bayesian analysis}}
| awards =
| website = {{URL|https://jkkweb.sitehost.iu.edu/}}
| footnotes =
}}
John Kendall Kruschke is an American psychologist and statistician known for his work in connectionist models of human learning,{{cite journal |last1=Kruschke |first1=John K. |title=ALCOVE: An exemplar-based connectionist model of category learning |journal=Psychological Review |date=1992 |volume=99 |issue=1 |pages=22–44 |doi=10.1037/0033-295X.99.1.22|pmid=1546117 }}
and in Bayesian statistical analysis.{{cite book |last1=Kruschke |first1=John K. |title=Doing Bayesian Data Analysis: A tutorial with R, JAGS, and Stan |date=2015 |publisher=Academic Press |isbn=9780124058880 |edition=2nd}}
He is Provost Professor Emeritus
in the [https://psych.indiana.edu/ Department of Psychological and Brain Sciences] at Indiana University Bloomington.
He won the Troland Research Award from the National Academy of Sciences in 2002.{{cite web |title=Troland Research Awards |url=http://www.nasonline.org/programs/awards/troland-research-awards.html |website=National Academy of Sciences |access-date=22 January 2022}}
Research
= Bayesian statistical analysis =
== Dissemination ==
Kruschke's popular textbook, Doing Bayesian Data Analysis,
was notable for its accessibility and unique scaffolding of concepts. The first half of the book used the simplest type of data (i.e., dichotomous values) for presenting all the fundamental concepts of Bayesian analysis, including generalized Bayesian power analysis and sample-size planning. The second half of the book used the generalized linear model as a framework for explaining applications to a spectrum of other types of data.
Kruschke has written many tutorial articles about Bayesian data analysis, including an open-access article that explains Bayesian and frequentist concepts side-by-side.{{cite journal | last1 = Kruschke | first1 = John K. | last2 = Liddell | first2 = Torrin M. | date = 2018 | title = The Bayesian new statistics: hypothesis testing, estimation, meta-analysis, and power analysis from a Bayesian perspective | journal = Psychonomic Bulletin & Review | volume = 25 | number = 1 | pages = 178–206 | doi = 10.3758/s13423-016-1221-4 | pmid = 28176294 | s2cid = 4523799 | doi-access = free }}
There is an accompanying [https://jkkweb.sitehost.iu.edu/KruschkeFreqAndBayesAppTutorial.html online app] that interactively does frequentist and Bayesian analyses simultaneously.
Kruschke gave a video-recorded plenary talk on this topic at the [https://www.causeweb.org/cause/uscots/uscots19/keynote/3 United States Conference on Teaching Statistics (USCOTS)].
== Bayesian analysis reporting guidelines ==
Bayesian data analyses are increasing in popularity but are still relatively novel in many fields, and guidelines for reporting Bayesian analyses are useful for researchers, reviewers, and students. Kruschke's open-access Bayesian analysis reporting guidelines (BARG)
provide a step-by-step list with explanation. For instance, the BARG recommend that if the analyst uses Bayesian hypothesis testing, then the report should include not only the Bayes factor but also the minimum prior model probability for the posterior model probability to exceed a decision criterion.
== Assessing null values of parameters ==
Kruschke proposed a decision procedure for assessing null values of parameters, based on the uncertainty of the posterior estimate of the parameter.{{cite journal | last1 = Kruschke | first1 = John K. | date = 2018 | title = Rejecting or Accepting Parameter Values in Bayesian Estimation | journal = Advances in Methods and Practices in Psychological Science | volume = 1 | number = 2 | pages = 270–280 | doi = 10.1177/2515245918771304 | s2cid = 125788648 | url = https://jkkweb.sitehost.iu.edu/articles/Kruschke2018RejectingOrAcceptingParameterValuesWithSupplement.pdf }}
This approach contrasts with Bayesian hypothesis testing as model comparison
== Ordinal data ==
Liddell and Kruschke
showed that the common practice of treating ordinal data (such as subjective ratings) as if they were metric values can systematically lead to errors of interpretation, even inversions of means. The problems were addressed by treating ordinal data with ordinal models, in particular an ordered-probit model. Frequentist techniques can also use ordered-probit models, but the authors favored Bayesian techniques for their robustness.
= Models of learning =
An overview of Kruschke's models of attentional learning through 2010 is provided in reference.{{cite book | chapter = Models of attentional learning | author-last1 = Kruschke | author-first1 = John K. | title = Formal approaches to categorization | publisher = Cambridge University Press | year = 2011 | pages = 120–152 | editor-last1 = Pothos | editor-first1 = E. M. | editor-last2 = Wills | editor-first2 = A. J. | isbn = 9781139493970 | url = https://jkkweb.sitehost.iu.edu/articles/Kruschke2011PWed.pdf}}
That reference summarizes numerous findings from human learning that suggest attentional learning. That reference also summarizes a series of Kruschke's models of learning under a general framework.
== Dimensionality in backpropagation networks ==
Back-propagation networks are a type of connectionist model, at the core of deep-learning neural networks. Kruschke's early work with back-propagation networks created algorithms for expanding or contracting the dimensionality of hidden layers in the network, thereby affecting how the network generalized from training cases to testing cases
The algorithms also improved the speed of learning.{{cite journal | last1 = Kruschke | first1 = John K. | last2 = J. R. Movellan | first2 = J. R. Movellan | date = 1991 | title = Benefits of Gain: Speeded learning and minimal hidden layers in back-propagation networks | journal = IEEE Transactions on Systems, Man, and Cybernetics | volume = 21 | pages = 273–280 | doi = 10.1109/21.101159 | url = https://jkkweb.sitehost.iu.edu/articles/KruschkeMovellan1991BenefitsOfGain.pdf }}
== Exemplar-based models and learned attention ==
The ALCOVE model of associative learning
used gradient descent on error, as in back-propagation networks, to learn what stimulus dimensions to attend to or to ignore. The ALCOVE model was derived from the generalized context model
of [https://scholar.google.com/citations?user=559JfocAAAAJ&hl=en&oi=sra R. M. Nosofsky]. These models mathematically represent stimuli in a multi-dimensional space based on human perceived dimensions (such as color, size, etc.), and assume that training examples are stored in memory as complete exemplars (that is, as combinations of values on the dimensions). The ALCOVE model is trained with input-output pairs and gradually associates exemplars with trained outputs while simultaneously shifting attention toward relevant dimensions and away from irrelevant dimensions.
An enhancement of the ALCOVE model, called RASHNL, provided a mathematically coherent mechanism for gradient descent with limited-capacity attention.{{cite journal | last1 = Kruschke | first1 = John K. | last2 = Johansen | first2 = M. K. | date = 1999 | title = A model of probabilistic category learning | journal = Journal of Experimental Psychology: Learning, Memory, and Cognition | volume = 25 | number = 5 | pages = 1083–1119 | doi = 10.1037/0278-7393.25.5.1083 | pmid = 10505339 }}
The RASHNL model assumed that attention is shifted rapidly when a stimulus is presented, while learning of attention across trials is more gradual.
These models were fitted to empirical data from numerous human learning experiments, and provided good accounts of relative difficulties of learning different types of associations, and of accuracies of individual stimuli during training and generalization. Those models can not explain all aspects of learning; for example, an additional mechanism was needed to account for the rapidity of human learning of reversal shift (i.e., what was "A" is now "B" and vice versa).{{cite journal | last1 = Kruschke | first1 = John K. | date = 1996 | title = Dimensional relevance shifts in category learning | journal = Connection Science | volume = 8 | issue = 2 | pages = 201–223 | doi = 10.1080/095400996116893 }}
== The highlighting effect ==
When people learn to categorize combinations of discrete features successively across a training session, people will tend to learn about the distinctive features of the later-learned items instead of learning about their complete combination of features. This attention to distinctive features of later-learned items is called "the highlighting effect", and is derived from an earlier finding known as "the inverse base-rate effect".{{cite journal | last1 = Medin | first1 = D. L. | last2 = Edelson | first2 = S. M. | date = 1988 | title = Problem structure and the use of base-rate information from experience | journal = Journal of Experimental Psychology: General | volume = 117 | number = 1 | pages = 68–85 | doi = 10.1037/0096-3445.117.1.68 | pmid = 2966231 }}
Kruschke conducted an extensive series of novel learning experiments with human participants, and developed two connectionist models to account for the findings. The ADIT model
learned to attend to distinctive features, and the EXIT model
used rapid shifts of attention on each trial.
A canonical highlighting experiment and a review of findings was presented in reference.{{cite book | last1 = Kruschke | first1 = John K. |chapter=Highlighting: A canonical experiment |editor-last1=Ross |editor-first1=Brian |title=The Psychology of Learning and Motivation, Volume 51 |year = 2009 | pages = 153–185 | volume = 51 | doi = 10.1016/S0079-7421(09)51005-5 | publisher = Academic Press | url = https://jkkweb.sitehost.iu.edu/articles/Kruschke2009PLM.pdf }}
== Hybrid representation models for rules or functions with exceptions ==
People can learn to classify stimuli according to rules such as "a container for liquids that is wider than it is tall is called a bowl", along with exceptions to the rule such as "unless it is this specific case that is called a mug". A series of experiments demonstrated that people tend to classify novel items, that are relatively close to an exceptional case, according to the rule more than would be predicted by exemplar-based models. To account for the data, Erickson and Kruschke developed hybrid models that shifted attention between rule-based representation and exemplar-based representation.{{cite journal | last1 = Erickson | first1 = M. A. | last2 = Kruschke | first2 = John K. | date = 1998 | title = Rules and exemplars in category learning | journal = Journal of Experimental Psychology: General | volume = 127 | number = 2 | pages = 107–140 | doi = 10.1037/0096-3445.127.2.107 | pmid = 9622910 }}{{cite journal | last1 = Erickson | first1 = M. A. | last2 = Kruschke | first2 = John K. | date = 2002 | title = Rule-based extrapolation in perceptual categorization | journal = Psychonomic Bulletin & Review | volume = 9 | number = 1 | pages = 160–168 | doi = 10.3758/BF03196273 | pmid = 12026949 | s2cid = 2388327 | doi-access = free }}{{cite journal | last1 = Denton | first1 = S. E. | last2 = Kruschke | first2 = John K. | last3 = Erickson | first3 = M. A. | date = 2008 | title = Rule-based extrapolation: A continuing challenge for exemplar models | journal = Psychonomic Bulletin & Review | volume = 15 | number = 4 | pages = 780–786 | doi = 10.3758/PBR.15.4.780 | pmid = 18792504 | s2cid = 559864 | doi-access = free }}
People can also learn continuous relationships between variables, called functions, such as "a page's height is about 1.5 times its width". When people are trained with examples of functions that have exceptional cases, the data are accounted for by hybrid models that combine locally applicable functional rules.{{cite journal | last1 = Kalish | first1 = M. L. | last2 = Lewandowsky | first2 = S. | date = 2004 | title = Population of Linear Experts: Knowledge Partitioning and Function Learning | journal = Psychological Review | volume = 111 | number = 4 | pages = 1072–1099 | doi = 10.1037/0033-295X.111.4.1072 | pmid = 15482074 }}
== Bayesian models of learning ==
Kruschke also explored Bayesian models of human-learning results that were addressed by his connectionist models. The effects of sequential or successive learning (such as highlighting, mentioned above) can be especially challenging for Bayesian models, which typically assume order-independence. Instead of assuming that the entire learning system is globally Bayesian, Kruschke developed models in which layers of the system are locally Bayesian.{{cite journal | last1 = Kruschke | first1 = John K. | date = 2006 | title = Locally Bayesian Learning with Applications to Retrospective Revaluation and Highlighting | journal = Psychological Review | volume = 113 | number = 4 | pages = 677–699 | doi = 10.1037/0033-295X.113.4.677 | pmid = 17014300 }}
This "locally Bayesian learning" accounted for combinations of phenomena that are difficult for non-Bayesian learning models or for globally-Bayesian learning models.
Another advantage of Bayesian representations is that they inherently represent uncertainty of parameter values, unlike typical connectionist models that save only a single value for each parameter. The representation of uncertainty can be used to guide active learning in which the learner decides which cases would be most useful to learn about next.{{cite journal | last1 = Kruschke | first1 = John K. | date = 2008 | title = Bayesian approaches to associative learning: From passive to active learning | journal = Learning & Behavior | volume = 36 | number = 3 | pages = 210–226 | doi = 10.3758/LB.36.3.210 | pmid = 18683466 | s2cid = 16668044 | doi-access = free }}
Career
Kruschke joined the faculty of the [https://psych.indiana.edu/ Department of Psychological and Brain Sciences] at Indiana University Bloomington as a lecturer in 1989. He remained at IU until he retired as Provost Professor Emeritus in 2022.
= Education =
Kruschke attained a B.A. in mathematics, with High Distinction in General Scholarship, from the [https://www.berkeley.edu/ University of California at Berkeley] in 1983. In 1990, he received a Ph.D. in Psychology also from U. C. Berkeley.
Kruschke attended the 1978 [https://summerscience.org/ Summer Science Program] at The Thacher School in Ojai CA, which focused on astrophysics and celestial mechanics. He attended the 1988 Connectionist Models Summer School
at Carnegie Mellon University.
= Awards =
- [https://www.pbk.org/ Phi Beta Kappa] (academic honor society), 1982.
- [https://www.nsfgrfp.org/ National Science Foundation Graduate Fellowship], 1983.
- [https://grants.nih.gov/grants/policy/r29.htm National Institute of Mental Health FIRST Award], 1994.
- Indiana University Trustees Teaching Excellence Recognition Awards: 1997, 1998, 1999, 2008, 2009, 2010, 2011, 2012.{{cite web | url=https://vpfaa.indiana.edu/faculty-resources/awards-lectures/awards/trustees-award.html |title=Office of the Vice Provost for Faculty and Academic Affairs: Trustees Teaching Award |author= }}
- Troland Research Award, National Academy of Sciences, 2002.
- Remak Distinguished Scholar Award, Indiana University, 2012.
- Provost Professor, Indiana University, 2018.
References
{{Reflist}}
External links
- {{Official website|https://jkkweb.sitehost.iu.edu/}}
- [https://psych.indiana.edu/directory/faculty/kruschke-john.html Faculty page]
- {{Google Scholar id|Im5IIiMAAAAJ}}
{{Authority control}}
{{DEFAULTSORT:Kruschke, John Kendall}}
Category:21st-century American psychologists
Category:American statisticians
Category:Indiana University faculty
Category:University of California, Berkeley alumni