top banner top banner
index
RegularArticles
ReplicationStudies
SpecialIssues
Vignettes
EditorialBoard
Instructions4Authors
JournalGuidelines
Messages
Submission

Search publications

Assessing Conceptual Complexity and Compressibility Using Information Gain and Mutual Information

Full text PDF
Bibliographic information: BibTEX format RIS format XML format APA style
Cited references information: BibTEX format APA style
Doi: 10.20982/tqmp.06.1.p016

Mathy, Fabien
16-30
Keywords: Information theory , Information gain and mutual information
(no sample data)   (no appendix)

In this paper, a few basic notions stemming from information theory are presented with the intention of modeling the abstraction of relevant information in categorization tasks. In a categorization task, a single output variable is the basis for performing a dichotomic classification of objects that can be distinguished by a set of input variables which are more or less informative about the category to which the objects belong. At the beginning of the experiment, the target classification is unknown to learners who must select the most informative variables relative to the class in order to succeed in classifying the objects efficiently. I first show how the notion of entropy can be used to characterize basic psychological processes in learning. Then, I indicate how a learner might use information gain and mutual information –both based on entropy– to efficiently induce the shortest rule for categorizing a set of objects. Several basic classification tasks are studied in succession with the aim of showing that learning can improve as long as subjects are able to compress information. Referring to recent experimental results, I indicate in the Conclusion that these notions can account for both strategies and performance in subjects trying to simplify a learning process.


Pages © TQMP;
Website last modified: 2025-02-11.
Template last modified: 2022-03-04 18h27.
Page consulted on .
Be informed of the upcoming issues with RSS feed: RSS icon RSS