TechRxiv
AGI23_PREPRINT_ARCHIVE_VERSION_Part_2___The_Optimal_Choice_of_Hypothesis_Is_the_Weakest__Not_the_Shortest(1).pdf (542.86 kB)
Download file

The Optimal Choice of Hypothesis Is the Weakest, Not the Shortest

Download (542.86 kB)
preprint
posted on 2023-04-28, 20:04 authored by Michael Timothy BennettMichael Timothy Bennett

Accepted for poster presentation at the 16th Conference on Artificial General Intelligence, taking place in Stockholm, 2023.


If A and B are sets such that A is a subset of B, generalisation may be understood as the inference from A of a hypothesis sufficient to construct B. One might infer any number of hypotheses from A, yet only some of those may generalise to B. How can one know which are likely to generalise? One strategy is to choose the shortest, equating the ability to compress information with the ability to generalise (a ``proxy for intelligence”). We examine this in the context of a mathematical formalism of enactive cognition. We show that compression is neither necessary nor sufficient to maximise performance (measured in terms of the probability of a hypothesis generalising). We formulate a proxy unrelated to length or simplicity, called weakness. We show that if tasks are uniformly distributed, then there is no choice of proxy that performs at least as well as weakness maximisation in all tasks while performing strictly better in at least one. In experiments comparing maximum weakness and minimum description length in the context of binary arithmetic, the former generalised at between 1.1 and 5 times the rate of the latter. We argue this demonstrates that weakness is a far better proxy, and explains why Deepmind's Apperception Engine is able to generalise effectively.

History

Email Address of Submitting Author

michael.bennett@anu.edu.au

ORCID of Submitting Author

0000-0001-6895-8782

Submitting Author's Institution

The Australian National University

Submitting Author's Country

  • Australia