a Tautology
Peter Nordin
The Answers is:
The asymmetry of similarity
1. Pony?
1. Pony?
2. Cow?
2. Cow?
3. Dog?
3. Dog?
…
…
345. Pegasus!
345. Pegasus!
345
!
Science is Prediction
The Model
The Turkey and the issue
with inductive predictions
(1)
The Turkey and the issue
with inductive predictions
(2)
Mandatory Reading
All Real Science is
Predictive Science
! No risk - no gain
Recent progress
Recent advances:
Universal Learning
Algorithms. There is a theoretically
optimal way of predicting the future,
given the past. It can be used to define
an optimal (though noncomputable)
rational agent that maximizes its
expected reward in almost arbitrary
environments sampled from
computable probability distributions.
Recent advances:
! …Is compression
Optimal Cognition
Optimal prediction
Exprerimental planning
Turingcompete repr.
Bayes etc
! Information-theoretic,
! Statistical, and
! Philosophical,
! Foundations of
! Artificial Intelligence
Universal AI
= =
+ +
!"
Pieces of the puzzle
! Induction →Prediction→Decision→Action
! Example 1:
! (d = past # days sun rose, 1 =sun rises. 0 = sun will not rise) •
p is undefined, because there has never been an experiment that
tested the existence of the sun tomorrow (ref. class problem).
! • Derive p from the type, age, size and temperature of the sun,
even though we never observed another star with those exact
properties. Conclusion: We predict that the sun will rise
tomorrow with high probability independent of the
justification.
Sequence
! Example 2:
! Example 3:
! Kolmogorov complexity
! Universal Distribution
! Inductive Learning
Principle of
Indifference
(Epicurus)
! P(H|D) = P(D|H)*P(H)/P(D)
! k(s)<=|s|
! P(s) = 2-k(s)
! "+1/3+….>1
Levin’s improvement
! Kraft’s inequality
! Let L1, L2,… be a sequence of natural
numbers. There is a prefix-code with this
sequence as lengths of its binary code words
iff !n2-ln<=1
Multiplicative
domination
! Is Bob honest?
! Side bet: flip coin 1000 times, record the
result as a string s
! Alice pay 1$, Bob pay Alice 21000-k(s) $
! Good offer:
! !|s|=1000 2-1000 21000-k(s)=! |s|=1000 2-k(s)<=1
! Then, this
multiplicatively
dominates all priors
! Suppose we want a ! though neutral priors
prior so neutral that will mean slow
it never rules out a learning
model
! m(x) are “universal”
! Possible, if limit to priors
computable models
! Mixture of all
(computable) priors,
with weights, "i, that
decline fairly fast:
The most neutral possible coding
language
! Invariance theorem
! any languages L1, L2, #c,
! $x |KL1(x)-KL2(x)| # c
! Not quite!
! c can be large