Concretely, processes which are apparently so different as induction
and deduction can be
explained in a computational framework as inference processes that
{\em just} generate an output
from an input, which must follow some semantical restrictions and/or
selection criteria, widely studied
in philosophy of science and mathematical logic, respectively. The
term information is seen as
the result of a computational effort, analogically to the way energy
is seen as the result of
a physical work. This suggests many questions, especially how to measure
this computational effort.
The {\em answer} was given by Levin in the seventies, proving that
the weighting
$LT(x) = length(x) + log Cost(x)$ between space and time was optimal
in the sense of universal
search problems. Given two objects, the effort from $x$ to $y$ in a
computational system $\phi$ is then measured as the
relative Levin-descriptional complexity $Kt(y | x) = min \{ LT(p) :
\phi(<p, x>) = y\}$.
The Information Gain of object $y$ wrt. object $x$ is then defined as
the quotient between the effort
which is necessary to describe $y$ from $x$ and the effort which is
necessary to describe $y$ alone.
More formally, $G(y | x) = Kt(y | x) / Kt(y)$. Some properties of this
measure are shown before applying it to inference processes.
In the case of induction, information gain represents how informative
is the hypothesis wrt.
the evidence (in Popper's sense) and it is compared with other
selection criteria, especially simplicity. This leads to the notion
of {\em authentic learning},
quite different from Gold's identification. In the case of deduction,
the measure also represents
how informative is the conclusion from the premises. This establishes
a generic measure of the gain
which is obtained from making explicit something that was implicit,
provided that the system is not
omniscient and resource limited, where there is a clear difference
between the explicit or
surface information, and implicit or depth information, as it was highlighted
by Hintikka
for first-order logic.
We study optimal compromises between the size of a theory and its explicitness,
formalising
the necessity of lemmata and the use of extensional properties for
mathematical practice,
in order to avoid difficult derivations that were already done (while
still maintaining under control
the whole size of the theory).
Keywords: Informativeness, Learning, Descriptional Complexity, Inference Processes, Deduction, Induction.
Go back to my home page .