Wednesday, January 15, 2014

Probabilistic Language Processing
Using logic to represent and reason we can represent knowledge about the world with
facts and rules, like the following ones:
bird(tweety).
fly(X) :- bird(X).

We can also use a theorem-prover to reason about the world and deduct new facts about
the world, for e.g.,
?- fly(tweety).
Yes

However, this often does not work outside of toy domains - non-tautologous certain
rules are hard to find.
A way to handle knowledge representation in real problems is to extend logic by using
certainty factors.
 In other words, replace
 IF condition THEN fact
with
IF condition with certainty x THEN fact with certainty f(x)
Unfortunately cannot really adapt logical inference to probabilistic inference, since the
latter is not context-free.
Replacing rules with conditional probabilities makes inferencing simpler.
Replace
smoking -> lung cancer
or
lotsofconditions, smoking -> lung cancer
with
P(lung cancer | smoking) = 0.6

Uncertainty is represented explicitly and quantitatively within probability theory, a
formalism that has been developed over centuries.
A probabilistic model describes the world in terms of a set S of possible states - the
sample space. We don’t know the true state of the world, so we (somehow) come up with
a probability distribution over S which gives the probability of any state being the true
one. The world usually described by a set of variables or attributes.
Consider the probabilistic model of a fictitious medical expert system. The ‘world’ is
described by 8 binary valued variables:

Review of Probability Theory
The primitives in probabilistic reasoning are random variables. Just like primitives in
Propositional Logic are propositions. A random variable is not in fact a variable, but a
function from a sample space S to another space, often the real numbers.
For example, let the random variable Sum (representing outcome of two die throws) be
defined thus: Sum(die1, die2) = die1 +die2

Each random variable has an associated probability distribution determined by the
underlying distribution on the sample space

Continuing our example : P(Sum = 2) = 1/36,
P(Sum = 3) = 2/36, . . . , P(Sum = 12) = 1/36
Conditional probability is defined as:
It means for any value x of A and any value y of B


No comments:

Post a Comment