1. Home
  2. Docs
  3. Artificial Intelligence
  4. Knowledge Representation
  5. Handling Uncertain Knowledge

Handling Uncertain Knowledge

In artificial intelligence, decision-making often involves uncertain knowledge, where information is incomplete, noisy, or probabilistic. To handle uncertainty, AI systems use probability theory, Bayesian reasoning, and graphical models like Bayesian networks to infer and update beliefs based on new evidence.

A random variable is a variable that can take different values, each with an associated probability. These variables help model uncertainty in real-world problems.

    Types of Random Variables:

    • Discrete Random Variables: Take a finite number of values (e.g., rolling a die: {1, 2, 3, 4, 5, 6}).
    • Continuous Random Variables: Take infinite values within a range (e.g., temperature measurements: {0°C to 100°C}).

    Example:
    Let X be a random variable representing the weather.

    • X=”Sunny” → 0.7 probability
    • X=”Rainy” → 0.3 probability

    These probabilities help AI systems predict outcomes in uncertain environments.

    Prior Probability (P(A)): The initial probability of an event before considering new evidence.
    Posterior Probability (P(A|B)): The updated probability of an event after considering new evidence.

    Example:

    • Suppose 10% of students have the flu (Prior Probability = 0.1).
    • A student has a fever (new evidence).
    • The probability that a student actually has the flu given they have a fever is the posterior probability, calculated using Bayes’ Theorem.

    A Full Joint Distribution lists the probabilities of all possible combinations of random variables in a system.

      Example:
      If we have two variables:

      • Weather W: {Sunny, Rainy}
      • Traffic T: {Heavy, Light}

      The full joint probability distribution:

      The full joint probability distribution

      Bayes’ Rule is a mathematical formula for updating probabilities when new evidence is observed. It relates prior and posterior probabilities as:

      Bayes’ Rule

      Explanation:

      • P(A|B): Probability of event A given B (Posterior)
      • P(B|A): Probability of B given A (Likelihood)
      • P(A): Probability of A (Prior)
      • P(B): Probability of B (Evidence)

      Example: (Medical Diagnosis)

      • A: Patient has disease D.
      • B: Patient tests positive.

      Given:

      • P(D)=0.01 (1% of people have disease)
      • P(Positive∣D)=0.9 (90% of sick patients test positive)
      • P(Positive)=0.05 (5% of all patients test positive)

      Compute Posterior Probability:

      Compute Posterior Probability

      Thus, the probability that a patient actually has the disease after a positive test is 18%, even though the test detects 90% of cases.

      A Bayesian Network is a graphical model that represents probabilistic relationships among a set of variables using nodes and directed edges.

        Components:

        • Nodes: Represent random variables.
        • Edges: Represent probabilistic dependencies.
        • Conditional Probability Tables (CPTs): Define relationships between connected nodes.

        Example: Medical Diagnosis Bayesian Network

               Cold  -----> Fever
                 |           |
                 ↓           ↓
             Cough       Test Result
        • Fever depends on Cold.
        • Test Result depends on Fever.
        • Cough depends on Cold.

        Using Bayesian Networks, we can infer the probability of a cold given a cough without computing the full joint distribution.

        Belief networks (Bayesian networks) are used to infer probabilities and make decisions under uncertainty.

          Types of Reasoning:

          • Diagnostic Reasoning (Effect → Cause): Given symptoms, infer the disease.
          • Causal Reasoning (Cause → Effect): Given a disease, predict symptoms.
          • Intercausal Reasoning: Adjust beliefs when new evidence is introduced.

          Example: Disease Diagnosis

          • Known: The patient has cough.
          • Infer: What is the probability they have a cold?
          • New Evidence: The patient also has fever → Updated Probability using Bayes’ Rule.

          How can we help?

          Leave a Reply

          Your email address will not be published. Required fields are marked *