Bayesian Causal Networks for Probabilistic Logic

chevron-icon
RFP Proposals
Top
chevron-icon
project-presentation-img
Expert Rating 3.1
evanluoicecream
Project Owner

Bayesian Causal Networks for Probabilistic Logic

Expert Rating

3.1

Overview

This research aims to improve Probabilistic Logic Networks (PLN) by integrating causal inference mechanisms to improve differentiation between causation and correlation. The project focuses on two primary goals: guiding PLN inference control using causal networks and exploring PLN’s capability to learn causal relationships. By employing causal learning systems, such as Bayesian networks and Pearl’s Do-Calculus, the research seeks to validate enhanced reasoning accuracy in domains like bioinformatics and AI planning. In sum, this work will refine decision-making processes in AI, leading to more robust and accurate predictions in complex environments.

RFP Guidelines

Causal learning guided PLN inference control

Complete & Awarded
  • Type SingularityNET RFP
  • Total RFP Funding $150,000 USD
  • Proposals 3
  • Awarded Projects n/a
author-img
SingularityNET
Oct. 4, 2024

Implementing causal, versus strictly correlative, inference can lead to a deeper understanding of many complex problems. Probabilistic Logic Networks (PLN), with their robust mathematical foundations for handling uncertain and incomplete knowledge—such as induction, abduction, analogy, and reasoning about time—offer a strong framework for this task. This project has two main goals:

  1. To guide PLN inference control using causal networks.
  2. To explore how PLN rules can distinguish causal from correlational relationships.

Proposals can address either of these goals separately or explore both together and the feedback between them. Submissions should include testable examples to validate the approach and demonstrate how causal reasoning improves PLN’s overall performance.

Proposal Description

Company Name (if applicable)

Protractor LLC

Project details

At the core of our research is the construction of causal networks, which provide a structured representation of the dependencies between variables. We will use Bayesian networks, structural causal models (SCMs), and Pearl’s Do-Calculus to formalize these relationships. The Bayesian network serves as the probabilistic framework for encoding causal relationships, where nodes represent random variables, and directed edges signify causal influences. This representation lends itself to efficient inference on the probability of various outcomes based on observed data. By inferring the causal structure from observational data, we will establish a causal graph that delineates how changes in one variable can effectuate changes in others. 

The integration of these causal networks with PLN will entail developing an inference control mechanism that uses the causal structure to prioritize pathways during reasoning tasks. We will design algorithms that operate on causal graphs to dynamically adjust the inference process based on causal information. This will include formulating causal intervention strategies that simulate interventions in the causal model to predict the resultant effects, thereby refining the PLN’s decision-making processes. The key here is to enable PLN to use causal information not merely as additional context but as a primary guiding principle for inference, improving the robustness of the conclusions drawn. 

Guiding PLN Inference

Integrating causal networks into PLNs improves inference control and provides a robust framework for reasoning under uncertainty. This section outlines a methodology for using causal structures to optimize PLN’s inference processes, focusing on developing algorithms that use causal information to guide decision-making in AI services, particularly in dynamic environments like bioinformatics and AI planning. 

A Bayesian network is structured as a directed acyclic graph where nodes represent random variables and directed edges indicate causal relationships. Each node is linked to a conditional probability distribution that describes how the variable behaves based on its parent nodes. Structural learning involves using algorithms like the PC or GES algorithms to identify the most probable causal graph from observational data. The process tests independent relationships between variables to construct the network structure. In parameter learning, techniques such as maximum likelihood estimation or Bayesian estimation are applied to derive the conditional probability distributions, which is important for accurately modeling real-world phenomena. 

Structural causal models (SCMs) extend Bayesian networks by incorporating potential outcomes for each variable. Each variable can be expressed in terms of its causal parents and an error term, which supports causal interventions. By simulating interventions, the causal graph can be modified to evaluate the effects of changes in variables, which permits robust predictions regarding how alterations influence outcomes. 

To guide PLN inference using causal networks, we propose a set of algorithms that prioritize causal pathways during reasoning tasks:

  1. Causal Pathway Prioritization: This mechanism evaluates the strength of causal relationships in the network, determining the relevance of each path to the inference task. 

  2. Dynamics Inference Adjustment: This algorithm adaptively modifies inference strategies based on new evidence. For instance, if evidence is provided for a variable, the inference pathways are recalibrated to emphasize nodes directly affected by that variable. 

  3. Causal Intervention Simulation: The system can simulate hypothetical interventions within the PLN framework, predicting the effects of changes and informing decision-making processes. 

Causality

Predictive implication—a framework that supports discussions of temporal correlation—is useful for addressing pragmatic aspects of causation.  To illustrate, consider the classic example: a rooster crowing before dawn. Even though the rooster crows consistently before the sun rises, it would be an error to conclude causality in this sequence. The distinction lies in understanding that a third variable could be influencing both events, or, as in this case, other contextual knowledge dismisses the idea that the rooster causes the sun to rise. To avoid such erroneous conclusions, a reasoning system must be able to recognize and test for alternative explanations. In this case, both “rooster crows” and “sun rises” may hold a strong temporal and probabilistic link. 

This differentiation is formalized through these predictive implication relationships. Here, extensional relationships arise from direct observation (i.e. the observation that roosters crow before sunrise) whereas intensional relationships draw on the background knowledge that provides context to these observations. For instance, while extensional evidence suggests that “rooster crows imply sun rises” with high confidence, the intensional knowledge that roosters lack the physical capability to influence sunrise reduces the confidence in this causal implication. Thus, a reasoning system could assign low overall confidence to a causal inference between these events.

As shown in Goertzel et. al, one could model a network as:

  • PredictiveImplication <0.00, 0.99> between “small physical force” and “movement of a large object”

  • PredictiveImplication <0.99, 0.99> between “rooster crows” and “small physical force”

  • PredictiveImplication <0.99, 0.99> between “sun rises” and “movement of a large object”

  • PredictiveImplication <0.00, 0.99> between “rooster crows” and “sun rises

In clearer terms, here’s what the evidence suggests:

  • There’s virtually no chance (0.00) that a small force, like a rooster’s crow, could move something as massive as the sun, though we’re very confident (0.99) in this basic observation.

  • We can be highly confident (0.99) that a rooster crowing involves only a small physical force, aligning well with our understanding of its abilities.

  • We’re similarly confident (0.99) that the sunrise involves the movement of a large object, consistent with our knowledge of planetary motions.

  • So, while we often observe roosters crowing before sunrise, there’s almost no likelihood (0.00) that the crowing causes the sunrise. Instead, our background knowledge keeps us from mistaking this correlation for actual causation.

AI Service

Our research culminates in the Causal Inference Engine (CIE), which is an AI service designed to improve decision-making processes by integrating causal mechanisms into PLNs. This service uses Bayesian causal networks to differentiate between causation and correlation, providing robust predictive capabilities in complex environments. 

Central to the CIE is the construction of causal networks represented as directed acyclic graphs (DAGs), where nodes denote random variables and directed edges encapsulate causal relationships. This framework supports the encoding of intricate dependencies among variables, allowing for efficient probabilistic inference. The engine employs a structural learning algorithm to derive the most probable causal graph from observational data. 

The integration of Structural Causal Models (SCMs) enriches the CIE’s capabilities by enabling comprehensive analysis of potential outcomes. In this framework, each variable is articulated in terms of its causal parents alongside a stochastic error term, allowing for the evaluation of counterfactual scenarios. 

To optimize inference control, the causal pathway prioritization algorithm assesses the strength and relevance of causal relationships. Additionally, the dynamics inference adjustment module adaptively recalibrates inference strategies in real-time based on incoming evidence. Moreover, the CIE supports causal intervention simulations, allowing users to conduct hypothetical interventions within the causal framework to evaluate potential outcomes and inform strategic decisions.

Open Source Licensing

MIT - Massachusetts Institute of Technology License

Proposal Video

Not Avaliable Yet

Check back later during the Feedback & Selection period for the RFP that is proposal is applied to.

  • Total Milestones

    4

  • Total Budget

    $45,000 USD

  • Last Updated

    2 Nov 2024

Milestone 1 - Literature Review & Problem Specification

Description

Conduct literature review on Bayesian networks, Pearl’s Do-Calculus, and Structural Causal Models (SCMs). Define project objectives and specify goals for the Causal Inference Engine (CIE), including functional requirements and targeted applications (e.g., bioinformatics, AI planning). Develop initial theoretical framework for integrating causal networks with PLN for inference control.

Deliverables

A report for the literature review containing the produced work.

Budget

$5,000 USD

Milestone 2 - Data Collection & Preprocessing

Description

Identify datasets relevant to bioinformatics and AI planning domains. Establish data pipelines for data preprocessing and feature engineering, focusing on variables suited to causal inference. Construct preliminary Bayesian network structures based on domain knowledge.

Deliverables

A sample of the dataset.

Budget

$10,000 USD

Milestone 3 - Bayesian Network Development

Description

Preliminary Bayesian network models structured from collected datasets, including initial causal relationships. Algorithm documentation for structural learning techniques used (e.g., PC, GES). Parameter estimation report with results from maximum likelihood or Bayesian estimation.

Deliverables

Example code

Budget

$20,000 USD

Milestone 4 - Pilot Testing & Evaluation of Causal Models

Description

Evaluation report documenting findings from pilot testing of Bayesian networks, including validation results and feedback from experts. Refined Bayesian network models based on pilot test outcomes. Project summary report covering foundational development and initial network validations.

Deliverables

The evaluation report.

Budget

$10,000 USD

Join the Discussion (0)

Expert Ratings

Reviews & Ratings

Group Expert Rating (Final)

Overall

3.1

  • Compliance with RFP requirements 3.7
  • Solution details and team expertise 3.3
  • Value for money 3.0
  • Expert Review 1

    Overall

    3.0

    • Compliance with RFP requirements 3.0
    • Solution details and team expertise 3.0
    • Value for money 0.0
    Good fitting plan and project details, overall promising

    This proposal seems to be the promising in adding causality-related ideas to PLN and in a way that inference control can be guided by taking strength of causal relations into account. In addition to Bayesian Networks, the author has an understanding of PLN and also a deeper understanding of causality and the role background knowledge and intensionality plays in causal reasoning.

  • Expert Review 2

    Overall

    1.0

    • Compliance with RFP requirements 1.0
    • Solution details and team expertise 1.0
    • Value for money 0.0
    Naive proposal with no detailed explanations and does not discusses the integration with PLN

    Not Recommend! The proposal is naive and on the surface. The researcher falsely thinks that introducing a causal Bayesian network will necessarily improve efficiency of inference control, while in reality it will at least increase a complexity of already complex inference engine and actual casual inference still remains questionable. I disregard the proposal because of these main points: 1. Lack of details. It only provides very generic overview of Bayesian networks and that it should help to guide PLN to capture causal relations but never mentions how. It does not provide specific details and only provides some assertions which already looks very false for me. 2. Lack of PLN integration. The RFP specifically asks for two goals either (1) integrate with current reasoning control of PLN or (2) tweak PLN rules or control procedures to capture causal relations. However, in the proposal nor in the Milestones PLN integration is not even mentioned, while the researcher allocates time and budget for literature review for Bayesian networks and related algorithms. PLN integration is the major thing here and the person working with it must know the specifics of the rules and control of PLN. 3. Questionable skills. I question the experience and skills of the researcher since there is no paper reference or any information, I could find about the person. 4. Lack of domain/dataset. The proposal does not provide specific domain or dataset which will serve for benchmarking or testing the system which is crucial since causality/correlation vastly depend on the data chosen.

  • Expert Review 3

    Overall

    5.0

    • Compliance with RFP requirements 5.0
    • Solution details and team expertise 4.0
    • Value for money 0.0
    In favor

    Good alignment with RFP goals. Uses Bayesian networks and Pearl's Do-Calculus to guide PLN inference and distinguish causation vs. correlation. Promises dynamic inference control and intervention simulations. Little evidence of expertise with PLN or Hyperon but the approach described seems credible. Good value. In favor of funding.

  • Expert Review 4

    Overall

    4.0

    • Compliance with RFP requirements 5.0
    • Solution details and team expertise 4.0
    • Value for money 0.0
    Good proposal, I like the suggestion of integrating the do-calculus.

    The proposal attempts to address the two sides of that RFP (1. causal vs correlative relationship, 2. guiding PLN) in a somewhat unified way. I don't know how well it is going to unfold but certainly trying to integrate BN (and do-calculus in particular) to PLN and inference control should be interesting. Please be aware that PLN might not have the level of maturity that would be expected for that task by the time the work begins, we'll see. But hopefully one should be able to do some progress by using an incomplete and imperfect version of PLN.

  • Expert Review 5

    Overall

    3.0

    • Compliance with RFP requirements 3.0
    • Solution details and team expertise 3.0
    • Value for money 0.0
    This one is a little odd -- the Overview is about causal guidance of PLN, but the milestones are purely about development of BNs

    I was excited about this one from reading the overview, but then I noted the milestones omit everything about PLN guidance and just talk about BN development, which is not really the point of the RFP

  • Expert Review 6

    Overall

    3.0

    • Compliance with RFP requirements 5.0
    • Solution details and team expertise 3.0
    • Value for money 0.0

    Bayesian Causal Networks and SCMs could certainly inform PLN. Well thought out with enough detail with emphases on bioinformatics and AI planning. Not enough team information.

feedback_icon