# Contents

# Context

#### Learning | Structural Priors Learning

The **Structural Priors Learning** is a Meta-Learning algorithm that can be compared to **Bagging**. It is based on **Data Perturbation** (Smoothed Bootstrapping) for learning a bag of networks and gathering statistics about the relationships that have been found. These statistics are then used for automatically defining **Structural Priors**.

Once defined, the Structural Priors are used for learning a network on the original unperturbed data set.

**Example**

Let's use a small data set that contains only 91 particles. Each particle is a physically active individual (several hours of exercise a week), described with body girth measurements and skeletal diameter measurements, as well as age, weight, height and gender.

All the variables are continuous, except Gender. We discretized all continuous variables with **R2GenOpt* 3**.

We define Temporal Indices to specify that * Age* and

*have to be root nodes.*

**Gender**Below is the network learned with EQ.

Let's use **Data Perturbation** in order to try to escape from the local optima found with EQ.

The obtained network has a better score that the one obtained with EQ. It also makes more sense as we can see that it captures the relationship between * Weight, Height *and

*.*

**BMI**However, we still have 4 "orphan" nodes: * Ankle diameter, Ankle girth, Age *and

*.*

**Knee diameter**There are 2 scenarios:

- either these nodes are really marginally independent of all the other nodes, or
- their relationships with the other nodes is above the "significance threshold" implicitly defined by the MDL score.

Keeping in mind that we just have 91 particles, and given that the MDL score is conservative, we can try decreasing the Structural Coefficient.

Instead of manually testing the structural coefficient, we use the Structural Coefficient Analysis tool.

The analysis of these curves suggest the utilization of a Structural Coefficient = 0.6

The network below has been learned with Data Perturbation - EQ.

As we can see, there is now only 1 orphan: * Ankle diameter*. Reducing the value of the Structural Coefficient was indeed efficient for connecting the orphans. However, this coefficient has a global impact on the MDL score. It reduces the cost of adding arcs for all the nodes. We can see for example that

*(highlighted in red) has now 3 parents, which is probably too much given the amount of data available. There are also now 5 nodes with 2 parents (highlighted in pink), instead of one with the Structural Coefficient = 1.0.*

**Biacromial diameter**Instead of reducing the Structural Coefficient globally, we now use **Structural Priors Learning** for exploring a range of coefficients [0.2 ; 1.0] and automatically get Structural Priors.

In addition to the MDL score, the **Console** returns the **Min**, **Max** and **Mean** of the arcs' Priors that are represented in the learned network.

As we can see, there is also only 1 orphan. However, the complexity of the obtained network is lower than the one learned with a Structural Coefficient = 0.6. The * Biacromial diameter* has only one parent, which makes much more sense given the amount of data available, and there are only 3 nodes with 2 parents (highlighted in pink).

Clicking in the lower right corner of the graph window opens the editor of the Structural Priors:

The **Preview** returns a graph with all the connections that have been learned on the perturbed data sets.

**Store Priors on Arcs** associates the priors with the arcs of the network. It also a way to get in the **Console** the **Min**, **Max** and **Mean** of the arcs' Priors that are represented in the current network.