Machine Learning with BayesiaLab
BayesiaLab features a comprehensive array of highly optimized learning algorithms that can quickly uncover structures in datasets. The optimization criteria in BayesiaLab’s learning algorithms are based on information theory (e.g. the Minimum Description Length). With that, no assumptions regarding the variable distributions are made. These algorithms can be used for all kinds and all sizes of problem domains, sometimes including thousands of variables with millions of potentially relevant relationships.
Supervised Learning in BayesiaLab has the same objective as many traditional modeling methods, i.e. to develop a model for predicting a target variable. Note that numerous statistical packages also offer “Bayesian Networks” as a predictive modeling technique. However, in most cases, these packages are restricted in their capabilities to a one type of network, i.e. the Naive Bayes network. BayesiaLab offers a much greater number of Supervised Learning algorithms to search for the Bayesian network that best predicts the target variable while also taking into account the complexity of the resulting network.
We should highlight the Markov Blanket algorithm for its speed, which is particularly helpful when dealing with a large number of variables. In this context, the Markov Blanket algorithm can serve as an efficient variable selection algorithm. An example of Supervised Learning using this algorithm, and the closely-related Augmented Markov Blanket algorithm.
- Chapter 6 in Bayesian Networks & BayesiaLab
- Example 2 in Artificial Intelligence for Research, Analytics, and Reasoning