next up previous
Next: Definition and Basic Properties Up: Dependency Networks for Inference, Previous: Probabilistic Inference


General Dependency Networks

As we have discussed, consistent dependency networks and Markov networks are interchangeable representations: given one, we can compute the other. In this section, we consider a more general class of dependency networks and part company with Markov networks. Our extension of consistent dependency networks is motivated by computational concerns. For domains having a large number of variables and many dependencies among those variables, it is computationally expensive to learn a Markov network from data and then convert that network to a consistent dependency network. As an alternative, we start with the observation that the local distribution for variable $X_i$ in a dependency network is the conditional distribution $p(x_i\vert{\bf x}\setminus x_i)$, which can be estimated by any number of probabilistic classification techniques (or regression techniques, if we were to consider continuous variables) such as methods using a probabilistic decision tree (e.g., Buntine, 1991), a generalized linear model (e.g., McCullagh and Nelder, 1989), a neural network (e.g., Bishop, 1995), a probabilistic support-vector machine (e.g., Platt, 1999), or an embedded regression/classification model (Heckerman and Meek, 1997). This observation suggests a simple, heuristic approach for learning the structure and probabilities of a dependency network from exchangeable (i.i.d.) data. Namely, for each variable $X_i$ in domain ${\bf X}$, we independently estimate its local distribution from data using a classification algorithm. Once we have all estimates for the local distributions, we then construct the structure of the dependency network from the (in)dependencies encoded in these estimates. For example, suppose we wish to construct a dependency network for the domain ${\bf X}=(X_1,X_2,X_3)$. To do so, we need to estimate three conditional probability distributions: $p(x_1\vert x_2,x_3)$, $p(x_2\vert x_1,x_3)$, and $p(x_3\vert x_1,x_2)$. First, we use prior knowledge to decide how each distribution is to be modeled. We may decide, for example, that a logistic regression, a multilayer neural net, and a probabilistic decision tree are appropriate models for $p(x_1\vert x_2,x_3)$, $p(x_2\vert x_1,x_3)$, and $p(x_3\vert x_1,x_2)$, respectively. (There is no requirement that each distribution be chosen from the same model class.) In addition, for each estimation, we may choose to use a feature-selection algorithm that can discard some of the inputs. Next, we apply the estimation/learning procedures. For the sake of illustration, suppose we discover that $X_1$ is not a significant predictor of $X_3$ and that $X_3$ is not a significant predictor of $X_1$. Finally, we construct the dependency network structure--in this case, $X_1 \leftrightarrow X_2
\leftrightarrow X_3$--and populate the local distributions with the conditional-probability estimates. Note that feature selection in the classification/regression process governs the structure of the dependency network. This algorithm will be computationally efficient in many situations where learning a Markov network and converting it to a dependency network is not. Nonetheless, this procedure has the drawback that, due to--for example--heuristic search and finite-data effects, the resulting local distributions are likely to be inconsistent in that there is no joint distribution $p({\bf x})$ from which each of the local distributions may be obtained via the rules of probability. For example, in the simple domain ${\bf X}=(X_1,X_2)$ it is possible that the estimator of $p(x_1\vert x_2)$ discards $X_2$ as an input whereas the estimator of $p(x_2\vert x_1)$ retains $X_1$ as an input. The result is the structural inconsistency that $X_1$ helps to predict $X_2$, but $X_2$ does not help to predict $X_1$. Numeric inconsistencies are also likely to result. Nonetheless, in situations where the data set contains many samples, strong inconsistencies will be rare because each local distribution is learned from the same data set, which we assume is generated from a single underlying joint distribution. In other words, although dependency networks learned using our procedure will be inconsistent, they will be ``almost'' consistent when learned from data sets with large sample sizes. This observation assumes that the model classes used for the local distributions can closely approximate the conditional distributions consistent with the underlying joint distribution. For example, probabilistic decision trees, which we shall use, satisfy this assumption for variables with finite domains. Because a dependency network learned in this manner is almost consistent, we can imagine--as a heuristic--using procedures resembling Gibbs sampling to extract a joint distribution and to answer probabilistic queries. In the remainder of this section, we formalize these ideas and evaluate them with experiments on real data.

Subsections
next up previous
Next: Definition and Basic Properties Up: Dependency Networks for Inference, Previous: Probabilistic Inference
Journal of Machine Learning Research, 2000-10-19