“…MFNets accommodate noisy data and are trained via gradient-based minimization of a nonlinear least squares objective. While MFNets can learn non-hierarchical relations between data sources, they: (1) rely on having prior knowledge on a set of latent variables that explain the relations between the sources, (2) assume each source can be surrogated via a linear subspace model, (3) are not probabilistic and also require regularization, (4) impose independence assumption among the data sources to derive the likelihood (i.e., the objective) function, and (5) rely on iterative approaches for finding the optimal graph structure.…”