Having introduced Graph Diffusion Distance in Chapter 2, we proceed to prove several of properties of various instances of our graph dissimilarity score, including triangle inequalities for some specific versions and an upper bound on the distance between two graph products. We will here rely heavily on various properties of the Kronecker sum and product of matrices which may be found in (Hogben 2006), Section 11.4.
For the purpose of the calculations in this section, we restrict ourselves to the “diffusion” term of our objective function [eqn:objfunction] (the term which coerces two diffusion processes to agree), which we will write as Because and are each real and symmetric, they may both be diagonalized as where is a rotation matrix and is a diagonal matrix with the eigenvalues of on the diagonal. Substituting into [eqn:distancedefn], and letting , we have where is an orthogonal matrix if and only if is as well. Since the Frobenius norm is invariant under multiplication by rotation matrices, [eqn:spectral_form] is a re-formulation of our original Laplacian matrix objective function in terms of the spectra of the two graphs. Optimization of this modified form of the objective function subject to orthogonality constraints on is upper-bounded by optimization over matchings of eigenvalues: for any fixed the eigenvalue-matching problem has the same objective function, but our optimization is over all real valued orthogonal . The orthogonality constraint is a relaxed version of the constraints on matching problems (Equation [eqn:matchingconstraints]) discussed in subsection [subsub:definitions], since matching matrices M are also orthogonal . Many algorithms exist for solving the inner partial and 0-1 constrained minimum-cost assignment problems, such as the Munkres algorithm (Munkres 1957) (also in subsection [subsub:definitions]).
We note three corollaries of the above argument. Namely, because the Frobenius norm is invariant under the mapping to and from eigenspace:
Optimal or near-optimal in eigenvalue-space maintain their optimality through the mapping back to graph-space.
Solutions which are within of the optimum in -space are also within of the optimum in -space; and
More precisely, if they exist, zero-cost eigenvalue matchings correspond exactly with zero-cost .
A natural next question would be why it might be worthwhile to work in the original graph-space, rather than always optimizing this simpler eigenvalue-matching problem instead. In many cases (path graphs, cycle graphs) the spectrum of a member of a graph lineage is a subset of that of , guaranteeing that zero-cost eigenvalue matchings (and thus, by the argument above, prolongations with zero diffusion cost) exist. However, when this is not the case, the above argument only upper bounds the true distance, since the matching problem constraints are more strict. Thus, numerical optimization over , with orthogonality constraints only, may find a better bound on .
In this section, we show that both the linear and exponential versions of diffusion distance satisfy the triangle inequality when .
[lem:p_lem_1] For any matrices and , with satisfying ,
and .
Suppose without loss of generality that . Then:
${\left| \left| P M\right| \right|}^2_F = \Tr [ M^T P^T P M ] = \Tr[M^T M] = {\left| \left|M\right| \right|}^2_F$
If , then letting , is a projection operator satisfying . Then,
$$\begin{split} {\left| \left| M \right| \right|}^2_F = \Tr [M^T M ] &= \Tr[M^T M (\Pi + (I - \Pi))] \\ &= \Tr[M^T M \Pi] + \Tr[M^T M (I - \Pi)] \\ &= \Tr[M^T M P P^T] + \Tr[M^T M {(I - \Pi)}^2] \\ &= {\left| \left| M P \right| \right|}_F^2 + {\left| \left| M (I - \Pi) \right| \right|}_F^2 \\ &\geq {\left| \left| M P \right| \right|}_F^2 \end{split}$$
[thm:constant_alpha_thm] satisfies the triangle inequality for .
Let be simple graphs, with Laplacians . Let is guaranteed to exist for constraints which form a compact space of matrices; orthogonality constraints are an example, since the space of orthogonal matrices is closed and bounded. Then where we write to signify that the product satisfies the original transitive constraints on , e.g. orthogonality and/or sparsity. Since the constraint predicate satisfies Equation [eqn:tran_constraint_defn], then so does their product, so we may write By Lemma [lem:p_lem_1],
We note that in this proof we use and (making this the small- or linear version of the objective function), but the same argument holds when all three are replaced with , so we also have
[cor:constant_alpha_exp] satisfies the triangle inequality for .
By the same calculation as in Theorem [thm:constant_alpha_thm], with all replaced by , we have for any constant . Then, letting we have:
Note that in the proofs of Theorem [thm:constant_alpha_thm], Theorem [thm:tsgdd_triangle], and Corollary [cor:constant_alpha_exp], we assume that the constraint predicate includes at least orthogonality (so that we may apply Lemma [lem:p_lem_1]). However, this constraint predicate could be more strict, e.g. include both orthogonality and sparsity. Hence these statements also apply to the cases in Table [tab:dist_versions], which we do not otherwise consider in this work: in our numerical experiments we (for reasons of computational simplicity) only require our optimization over be orthogonally constrained.
For any graphs and , and some real nuber , we define the Time-Scaled Graph Diffusion Distance (TSGDD) as a scaled version of the linear distance, with fixed. Namely, let The intuition for this version of the distance measure is that we are constraining the time dilation, , between and to be a power of the ratio of the two graph sizes. The factor is needed to ensure this version of the distance satisfies the triangle inequality, as seen in Theorem [thm:tsgdd_triangle].
[thm:tsgdd_triangle] The TSGDD, as defined above, satisfies the triangle inequality.
As above, let be three graphs with and , and let be the Laplacian of . Let represent a transitive constraint predicate, also as described previously. Then, for a constant , we have: $$\begin{aligned} &\tilde{D}_r(G_1, G_3) = \\ & \qquad \inf_{P|\mathcal{C}(P)} {\left( n_1 n_3\right)}^{-r} {\left|\left| {\left( \frac{n_1}{n_3} \right)}^{-r} P L_1 - {\left( \frac{n_1}{n_3} \right)}^{r} L_3 P \right|\right|}_F \nonumber \\ & \qquad \leq \inf_{P_{32}, P_{21} | \mathcal{C}(P_{32} P_{21})} {\left( n_1 n_3\right)}^{-r} {\left|\left| {\left( \frac{n_1}{n_3} \right)}^{-r} P_{32} P_{21} L_1 - {\left( \frac{n_1}{n_3} \right)}^{r} L_3 P_{32} P_{21} \right|\right|}_F \nonumber \\ \intertext{under the assumption, as in Equation \eqref{eqn:tran_constraint_defn}, that $ \mathcal{C}(P_{32}) \wedge \mathcal{C}(P_{21}) \implies \mathcal{C}(P_{32} P_{21}) $, } \end{aligned}$$ $$\begin{aligned} % &\tilde{D}_r(G_1, G_3) \leq \\ &\inf_{P_{32}, P_{21} | \mathcal{C}(P_{32}) \wedge \mathcal{C}(P_{21})} {\left( n_1 n_3\right)}^{-r} {\left|\left| {\left( \frac{n_1}{n_3} \right)}^{-r} P_{32} P_{21} L_1 - {\left( \frac{n_1}{n_3} \right)}^{r} L_3 P_{32} P_{21} \right|\right|}_F \nonumber \\ % &= \inf_{P_{32}, P_{21} | \mathcal{C}(P_{32}) \wedge \mathcal{C}(P_{21})} {\left( n_1 n_3\right)}^{-r} {\left|\left| {\left( \frac{n_1}{n_3} \right)}^{-r} P_{32} P_{21} L_1 - {\left( \frac{n_1 n_3}{n_2^2} \right)}^{r} P_{32} L_2 P_{21} \right. \right. } \nonumber \\* &\null \qquad + \qquad {\left. \left. {\left( \frac{n_1 n_3}{n_2^2} \right)}^{r} P_{32} L_2 P_{21} - {\left( \frac{n_1}{n_3} \right)}^{r} L_3 P_{32} P_{21} \right|\right|}_F \\ % &\leq \inf_{P_{32}, P_{21} | \mathcal{C}(P_{32}) \wedge \mathcal{C}(P_{21})} {\left( n_1 n_3\right)}^{-r} {\left|\left| {\left( \frac{n_1}{n_3} \right)}^{-r} P_{32} P_{21} L_1 - {\left( \frac{n_1 n_3}{n_2^2} \right)}^{r} P_{32} L_2 P_{21} \right|\right|}_F \\ &\null \qquad + \qquad {\left( n_1 n_3\right)}^{-r} {\left|\left| {\left( \frac{n_1 n_3}{n_2^2} \right)}^{r} P_{32} L_2 P_{21} - {\left( \frac{n_1}{n_3} \right)}^{r} L_3 P_{32} P_{21} \right|\right|}_F \\ % &= \inf_{P_{32}, P_{21} | \mathcal{C}(P_{32}) \wedge \mathcal{C}(P_{21})} {\left( n_1 n_3\right)}^{-r} {\left( \frac{n_3}{n_2} \right)}^{r} {\left|\left| {\left( \frac{n_1}{n_2} \right)}^{-r} P_{32} P_{21} L_1 - {\left( \frac{n_1}{n_2} \right)}^{r} P_{32} L_2 P_{21} \right|\right|}_F \\ &\null \qquad + \qquad {\left( n_1 n_3\right)}^{-r} {\left( \frac{n_1}{n_2} \right)}^{r} {\left|\left| {\left( \frac{n_2}{n_3} \right)}^{-r} P_{32} L_2 P_{21} - {\left( \frac{n_2}{n_3} \right)}^{r} L_3 P_{32} P_{21} \right|\right|}_F \\ % &= \inf_{P_{32}, P_{21} | \mathcal{C}(P_{32}) \wedge \mathcal{C}(P_{21})} {\left( n_1 n_2\right)}^{-r} {\left|\left| {\left( \frac{n_1}{n_2} \right)}^{-r} P_{32} P_{21} L_1 - {\left( \frac{n_1}{n_2} \right)}^{r} P_{32} L_2 P_{21} \right|\right|}_F \\ &\null \qquad + \qquad {\left( n_2 n_3\right)}^{-r} {\left|\left| {\left( \frac{n_2}{n_3} \right)}^{-r} P_{32} L_2 P_{21} - {\left( \frac{n_2}{n_3} \right)}^{r} L_3 P_{32} P_{21} \right|\right|}_F \intertext{By Lemma \ref{lem:p_lem_1},} \end{aligned}$$ $$\begin{aligned} \tilde{D}_r(G_1, G_3) &\leq \inf_{P_{32}, P_{21} | \mathcal{C}(P_{32}) \wedge \mathcal{C}(P_{21})} {\left( n_1 n_2\right)}^{-r} {\left|\left| {\left( \frac{n_1}{n_2} \right)}^{-r} P_{21} L_1 - {\left( \frac{n_1}{n_2} \right)}^{r} L_2 P_{21} \right|\right|}_F \\* &\null \qquad + \qquad {\left( n_2 n_3\right)}^{-r} {\left|\left| {\left( \frac{n_2}{n_3} \right)}^{-r} P_{32} L_2 - {\left( \frac{n_2}{n_3} \right)}^{r} L_3 P_{32} \right|\right|}_F \\ % &= \inf_{P_{21} | \mathcal{C}(P_{21})} {\left( n_1 n_2\right)}^{-r} {\left|\left| {\left( \frac{n_1}{n_2} \right)}^{-r} P_{21} L_1 - {\left( \frac{n_1}{n_2} \right)}^{r} L_2 P_{21} \right|\right|}_F \\* &\null \qquad + \qquad \inf_{P_{32} | \mathcal{C}(P_{32})} {\left( n_2 n_3\right)}^{-r} {\left|\left| {\left( \frac{n_2}{n_3} \right)}^{-r} P_{32} L_2 - {\left( \frac{n_2}{n_3} \right)}^{r} L_3 P_{32} \right|\right|}_F \\ &= \tilde{D}_r(G_1, G_2) + \tilde{D}_r(G_2, G_3) \nonumber \intertext{and so} &\tilde{D}_r(G_1, G_3) \leq \tilde{D}_r(G_1, G_2) + \tilde{D}_r(G_2, G_3)\end{aligned}$$ for any fixed .
Recall that we use the notation for a constraint predicate that must be satisfied by prolongation matrix , which is transitive in the sense that: The simplest example is . Let is the total number of nonzero entries in row or column of . Sparsity can be introduced in transitive form by where for some real number . Here, and are the dimensions of . This predicate is transitive since and since cancels out from the numerator and denominator of the product of the fanout bounds. This transitive sparsity constraint depends on a power-law parameter . When , there is no sparsity constraint.
Another form of sparsity constraints are those which specify a pattern on matrix entries which are allowed to be nonzero. Two simple examples (which are also transitive) are matrices which are constrained to be upper triangular, as well as matrices which are constrained to be of the form where and are themselves both constrained to be sparse. More complicated are matrices which are constrained to be banded for some specified pattern of bands: more specifically, that there is a reordering of the rows and columns that the number of diagonal bands (of width 1, slope ) with nonzero entries is less than for some . For example, linear interpolation matrices between d-dimensional grids, with non-overlapping source regions, follow this constraint.
As a final note on sparsity, we observe that any of the optimizations detailed in this work could also be performed including a sparsity term (for example, the -norm of the matrix , calculated as is one possibility, as are terms which penalize or far from 1), rather than explicit sparsity constraints. A potential method of performing this optimization would be to start by optimizing the non-sparse version of the objective function (as detailed in Section 4.1) and then slowly increasing the strength of the regularization term.
We next consider the problem of finding optimal prolongations between two graphs and when optimal prolongations are known between and , and and . We show that under some reasonable assumptions, these two prolongation optimizations decouple - we may thus solve them separately and combine the solutions to obtain the optimal prolongations between the two product graphs.
From the definition of graph box product, we have where is the Kronecker sum of matrices as previously defined. See (Fiedler 1973), Item 3.4 for more details on Laplacians of graph products. We calculate
$$\begin{aligned} D^{P,\alpha} \left( G_\Box^{(1)}, G_\Box^{(2)} \right) &= \left| \left| \frac{1}{\sqrt{\alpha}} P L_\Box^{(1)} - \sqrt{\alpha} L_\Box^{(2)} P \right| \right|_F\\ &= \left| \left| \frac{1}{\sqrt{\alpha}} P \left( \left( L_1^{(1)} \otimes I_2^{(1)} \right) + \left(I_1^{(1)} \otimes L_2^{(1)}\right) \right) \right. \right. \\ & \left. \left. \qquad - \sqrt{\alpha} \left( \left( L_1^{(2)} \otimes I_2^{(2)}\right) + \left(I_1^{(2)} \otimes L_2^{(2)} \right) \right) P \right| \right|_F \\ &= \left| \left| \left( \frac{1}{\sqrt{\alpha}} P \left(L_1^{(1)} \otimes I_2^{(1)}\right) - \sqrt{\alpha}\left(L_1^{(2)} \otimes I_2^{(2)}\right) P \right) \right. \right. \\ & \left. \left. \qquad + \left( \frac{1}{\sqrt{\alpha}} P \left( I_1^{(1)} \otimes L_2^{(1)} \right) - \sqrt{\alpha} \left( I_1^{(2)} \otimes L_2^{(2)} \right) P\right) \right| \right|_F \\ \intertext{Now we try out the assumption that $P = P_1 \otimes P_2$, which restricts the search space over $P$ and may increase the objective function:} D^{P,\alpha} \left( G_\Box^{(1)}, G_\Box^{(2)} \right) &= \left| \left| \left[ \frac{1}{\sqrt{\alpha}} \left( P_1 \otimes P_2 \right)\left(L_1^{(1)} \otimes I_2^{(1)}\right) \right. \right. \right. \\ & \qquad \qquad - \left. \sqrt{\alpha}\left(L_1^{(2)} \otimes I_2^{(2)}\right) \left( P_1 \otimes P_2 \right) \right] \\ & \qquad + \left[ \frac{1}{\sqrt{\alpha}} \left( P_1 \otimes P_2 \right) \left( I_1^{(1)} \otimes L_2^{(1)} \right) \right. \\ & \left. \left. \left. \qquad \qquad - \sqrt{\alpha} \left( I_1^{(2)} \otimes L_2^{(2)} \right) \left( P_1 \otimes P_2 \right) \right] \right| \right|_F \\ &= \left| \left| \left( \frac{1}{\sqrt{\alpha}} \left( P_1 L_1^{(1)} \otimes P_2 \right) - \sqrt{\alpha}\left(L_1^{(2)} P_1 \otimes P_2 \right) \right) \right. \right. \\ & \left. \left. \qquad + \left( \frac{1}{\sqrt{\alpha}} \left( P_1 \otimes P_2 L_2^{(1)} \right) - \sqrt{\alpha} \left( P_1 \otimes L_2^{(2)} P_2 \right) \right) \right| \right|_F \\ &= \left| \left| \left( \left( \frac{1}{\sqrt{\alpha}} P_1 L_1^{(1)} - \sqrt{\alpha} L_1^{(2)} P_1 \right) \otimes P_2 \right) \right. \right. \\ & \left. \left. \qquad + \left( P_1 \otimes \left( \frac{1}{\sqrt{\alpha}} P_2 L_2^{(1)} - \sqrt{\alpha} L_2^{(2)} P_2 \right) \right) \right| \right|_F \intertext{Since $|| A + B ||_F \leq ||A||_F + ||B||_F$,} &\leq \left| \left| \left( \left( \frac{1}{\sqrt{\alpha}} P_1 L_1^{(1)} - \sqrt{\alpha} L_1^{(2)} P_1 \right) \otimes P_2 \right) \right| \right| \\ & \qquad + \left| \left| \left( P_1 \otimes \left( \frac{1}{\sqrt{\alpha}} P_2 L_2^{(1)} - \sqrt{\alpha} L_2^{(2)} P_2 \right) \right) \right| \right|_F \\ &= \left| \left| \frac{1}{\sqrt{\alpha}} P_1 L_1^{(1)} - \sqrt{\alpha} L_1^{(2)} P_1 \right| \right|_F \left| \left| P_2 \right| \right|_F \\ & \qquad + \left| \left| P_1 \right| \right|_F \left| \left| \frac{1}{\sqrt{\alpha}} P_2 L_2^{(1)} - \sqrt{\alpha} L_2^{(2)} P_2 \right| \right|_F , \\ \intertext{Thus assuming $P = P_1 \otimes P_2$} D^{P,\alpha} \left( G_\Box^{(1)}, G_\Box^{(2)} \right) &\leq \left| \left| \tilde{P}_2 \right| \right|_F D_{\alpha, P_1} \left(G_1^{(1)},G_1^{(2)} \right) \\ & \qquad + \left| \left| \tilde{P}_1 \right| \right|_F D_{\alpha, P_2} \left( G_2^{(1)},G_2^{(2)} \right), %\end{aligned}$$ which is a weighted sum of objectives of the optimizations for prolongation from to and to . Recall that our original constraint on was that ; since this is equivalent (by a property of the Kronecker product; see Corollary 13.8 in (Laub 2005)) to the coupled constraints on and : for some . For any which obey [eqn:p1p2const], we may rescale them by to make them orthogonal without changing the value of the objective, so we take in subsequent calculations. Noting that , we see that
Thus, we have proven the following:
[thm:decompthm] Assuming that decomposes as , the diffusion distance between and is bounded above by the strictly monotonically increasing function of the two distances and : $$\begin{aligned} \mathcal{F}(D_{P_1,\alpha}, D_{P_2,\alpha}) &= \sqrt{n_2^{(1)}} D_{P_1,\alpha} + \sqrt{n_1^{(1)}} D_{P_2,\alpha} , \intertext{Namely,} D_{P, \alpha} \left(G_\Box^{(1)}, G_\Box^{(2)} \right) &\leq \mathcal{F} \left( D_{P_1,\alpha} \left( G_1^{(1)}, G_1^{(2)} \right), D_{P_2,\alpha} \left( G_2^{(1)}, G_2^{(2)} \right) \right)\end{aligned}$$
Thus, the original optimization over the product graphs decouples into separate optimizations over the two sets of factors, constrained to have the same value of . Additionally, since the requirement that is an additional constraint,
[thm:decompcorollary] If and , subject to orthogonality constraints, are optima of and , and furthermore if , then the value of for an optimal is bounded above by .
This upper bound on the original objective function is a monotonically increasing function of the objectives for the two smaller problems. A consequence of this upper bound is that if and , then the composite solution must have . Thus if both of these distances are arbitrarily small then the composite distance must also be small. Furthermore, if only one of these is small, so that or , then or , respectively.
We have experimentally found that many families of graphs do not require scaling between the two diffusion processes: the optimal pair has . In particular, prolongation between path (cycle) graphs of size and size always have , since the spectrum of the former graph is a subset of that of the larger - therefore, there is a matching solution of cost 0 which by the argument above can be mapped to a graph-space with objective function value 0 (we prove this in Section 3.7). In this case, the two terms of the upper bound are totally decoupled and may each be optimized separately (whereas in the form given above, they both depend on a ).
We now consider the case where we want to compute the distance of two graph box products, i.e. where and
are known for some .
[thm:sq_grid_lim_lem] Let and be graph box products as described above, and for a graph let be its Laplacian. For fixed , , as given above, for any , we have where all distances are evaluated at , , but we have omitted that notation for simplicity.
For graph products , we have (this fact can be easily verified from the formula for the adjacency matrix of a graph box product, given in the definition in Section 1.4.2), and so Because and commute for any and , We will make the following abbreviations: Then, By Lemma [lem:p_lem_1], $$\begin{aligned} \intertext{If we instead use $\left( E^{(1)}_2 P^{(1)} \otimes P^{(2)} E^{(2)}_1 \right)$ as the cross term in Equation \eqref{eqn:cross_term_eqn}, we have } \inf_P D \left(\mathbf{G}_1, \mathbf{G}_2 \right) \leq {\left| \left| E^{(1)}_2 \right| \right|}_F {\left| \left| P^{(2)} E^{(2)}_1 - E^{(2)}_2 P^{(2)} \right| \right|}_F \\ \quad \quad + \quad {\left| \left| P^{(1)} E^{(1)}_1 - E^{(1)}_2 P^{(1)} \right| \right|}_F {\left| \left| E^{(2)}_1 \right| \right|}_F \nonumber\end{aligned}$$ A linear combination of these two bounds gives us the desired bound.
This has the additional consequence that Additionally, if This reduces further to and so $$\begin{aligned} & D \left( \left. G_1^{(1)} \Box G_1^{(1)}, G_2^{(1)} \Box G_2^{(1)} \right| t_c, \alpha_c \right) \label{eqn:graph_product_special_case} \\ &\null \quad \quad \quad \leq \min \left( {\left| \left| e^{\frac{t_c}{a_c} L(G^{(1)}_1)} \right| \right|}_F , {\left| \left| e^{t_c a_c L(G^{(1)}_2)} \right| \right|}_F \right) D \left( G_1^{(1)}, G_2^{(1)} \left. \right| t_c, \alpha_c \right) \nonumber \end{aligned}$$ An example of such a graph sequence is the sequence of two-dimensional square grids, which are each the box product of two identical one-dimensional grids i.e., path graphs: .
[theorem:spectralzero] In this section, we give an example closed-form solution for a prolongation matrix which achieves zero error when prolonging between cycle graphs. Let and be graphs with spectra and , respectively, with . Suppose that for every of multiplicity , is also an eigenvalue of of multiplicity . Then there is a zero-cost eigenvalue matching between and .
Let be a list of pairs of indices such that the following hold:
All of the are unique.
All of the are unique.
For any pair , .
Define as follows: is clearly orthogonal, since it has exactly one in each row and each column and zeros elsewhere ( is a permutation matrix for and a subpermutation matrix otherwise). Furthermore, we must have and therefore
For any , there exist zero-error matchings between cycle graphs and .
The spectra of are given by the formula (see (Hogben 2006) Section 39.3): Thus and clearly satisfy the conditions of Theorem [theorem:spectralzero] above. In particular, the matrix has 0 cost as in Equation [eqn:p_zero_cost]
In this section we derive an eigenspace version of the bound derived in Section 3.5.
The eigenspace version of the diffusion term of the objective function of a graph product prolongation also decouples into two smaller prolongation problems.
As in section 3.1, we adopt the following notation: $$\begin{aligned} \null && L_i^{(j)} = L(G_i^{(j)}) = {U_i^{(j)}}^T \Lambda_i^{(j)} U_i^{(j)} && \text{for $i \in \{1,2\}$} \\ \null && L_\Box^{(j)} = L(G_\Box^{(j)}) ={U_\Box^{(j)}}^T \Lambda_\Box^{(j)} U_\Box^{(j)} && \null\end{aligned}$$
For the graph box product, we have Where is the Kronecker sum of matrices as previously defined. Additionally, if diagonalizes and diagonalizes , then diagonalizes , since The Kronecker product of two diagonal matrices is itself diagonal, and so this is a sum of two diagonal matrices. Furthermore, the Kronecker product of two rotation matrices is itself a rotation matrix, and therefore is a rotation matrix.
Applying this diagonalization to map to eigenspace in the same manner as subsection 3.1, we have $$\begin{aligned} D^2_{P,\alpha} \left( G_\Box^{(1)}, G_\Box^{(2)} \right) &= \left| \left| \frac{1}{\sqrt{\alpha}} P L_\Box^{(1)} - \sqrt{\alpha} L_\Box^{(2)} P \right| \right|_F\\ &= \left| \left| \frac{1}{\sqrt{\alpha}} \tilde{P} \Lambda_\Box^{(1)} - \sqrt{\alpha} \Lambda_\Box^{(2)} \tilde{P} \right| \right|_F \\ &= \left| \left| \frac{1}{\sqrt{\alpha}} \tilde{P} \left( \left( \Lambda_1^{(1)} \otimes I_2^{(1)} \right) + \left(I_1^{(1)} \otimes \Lambda_2^{(1)}\right) \right) \right. \right. \\ & \left. \left. \qquad - \sqrt{\alpha} \left( \left( \Lambda_1^{(2)} \otimes I_2^{(2)}\right) + \left(I_1^{(2)} \otimes \Lambda_2^{(2)} \right) \right) \tilde{P} \right| \right|_F \\ &= \left| \left| \left( \frac{1}{\sqrt{\alpha}} \tilde{P} \left(\Lambda_1^{(1)} \otimes I_2^{(1)}\right) - \sqrt{\alpha}\left(\Lambda_1^{(2)} \otimes I_2^{(2)}\right) \tilde{P} \right) \right. \right. \\ & \left. \left. \qquad + \left( \frac{1}{\sqrt{\alpha}} \tilde{P} \left( I_1^{(1)} \otimes \Lambda_2^{(1)} \right) - \sqrt{\alpha} \left( I_1^{(2)} \otimes \Lambda_2^{(2)} \right) \tilde{P} \right) \right| \right|_F \\ \intertext{Now we make the assumption that $P = P_1 \otimes P_2$ and therefore $\tilde{P} = \tilde{P}_1 \otimes \tilde{P}_2$ for some $\tilde{P}_1$ and $\tilde{P}_2$,} &= \left| \left| \left( \frac{1}{\sqrt{\alpha}} \left( \tilde{P}_1 \otimes \tilde{P}_2 \right)\left(\Lambda_1^{(1)} \otimes I_2^{(1)}\right) - \sqrt{\alpha}\left(\Lambda_1^{(2)} \otimes I_2^{(2)}\right) \left( \tilde{P}_1 \otimes \tilde{P}_2 \right) \right) \right. \right. \\ & \left. \left. \qquad + \left( \frac{1}{\sqrt{\alpha}} \left( \tilde{P}_1 \otimes \tilde{P}_2 \right) \left( I_1^{(1)} \otimes \Lambda_2^{(1)} \right) - \sqrt{\alpha} \left( I_1^{(2)} \otimes \Lambda_2^{(2)} \right) \left( \tilde{P}_1 \otimes \tilde{P}_2 \right) \right) \right| \right|_F \\ &= \left| \left| \left( \frac{1}{\sqrt{\alpha}} \left( \tilde{P}_1 \Lambda_1^{(1)} \otimes \tilde{P}_2 \right) - \sqrt{\alpha}\left(\Lambda_1^{(2)} \tilde{P}_1 \otimes \tilde{P}_2 \right) \right) \right. \right. \\ & \left. \left. \qquad + \left( \frac{1}{\sqrt{\alpha}} \left( \tilde{P}_1 \otimes \tilde{P}_2 \Lambda_2^{(1)} \right) - \sqrt{\alpha} \left( \tilde{P}_1 \otimes \Lambda_2^{(2)} \tilde{P}_2 \right) \right) \right| \right|_F \\ &= \left| \left| \left( \left( \frac{1}{\sqrt{\alpha}} \tilde{P}_1 \Lambda_1^{(1)} - \sqrt{\alpha} \Lambda_1^{(2)} \tilde{P}_1 \right) \otimes \tilde{P}_2 \right) \right. \right. \\ & \left. \left. \qquad + \left( \tilde{P}_1 \otimes \left( \frac{1}{\sqrt{\alpha}} \tilde{P}_2 \Lambda_2^{(1)} - \sqrt{\alpha} \Lambda_2^{(2)} \tilde{P}_2 \right) \right) \right| \right|_F \\ \intertext{Since $|| A + B ||_F \leq ||A||_F + ||B||_F$,} \\ &\leq \left| \left| \left( \left( \frac{1}{\sqrt{\alpha}} \tilde{P}_1 \Lambda_1^{(1)} - \sqrt{\alpha} \Lambda_1^{(2)} \tilde{P}_1 \right) \otimes \tilde{P}_2 \right) \right| \right| \\ & \qquad + \left| \left| \left( \tilde{P}_1 \otimes \left( \frac{1}{\sqrt{\alpha}} \tilde{P}_2 \Lambda_2^{(1)} - \sqrt{\alpha} \Lambda_2^{(2)} \tilde{P}_2 \right) \right) \right| \right|_F \\ &= \left| \left| \frac{1}{\sqrt{\alpha}} \tilde{P}_1 \Lambda_1^{(1)} - \sqrt{\alpha} \Lambda_1^{(2)} \tilde{P}_1 \right| \right|_F \left| \left| \tilde{P}_2 \right| \right|_F \\ & \qquad + \left| \left| \tilde{P}_1 \right| \right|_F \left| \left| \frac{1}{\sqrt{\alpha}} \tilde{P}_2 \Lambda_2^{(1)} - \sqrt{\alpha} \Lambda_2^{(2)} \tilde{P}_2 \right| \right|_F \\\end{aligned}$$ From Theorem [thm:decompthm] of the main manuscript, we know that for and assuming , Trivially, we can rewrite each of these Frobenius norms to be their spectral version, as in Equation [eqn:distancedefn]. Thus, which is a weighted sum of objectives of the two spectral prolongation problems for the two factor lineages. We have thus also decoupled this eigenvalue-matching version of the objective function into two separate prolongation problems.
Finally, we show that if and are solutions to the eigenvalue matching problem and respectively, then is a valid, but not necessarily optimal, solution to the eigenvalue matching problem . By valid we mean that satisfies the constraints given in the definition of matching problems in Section [subsub:definitions].
This fact follows directly from the constraints on and . A matrix is a valid matching matrix iff its entries are in and it is orthogonal (this is an equivalent expression of the constraints given in Section [subsub:definitions]. If and are both orthogonal and -valued, then we observe the following facts about their Kronecker product :
it is also -valued, since any of its entries is the product of an entry of and one of .
it is orthogonal, since is orthogonal iff and for some (Laub 2005). and satisfy these conditions with .
So satisfies the constraints given for the eigenvalue matching problem .
We can add a term to the graph diffusion distance which penalizes large distortions induced by , as follows: define We can show analytically that this distance satisfies the triangle inequality:
satisfies the triangle inequality.
For graphs and Laplacians , for any fixed , we have: Suppose that Then, $$\begin{aligned} \inf_{P|\mathcal{C}(P)} {\left| \left| P e^{t L_1} - e^{t L_3 } P \right| \right|}_F &\leq {\left| \left| P_{32} P_{21} e^{t L_1} - e^{t L_3 } P_{32} P_{21} \right| \right|}_F \\ \inf_{P|\mathcal{C}(P)} {\left| \left| P e^{t L_1} - e^{t L_3 } P \right| \right|}_F &\leq {\left| \left| P_{32} P_{21} e^{t L_1} - P_{32} P_{21} e^{\frac{t}{\alpha_{21}} L_1} \right. \right. } + P_{32} P_{21} e^{\frac{t}{\alpha_{21}} L_1} \\* & \quad - P_{32} e^{t \alpha_{21} L_2 } P_{21} + P_{32} e^{t \alpha_{21} L_2 } P_{21} - P_{32} e^{t L_2 } P_{21} \\* & \quad + P_{32} e^{t L_2 } P_{21} - P_{32} e^{\frac{t}{\alpha_{32}} L_2 } P_{21} + P_{32} e^{\frac{t}{\alpha_{32}} L_2 } P_{21} \\* & \quad - e^{t \alpha_{32} L_3 } P_{32} P_{21} {\left. \left. + e^{t \alpha_{32} L_3 } P_{32} P_{21} - e^{t L_3 } P_{32} P_{21} \right| \right|}_F \\ &\leq {\left| \left| P_{32} P_{21} e^{t L_1} - P_{32} P_{21} e^{\frac{t}{\alpha_{21}} L_1}\right| \right|}_F \\* & \quad + {\left| \left| P_{32} P_{21} e^{\frac{t}{\alpha_{21}} L_1} - P_{32} e^{t \alpha_{21} L_2 } P_{21} \right| \right|}_F \\* & \quad + {\left| \left| P_{32} e^{t \alpha_{21} L_2 } P_{21} -P_{32} e^{t L_2 } P_{21} \right| \right|}_F \\* & \quad + {\left| \left| P_{32} e^{t L_2 } P_{21} - P_{32} e^{\frac{t}{\alpha_{32}} L_2 } P_{21} \right| \right|}_F \\* & \quad + {\left| \left| P_{32} e^{\frac{t}{\alpha_{32}} L_2 } P_{21} - e^{t \alpha_{32} L_3 } P_{32} P_{21} \right| \right|}_F \\* & \quad + {\left| \left| e^{t \alpha_{32} L_3 } P_{32} P_{21} - e^{t L_3 } P_{32} P_{21} \right| \right|}_F \\ \intertext{by Lemma \ref{lem:p_lem_1},} \inf_{P|\mathcal{C}(P)} {\left| \left| P e^{t L_1} - e^{t L_3 } P \right| \right|}_F &\leq {\left| \left| e^{t L_1} - e^{\frac{t}{\alpha_{21}} L_1}\right| \right|}_F + {\left| \left| P_{21} e^{\frac{t}{\alpha_{21}} L_1} - e^{t \alpha_{21} L_2 } P_{21} \right| \right|}_F \\ & \quad + {\left| \left| e^{t \alpha_{21} L_2 } P_{21} - e^{t L_2 } P_{21} \right| \right|}_F \\ &+ {\left| \left| e^{t L_2 } - e^{\frac{t}{\alpha_{32}} L_2 } \right| \right|}_F + {\left| \left| P_{32} e^{\frac{t}{\alpha_{32}} L_2 } - e^{t \alpha_{32} L_3 } P_{32} \right| \right|}_F \\ & \quad + {\left| \left| e^{t \alpha_{32} L_3 } P_{32} - e^{t L_3 } P_{32} \right| \right|}_F \\ &= D_\text{reg}(G_1, G_2 | t = c) + D_\text{reg}(G_2, G_3 | t = c)\end{aligned}$$ Since this is true for any fixed , let Then
We can construct a similar regularized version of the linear objective function: $$\begin{aligned} \tilde{D}_\text{reg}(G_1, G_2) &= {\left| \left| \frac{1}{\alpha} P L_1 - \alpha L_2 P \right| \right|} + {\left| \left| \frac{1}{\alpha} L_1 - L_1 \right| \right|} + {\left| \left| \frac{\null}{\null} P L_2 - \alpha L_2 P \right| \right|}\end{aligned}$$
The additional terms included in and penalize distorting the respective Laplacians far from their original values. In practice, many of the theoretical guarantees provided earlier in this manuscript may not apply to optimization of the augmented objective function. Hence, a major area of future work will be modification of our optimization procedure to compute this form of distance.
Triangle inequalities are proven for some members of the proposed family of graph distortion or “distance” measures, including infinitesimal and finite diffusion time, a power law for sparsity, and/or a power law for the time scaling factor between coarse and fine scales. However, the case of an optimal (not power law) time conversion factor needs to be investigated by numerical experiment, and that requires new algorithms, introduced in Section 4.3. We also show that in the case of distances between graph box products, optimization over for the product graphs is bounded above by a monotonic function of the optimum over the component graphs.