Intereting Posts

When does independence imply conditional independence, and vice versa?
Finitely generated idempotent ideals are principal: proof without using Nakayama's lemma
Fixed Point of $x_{n+1}=i^{x_n}$
Solving the equation $x^{2}-7^{y}=2$
A formal name for “smallest” and “largest” partition
Geometric intuition behind the Lie bracket of vector fields
Distribution of sums
Is there a notation for being “a finite subset of”?
Graph theory problem (edge-disjoint matchings)
Polynomial GCD in the presence of floating-point errors
Convergence of Integral Implies Uniform convergence of Equicontinuous Family
What are $E_\infty$-rings?
A natural proof of the Cauchy-Schwarz inequality
Smallest integer $k$ so that no Sudoku grid has exactly $k$ solutions
Ordered binary sequences of length n with no two consecutive 0's without using recursion

Let $T$ be a linear operator on the $n$ dimensional vector space $\mathbb{V}$. Suppose that $\mathbb{V} = \sum_{i=1}^{k}W_i$ where each $W_i$ is $T$ – invariant. Let $\mu_{T_i}$ be the minimal polynomial of the operator restricted to $W_i$. If the minimal polynomials of all restrictions are coprime, that is if $\gcd(\mu_{T_i}, \mu_{T_j}) = 1$ for $i\neq j$, is it true that the subspaces $W_i$ are independent? That is, $\mathbb{V} = \bigoplus_{i=1}^k W_i$. Perhaps more generally, is it true that if the minimal polynomials of two subspaces are coprime, then the two subspaces are independent?

The converse is easily seen to be false, the identity operator is a clear counterexample. I haven’t been able to find an easy counterexample to the above statement however. A proof or counterexample would be appreciated.

- covariant and contravariant components and change of basis
- Can we prove that all eigenvalues of $A$ are positive?
- Prove that the trace of the matrix product $U'AU$ is maximized by setting $U$'s columns to $A$'s eigenvectors
- Why do we care about dual spaces?
- Proving the determinant of a tridiagonal matrix with $-1, 2, -1$ on diagonal.
- Matrix characterization of surjective and injective linear functions

- Find the eigenvalues of a matrix with ones in the diagonal, and all the other elements equal
- What is the transformation representation/interpretation of symmetric matrices?
- Counting diagonalizable matrices in $\mathcal{M}_{n}(\mathbb{Z}/p\mathbb{Z})$
- Does the shift operator on $\ell^2(\mathbb{Z})$ have a logarithm?
- Basis for the Space of Quadratic Polynomials $P^{(2)}$ — Homework Help
- Lattice generated by vectors orthogonal to an integer vector
- Minimum eigenvalue and singular value of a square matrix
- Linear Algebra - Rank of a matrix
- Why is the determinant defined in terms of permutations?
- Use of determinants

Yes it is true. Here is a sketch of a proof by contraposition.

If $W:=W_j\cap \sum\limits_{i\neq j}W_i\neq\{0\}$ for some $j$, then $W$ is a nonzero invariant subspace for $T$. The minimial polynomial $p$ of the restriction of $T$ to $W$ must divide both $\mu_{T_j}$ and the minimal polynomial $q$ of $T$ restricted to $\sum\limits_{i\neq j}W_i$. But $q$ divides $\prod\limits_{i\neq j}\mu_{T_I}$, and this implies that each irreducible factor of $p$ must divide both $\mu_{T_j}$ and $\mu_{T_i}$ for some $i\neq j$. Therefore $\gcd(\mu_{T_i}, \mu_{T_j}) \neq 1$.

Here is a variation of Jonas’s proof.

Let $\mu$ be the product of the $\mu_{T_i}$, which I will denote $\mu_i$, and let $K$ be the ground field.

Since $\mu(T)=0$, the ring $K[T]$ is a quotient of $A:=K[X]/(\mu)$, and $\mathbb{V}$ is an $A$-module.

By the Chinese Remainder Theorem, $A$ is isomorphic to the product of the $A_i:=K[X]/(\mu_i)$.

The element $e_i$ of $A$ whose $j$-th component is $\delta_{ij}$ acts by the identity on $W_i$, and by $0$ on $W_j$ for $j\neq i$.

For $i=1,\dots,k$, let $w_i$ be in $W_i$, and assume $\sum w_i=0$.

By applying $e_j$ to the above equality, we get $w_j=0$.

Here is one more variant. Due to the relative primality, $\mu_i[T]$ acts invertibly on each subspace $W_j$ with $j\neq i$: if $r,s\in K[T]$ are Bezout coefficients for $\mu_i,\mu_j$, in other words $r\mu_i+s\mu_j=1$, then restricted to$~W_j$ the operator $r[T]\mu_i[T]$ acts as the identity. This is the kind of situation that easily give directness of the sum of the subspaces $W_i$: given a hypothetical non-trivial relation $0=w_1+\cdots+w_k$ with $w_i\in W_i$ for all$~i$, take one for which a minimal number of *nonzero* vectors among the $w_i$ (but of course at least one; in fact there now clearly must be at least two nonzero among them); with $w_i\neq0$, apply $\mu_i[T]$ to the relation to find a new relation with exactly one less nonzero vector, contradicting minimality.

- Diophantine equation: $n^p+3^p=k^2$
- Hessian Related convex optimization question
- Proving that any permutation in $S_n$ can be written as a product of disjoint cycles
- What happens to small squares in Riemann mapping?
- Probability that a normal distribution is greater than two others
- What are the requirements for separability inheritance
- Markovian Gaussian stationary process with continuous paths
- Finding the remainder of $\overbrace{11\ldots1}^{123 \text{ times}}$ divided by $271$
- Separation in dual space
- How do I show that the integral of $e^{inx}$ over a set of measure $1$ is nonzero for some nonzero $n$?
- Simplifying Ramanujan-type Nested Radicals
- $2^{k} \mid {2k \choose 0} \cdot 3^{0} + {2k \choose 2} \cdot 3^{1} + \cdots + {2k \choose 2i} \cdot 3^{i} + \cdots + {2k \choose 2k} \cdot 3^{k}$
- What does the factorial of a negative number signify?
- Is there a definitive guide to speaking mathematics?
- Ring with 10 elements is isomorphic to $\mathbb{Z}/10 \mathbb{Z} $