Intereting Posts

Weak Gâteaux Derivative
Inequality. $\frac{x^3}{1+9y^2xz}+\frac{y^3}{1+9z^2yx}+\frac{z^3}{1+9x^2yz} \geq \frac{(x+y+z)^3}{18}$
Show that the polynomial $x^3-xy^2+y+1$ is irreducible in $\mathbb{Q}$
Can you prove that $\frac{1}{n}+\frac{1}{n^2}+\frac{1}{n^3}… = \frac{1}{n-1}$?
An elementary way to show any bounded subset of $\Bbb{R}^k$ is totally bounded
Another property of the function $\phi : p \mapsto \int |f|^p d\mu$.
Why is Gödel's Second Incompleteness Theorem important?
Please suggest a functional analysis book to refresh my knowledge
How to show that a root of the equation $x (x+1)(x+2) … (x+2009) = c $ can have multiplicity at most 2?
Number theory problem – show $N$ is a square when…
euclidean geometry books…
Openness condition in Seifert-van Kampen Theorem
Showing the square of a Markov process is or isn't Markov
bounded vs. unbounded linear programs
How to learn commutative algebra?

Let $V(\mathbb{R})$ be the vector space of $m\times n$ real matrices such that each row sum and each column sum is zero. What is the dimension of $V(\mathbb{R})$? I know by General thinking that its dimension is $(m-1)(n-1)$. But I don’t know what is the method to find its dimension. Please tell me how to think about its dimension. Thanks a lot.

- Analogy between linear basis and prime factoring
- Transformation Matrix representing $D: P_2 \to P_2$ with respect to the basis $B$.
- Prove this consequence of Cramer's theorem.
- What's the condition for a matrix $A$ ($2N\times 2N$ dimension) to have eigenvalues in pairs $\pm\lambda$?
- incidence matrix of a digraph with a self loop
- Does the set of matrix commutators form a subspace?
- Pseudoinverse and norm
- Proof of Vandermonde Matrix Inverse Formula
- Efficient Algorithm for Generalized Sylvester's Equation
- How to intuitively understand eigenvalue and eigenvector?

Let $K$ be a field. For $A \in \mathrm{M}(m \times n, K)$ let

$$

R_i(A) = \sum_{j=1}^n A_{ij} \quad \text{for every $1 \leq i \leq m$}

$$

and

$$

C_j(A) = \sum_{i=1}^m A_{ij} \quad \text{for every $1 \leq j \leq n$},

$$

and set

$$

V_{m,n}(K) = \{A \in \mathrm{M}(m \times n, K) \mid R_1(A) = \dotsb R_m(A) = C_1(A) = \dotsb = C_n(A)\}.

$$

We show that the map

\begin{align*}

\Phi \colon V_{m,n}(K) &\to \mathrm{M}((m-1) \times (n-1), K), \\

\quad (a_{ij})_{1 \leq i \leq n, 1 \leq j \leq m} &\mapsto (a_{ij})_{1 \leq i \leq n-1, 1 \leq j \leq m-1}

\end{align*}

is an isomorphism; it is clearly linear.

First **surjectivity**: Let $A = (a_{ij})_{1 \leq i \leq n-1, 1 \leq j \leq m-1} \in \mathrm{M}((m-1) \times (n-1), K)$. For all $1 \leq i \leq m-1$ let $a_{in} = -R_i(A)$ and for all $1 \leq j \leq n-1$ let $a_{mj} = -C_j(A)$. Also let

$$

a_{mn}

= \sum_{\substack{1 \leq i \leq m-1 \\ 1 \leq j \leq n-1}} a_{ij}.

$$

For $\hat{A} = (a_{ij})_{1 \leq i \leq n, 1 \leq j \leq m} \in \mathrm{M}(m \times n, K)$ we have that

$$

R_i(\hat{A}) = \sum_{j=1}^n a_{ij} = R_i(A) + a_{in} = 0

\quad \text{for every $1 \leq i \leq m-1$}

$$

as well as

\begin{align*}

R_m(\hat{A})

&= \sum_{j=1}^n a_{mj}

= \sum_{j=1}^{n-1} a_{mj} + a_{mn} \\

&= -\sum_{j=1}^{n-1} C_j(A) + a_{mn}

= -\sum_{j=1}^{n-1} \sum_{i=1}^{m-1} a_{ij} + \sum_{\substack{1 \leq i \leq m-1 \\ 1 \leq j \leq n-1}} a_{ij}

= 0.

\end{align*}

So all row sums of $\hat{A}$ are zero. Simililarly we find that all column sums of $\hat{A}$ are zero. So $\hat{A} \in V_{m,n}(K)$. Because $\Phi(\hat{A}) = A$ this shows the surjectivity of $\Phi$.

For the **injecitvity** we argue the other way around: For every $A \in V_{m,n}(K)$ we have $A_{in} = -R_i(\Phi(A))$ for every $1 \leq i \leq m-1$ and $A_{mj} = -C_j(\Phi(A))$ for every $1 \leq j \leq n-1$, as well as

$$

A_{mn}

= -\sum_{j=1}^{n-1} A_{mj}

= \sum_{j=1}^{n-1} C_j(\Phi(A)),

$$

So $A$ is uniquely determined by $\Phi(A)$, showing that $\Phi$ in injective.

Take variable $x_{ij},\ i=1,2,\ldots,m,\ j=1,2\ldots,n$ that correspod to the entries of the matrix. Row sum being eeuqal to zero gives rise to the conditions $s\sum_{j=1}^n x_{1j}=0, \sum_{j=1}^n x_{2j}=0, \ldots$. Similarly one has to get equations corresponding to columns sums being zero. Now consider the rank of this system on $mn$ variables and $m+n$ equations.

Now use rank and nullity theorem.

A variant on Jendrik Stelzner’s excellent answer is to note that if a matrix $a$ has row and column sum zero, then for each $i$

$$

a_{i,n} = – \sum_{s=1}^{n-1} a_{i,s},

$$

for each $j$

$$

a_{m,j} = – \sum_{t=1}^{m-1} a_{t,j},

$$

and finally

$$

a_{m,n} = \sum_{s=1}^{n-1} \sum_{t=1}^{m-1} a_{t, s}.

$$

Conversely, these three conditions imply that $a$ has row and column sum zero.

Now it is immediate that a basis of the space of this matrices is given by the $(m-1)(n-1)$ matrices

$$

e_{i,j} – e_{i,n} – e_{m, j} + e_{m, n},

$$

for $0 \le i < m$, $0 \le j < n$, where $e_{s, t}$ is the usual matrix which has zero everywhere except for a $1$ in position $s, t$. The above show that they span the space, and it suffices to look at the first $n-1$ and $m-1$ rows to see that they are independent.

For instance, when $m = n = 3$ you get

$$

\begin{bmatrix}

1 & 0 & -1\\

0 & 0 & 0\\

-1 & 0 & 1\\

\end{bmatrix},

\quad

\begin{bmatrix}

0 & 0 & 0\\

1 & 0 & -1\\

-1 & 0 & 1\\

\end{bmatrix},

\quad

\begin{bmatrix}

0 & 0 & 0\\

0 & 1 & -1\\

0 & -1& 1\\

\end{bmatrix},

\quad

\begin{bmatrix}

0 & 1 & -1\\

0 & 0 & 0\\

0 & -1& 1\\

\end{bmatrix}.

$$

Note that if you look at the first two rows and columns, you get the usual base of the space of $2 \times 2$ matrices.

consider any one row. since row sum is zero, so at least one component of that row can be written as the linear combination of other components. so that one element can be removed or can be replaced by zero. do this process for each row.

Similarly, do this process for each column. you will find finally (m-1).(n-1) remaining elements. so the dimension is (m-1).(n-1).

- Number of variables and dimension of a function
- Elementary proof of Zsigmondy's theorem
- A and B is similar ⇒ $A^T$ is similar to $B^T$.
- How to solve this inequality.
- Find a formula in terms of k for the entries of Ak, where A is the diagonalizable matrix:
- Ratio test with limsup vs lim
- Group of order $p^{n}$ has normal subgroups of order $p^{k}$
- Euler function $\phi(mn) = \phi(n)$
- Sum of eigenvalues and singular values
- When do the multiples of two primes span all large enough natural numbers?
- In combinatorics, how can one verify that one has counted correctly?
- Relating the Künneth Formula to the Leray-Hirsch Theorem
- Proving connectedness of the $n$-sphere
- Bounded operators that are not closed.
- Intuition behind Snake Lemma