User:J D Bowen/Math725 HW4

1) Let $V, W \ $ be finite-dimensional vector spaces. We aim to show that the following three statements are equivalent:
 * a) $\text{dim}(V)\geq \text{dim}(W) \ $
 * b) There is a surjective linear map $V \to W \ $.
 * c) There is an injective linear map $W \to V \ $.

(a)$\implies \ $ (b):

Suppose we have $\text{dim}(V)\geq \text{dim}(W) \ $. There exists basis sets for these spaces $\left\{{\vec{v_1}, \dots, \vec{v_m}}\right\}\subset V, \left\{{\vec{w_1}, \dots, \vec{w_n} }\right\} \subset W \ $, where $m=\text{dim}(V), n=\text{dim}(W) \ $. We have $m\geq n \ $.

Define a linear transformation $\phi:V\to W \ $ as $\sum_{i=1}^m a_i\vec{v_i} \mapsto \sum_{i=1}^n a_i\vec{w_i} \ $. Since $m\geq n \ $, every vector $\vec{w}\in W \ $ has a preimage in $V \ $. Hence $\phi \ $ is a surjection.

(b)$\implies \ $ (a):

Suppose there exists at least one surjective linear map $\phi:V\to W \ $. If we let $\left\{{\vec{v_1}, \dots, \vec{v_m}}\right\}\subset V \ $ be any basis for $V \ $, we can be sure that every vector in $V \ $ can be represented as $\vec{v} = \sum_{i=1}^m a_i\vec{v_i} \ $ for some constants $a_i \ $.

Since $\phi \ $ is linear, we have

$\phi(\vec{v}) = \sum_{i=1}^m a_i\phi(\vec{v_i}) \ $.

Since $\phi \ $ is surjective, for every $\vec{w}\in W \ $, there is at least one $\vec{v}\in V \ $ such that $\phi(\vec{v})=\vec{w} \ $. Therefore, for any $\vec{w}\in W \ $, there are constants $a_i \ $ such that

$\vec{w}=\sum_{i=1}^m a_i\phi(\vec{v_i}) \ $.

Hence the $\phi(\vec{v_i}) \ $ form a spanning set for $W \ $, implying that $\text{dim}(W)\leq \text{dim}(V) \ $.

(a)$\implies \ $ (c):

Suppose we have $\text{dim}(V)\geq \text{dim}(W) \ $. There exists basis sets for these spaces $\left\{{\vec{v_1}, \dots, \vec{v_m}}\right\}\subset V, \left\{{\vec{w_1}, \dots, \vec{w_n} }\right\} \subset W \ $, where $m=\text{dim}(V), n=\text{dim}(W) \ $. We have $m\geq n \ $.

Define a linear transformation $\phi:W\to V \ $ as $\sum_{i=1}^n a_i\vec{w_i} \mapsto \sum_{i=1}^n a_i\vec{v_i} \ $. Since $n\leq m \ $, every $\vec{v}\in V \ $ has at most one pre-image (1 or 0, depending on whether or not $\vec{v} \ $ is in the subspace $a_{n+1}=\dots=a_m =0 \ $ or not). Hence $\phi \ $ is an injection.

(c)$\implies \ $(a):

Suppose there exists at least one injective linear map $\phi:W\to V \ $. Let a basis for $W \ $ be $\left\{{\vec{w_1}, \dots, \vec{w_n} }\right\} \subset W \ $.

We have $\sum_{i=1}^n a_i\vec{w_i}=\vec{0} \implies (\forall i, \ a_i=0) \ $.

Since $\phi \ $ is linear, we have two important facts:


 * 1) $\phi(\vec{0_W})=\vec{0_V} \ $


 * 2) $\phi \left({\sum_{i=1}^n a_i\vec{w_i} }\right) = \sum_{i=1}^n a_i\phi(\vec{w_i}) \ $

Now, since $\phi \ $ is an injection, for $\vec{u},\vec{v}\in V, \ \vec{u}\neq\vec{v} \ $, we have $\phi(\vec{u})\neq\phi(\vec{v}) \ $. This, combined with fact 1, means that the  only  vector which maps to $\vec{0_V} \ $ is $\vec{0_W} \ $.

Now we can use fact 2 to investigate if the image of the basis vectors in $W \ $ is linearly independent in $V \ $:

$\sum_{i=1}^n a_i\phi(\vec{w_i}) = \vec{0_V} \implies \phi \left({\sum_{i=1}^n a_i\vec{w_i} }\right) = \vec{0_V} \implies \sum_{i=1}^n a_i\vec{w_i}=\vec{0_W} \implies (\forall i, \ a_i=0) $

Therefore, the $\phi(\vec{w_i}) \ $ are linearly independent in $V \ $, which means $\text{dim}(V)\geq n = \text{dim}(W) \ $.

work space
2) Let $U, W \ $ be subspaces of a vector space $V \ $. Define $r \ $ as the map taking any map from $V\to W \ $ to its restriction $U\to W \ $.

a) Let $\phi,\psi \ $ be any two linear maps $V\to W \ $. Then $\forall \vec{u}\in U, a,b\in\mathbb{C} \ $, we have

$r(a\phi+b\psi)(\vec{u}) = (a\phi+b\psi)(\vec{u})=a\phi(\vec{u})+b\psi(\vec{u})=a (r(\phi))(\vec{u})+b(r(\psi))(\vec{u}) \ $

since these two maps are linear. Since $r(\psi),r(\phi) \ $ can take no other values than vectors in $U \ $ this is the only case we need to consider. Hence, $r \ $ is linear.

b) Let $\sigma:U\to W \ $ be a linear map.

c)

end workspace
3) Let $S,T \ $ be linear operators on a vector space $V \ $.

a) Let $\vec{x}\in\text{ker}(T) \ $, that is, $T(\vec{x})=\vec{0} \ $. Then $ST(\vec{x})=S(T(\vec{x})) = S(\vec{0}) \ $.

But $S \ $ is linear, so $S(\vec{y})=S(\vec{0}+\vec{y})=S(\vec{0})+S(\vec{y}) \implies S(\vec{0})=\vec{0} \ $.

Hence $T(\vec{x})=\vec{0}\implies ST(\vec{x})=\vec{0} \ $, so $\text{ker}(T)\subseteq \text{ker}(ST) \ $.

b) Let $T:\mathbb{R}^2 \to \mathbb{R}^2 \ $ be the linear map defined $x\vec{e}_1+y\vec{e_2}\mapsto -y\vec{e}_1+x\vec{e_2} \ $, and let $S:\mathbb{R}^2\to\mathbb{R}^2 \ $ be defined $x\vec{e}_1+y\vec{e_2}\mapsto x\vec{e}_1 \ $. Then $\text{ker}(S)=\left\{{ c\vec{e_2}:c\in\mathbb{R} }\right\}, \text{ker}(ST)=\left\{{c\vec{e_1}:c\in\mathbb{R}}\right\} \ $, and so $\text{ker}(S)\not\subset\text{ker}(ST) \  $.

c) Suppose $\exists \vec{y}\in\text{Im}(ST) \ $.  Then $\exists \vec{x}:ST(\vec{x})=\vec{y} \ $, observe that

$ST(\vec{x})=S( T(\vec{x})) \ $

So there exists a $T(\vec{x}) \ $ such that $\vec{y}=S(T(\vec{x})) \ $. Hence, $y\in\text{Im}(S) \ $.

We have $\text{Im}(ST)\subseteq\text{Im}(S) \ $.

d) Let $T:\mathbb{R}^2 \to \mathbb{R}^2 \ $ be the linear map defined $x\vec{e}_1+y\vec{e_2}\mapsto x\vec{e}_1 \ $, and let $S:\mathbb{R}^2\to\mathbb{R}^2 \ $ be defined $x\vec{e}_1+y\vec{e_2}\mapsto -y\vec{e}_1+x\vec{e_2} \ $. Then $\text{Im}(ST) = \left\{{ c\vec{e_2}:c\in\mathbb{R}}\right\}, \text{Im}(T)=\left\{{ c\vec{e_1}:c\in\mathbb{R}}\right\} \ $, and we note $\text{Im}(ST)\not\subset\text{Im}(T) \ $

4) Let $\mathbf{P} \ $ be the space of polynomials over a field $\mathbf{F} \ $.

a) Let $P,Q \in \mathbf{P} \ $ be defined $P=p_m x^m + \dots +p_0, Q = q_n x^n + \dots + 0 \ $.

If we let the constants be zero for subscripts not summed over in the relevant polynomial, then observe that for the differential operator,

$d_x(aP+bQ)= d_x \left({ \sum_{i=0}^{\text{max}(m,n)} (ap_i+bq_i)x^i }\right) =  \sum_{i=0}^{\text{max}(m,n)}d_x\left({  (ap_i+bq_i)x^i }\right) =  \sum_{i=1}^{\text{max}(m,n)} i\left({  (ap_i+bq_i)x^{i-1} }\right)$

$=\sum_{i=1}^{\text{max}(m,n)} ap_iix^{i-1}+bq_iix^{i-1} = \sum_{i=1}^m ap_iix^{i-1} + \sum_{i=1} bq_iix^{i-1} = a\sum_{i=1}^m p_iix^{i-1} + b\sum_{i=1} q_iix^{i-1}= ad_x(P)+bd_x(Q) \ $

So, the differential operator is linear. Note that the two polynomials $x\mapsto 1, x\mapsto 2 \ $ both have derivative $x\mapsto 0 \ $, and so the differential operator is not injective. Note also that for any polynomial $P = p_m x^m + \dots +p_0 \ $, there is another polynomial $Q= (p_m/(m+1))x^{m+1} + \dots p_0 x \ $ such that $d_x Q = P \ $. Since $\mathbf{F}= \mathbb{R} \ $ or $\mathbb{C} \ $, we can be sure that $p_k/(k+1) \ $ is an element of $\mathbf{F} \ $. Hence the differential operator is surjective.

b) Fix $a\in\mathbf{F} \ $ and let $I_a:\mathbf{P}\to\mathbf{P} \ $ be

$f\mapsto \int_a^x f \ $

Observe that for field elements $r,s \ $, and polynomials $P, Q \ $, we have

$I_a(rP+sQ)=\int_a^x (rP+sQ)dx= \int_a^x rPdx+\int_a^x sQdx = r\int_a^xPdx+s\int_a^xQdx = rI_a(P)+sI_a(Q) \ $,

so $I_a \ $ is linear.

c) If we let $P=\sum_{i=0}^n p_i x^i \ $ be any polynomial, then note that

$\frac{d}{dx} \int_a^x P = \frac{d}{dx} \int_a^x \left({ \sum_{i=0}^n p_i x^i }\right) dx= \frac{d}{dx} \left({ \sum_{i=0}^n \frac{p_i}{i+1} x^{i+1} }\right) = \sum_{i=0}^n p_i x^i= P \ $

and so $I_a \ $ is a right inverse for $d_x \ $. Note, however, that for that same polynomial,

$\int_a^x \frac{d}{dt} P = \int_a^x\frac{d}{dt}  \left({ p_0+p_1t+\dots+p_nt^n }\right) dt= \int_a^x \left({ p_1+p_2t + \dots +np_nt^{n-1} }\right) $

$=\left[{ p_1t + p_2t^2+\dots+p_nt^n }\right]^x_a = p_1x + p_2x^2+\dots+p_nx^n + C \ $

where $C=-P(a) \ $ is some constant. Note that this means for some polynomial $P \ $ with constant term $p_0 \ $, the expression $I_a \ $ is a left inverse of the differential operator if and only if $P(a)=-p_0 \ $. Therefore, it is not possible for any one $I_a \ $ to be a left inverse of the differential operator for all polynomials.