Theoretical Math: Proving an injection when it's countably infinite.

mmilton
Messages
8
Reaction score
0

Homework Statement



Let f: A --> B be an injection and suppose that the set A is countably infinite; how can I prove that there is an injection from B to A if and only if B is countably infinite?

Also, if we would suppose that A is uncountable, can B be countable?


Homework Equations





The Attempt at a Solution



Here is what I have thus far,

First direction:
Suppose B is countably infinite. Then, by definition, there is a
bijection from B to the naturals. Since A is also countably infinite, there is a bijection
from A to the naturals, and hence a bijection between B and A (and hence injection from B to A).

Next direction:
First show that since f is an injection of a countably infinite set to B, then B must be infinite.
Now, if there is an injection from B to A, then there is a bijection from B to a subset of A, call this subset S. So B and S have the same cardinality. But any infinite subset S of a countably infinite set A is countably infinite, so B has the same cardinality as a countably infinite set S.
 
Physics news on Phys.org
For your first direction, I would say instead that since A and B can be put into a 1-1 correspondence with \mathbb{N} we can map the first element of B which is mapped to 1 to the element in A which is mapped to 1 and so forth. Hence we have an injective map from B to A. You don't need bijection since you are only asked about injection.

We are given f is injective from A to B, A\mapsto\mathbb{N} i.e. A is countable infinite, and g:B->A is injective. So since f is injective every element of A maps to an element of B. However, there could be more elements in B but we have that g is also injective so every element of B maps to an element of A and A is countable infinite so B must be what?

I think these are pretty much just straight forward. I don't think you need to worry about a subset for the second one.
 
There are two things I don't understand about this problem. First, when finding the nth root of a number, there should in theory be n solutions. However, the formula produces n+1 roots. Here is how. The first root is simply ##\left(r\right)^{\left(\frac{1}{n}\right)}##. Then you multiply this first root by n additional expressions given by the formula, as you go through k=0,1,...n-1. So you end up with n+1 roots, which cannot be correct. Let me illustrate what I mean. For this...
Back
Top