Abstract algebra: monic gcd of polynomials in a subfield problem

rachellcb
Messages
10
Reaction score
0

Homework Statement


Let K \subseteq L be fields. Let f, g \in K[x] and h a gcd of f and g in L[x].

To show: if h is monic then h \in K[x].

The Attempt at a Solution



Assume h is monic.

Know that: h = xf + yg for some x, y \in K[x].
So the ideal generated by h, (h) in L[x] equals the ideal (f,g) in L[x]. Also, since K[x] is a principal ideal domain, (f,g)=(d) in K[x] for some d \in K[x], so d = af + bg for some a, b \in K[x]. So d is a gcd of f and g in K[x].

Now I'm not sure where to go... I know that h is monic and therefore the unique monic gcd of f and g in L[x], but not sure how this is useful. Do I need to show that h = d? How can I use the monic assumption to show this?

Thanks!
 
Physics news on Phys.org
Hmm, well, you the monic assumption is usually used to prove the uniqueness of the gcd. In this case, I'd say the monic assumption is useful since if you do not define the gcd to be monic, then h \left( x \right) a gcd of f \left( x \right) and g \left( x \right) means that r h \left( x \right) where r \neq 0 \in L is also a greatest common divisor of f \left( x \right) and g \left( x\right). Well, if you let h \left( x \right) be monic and r \in L but r \not \in K, then it is easy to see the r h \left( x \right) \not \in K \left[ x \right].

I believe you could let d \left( x \right) be monic as well and you'd want to show that d \left( x \right) = h \left( x \right).

I will premise this by saying that I'm not sure if this is the easiest way to do this problem, but suppose that h \left( x \right) and d \left( x \right) are the unique monic gcd of f \left( x \right) and g \left( x \right) in L \left[ x \right] and K \left[ x \right] respectively. Well, then it can be seen that d \left( x \right) is at least still a common divisor of f \left( x \right) and g \left( x \right) in L \left[ x \right], and d \left( x \right) = p \left( x \right) f \left( x \right) + q \left( x \right) g \left( x \right) where p \left( x \right), q \left( x \right) \in K \left[ x \right] \subset L \left[ x \right]. Can you use this and the fact that h \left( x \right) and d \left( x \right) both monic to show that h \left(x \right) = d \left( x \right)?
 
Thanks! I think I can,... does it involve using the fact that a gcd of f and g is the smallest degree polynomial that can be written as a linear combination of f and g, and that d, being a common factor of f and g in L[x], must divide h?
 
Yea, pretty much. I also believe you may have seen somewhere that a gcd is the only common divisor which can be written as a linear combination. And the monic condition means it is the unique monic gcd.

Then you've just shown that making the field bigger doesn't change the gcd.
 
There are two things I don't understand about this problem. First, when finding the nth root of a number, there should in theory be n solutions. However, the formula produces n+1 roots. Here is how. The first root is simply ##\left(r\right)^{\left(\frac{1}{n}\right)}##. Then you multiply this first root by n additional expressions given by the formula, as you go through k=0,1,...n-1. So you end up with n+1 roots, which cannot be correct. Let me illustrate what I mean. For this...
Back
Top