A Factorization Algorithm
Consider the following true statement. Given natural numbers n, m, s
(1) [tex]\sum_{i=0}^n a^i = \sum_{i=0}^s a^{(m+1)i} \sum_{i=0}^m a^i[/tex] iff (2) [tex]n+1=(m+1)(s+1)[/tex] Now suppose that [tex]N[/tex] is a natural number where (3) [tex]N = \sum_{i=0}^n a^i [/tex] Now restrict [tex]a[/tex] to the natural numbers. Clearly for certain special numbers N we have a one to one map of factor pairs of [tex]N[/tex] and factor pairs of [tex]n+1[/tex] which is much smaller. In order for this thereom to give nontrivial factorizations [tex]n+1[/tex] must not be prime. If this is true and (3) holds then [tex]N[/tex] can be factored easily. If one thinks carefully about the algorithm for polynomial multiplication you can see that numbers defined by (3) above are quite special with respect to the base [tex]a[/tex]. In order to grasp the full implications of this theorem we need to think about two common algorithms. 1) The first is the base [tex]a[/tex] expansion algorithm used to write a natural number as a sum of powers of a. For example the number [tex]14[/tex] can be written as a sum of powers of [tex]3[/tex] as [tex]14 = 3^2+3+2[/tex]. In general this algorithm can be used to write (4) [tex] N_a = c_k a^k + c_{k1} a^{k1} + ... + c_1 a + c_0[/tex] Where some of the [tex]c_j[/tex] may be zero and all are less than [tex]a[/tex]. I will denote this algorithm by the capital letter [tex]\beta[/tex] so that [tex]\beta (N,a)[/tex] is equivalent to (4) above. 2) The second algorithm is polynomial multiplication. Now this algorithm proceeds by steps. We first multiply every term by every other term, then we collect terms. Denote this by [tex]P[/tex]. The difference between the result of this algorithm and the base [tex]a[/tex] expansion algorithm is best understood by looking closely at a third related algorithm. 3) The multiplication algorithm for natural numbers is really equivalent to polynomial multiplication taken one step further by carrying so that none of the digits are greater than the base and then evaluation of the resulting polynomial at the base value [tex]a[/tex]. Denote this carry step by [tex]C[/tex]. It follows that if [tex]N = ST[/tex] then [tex]\beta (N,a) = C(P(\beta (S,a),\beta (T,a)))[/tex] and under certain special conditions [tex]\beta(N,a)=P(\beta(S,a),\beta(T,a))[/tex]. One of these special conditions is the thereom stated above. This happens because of two things. First all coefficients of the summation terms on the right are equal to one in the base and the powers that do appear do not cause any duplication of powers, thus there is no carrying. In the posts that follow I will try to highlight the implications of this. One of the first is that though it would appear that we are looking at a very restricted class of factorizations, in reality a shift operator can often be applied to [tex]\beta(N,a)[/tex] to arrive at a new base where coefficients are all one and the theorem applies. This is typically possible when [tex]\beta(N,a)[/tex] is not the result of carrying, or more formally there exists a pair of factors of [tex]N[/tex] such that the base [tex]a[/tex] expansion and subsequent multiplication produces a polynomial in [tex]a[/tex] with precisely the same coefficients as the base [tex]a[/tex] expansion of [tex]N[/tex]. Clearly this all plays around in an uncomfortable way with our notions of variables and constants. However if you look at it carefully you will see that there is merit here. Furthermore there exist shift operators that can transform certain base [tex]a[/tex] expansions into the form given in (1) above. I will talk about those next. 
Shift Operators
Given natural numbers n, m, s
(1) [tex]\sum_{i=0}^n a^i = \sum_{i=0}^s a^{(m+1)i} \sum_{i=0}^m a^i[/tex] iff (2) [tex]n+1=(m+1)(s+1)[/tex] Now suppose we look let [tex]n=3[/tex] then [tex]n+1=4[/tex] the only nontrivial solution of (2) above is [tex]m=s=1[/tex]. So (1) above becomes (3) [tex]\sum_{i=0}^3 a^i = \sum_{i=0}^1 a^{2i} \sum_{i=0}^1 a^i[/tex] The expansion gives. (4) [tex]a^3+a^2+a+1 = (a^2+1)(a+1)[/tex] Thinking of the left hand side as a polynomial in [tex]a[/tex], for every natural number value of [tex]a[/tex] there is a corresponding natural number [tex]N[/tex] which is easily factored using (4) above. The following table lists the first seven numbers and there factors. [tex]\begin{array}{ cccc } a&N&S&T\\ \hline0&1&1&1\\ \hline1&4&2&2\\ \hline2&15&5&3\\ \hline3&40&10&4\\ \hline4&85&17&5\\ \hline5&156&26&6\\ \hline6&259&37&7\\ \hline7&400&50&8\\ \end{array}[/tex] The easiest related polynomial can be found using the values in the fourth column. Notice that the second differences of these values are all equal to [tex]2[/tex]. So if we wish a relationship where [tex]k=a1[/tex] we must solve the system [tex]f(k)=\int\int C_0 dk[/tex] [tex]f(0)=2[/tex] [tex]f(6)=50[/tex] [Note: the choice of initial conditions here is somewhat arbitrary, we only require that the resulting coefficients be greater than or equal to zero.] Solving this system is simple enough and yields [tex]f(k)=k^2+C_1k+C_0[/tex] [tex]f(0)=2[/tex] [tex]f(6)=50[/tex] [tex]C_0 = 2[/tex] [tex]C_1=2[/tex] Working this back into (4) above gives (5) [tex]k^3+4k^2+6k+4=(k^2+2k+2)(k+2)[/tex] Thinking about the base [tex]k[/tex] expansion algorithm suppose we have (6) [tex]\beta(N,k)=k^3+4k^2+6k+4[/tex] Then so long as [tex]k[/tex] is one of those special natural numbers for which [tex]\beta(N,k)=P(\beta(S,k),\beta(T,k))[/tex] for some factor pair of [tex]N[/tex] we can find a translation of the polynomial (5) which transforms it into the form (4). There is a vector space approach to this but we will not look at it now. In general we want to look at translations of (5) where [tex]k=as[/tex] for some integer [tex]s[/tex]. It should be clear that we can find desirable values for [tex]s[/tex] by solving [tex](as)^3+4(as)^2+6(as)+4=1[/tex] when [tex]a=0[/tex] that is (7) [tex]s^3+4s^26s+3=0[/tex] I used Excel to see that [tex]s=1[/tex] is the solution but Newton's Method is a good first start in general. So what does this suggest? Firstly there are deep and magnificent connections between the multiplicative structure of the natural numbers and that of the polynomials taken over the natural numbers. Important questions remain to be answered. I cannot claim to be able to answer them and perhaps some higher level of abstraction has already done so. One question that is important is this. Given a natural number [tex]N[/tex] does there always exist a base [tex]a[/tex] such that at least one factor pair of [tex]N[/tex] both have base [tex]a[/tex] expansions which multiply without the need to carry? If this is true then the ideas laid out here can factor any number. A second important question here is how fast are the algorithms determined by this construct. Clearly for (1) above we have the luxury of factoring a much smaller number than the one we started with. But in cases where we do not start with all of the coefficients equal to one, how long will it take us to find solutions to the translation equation (7)? Finally, if there exist composite natural numbers for which all factor pairs result in carrying for all possible bases is there an extension of the ideas here that describes efficient algorithms for undoing the carrying process? I will try to answer some of these questions as time goes on, right now I am hoping to find my first real job and looking for it is taking up most of my time. Cheers LAC 
A Reducibility Theorem
For every composite natural number [tex]N[/tex] there must be a reducible polynomial [tex]p(x)[/tex] over the natural numbers and a natural number [tex]a[/tex] for which [tex]N=p(a)[/tex] and at least one factor pair [tex]s[/tex] and [tex]t[/tex] of [tex]p[/tex] evaluated at [tex]a[/tex] satisfy [tex]N=p(a)=s(a)t(a)[/tex].
Proof Let [tex]\beta[/tex] be the base [tex]a[/tex] expansion algorithm, [tex]P[/tex] be polynomial multiplication, and [tex]C[/tex] be the carrying algorithm that is applied in standard multiplication of natural numbers over a base. Since [tex]N[/tex] is composite it has at least one nontrivial factor pair. It can be shown that [tex]\beta (N,a) = C(P(\beta(S,a),\beta(T,a)))[/tex] in all cases. Therefore [tex]C^{1}(\beta(N,a))[/tex] is a subset of the polynomials over the natural numbers evaluated at [tex]a[/tex] and contains only reducible polynomials. Clearly [tex]P(\beta(S,a),\beta(T,a))[/tex] must be an element of this set. 
This last thereom is somewhat deceptive because the carrying algorithm [tex]C[/tex] actually has many steps and possible paths so that there is actually a graph that must be searched. Although we generally proceed by carrying from smallest power to largest, when we start with the base [tex]a[/tex] expansion of [tex]N[/tex] we do not know which of the many possible reductions corresponds to the original polynomial multiplication coefficients in general. The point of (1) in the first post is that it guarantees that no carrying happens at all so [tex]C[/tex] is and identify operator under those circumstances. The complexity of [tex]C[/tex] and [tex]C_{1}[/tex] are both important generally. We can assume that a base [tex]a[/tex] expansion of any factor pair starts us off with cofficients that are less than [tex]a[/tex]. Consider the elementary example
[tex](b_1a+b_0)(c_1a+c_0)=b_1c_1a^2+(c_1b_0+b_1c_0)a+b_0c_0[/tex] If we have [tex]\beta(N,a)=C_2a^2+C_1a+C_0[/tex] equating coefficients gives the system [tex]C_2=b_1c_1[/tex] [tex]C_1=c_1b_0+b_1c_0[/tex] [tex]C_0=c_0b_0[/tex] This system is indeterminate having three equations and four unknowns. However if [tex]C_0=1[/tex] then by the rules of the game the system becomes [tex]C_2=b_1c_1[/tex] [tex]C_1=c_1+b_1[/tex] And is quite easily solvable. The shift operator sa mention above does precisely this. Also if [tex]C_0[/tex] is confirmed to be prime the system is easily solvable. But suppose that [tex]\beta(N,a)=C_3a^3+C_1a+C_0[/tex] Now the carrying algorithm must be inverted and the system that must be solved looks like [tex]C_3=r_2[/tex] [tex]C_2=b_1c_1+r_1ar_2[/tex] [tex]C_1=c_1b_0+b_1c_0+r_0ar_1[/tex] [tex]C_0=c_0b_0ar_0[/tex] Really three equations in five unknowns leaving a very uncomfortable two dimensional search ahead of us. Some advance can be made by noting that under the rules of the game [tex]c_0b_0 \leq (a1)^2[/tex] so that [tex]0 \leq r_0 \leq a1[/tex]. Furthermore chosing [tex]a[/tex] relatively prime to [tex]N[/tex] offers some advantages if the reduced residue system of [tex]a[/tex] is easily obtainable. Coming from the other side factoring [tex]C_2+ar_2[/tex] can also break the forutne cookie wide open so to speak. These arguments are know by every high school precal student. The more interesting set of questions centers around particular natural numbers. 
The Power of One
The preceding arguments show that the number one is our ally when it appears in certain places in an expansion. We like to see the number one appear in the leading and/or the trailing coefficient. It is also good see primes appear in the leading and trailing coefficients. So when does this happen? The trailing coefficient is actually calculable as [tex]N modulo a[/tex]. We need to solve the equation [tex] N \equiv 1 mod a[/tex].
Clearly [tex] N \equiv 1 mod N1[/tex] but no meaningful factors of [tex]N[/tex] can be expanded in this base. We need to find bases less than the square root of [tex]N[/tex]. But look suppose [tex]a_1[/tex] is such a base then [tex]N=a_1x+1[/tex] and is is quite clear that [tex]a_1N1[/tex]. So perhaps we should factor [tex]N1[/tex]? Assuming [tex]N[/tex] was odd we now have to find [tex]a_2(N1)/2 1[/tex]. In this fashion we can reduce the number to be factored until it is clearly trivial. Example: Factor 201 I first try to factor 200 by removing powers of two to obtain [tex]200=2^825[/tex]. now supposing that twentyfive is still too large I try to factor 24 by removing powers of two and obtain [tex]24=2^33[/tex]. So now I can write twentyfive in terms of any of the cators of twentyfour as [tex]{2*12+1, 4*6+1, 6*4+1, 8*3+1, 12*2+1}[/tex]. Because I want all factor pairs of 25 to be grater than whatever base I choose I consider those significantly less than the square root of twentyfive. So I set up the equation (1) [tex]25 = 2*12+1= 2z+1 = (2x+1)(2y+1)[/tex] This can be simplified to obtain (2) [tex] 12 =2xy+x+y[/tex] (3) [tex] y = \frac {12x} {2x+1}[/tex] Starting with [tex]x=1[/tex] and working upward I find that [tex]x=2[/tex] gives a whole number solution and twenty five is factored as five times five. So the factorization of 200 is [tex]200=2^35^2[/tex]. Applying this to the factorization of 201 I note that the square root of 201 is smaller than 15 but larger than 14. Set up the equation (4) [tex] 201 = 2*100+1 = 2z+1= (2x+1)(2y+1)[/tex] (5) [tex]y = \frac {100x} {2x+1}[/tex] Luckily for us w find that [tex]x=1[/tex] if a whole number solution. And 201 = 3*67. If we had gotten greedy and skipped the base two expansion we would have missed the factor three, or would we? Suppose w had chosen the base five, giving (6) [tex] 201 = 5*40+1 = 5z+1= (5x+s)(5y+t)[/tex] Now there are several pairs (s,t) such that st is congruent to one mod five (four pairs to be precise) they are {(4,4),(2,3),(3,2) (1,1)}. Since x and y remain free until later we can consider (2,3) and (3,2) to be the same choice. So picking at random we have only one third chance of getting it right in which case we also need to start with x=0 or we miss the boat all together. These difficulties suggest that we want to use bases that take care of the possible small prime factors right off or else simply divide by the first so many small primes. This is usually recommended even by fellows like Pollard. How is this related to the formulation that started the thread? Consider the following true statement. Given natural numbers n, m, s (1) [tex]\sum_{i=0}^n a^i = \sum_{i=0}^s a^{(m+1)i} \sum_{i=0}^m a^i[/tex] iff (2) [tex]n+1=(m+1)(s+1)[/tex] If you can, extend this summation to cover cases where [tex]\sum a^kb^i = \sum a^i \sum b^k[/tex] Remember the solution must resolve to (1) above when a=b. Good luck! 
Of course it's clear that
[tex]\sum_{i=0}^n (ab)^i = \sum_{i=0}^s (ab)^{(m+1)i} \sum_{i=0}^m (ab)^i[/tex] And it follows that [tex]\sum_{i=0}^n a^ib^i = \sum_{i=0}^s a^{(m+1)i}b^{(m+1)i} \sum_{i=0}^m a^ib^i[/tex] Implying that both a and b divide [tex]\sum_{i=1}^n a^ib^i = \sum_{i=1}^s a^{(m+1)i}b^{(m+1)i} \sum_{i=1}^m a^ib^i[/tex] So that ab must be factor of [tex]\sum_{i=0}^{n1} a^ib^i[/tex] If it is still true that n1+1 is composite then we can write [tex]\sum_{i=1}^{n1} a^ib^i = \sum_{i=1}^{s_1} a^{(m_1+1)i}b^{(m_1+1)i} \sum_{i=1}^{m_1} a^ib^i[/tex] iff [tex]n =(m_1+1)(s_1+1)[/tex] this reduction can continue to happen step by step as long as the necessary conditions exist. These things are very easy to see but I am looking for something more complicated I think. I could be wrong. That might seem a waste of time but it says we can build factorizations of larger numbers from factorizations of smaller numbers using (1) above so long as the upper limit of summation n and n+1 are both composite. This does not mean much for very small n but when n becomes larger the number of primes in any randomly chosen interval is much reduced. 
"Now there are several pairs (s,t) such that st is congruent to one mod five (four pairs to be precise) they are {(4,4),(2,3),(3,2) (1,1)}. Since x and y remain free until later we can consider (2,3) and (3,2) to be the same choice. So picking at random we have only one third chance of getting it right in which case we also need to start with x=0 or we miss the boat all together.
These difficulties suggest that we want to use bases that take care of the possible small prime factors right off or else simply divide by the first so many small primes. This is usually recommended even by fellows like Pollard." The sequence of bases that is formed as follows has special properties with regard to size of reduced residue systems locally in the natural numbers. (1a) [tex]\{\Pi_{i=1}^k p_i  k = 1 ... L\}[/tex] Here L is determined by the natural number N we wish to factor. This sequence has very special properties. Firstly the bases [tex]a[/tex] are products of the first k primes, so by first confirming that N and a are relatively prime we have cleared out those small factors. If it turns out other wise (the Euclidean Algorithm is fast by the way) we simply remove the common factor and start again. Because the reduced residue system is closed under multiplication modulo a we can factor the trailing coefficient of any base a coefficient in a finite number of steps simply by looking at the tables for that base. Consider base 30, where thirty is the product of 2,3 and 5. Suppose N is congruent to 1 mod 30 then any two factor pairs of N must be congruent to one of the following ordered pairs modulo thirty. (2a) {(1,1),(7,13),(11,11),(17,23),(19,19),(29,29)} There would be eight except symmetry reduces this slightly. The size of the reduced residue system for elements of (1a) above is simply to calculate as Euler's totient (3a) [tex]\phi(p_kp_{k1}...3*2)=\Pi_{i=1}^k (p_i1)[/tex] Furthermore one can look at the set (2a) as the kernel of the map (4a) [tex]mod 30 :rrs_{30} \otimes rrs_{30} \Rightarrow rrs_{30}[/tex] Interestingly we only need to store the kernel of this map in it's full form ordered by first entries as (5a) {(1,1),(7,13),(11,11),(13,7),(17,23),(19,19),(23,17),(29,29)} Because from this we can obtain the relevant factorization of any reduced residue modulo thirty by simply multiplying the first element of each ordered pair by the residue. Suppose rather than being congruent to one modulo thirty our natural number was congruent to seven mod thirty. Then we obtain the possible factorizations as mentioned (6a) {(7,1),(19,13),(17,11),(1,7),(29,23),(13,19),(11,17),(23,29)} Reordering and reducing by symmetry we get (7a) {(1,7),(13,19),(11,17),(23,29)} These coorespond to factorization equations for example as (8a) [tex]N=30z+7=(30x+13)(30y+19)[/tex] The search area here is calculable so the number of steps in the whole algorithm can be accurately estimated. Now the sequence in (3a) is also special because of the relationship between reduced residue systems of consecutive elements. Consider 6=2*3 and 30=2*3*5. It is clear that every element of the rrs mod thirty is relatively prime to six. So we can generate the rrs mod thirty from the rrs mod 6 as follows. Simply produce the table [tex]\begin{array}{ 1cccc } x/r&1&2&3&4\\ \hline1&7&13&19&25\\ \hline5&11&17&23&29\\ \end{array}[/tex] The elements are calculated as 6x+r. Clearly 25 is not in rrs mod thirty but we can quickly remove this in this case because it is not relatively prime to 5 which will be a factor of the new base (30). One must remain so we have quickly obtained the set rrs mod 30 ={1,7,11,13,17,19,23,29}. In general the elements of the array that must be removed can actually be generated from the rrs of the smaller base by simply multiplying the rrs by the new prime factor under standard multiplication. In the above example we want (9a) [tex]5\{1, 5\} = \{5,25\}[/tex] So in general using sequence (1a) (10a) [tex]rrs_{k+1}=\{x \Pi_{i=1}^k p_i+rrs_k x = 1...p_{k+1}1\}  \{p_{k+1}rrs_k\}[/tex] These arguments define a class of sieves that can approach the efficiency one would get if we knew all primes less than the square root of a natural number. The beauty of course is we do not need to know those primes. So we trade some of the work necessary to find those primes for a task we can do in a knowable number of steps. Some theorems that can be proved as excercises. 1) Show that [tex]\phi(p_kp_{k1}...3*2)<\phi(n) \forall n > p_kp_{k1}...3*2[/tex] 2) Show that if k is a square free natural number [tex]\phi(k)\leq\phi(n) \forall n > k[/tex] Notice that there are infinitely many 'ladders' of square free natural numbers that can be used to build sieves just as the sequence (1a) was used and that these sieves are 'locally' minimal as a choice of base for factoring natural numbers by these methods. The final step in this whole process is finding a way to put the approach looked at in the earlier posts of this thread and this last approach into one comprehensive theory. 
Quote:
Also, for your interest, [itex]\beta[/itex] is not a capital letter. Capital Beta looks just like the captial roman B. When using LaTeX, if you write a greek letter in lower case, like so \pi, then you'll get the lower case [itex]\pi[/itex] and if you write it like \Pi then you'll get [itex]\Pi[/itex]. Also, when using LaTeX in the same line as your text, use "itex" tags instead of "tex". So you'd write [ itex ]\phi ^2  \phi  1 = 0[ /itex ] to have the markup right in line [itex]\phi ^2  \phi  1 = 0[/itex]. To have it bigger and on a separate line: [ tex ]\phi ^2  \phi  1 = 0[ /tex ] [tex]\phi ^2  \phi  1 = 0[/tex] Quote:
Let N be composite, say N = nm for n, m > 1. If you define 0 to be a natural, let a be any natural less than or equal to min{n,m}. Otherwise, let a be any natural strictly less than min{n,m}. Let p(x) = (x + (na))(x + (ma)). I'm in a rush right now, but I'll look at posts 2, 4, and 5 later. 
Thanks AKG for the Latex help. The thereom itself in the first post is correct. Whenever the conditions are met some subset of factor pairs of [itex]N[/itex] is mapped in a one to one and onto the full set of factor pairs of [itex]n+1[/itex]. I realize there is formatting that needs to be done to these posts. I will get to it as time goes on. But I do see your response being similar to others I have seen about the first theorem. It is not at all the way number theory teaches us to think about factorization. But in fact for special [itex]N[/itex] it accomplishes the desirable task of lettings us try to factor something much smaller. The reducibility theorem does not say that every composite [itex]N[/itex] can be put into a form satisfying the first theorem. Rather, is simply says we can know that there must exist a base for which the expansion when thought of as a polynomial in a is reducible. You'll see what I mean if you look at all the posts together, but I warn you I have never seen anyone take the time so far to think through these things carefully. The whole project is much more intricate than it first appears. Later I will start doing posts where I use these ideas to shed light on results in elementary number theory. It gets very interesting. You see until now numbe theory is mostly treated as collection of almost unrelated theories. Very few books try to identify unifying principles.
Oh here are some important and little understood questions. The distribution of primes is known to satisfy the Prime Number Theorem, this is a limit theorem. For any finite [itex]n[/itex], [itex]\pi(n)[/tex] can only be approximated at best by the real valued functions currently used. In reality the number of primes seems to jump around [itex]\frac {x} {log x}[/itex]. Distances between primes also seem to be randomly distributed but generally beomce larger as we go further out. The facts are that the distribution of primes is formed in complement with the multiplication tables of natural numbers, the are a discrete phenomenon that cannot be precisely modeled using anything save perhaps a fourier series of some type. I know everyone is looking at Reimann and it is nice, but trying to stay within the descrete nature of the problem as much as possible is the avenue I am pursuing. Are you a person who can visualize things? Some math folks think in euqations. But visualize this. Every prime is the seed of a wave form that crosses the positive axis (n) only when p divides n in the strict natural number sense. Now if n=30 we are looking at the place on the positive axis where the wave forms of 2 3 and 5 are all zero at the same time. There is a purpose to what I am saying here. The prime numbers can only exist in wholes in the pattern. that is once a prime is discovered think of it as this wave form. Every place it touches the axis no prime can exist there. As we go further out more and more primes are discovered each with a unique period. These periodic waves 'punch wholes' in the number line every place n is composite. The places where no wave form touched the axis are prime. Amazingly we can reach forward a little bit so that we can actually find larger primes by knowing primes. Let me see if I can write this function down (Lagrange did something similar to this I think). (A) [tex]\Psi(n)= K \Pi_{i=1}^{k p_k<\sqrt n} \sin^2\frac {n\pi}{p_i}[/tex] Now clearly only those terms formed by prime divisors of n will be zero. K is a scaling factor. The function is positive definite. If you pick the set of primes ahead of time then you see that for instance if we use 2 3 and 5 (B) [tex]\Psi_{30}(n)=K\sin^2\frac {n\pi}{2} \sin^2\frac {n\pi}{3} \sin^2\frac {n\pi}{5}[/tex] [Note: Plot this for natural numbers n using Excel and notice the pattern that arises.] If you plot (B) for n between 7 and 49, every nonzero result is prime (Why is this true? it is true, but why?). if we add a term for eleven and plot for n between 11 and 121 same thing every nonzero value corresponds to prime n. We can do this but it saves us no time checking each one. What we want to do is start at seven, then use it to find every prime up to 49, now add terms for each of these. Now the interval of interst is 47 to 47^2 etc. you see that if we are good at this it gets faster. Still we are talking about checking every number apparently. However in a future post I will show how to use the modular concepts mentioned above to speed this puppy up massively, to aobtain a 'small' set of candidates fomr the set {p_k, ..., p_k^2} and to determine just how many numbers we need to look at to factor a number N. Here is how it works, for a number with only two large prime divisors, the single divisor that is less than the square root of N is relatively speaking quite close to root N. The class of functions defined in (A) above will require that I know all of the primes up to the fourth root of N or the root of root N. For your typical RSA numbers N may have 1000 digits. Root N will have something like 500 ish digits, and root root N will have about 250 digits. So to even try to attack an RSA we would need to know all the primes up to say 10^250, write out the rrs modulo the product of these primes in the neighbor hood of root N. (Of course this is impossible.) Then once we have applied the modular sieve, we test each of those numbers using the euclidean algorithm compared to reasonable sized products of those primes. We might even like to categorize intervals near root N using some of these probability methods which tell us likelihoods of a factor in those areas etc. In the posts that follow I will develop this algorithm. The beauty of it of course is that we can give upper bounds on time for the algorithm to complete. I am not at all claiming it is polynomial time but all that will have to be looked at. Using a distributed programming model over the internet I think time wise we might be quite surprised. NOTE: This thread is work in progress, I am using it as a scratchpad to investigate how a suite of observations and ideas I have had over the last ten years might relate to one another to produce factorization algorithms, and the usefulness of those algorithms. The over arching goal is to be true to the traditional confines of elementary number theory as much as possible, this is an ethic I am holding myself to mostly because I do not believe that the methods have been fully appreciated or developed in our mad rush to conquer the universe using ever more powerful forms of abstraction and functional reasoning. 
My attention span isn't what it's used to be. It's hard for me to follow your posts, Playdo, without some sort of summary or outline that describes the plot. :frown:
P.S. if we write N=PQ in base P, then we don't have to do any carrying when we multiply P and Q. 
Somehow, I missed your latest post when I wrote my previous reply!
If you're interested in the general problem if finding all the primes up to a given size, you might consider looking for info on wheel sieves  I think they're the current state of the art. If, to factor N, your method requires you to know all of the primes up to N^(1/4), then I'm sorry to say that in its current form, it is not useful for factoring a very large number. Any of the sieving algorithms will have factored N in much less time than it takes to simply write down that list of primes... let alone compute it! (Of course, I have that sneaky feeling that I've just delivered a setup line! :smile:) 
A few random comments:
Quote:
Quote:
Quote:
Quote:
Quote:
Why not implement this in some way to test it out? You'll have a much easier time convvincing people it's worthwhile if you have a working program. Or just write down the steps you'd use to factor a number completely in one cleanly formatted post. You don't have to explain the "why" of the steps, just give the bare bones algorithm. An estimate for the runtime would go a long way as well (you've mentioned a few places where you can count the steps please do it) 
[QUOTE=shmoe]
This is not an equality[\quote] The caveat is [itex]n+1=(m+1)(s+1)[/itex] in which case it definitely is an equality as [itex]a^ib^i = (ab)^i[/itex] so technically it simply starts to ask the implications of the base in (1) of the first post being composite. A general useful goal in factoring is finding a way to exchange factorization of a large number for the factorization of a smaller one. If you keep that in mind you can chain together any number of apparently unrelated algorithms and in whatever order, so long as at the link you can get back and forth (think in pictures). I admit its a bunch of scratch paper so far. It will improve over time. Your issue with the totient, your are right I made a mistake. I will fix that shortly. In fact primes have maximal totients compared to all numbers less than them. The result applies when [itex]\tau(k)>1[/itex]. That will be corrected. In fact it can be easily seen to be true by examining the formula for the totient function (You have to rewrite it, most books have it in a form where the terms are [itex](1\frac{1}{p_k})[/itex]. I will go back and clarify that when I get a chance. Over time I will work out examples. My understanding is that the two most popular sieve forms are modular sieves and quadratic sieves, the last two RSA numbers to be had fell to these and three months worth of massive parallel computing. The story line here is simple. It is an adventure, nothing more. But there is a specific problem that is being looked at using a few basic concepts. The issue is how are composites and primes related? Clearly no number can be prime unless the primes below its root fail to divide it. Why there would exist algorithms that are faster than this is of interest. Behind all of this is a question of complexity as Chaitin defines it. Eventually I will flesh this thread out, but you may have to wait for a while. I don't get paid to do this, it is a hobby that I started enjoying when I first became intersted in mathematics. Although I do know some of the latest in cryptography and have a good backgorund in number theory and math in general, I am not pursuing this in the interst of learning what 'they' say is possible. This is experimental mathematics done in the spirit of the early number theorists. We have a toolbox of mathematical tricks and a set of objects to be investigated. I will take pleasure in revealing a picture that is eventually understandable to people outside the ivory tower. The important thing for me is that this is experimental mathematics and it is recreational. I do not expect to win a fields medal here. I do it because I like it, and therefore I do it the way I want to do it. That said I know the difference between a meaningful result and triviality. Sometimes even things that appear trivial require something more then 'it's obvious that' which is so popular among mathematicians and have started to pervade our textbooks to the point that many proofs cannot be followed confidently even by those with great skill. IMHO you might as well not even include a proof if you are only going to give half of it. 
Quote:

Quote:
Certain classes of sieves use mappings that for special numbers are faster than if we just knew the number of primes less than root N. But in general sieves can only hope to approach the speed of knowing all the primes less than root N apriori. The point is that we do not know those primes less than root N for large N because they are very large. If we did knwo them we simply do the euclidean algorithm on N and the product of the primes less than root N. It takes much memory but the time involved can be stimated and we are done. But since for very large N we cannot know all the primes less than root N, we must find ways of either cleverly using more abstract spaces than the natural numbers under multiplication or we have to find methods for developing proxy sets for the set of primes less than N. Modular sieves do precisely this last. Notice that for the sequence of products of primes b[k] = 2, 6, 30, 210 ,2310 etc. [itex]\phi(b[k])k[/itex] is a decent approximation of the number of primes less than b[k] not already included as factors of b[k]. Now the theorem that starts this thread highlights certain very special natural numbers N for which a subset of divisors base a can be mapped onto the set of divisors of n+1 base 10. It would stand to reason that we might want to generalize this by thinking of the conditional [itex]n+1=(m+1)(s+1)[/itex] not simply in base ten but in any base whatever. So suppose we are looking at base a and base b, where in base b the condition holds. But how does summation work using indexes from other bases than ten? 
Quote:
[tex]\sum_{i=1}^n a^ib^i = \sum_{i=1}^s a^{(m+1)i}b^{(m+1)i} \sum_{i=1}^m a^ib^i[/tex] is still not an equality. If you meant the indicies to start at 0 and not 1, then it's an equality, but then your statement that a and b divide the above is not in general true. Quote:
Quote:
Quote:
Quote:

Quote:
The result is pi(p#)/phi(p#)>0 as p> infinity, not a good approximation at all. In vague words, most of the numbers less than p# are not prime because they have a prime divisor greater than p, these aren't removed when you sieve out by primes less than p. This is just a consequence of the exponential rate of growth of p# that makes the composites removed by sieving out by primes less than p somewhat inconsequential. You don't have to go far to see this approximation get bad. pi(11#)=pi(2310)=343 but phi(11#)5=475 pi(13#)=pi(30030)=3248 but phi(13#)6=5754 
Quote:
Quote:
There are O(N / lg N) primes less than the square root of N. Each of them is at least as big as 2. The product of those primes is 2^O(N / lg N). The number of iterations required by the Euclidean algorithm is on the order of the logarithm of the arguments. Thus, we require O(N / lg N) iterations. So, even if you were handed on a silver platter the product of the primes less than the root of N, and you had a special computer that was capable of performing the necessary arithmetic, you still can't beat the sieving algorithms when looking at very large numbers. (For reference, see Wikipedia for the running time of the GNFS) Incidentally, performing one giant GCD is not the most efficient way to trial divide by all of your primes  even if you got their product for free. I think you get the best performance when you partition your primes so that the product within each partition is roughly the same size as N. None of the high powered factoring algorithms are based on division at all: The Pollard rho method iteratively applies a polynomial map. When that map, viewed modulo some prime factor of N, starts cycling, that factor can be extracted through a gcd computation. (The elliptic curve method is better, and I think it does essentially the same thing on an elliptic curve, but I don't have reference material handy to make sure) The sieves are based on the fact that numbers that squaring is a 41 (or more) function modulo N. Given a pair of numbers that square to the same thing (and aren't negations of each other), you can generally extract a prime factor of N through a gcd computation. 
All times are GMT 5. The time now is 01:25 PM. 
Powered by vBulletin Copyright ©2000  2014, Jelsoft Enterprises Ltd.
© 2014 Physics Forums