Relation between Sparse Primes and Curse of Dimensionality?

In summary, the conversation discusses the relationship between primes and dimensions in a vector space. It is suggested that as more primes are discovered, the dimension of the number space increases, making it more likely for a number to be composed of older primes rather than a new prime. This is similar to the curse of dimensionality in which adding more unrelated vectors increases the dimension of the space. The idea of primes as basis vectors is also explored and it is proposed that the number of common primes between two numbers can be used as a metric for distance or correlation. However, it is noted that a different metric is needed to account for duplicates.
  • #1
FallenApple
566
61
Is there a relation between these concepts? We know that as we move down the number line, the primes become less common. This makes sense: as we get more and more numbers, they could be constituted using all the primes that came before early on and less likely to be constituted by all the recent primes.

Although the curse of dimensionality is more of an issue related to basis vectors. The curse is this: as more and more unrelated( i.e independent) vectors are added, the dimension of the spanned space goes up while each increment of data(i.e column vector) stays the same, hence it has to occupy more hyper volume for same increment of data. Hence the data becomes more sparse for each basis added.

This is analogous because each new number can be written as a multiplicative combination of older primes or it can be a new prime, with unknown chance. If there is a new prime, then this is analogous to adding a dimension(new building block) which could in principle have a non zero probability to build another number down the line.

But I don't know if we can think of a multiplicative building block in terms of a basis. Unless there is some way to code each prime in terms of 1s and 0s in vectors that somehow make the prime vectors orthogonal also add up to build composite numbers(in vector 0s and 1s form as well). Perhaps there's a way to use log transformations to map from multiplication to addition? But if we were to do this, how would we incorporate the randomness(chance of being prime or not) of the next prime into this picture since there isn't really a fixed probability of nth number being prime. Maybe we can use Bayesian Statistics to assign some weight to a prime based on past occurrences of primes?

In short, are the ideas analogous?
 
Last edited:
Mathematics news on Phys.org
  • #2
What about the primes as basis vectors? You don't get a proper vector space unless you include all rational numbers (apart from 0), but let's ignore that point.
1 is (0,0,0,...)
2 is (1,0,0,...)
3 is (0,1,0,...)
4 is (2,0,0,...)
60 is (2,1,1,0,0,...)
 
  • Like
Likes FallenApple
  • #3
mfb said:
What about the primes as basis vectors? You don't get a proper vector space unless you include all rational numbers (apart from 0), but let's ignore that point.
1 is (0,0,0,...)
2 is (1,0,0,...)
3 is (0,1,0,...)
4 is (2,0,0,...)
60 is (2,1,1,0,0,...)

Ah, so then since each number is now a vector of data composed of sums of basis vectors which themselves are vectors, the dimension of the number space keeps increasing as we keep discovering primes. So say we find that the n-th number is prime. Since the next number, n+1, after is not likely to have the recently discovered prime in it, but most likely made of much earlier ones. Maybe its just that for the next number, there is just so many more ways that the previous basis vectors can add to make it that it's just more likely to find a composite instead of a prime. That is probably the generative process here.

But if we look at cross sectional neighborhoods of the discrete natural number line, we can see the sparseness.

say that m>n, for some m and n that are natural numbers. Let's scan the number line left to right and add a extra basis to the basis set every time we discover a prime. This is for a fair comparison, looking at the dimension for the ith number only.

From this, we see that n is more likely to made of less basis vectors than m. Not always, but just more likely for the reasons described above. This is the same as saying that n is more likely to live in a lower dimensional space than m. So at least heuristically, from this argument, the neighborhood of m is more likely to be more sparse than the neighborhood of n. For m>>>n, this is even more likely.
 
Last edited:
  • #4
Also, since new prime numbers are just surrounded by redundant information made from previous basis primes, can we compress this data ,that is the natural number line, down to just the primes by removing the composites?

Notice something interesting. Since 1=(0,0,0,...), the entire notion of ordering is encoded on the prime basis vectors. So we do not get new info until a new prime pops up, because every time we use 1, its just adding the zero vector.
 
Last edited:
  • #5
We can even define an inner product to be the number of primes they have in common, multiplicity included.

6=2+3=(1,1,0,0,...)
15=3+5=(0,0,1,1,...)

<6,15>=( 2+3)(3+5)=<2,3>+<3,3>+<2,5>+<3,5>=0+1+0+0=1

This makes sense since the only prime factor in common between 6 and 15 is 3. So we can define the metric to be the number of common denominator that is prime. That is, the number of overlapping basis vectors.

So the notion of distance, or correlation, mutual information, or whatever, between two vectors on this space is just the number of primes they have in common.

With some math, we can verify that we have a complete inner product space over the field of integers(negatives correspond to division).
<1,1>=0. iff 1=(0,0,0,...) which is true. <10,3>=<2+5,3>=<2,3>+<5,3>=0+0=0 which is True and so on. So the axioms are satisfied, making this an inner product space.

Then with this, it makes sense that if we analyze two neighborhoods, once for a cluster centered at n and once for an equally sized cluster centered at m where m>n, we see that the amount of mutual information should be higher on average for the cluster around m just because there are more combinations of primes available, at least asmptotically. So for computations, if we think of the ith number as time(which is reasonable for a computer), we see that the entropy increases with time. For numbers, the entropy of the state space(cross section at a neighborhood) increases with number order.

I'm not sure how to get rid of the multiplicity though. The inner product as I have defined only retrieves overlapping information. I need another metric does not involve duplicates to give an honest assessment of increasing dimensionality for increasing n. <8,2>=<2+2+2 , 2>=<2,2>+<2,2>+<2,2>=3 but clearly there is redundant information. So we a metric that gives only the amount of unique basis vectors between two vectors. Maybe divide by sqrt(<8,8>)?
<8,2>/sqrt(<8,8>)=1. It seems to work.
 
Last edited:

1. What is the relation between sparse primes and curse of dimensionality?

The relation between sparse primes and curse of dimensionality is that both concepts are related to the distribution of numbers. Sparse primes refer to prime numbers that are relatively far apart from each other, while the curse of dimensionality refers to the difficulty of accurately representing and analyzing data in high-dimensional spaces.

2. How does the distribution of sparse primes affect the curse of dimensionality?

The distribution of sparse primes can exacerbate the curse of dimensionality because as the number of dimensions increases, the likelihood of encountering sparse primes also increases. This can lead to difficulties in accurately representing and analyzing data, as there may be gaps or lack of information in certain areas of the data.

3. Can the curse of dimensionality be avoided by using sparse primes?

No, the curse of dimensionality cannot be completely avoided by using sparse primes. While the distribution of sparse primes may help alleviate some of the effects of high-dimensional data, it does not completely solve the problem. Other techniques, such as dimensionality reduction and feature selection, are often necessary to effectively analyze high-dimensional data.

4. Are there any benefits to using sparse primes in data analysis?

Yes, there can be benefits to using sparse primes in data analysis. Sparse primes can help reduce the computational complexity of certain algorithms, making them more efficient to use on high-dimensional data. Additionally, using sparse primes can also help identify important features or patterns in the data that may have been overlooked when using other methods.

5. How can the curse of dimensionality be mitigated when working with high-dimensional data?

The curse of dimensionality can be mitigated by using various techniques such as dimensionality reduction, feature selection, and data preprocessing. These methods can help reduce the number of dimensions and focus on the most relevant features in the data, making it easier to analyze and interpret. Additionally, using specialized algorithms and techniques specifically designed for high-dimensional data can also help mitigate the effects of the curse of dimensionality.

Similar threads

Replies
3
Views
272
Replies
7
Views
2K
  • General Math
Replies
4
Views
3K
  • Engineering and Comp Sci Homework Help
Replies
1
Views
832
  • Calculus and Beyond Homework Help
Replies
1
Views
1K
  • Calculus and Beyond Homework Help
Replies
15
Views
953
  • Differential Geometry
Replies
0
Views
629
  • Special and General Relativity
5
Replies
146
Views
6K
  • Linear and Abstract Algebra
Replies
14
Views
5K
  • General Math
Replies
1
Views
1K
Back
Top