## Fastest computer

there must be a theoretical limit of how fast a computer can be, since the maximum speed is the speed of light. so what would happen after that? Everyone will just buy one computer with the best stuff, i.e. graphics card (1*10^999999999999 MB) and ram, and pass it on to the next generations forever?

 PhysOrg.com science news on PhysOrg.com >> Heat-related deaths in Manhattan projected to rise>> Dire outlook despite global warming 'pause': study>> Sea level influenced tropical climate during the last ice age
 One thing you could do if you got a fastest processor is just start putting them in parallel to make yet faster clusters. In fact they could just keep adding this parallel capability within a single unit, so even if you got the fastest fabrication procedure you might still be able to parallize it indefinently.

 Originally posted by garytse86 there must be a theoretical limit of how fast a computer can be, since the maximum speed is the speed of light. so what would happen after that? Everyone will just buy one computer with the best stuff, i.e. graphics card (1*10^999999999999 MB) and ram, and pass it on to the next generations forever?

If a processor is processing at the speed of light then just use paralell ones forever until each one is computing the least possible smallest amount of information........

And bTW - no computer will ever calculate anything near the speed of light. It would cause the mass to be so much it would fall through the earth.

Recognitions:

## Fastest computer

I thought a quantum computer would calculate over the speed of light, simply because of the nature of quantum uhm, stuff.

I think there is a limit to hardrive space. But even then, just buy another harddrive.

 I don't see what you all mean by calculate at speed of light. What does this mean?
 how can something process faster than light? (is it even possible?)
 If we can slow down time in the future we could contain the light computer inside a time field after giving it something to do and it would calculate it faster relative to our time. taadaa.
 Another way to speed the computing process is to shorten the time it takes to input data. Right now computers are very fast, but it still can take hours to input data for a spread sheet or write a letter, or draw a building using CAD. I think that the future of computing is making the interface invisible to the user, by just talking, hand drawing, or thinking the commands. We have voice recognition now, but it still needs refinement; drawing pads are okay, but the computer has trouble deciding what it is seeing; and cache has the ability built in to anticipate which pieces of program to hold that will be used again soon, but this is still limited. The human to machine interface is where real leaps in speed still need to occur. Do we need a lightspeed computer when the average user can only type at twenty or thirty words a minute?
 Recognitions: Science Advisor I agree with ya, but I also think that the interface will come along with the technology. I do know that one of the major reasons why computers have problems interpreting analog information (drawing, face recognition, etc) is because computers process digital information. Life is analog, so unless we can figure out a way to build analog computers (us) that process on the same speed as current pcs, I'd say we will reach a limit in a short time as to just what a computer can do. And that, is why more power is needed. You know, voice technologies have increased like crazy in the past 5 years. Why? I'd say due to the massive jump in cpu speed, increase in fsb, better ram, etc. And I'm pretty sure Microsoft is working on technology to give total control to your computer through a mind link type device. Maybe it was on new scientist or something of the sort.
 I think the limit to be reached is not what the hardware is capable of - superconductors, qubits etc allow computers that are staggeringly powerful. The limit would be from us - what we want, what we can take, what information we can feed it. The fastest computer in the universe is just a heap of junk if you have no programs to run on it. And the output from computers would very likely exceed what the human body can experience. An excess of detail, perhaps?

 Originally posted by megashawn I agree with ya, but I also think that the interface will come along with the technology. I do know that one of the major reasons why computers have problems interpreting analog information (drawing, face recognition, etc) is because computers process digital information. Life is analog, so unless we can figure out a way to build analog computers (us) that process on the same speed as current pcs, I'd say we will reach a limit in a short time as to just what a computer can do. And that, is why more power is needed. You know, voice technologies have increased like crazy in the past 5 years. Why? I'd say due to the massive jump in cpu speed, increase in fsb, better ram, etc. And I'm pretty sure Microsoft is working on technology to give total control to your computer through a mind link type device. Maybe it was on new scientist or something of the sort.
Yes. I agree.

 I've read that scientist are working on making a Quantum Computer that can perforn a task in seconds that takes a supercomputer years to do. I don't know if that's anywhere near the speed of light, but it is pretty fast.
 Actually, I'd like a reply to grady's question too... how do you relate processing speeds (which is usually given in terms of processor clock cycles per second, Hz) to the speed of light? If you are talking about the carrier of information within the chips, then you would notice that the "speed" of computers has never changed - the distance travelled to go through each switch has. In the same way, c wouldn't be a practical barrier to computing, as it can be circumvented with advanced switchs and miniaturisation. Quantum computers don't get their speed from going over c. They do it because they don't have the practical weaknesses of silicon computing, and that qubits allow multiple data storage, and perhaps even fuzzy logic computing, allowing switches that are much more efficient.

 Originally posted by garytse86 there must be a theoretical limit of how fast a computer can be, since the maximum speed is the speed of light. so what would happen after that? Everyone will just buy one computer with the best stuff, i.e. graphics card (1*10^999999999999 MB) and ram, and pass it on to the next generations forever?
it defintly woulden go that high not becuase of physics but becuase we are too slow to notice like i really cant tell the diffrenace betwwen a 1.8 ghz and a 2.4 ghz

 i think the main limiting factor for traditional computer model will be the bus paradigm. no matter how fast the cpu is, no matter many cpus are connected in parallel, it will always bog down to how the data that is to be processed transported to the processor and the processed result returned for display/stored/etc. is there a way to make N-N connections amongs computer components? there are numerous tha have been studied (mostly in the family of hypercubes). however i imagine that someday (or someone has already implemented it?) the wired connections will be replaced by lasers (and further on by protein-based/biological computer?). the connections topology will be in 3-D where the restriction of overlapping electrical circuits will non-issue. switching from one connection to another will be performed with mems or other to-be invented optical switching methods. granted these computers will be much more complex than today's ones. but so is einstein's relativity (and later quantum mechanics) compared to newtonian mechanics.
 I would like to stress that there is no limit to the calculation speed of a computer, as the speed of processing is not related to the speed of light.

Mentor
 Originally posted by fffbone I would like to stress that there is no limit to the calculation speed of a computer, as the speed of processing is not related to the speed of light.
There may be no limit because of the speed of light, but there are limits for other reasons - like the size of the atom. There are only a certain number of transistors you can pack into a silicon chip. And then there's heat dissipation...

5 years is what I give it before we see a major slowdown in Moore's Law (which predicts doubling of the number of transistors and by implication computing power every 18 months).