Moore's Law is generally known as the observation attributed to Gordon E. Moore (one of the co-founders of Intel, not to be confused with the philosopher G. E. Moore) that, at our rate of technological development and advances in the semiconductor industry, the number of transistors on integrated circuits (a rough measure of computer processing power) doubles every 18 months.
Other versions of Moore's law claim that RAM storage capacity increases at the same rate as processing power. However, memory speeds have not increased as fast as CPU speeds in recent years, leading to a heavy reliance on caching in current computer systems.
Moore first outlined the "law" in 1965. In its original form, it claimed that the number of components on lowest-cost semiconductor chips doubles roughly every 12 months. In 1975, Moore revised his estimate for the expected doubling time, arguing that it was slowing down to about two years. At the end of the 1970s, Moore's law became known as the limit for the number of transistors on the most complex chips.
Recent computer industry technology 'roadmaps' predict (as of 2001) that Moore's law will continue for several chip generations. Depending on the doubling time used in the calculations, this could mean up to 100 fold increase in transistor counts on a chip in a decade. The semiconductor industry technology roadmap uses a three-year doubling time for microprocessors, leading to about nine-fold increase in a decade.
Since the rapid exponential improvement could put 100 GHz personal computers in every home and 20 GHz devices in every pocket, some commentators have speculated that sooner or later computers will meet or exceed any conceivable need for computation. This is only true for some problems - there are others where exponential increases in processing power are matched or exceeded by exponential increases in complexity as the problem size increases. See computational complexity theory and complexity classes P and NP for a (somewhat theoretical) discussion of such problems, which occur very commonly in applications such as scheduling.
Although Moore's law has since the 1970s been defined in terms of the number of transistors on a chip, it is common to refer to Moore's law in reference to the rapid continuing advance in computing power per dollar cost.
A similar progression has held for hard disk storage available per dollar cost - in fact, the rate of progression in disk storage over the past 10 years or so has actually been faster than for semiconductors—although, largely because of production cost issues, hard drive performance increases have lagged significantly.
Extrapolation based on Moore's Law has led futurists such as Vernor Vinge and Bruce Sterling to speculate about a technological singularity. Historical analysis of Moore's law has, however, shown that its interpretations have qualitatively changed over the years and that it has not very accurately described developments in semiconductor technology. For a detailed study on Moore's law and its historical evolution, see 'The lives and death of Moore's law,' First Monday, November 2002, at http://firstmonday.org/issues/issue7-11/tuomi/. The effect of the operation of Moore's Law on computer component suppliers is profound. A typical major design project (such as an all-new CPU or hard drive) takes between two and five years to reach production-ready status: in consequence, component manufacturers face enormous timescale pressures: just a few weeks delay in a major project can spell the difference between great success and massive losses, even bankruptcy. Expressed as "a doubling every 18 months", Moore's Law suggests the phenomenal progress of technology in recent years. Expressed on a shorter timescale, however, Moore's Law equates to an average performance improvement in the industry as a whole of over 1% a week. For a manufacturer competing in the cut-throat CPU, hard drive or RAM markets, a new product that is expected to take three years to develop and is just two or three months late is 10 to 15% slower or smaller than the directly competing products, and is usually unsaleable.
Note that not all aspects of computing technology develop in capacities and speed according to Moore's Law. Random Access Memory speeds, and hard drive seek times improve at best at a few percentage points per year.
Another sometimes misunderstood point is that exponentially improved hardware does not necessarily imply exponentially-improved software to go with it. The productivity of software developers most assuredly does not increase exponentially with the improvement in hardware, but by most measures has increased only slowly and fitfully over the decades.