The history of computing hardware starting at 1960 is marked by the conversion from vacuum tube to solid-state devices such as transistors and then integrated circuit (IC) chips. Around 1953 to 1959, discrete transistors started being considered sufficiently reliable and economical that they made further vacuum tube computers uncompetitive. Metal–oxide–semiconductor (MOS) large-scale integration (LSI) technology subsequently led to the development of semiconductor memory in the mid-to-late 1960s and then the microprocessor in the early 1970s. This led to primary computer memory moving away from magnetic-core memory devices to solid-state static and dynamic semiconductor memory, which greatly reduced the cost, size, and power consumption of computers. These advances led to the miniaturized personal computer (PC) in the 1970s, starting with home computers and desktop computers, followed by laptops and then mobile computers over the next several decades.
List of transistorized computers and Transistor computer
For the purposes of this article, the term "second generation" refers to computers using discrete transistors, even when the vendors referred to them as "third-generation". By 1960 transistorized computers were replacing vacuum tube computers, offering lower cost, higher speeds, and reduced power consumption. The marketplace was dominated by IBM and the seven dwarfs:
IBM
The BUNCH
Burroughs
UNIVAC
NCR
Control Data Corporation (CDC)
Honeywell
General Electric
RCA.
Some examples of 1960s second generation computers from those vendors are:
the IBM 1401, the IBM 7090/7094, and the IBM System/360;
the Burroughs 5000 series;
the UNIVAC 1107;
the NCR 315;
the CDC 1604 and the CDC 3000 series;
the Honeywell 200, Honeywell 400, and Honeywell 800;
the GE-400 series and the GE-600 series;
the RCA 301, 3301 and the Spectra 70 series.
However, some smaller companies made significant contributions. Also, towards the end of the second generation Digital Equipment Corporation (DEC) was a serious contender in the small and medium machine marketplace.
This page is automatically generated and may contain information that is not correct, complete, up-to-date, or relevant to your search query. The same applies to every other page on this website. Please make sure to verify the information with EPFL's official sources.
Driven by the demand for real-time processing and the need to minimize latency in AI algorithms, edge computing has experienced remarkable progress. Decision-making AI applications stand out for their heavy reliance on data-centric operations, predominantl ...
Covers the historical evolution and theoretical foundations of Information, Calculus, and Communication, exploring topics such as computing roots, technology acceleration, and system architecture.
Home computers were a class of microcomputers that entered the market in 1977 and became common during the 1980s. They were marketed to consumers as affordable and accessible computers that, for the first time, were intended for the use of a single nontechnical user. These computers were a distinct market segment that typically cost much less than business, scientific or engineering-oriented computers of the time such as those running CP/M or the IBM PC, and were generally less powerful in terms of memory and expandability.
The first microprocessors were designed and manufactured in the 1970s. Intel's 4004 of 1971 is widely regarded as the first commercial microprocessor. Designers predominantly used MOSFET transistors with pMOS logic in the early 1970s, switching to nMOS logic after the mid-1970s. nMOS had the advantage that it could run on a single voltage, typically +5V, which simplified the power supply requirements and allowed it to be easily interfaced with the wide variety of +5V transistor-transistor logic (TTL) devices.
The history of computing hardware covers the developments from early simple devices to aid calculation to modern day computers. The first aids to computation were purely mechanical devices which required the operator to set up the initial values of an elementary arithmetic operation, then manipulate the device to obtain the result. Later, computers represented numbers in a continuous form (e.g. distance along a scale, rotation of a shaft, or a voltage). Numbers could also be represented in the form of digits, automatically manipulated by a mechanism.
Drawing from a fieldwork conducted at COMPUTEX Taipei, one of the largest computer expo in the world, this contribution proposes to zoom-in at the level of Graphical Processing Units (GPU) manufacturers and their interactions with computer hardware hobbyis ...
Quantum random number generators (QRNGs) are a burgeoning technology used for a variety of applications, including modern security and encryption systems. Typical methods exploit an entropy source combined with an extraction or bit generation circuit in or ...