first computer |
First Generation
With the onset of the Second World War, the countries involved in the war sought to develop computers to exploit their potential strategic importance computer. This increased funding to accelerate the progress of the development of computer and computer engineering. In 1941, Konrad Zuse, a German engineer to build a computer, the Z3, to design airplanes and missiles.
Party allies also made other progress in the development of computer power. In 1943, the British completed the secret code-breaking computer called Colossus to break the secret code used by Germany. The Colossus's impact affecting the development of the computer industry because of two reasons. First, the Colossus is not a versatile computer (general-purpose computer), it was only designed to decode secret messages. Second, the existence of these machines kept confidential until a decade after the war ended.
Work done by the Americans at that time produced some other advancement. Howard H. Aiken (1900-1973), a Harvard engineer working with IBM, succeeded in producing electronic calculators for the U.S. Navy. The calculator is a length of half a football field and has a range of 500 miles of cable. The Harvard-IBM Automatic Sequence Controlled Calculator, or Mark I, an electronic relay computer. He uses electromagnetic signals to move the mechanical components. The machine beropreasi with slow (it takes 3-5 seconds for each calculation) and inflexible (the order of calculations can not be changed). The calculator can perform basic arithmetic calculations and equations are more complex.
Another computer development at the present time is the Electronic Numerical Integrator and Computer (ENIAC), which is made by the cooperation between the governments of the United States and the University of Pennsylvania. Consisting of 18,000 vacuum tubes, 70,000 resistors and 5 million soldered joints, the computer is a machine that consume enormous power of 160kW.
This computer was designed by John Presper Eckert (1919-1995) and John W. Mauchly (1907-1980), ENIAC is a versatile computer (general purpose computer) that work 1000 times faster than Mark I.
In the mid-1940s, John von Neumann (1903-1957) joined the team of University of Pennsylvania in an effort to build a computer design concept that up to 40 years is still used in computer engineering. Von Neumann designed the Electronic Discrete Variable Automatic Computer (EDVAC) in 1945 with a memory to accommodate both programs or data. This technique allows the computer to stop at some point and then resume her job back. The key factor von Neumann architecture is the central processing unit (CPU), which allowed all computer functions to be coordinated through a single source. In 1951, UNIVAC I (Universal Automatic Computer I) made by Remington Rand, became the first commercial computer that uses the Von Neumann architecture model.
Both the United States Census Bureau and General Electric have UNIVAC. One of the impressive results achieved by the UNIVAC dalah success in predicting victory Dwilight D. Eisenhower in the 1952 presidential election.
First generation computers were characterized by the fact that operating instructions are made specifically for a particular task. Each computer has a different binary code program called "machine language" (machine language). This causes the computer is difficult to be programmed and the speed limit. Another feature is the use of first generation computer vacuum tube (which makes the computer at that time are very large) and magnetic cylinders for the storage of data.
Second Generation
In 1948, the invention of the transistor greatly influenced the development of computers. Transistors replaced vacuum tubes in televisions, radios, and computers. As a result, the size of electronic machinery has reduced drastically.
The transistor was used in computers began in 1956. In other findings in the form of magnetic core memory-assisting the development of second generation computers smaller, faster, more reliable, and more energy efficient than their predecessors. The first machine that utilizes this new technology is a supercomputer. IBM makes supercomputers, Stretch and Sprery-Rand makes a computer named LARC. These computers, which was developed for atomic energy laboratories, could handle large amounts of data, a capability that is needed by researchers atoms. The machine was very expensive and tend to be too complex for business computing needs, thereby limiting its popularity. There are only two LARC has ever installed and used: one at the Lawrence Radiation Labs in Livermore, California, and the other in the U.S. Navy Research and Development Center in Washington DC Second generation computers replaced machine language with assembly language. Assembly language is a language that uses abbreviations to replace the binary code.
In the early 1960s, began to appear successful second generation computers in business, in universities and in government. The computers of this second generation is fully computer using transistors. They also have components that can be associated with the computer at this time: a printer, storage, disk, memory, operating system and programs.
One important example of the computer at this time is 1401 which is widely accepted in the industry. In 1965, nearly all big businesses use computers second generation to financial memprosesinformasi.
The program stored in the computer and programming language that is in it gives flexibility to the computer. Flexibility is increased performance at a reasonable price for business use. With this concept, the computer could print customer purchase invoices and then run a product design or calculate payroll. Some programming languages began to appear at that time. Programming language Common Business-Oriented Language (COBOL) and FORTRAN (Formula Translator) came into common use. This programming language replaces the complicated machine with the words, sentences, and mathematical formulas are more easily understood by humans. This allows a person to program and manage the computer. A variety of emerging careers (programmer, systems analyst, and expert computer systems). Industr software also began to emerge and evolve during this second-generation computer.
Third Generation
Although the transistors in many respects the vacuum tube, but transistors generate considerable heat, which can potentially damage the computer's internal parts. Quartz stone (quartz rock) to eliminate this problem. Jack Kilby, an engineer at Texas Instruments, developed the integrated circuit (IC: integrated circuit) in 1958. IC combined three electronic components in a small silicon disc made of quartz sand. Scientists later managed to fit more components into a single chip called a semiconductor. As a result, computers became ever smaller as the components can be squeezed onto the chip. Other third-generation development is the use of the operating system (operating system) that allows the engine to run many different programs at once with a central program that monitored and coordinated the computer's memory.
Fourth Generation
After IC, the goal of development becomes more obvious: reduce the size of circuits and electrical components. Large Scale Integration (LSI) could fit hundreds of components on a chip. In the 1980's, the Very Large Scale Integration (VLSI) contains thousands of components on a single chip.
Ultra-Large Scale Integration (ULSI) increased that number into the millions. Ability to install so many components in a chip half the size of coins to encourage lower prices and the size of a computer. It also increased power, efficiency and reliability. Intel 4004 chip made in 1971membawa progress in IC by putting all the components of a computer (central processing unit, memory, and control input / output) in a very small chip. Previously, the IC is made to do a certain task specific. Now, a microprocessor can be manufactured and then programmed to meet all the requirements. Soon, every household devices like microwave ovens, televisions, and automobiles with electronic fuel injection (EFI) is equipped with a microprocessor.
Such developments allow ordinary people to use a regular computer. Computers no longer a dominance of large companies or government agencies. In the mid-1970s, computer assemblers offer their computer products to the general public. These computers, called minicomputers, sold with a software package that is easy to use by the layman. The most popular software at the time was word processing and spreadsheet programs. In the early 1980s, such as the Atari 2600 video game attracted the attention of consumers on a more sophisticated home computer and can be programmed.
In 1981, IBM introduced the use of Personal Computer (PC) for use in homes, offices, and schools. The number of PCs that use jumped from 2 million units in 1981 to 5.5 million units in 1982. Ten years later, 65 million PCs in use. Computers continued evolution towards smaller size, of computers that are on the table (desktop computer) into a computer that can be inserted into the bag (laptop), or even a computer that can be held (palmtop).
IBM PC to compete with Apple's Macintosh in getting the computer market. Apple Macintosh became famous for popularizing the graphical system on his computer, while his rival was still using a text-based computer. Macintosh also popularized the use of mouse devices.
At the present time, we know the journey IBM compatible with CPU usage: IBM PC/486, Pentium, Pentium II, Pentium III, Pentium IV (series of CPUs made by Intel). Also we know AMD K6, Athlon, etc.. This is all included in the class of fourth-generation computers.
Along with the proliferation of computer usage in the workplace, new ways to explore the potential to be developed. Along with the increased strength of a small computer, these computers can be connected together in a network to share a memory, software, information, and also to be able to communicate with each other. Computer networks allow computers to form a single electronic cooperation to complete a process task. By using direct wiring (also known as Local Area Network or LAN), or [cable telephone, this network can become very large.
Fifth Generation
Defining a fifth-generation computer becomes quite difficult because this stage is still very young. An example is the fifth generation computer imaginative fictional HAL9000 computer from the novel by Arthur C. Clarke titled 2001: Space Odyssey. HAL displays all the desired function from a fifth-generation computer. With artificial intelligence (artificial intelligence or AI), the HAL may have enough reason to do percapakan with humans, using visual feedback, and learn from his own experience.
Although it may be the realization of HAL9000 is still far from reality, many of the functions that had been established. Some computers can receive verbal instructions and are capable of imitating human reasoning. The ability to translate a foreign language also becomes possible. This facility seemed simple. However, such facilities become much more complicated than expected when programmers realized that human understanding relies heavily on context and understanding rather than just translate the words directly.
Many advances in the field of computer design and technology that increasingly allows the manufacture of fifth generation computers. Two engineering advances which are mainly parallel processing capability, which will replace the non-Neumann model. Non Neumann model will be replaced with a system capable of coordinating multiple CPUs to work in unison. Another advancement is the superconducting technology that permit the flow of electrically without any obstacles, which in turn can accelerate the speed of information.
Japan is a country known in the jargon of socialization and the fifth generation computer project. Institutions ICOT (Institute for New Computer Technology) was also formed to make it happen. Many newspapers stating that this project has failed, but some other information that the success of this fifth generation computer project will bring new changes to the paradigm of computerization in the world.
With the onset of the Second World War, the countries involved in the war sought to develop computers to exploit their potential strategic importance computer. This increased funding to accelerate the progress of the development of computer and computer engineering. In 1941, Konrad Zuse, a German engineer to build a computer, the Z3, to design airplanes and missiles.
Party allies also made other progress in the development of computer power. In 1943, the British completed the secret code-breaking computer called Colossus to break the secret code used by Germany. The Colossus's impact affecting the development of the computer industry because of two reasons. First, the Colossus is not a versatile computer (general-purpose computer), it was only designed to decode secret messages. Second, the existence of these machines kept confidential until a decade after the war ended.
Work done by the Americans at that time produced some other advancement. Howard H. Aiken (1900-1973), a Harvard engineer working with IBM, succeeded in producing electronic calculators for the U.S. Navy. The calculator is a length of half a football field and has a range of 500 miles of cable. The Harvard-IBM Automatic Sequence Controlled Calculator, or Mark I, an electronic relay computer. He uses electromagnetic signals to move the mechanical components. The machine beropreasi with slow (it takes 3-5 seconds for each calculation) and inflexible (the order of calculations can not be changed). The calculator can perform basic arithmetic calculations and equations are more complex.
Another computer development at the present time is the Electronic Numerical Integrator and Computer (ENIAC), which is made by the cooperation between the governments of the United States and the University of Pennsylvania. Consisting of 18,000 vacuum tubes, 70,000 resistors and 5 million soldered joints, the computer is a machine that consume enormous power of 160kW.
This computer was designed by John Presper Eckert (1919-1995) and John W. Mauchly (1907-1980), ENIAC is a versatile computer (general purpose computer) that work 1000 times faster than Mark I.
In the mid-1940s, John von Neumann (1903-1957) joined the team of University of Pennsylvania in an effort to build a computer design concept that up to 40 years is still used in computer engineering. Von Neumann designed the Electronic Discrete Variable Automatic Computer (EDVAC) in 1945 with a memory to accommodate both programs or data. This technique allows the computer to stop at some point and then resume her job back. The key factor von Neumann architecture is the central processing unit (CPU), which allowed all computer functions to be coordinated through a single source. In 1951, UNIVAC I (Universal Automatic Computer I) made by Remington Rand, became the first commercial computer that uses the Von Neumann architecture model.
Both the United States Census Bureau and General Electric have UNIVAC. One of the impressive results achieved by the UNIVAC dalah success in predicting victory Dwilight D. Eisenhower in the 1952 presidential election.
First generation computers were characterized by the fact that operating instructions are made specifically for a particular task. Each computer has a different binary code program called "machine language" (machine language). This causes the computer is difficult to be programmed and the speed limit. Another feature is the use of first generation computer vacuum tube (which makes the computer at that time are very large) and magnetic cylinders for the storage of data.
Second Generation
second computer |
In 1948, the invention of the transistor greatly influenced the development of computers. Transistors replaced vacuum tubes in televisions, radios, and computers. As a result, the size of electronic machinery has reduced drastically.
The transistor was used in computers began in 1956. In other findings in the form of magnetic core memory-assisting the development of second generation computers smaller, faster, more reliable, and more energy efficient than their predecessors. The first machine that utilizes this new technology is a supercomputer. IBM makes supercomputers, Stretch and Sprery-Rand makes a computer named LARC. These computers, which was developed for atomic energy laboratories, could handle large amounts of data, a capability that is needed by researchers atoms. The machine was very expensive and tend to be too complex for business computing needs, thereby limiting its popularity. There are only two LARC has ever installed and used: one at the Lawrence Radiation Labs in Livermore, California, and the other in the U.S. Navy Research and Development Center in Washington DC Second generation computers replaced machine language with assembly language. Assembly language is a language that uses abbreviations to replace the binary code.
In the early 1960s, began to appear successful second generation computers in business, in universities and in government. The computers of this second generation is fully computer using transistors. They also have components that can be associated with the computer at this time: a printer, storage, disk, memory, operating system and programs.
One important example of the computer at this time is 1401 which is widely accepted in the industry. In 1965, nearly all big businesses use computers second generation to financial memprosesinformasi.
The program stored in the computer and programming language that is in it gives flexibility to the computer. Flexibility is increased performance at a reasonable price for business use. With this concept, the computer could print customer purchase invoices and then run a product design or calculate payroll. Some programming languages began to appear at that time. Programming language Common Business-Oriented Language (COBOL) and FORTRAN (Formula Translator) came into common use. This programming language replaces the complicated machine with the words, sentences, and mathematical formulas are more easily understood by humans. This allows a person to program and manage the computer. A variety of emerging careers (programmer, systems analyst, and expert computer systems). Industr software also began to emerge and evolve during this second-generation computer.
Third Generation
third computer |
Although the transistors in many respects the vacuum tube, but transistors generate considerable heat, which can potentially damage the computer's internal parts. Quartz stone (quartz rock) to eliminate this problem. Jack Kilby, an engineer at Texas Instruments, developed the integrated circuit (IC: integrated circuit) in 1958. IC combined three electronic components in a small silicon disc made of quartz sand. Scientists later managed to fit more components into a single chip called a semiconductor. As a result, computers became ever smaller as the components can be squeezed onto the chip. Other third-generation development is the use of the operating system (operating system) that allows the engine to run many different programs at once with a central program that monitored and coordinated the computer's memory.
Fourth Generation
fourth computer |
After IC, the goal of development becomes more obvious: reduce the size of circuits and electrical components. Large Scale Integration (LSI) could fit hundreds of components on a chip. In the 1980's, the Very Large Scale Integration (VLSI) contains thousands of components on a single chip.
Ultra-Large Scale Integration (ULSI) increased that number into the millions. Ability to install so many components in a chip half the size of coins to encourage lower prices and the size of a computer. It also increased power, efficiency and reliability. Intel 4004 chip made in 1971membawa progress in IC by putting all the components of a computer (central processing unit, memory, and control input / output) in a very small chip. Previously, the IC is made to do a certain task specific. Now, a microprocessor can be manufactured and then programmed to meet all the requirements. Soon, every household devices like microwave ovens, televisions, and automobiles with electronic fuel injection (EFI) is equipped with a microprocessor.
Such developments allow ordinary people to use a regular computer. Computers no longer a dominance of large companies or government agencies. In the mid-1970s, computer assemblers offer their computer products to the general public. These computers, called minicomputers, sold with a software package that is easy to use by the layman. The most popular software at the time was word processing and spreadsheet programs. In the early 1980s, such as the Atari 2600 video game attracted the attention of consumers on a more sophisticated home computer and can be programmed.
In 1981, IBM introduced the use of Personal Computer (PC) for use in homes, offices, and schools. The number of PCs that use jumped from 2 million units in 1981 to 5.5 million units in 1982. Ten years later, 65 million PCs in use. Computers continued evolution towards smaller size, of computers that are on the table (desktop computer) into a computer that can be inserted into the bag (laptop), or even a computer that can be held (palmtop).
IBM PC to compete with Apple's Macintosh in getting the computer market. Apple Macintosh became famous for popularizing the graphical system on his computer, while his rival was still using a text-based computer. Macintosh also popularized the use of mouse devices.
At the present time, we know the journey IBM compatible with CPU usage: IBM PC/486, Pentium, Pentium II, Pentium III, Pentium IV (series of CPUs made by Intel). Also we know AMD K6, Athlon, etc.. This is all included in the class of fourth-generation computers.
Along with the proliferation of computer usage in the workplace, new ways to explore the potential to be developed. Along with the increased strength of a small computer, these computers can be connected together in a network to share a memory, software, information, and also to be able to communicate with each other. Computer networks allow computers to form a single electronic cooperation to complete a process task. By using direct wiring (also known as Local Area Network or LAN), or [cable telephone, this network can become very large.
Fifth Generation
fifth computer |
Defining a fifth-generation computer becomes quite difficult because this stage is still very young. An example is the fifth generation computer imaginative fictional HAL9000 computer from the novel by Arthur C. Clarke titled 2001: Space Odyssey. HAL displays all the desired function from a fifth-generation computer. With artificial intelligence (artificial intelligence or AI), the HAL may have enough reason to do percapakan with humans, using visual feedback, and learn from his own experience.
Although it may be the realization of HAL9000 is still far from reality, many of the functions that had been established. Some computers can receive verbal instructions and are capable of imitating human reasoning. The ability to translate a foreign language also becomes possible. This facility seemed simple. However, such facilities become much more complicated than expected when programmers realized that human understanding relies heavily on context and understanding rather than just translate the words directly.
Many advances in the field of computer design and technology that increasingly allows the manufacture of fifth generation computers. Two engineering advances which are mainly parallel processing capability, which will replace the non-Neumann model. Non Neumann model will be replaced with a system capable of coordinating multiple CPUs to work in unison. Another advancement is the superconducting technology that permit the flow of electrically without any obstacles, which in turn can accelerate the speed of information.
Japan is a country known in the jargon of socialization and the fifth generation computer project. Institutions ICOT (Institute for New Computer Technology) was also formed to make it happen. Many newspapers stating that this project has failed, but some other information that the success of this fifth generation computer project will bring new changes to the paradigm of computerization in the world.
|