ArtsAutosBooksBusinessEducationEntertainmentFamilyFashionFoodGamesGenderHealthHolidaysHomeHubPagesPersonal FinancePetsPoliticsReligionSportsTechnologyTravel
  • »
  • Technology»
  • Computers & Software

History of Computers - Computer History Timeline

Updated on February 6, 2014

A computer is an electronic machine that accepts massive data, organises it and arrives at solutions as per the programme fed into it, and performs various other functions like word processing etc.

Computers promote efficiency by saving time and labour, by giving accurate solutions, and by quickly handling massive and complicated data that is difficult to handle manually. Computers are being used in defence, railways, airways, industry, research, administration, education, etc.

In the 19th century Charles Babbage, an English scientist and mathematician attempted to mechanise sequence of calculations in a machine, so that it would perform all the necessary operations in a predetermined sequence without the intervention of an operator. The machine designed by Babbage used card-boards and cards with holes punched in them, to introduce both instructions and data into machine. The machine was to perform the instructions dictated by the card holes automatically until an entire sequence of instructions is completed. Babbage succeeded in establishing the basic principles on which the modern computers are constructed.

First Generation: In 1937, Howard Aiken at Harvard, proposed to IBM that a machine could be constructed by using some parts and technique from the punch card machinery which could automatically sequence the operations and hence calculations can be performed. This machine when completed used a combination of electro-mechanical devices. The machine was in operation for some time, generating many tables of mathematical functions (particularly Bessel functions) and was used for trajectory calculations in World War II.

Second Generation: The replacement of large expensive (hot) vacuum tubes with small inexpensive, reliable, comparatively low heat dissipating transistors led to the second generation computers. The vacuum tube computers were known as first generation computers whereas the computers using transistors were called second generation computers.

Third Generation: By 1965, third generation computers were introduced. The third generation computers of this period began making heavy use of integrated circuits, in which many transistors and other components were fabricated and packed together in a single small container.

Fourth Generation: The manufacture of integrated circuits has become so advanced so as to incorporate hundreds of thousands of active components in a volume which is a very small fraction of an inch. This new technique was called large scale integration and very large scale integration. This has led to smaller size, lower cost, larger memory and ultra fast processing computers. Present day computers are fourth generation computers.

Fifth Generation: The developments in computer technology which are being used for "Artificial Intelligence" are being termed as fifth generation computers.


    0 of 8192 characters used
    Post Comment

    • profile image

      ashisa kumar meher 6 years ago

      thandks for providing the knowledge about of computer.

    • Kevin Hungerford profile image

      Kevin Hungerford 7 years ago from Florida

      It's definitely important to know about a device that we dedicate so much of our time to. Thanks.

    • profile image

      Chawla & Co Ltd 7 years ago

      Its very interesting and important to know, how computers made? In day to day work, computers, laptops, Notebooks etc are very essential for us.

      We really like your HUB and hoping, in future you will be writing some of more interesting concepts about computers.