0% found this document useful (0 votes)
105 views

Ass in Computer

A computer is an electronic device that can be programmed to perform mathematical and logical operations automatically. It accepts digital data as input, processes it according to programmed instructions, and produces results as output. The key components of a computer are the central processing unit (CPU), memory, display adapters, and input/output devices. Computer hardware refers to the physical parts of a computer system, while software refers to encoded instructions for processing data. Over time, computers have become faster, more powerful and versatile.
Copyright
© © All Rights Reserved
Available Formats
Download as DOCX, PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
105 views

Ass in Computer

A computer is an electronic device that can be programmed to perform mathematical and logical operations automatically. It accepts digital data as input, processes it according to programmed instructions, and produces results as output. The key components of a computer are the central processing unit (CPU), memory, display adapters, and input/output devices. Computer hardware refers to the physical parts of a computer system, while software refers to encoded instructions for processing data. Over time, computers have become faster, more powerful and versatile.
Copyright
© © All Rights Reserved
Available Formats
Download as DOCX, PDF, TXT or read online on Scribd
You are on page 1/ 10

A computer is a device that accepts information (in the form

of digitalized data) and manipulates it for some result based on a program or


sequence of instructions on how the data is to be processed.

The housing is the part of your computer that holds all the parts that make the computer
operate. In a desktop tower, the housing does not include a monitor, keyboard or mouse.

The motherboard is the part of your computer that wires all of the other parts together. It is a
printed circuit board and includes the CPU and memory. Every component of your computer
connects directly into the motherboard. It typically includes expansion options for high-end
graphic displays and additional devices. Motherboard specifications are designed around which
CPU your computer uses. For example, if your computer uses an Intel processor, you need an
Intel-compatible motherboard.

The processor, or CPU, of a computer is basically the brains of the computer. The processor is
responsible for interpreting every code it receives from the other computer components, and
making it usable to your operating system. Though it's simply a single chip, the processor's
speed determines the overall speed of your computer. The CPU contains a control unit and the
Arithmetic Logic Unit, or ALU.

The hard drive part of your computer is also called the hard disk. Essentially, it is the part of the
computer that stores your information. If you upload a picture to your computer, it is stored on
the hard drive. Hard drive capacity is measure in megabytes, or MB, though some massive
computer store in terrabytes, or TB. The hard drive is set of disk platters that sit inside the
computer housing. Though other devices may store data, such as thumb drives, the hard drive
is resident in your computer and cannot be removed.

Whether a graphic card or part of the motherboard, the display adapter is the component of your
computer that interprets video signal. They define how frequently your monitor refreshes, how
many colors the display supports and the maximum resolutions. Gaming adapters are the most
high-end type; they offer 3D graphics and require significant power and cooling.

Random Access Memory, or RAM, is the part of your computer that is used as instant memory.
RAM is always in use as a temporary location to execute instructions from your computer. For
example, if you run many programs at once, you may need to upgrade your RAM because your
computer is using more memory. RAM comes in many different configurations, such as
Dynamic RAM, or DRAM, and SRAM, which runs at a higher speed.

An optical drive uses light to read and write data. It is the drive you use to burn CDs, DVDs and
Blu-Ray discs. Optical drives can be either slot loading, or tray loading and come in many
different configurations. Optical drives connect directly to the motherboard of the computer.
A computer is a device that can be instructed to carry out an arbitrary set
of arithmetic or logical operations automatically. The ability of computers to follow generalized
sequences of operations, called programs, enable them to perform a wide range of tasks.
Such computers are used as control systems for a very wide variety of industrial and consumer
devices. This includes simple special purpose devices like microwave ovens and remote controls,
factory devices such as industrial robots and computer assisted design, but also in general purpose
devices like personal computers and mobile devices such as smartphones. The Internet is run on
computers and it connects millions of other computers.
Since ancient times, simple manual devices like the abacus aided people in doing calculations. Early
in the Industrial Revolution, some mechanical devices were built to automate long tedious tasks,
such as guiding patterns for looms. More sophisticated electrical machines did
specialized analog calculations in the early 20th century. The first digital electronic calculating
machines were developed during World War II. The speed, power, and versatility of computers has
increased continuously and dramatically since then.
Conventionally, a modern computer consists of at least one processing element, typically a central
processing unit (CPU), and some form of memory. The processing element carries out arithmetic
and logical operations, and a sequencing and control unit can change the order of operations in
response to stored information. Peripheral devices include input devices (keyboards, mice, joystick,
etc.), output devices (monitor screens, printers, etc.), and input/output devices that perform both
functions (e.g., the 2000s-era touchscreen). Peripheral devices allow information to be retrieved
from an external source and they enable the result of operations to be saved and retrieved.

Computer software, or simply software, is a part of a computer system that consists of data or
computer instructions, in contrast to the physical hardware from which the system is built.

Computer hardware is the collection of physical components that constitute a computer system.
Computer hardware is the physical parts or components of a computer, such
as monitor, keyboard, computer data storage, graphic card, sound card, motherboard, and so on, all
of which are tangible objects.[1] By contrast, software is instructions that can be stored and run by
hardware.

Microsoft PowerPoint is a powerful presentation software developed by Microsoft. It is a standard


component of the company's Microsoft Office suite software, and is bundled together with Word,
Excel and other office productivity tools. The program uses slides to convey information rich in
multimedia. The term slide refers to the old slide projector, which this software effectively replaces.
History
Main article: History of computing hardware

Pre-20th century

The Ishango bone

Devices have been used to aid computation for thousands of years, mostly using one-to-one
correspondence with fingers. The earliest counting device was probably a form of tally stick. Later
record keeping aids throughout the Fertile Crescent included calculi (clay spheres, cones, etc.)
which represented counts of items, probably livestock or grains, sealed in hollow unbaked clay
containers.[3][4] The use of counting rods is one example.

The Chinese Suanpan () (the number represented on this abacus is 6,302,715,408)

The abacus was initially used for arithmetic tasks. The Roman abacus was developed from devices
used in Babylonia as early as 2400 BC. Since then, many other forms of reckoning boards or tables
have been invented. In a medieval European counting house, a checkered cloth would be placed on
a table, and markers moved around on it according to certain rules, as an aid to calculating sums of
money.
The ancient Greek-designed Antikythera mechanism, dating between 150 and 100 BC, is the world's oldest
analog computer.

The Antikythera mechanism is believed to be the earliest mechanical analog "computer", according
to Derek J. de Solla Price.[5] It was designed to calculate astronomical positions. It was discovered in
1901 in the Antikythera wreck off the Greek island of Antikythera, between Kythera and Crete, and
has been dated to circa 100 BC. Devices of a level of complexity comparable to that of the
Antikythera mechanism would not reappear until a thousand years later.
Many mechanical aids to calculation and measurement were constructed for astronomical and
navigation use. The planisphere was a star chart invented by Ab Rayhn al-Brn in the early 11th
century.[6] The astrolabe was invented in the Hellenistic world in either the 1st or 2nd centuries BC
and is often attributed to Hipparchus. A combination of the planisphere and dioptra, the astrolabe
was effectively an analog computer capable of working out several different kinds of problems
in spherical astronomy. An astrolabe incorporating a mechanical calendar computer[7][8] and gear-
wheels was invented by Abi Bakr of Isfahan, Persia in 1235.[9] Ab Rayhn al-Brn invented the first
mechanical geared lunisolar calendar astrolabe,[10] an early fixed-wired knowledge
processing machine[11] with a gear train and gear-wheels,[12] circa 1000 AD.
The sector, a calculating instrument used for solving problems in proportion, trigonometry,
multiplication and division, and for various functions, such as squares and cube roots, was
developed in the late 16th century and found application in gunnery, surveying and navigation.
The planimeter was a manual instrument to calculate the area of a closed figure by tracing over it
with a mechanical linkage.

A slide rule

The slide rule was invented around 16201630, shortly after the publication of the concept of
the logarithm. It is a hand-operated analog computer for doing multiplication and division. As slide
rule development progressed, added scales provided reciprocals, squares and square roots, cubes
and cube roots, as well as transcendental functions such as logarithms and exponentials, circular
and hyperbolictrigonometry and other functions. Aviation is one of the few fields where slide rules
are still in widespread use, particularly for solving timedistance problems in light aircraft. To save
space and for ease of reading, these are typically circular devices rather than the classic linear slide
rule shape. A popular example is the E6B.
In the 1770s Pierre Jaquet-Droz, a Swiss watchmaker, built a mechanical doll (automata) that could
write holding a quill pen. By switching the number and order of its internal wheels different letters,
and hence different messages, could be produced. In effect, it could be mechanically "programmed"
to read instructions. Along with two other complex machines, the doll is at the Muse d'Art et
d'Histoire of Neuchtel, Switzerland, and still operates.[13]
The tide-predicting machine invented by Sir William Thomson in 1872 was of great utility to
navigation in shallow waters. It used a system of pulleys and wires to automatically calculate
predicted tide levels for a set period at a particular location.
The differential analyser, a mechanical analog computer designed to solve differential
equations by integration, used wheel-and-disc mechanisms to perform the integration. In 1876 Lord
Kelvin had already discussed the possible construction of such calculators, but he had been stymied
by the limited output torque of the ball-and-disk integrators.[14] In a differential analyzer, the output of
one integrator drove the input of the next integrator, or a graphing output. The torque amplifier was
the advance that allowed these machines to work. Starting in the 1920s, Vannevar Bush and others
developed mechanical differential analyzers.

First computing device

A portion of Babbage's Difference engine.

Charles Babbage, an English mechanical engineer and polymath, originated the concept of a
programmable computer. Considered the "father of the computer",[15] he conceptualized and invented
the first mechanical computer in the early 19th century. After working on his revolutionary difference
engine, designed to aid in navigational calculations, in 1833 he realized that a much more general
design, an Analytical Engine, was possible. The input of programs and data was to be provided to
the machine via punched cards, a method being used at the time to direct mechanical looms such as
the Jacquard loom. For output, the machine would have a printer, a curve plotter and a bell. The
machine would also be able to punch numbers onto cards to be read in later. The Engine
incorporated an arithmetic logic unit, control flow in the form of conditional branching and loops, and
integrated memory, making it the first design for a general-purpose computer that could be
described in modern terms as Turing-complete.[16][17]
The machine was about a century ahead of its time. All the parts for his machine had to be made by
hand this was a major problem for a device with thousands of parts. Eventually, the project was
dissolved with the decision of the British Government to cease funding. Babbage's failure to
complete the analytical engine can be chiefly attributed to difficulties not only of politics and
financing, but also to his desire to develop an increasingly sophisticated computer and to move
ahead faster than anyone else could follow. Nevertheless, his son, Henry Babbage, completed a
simplified version of the analytical engine's computing unit (the mill) in 1888. He gave a successful
demonstration of its use in computing tables in 1906.

Analog computers
Sir William Thomson's third tide-predicting machine design, 187981

During the first half of the 20th century, many scientific computing needs were met by increasingly
sophisticated analog computers, which used a direct mechanical or electrical model of the problem
as a basis for computation. However, these were not programmable and generally lacked the
versatility and accuracy of modern digital computers.[18] The first modern analog computer was a tide-
predicting machine, invented by Sir William Thomson in 1872. The differential analyser, a
mechanical analog computer designed to solve differential equations by integration using wheel-and-
disc mechanisms, was conceptualized in 1876 by James Thomson, the brother of the more famous
Lord Kelvin.[14]
The art of mechanical analog computing reached its zenith with the differential analyzer, built by H.
L. Hazen and Vannevar Bush at MIT starting in 1927. This built on the mechanical integrators
of James Thomson and the torque amplifiers invented by H. W. Nieman. A dozen of these devices
were built before their obsolescence became obvious. By the 1950s the success of digital electronic
computers had spelled the end for most analog computing machines, but analog computers
remained in use during the 1950s in some specialized applications such as education (control
systems) and aircraft (slide rule).

Digital computers
It has been suggested that this section be split out into another article
titled Digital computer. (Discuss) (May 2017)

Electromechanical
By 1938 the United States Navy had developed an electromechanical analog computer small
enough to use aboard a submarine. This was the Torpedo Data Computer, which used trigonometry
to solve the problem of firing a torpedo at a moving target. During World War II similar devices were
developed in other countries as well.
Replica of Zuse's Z3, the first fully automatic, digital (electromechanical) computer.

Early digital computers were electromechanical; electric switches drove mechanical relays to
perform the calculation. These devices had a low operating speed and were eventually superseded
by much faster all-electric computers, originally using vacuum tubes. The Z2, created by German
engineer Konrad Zuse in 1939, was one of the earliest examples of an electromechanical relay
computer.[19]
In 1941, Zuse followed his earlier machine up with the Z3, the world's first
working electromechanical programmable, fully automatic digital computer.[20][21] The Z3 was built with
2000 relays, implementing a 22 bit word length that operated at a clock frequency of about 5
10 Hz.[22]Program code was supplied on punched film while data could be stored in 64 words of
memory or supplied from the keyboard. It was quite similar to modern machines in some respects,
pioneering numerous advances such as floating point numbers. Rather than the harder-to-implement
decimal system (used in Charles Babbage's earlier design), using a binary system meant that Zuse's
machines were easier to build and potentially more reliable, given the technologies available at that
time.[23] The Z3 was Turing complete.[24][25]
Vacuum tubes and digital electronic circuits
Purely electronic circuit elements soon replaced their mechanical and electromechanical
equivalents, at the same time that digital calculation replaced analog. The engineer Tommy Flowers,
working at the Post Office Research Station in London in the 1930s, began to explore the possible
use of electronics for the telephone exchange. Experimental equipment that he built in 1934 went
into operation five years later, converting a portion of the telephone exchange network into an
electronic data processing system, using thousands of vacuum tubes.[18] In the US, John Vincent
Atanasoff and Clifford E. Berry of Iowa State University developed and tested the AtanasoffBerry
Computer (ABC) in 1942,[26] the first "automatic electronic digital computer".[27] This design was also
all-electronic and used about 300 vacuum tubes, with capacitors fixed in a mechanically rotating
drum for memory.[28]

Colossus was the first electronicdigital programmable computing device, and was used to break German
ciphers during World War II.

During World War II, the British at Bletchley Park achieved a number of successes at breaking
encrypted German military communications. The German encryption machine, Enigma, was first
attacked with the help of the electro-mechanical bombes. To crack the more sophisticated
German Lorenz SZ 40/42 machine, used for high-level Army communications, Max Newman and his
colleagues commissioned Flowers to build the Colossus.[28] He spent eleven months from early
February 1943 designing and building the first Colossus.[29] After a functional test in December 1943,
Colossus was shipped to Bletchley Park, where it was delivered on 18 January 1944[30] and attacked
its first message on 5 February.[28]
Colossus was the world's first electronic digital programmable computer.[18] It used a large number of
valves (vacuum tubes). It had paper-tape input and was capable of being configured to perform a
variety of boolean logical operations on its data, but it was not Turing-complete. Nine Mk II Colossi
were built (The Mk I was converted to a Mk II making ten machines in total). Colossus Mark I
contained 1,500 thermionic valves (tubes), but Mark II with 2,400 valves, was both 5 times faster and
simpler to operate than Mark I, greatly speeding the decoding process.[31][32]

ENIAC was the first electronic, Turing-complete device, and performed ballistics trajectory calculations for
the United States Army.

The U.S.-built ENIAC[33] (Electronic Numerical Integrator and Computer) was the first electronic
programmable computer built in the US. Although the ENIAC was similar to the Colossus, it was
much faster, more flexible, and it was Turing-complete. Like the Colossus, a "program" on the
ENIAC was defined by the states of its patch cables and switches, a far cry from the stored
program electronic machines that came later. Once a program was written, it had to be mechanically
set into the machine with manual resetting of plugs and switches.
It combined the high speed of electronics with the ability to be programmed for many complex
problems. It could add or subtract 5000 times a second, a thousand times faster than any other
machine. It also had modules to multiply, divide, and square root. High speed memory was limited to
20 words (about 80 bytes). Built under the direction of John Mauchly and J. Presper Eckert at the
University of Pennsylvania, ENIAC's development and construction lasted from 1943 to full operation
at the end of 1945. The machine was huge, weighing 30 tons, using 200 kilowatts of electric power
and contained over 18,000 vacuum tubes, 1,500 relays, and hundreds of thousands of resistors,
capacitors, and inductors.[34]

Modern computers
Concept of modern computer
The principle of the modern computer was proposed by Alan Turing in his seminal 1936 paper,[35] On
Computable Numbers. Turing proposed a simple device that he called "Universal Computing
machine" and that is now known as a universal Turing machine. He proved that such a machine is
capable of computing anything that is computable by executing instructions (program) stored on
tape, allowing the machine to be programmable. The fundamental concept of Turing's design is
the stored program, where all the instructions for computing are stored in memory. Von
Neumann acknowledged that the central concept of the modern computer was due to this
paper.[36] Turing machines are to this day a central object of study in theory of computation. Except
for the limitations imposed by their finite memory stores, modern computers are said to be Turing-
complete, which is to say, they have algorithm execution capability equivalent to a universal Turing
machine.
Stored programs
A section of the Manchester Small-Scale Experimental Machine, the first stored-program computer.

Early computing machines had fixed programs. Changing its function required the re-wiring and re-
structuring of the machine.[28] With the proposal of the stored-program computer this changed. A
stored-program computer includes by design an instruction set and can store in memory a set of
instructions (a program) that details the computation. The theoretical basis for the stored-program
computer was laid by Alan Turing in his 1936 paper. In 1945 Turing joined the National Physical
Laboratory and began work on developing an electronic stored-program digital computer. His 1945
report "Proposed Electronic Calculator" was the first specification for such a device. John von
Neumann at the University of Pennsylvania also circulated his First Draft of a Report on the
EDVAC in 1945.[18]
The Manchester Small-Scale Experimental Machine, nicknamed Baby, was the world's first stored-
program computer. It was built at the Victoria University of Manchester by Frederic C. Williams, Tom
Kilburn and Geoff Tootill, and ran its first program on 21 June 1948.[37] It was designed as
a testbed for the Williams tube, the first random-access digital storage device.[38] Although the
computer was considered "small and primitive" by the standards of its time, it was the first working
machine to contain all of the elements essential to a modern electronic computer.[39] As soon as the
SSEM had demonstrated the feasibility of its design, a project was initiated at the university to
develop it into a more usable computer, the Manchester Mark 1.
The Mark 1 in turn quickly became the prototype for the Ferranti Mark 1, the world's first
commercially available general-purpose computer.[40] Built by Ferranti, it was delivered to
the University of Manchester in February 1951. At least seven of these later machines were
delivered between 1953 and 1957, one of them to Shell labs in Amsterdam.[41] In October 1947, the
directors of British catering company J. Lyons & Company decided to take an active role in
promoting the commercial development of computers. The LEO I computer became operational in
April 1951[42] and ran the world's first regular routine office computer job.
Transistors

A bipolar junction transistor


The bipolar transistor was invented in 1947. From 1955 onwards transistors replaced vacuum
tubes in computer designs, giving rise to the "second generation" of computers. Compared to
vacuum tubes, transistors have many advantages: they are smaller, and require less power than
vacuum tubes, so give off less heat. Silicon junction transistors were much more reliable than
vacuum tubes and had longer, indefinite, service life. Transistorized computers could contain tens of
thousands of binary logic circuits in a relatively compact space.
At the University of Manchester, a team under the leadership of Tom Kilburn designed and built a
machine using the newly developed transistors instead of valves.[43] Their first transistorised
computer and the first in the world, was operational by 1953, and a second version was completed
there in April 1955. However, the machine did make use of valves to generate its 125 kHz clock
waveforms and in the circuitry to read and write on its magnetic drum memory, so it was not the first
completely transistorized computer. That distinction goes to the Harwell CADET of 1955,[44] built by
the electronics division of the Atomic Energy Research Establishment at Harwell.[44][45]
Integrated circuits
The next great advance in computing power came with the advent of the integrated circuit. The idea
of the integrated circuit was first conceived by a radar scientist working for the Royal Radar
Establishment of the Ministry of Defence, Geoffrey W.A. Dummer. Dummer presented the first public
description of an integrated circuit at the Symposium on Progress in Quality Electronic Components
in Washington, D.C. on 7 May 1952.[46]
The first practical ICs were invented by Jack Kilby at Texas Instruments and Robert
Noyce at Fairchild Semiconductor.[47] Kilby recorded his initial ideas concerning the integrated circuit
in July 1958, successfully demonstrating the first working integrated example on 12 September
1958.[48] In his patent application of 6 February 1959, Kilby described his new device as "a body of
semiconductor material ... wherein all the components of the electronic circuit are completely
integrated".[49][50] Noyce also came up with his own idea of an integrated circuit half a year later than
Kilby.[51] His chip solved many practical problems that Kilby's had not. Produced at Fairchild
Semiconductor, it was made of silicon, whereas Kilby's chip was made of germanium.
This new development heralded an explosion in the commercial and personal use of computers and
led to the invention of the microprocessor. While the subject of exactly which device was the first
microprocessor is contentious, partly due to lack of agreement on the exact definition of the term
"microprocessor", it is largely undisputed that the first single-chip microprocessor was the Intel
4004,[52] designed and realized by Ted Hoff, Federico Faggin, and Stanley Mazor at Intel.[53]

Mobile computers become dominant


With the continued miniaturization of computing resources, and advancements in portable battery
life, portable computers grew in popularity in the 2000s.[54] The same developments that spurred the
growth of laptop computers and other portable computers allowed manufacturers to integrate
computing resources into cellular phones. These so-called smartphones and tablets run on a variety
of operating systems and have become the dominant computing device on the market, with
manufacturers reporting having shipped an estimated 237 million devices in 2Q 2013.[55]

You might also like