Computer, Analog
A digital computer employs physical device-states as symbols, but an analog computer employs them as models. That is, in a digital computer the on-off states of transistor devices are used to stand for 0s and 1s, which are made stand for phenomena of interest, such as words, data, pixels, or the like; in an analog computer, the continuously variable states of various electronic devices are made to behave like some physical system of interest. A rough parallel would be using pencil-and-paper mathematics to determine the carrying capacity of an arch design (symbolic computing) versus building a scale model out of balsawood and seeing how much weight it will bear (analog computing).
A more precise definition of analog computing is as follows: An analog computer models the behaviors of smoothly varying mathematical variables—usually representing physical phenomena such as temperatures, pressures, or velocities—by translating these variables into (usually) voltages or gear movements. It then manipulates these physical quantities so as to solve the equations describing the original phenomena. Thermometers and scales can be viewed as rudimentary analog computers: they translate an unwieldy physical phenomenon (i.e., a patient’s temperature or weight) into a manageable physical model or analog (i.e., the volume of a fixed quantity of colored alcohol or the displacement of a spring) that has been designed to vary linearly with the phenomenon to be measured; the device then derives a numerical result from the model (i.e., by aligning alcohol level or a pointer with a printed scale). Another example: pouring three equal units of water into an empty vertical tube and measuring the height of the result would be an analog method of computing that x + x + x = 3 x.
The earliest known analog computer is the astrolabe. First built in Greece during the first century BC, this device used pointers and scales on its face and a complex arrangement of bronze gears to predict the motions of the sun, planets, and stars.
Other early measuring devices were also analog computers. Sundials, for example, traced a shadow’s path to show the time of day. Spring-operated weight scales, which have been used for centuries, convert the pull on a stretched spring to numerical units of weight. The slide rule was invented about 1620 and was used until superseded by the electronic calculator in the late twentieth century.
In 1905, Rollin Harris (1863–1918) and E. G. Fisher (1852–1939) of the United States Coast and Geodetic Survey started work on a calculating device that would forecast tides. It was not the first such device, but was the most complex to ever be built. Dubbed the Great Brass Brain, it was 11 ft (3.35 m) long and 7 ft (2.1 m) high, weighed 2, 500 lb (1135 kg), and contained a maze of cams, gears, and rotating shafts. Completed in 1910, the machine worked as follows: an operator set 37 dials (each representing a particular geological or astronomical variable) and turned a crank. The computer then drew up tidal charts for as far into the future as the operator wished. It made accurate predictions and was used for 56 years before being retired in 1966.
(1890–1974), an electrical engineer at the Massachusetts Institute of Technology, created what is considered to be the first modern analog computer in the 1930s. Bush, with a team from MIT’s electrical engineering staff, discouraged by the time-consuming mathematical computations of differential equations that were required to solve certain engineering problems, began work on a device to solve these equations automatically. The first version of their device, dubbed the differential analyzer, was unveiled in 1930, the second in 1935. The latter weighed 100 tons, contained 150 motors, and hundreds of miles of wires connecting relays and vacuum tubes; instructions could be fed to the machine using hole-punched paper tape. Three copies of the machine were built for military and research use. Over the next 15 years, MIT built several new versions of the computer. By present standards the machine was slow, only about 100 times faster than a human operator using a desk calculator. Like most analog computers since, the MIT machines modeled phenomena using voltages, and contained a number of standard voltage-manipulating modules—integrators, differentiators, adders, inverters, and so forth— whose connections could be reconfigured to model, within limits, any desired equation.
Comments
Post a Comment