Personal Computers: History and Development
Overview
The personal computer (PC) has revolutionized business and
personal activities and even the way people talk and think; however, its development has been less of a revolution than an
evolution and convergence of three critical elements - thought, hardware, and software. Although the PC traces its lineage
to the mainframe and minicomputers of the 1950s and 1960s, the conventional thought that was prevalent during the first thirty
years of the computer age saw no value in a small computer that could be used by individuals.
A PC is a microcomputer, so named because it is smaller than
a minicomputer, which in turn is smaller than a mainframe computer. While early mainframes and their peripheral devices often
took up the floor space of a house, minicomputers are about the size of a refrigerator and stove. The microcomputer, whose
modern development traces back to the early 1970s, and fits on a desk.
From the start, the creation of the computer was centered around the concept that a single unit would be
used to perform complex calculations with greater speed and accuracy than humans could achieve.
The Transistor
On December 23, 1947, one of the most far-reaching technologies
of the 20th Century was developed at Bell Laboratories by John Bardeen, Walter Brattain, and William Shockley - the
transistor. But the transistor wasn't available to U.S. manufacturers
until 1956, when a seven year-old antitrust law suit against AT&T, the owners of Bell Labs, was settled. The judgment required
that AT&T give away licenses to manufacture the transistor to American companies. Following this decision, the transistor
was used to replace thousands of vacuum tubes in computers and began the miniaturization of electronics.
Because it drastically reduced the size and heat considerations of the large vacuum tubes, the transistor enabled the computer
to become a viable tool for business and government.
The Computer Mystique
From the beginning, computers baffled the populous with their
capability. In corporate and government offices and on university campuses, information processing departments sprouted up
to serve the computer. The IBM 701, which was introduced in 1952 as a business computer, was comprised of several units that
could be shipped and connected at a customer's location, rather than the earlier massive units that had to be assembled on
site. In 1953, IBM began shipping the first mass-produced computer, the IBM 650.
IBM introduced the first solid-state (transistorized) computer in 1959, the IBM 7090. Then in 1964, IBM culminated over $1
billion in research when it brought out the System/360 series of computers. Unlike other mainframes, the System/360 computers
were compatible with each other.
By 1960, the computer was king. Companies hired armies of
technicians and programmers to write its operating programs and software, fix it, and allocate the precious computer time.
The capability of the machines was more than a mere mortal could fathom, but gathering raw data and "keying" it in so the
computer could "crunch the numbers" was a complicated and time-consuming task.
Frustrations abounded, computer errors were called "glitches,"
and the phrases "garbage in/garbage out," "It's a computer mistake," and "Sorry, the computer's down and we can't do anything,"
were introduced into the lexicon.
On college campuses in the 1960s, students carried bundles of computer cards to
and from class, hoping that their share of the valuable computer time would not be bumped or allocated to someone else. The
term, "Do not fold, spindle or mutilate," was coined so people wouldn't disable the process of feeding the punched computer
cards into punch card readers, where the intricate patterns of holes were decoded.
The computer mystique was reinforced in people every time
they heard of some new accomplishment. In 1961, a computer calculated the value of pi to 100,000 decimal places. A computer
could play checkers, and in 1967 a chess playing computer program was made an honorary member of the
United States Chess Federation. Banks began printing checks with magnetic ink so they could be processed by the computers.
A Small Change in Thought
Until 1971, nobody even thought of a computer as anything
but a big, fast, electronic brain that resided in a climate-controlled room and consumed data and electricity in massive quantities.
In 1971, an Intel 4004 chip containing 4004 transistors was programmed to perform
complex mathematical calculations; the hand-held calculator was born. Suddenly, scientists and engineers could carry the computational
power of a computer with them to job sites, classrooms, and laboratories; but the hand-held calculator, like the ENIAC before it, was not yet a computer. The microprocessor
was developed by Robert Noyce, the founder of Intel and one of the inventors of the integrated circuit, and brought with it
a change in the way people worked.
New Technologies and New Ideas
Small, hand-held calculators had provided an idea, or at
least a "what if," to some people. Still, in the early 1970s, computers were used for number crunching and printing out long
streams of green and white paper. IBM Selectric typewriters were the top of the line "word processors" and Xerox copiers churned out photocopies. Most people never imagined
that a computer could process data in real time, be used to write letters, or fit on a desk.
In 1972, Intel brought out its 8008 chip, capable of processing
8-bits of data, enough to convey numbers and letters of the alphabet. In that same year, Xerox began working on a personal
computer at their Palo Alto Research Center. For the next several years, a team of Xerox scientists worked
on the "Alto," a small computer that would have become the first PC if only the development team had been able to convince
someone of its usefulness.
Likewise, in 1972 Digital Electronics Corporation (DEC), a minicomputer manufacturing company headed by Kenneth Olsen,
had a group of product engineers developing the DEC Datacenter. This PC incorporated not only the computer hardware but the
desk as well. The DEC Datacenter could have put tremendous computing capability in the home or at work, but management saw
no value to the product and halted its development.
In the end, none of the giant companies whose names had been
synonymous with computers would introduce the PC to the world. There seemed to be no future in an inexpensive product that
would replace the million dollar "Big Iron" that they were selling as fast as they could make them.
The people who eventually introduced the PC were rebels.
Many had spent time in the bowels of the big companies and were frustrated by the lack of vision they encountered. They retreated
into their own garages and attended meetings with other "computer nuts" who saw a much different future than the one laid
out over the previous 30 years by the giants of the computer industry.
The PC is Born
In 1975, Rubik's Cube was put on store shelves and proved to many that the human
brain was incapable of complex problem solving. But a ray of hope also appeared; the first PC was introduced. Micro Instrumentation
and Telemetry Systems, Inc. (MITS) sold a kit for the MITS Altair 8800 that enabled computer hobbyists to assemble their own
computers. It had no monitor, no keyboard, no printer, and couldn't store data, but the demand for it, like Rubik's Cube,
was overwhelming.
The Altair proved that a PC was both possible and popular,
but only with those people who would spend hours in their basements with soldering irons and wire strippers. The Altair, which
looked like a control panel for a sprinkler system, didn't last, but it helped launch one of the largest companies in the
computer world and gave a couple of young software programmers a start. In 1974, Bill Gates and Paul Allen wrote a version
of BASIC for the Altair and started a company called Microsoft Corporation.
In 1976, another computer kit was sold to hobbyists - the
Apple I. Stephen Wozniak sold his Volkswagen and Steve Jobs sold his programmable calculator to get enough
money to start Apple. In 1977, they introduced the Apple II, a pre-assembled PC
with a color monitor, sound, and graphics. It was popular, but everyone knew that a serious computer didn't need any of this.
The kits were just a hobby and the Apple II was seen as a toy. Even the Apple name wasn't a serious, corporate sounding name
like IBM, Digital Equipment Corporation, or Control Data.
But 1977 also brought competition. The Zilog Z-80 microprocessor,
which had been introduced in 1975, was used in the Tandy Radio Shack TRS-80, affectionately called the "Trash 80." Apple, Commodore,
and Tandy dominated the PC marketplace. The Apple II had 16K bytes of RAM and 16K bytes of ROM; Commodore Business Machines' Personal Electronic Transactor (PET) included 4K RAM and 14K
ROM; and the TRS-80 had 4K RAM and 4K ROM.
Also in 1977, the Central Program for Microprocessors (CP/M) operating system was developed by Digital Research and Gary
Kildall. From its introduction until 1980, CP/M was used in most PCs, but even that did not guarantee that a program or document
could be written on one machine and read on another because each manufacturer used different floppy disk drives.
Apple introduced the floppy disk drive in 1978, allowing
Apple II users to store data on something other than the cumbersome and unreliable tape cassettes that had been used up to
that point.
But despite the popularity of the three PCs, non-computer
people still saw little reason to buy an expensive calculator when there were other ways to do the same things. In 1979, that
all changed.
When VisiCalc was introduced for the Apple II, non-computer
people suddenly saw a reason to buy a computer. VisiCalc, a spreadsheet program created by Dan Bricklin and Bob Frankston,
allowed people to change one number in a budget and watch the effect it had on the entire budget. It was something new and
valuable that could only be done with a computer. For thousands of people, the toy, the computer few could find a use for,
had been transformed into a device that could actually do something worthwhile.
Microprocessors and high-tech gadgets were gradually worming
their way into people's lives. In 1978, Sony introduced the Beta format video cassette recorder, and a year
later the VHS video recorder and the Sony Walkman. And to remind everyone of how far we had to go, Star Trek: The Motion Picture came to theaters in 1979.
The Sinclair ZX-80 PC, which hit the market in 1980, used
the same Z-80 chip as Commodore's PET and the Tandy TRS-80. The ZX-80 had 1K RAM and 4K ROM. Developed by British entrepreneur
Clive Sinclair, the ZX-80 meant that people could enter the computer revolution for under $200. Its small size and price attracted
people who had never thought about owning a PC.
The Commodore VIC-20, also introduced in 1980, had a color
monitor and would eventually become the first PC to sell more than one million units.
Even with all of the success the early PC manufacturers had
in the late 1970s and early 1980s, the advances in microprocessor speeds, and the creation of software, the PC was still not
seen as a serious business tool. Unknown to everyone in the computer industry; however, a huge oak tree was about to drop
an acorn that would fall close to the tree and change everything.
Validation of the PC
Two events occurred in 1981 that would have a tremendous
impact on the future of the PC. In 1980, IBM had started a secret project in Boca Raton, Florida called "Acorn." Thirteen months later, in 1981, IBM introduced
the IBM PC, a product that validated the PC as a legitimate business tool. For many people, even those who prided themselves
on being able to operate the "Big Iron," if IBM was making PCs then the small desk-top units were worthy of respect.
When the IBM PC hit the market, it was a compete system.
Secretly, IBM had provided software developers with prototypes of their PC so they could develop an array of programs that
would be available when the machine hit the streets. IBM also developed printers, monitors, and expansion cards for the PC
and made it an open system so other manufacturers could develop peripherals for it.
The IBM PC used an Intel 8088 microprocessor, had 16K of
RAM, was expandable to 256K, came with one 5.25-inch disk drive and room for a second, and was available with a choice of
operating systems; CP/M-86 or IBM PC-DOS, which had been developed by Microsoft.
The second major event of 1981 was the introduction of the
first luggable computer, the Osborne 1. This self-contained, suitcase-sized PC, developed by Adam Osborne, was not only the
first portable PC, but also the first to be sold with software. The Osborne I came with BASIC, CBASIC, WordStar for word processing,
and the SuperCalc spreadsheet program. Over the next two years, the Osborne Computing Company would go from nothing to a company
with $70 million in annual revenue and then into bankruptcy.
Prior to 1980, the most common method of storing data was
to connect an audio tape recorder to the PC and dump data to standard tape cassettes. Large word processors and some PCs had
8-inch disk drives, but in 1980 Al Shugart introduced the Winchester hard-disk drive.
The Race was On
Now that the PC had been validated, it began appearing on
desk-tops in large and small companies to produce work schedules and payrolls, write letters and memos, and generate budgets.
Software enabled people to do more in less time and business was promised the "paperless office" as an added benefit of the
PC.
Managers attended classes and began writing memos and letters,
but many felt that the work they could now do themselves on a PC was demeaning; it was the work that secretaries and clerks
had always done. For some, having a PC on the desk meant that they now had to do the work, not just delegate it, and for others
it meant they no longer supervised a person, but a machine.
There was also a strong fear factor. The PCs were expensive
and many people were afraid they would damage the units or erase everything in one keystroke. People who had always worked
with things they could see and understand were suddenly putting their faith in chips and hard drives that they not only couldn't
see or touch, but they also didn't understand. Suddenly it was permissible to make a mistake in spelling or grammar; it could
be changed and rewritten until it was correct. The whole thought process didn't set well with some, for others it freed them
from the drudgery of using white correction fluid to cover up mistakes on printed documents.
The early 1980s were a time of furious change in the computer
industry. More than 100 companies were manufacturing PCs, each with its own unique features, each with its own software. When
IBM entered the market in 1981, software companies knew that writing IBM compatible software would be profitable. Software
for the Apple II had exploded to 16,000 titles and IBM would do the same. New software in the 1980s included WordStar, Lotus
1-2-3, Microsoft Word, and Word Perfect.
In 1981, Hayes Micromodem brought the MOdulator/DEModulator (MODEM) to the
market for PCs. The modem had been invented at AT&T Bell Labs in 1960 to connect mainframes and minicomputers. Hayes'
modem allowed PCs to communicate with each other and access CompuServe and The Source, the online services that started up in 1979.
CompuServe showed people what to do with their 300 baud modems by offering them an array of services and databases to connect
with.
In 1982 Compaq introduced the first IBM compatible machine. Until Compaq,
most manufacturers feared IBM and would not bring out a machine that was compatible with the PC. Later the compatibles would
be termed "clones."
Also in 1982, Tandy brought out the TRS-80 Model 16, which
was based on the Motorola 68000 and Z-80 microprocessors. The Model 16 retailed for $5,000 and included 128K RAM, an 8-inch
floppy disk drive, as well as the Xenix operating system, a derivative of UNIX.
In January, 1983 Time magazine anointed the PC as the "Man of the Year," a designation
by the editors that the computer had been the most influential newsmaker of 1982. The magazine estimated 80 million PCs would
be in use by the end of the century. Industry leaders included Texas Instruments, Timex, Commodore, Atari, Apple, IBM, and Tandy, with Osborne leading the way in the
portable market.
The individuals pushing the PC into the future were John
Opel at IBM, Adam Osborne of Osborne Computers, VisiCalc creator Dan Bricklin, Jack Tramiel of Commodore, and Clive Sinclair
who founded Sinclair Research.
The leading products of 1982 and their sales figures included
the Timex/Sinclair 1000 - 600,000; Commodore VIC-20 - over 600,000, Atari 400 and Atari 800 - 600,000; Texas Instruments 99/4A
- 530,000; TRS-80 Model III - 300,000; Apple II Plus - 270,000; IBM PC - 200,000; and Osborne 1 - 55,000. These computers
ranged in price from the $99 Timex/Sinclair to the Osborne 1 at $1,795 with bundled software. In the opinion of Time,
computers priced over $2,000 would appeal to a market of "…growing small businesses and big corporate clients…"
Manufacturers of these higher end PCs included Altos, Corvus, Cromemco, Control Data, Digital Equipment, Hewlett-Packard, North Star, Olivetti, Tele Video, Toshiba, Xerox, and Zenith.
But in 1983 there was once again a wind of change blowing
across the PC landscape.
The Mac Attack
In 1983, Apple brought out a machine that failed to sell
but nonetheless showed consumers and manufacturers a new direction for the PC. The Lisa, an expensive PC with a graphical
user interface (GUI), hit the market with a thud. At $10,000, it had few friends and even fewer buyers.
Also in 1983, IBM introduced IBM XT with a 10MB hard drive,
three additional expansion slots, 128K RAM, and a 360K floppy drive. To many buyers, the 10MB storage capacity seemed large
enough to last a lifetime.
Immediately after the failure of Lisa, Steven Jobs rethought
the machine and in 1984, out came the Macintosh. The Macintosh was powered by Motorola's 68000 processor and came with 128K
of RAM. It was so radically different from any other PC that it split the PC world into two halves that would not be rejoined
for another decade. In addition to the GUI that made the computer an "intuitive" extension of the user, the "Mac" had its
own operating system that was incompatible with IBM's MS-DOS system. Suddenly PC meant DOS-based and IBM compatible and Mac
meant GUI and mouse.
The Mac was introduced to the world in an extravagant television
commercial that was shown only once during half-time of the NFL Super Bowl. The commercial changed the advertising industry
almost as much as the Mac changed computing.
Suffering from the failure of the Apple III and Lisa, Apple
was literally saved by the Mac. People who hated computers loved the simplicity of Mac. The GUI allowed the user to click
a mouse button on an icon to launch a program, print a document, or copy a file. No longer did users have to know combinations
of keys or special codes to get the computer to do what they wanted it to do. The Mac was "user friendly."
Although not the first PC with a mouse or GUI (that distinction
went to Xerox's $50,000 Star that came out in 1981 and immediately failed), the Mac did set the computer world on its ear
because of its ease of operation and its operating system.
When Apple came out with the Apple LaserWriter in 1985 it
was with Adobe Systems Inc.'s PostScript page description language. By 1986, with its what-you-see-is-what-you-get (WYSIWYG)
display and printing, desk-top publishing was born. WYSIWYG meant that a person could format a document with special fonts
and spacing and be assured that what came out of the printer would look like what they had created on the screen.
Adobe, founded in 1982 by John Warnock and Charles Geschke,
turned the printed page into a graphic image. The bit map made each pixel on the screen a definable image that could be moved
and changed without the limitations of a standard text format. PostScript changed the way people thought about fonts, page
layout, and the visual impact of the documents they produced with their PC. Printers like the Apple LaserWriter and the Hewlett-Packard
HP LaserJet made every document look like it had been professionally typeset and printed.
In 1985, the Commodore Amiga 1000, which featured multitasking,
graphics, sound, and video in a windowing operating system, exposed people to multimedia. At the same time Toshiba came out
with the T1100 laptop, Tandy introduced the Tandy 200 laptop, and AT&T introduced the UNIX PC. Intel took the microprocessor
to a new level when it brought out the 386 microprocessor in 1985, proving that PCs were not only getting better, they were
getting faster.
The 1980s were very active times for hardware manufacturers
and software producers. Small software companies locked in with either IBM or Macintosh, but large companies like Microsoft
were able to create new applications for both operating systems. While Aldus brought out PageMaker, and Lotus introduced Jazz,
Microsoft announced Excel for the Mac, C 3.0, and finally shipped a long-awaited program called Windows.
Bill Gates, a founder of Microsoft, tried three times to
interest IBM in Windows but was turned down each time. Although the Mac operating system had changed the interface between
users and their PCs, many DOS users continued to hang on to their command line-driven MS-DOS operating system, and it would
be several more years until the Windows concept caught on.
With the availability of hundreds of software programs, hard
disk space became valuable real estate. The 10MB hard disk on the IBM XT began to fill up so hard drive manufacturers started
the process of doubling their capacity.
As modems proliferated and the Hayes Smartmodem was accepted
as the standard for modems, just about everyone either knew someone they could get online with, subscribed to an online service
such as CompuServe, or wanted to access the 1000 host sites on the Internet.
But PCs that were connected to the outside world were also
vulnerable to a new phenomenon called viruses. Once downloaded, these programs could attach themselves without
warning to a PC's hard drive and gradually or in the blink of an eye destroy or overwrite files. Virus checkers became the
rage for anyone who received data over telephone lines.
By 1987 enough people were writing their own software and
sharing it that the Association of Shareware Professionals was formed to market and protect the inexpensive software.
In 1987 a new computer language, C++, stimulated the growth of object-oriented programming (OOP).
Out of the Box and Obsolete
For consumers, the late 1980s were a time of frustration.
No sooner had they learned to run their new PC and Macs than a new, better, larger, faster model was on the shelf. New versions
of software, printers, and modems made it impossible to have the latest of anything.
In 1990, Intel's 386 and Motorola's 68030 microprocessors were at the top, then in 1991 Intel
brought out the i486SX 20 MHz chip and Motorola introduced the 68040. Less than a year later Intel introduced the 50MHz 486
chip and Tandy brought out its $400 CD-ROM drive for PCs. Then, just to make everyone wonder what was going on, in 1991 Apple
and IBM agreed to share technology by integrating the Mac into IBM's systems and using the IBM Power PC chip.
In 1992, Apple brought out the Apple PowerBook, a laptop
that made everyone wonder just how small a full-function computer could get. A year later everyone knew the answer when Apple
introduced the Newton Personal Digital Assistant (PDA). The Newton was supposed to be able to recognize hand-written notes
and Apple sold 50,000 of them in 10 weeks.
In 1993, Intel introduced the 60MHz Pentium chip, the next
generation of chips. The Pentium; however, had a nasty mathematical bug and its acceptance was slowed. Apple discontinued
the workhorse of its fleet, the Apple II, which, despite the mind boggling changes in the industry, had lasted 17 years.
Not only were hardware and software obsolete, people were
also getting caught up in their own obsolescence. For years, employers had included the operating systems and software names
in their advertising for clerical and secretarial positions. As companies used more temporary workers and included both IBM
clones and Macintosh's in their operations, proficiency with only one slammed the door on employment opportunities.
Many people enrolled in classes to learn the latest software
or update their computer skills. A good, well-rounded employee needed to know desktop publishing, two or more word processing
programs, at least one spreadsheet program, and a graphics package. They had to be able to access the company local area network
(LAN), send and receive E-mail using high-speed (28,800bps) modems, and solve problems with hardware and software to maximize
their output. Microprocessor-driven telephones, cellular phones, and pagers added to the complexity of the job, and repetitive
motion syndrome from using keyboards hour after hour created an army of people wearing wrist braces.
Many people left a job where their day was spent working
at a computer terminal or PC and went home to enjoy the quite, relaxing camaraderie they found in Internet chat rooms, by visiting the World Wide Web, or reading their favorite newspapers and electronic magazines (ezines).
From its inception in 1975, the PC has become a focal point
of business, education, and home life. The microprocessor, an amazing technology when it had 4000 transistors on a single
chip, is now even more amazing when it has over 3 billion transistors on an even smaller chip. In 1982, when Time magazine
made the computer its "Man of the Year," the PC was still in its infancy. "Big Iron" still dominated the high-tech environment
and having a personal computer was a luxury.
The creation and success of the PC would not have been possible
without the elimination of the concept that a computer was a large, centralized, data processor and number cruncher. Today
the PC is a communication channel more than it is a computational tool. Millions of people work in their "electronic cottages,"
either operating their own business from home or telecommuting to work. It is ironic that one of the first Intel 4004 microprocessors
made continues to operate and lead the world to the outer edges of time and space. In 1972 one of the small chips was installed
in the Pioneer spacecraft. Today it continues to operate over 5 billion miles from earth.