More Treasured Texts

From the 1940s to the 1980s, these books led engineers all the way to semiconductors and the Information Age

12 min read
More Treasured Texts

computing calculator

The HP 9100A computing calculator allowed users to enter programs with a magnetic program card. From Computer Structures: Readings and Examples by C. Gordon Bell and Allen Newell.

It’s a cloudy night on the Isle of Wight. The year is 1940, and a soldier sitting in an underground command center is peering intently at the glow of a cathode-ray tube. Suddenly, echo pulses appear on the screen, as they have almost every night for the past two months. Leaping for the radiophone, the soldier alerts his commanding officer. In minutes, Spitfire and Hurricane fighters are dispatched to intercept the incoming bombers over the English Channel.

That drama played itself out countless times during World War II. But regardless of who was attacking whom, the star was always radar. And radar, along with the proximity fuze and digital electronic computing, was part of a series of important inventions that transformed electrical engineering, EE education, and, particularly EE textbooks, in the middle of the 20th century. The transformation began before the war had even ended, when EE educators, particularly in the United States, began confronting the fact that it had been physicists, more than engineers, who had achieved many of the key breakthroughs in electronics.

The academics concluded that the EE curriculum had become too focused on practice and technique. What’s more, it had paid too little attention to the mathematical and physical fundamentals underlying electrical technologies, in particular, Maxwell’s equations. This realization prompted renewed emphasis on theory.

One of the first EE textbooks to adopt a strongly theoretical approach did such a good job that it remains in print to this day. This intermediate-level textbook, Fields and Waves in Modern Radio (1944) by General Electric Co. engineers Simon Ramo and John R. Whinnery, was written not for university students but for engineers in GE’s advanced training program. It covers electromagnetic theory so thoroughly that the reader becomes adept at applying Maxwell’s equations to a wide range of problems.

The book became the best-selling textbook ever for John Wiley & Sons. Recent editions have added Theodore Van Duzer as a third author, and the title was changed to Fields and Waves in Communication Electronics. Whinnery became a professor at the University of California, Berkeley. Ramo became a missile magnate, founding with Dean E. Wooldridge in 1953 a company that was the driving force behind the early U.S. intercontinental ballistic missiles program. In 1958 Ramo-Wooldridge merged with Thompson Products and eventually became TRW Inc. (now part of Northrop Grumman Corp.).

The intent of Ernst A. Guillemin’s 1949 textbook is clear from its title: The Mathematics of Circuit Analysis: Extensions to the Mathematical Training of Electrical Engineers [see figure]. Although topics include matrices, quadratic forms, vector analysis, complex variables, and Fourier series, Guillemin took an engineer’s approach, emphasizing plausibility and general understanding of the techniques rather than mathematical rigor. He was an important contributor to linear network analysis and synthesis, and his Communication Networks (1931, 1935) is also a classic.

Japanese professors have so cherished Naohei Yamada’s Denki-jikigaku (Electromagnetic theory) that the book, first published in 1950, came out in its latest edition just months ago. The book is famous for its careful explanations of the basic equations of electromagnetic theory and for its many exercises dealing with actual devices and equipment. Makoto Katsurai, professor of electrical engineering at Tokyo University, who updated the latest edition, agreed to take on the job even though he was the author of a textbook on the very same subject.

In France, Yves Rocard, a professor at the École Normale Supérieure, took up where Yamada in Japan and Ramo and Whinnery in the United States left off. His Électricité, first published in 1951, provided a broad foundation for all types of electrical and electronic engineering (giving much attention to l’électron libre), and sought to enhance students’ understanding of the science behind electrical engineering.

Rocard’s book is often compared to one by Karl Küpfmüller, Einführung in die theoretische Elektrotechnik (Introduction to theoretical electrotechnology). Both books analyzed the design and performance of actual devices, such as dynamos, galvanometers, radio receivers, and television cameras. [For a discussion of the Küpfmüller book, see ”Treasured Texts,” IEEE Spectrum, April 2003, pp. 44-49.] But Électricité went further, including discussions of magnetrons, waveguides, and other developments. For two or three decades, the book was the standard for French EE students, and it was translated into German. Despite having written such a successful book, Rocard is known today mainly as the father of the political leader Michel Rocard, who was prime minister of France from 1988 to 1991.

The art and science of circuit design

The invention of the transistor in 1947 and of the integrated circuit in 1959 kicked the postwar electronics boom into high gear. Both events set the stage for a vigorous expansion in applications. By the 1960s, scientists and engineers of all sorts were using electronics for instrumentation, control, and information processing. As the demand for new applications grew, designing electronic circuits became an important skill for large numbers of people.

One of the most valuable circuit design books had humble beginnings. In 1969 two students at the University of Stuttgart, Ulrich Tietze and Christoph Schenk, took a systematic approach to preparing for the E-Technik Prüfung (Germany’s exam for a master’s degree). They collected material on all aspects of circuit design and wrote detailed summaries. Fellow students clamored for copies of this comprehensive overview, but photocopying machines at that time were few. So a professor, who had connections with German publisher Springer-Verlag, urged its publication.

The result was astounding: enormous sales, 12 editions (so far) in German, many editions in six other languages, and one of the most successful technical books ever for Springer-Verlag. Halbleiter-Schaltungstechnik (the English version is titled Electronic Circuits: Design and Applications), expanded and updated over the years, teaches students not only to analyze circuits mathematically, but also to design circuits for particular applications. Offering many great how-to examples, it covers analog and digital circuits and emphasizes the use of commercially available ICs.

The Book I Remember


I am an avid reader and so it’s tough to come up with a single favorite text. Forced to do so, however, I suspect it might be William L. Shirer’s The Rise and Fall of the Third Reich (1960). Somehow this book awakened an idealistic young man to the fact that, sadly, there are folks in the world that wish to do others great ill—and that the answer in dealing with them is not one of appeasement.

* The author is chairman of the executive committee of the U.S. National Academy of Engineering (Washington, D.C.), former president and CEO of Lockheed Martin Corp. (Bethesda, Md.), and an adjunct professor at Princeton University (New Jersey)


The book that literally changed my life is Structured Computer Organization (1976), by Andrew S. Tanenbaum. It was used in the first class I had ever taken in computer architecture. I was very excited about the class and bought the book at the end of the previous school year. By the start of class in the fall semester, I had finished taking notes on the book and had worked every problem. To my surprise, the professor was using the text for the first time and had only worked his way through chapter 2. My notes and solved problems became the basis for the entire class. From this point forward, I knew I wanted to be in computer design and architecture.

* The author is chief technology officer of Intel Corp. (Santa Clara, Calif.)


In the development of Lotus Notes and now Groove, I have spent most of my career at the intersection of people, organizations, and technology. Computer Lib: Dream Machines (1987), by Ted Nelson, reinforced my belief that computers could be utilized for more than just data processing. It also influenced a philosophy that I have followed throughout my career: that it’s O.K. to think differently.

* The author is founder, chairman, and chief executive officer of Groove Networks Inc. (Beverly, Mass.)

Another electronics textbook that became an international best-seller—and how likely is that?—was The Art of Electronics (1980) by Paul Horowitz and Winfield Hill. Because both authors were themselves prolific in designing new electronic instruments, it is not surprising that the book is renowned for presenting the techniques that circuit designers actually use. Besides basic principles and general guidance, the text points out, there are ”back-of-the-envelope techniques,” ”nuts-and-bolts recommendations,” and ”a large bag of tricks.” Since the authors meant to convey the art, rather than the theory, of electronics, the title was carefully chosen. Translated into a dozen languages, the book became a standard textbook in countries around the world. A second edition appeared in 1989.

In Latin America, one of the most important textbooks on circuit design is the two-volume Análisis de Modelos Circuitales (Analysis of circuit models) (1981, 1982) by Héctor O. Pueyo and Carlos Marco, professors at the Universidad Tecnológica Nacional in Buenos Aires. The book, which has been used in numerous Latin American countries, presents circuit theory and shows, through many examples and exercises, the application of the theory in practical devices. The two volumes have gone through several editions, and a third volume, dealing with computer simulation, is scheduled for publication this year.

Designing very large ICs

The 1960s and 1970s were a heady time for electronics, as the IC came into its own and lived up to the prophecy first articulated by Intel Corp. founder Gordon E. Moore in his famous Moore’s Law. Just as Moore had predicted, in a 1965 article in Electronics magazine, the number of transistors that engineers could put on an IC doubled periodically, roughly every year in those days.

By the end of the 1970s, IC design engineers were dealing with many thousands of transistors, and it was clear that a systematic methodology for designing circuits was sorely needed. A breakthrough, both in EE education and in industry practice, was Carver Mead and Lynn Conway’s Introduction to VLSI Systems (1980), which led the way in showing students how to design complex ICs. (VLSI stands for very large scale integration.)

In 1975, Mead, at the California Institute of Technology (Pasadena), and Conway, at the Xerox Palo Alto Research Center (California), began a collaboration to find simplified methods of VLSI design. At the time, Conway was working on a special-purpose architecture for image processing, and she was bothered by ”the gap between the sorts of systems we could visualize and what we could actually get into hardware in a timely way.” Mead had for some years worked at computer-aided design, that is, automating, to some degree, the design of integrated circuits.

The two developed a methodology of VLSI design and tested it in courses at the California Institute of Technology and the Massachusetts Institute of Technology (Cambridge), where Conway was a visiting professor in 1978. Just a year after its publication, Introduction to VLSI Systems was in use in more than a hundred universities.

The new technology of computers

Probably the most momentous technical advance of World War II was the electronic digital computer, notably the Colossus in England for code breaking and the Eniac in the United States for calculating ballistics tables. Yet one of the first textbooks on computers came out of war-devastated Japan.

The Eniac, shrouded in secrecy during the war and immediately afterward, was publicly inaugurated on 15 February 1946. Newsweek magazine then carried an article about the machine, and a particularly interested reader was Kenzo Jo, a professor at Osaka Imperial University (later Osaka University). Jo, who had been doing research on calculating machines since 1929, immediately began developing an electronic computer.

He first built an Eniac-type decimal arithmetic unit and then worked on a stored-program binary computer. One of his assistants was Saburo Makinouchi, who became an associate professor at Osaka University in 1951, and in 1953 Jo and Makinouchi published Keisan Kikai (Computing machines). The book, which remained in use for many years, surveyed computing technologies—electromechanical calculators, punch-card systems, differential analyzers, and electronic digital computers—and presented, in considerable detail, binary computer circuits.

One of the most influential books on computers was Computer Structures: Readings and Examples [see figure] (1971) by C. Gordon Bell and Allen Newell, professors at Carnegie Mellon University (Pittsburgh). The authors surveyed computer structures—both in component configurations and in instruction sets—through actual computers rather than abstractions, and their book introduced a highly useful taxonomy—processor-memory-switch (PMS) notation for overall system structure and instruction-set-processor (ISP) notation for processor architectures.

Teaching can be a stimulus to invention, and it certainly was with this book. While working on it, Bell and Newell came up with two seminal ideas in computer design. One was the general-registers idea, that is, registers that could operate in various ways, as stack pointers, index registers, accumulators, program counters, and so on. Another was the Unibus protocol, which was introduced on the PDP-11, Digital Equipment Corp.’s hugely successful minicomputer. An updated version of the book, prepared by Daniel P. Siewiorek, Carnegie Mellon professor of electrical and computer engineering, appeared in 1982.

A dominant trend of late-20th century computing was the linking of large numbers of computers into networks. Probably the most influential textbook dealing with this trend has been Computer Networks (1981) by Andrew S. Tanenbaum, professor at Vrije Universiteit, in Amsterdam, the Netherlands. Written in English, the book has been translated into 17 languages and has sold more than half a million copies. Because of the continual changes in network hardware and software, subsequent editions in 1988 and 1996 were largely new books.

As Tanenbaum points out in the book’s third edition, the three editions correspond to different phases in the development of networks. In 1981, networks were an academic curiosity; in 1988, they were used by universities and large businesses; and in 1996, networks, and especially the Internet, were a part of daily life for hundreds of millions of people. Wireless networks, the latest phase, is part of the subject matter of a fourth edition of Computer Networks, to be published later this year.

Making computers useful

Programs for the first computers were necessarily written in machine instructions, but in the early 1950s, John Backus, Grace Murray Hopper, and others promoted ”automatic programming,” the process of writing programs in a higher-level code that the computer would convert into machine instructions. Fortran became the standard for scientific computing, Cobol for business computing. The most successful teacher of the new programming languages was the self-employed author and consultant Daniel D. McCracken.

In 1961, McCracken’s A Guide to FORTRAN IV Programming appeared [see figure]. It and later editions were used in technical schools and universities around the world. A statement in the introduction reminds us that in 1961 computers were few and far between: a computer is not necessary, ”but if one is available it should be utilized at every opportunity.” The books are clearly written and amply provided with flow charts, sample programs, and exercises (with answers in the back, something I remember appreciating as a user of the book ages ago).

In the early 1970s, Kenneth Thompson and Dennis M. Ritchie at Bell Telephone Laboratories were working on what became the Unix operating system. They wanted an easy and clear language for implementing Unix, but one that would retain the power of assembly language. Thompson created the first version, called B; Ritchie, the second version, called C. (There was never an A; one explanation of the name B is that it comes from BCPL, the name of a language designed by Martin Richards that was Thompson’s point of departure. BCPL got its name from an offshoot of CPL, the Combined Programming Language.)

Brian W. Kernighan, also at Bell Labs, collaborated with Ritchie in producing The C Programming Language (1978). The growing popularity of C and Unix and the success of the book reinforced each other. Indeed, the book served as the C programming language reference until a formal standard was adopted a decade later. A second edition appeared in 1988, and the book has been translated into 22 other (human) languages.

Typesetting books

It is fitting that books about computers introduced computer-based techniques to the publishing industry. Prior to the 1970s, a book was prepared as a typewritten manuscript, then set in metal type. After proofreading and correction, the type was used to print one copy of the book’s pages, which were photographed for the offset process. All this took a year or two in most cases and cost roughly US $20 a page.

Things changed in the 1970s, when Bell Labs developed Unix-based software to control typesetters, making it possible for authors to typeset their own books and produce camera-ready copy. This development reduced production cycles to a couple of months and cost very little. Kernighan and Ritchie’s The C Programming Language and Tanenbaum’s Computer Networks were early examples of texts produced in this way.

Even as Thompson and Ritchie were developing Unix, Niklaus Wirth, then a professor of computer science at the Eidgenössische Technische Hochschule (ETH, the Swiss Federal Institute of Technology) in Zurich, was creating Pascal and other computer languages. He was also an innovative teacher. In 1972 he published Systematic Programming (German edition in 1972, English edition in 1973) and, three years later, the more advanced Algorithms + Data Structures = Programs. The latter contains chapters on sorting, recursive algorithms, dynamic data structures, and language structures and compilers. Written from a practical viewpoint, the book teaches stepwise development of well-structured programs and emphasizes performance analysis. It has appeared in a dozen languages and five editions.

Final thoughts on great texts

The character of a profession depends to a large degree on the education its members receive, and that education also depends on the textbooks that convey the technical knowledge, problem-solving techniques, and vision of the profession. Textbooks not only hold a great deal of useful information, but also present an intellectual framework for subsequent studies and experience.

What makes a textbook great is the lucid explanation it provides, emphasizing timeless principles but relating them to useful techniques. Great textbooks challenge the student to turn understanding into a tool to solve practical problems and convey a comprehensive vision of a technical area, inspiring the student to master large parts of it and even to extend the field.

How the Books Were Chosen

Spectrum’s review of classic textbooks is not absolute or all encompassing. It is rather a subjective sampling of great texts that have had many readers over a long period of time and have profoundly influenced those readers. Whether a text is a classic might be measured, at least roughly, by how many copies were printed and how long the book remained in print. Another metric is nothing less than the warmth of readers’ testimonials. Dozens of people have told me what they regard as classic textbooks; and their testimonials helped me to make my selections.

Also, I wanted the selection of textbooks to illustrate the development of electrical technologies from the 1880s to the 1980s. So the books date from 1884 to 1981 and cover a wide range of technologies. (In the April issue of IEEE Spectrum [pp. 44-49], I described texts written during the formative years of the electrical engineering profession—from 1884 to the mid-20th century.) Some more recent books, which, in coming years, may well become classics, are listed in ”Up-and-Coming Classics?”.

I wanted, too, to indicate the transnational character of the technology, so although there is an emphasis on books from the United States, almost half come from other countries. Finally, there was space for only a couple of dozen textbooks, so scores of unquestionable classics have not been named. I know that Spectrum readers have their own cherished texts, and I invite them to add their opinions to this discussion.

The Conversation (0)

Q&A With Co-Creator of the 6502 Processor

Bill Mensch on the microprocessor that powered the Atari 2600 and Commodore 64

5 min read
Bill Mensch

Few people have seen their handiwork influence the world more than Bill Mensch. He helped create the legendary 8-bit 6502 microprocessor, launched in 1975, which was the heart of groundbreaking systems including the Atari 2600, Apple II, and Commodore 64. Mensch also created the VIA 65C22 input/output chip—noted for its rich features and which was crucial to the 6502's overall popularity—and the second-generation 65C816, a 16-bit processor that powered machines such as the Apple IIGS, and the Super Nintendo console.

Many of the 65x series of chips are still in production. The processors and their variants are used as microcontrollers in commercial products, and they remain popular among hobbyists who build home-brewed computers. The surge of interest in retrocomputing has led to folks once again swapping tips on how to write polished games using the 6502 assembly code, with new titles being released for the Atari, BBC Micro, and other machines.

Keep Reading ↓ Show less

Spot’s 3.0 Update Adds Increased Autonomy, New Door Tricks

Boston Dynamics' Spot can now handle push-bar doors and dynamically replan in complex environments

5 min read
Boston Dynamics

While Boston Dynamics' Atlas humanoid spends its time learning how to dance and do parkour, the company's Spot quadruped is quietly getting much better at doing useful, valuable tasks in commercial environments. Solving tasks like dynamic path planning and door manipulation in a way that's robust enough that someone can buy your robot and not regret it is, I would argue, just as difficult (if not more difficult) as getting a robot to do a backflip.

With a short blog post today, Boston Dynamics is announcing Spot Release 3.0, representing more than a year of software improvements over Release 2.0 that we covered back in May of 2020. The highlights of Release 3.0 include autonomous dynamic replanning, cloud integration, some clever camera tricks, and a new ability to handle push-bar doors, and earlier today, we spoke with Spot Chief Engineer at Boston Dynamics Zachary Jackowski to learn more about what Spot's been up to.

Keep Reading ↓ Show less

This Gift Will Help Your Aspiring Engineer Learn Technology

Know someone that is hard to shop for? We have the perfect gift for you.

4 min read

Trending Stories

The most-read stories on IEEE Spectrum right now