Sputnik: 50 Years Later

A story could start almost anywhere. This one begins at a moment startled by a rocket.

In the autumn of 1957, America was not at war … or at peace. The threat of nuclear annihilation shadowed every day, flickering with visions of the apocalyptic. In classrooms, “duck and cover” drills were part of the curricula. Underneath any Norman Rockwell painting, the grim reaper had attained the power of an ultimate monster.

Dwight Eisenhower was most of the way through his fifth year in the White House. He liked to speak reassuring words of patriotic faith, with presidential statements like: “America is the greatest force that God has ever allowed to exist on His footstool.” Such pronouncements drew a sharp distinction between the United States and the Godless Communist foe.

But on October 4, 1957, the Kremlin announced the launch of Sputnik, the world’s first satellite. God was supposed to be on America’s side, yet the Soviet atheists had gotten to the heavens before us. Suddenly the eagle of liberty could not fly nearly so high.

Sputnik was instantly fascinating and alarming. The American press swooned at the scientific vistas and shuddered at the military implications. Under the headline “Red Moon Over the U.S.,” Time quickly explained that “a new era in history had begun, opening a bright new chapter in mankind’s conquest of the natural environment and a grim new chapter in the cold war.” The newsmagazine was glum about the space rivalry: “The U.S. had lost its lead because, in spreading its resources too thin, the nation had skimped too much on military research and development.”

The White House tried to project calm; Eisenhower said the satellite “does not raise my apprehension, not one iota.” But many on the political spectrum heard Sputnik’s radio pulse as an ominous taunt.

A heroine of the Republican right, Clare Boothe Luce, said the satellite’s beeping was an “outer-space raspberry to a decade of American pretensions that the American way of life was a gilt-edged guarantee of our material superiority.” Newspaper readers learned that Stuart Symington, a Democratic senator who’d been the first secretary of the air force, “said the Russians will be able to launch mass attacks against the United States with intercontinental ballistic missiles within two or three years.”

A New York Times article matter-of-factly referred to “the mild panic that has seized most of the nation since Russia’s sputnik was launched two weeks ago.” In another story, looking forward, Times science reporter William L. Laurence called for bigger pots of gold at the end of scientific rainbows: “In a free society such as ours it is not possible to channel human efforts’ without the individual’s consent and wholehearted willingness. To attract able and promising young men and women into the fields of science and engineering it is necessary first to offer them better inducements than are presently offered.”

At last, in early February 1958, an American satellite — the thirty-pound Explorer — went into orbit. What had succeeded in powering it into space was a military rocket, developed by a U.S. Army research team. The head of that team, the rocket scientist Wernher von Braun, was boosting the red-white-and-blue after the fall of his ex-employer, the Third Reich. In March 1958 he publicly warned that the U.S. space program was a few years behind the Russians.

* * *

Soon after dusk, while turning a skate key or playing with a hula hoop, children might look up to see if they could spot the bright light of a satellite arching across the sky. But they could not see the fallout from nuclear bomb tests, underway for a dozen years by 1958. The conventional wisdom, reinforced by the press, downplayed fears while trusting the authorities; basic judgments about the latest weapons programs were to be left to the political leaders and their designated experts.

On the weekly prime-time Walt Disney television show, an animated fairy with a magic wand urged youngsters to drink three glasses of milk each day. But airborne strontium-90 from nuclear tests was falling on pastures all over, migrating to cows and then to the milk supply and, finally, to people’s bones. Radioactive isotopes from fallout were becoming inseparable from the human diet.

Young people — dubbed “baby boomers,” a phrase that both dramatized and trivialized them — were especially vulnerable to strontium-90 as their fast-growing bones absorbed the radioactive isotope along with calcium. The children who did as they were told by drinking plenty of milk ended up heightening the risks — not unlike their parents, who were essentially told to accept the bomb fallout without complaint.

Under the snappy rubric of “the nuclear age,” the white-coated and loyal American scientist stood as an icon, revered as surely as the scientists of the enemy were assumed to be pernicious. And yet the mutual fallout, infiltrating dairy farms and mothers’ breast milk and the bones of children, was a type of subversion that never preoccupied J. Edgar Hoover.

The more that work by expert scientists endangered us, the more we were informed that we needed those scientists to save us. Who better to protect Americans from the hazards of the nuclear industry and the terrifying potential of nuclear weapons than the best scientific minds serving the industry and developing the weapons?

In June 1957 — the same month Nobel Prize-winning chemist Linus Pauling published an article estimating that ten thousand cases of leukemia had already occurred due to U.S. and Soviet nuclear testing — President Eisenhower proclaimed that the American detonations would result in nuclear warheads with much less radioactivity. Ike said that “we have reduced fallout from bombs by nine-tenths,” and he pledged that the Nevada explosions would continue in order to “see how clean we can make them.” The president spoke just after meeting with Edward Teller and other high-powered physicists. Eisenhower assured the country that the scientists and the U.S. nuclear test operations were working on the public’s behalf. “They say: ‘Give us four or five more years to test each step of our development and we will produce an absolutely clean bomb.'”

But sheer atomic fantasy, however convenient, was wearing thin. Many scientists actually opposed the aboveground nuclear blasts. Relying on dissenters with a range of technical expertise, Democratic nominee Adlai Stevenson had made an issue of fallout in the 1956 presidential campaign. During 1957 — a year when the U.S. government set off thirty-two nuclear bombs over southern Nevada and the Pacific — Pauling spearheaded a global petition drive against nuclear testing; by January 1958 more than eleven thousand scientists in fifty countries had signed. Clearly, the views and activities of scientists ran the gamut. But Washington was pumping billions of tax dollars into massive vehicles for scientific research. These huge federal outlays were imposing military priorities on American scientists without any need for a blatant government decree.

* * *

What was being suppressed might suddenly pop up like some kind of jack-in-the-box. Righteous pressure against disruptive or “un-American” threats was internal and also global, with a foreign policy based on containment. Control of space, inner and outer, was pivotal. What could not be controlled was liable to be condemned.

The ’50s and early ’60s are now commonly derided as unbearably rigid, but much in the era was new and stylish at the time. Suburbs boomed along with babies. Modern household gadgets and snazzier cars appeared with great commercial fanfare while millions of families, with a leg up from the GI Bill, climbed into some part of the vaguely defined middle class. The fresh and exciting technology called television did much to turn suburbia into the stuff of white-bread legends — with scant use for the less-sightly difficulties of the near-poor and destitute living in ghettos or rural areas where the TV lights didn’t shine.

On the surface, most kids lived in a placid time, while small screens showed entertaining images of sanitized life. One among many archetypes came from Betty Crocker cake-mix commercials, which were all over the tube; the close-ups of the icing could seem remarkable, even in black and white. Little girls who had toy ovens with little cake-mix boxes could make miniature layer cakes.

Every weekday from 1955 to 1965 the humdrum pathos of women known as housewives could be seen on Queen for a Day. The climax of each episode came as one of the competitors, often sobbing, stood with a magnificent bouquet of roses suddenly in her arms, overcome with joy. Splendid gifts of brand-new refrigerators and other consumer products, maybe even mink stoles, would elevate bleak lives into a stratosphere that America truly had to offer. The show pitted women’s sufferings against each other; victory would be the just reward for the best, which was to say the worst, predicament. The final verdict came in the form of applause from the studio audience, measured by an on-screen meter that jumped with the decibels of apparent empathy and commiseration, one winner per program. Solutions were individual. Queen for a Day was a nationally televised ritual of charity, providing selective testimony to the goodness of society. Virtuous grief, if heartrending enough, could summon prizes, and the ecstatic weeping of a crowned recipient was vicarious pleasure for viewers across the country, who could see clearly America’s bounty and generosity.

That televised spectacle was not entirely fathomable to the baby-boom generation, which found more instructive role-modeling from such media fare as The Adventures of Spin and Marty and Annette Funicello and other aspects of the Mickey Mouse Club show — far more profoundly prescriptive than descriptive. By example and inference, we learned how kids were supposed to be, and our being more that way made the media images seem more natural and realistic. It was a spiral of self-mystification, with the authoritative versions of childhood green-lighted by network executives, producers, and sponsors. Likewise with the sitcoms, which drew kids into a Potemkin refuge from whatever home life they experienced on the near side of the TV screen. Dad was apt to be emotionally aloof in real life, but on television the daddies were endearingly quirky, occasionally stern, essentially lovable, and even mildly loving. Despite the canned laugh tracks, for kids this could be very serious — a substitute world with obvious advantages over the starker one around them. The chances of their parents measuring up to the moms and dads on Ozzie and Harriet or Father Knows Best were remote. As were, often, the real parents. Or at least they seemed real. Sometimes.

Father Knows Best aired on network television for almost ten years. The first episodes gained little momentum in 1954, but within a couple of years the show was one of the nation’s leading prime-time psychodramas. It gave off warmth that simulated intimacy; for children at a huge demographic bulge, maybe no TV program was more influential as a family prototype.

But seventeen years after the shooting stopped, the actor who had played Bud, the only son on Father Knows Best, expressed remorse. “I’m ashamed I had any part of it,” Billy Gray said. “People felt warmly about the show and that show did everybody a disservice.” Gray had come to see the program as deceptive. “I felt that the show purported to be real life, and it wasn’t. I regret that it was ever presented as a model to live by.” And he added: “I think we were all well motivated but what we did was run a hoax. We weren’t trying to, but that is what it was. Just a hoax.”

* * *

I went to the John Glenn parade in downtown Washington on February 26, 1962, a week after he’d become the first American to circle the globe in a space capsule. Glenn was a certified hero, and my school deemed the parade a valid excuse for an absence. To me, a fifth grader, that seemed like a good deal even when the weather turned out to be cold and rainy.

For the new and dazzling space age, America’s astronauts served as valiant explorers who added to the elan of the Camelot mythos around the presidential family. The Kennedys were sexy, exciting, modern aristocrats who relied on deft wordsmiths to produce throbbing eloquent speeches about freedom and democracy. The bearing was American regal, melding the appeal of refined nobility and touch football. The media image was damn-near storybook. Few Americans, and very few young people of the era, were aware of the actual roles of JFK’s vaunted new “special forces” dispatched to the Third World, where — below the media radar — they targeted labor-union organizers and other assorted foes of U.S.-backed oligarchies. But a confrontation with the Soviet Union materialized that could not be ignored. Eight months after the Glenn parade, in tandem with Nikita Khrushchev, the president dragged the world to a nuclear precipice. In late October 1962, Kennedy went on national television and denounced “the Soviet military buildup on the island of Cuba,” asserting that “a series of offensive missile sites is now in preparation on that imprisoned island.” Speaking from the White House, the president said: “We will not prematurely or unnecessarily risk the costs of worldwide nuclear war in which even the fruits of victory would be ashes in our mouth — but neither will we shrink from that risk at any time it must be faced.”

Early in the next autumn, President Kennedy signed the Limited Test Ban Treaty, which sent nuclear detonations underground. The treaty was an important public health measure against radioactive fallout. Meanwhile, the banishment of mushroom clouds made superpower preparations for blowing up the world less visible. The new limits did nothing to interfere with further development of nuclear arsenals.

Kennedy liked to talk about vigor, and he epitomized it. Younger than Eisenhower by a full generation, witty, with a suave wife and two adorable kids, he was leading the way to open vistas. Store windows near Pennsylvania Avenue displayed souvenir plates and other Washington knickknacks that depicted the First Family — standard tourist paraphernalia, yet with a lot more pizzazz than what Dwight and Mamie had generated.

A few years after the Glenn parade, when I passed the same storefront windows along blocks just east of the White House, the JFK glamour had gone dusty, as if suspended in time, facing backward. I thought of a scene from Great Expectations. The Kennedy era already seemed like the room where Miss Havisham’s wedding cake had turned to ghastly cobwebs; in Dickens’ words, “as if a feast had been in preparation when the house and the clocks all stopped together.”

The clocks all seemed to stop together on the afternoon of November 22, 1963. But after the assassination, the gist of the reputed best-and-brightest remained in top Cabinet positions. The distance from Dallas to the Gulf of Tonkin was scarcely eight months as the calendar flew. And soon America’s awesome scientific capabilities were trained on a country where guerrilla fighters walked on the soles of sandals cut from old rubber tires.

Growing up in a mass-marketed culture of hoax, the baby-boom generation came of age in a warfare state. From Vietnam to Iraq, that state was to wield its technological power with crazed dedication to massive violence.

This column is excerpted from NORMAN SOLOMON’s new book Made Love, Got War.

 

Norman Solomon is the national director of RootsAction.org and executive director of the Institute for Public Accuracy. His latest book, War Made Invisible: How America Hides the Human Toll of Its Military Machine, is published by The New Press.