First let me address the conclusion of King's essay: "We are not even close to the danger point.... Language does not threaten American unity." He prescribes a policy of "benign neglect." Even if we grant his premise that we are not yet in a dangerous situation, it does not make sense to postpone setting a language policy for the United States until we are in the midst of a linguistic crisis -- by then it may be too late. By acting now to make English the official language, Congress may prevent such a crisis from ever occurring.

However, even benign neglect would be preferable to the current policy of active expansion of multilingual government services. One can take a driver's-license exam in twenty languages in Michigan, twenty-three in New York, twenty-five in Massachusetts, and thirty-three in California. And these are only a fraction of the 329 languages spoken in the United States in 1990.

The policy of benign neglect toward language worked for nearly 200 years. But as King himself acknowledges, things began to change about thirty years ago. Extremists began pushing an activist multilingual policy for our government, attacking one of the great unifying factors in our nation's history. Yet in the Orwellian doublespeak that is the official language of the politically correct, it is the movement in favor of a common language that is now called divisive.

Mauro E. Mujica

Robert D. King is disingenuous in concluding that "benign neglect is a good policy for any country when it comes to language." The social policies mandating bilingual education, salary differentials, and hiring preferences for bilingual employees, and special services for non-English-speaking people, are anything but benign neglect. These policies, along with affirmative action, multiculturalism, and so forth, are in fact part of the dominant liberal social-engineering paradigm that has been in ascendance for some thirty years in this country.

Bob Fromuth

There is nothing particularly offensive in what Robert D. King says in his recent article. It's what he doesn't say that presents the problem. Nowhere in his piece, for example, do the words "bilingual education" appear. He does not even hint that he is aware that bilingual education threatens to leave more than a million Hispanic youngsters functionally illiterate in English, the language they will need if they are ever to integrate fully into the economic, social, and political mainstream of this society.

I would welcome a policy of "benign neglect" when it comes to language. Ever since the Supreme Court's Lau decision, in 1974, however, the federal government has coerced literally hundreds of school districts into instituting full-fledged native-language-maintenance programs, even when local educators would have preferred to adopt an alternative such as English as a Second Language. The result is a legacy of half-educated children&nbsp -- almost all Hispanic&nbsp -- who have not learned English as well as they might have precisely because the federal government intervened.

King ignores other federal actions that similarly abuse his rule of benign neglect. The U.S. Department of Housing and Urban Development has threatened to withhold millions of dollars in federal housing funds from cities that refuse to print documents in languages other than English. The Voting Rights Act continues to force localities to print foreign-language ballots. Last year the Yuba County, California, registrar of voters complained in congressional testimony that this mandate would cost her constituents nearly $30,000 in 1996. She has received only one request in sixteen years from a constituent seeking foreign-language voting assistance.

With meddlesome policies like these, is it any wonder that so many Americans want to make English their country's official language?

Linda Chavez

There is little difference between what my critics want and what I want&nbsp -- that every citizen of the United States speak, understand, read, and write English. But there is a gulf of difference between us regarding means. U.S. English thinks that passing a law making English the official language of the country will do the job. I don't. Linguistic history teaches painfully unambiguous lessons about language laws: they are divisive, they are unenforceable, they make people hate each other, they encourage snitches ("Juanita talked Spanish during the break!"). Worse, they don't work.

So what might work? Linda Chavez's letter points the way. If you don't like bilingual official documents (it's not a big issue with me), then repeal the laws that mandated them in the first place. Bilingual education? I've seen school districts in Texas where it works just the way it is supposed to: little kids enter kindergarten speaking no English, and by the second grade everything is in English. But I know of other places where bilingual education continues far too long, where it is ruinously bureaucratized or politicized, where it has become zany (for example, demanding instruction in ancient Aztec, I'm told by a caller on a radio talk show). For me it's empirical, not ideological: if bilingual ed works, great; if it doesn't, change it. But don't pass a national law making English our official language (with unintended but predictably bad consequences) in order to remedy bilingual ballots and bilingual education.

The letters printed here make reasonable points. Most of the others I have received have been overwhelmingly vituperative&nbsp -- which reinforces my conviction that only foolish governments mess with an icon of such extraordinary power and intensity as language.

Mark Halpern's defense of usage orthodoxies "A War That Never Ends," March Atlantic) offers my own Atlantic article "Decline of Grammar" December, 1983) as a paradigm of "leave-your-language-alone" descriptivism, along with Steven Pinker's book The Language Instinct. Pinker can speak for himself, but only a willful misreading of my article would lead someone to describe it in this way.

Halpern characterizes my views by saying, "Nunberg objects to the prescriptivist approach on two grounds: it is futile, since language will follow its natural destiny despite all the efforts of the prescriptivists; and it is somehow wrong -- immoral? unethical? -- to try to interfere...." He is right about the first part, dead wrong about the second. As he notes, I did liken prescriptive grammarians to landscape gardeners trying to arrest the processes of continental drift, and I did say that in the long run English will survive whatever abuses its critics complain of. But I went on to argue that this by no means invalidates the prescriptivist program.

Yet it is hard to take comfort in the scholars' sanguine detachment.... It may be that my children will use gift and impact as verbs without the slightest compunction (just as I use contact, wondering that anyone ever bothered to object to it). But I can't overcome the feeling that it is wrong for me to use them in that way and that people of my generation who say "We decided to gift them with a desk set" are in some sense guilty of a moral lapse, whether because they are ignorant or because they are weak. In the face of that conviction, it really doesn't matter to me whether to gift will eventually prevail, carried on the historical tide. Our glory, Silone said, lies in not having to submit to history.

It is hard to see how anyone could read this as a descriptivist credo, and in fact my article defended a reasonable program of language criticism, one sensitive to the complexities of both received opinion and modern linguistic research. Fortunately, it isn't necessary to belabor the point here, since the editors of The Atlantic have thoughtfully made my original article available on the magazine's Web site (www.TheAtlantic.com), where interested readers can compare what I actually said with Halpern's version of it.

Another point is more important, though: it's a remarkable feature of Halpern's article that he discusses no particular examples of grammatical rules or prescriptive dicta, as if the entire controversy could be resolved in the abstract by appeals to first principles -- as if you could make up your mind about language without ever talking about words. In the event, though, the grammar "war" to which his title alludes is not about to be won by aerial bombardment; what's called for is something more like street-to-street combat. Some prescriptive dicta are mere superstitions, as Fowler put it, while others are solidly based in linguistic principle. Some are the manifestations of covert racism and class prejudice, while others embody a conception of the language as an instrument that overcomes these distinctions. Like any form of criticism, prescriptivism is all a matter of making picky and tiresome discriminations.

It's true that one can find the same sort of programmatic oversimplifications in many of the descriptivist linguists whose views Halpern is attacking (in fact, as one reads his article, one is struck by how much his neo-orthodoxy shares with their heterodoxy). The effect is to turn the debate into just another shrill chapter of the culture wars, as people lose sight of all the subtleties that make a critic like Fowler still worth reading. This is the danger I was alluding to in the title of my own article: "The Decline of Grammar" referred not to worsening standards of usage but to the trivialization of modern discussions of the problem. Unfortunately, Halpern's article does little to allay this concern.

Geoffrey Nunberg

Mark Halpern does little to elucidate the complex interrelations between prescriptive grammar and unruly, evolving usage. It would help enormously if he could acknowledge that standards of "correct" grammar evolve along with language itself -- that "good English" in 1997 is not the same as good English in 1597 or, in all probability, good English in 2397. His article also demonstrates the severely limited power of analogies -- his own as well as those he takes up from others -- to resolve complex questions. For the clincher to his prescriptivist argument he gives us a quotation from Samuel Johnson beginning thus: "Tongues [languages], like governments, have a natural tendency to degeneration." Can this really be Halpern's own understanding of language change -- perpetual decline staved off only by the efforts of prescriptive grammarians?

Languages don't degenerate, but they do reflect and embody social change. Consider the impact of increased sensitivity to gender issues, as exemplified by discontent with the so-called "generic" pronoun "he," used to refer to both males and females. Usage change has been influenced by art in the form of "s/he" and similar coinages, but nature has provided a sort of grassroots rejection of gender bias: most speakers now regularly (and unselfconsciously) use plural pronominal forms ("they," "them," "their"), which do not specify gender, as in "Everyone should follow their own conscience." The existing situation creates a problem in linguistic etiquette of concern to English teachers and, presumably, technical editors like Halpern, but one's preference in such matters ought not to be glorified by high-toned appeals to "the clarity and richness of communication." The American Heritage Dictionary's lengthy usage note on "he," which concludes with the observation that "the entire question is unlikely to be resolved in the near future," reveals much more about the problematic relations between changing usage and prescriptive grammar than the weary, oversimplified dichotomy that Halpern offers your readers.

Phillip J. Hanse

As Geoffrey Nunberg is well aware, I regard his position on usage as far more complex, interesting, and therefore dangerous than that of the simple leave-the-language-alone school.

He objects that I offer no examples of "grammatical rules or prescriptive dicta"; he is missing my point. My article is an attack on the claims of linguistic scholars to a privileged place at the table where usage decisions are made, not a plea for any particular rules or an argument that rules in general will save us. Rules have their uses, but only for the already literate; in the absence of a foundation of taste and judgment, rules may do more harm than good ("the wise man points at the stars; the fool looks at his finger").

While asking me to offer concrete examples in support of a position I do not hold, Numberg fails to support his own main contention: that linguistic science can settle or at least shed much light on usage issues. What linguists can often do is demolish the pretensions to scholarship of "pop grammarians" or "mavens" when they are so foolish as to try to buttress their judgments on usage with their usually skimpy and obsolete stock of linguistic knowledge. But once the linguists have done that demolition work and the dust has settled, the usage issues remain -- and linguistic science does nothing to settle them.

Phillip Hanse is correct in stating that I did "little to elucidate the complex interrelations between prescriptive grammar and unruly, evolving usage." It was no part of my intention to do so. If it would "help enormously" for me to acknowledge that the standards of good usage change as time goes by, I will gladly do so -- in fact I did just that when I wrote, "The rules we [prescriptivists] propose will need continual review and revision."

Mr. Hanse wants me to understand the "severely limited power of analogies ....to resolve complex questions." I don't think analogies have any power to resolve complex or even simple questions -- they can only illuminate and clarify issues. The analogies that I discussed were all introduced by Nunberg and Pinker, not by me; my sole concern with them was to point out that they support my position, not their authors'.

My final quotation from Samuel Johnson was not meant to clinch my argument -- it has nothing to do with my main argument in fact -- but is there only to show that Pinker was wrong in presenting Johnson as a champion of his position.

Most Popular

His paranoid style paved the road for Trumpism. Now he fears what’s been unleashed.

Glenn Beck looks like the dad in a Disney movie. He’s earnest, geeky, pink, and slightly bulbous. His idea of salty language is bullcrap.

The atmosphere at Beck’s Mercury Studios, outside Dallas, is similarly soothing, provided you ignore the references to genocide and civilizational collapse. In October, when most commentators considered a Donald Trump presidency a remote possibility, I followed audience members onto the set of The Glenn Beck Program, which airs on Beck’s website, theblaze.com. On the way, we passed through a life-size replica of the Oval Office as it might look if inhabited by a President Beck, complete with a portrait of Ronald Reagan and a large Norman Rockwell print of a Boy Scout.

Should you drink more coffee? Should you take melatonin? Can you train yourself to need less sleep? A physician’s guide to sleep in a stressful age.

During residency, Iworked hospital shifts that could last 36 hours, without sleep, often without breaks of more than a few minutes. Even writing this now, it sounds to me like I’m bragging or laying claim to some fortitude of character. I can’t think of another type of self-injury that might be similarly lauded, except maybe binge drinking. Technically the shifts were 30 hours, the mandatory limit imposed by the Accreditation Council for Graduate Medical Education, but we stayed longer because people kept getting sick. Being a doctor is supposed to be about putting other people’s needs before your own. Our job was to power through.

The shifts usually felt shorter than they were, because they were so hectic. There was always a new patient in the emergency room who needed to be admitted, or a staff member on the eighth floor (which was full of late-stage terminally ill people) who needed me to fill out a death certificate. Sleep deprivation manifested as bouts of anger and despair mixed in with some euphoria, along with other sensations I’ve not had before or since. I remember once sitting with the family of a patient in critical condition, discussing an advance directive—the terms defining what the patient would want done were his heart to stop, which seemed likely to happen at any minute. Would he want to have chest compressions, electrical shocks, a breathing tube? In the middle of this, I had to look straight down at the chart in my lap, because I was laughing. This was the least funny scenario possible. I was experiencing a physical reaction unrelated to anything I knew to be happening in my mind. There is a type of seizure, called a gelastic seizure, during which the seizing person appears to be laughing—but I don’t think that was it. I think it was plain old delirium. It was mortifying, though no one seemed to notice.

Why did Trump’s choice for national-security advisor perform so well in the war on terror, only to find himself forced out of the Defense Intelligence Agency?

How does a man like retired Lieutenant General Mike Flynn—who spent his life sifting through information and parsing reports, separating rumor and innuendo from actionable intelligence—come to promote conspiracy theories on social media?

Perhaps it’s less Flynn who’s changed than that the circumstances in which he finds himself—thriving in some roles, and flailing in others.

In diagnostic testing, there’s a basic distinction between sensitivity, or the ability to identify positive results, and specificity, the ability to exclude negative ones. A test with high specificity may avoid generating false positives, but at the price of missing many diagnoses. One with high sensitivity may catch those tricky diagnoses, but also generate false positives along the way. Some people seem to sift through information with high sensitivity, but low specificity—spotting connections that others can’t, and perhaps some that aren’t even there.

“Well, you’re just special. You’re American,” remarked my colleague, smirking from across the coffee table. My other Finnish coworkers, from the school in Helsinki where I teach, nodded in agreement. They had just finished critiquing one of my habits, and they could see that I was on the defensive.

I threw my hands up and snapped, “You’re accusing me of being too friendly? Is that really such a bad thing?”

“Well, when I greet a colleague, I keep track,” she retorted, “so I don’t greet them again during the day!” Another chimed in, “That’s the same for me, too!”

Unbelievable, I thought. According to them, I’m too generous with my hellos.

When I told them I would do my best to greet them just once every day, they told me not to change my ways. They said they understood me. But the thing is, now that I’ve viewed myself from their perspective, I’m not sure I want to remain the same. Change isn’t a bad thing. And since moving to Finland two years ago, I’ve kicked a few bad American habits.

Why the ingrained expectation that women should desire to become parents is unhealthy

In 2008, Nebraska decriminalized child abandonment. The move was part of a "safe haven" law designed to address increased rates of infanticide in the state. Like other safe-haven laws, parents in Nebraska who felt unprepared to care for their babies could drop them off in a designated location without fear of arrest and prosecution. But legislators made a major logistical error: They failed to implement an age limitation for dropped-off children.

Within just weeks of the law passing, parents started dropping off their kids. But here's the rub: None of them were infants. A couple of months in, 36 children had been left in state hospitals and police stations. Twenty-two of the children were over 13 years old. A 51-year-old grandmother dropped off a 12-year-old boy. One father dropped off his entire family -- nine children from ages one to 17. Others drove from neighboring states to drop off their children once they heard that they could abandon them without repercussion.

Democrats who have struggled for years to sell the public on the Affordable Care Act are now confronting a far more urgent task: mobilizing a political coalition to save it.

Even as the party reels from last month’s election defeat, members of Congress, operatives, and liberal allies have turned to plotting a campaign against repealing the law that, they hope, will rival the Tea Party uprising of 2009 that nearly scuttled its passage in the first place. A group of progressive advocacy groups will announce on Friday a coordinated effort to protect the beneficiaries of the Affordable Care Act and stop Republicans from repealing the law without first identifying a plan to replace it.

They don’t have much time to fight back. Republicans on Capitol Hill plan to set repeal of Obamacare in motion as soon as the new Congress opens in January, and both the House and Senate could vote to wind down the law immediately after President-elect Donald Trump takes the oath of office on the 20th.

Trinidad has the highest rate of Islamic State recruitment in the Western hemisphere. How did this happen?

This summer, the so-called Islamic State published issue 15 of its online magazine Dabiq. In what has become a standard feature, it ran an interview with an ISIS foreign fighter. “When I was around twenty years old I would come to accept the religion of truth, Islam,” said Abu Sa’d at-Trinidadi, recalling how he had turned away from the Christian faith he was born into.

At-Trinidadi, as his nom de guerre suggests, is from the Caribbean island of Trinidad and Tobago (T&T), a country more readily associated with calypso and carnival than the “caliphate.” Asked if he had a message for “the Muslims of Trinidad,” he condemned his co-religionists at home for remaining in “a place where you have no honor and are forced to live in humiliation, subjugated by the disbelievers.” More chillingly, he urged Muslims in T&T to wage jihad against their fellow citizens: “Terrify the disbelievers in their own homes and make their streets run with their blood.”

A professor of cognitive science argues that the world is nothing like the one we experience through our senses.

As we go about our daily lives, we tend to assume that our perceptions—sights, sounds, textures, tastes—are an accurate portrayal of the real world. Sure, when we stop and think about it—or when we find ourselves fooled by a perceptual illusion—we realize with a jolt that what we perceive is never the world directly, but rather our brain’s best guess at what that world is like, a kind of internal simulation of an external reality. Still, we bank on the fact that our simulation is a reasonably decent one. If it wasn’t, wouldn’t evolution have weeded us out by now? The true reality might be forever beyond our reach, but surely our senses give us at least an inkling of what it’s really like.

The same part of the brain that allows us to step into the shoes of others also helps us restrain ourselves.

You’ve likely seen the video before: a stream of kids, confronted with a single, alluring marshmallow. If they can resist eating it for 15 minutes, they’ll get two. Some do. Others cave almost immediately.

This “Marshmallow Test,” first conducted in the 1960s, perfectly illustrates the ongoing war between impulsivity and self-control. The kids have to tamp down their immediate desires and focus on long-term goals—an ability that correlates with their later health, wealth, and academic success, and that is supposedly controlled by the front part of the brain. But a new study by Alexander Soutschek at the University of Zurich suggests that self-control is also influenced by another brain region—and one that casts this ability in a different light.

A new survey suggests many might prefer a kind of multipolar Washington, with three distinct orbits of power checking each other.

Does Donald Trump have a mandate?

Though last month’s election provided Trump and his fellow Republicans unified control of the White House, House of Representatives, and Senate for the first time since 2006, the latest Allstate/Atlantic Media Heartland Monitor Poll shows the country remains closely split on many of the key policy challenges facing the incoming administration—and sharply divided on whether they trust the next president to take the lead in responding to them.

In addition, on several important choices facing the new administration and Congress, the survey found that respondents who voted for Trump supported a position that was rejected by the majority of adults overall. That contrast may simultaneously encourage Trump to press forward on an agenda that energizes his coalition, while emboldening congressional Democrats to resist him.