January
28, 2007
By
MICHAEL POLLAN
Eat food. Not
too much. Mostly plants.
That, more or
less, is the short answer to the supposedly incredibly complicated and
confusing question of what we humans should eat in order to be maximally
healthy. I hate to give away the game right here at the beginning of a long
essay, and I confess that I'm tempted to complicate matters in the interest of
keeping things going for a few thousand more words. I'll try to resist but will
go ahead and add a couple more details to flesh out the advice. Like: A little
meat won't kill you, though it's better approached as a side dish than as a
main. And you're much better off eating whole fresh foods than processed food
products. That's what I mean by the recommendation to eat ''food.'' Once, food
was all you could eat, but today there are lots of other edible foodlike substances in the supermarket. These novel
products of food science often come in packages festooned with health claims,
which brings me to a related rule of thumb: if you're concerned about your
health, you should probably avoid food products that make health claims. Why?
Because a health claim on a food product is a good indication that it's not
really food, and food is what you want to eat.
Uh-oh. Things are suddenly sounding a little more complicated,
aren't they? Sorry. But that's how it goes as soon as you try to get to the
bottom of the whole vexing question of food and health. Before long, a dense
cloud bank of confusion moves in. Sooner or later, everything solid you thought
you knew about the links between diet and health gets blown away in the gust of
the latest study.
Last winter
came the news that a low-fat diet, long believed to protect against breast
cancer, may do no such thing -- this from the monumental, federally financed
Women's Health Initiative, which has also found no link between a low-fat diet
and rates of coronary disease. The year before we learned that dietary fiber
might not, as we had been confidently told, help prevent colon cancer. Just
last fall two prestigious studies on omega-3 fats published at the same time
presented us with strikingly different conclusions. While the Institute of
Medicine stated that ''it is uncertain how much these omega-3s contribute to
improving health'' (and they might do the opposite if you get them from
mercury-contaminated fish), a Harvard study declared that simply by eating a
couple of servings of fish each week (or by downing enough fish oil), you could
cut your risk of dying from a heart attack by more than a third -- a stunningly
hopeful piece of news. It's no wonder that omega-3 fatty acids are poised to
become the oat bran of 2007, as food scientists micro-encapsulate fish oil and
algae oil and blast them into such formerly all-terrestrial foods as bread and
tortillas, milk and yogurt and cheese, all of which will soon, you can be sure,
sprout fishy new health claims. (Remember the rule?)
By now you're
probably registering the cognitive dissonance of the supermarket shopper or
science-section reader, as well as some nostalgia for the simplicity and
solidity of the first few sentences of this essay. Which I'm still prepared to
defend against the shifting winds of nutritional science and food-industry
marketing. But before I do that, it might be useful to figure out how we
arrived at our present state of nutritional confusion and anxiety.
The story of
how the most basic questions about what to eat ever got so complicated reveals
a great deal about the institutional imperatives of the food industry,
nutritional science and -- ahem -- journalism, three parties that stand to gain
much from widespread confusion surrounding what is, after all, the most
elemental question an omnivore confronts. Humans deciding what to eat without
expert help -- something they have been doing with notable success since coming
down out of the trees -- is seriously unprofitable if you're a food company,
distinctly risky if you're a nutritionist and just plain boring if you're a
newspaper editor or journalist. (Or, for that matter, an
eater. Who wants to hear, yet again, ''Eat more fruits and
vegetables''?) And so, like a large gray fog, a great Conspiracy of Confusion
has gathered around the simplest questions of nutrition -- much to the
advantage of everybody involved. Except perhaps the ostensible
beneficiary of all this nutritional expertise and advice: us, and our health
and happiness as eaters.
FROM FOODS TO NUTRIENTS
It was in the
1980s that food began disappearing from the American supermarket, gradually to
be replaced by ''nutrients,'' which are not the same thing. Where once the
familiar names of recognizable comestibles -- things like eggs or breakfast
cereal or cookies -- claimed pride of place on the brightly colored packages
crowding the aisles, now new terms like ''fiber'' and ''cholesterol'' and
''saturated fat'' rose to large-type prominence. More important than mere
foods, the presence or absence of these invisible substances was now generally
believed to confer health benefits on their eaters. Foods by comparison were
coarse, old-fashioned and decidedly unscientific things -- who could say what
was in them, really? But nutrients -- those chemical compounds and minerals in
foods that nutritionists have deemed important to health -- gleamed with the
promise of scientific certainty; eat more of the right ones, fewer of the
wrong, and you would live longer and avoid chronic diseases.
Nutrients
themselves had been around, as a concept, since the early 19th century, when
the English doctor and chemist William Prout
identified what came to be called the ''macronutrients'': protein, fat and
carbohydrates. It was thought that that was pretty much all there was going on
in food, until doctors noticed that an adequate supply of the big three did not
necessarily keep people nourished. At the end of the 19th century, British
doctors were puzzled by the fact that Chinese laborers in the Malay states were
dying of a disease called beriberi, which didn't seem to afflict Tamils or
native Malays. The mystery was solved when someone pointed out that the Chinese
ate ''polished,'' or white, rice, while the others ate rice that hadn't been
mechanically milled. A few years later, Casimir Funk,
a Polish chemist, discovered the ''essential nutrient'' in rice husks that
protected against beriberi and called it a ''vitamine,''
the first micronutrient. Vitamins brought a kind of glamour to the science of
nutrition, and though certain sectors of the population began to eat by its
expert lights, it really wasn't until late in the 20th century that nutrients
managed to push food aside in the popular imagination of what it means to eat.
No single event
marked the shift from eating food to eating nutrients, though in retrospect a
little-noticed political dust-up in
Naïvely putting
two and two together, the committee drafted a straightforward set of dietary
guidelines calling on Americans to cut down on red meat and dairy products.
Within weeks a firestorm, emanating from the red-meat and dairy industries,
engulfed the committee, and Senator McGovern (who had a great many cattle
ranchers among his
A subtle change
in emphasis, you might say, but a world of difference just the same. First, the
stark message to ''eat less'' of a particular food has been deep-sixed; don't look for it ever again in any official
The linguistic
capitulation did nothing to rescue McGovern from his blunder; the very next
election, in 1980, the beef lobby helped rusticate the three-term senator,
sending an unmistakable warning to anyone who would challenge the American
diet, and in particular the big chunk of animal protein sitting in the middle
of its plate. Henceforth, government dietary guidelines would shun plain talk
about whole foods, each of which has its trade association on Capitol Hill, and
would instead arrive clothed in scientific euphemism and speaking of nutrients,
entities that few Americans really understood but that lack powerful lobbies in
Washington. This was precisely the tack taken by the National Academy of
Sciences when it issued its landmark report on diet and cancer in 1982.
Organized nutrient by nutrient in a way guaranteed to offend no food group, it
codified the official new dietary language. Industry and media followed suit,
and terms like polyunsaturated, cholesterol, monounsaturated, carbohydrate,
fiber, polyphenols, amino acids and carotenes soon
colonized much of the cultural space previously occupied by the tangible
substance formerly known as food. The Age of Nutritionism
had arrived.
THE RISE OF NUTRITIONISM
The first thing
to understand about nutritionism -- I first
encountered the term in the work of an Australian sociologist of science named Gyorgy Scrinis -- is that it is
not quite the same as nutrition. As the ''ism'' suggests, it is not a
scientific subject but an ideology. Ideologies are ways of organizing large
swaths of life and experience under a set of shared but unexamined assumptions.
This quality makes an ideology particularly hard to see, at least while it's
exerting its hold on your culture. A reigning ideology is a little like the
weather, all pervasive and virtually inescapable. Still, we can try.
In the case of nutritionism, the widely shared but unexamined assumption
is that the key to understanding food is indeed the nutrient. From this basic
premise flow several others. Since nutrients, as compared with foods, are
invisible and therefore slightly mysterious, it falls to the scientists (and to
the journalists through whom the scientists speak) to explain the hidden
reality of foods to us. To enter a world in which you dine on unseen nutrients,
you need lots of expert help.
But expert help
to do what, exactly? This brings us to another unexamined assumption: that the
whole point of eating is to maintain and promote bodily health. Hippocrates's
famous injunction to ''let food be thy medicine'' is ritually invoked to
support this notion. I'll leave the premise alone for now, except to point out
that it is not shared by all cultures and that the experience of these other
cultures suggests that, paradoxically, viewing food as being about things other
than bodily health -- like pleasure, say, or socializing -- makes people no
less healthy; indeed, there's some reason to believe that it may make them more
healthy. This is what we usually have in mind when we speak of the ''French
paradox'' -- the fact that a population that eats all sorts of unhealthful
nutrients is in many ways healthier than we Americans are. So there is at least
a question as to whether nutritionism is actually any
good for you.
Another
potentially serious weakness of nutritionist ideology is that it has trouble
discerning qualitative distinctions between foods. So fish, beef and chicken
through the nutritionists' lens become mere delivery systems for varying
quantities of fats and proteins and whatever other nutrients are on their
scope. Similarly, any qualitative distinctions between processed foods and
whole foods disappear when your focus is on quantifying the nutrients they
contain (or, more precisely, the known nutrients).
This is a great
boon for manufacturers of processed food, and it helps explain why they have
been so happy to get with the nutritionism program.
In the years following McGovern's capitulation and the 1982
By comparison,
the typical real food has more trouble competing under the rules of nutritionism, if only because something like a banana or an
avocado can't easily change its nutritional stripes (though rest assured the
genetic engineers are hard at work on the problem). So far, at least, you can't
put oat bran in a banana. So depending on the reigning nutritional orthodoxy,
the avocado might be either a high-fat food to be avoided (Old Think) or a food
high in monounsaturated fat to be embraced (New Think). The fate of each whole
food rises and falls with every change in the nutritional weather, while the
processed foods are simply reformulated. That's why when the Atkins mania hit
the food industry, bread and pasta were given a quick redesign (dialing back
the carbs; boosting the protein), while the poor
unreconstructed potatoes and carrots were left out in the cold.
Of course it's
also a lot easier to slap a health claim on a box of sugary cereal than on a
potato or carrot, with the perverse result that the most healthful foods in the
supermarket sit there quietly in the produce section, silent as stroke victims,
while a few aisles over, the Cocoa Puffs and Lucky Charms are screaming about
their newfound whole-grain goodness.
EAT RIGHT, GET FATTER
So nutritionism is good for business. But is it good for us?
You might think that a national fixation on nutrients would lead to measurable
improvements in the public health. But for that to happen, the underlying
nutritional science, as well as the policy recommendations (and the journalism)
based on that science, would have to be sound. This has seldom been the case.
Consider what
happened immediately after the 1977 ''Dietary Goals'' -- McGovern's masterpiece
of politico-nutritionist compromise. In the wake of the panel's recommendation
that we cut down on saturated fat, a recommendation seconded by the 1982
This story has
been told before, notably in these pages (''What if It's All Been a Big Fat
Lie?'' by Gary Taubes, July 7, 2002), but it's a
little more complicated than the official version suggests. In that version,
which inspired the most recent Atkins craze, we were told that
But there are a
couple of problems with this revisionist picture. First, while it is true that
Americans post-1977 did begin binging on carbs, and
that fat as a percentage of total calories in the American diet declined, we
never did in fact cut down on our consumption of fat. Meat consumption actually
climbed. We just heaped a bunch more carbs onto our
plates, obscuring perhaps, but not replacing, the expanding chunk of animal
protein squatting in the center.
How did that
happen? I would submit that the ideology of nutritionism
deserves as much of the blame as the carbohydrates themselves do -- that and
human nature. By framing dietary advice in terms of good and bad nutrients, and
by burying the recommendation that we should eat less of any particular food,
it was easy for the take-home message of the 1977 and 1982 dietary guidelines
to be simplified as follows: Eat more low-fat foods. And that is what we did.
We're always happy to receive a dispensation to eat more of something (with the
possible exception of oat bran), and one of the things nutritionism
reliably gives us is some such dispensation: low-fat cookies then, low-carb beer now. It's hard to imagine the low-fat craze
taking off as it did if McGovern's original food-based recommendations had
stood: eat fewer meat and dairy products. For how do you get from that stark
counsel to the idea that another case of Snackwell's
is just what the doctor ordered?
BAD SCIENCE
But if nutritionism leads to a kind of false consciousness in the
mind of the eater, the ideology can just as easily mislead the scientist. Most
nutritional science involves studying one nutrient at a time, an approach that
even nutritionists who do it will tell you is deeply
flawed. ''The problem with nutrient-by-nutrient nutrition science,'' points out
Marion Nestle, the
If nutritional
scientists know this, why do they do it anyway? Because a nutrient bias is
built into the way science is done: scientists need individual variables they
can isolate. Yet even the simplest food is a hopelessly complex thing to study,
a virtual wilderness of chemical compounds, many of which exist in complex and
dynamic relation to one another, and all of which together are in the process
of changing from one state to another. So if you're a nutritional scientist,
you do the only thing you can do, given the tools at your disposal: break the
thing down into its component parts and study those one by one, even if that
means ignoring complex interactions and contexts, as well as the fact that the
whole may be more than, or just different from, the sum of its parts. This is
what we mean by reductionist science.
Scientific
reductionism is an undeniably powerful tool, but it can mislead us too,
especially when applied to something as complex as, on the one side, a food,
and on the other, a human eater. It encourages us to take a mechanistic view of
that transaction: put in this nutrient; get out that physiological result. Yet
people differ in important ways. Some populations can metabolize sugars better
than others; depending on your evolutionary heritage, you may or may not be
able to digest the lactose in milk. The specific ecology of your intestines
helps determine how efficiently you digest what you eat, so that the same input
of 100 calories may yield more or less energy depending on the proportion of Firmicutes and Bacteroidetes
living in your gut. There is nothing very machinelike about the human eater,
and so to think of food as simply fuel is wrong.
Also, people
don't eat nutrients, they eat foods, and foods can behave very differently than
the nutrients they contain. Researchers have long believed, based on
epidemiological comparisons of different populations, that a diet high in
fruits and vegetables confers some protection against cancer. So naturally they
ask, What nutrients in those plant foods are
responsible for that effect? One hypothesis is that the antioxidants in fresh
produce -- compounds like beta carotene, lycopene,
vitamin E, etc. -- are the X factor. It makes good sense: these molecules
(which plants produce to protect themselves from the highly reactive oxygen
atoms produced in photosynthesis) vanquish the free radicals in our bodies,
which can damage DNA and initiate cancers. At least that's how it seems to work
in the test tube. Yet as soon as you remove these useful molecules from the
context of the whole foods they're found in, as we've done in creating
antioxidant supplements, they don't work at all. Indeed, in the case of beta
carotene ingested as a supplement, scientists have discovered that it actually
increases the risk of certain cancers. Big oops.
What's going on
here? We don't know. It could be the vagaries of human digestion. Maybe the
fiber (or some other component) in a carrot protects the antioxidant molecules
from destruction by stomach acids early in the digestive process. Or it could
be that we isolated the wrong antioxidant. Beta is just one of a whole slew of
carotenes found in common vegetables; maybe we focused on the wrong one. Or
maybe beta carotene works as an antioxidant only in concert with some other plant
chemical or process; under other circumstances, it may behave as a pro-oxidant.
Indeed, to look
at the chemical composition of any common food plant is to realize just how
much complexity lurks within it. Here's a list of just the antioxidants that
have been identified in garden-variety thyme:
4-Terpineol, alanine, anethole, apigenin, ascorbic acid, beta carotene, caffeic
acid, camphene, carvacrol, chlorogenic
acid, chrysoeriol, eriodictyol,
eugenol, ferulic acid, gallic acid, gamma-terpinene isochlorogenic acid, isoeugenol, isothymonin, kaempferol, labiatic acid, lauric acid, linalyl acetate, luteolin, methionine, myrcene, myristic acid, naringenin, oleanolic acid, p-coumoric acid,
p-hydroxy-benzoic acid, palmitic
acid, rosmarinic acid, selenium, tannin, thymol, tryptophan, ursolic acid, vanillic acid.
This is what
you're ingesting when you eat food flavored with thyme. Some of these chemicals
are broken down by your digestion, but others are going on to do undetermined
things to your body: turning some gene's expression on or off, perhaps, or
heading off a free radical before it disturbs a strand of DNA deep in some
cell. It would be great to know how this all works, but in the meantime we can
enjoy thyme in the knowledge that it probably doesn't do any harm (since people
have been eating it forever) and that it may actually do some good (since
people have been eating it forever) and that even if it does nothing, we like
the way it tastes.
It's also
important to remind ourselves that what reductive science can manage to
perceive well enough to isolate and study is subject to change, and that we
have a tendency to assume that what we can see is all there is to see. When
William Prout isolated the big three macronutrients,
scientists figured they now understood food and what the body needs from it;
when the vitamins were isolated a few decades later, scientists thought, O.K.,
now we really understand food and what the body needs to be healthy; today it's
the polyphenols and carotenoids
that seem all-important. But who knows what the hell else is going on deep in
the soul of a carrot?
The good news
is that, to the carrot eater, it doesn't matter. That's the great thing about
eating food as compared with nutrients: you don't need to fathom a carrot's
complexity to reap its benefits.
The case of the
antioxidants points up the dangers in taking a nutrient out of the context of
food; as Nestle suggests, scientists make a second, related error when they
study the food out of the context of the diet. We don't eat just one thing, and
when we are eating any one thing, we're not eating another. We also eat foods
in combinations and in orders that can affect how they're absorbed. Drink
coffee with your steak, and your body won't be able to fully absorb the iron in
the meat. The trace of limestone in the corn tortilla unlocks essential amino
acids in the corn that would otherwise remain unavailable. Some of those
compounds in that sprig of thyme may well affect my digestion of the dish I add
it to, helping to break down one compound or possibly stimulate production of
an enzyme to detoxify another. We have barely begun to understand the
relationships among foods in a cuisine.
But we do
understand some of the simplest relationships, like the zero-sum relationship:
that if you eat a lot of meat you're probably not eating a lot of vegetables.
This simple fact may explain why populations that eat diets high in meat have
higher rates of coronary heart disease and cancer than those that don't. Yet nutritionism encourages us to look elsewhere for the
explanation: deep within the meat itself, to the culpable nutrient, which
scientists have long assumed to be the saturated fat. So they are baffled when
large-population studies, like the Women's Health Initiative, fail to find that
reducing fat intake significantly reduces the incidence of heart disease or
cancer.
Of course
thanks to the low-fat fad (inspired by the very same reductionist
fat hypothesis), it is entirely possible to reduce your intake of saturated fat
without significantly reducing your consumption of animal protein: just drink
the low-fat milk and order the skinless chicken breast or the turkey bacon. So
maybe the culprit nutrient in meat and dairy is the animal protein itself, as
some researchers now hypothesize. (The Cornell nutritionist T. Colin Campbell
argues as much in his recent book, ''The China Study.'') Or, as the Harvard
epidemiologist Walter C. Willett suggests, it could be the steroid hormones
typically present in the milk and meat; these hormones (which occur naturally
in meat and milk but are often augmented in industrial production) are known to
promote certain cancers.
But people
worried about their health needn't wait for scientists to settle this question
before deciding that it might be wise to eat more plants and less meat. This is
of course precisely what the McGovern committee was trying to tell us.
Nestle also
cautions against taking the diet out of the context of the lifestyle. The
Mediterranean diet is widely believed to be one of the most healthful ways to
eat, yet much of what we know about it is based on studies of people living on
the
But if
confounding factors of lifestyle bedevil comparative studies of different
populations, the supposedly more rigorous ''prospective'' studies of large
American populations suffer from their own arguably even more disabling flaws.
In these studies -- of which the Women's Health Initiative is the best known --
a large population is divided into two groups. The intervention group changes
its diet in some prescribed manner, while the control group does not. The two
groups are then tracked over many years to learn whether the intervention
affects relative rates of chronic disease.
When it comes
to studying nutrition, this sort of extensive, long-term clinical trial is
supposed to be the gold standard. It certainly sounds sound. In the case of the
Women's Health Initiative, sponsored by the National Institutes of Health, the
eating habits and health outcomes of nearly 49,000 women (ages 50 to 79 at the
beginning of the study) were tracked for eight years. One group of the women were told to reduce their consumption of fat to 20 percent
of total calories. The results were announced early last year, producing
front-page headlines of which the one in this newspaper was typical: ''Low-Fat
Diet Does Not Cut Health Risks, Study Finds.'' And the cloud of nutritional
confusion over the country darkened.
But even a
cursory analysis of the study's methods makes you wonder why anyone would take
such a finding seriously, let alone order a Quarter Pounder
With Cheese to celebrate it, as many newspaper readers no doubt promptly went
out and did. Even the beginner student of nutritionism
will immediately spot several flaws: the focus was on ''fat,'' rather than on
any particular food, like meat or dairy. So women could comply simply by
switching to lower-fat animal products. Also, no distinctions were made between
types of fat: women getting their allowable portion of fat from olive oil or
fish were lumped together with woman getting their fat from low-fat cheese or
chicken breasts or margarine. Why? Because when the study was designed 16 years
ago, the whole notion of ''good fats'' was not yet on the scientific scope.
Scientists study what scientists can see.
But perhaps the
biggest flaw in this study, and other studies like it, is that we have no idea
what these women were really eating because, like most people when asked about
their diet, they lied about it. How do we know this? Deduction.
Consider: When the study began, the average participant weighed in at 170
pounds and claimed to be eating 1,800 calories a day. It would take an unusual
metabolism to maintain that weight on so little food. And it would take an even
freakier metabolism to drop only one or two pounds after getting down to a diet
of 1,400 to 1,500 calories a day -- as the women on the ''low-fat'' regimen
claimed to have done. Sorry, ladies, but I just don't buy it.
In fact, nobody
buys it. Even the scientists who conduct this sort of research conduct it in
the knowledge that people lie about their food intake all the time. They even
have scientific figures for the magnitude of the lie. Dietary trials like the
Women's Health Initiative rely on ''food-frequency questionnaires,'' and
studies suggest that people on average eat between a fifth and a third more
than they claim to on the questionnaires. How do the researchers know that? By
comparing what people report on questionnaires with interviews about their
dietary intake over the previous 24 hours, thought to be somewhat more
reliable. In fact, the magnitude of the lie could be much greater, judging by
the huge disparity between the total number of food calories produced every day
for each American (3,900 calories) and the average number of those calories
Americans own up to chomping: 2,000. (Waste accounts for some of the disparity,
but nowhere near all of it.) All we really know about how much people actually
eat is that the real number lies somewhere between those two figures.
To try to fill
out the food-frequency questionnaire used by the Women's Health Initiative, as
I recently did, is to realize just how shaky the data on which such trials rely
really are. The survey, which took about 45 minutes to complete, started off
with some relatively easy questions: ''Did you eat chicken or turkey during the
last three months?'' Having answered yes, I was then asked, ''When you ate
chicken or turkey, how often did you eat the skin?'' But the survey soon became
harder, as when it asked me to think back over the past three months to recall
whether when I ate okra, squash or yams, they were fried, and if so, were they
fried in stick margarine, tub margarine, butter, ''shortening'' (in which
category they inexplicably lump together hydrogenated vegetable oil and lard),
olive or canola oil or nonstick spray? I honestly didn't remember, and in the
case of any okra eaten in a restaurant, even a hypnotist could not get out of
me what sort of fat it was fried in. In the meat section, the portion sizes
specified haven't been seen in
This is the
sort of data on which the largest questions of diet and health are being
decided in
THE ELEPHANT IN THE ROOM
In the end, the
biggest, most ambitious and widely reported studies of diet and health leave
more or less undisturbed the main features of the Western diet: lots of meat
and processed foods, lots of added fat and sugar, lots of everything -- except
fruits, vegetables and whole grains. In keeping with the nutritionism
paradigm and the limits of reductionist science, the
researchers fiddle with single nutrients as best they can, but the populations
they recruit and study are typical American eaters doing what typical American
eaters do: trying to eat a little less of this nutrient, a little more of that,
depending on the latest thinking. (One problem with the control groups in these
studies is that they too are exposed to nutritional fads in the culture, so
over time their eating habits come to more closely resemble the habits of the
intervention group.) It should not surprise us that the findings of such
research would be so equivocal and confusing.
But what about the elephant in the room -- the Western diet? It might be useful, in
the midst of our deepening confusion about nutrition, to review what we do know
about diet and health. What we know is that people who eat the way we do in
America today suffer much higher rates of cancer, heart disease, diabetes and
obesity than people eating more traditional diets. (Four of the 10 leading
killers in
No one likes to
admit that his or her best efforts at understanding and solving a problem have
actually made the problem worse, but that's exactly what has happened in the
case of nutritionism. Scientists operating with the
best of intentions, using the best tools at their disposal, have taught us to
look at food in a way that has diminished our pleasure in eating it while doing
little or nothing to improve our health. Perhaps what we need now is a broader,
less reductive view of what food is, one that is at once more ecological and
cultural. What would happen, for example, if we were to start thinking about
food as less of a thing and more of a relationship?
In nature, that
is of course precisely what eating has always been: relationships among species
in what we call food chains, or webs, that reach all
the way down to the soil. Species co-evolve with the other species they eat,
and very often a relationship of interdependence develops: I'll feed you if you
spread around my genes. A gradual process of mutual adaptation transforms
something like an apple or a squash into a nutritious and tasty food for a
hungry animal. Over time and through trial and error, the plant becomes tastier
(and often more conspicuous) in order to gratify the animal's needs and
desires, while the animal gradually acquires whatever digestive tools (enzymes,
etc.) are needed to make optimal use of the plant. Similarly, cow's milk did
not start out as a nutritious food for humans; in fact, it made them sick until
humans who lived around cows evolved the ability to digest lactose as adults.
This development proved much to the advantage of both the milk drinkers and the
cows.
''Health'' is,
among other things, the byproduct of being involved in these sorts of
relationships in a food chain -- involved in a great many of them, in the case
of an omnivorous creature like us. Further, when the health of one link of the
food chain is disturbed, it can affect all the creatures in it. When the soil
is sick or in some way deficient, so will be the grasses that grow in that soil
and the cattle that eat the grasses and the people who drink the milk. Or, as
the English agronomist Sir Albert Howard put it in 1945 in ''The Soil and
Health'' (a founding text of organic agriculture), we would do well to regard
''the whole problem of health in soil, plant, animal and man as one great
subject.'' Our personal health is inextricably bound up with the health of the
entire food web.
In many cases,
long familiarity between foods and their eaters leads to elaborate systems of
communications up and down the food chain, so that a creature's senses come to
recognize foods as suitable by taste and smell and color, and our bodies learn
what to do with these foods after they pass the test of the senses, producing
in anticipation the chemicals necessary to break them down. Health depends on
knowing how to read these biological signals: this smells spoiled; this looks
ripe; that's one good-looking cow. This is easier to do when a creature has
long experience of a food, and much harder when a food has been designed
expressly to deceive its senses -- with artificial flavors, say, or synthetic
sweeteners.
Note that these
ecological relationships are between eaters and whole foods, not nutrients.
Even though the foods in question eventually get broken down in our bodies into
simple nutrients, as corn is reduced to simple sugars, the qualities of the
whole food are not unimportant -- they govern such things as the speed at which
the sugars will be released and absorbed, which we're coming to see as critical
to insulin metabolism. Put another way, our bodies have a longstanding and
sustainable relationship to corn that we do not have to high-fructose corn
syrup. Such a relationship with corn syrup might develop someday (as people
evolve superhuman insulin systems to cope with regular floods of fructose and
glucose), but for now the relationship leads to ill health because our bodies
don't know how to handle these biological novelties. In much the same way,
human bodies that can cope with chewing coca leaves -- a longstanding
relationship between native people and the coca plant in
Looking at
eating through this ecological lens opens a whole new perspective on exactly
what the Western diet is: a radical and rapid change not just in our foodstuffs
over the course of the 20th century but also in our food relationships, all the
way from the soil to the meal. The ideology of nutritionism
is itself part of that change. To get a firmer grip on the nature of those
changes is to begin to know how we might make our relationships to food
healthier. These changes have been numerous and far-reaching, but consider as a
start these four large-scale ones:
From Whole Foods to Refined. The case of corn points up one of the
key features of the modern diet: a shift toward increasingly refined foods,
especially carbohydrates. Call it applied reductionism. Humans have been
refining grains since at least the Industrial Revolution, favoring white flour
(and white rice) even at the price of lost nutrients. Refining grains extends
their shelf life (precisely because it renders them less nutritious to pests)
and makes them easier to digest, by removing the fiber that ordinarily slows
the release of their sugars. Much industrial food production involves an
extension and intensification of this practice, as food processors find ways to
deliver glucose -- the brain's preferred fuel -- ever more swiftly and
efficiently. Sometimes this is precisely the point, as when corn is refined
into corn syrup; other times it is an unfortunate byproduct of food processing,
as when freezing food destroys the fiber that would slow sugar absorption.
So fast food is fast in this other sense too: it is to a
considerable extent predigested, in effect, and therefore more readily absorbed
by the body. But while the widespread acceleration of the Western diet offers
us the instant gratification of sugar, in many people (and especially those
newly exposed to it) the ''speediness'' of this food overwhelms the insulin response
and leads to Type II diabetes. As one nutrition expert put it to me, we're in
the middle of ''a national experiment in mainlining glucose.'' To encounter
such a diet for the first time, as when people accustomed to a more traditional
diet come to
From Complexity to Simplicity. If there is one word
that covers nearly all the changes industrialization has made to the food
chain, it would be simplification. Chemical fertilizers simplify the chemistry
of the soil, which in turn appears to simplify the chemistry of the food grown
in that soil. Since the widespread adoption of synthetic nitrogen fertilizers
in the 1950s, the nutritional quality of produce in
Simplification
has occurred at the level of species diversity, too. The astounding variety of
foods on offer in the modern supermarket obscures the fact that the actual
number of species in the modern diet is shrinking. For reasons of economics,
the food industry prefers to tease its myriad processed offerings from a tiny
group of plant species, corn and soybeans chief among them. Today, a mere four
crops account for two-thirds of the calories humans eat. When you consider that
humankind has historically consumed some 80,000 edible species, and that 3,000
of these have been in widespread use, this represents a radical simplification
of the food web. Why should this matter? Because humans are omnivores,
requiring somewhere between 50 and 100 different chemical compounds and
elements to be healthy. It's hard to believe that we can get everything we need
from a diet consisting largely of processed corn, soybeans, wheat and rice.
From Leaves to Seeds. It's no coincidence that most of the plants we
have come to rely on are grains; these crops are exceptionally efficient at
transforming sunlight into macronutrients -- carbs,
fats and proteins. These macronutrients in turn can be profitably transformed
into animal protein (by feeding them to animals) and processed foods of every
description. Also, the fact that grains are durable seeds that can be stored
for long periods means they can function as commodities as well as food, making
these plants particularly well suited to the needs of industrial capitalism.
The needs of
the human eater are another matter. An oversupply of macronutrients, as we now
have, itself represents a serious threat to our health, as evidenced by soaring
rates of obesity and diabetes. But the undersupply of micronutrients may
constitute a threat just as serious. Put in the simplest terms, we're eating a
lot more seeds and a lot fewer leaves, a tectonic dietary shift the full
implications of which we are just beginning to glimpse. If I may borrow the
nutritionist's reductionist vocabulary for a moment,
there are a host of critical micronutrients that are harder to get from a diet
of refined seeds than from a diet of leaves. There are the antioxidants and all
the other newly discovered phytochemicals (remember
that sprig of thyme?); there is the fiber, and then there are the healthy
omega-3 fats found in leafy green plants, which may turn out to be most
important benefit of all.
Most people
associate omega-3 fatty acids with fish, but fish get them from green plants
(specifically algae), which is where they all originate. Plant leaves produce
these essential fatty acids (''essential'' because our bodies can't produce
them on their own) as part of photosynthesis. Seeds contain more of another
essential fatty acid: omega-6. Without delving too deeply into the
biochemistry, the two fats perform very different functions, in the plant as
well as the plant eater. Omega-3s appear to play an important role in
neurological development and processing, the permeability of cell walls, the
metabolism of glucose and the calming of inflammation. Omega-6s are involved in
fat storage (which is what they do for the plant), the rigidity of cell walls,
clotting and the inflammation response. (Think of omega-3s as fleet and
flexible, omega-6s as sturdy and slow.) Since the two lipids compete with each
other for the attention of important enzymes, the ratio between omega-3s and
omega-6s may matter more than the absolute quantity of either fat. Thus too
much omega-6 may be just as much a problem as too little omega-3.
And that might
well be a problem for people eating a Western diet. As we've shifted from
leaves to seeds, the ratio of omega-6s to omega-3s in our bodies has shifted,
too. At the same time, modern food-production practices have further diminished
the omega-3s in our diet. Omega-3s, being less stable than omega-6s, spoil more
readily, so we have selected for plants that produce fewer of them; further,
when we partly hydrogenate oils to render them more stable, omega-3s are
eliminated. Industrial meat, raised on seeds rather than leaves, has fewer
omega-3s and more omega-6s than preindustrial meat
used to have. And official dietary advice since the 1970s has promoted the
consumption of polyunsaturated vegetable oils, most of which are high in
omega-6s (corn and soy, especially). Thus, without realizing what we were
doing, we significantly altered the ratio of these two essential fats in our
diets and bodies, with the result that the ratio of omega-6 to omega-3 in the
typical American today stands at more than 10 to 1; before the widespread
introduction of seed oils at the turn of the last century, it was closer to 1
to 1.
The role of
these lipids is not completely understood, but many researchers say that these
historically low levels of omega-3 (or, conversely, high levels of omega-6)
bear responsibility for many of the chronic diseases associated with the
Western diet, especially heart disease and diabetes. (Some researchers
implicate omega-3 deficiency in rising rates of depression and learning
disabilities as well.) To remedy this deficiency, nutritionism
classically argues for taking omega-3 supplements or fortifying food products,
but because of the complex, competitive relationship between omega-3 and
omega-6, adding more omega-3s to the diet may not do much good unless you also
reduce your intake of omega-6.
From Food Culture to Food Science. The last important
change wrought by the Western diet is not, strictly speaking, ecological. But
the industrialization of our food that we call the Western diet is
systematically destroying traditional food cultures. Before the modern food era
-- and before nutritionism -- people relied for
guidance about what to eat on their national or ethnic or regional cultures. We
think of culture as a set of beliefs and practices to help mediate our
relationship to other people, but of course culture (at least before the rise
of science) has also played a critical role in helping mediate people's
relationship to nature. Eating being a big part of that relationship, cultures
have had a great deal to say about what and how and why and when and how much
we should eat. Of course when it comes to food, culture is really just a fancy
word for Mom, the figure who typically passes on the food ways of the group --
food ways that, although they were never ''designed'' to optimize health (we
have many reasons to eat the way we do), would not have endured if they did not
keep eaters alive and well.
The sheer
novelty and glamour of the Western diet, with its 17,000 new food products
introduced every year, and the marketing muscle used to sell these products,
has overwhelmed the force of tradition and left us where we now find ourselves:
relying on science and journalism and marketing to help us decide questions
about what to eat. Nutritionism, which arose to help
us better deal with the problems of the Western diet, has largely been co-opted
by it, used by the industry to sell more food and to undermine the authority of
traditional ways of eating. You would not have read this far into this article
if your food culture were intact and healthy; you would simply eat the way your
parents and grandparents and great-grandparents taught you to eat. The question
is, Are we better off with these new authorities than we were with the
traditional authorities they supplanted? The answer by now should be clear.
It might be
argued that, at this point in history, we should simply accept that fast food
is our food culture. Over time, people will get used to eating this way and our
health will improve. But for natural selection to help populations adapt to the
Western diet, we'd have to be prepared to let those whom it sickens die. That's
not what we're doing. Rather, we're turning to the health-care industry to help
us ''adapt.'' Medicine is learning how to keep alive the people whom the
Western diet is making sick. It's gotten good at extending the lives of people
with heart disease, and now it's working on obesity and diabetes. Capitalism is
itself marvelously adaptive, able to turn the problems it creates into
lucrative business opportunities: diet pills, heart-bypass operations, insulin
pumps, bariatric surgery. But while fast food may be
good business for the health-care industry, surely the cost to society --
estimated at more than $200 billion a year in diet-related health-care costs --
is unsustainable.
BEYOND NUTRITIONISM
To medicalize the diet problem is of course perfectly
consistent with nutritionism. So what might a more
ecological or cultural approach to the problem recommend? How might we plot our
escape from nutritionism and, in turn, from the
deleterious effects of the modern diet? In theory nothing could be simpler --
stop thinking and eating that way -- but this is somewhat harder to do in
practice, given the food environment we now inhabit and the loss of sharp
cultural tools to guide us through it. Still, I do think escape is possible, to
which end I can now revisit -- and elaborate on, but just a little -- the
simple principles of healthy eating I proposed at the beginning of this essay,
several thousand words ago. So try these few (flagrantly unscientific) rules of
thumb, collected in the course of my nutritional odyssey, and see if they don't
at least point us in the right direction.
1. Eat food. Though in our current state of confusion, this is much easier said
than done. So try this: Don't eat anything your great-great-grandmother
wouldn't recognize as food. (Sorry, but at this point Moms are as confused as
the rest of us, which is why we have to go back a couple of generations, to a
time before the advent of modern food products.) There are a great many foodlike items in the supermarket your ancestors wouldn't
recognize as food (Go-Gurt? Breakfast-cereal bars? Nondairy creamer?); stay away from these.
2. Avoid even
those food products that come bearing health claims. They're apt to be heavily
processed, and the claims are often dubious at best. Don't forget that
margarine, one of the first industrial foods to claim that it was more
healthful than the traditional food it replaced, turned out to give people
heart attacks. When Kellogg's can boast about its Healthy Heart Strawberry
Vanilla cereal bars, health claims have become hopelessly compromised. (The
American Heart Association charges food makers for their endorsement.) Don't
take the silence of the yams as a sign that they have nothing valuable to say
about health.
3. Especially
avoid food products containing ingredients that are a) unfamiliar, b)
unpronounceable c) more than five in number -- or that contain high-fructose
corn syrup.None of these characteristics are
necessarily harmful in and of themselves, but all of them are reliable markers
for foods that have been highly processed.
4. Get out of
the supermarket whenever possible. You won't find any high-fructose corn syrup
at the farmer's market; you also won't find food harvested long ago and far
away. What you will find are fresh whole foods picked at the peak of
nutritional quality. Precisely the kind of food your great-great-grandmother
would have recognized as food.
5. Pay more,
eat less. The American food system has for a century devoted its energies and
policies to increasing quantity and reducing price, not to improving quality.
There's no escaping the fact that better food -- measured by taste or
nutritional quality (which often correspond) -- costsmore,
because it has been grown or raised less intensively and with more care. Not
everyone can afford to eat well in
''Eat less'' is
the most unwelcome advice of all, but in fact the scientific case for eating a
lot less than we currently do is compelling. ''Calorie restriction'' has
repeatedly been shown to slow aging in animals, and many researchers (including
Walter Willett, the Harvard epidemiologist) believe it offers the single
strongest link between diet and cancer prevention. Food abundance is a problem,
but culture has helped here, too, by promoting the idea of moderation. Once one
of the longest-lived people on earth, the Okinawans
practiced a principle they called ''Hara Hachi Bu'':
eat until you are 80 percent full. To make the ''eat less'' message a bit more
palatable, consider that quality may have a bearing on quantity: I don't know
about you, but the better the quality of the food I eat, the less of it I need
to feel satisfied. All tomatoes are not created equal.
6. Eat mostly
plants, especially leaves. Scientists may disagree on what's so good about
plants -- the antioxidants? Fiber? Omega-3s?
-- but they do agree that they're probably really good
for you and certainly can't hurt. Also, by eating a plant-based diet, you'll be
consuming far fewer calories, since plant foods (except seeds) are typically
less ''energy dense'' than the other things you might eat. Vegetarians are
healthier than carnivores, but near vegetarians (''flexitarians'')
are as healthy as vegetarians. Thomas Jefferson was on to something when he
advised treating meat more as a flavoring than a food.
7. Eat more
like the French. Or the Japanese. Or
the Italians. Or the Greeks. Confounding factors
aside, people who eat according to the rules of a traditional food culture are
generally healthier than we are. Any traditional diet will do: if it weren't a
healthy diet, the people who follow it wouldn't still be around. True, food
cultures are embedded in societies and economies and ecologies, and some of
them travel better than others: Inuit not so well as Italian. In borrowing from
a food culture, pay attention to how a culture eats, as well as to what it
eats. In the case of the French paradox, it may not be the dietary nutrients
that keep the French healthy (lots of saturated fat and alcohol?!) so much as the dietary habits: small portions, no seconds or
snacking, communal meals -- and the serious pleasure taken in eating. (Worrying
about diet can't possibly be good for you.) Let culture be your guide, not
science.
8. Cook. And if
you can, plant a garden. To take part in the intricate and endlessly
interesting processes of providing for our sustenance is the surest way to
escape the culture of fast food and the values implicit in it: that food should
be cheap and easy; that food is fuel and not communion. The culture of the
kitchen, as embodied in those enduring traditions we call cuisines, contains
more wisdom about diet and health than you are apt to find in any nutrition
journal or journalism. Plus, the food you grow yourself contributes to your
health long before you sit down to eat it. So you might want to think about
putting down this article now and picking up a spatula or hoe.
9. Eat like an
omnivore. Try to add new species, not just new foods, to your diet. The greater
the diversity of species you eat, the more likely you are to cover all your
nutritional bases. That of course is an argument from nutritionism,
but there is a better one, one that takes a broader view of ''health.''
Biodiversity in the diet means less monoculture in the fields. What does that
have to do with your health? Everything. The vast
monocultures that now feed us require tremendous amounts of chemical
fertilizers and pesticides to keep from collapsing. Diversifying those fields
will mean fewer chemicals, healthier soils, healthier plants and animals and,
in turn, healthier people. It's all connected, which is another way of saying
that your health isn't bordered by your body and that what's good for the soil
is probably good for you, too.