The Origins of Agriculture – a Biological Perspective and a New Hypothesis

The Origins of Agriculture – a Biological Perspective and a New Hypothesis

agriculture-history

by Greg Wadley & Angus Martin

Department of Zoology, University of Melbourne

Published in Australian Biologist 6: 96 – 105, June 1993

Introduction

What might head a list of the defining characteristics of the human species? While our view of ourselves could hardly avoid highlighting our accomplishments in engineering, art, medicine, space travel and the like, in a more dispassionate assessment agriculture would probably displace all other contenders for top billing. Most of the other achievements of humankind have followed from this one. Almost without exception, all people on earth today are sustained by agriculture. With a minute number of exceptions, no other species is a farmer. Essentially all of the arable land in the world is under cultivation. Yet agriculture began just a few thousand years ago, long after the appearance of anatomically modern humans.

Given the rate and the scope of this revolution in human biology, it is quite extraordinary that there is no generally accepted model accounting for the origin of agriculture. Indeed, an increasing array of arguments over recent years has suggested that agriculture, far from being a natural and upward step, in fact led commonly to a lower quality of life. Hunter-gatherers typically do less work for the same amount of food, are healthier, and are less prone to famine than primitive farmers (Lee & DeVore 1968, Cohen 1977, 1989). A biological assessment of what has been called the puzzle of agriculture might phrase it in simple ethological terms: why was this behaviour (agriculture) reinforced (and hence selected for) if it was not offering adaptive rewards surpassing those accruing to hunter-gathering or foraging economies?

This paradox is responsible for a profusion of models of the origin of agriculture. ‘Few topics in prehistory’, noted Hayden (1990) ‘have engendered as much discussion and resulted in so few satisfying answers as the attempt to explain why hunter/gatherers began to cultivate plants and raise animals. Climatic change, population pressure, sedentism, resource concentration from desertification, girls’ hormones, land ownership, geniuses, rituals, scheduling conflicts, random genetic kicks, natural selection, broad spectrum adaptation and multicausal retreats from explanation have all been proffered to explain domestication. All have major flaws … the data do not accord well with any one of these models. ‘

Recent discoveries of potentially psychoactive substances in certain agricultural products – cereals and milk – suggest an additional perspective on the adoption of agriculture and the behavioural changes (‘civilisation’) that followed it. In this paper we review the evidence for the drug-like properties of these foods, and then show how they can help to solve the biological puzzle just described .

 

The emergence of agriculture and civilisation in the Neolithic

The transition to agriculture

From about 10,000 years ago, groups of people in several areas around the world began to abandon the foraging lifestyle that had been successful, universal and largely unchanged for millennia (Lee & DeVore 1968). They began to gather, then cultivate and settle around, patches of cereal grasses and to domesticate animals for meat, labour, skins and other materials, and milk.

Farming, based predominantly on wheat and barley, first appeared in the Middle East, and spread quickly to western Asia, Egypt and Europe. The earliest civilisations all relied primarily on cereal agriculture. Cultivation of fruit trees began three thousand years later, again in the MiddleEast, and vegetables and other crops followed (Zohari 1986). Cultivation of rice began in Asia about 7000 years ago (Stark 1986).

To this day, for most people, two-thirds of protein and calorie intake is cereal-derived. (In the west, in the twentieth century, cereal consumption has decreased slightly in favour of meat, sugar, fats and so on.) The respective contributions of each cereal to current total world production are: wheat (28 per cent), corn/maize (27 per cent), rice (25 per cent), barley (10 per cent), others (10 per cent) (Pedersen et al. 1989).

The change in the diet due to agriculture

The modern human diet is very different from that of closely related primates and, almost certainly, early hominids (Gordon 1987). Though there is controversy over what humans ate before the development of agriculture, the diet certainly did not include cereals and milk in appreciable quantities. The storage pits and processing tools necessary for significant consumption of cereals did not appear until the Neolithic (Washburn & Lancaster 1968). Dairy products were not available in quantity before the domestication of animals.

The early hominid diet (from about four million years ago), evolving as it did from that of primate ancestors, consisted primarily of fruits, nuts and other vegetable matter, and some meat – items that could be foraged for and eaten with little or no processing. Comparisons of primate and fossil-hominid anatomy, and of the types and distribution of plants eaten raw by modern chimpanzees, baboons and humans (Peters & O’Brien 1981, Kay 1985), as well as microscope analysis of wear patterns on fossil teeth (Walker 1981, Peuch et al.1983) suggest that australopithecines were ‘mainly frugivorous omnivores with a dietary pattern similar to that of modern chimpanzees’ (Susman 1987:171).

The diet of pre-agricultural but anatomically modern humans (from 30,000 years ago) diversified somewhat, but still consisted of meat, fruits, nuts, legumes, edible roots and tubers, with consumption of cereal seeds only increasing towards the end of the Pleistocene (e.g. Constantini 1989 and subsequent chapters in Harris and Hillman 1989).

The rise of civilisation

Within a few thousand years of the adoption of cereal agriculture, the old hunter-gatherer style of social organisation began to decline. Large, hierarchically organised societies appeared, centred around villages and then cities. With the rise of civilisation and the state came socioeconomic classes, job specialisation, governments and armies.

The size of populations living as coordinated units rose dramatically above pre-agricultural norms. While hunter-gatherers lived in egalitarian, autonomous bands of about 20 closely related persons, with at most a tribal level of organisation above that, early agricultural villages had 50 to 200 inhabitants, and early cities 10,000 or more. People ‘had to learn to curb deep-rooted forces which worked for increasing conflict and violence in large groups’ (Pfeiffer 1977:438).

Agriculture and civilisation meant the end of foraging – a subsistence method with shortterm goals and rewards – and the beginning (for most) of regular arduous work, oriented to future payoffs and the demands of superiors. ‘With the coming of large communities, families no longer cultivated the land for themselves and their immediate needs alone, but for strangers and for the future. They worked all day instead of a few hours a day, as hunter-gatherers had done. There were schedules, quotas, overseers, and punishments for slacking off’ (Pfeiffer 1977:21).

 

Explaining the origins of agriculture and civilisation

The phenomena of human agriculture and civilisation are ethologically interesting, because (1) virtually no other species lives this way, and (2) humans did not live this way until relatively recently. Why was this way of life adopted, and why has it become dominant in the human species?

Problems explaining agriculture

Until recent decades, the transition to farming was seen as an inherently progressive one: people learnt that planting seeds caused crops to grow, and this new improved food source led to larger populations, sedentary farm and town life, more leisure time and so to specialisation, writing, technological advances and civilisation. It is now clear that agriculture was adopted despite certain disadvantages of that lifestyle (e.g. Flannery 1973, Henry 1989). There is a substantial literature (e.g. Reed 1977), not only on how agriculture began, but why. Palaeopathological and comparative studies show that health deteriorated in populations that adopted cereal agriculture, returning to pre-agricultural levels only in modem times. This is in part attributable to the spread of infection in crowded cities, but is largely due to a decline in dietary quality that accompanied intensive cereal farming (Cohen 1989). People in many parts of the world remained hunter-gatherers until quite recently; though they were quite aware of the existence and methods of agriculture, they declined to undertake it (Lee & DeVore 1968, Harris 1977). Cohen (1977:141) summarised the problem by asking: ‘If agriculture provides neither better diet, nor greater dietary reliability, nor greater ease, but conversely appears to provide a poorer diet, less reliably, with greater labor costs, why does anyone become a farmer?’

Many explanations have been offered, usually centred around a particular factor that forced the adoption of agriculture, such as environmental or population pressure (for reviews see Rindos 1984, Pryor 1986, Redding 1988, Blumler & Byrne 1991). Each of these models has been criticised extensively, and there is at this time no generally accepted explanation of the origin of agriculture.

Problems explaining civilisation

A similar problem is posed by the post-agricultural appearance, all over the world, of cities and states, and again there is a large literature devoted to explaining it (e.g. Claessen & Skalnik 1978). The major behavioural changes made in adopting the civilised lifestyle beg explanation. Bledsoe (1987:136) summarised the situation thus:

‘There has never been and there is not now agreement on the nature and significance of the rise of civilisation. The questions posed by the problem are simple, yet fundamental. How did civilisation come about? What animus impelled man to forego the independence, intimacies, and invariability of tribal existence for the much larger and more impersonal political complexity we call the state? What forces fused to initiate the mutation that slowly transformed nomadic societies into populous cities with ethnic mixtures, stratified societies, diversified economies and unique cultural forms? Was the advent of civilisation the inevitable result of social evolution and natural laws of progress or was man the designer of his own destiny? Have technological innovations been the motivating force or was it some intangible factor such as religion or intellectual advancement?’

To a very good approximation, every civilisation that came into being had cereal agriculture as its subsistence base, and wherever cereals were cultivated, civilisation appeared. Some hypotheses have linked the two. For example, Wittfogel’s (1957) ‘hydraulic theory’ postulated that irrigation was needed for agriculture, and the state was in turn needed to organise irrigation. But not all civilisations used irrigation, and other possible factors (e.g. river valley placement, warfare, trade, technology, religion, and ecological and population pressure) have not led to a universally accepted model.

 

Pharmacological properties of cereals and milk

Recent research into the pharmacology of food presents a new perspective on these problems.

Exorphins: opioid substances in food

Prompted by a possible link between diet and mental illness, several researchers in the late 1970s began investigating the occurrence of drug-like substances in some common foodstuffs.

Dohan (1966, 1984) and Dohan et al. (1973, 1983) found that symptoms of schizophrenia were relieved somewhat when patients were fed a diet free of cereals and milk. He also found that people with coeliac disease – those who are unable to eat wheat gluten because of higher than normal permeability of the gut – were statistically likely to suffer also from schizophrenia. Research in some Pacific communities showed that schizophrenia became prevalent in these populations only after they became ‘partially westernised and consumed wheat, barley beer, and rice’ (Dohan 1984).

Groups led by Zioudrou (1979) and Brantl (1979) found opioid activity in wheat, maize and barley (exorphins), and bovine and human milk (casomorphin), as well as stimulatory activity in these proteins, and in oats, rye and soy. Cereal exorphin is much stronger than bovine casomorphin, which in turn is stronger than human casomorphin. Mycroft et al. (1982, 1987) found an analogue of MIF-1, a naturally occurring dopaminergic peptide, in wheat and milk. It occurs in no other exogenous protein. (In subsequent sections we use the term exorphin to cover exorphins, casomorphin, and the MIF-1 analogue. Though opioid and dopaminergic substances work in different ways, they are both ‘rewarding’, and thus more or less equivalent for our purposes.)

Since then, researchers have measured the potency of exorphins, showing them to be comparable to morphine and enkephalin (Heubner et al. 1984), determined their amino acid sequences (Fukudome &Yoshikawa 1992), and shown that they are absorbed from the intestine (Svedburg et al.1985) and can produce effects such as analgesia and reduction of anxiety which are usually associated with poppy-derived opioids (Greksch et al.1981, Panksepp et al.1984). Mycroft et al. estimated that 150 mg of the MIF-1 analogue could be produced by normal daily intake of cereals and milk, noting that such quantities are orally active, and half this amount ‘has induced mood alterations in clinically depressed subjects’ (Mycroft et al. 1982:895). (For detailed reviews see Gardner 1985 and Paroli 1988.)

Most common drugs of addiction are either opioid (e.g heroin and morphine) or dopaminergic (e.g. cocaine and amphetamine), and work by activating reward centres in the brain. Hence we may ask, do these findings mean that cereals and milk are chemically rewarding? Are humans somehow ‘addicted’ to these foods?

Problems in interpreting these findings

Discussion of the possible behavioural effects of exorphins, in normal dietary amounts, has been cautious. Interpretations of their significance have been of two types:

– where a pathological effect is proposed (usually by cereal researchers, and related to Dohan’s findings, though see also Ramabadran & Bansinath 1988), and

– where a natural function is proposed (by milk researchers, who suggest that casomorphin may help in mother-infant bonding or otherwise regulate infant development).

We believe that there can be no natural function for ingestion of exorphins by adult humans. It may be that a desire to find a natural function has impeded interpretation (as well as causing attention to focus on milk, where a natural function is more plausible) . It is unlikely that humans are adapted to a large intake of cereal exorphin, because the modern dominance of cereals in the diet is simply too new. If exorphin is found in cow’s milk, then it may have a natural function for cows; similarly, exorphins in human milk may have a function for infants. But whether this is so or not, adult humans do not naturally drink milk of any kind, so any natural function could not apply to them.

Our sympathies therefore lie with the pathological interpretation of exorphins, whereby substances found in cereals and milk are seen as modern dietary abnormalities which may cause schizophrenia, coeliac disease or whatever. But these are serious diseases found in a minority. Can exorphins be having an effect on humankind at large?

Other evidence for ‘drug-like’ effects of these foods

Research into food allergy has shown that normal quantities of some foods can have pharmacological, including behavioural, effects. Many people develop intolerances to particular foods. Various foods are implicated, and a variety of symptoms is produced. (The term ‘intolerance’ rather than allergy is often used, as in many cases the immune system may not be involved (Egger 1988:159). Some intolerance symptoms, such as anxiety, depression, epilepsy, hyperactivity, and schizophrenic episodes involve brain function (Egger 1988, Scadding & Brostoff 1988).

Radcliffe (1982, quoted in 1987:808) listed the foods at fault, in descending order of frequency, in a trial involving 50 people: wheat (more than 70 per cent of subjects reacted in some way to it), milk (60 per cent), egg (35 per cent), corn, cheese, potato, coffee, rice, yeast, chocolate, tea, citrus, oats, pork, plaice, cane, and beef (10 per cent). This is virtually a list of foods that have become common in the diet following the adoption of agriculture, in order of prevalence. The symptoms most commonly alleviated by treatment were mood change (>50 per cent) followed by headache, musculoskeletal and respiratory ailments.

One of the most striking phenomena in these studies is that patients often exhibit cravings, addiction and withdrawal symptoms with regard to these foods (Egger 1988:170, citing Randolph 1978; see also Radcliffe 1987:808-10, 814, Kroker 1987:856, 864, Sprague & Milam 1987:949, 953, Wraith 1987:489, 491). Brostoff and Gamlin (1989:103) estimated that 50 per cent of intolerance patients crave the foods that cause them problems, and experience withdrawal symptoms when excluding those foods from their diet. Withdrawal symptoms are similar to those associated with drug addictions (Radcliffe 1987:808). The possibility that exorphins are involved has been noted (Bell 1987:715), and Brostoff and Gamlin conclude (1989:230):

‘… the results so far suggest that they might influence our mood. There is certainly no question of anyone getting ‘high’ on a glass of milk or a slice of bread – the amounts involved are too small for that – but these foods might induce a sense of comfort and wellbeing, as food-intolerant patients often say they do. There are also other hormone-like peptides in partial digests of food, which might have other effects on the body.’

There is no possibility that craving these foods has anything to do with the popular notion of the body telling the brain what it needs for nutritional purposes. These foods were not significant in the human diet before agriculture, and large quantities of them cannot be necessary for nutrition. In fact, the standard way to treat food intolerance is to remove the offending items from the patient’s diet.

A suggested interpretation of exorphin research

But what are the effects of these foods on normal people? Though exorphins cannot have a naturally selected physiological function in humans, this does not mean that they have no effect. Food intolerance research suggests that cereals and milk, in normal dietary quantities, are capable of affecting behaviour in many people. And if severe behavioural effects in schizophrenics and coeliacs can be caused by higher than normal absorption of peptides, then more subtle effects, which may not even be regarded as abnormal, could be produced in people generally.

The evidence presented so far suggests the following interpretation.

The ingestion of cereals and milk, in normal modern dietary amounts by normal humans, activates reward centres in the brain. Foods that were common in the diet before agriculture (fruits and so on) do not have this pharmacological property. The effects of exorphins are qualitatively the same as those produced by other opioid and / or dopaminergic drugs, that is, reward, motivation, reduction of anxiety, a sense of wellbeing, and perhaps even addiction. Though the effects of a typical meal are quantitatively less than those of doses of those drugs, most modern humans experience them several times a day, every day of their adult lives.

 

Hypothesis: exorphins and the origin of agriculture and civilisation

When this scenario of human dietary practices is viewed in the light of the problem of the origin of agriculture described earlier, it suggests an hypothesis that combines the results of these lines of enquiry.

Exorphin researchers, perhaps lacking a long-term historical perspective, have generally not investigated the possibility that these foods really are drug-like, and have instead searched without success for exorphin’s natural function. The adoption of cereal agriculture and the subsequent rise of civilisation have not been satisfactorily explained, because the behavioural changes underlying them have no obvious adaptive basis.

These unsolved and until-now unrelated problems may in fact solve each other. The answer, we suggest, is this: cereals and dairy foods are not natural human foods, but rather are preferred because they contain exorphins. This chemical reward was the incentive for the adoption of cereal agriculture in the Neolithic. Regular self-administration of these substances facilitated the behavioural changes that led to the subsequent appearance of civilisation.

This is the sequence of events that we envisage.

Climatic change at the end of the last glacial period led to an increase in the size and concentration of patches of wild cereals in certain areas (Wright 1977). The large quantities of cereals newly available provided an incentive to try to make a meal of them. People who succeeded in eating sizeable amounts of cereal seeds discovered the rewarding properties of the exorphins contained in them. Processing methods such as grinding and cooking were developed to make cereals more edible. The more palatable they could be made, the more they were consumed, and the more important the exorphin reward became for more people.

At first, patches of wild cereals were protected and harvested. Later, land was cleared and seeds were planted and tended, to increase quantity and reliability of supply. Exorphins attracted people to settle around cereal patches, abandoning their nomadic lifestyle, and allowed them to display tolerance instead of aggression as population densities rose in these new conditions.

Though it was, we suggest, the presence of exorphins that caused cereals (and not an alternative already prevalent in the diet) to be the major early cultigens, this does not mean that cereals are ‘just drugs’. They have been staples for thousands of years, and clearly have nutritional value. However, treating cereals as ‘just food’ leads to difficulties in explaining why anyone bothered to cultivate them. The fact that overall health declined when they were incorporated into the diet suggests that their rapid, almost total replacement of other foods was due more to chemical reward than to nutritional reasons.

It is noteworthy that the extent to which early groups became civilised correlates with the type of agriculture they practised. That is, major civilisations (in south-west Asia, Europe, India, and east and parts of South-East Asia; central and parts of north and south America; Egypt, Ethiopia and parts of tropical and west Africa) stemmed from groups which practised cereal, particularly wheat, agriculture (Bender 1975:12, Adams 1987:201, Thatcher 1987:212). (The rarer nomadic civilisations were based on dairy farming.)

Groups which practised vegeculture (of fruits, tubers etc.), or no agriculture (in tropical and south Africa, north and central Asia, Australia, New Guinea and the Pacific, and much of north and south America) did not become civilised to the same extent.

Thus major civilisations have in common that their populations were frequent ingesters of exorphins. We propose that large, hierarchical states were a natural consequence among such populations. Civilisation arose because reliable, on-demand availability of dietary opioids to individuals changed their behaviour, reducing aggression, and allowed them to become tolerant of sedentary life in crowded groups, to perform regular work, and to be more easily subjugated by rulers. Two socioeconomic classes emerged where before there had been only one (Johnson & Earle 1987:270), thus establishing a pattern which has been prevalent since that time.

 

Discussion

The natural diet and genetic change

Some nutritionists deny the notion of a pre-agricultural natural human diet on the basis that humans are omnivorous, or have adapted to agricultural foods (e.g. Garn & Leonard 1989; for the contrary view see for example Eaton & Konner 1985) . An omnivore, however, is simply an animal that eats both meat and plants: it can still be quite specialised in its preferences (chimpanzees are an appropriate example). A degree of omnivory in early humans might have preadapted them to some of the nutrients contained in cereals, but not to exorphins, which are unique to cereals.

The differential rates of lactase deficiency, coeliac disease and favism (the inability to metabolise fava beans) among modern racial groups are usually explained as the result of varying genetic adaptation to post-agricultural diets (Simopoulos 1990:27-9), and this could be thought of as implying some adaptation to exorphins as well. We argue that little or no such adaptation has occurred, for two reasons: first, allergy research indicates that these foods still cause abnormal reactions in many people, and that susceptibility is variable within as well as between populations, indicating that differential adaptation is not the only factor involved. Second, the function of the adaptations mentioned is to enable humans to digest those foods, and if they are adaptations, they arose because they conferred a survival advantage. But would susceptibility to the rewarding effects of exorphins lead to lower, or higher, reproductive success? One would expect in general that an animal with a supply of drugs would behave less adaptively and so lower its chances of survival. But our model shows how the widespread exorphin ingestion in humans has led to increased population. And once civilisation was the norm, non-susceptibility to exorphins would have meant not fitting in with society. Thus, though there may be adaptation to the nutritional content of cereals, there will be little or none to exorphins. In any case, while contemporary humans may enjoy the benefits of some adaptation to agricultural diets, those who actually made the change ten thousand years ago did not.

Other ‘non-nutritional’ origins of agriculture models

We are not the first to suggest a non-nutritional motive for early agriculture. Hayden (1990) argued that early cultigens and trade items had more prestige value than utility, and suggested that agriculture began because the powerful used its products for competitive feasting and accrual of wealth. Braidwood et al. (1953) and later Katz and Voigt (1986) suggested that the incentive for cereal cultivation was the production of alcoholic beer:

‘Under what conditions would the consumption of a wild plant resource be sufficiently important to lead to a change in behaviour (experiments with cultivation) in order to ensure an adequate supply of this resource? If wild cereals were in fact a minor part of the diet, any argument based on caloric need is weakened. It is our contention that the desire for alcohol would constitute a perceived psychological and social need that might easily prompt changes in subsistence behaviour’ (Katz & Voigt 1986:33).

This view is clearly compatible with ours. However there may be problems with an alcohol hypothesis: beer may have appeared after bread and other cereal products, and been consumed less widely or less frequently (Braidwood et al. 1953). Unlike alcohol, exorphins are present in all these products. This makes the case for chemical reward as the motive for agriculture much stronger. Opium poppies, too, were an early cultigen (Zohari 1986). Exorphin, alcohol, and opium are primarily rewarding (as opposed to the typically hallucinogenic drugs used by some hunter-gatherers) and it is the artificial reward which is necessary, we claim, for civilisation. Perhaps all three were instrumental in causing civilised behaviour to emerge.

Cereals have important qualities that differentiate them from most other drugs. They are a food source as well as a drug, and can be stored and transported easily. They are ingested in frequent small doses (not occasional large ones), and do not impede work performance in most people. A desire for the drug, even cravings or withdrawal, can be confused with hunger. These features make cereals the ideal facilitator of civilisation (and may also have contributed to the long delay in recognising their pharmacological properties).

Compatibility, limitations, more data needed

Our hypothesis is not a refutation of existing accounts of the origins of agriculture, but rather fits alongside them, explaining why cereal agriculture was adopted despite its apparent disadvantages and how it led to civilisation.

Gaps in our knowledge of exorphins limit the generality and strength of our claims. We do not know whether rice, millet and sorghum, nor grass species which were harvested by African and Australian hunter-gatherers, contain exorphins. We need to be sure that preagricultural staples do not contain exorphins in amounts similar to those in cereals. We do not know whether domestication has affected exorphin content or-potency. A test of our hypothesis by correlation of diet and degree of civilisation in different populations will require quantitative knowledge of the behavioural effects of all these foods.

We do not comment on the origin of noncereal agriculture, nor why some groups used a combination of foraging and farming, reverted from farming to foraging, or did not farm at all. Cereal agriculture and civilisation have, during the past ten thousand years, become virtually universal. The question, then, is not why they happened here and not there, but why they took longer to become established in some places than in others. At all times and places, chemical reward and the influence of civilisations already using cereals weighed in favour of adopting this lifestyle, the disadvantages of agriculture weighed against it, and factors such as climate, geography, soil quality, and availability of cultigens influenced the outcome. There is a recent trend to multi-causal models of the origins of agriculture (e.g. Redding 1988, Henry 1989), and exorphins can be thought of as simply another factor in the list. Analysis of the relative importance of all the factors involved, at all times and places, is beyond the scope of this paper.

 

Conclusion

‘An animal is a survival machine for the genes that built it. We too are animals, and we too are survival machines for our genes. That is the theory. In practice it makes a lot of sense when we look at wild animals…. It is very different when we look at ourselves. We appear to be a serious exception to the Darwinian law…. It obviously just isn’t true that most of us spend our time working energetically for the preservation of our genes’ (Dawkins 1989:138).

Many ethologists have acknowledged difficulties in explaining civilised human behaviour on evolutionary grounds, in some cases suggesting that modern humans do not always behave adaptively . Yet since agriculture began, the human population has risen by a factor of 1000: Irons (1990) notes that ‘population growth is not the expected effect of maladaptive behaviour’.

We have reviewed evidence from several areas of research which shows that cereals and dairy foods have drug-like properties, and shown how these properties may have been the incentive for the initial adoption of agriculture. We suggested further that constant exorphin intake facilitated the behavioural changes and subsequent population growth of civilisation, by increasing people’s tolerance of (a) living in crowded sedentary conditions, (b) devoting effort to the benefit of non-kin, and (c) playing a subservient role in a vast hierarchical social structure.

Cereals are still staples, and methods of artificial reward have diversified since that time, including today a wide range of pharmacological and non-pharmacological cultural artifacts whose function, ethologically speaking, is to provide reward without adaptive benefit. It seems reasonable then to suggest that civilisation not only arose out of self-administration of artificial reward, but is maintained in this way among contemporary humans. Hence a step towards resolution of the problem of explaining civilised human behaviour may be to incorporate into ethological models this widespread distortion of behaviour by artificial reward.

 

References

Adams, W .M., 1987, Cereals before cities except after Jacobs, in M. Melko & L.R. Scott eds, The boundaries of civilizations in space and time, University Press of America, Lanham.

Bell, I. R., 1987, Effects of food allergy on the central nervous system, in J. Brostoff and S. J. Challacombe, eds, Food allergy and intolerance, Bailliere Tindall, London.

Bender, B., 1975, Farming in prehistory: from hunter-gatherer to food producer, John Baker, London.

Bledsoe, W., 1987, Theories of the origins of civilization, in M. Melko and L. R. Scott, eds, The boundaries of civilizations in space and time, University Press of America, Lanham.

Blumler, M., & Byrne, R., 1991, The ecological genetics of domestication and the origins of agriculture, Current Anthropology 32: 2-35.

Braidwood, R. J., Sauer, J.D., Helbaek, H., Mangelsdorf, P.C., Cutler, H.C., Coon, C.S., Linton, R., Steward J. & Oppenheim, A.L., 1953, Symposium: did man once live by beer alone? American Anthropologist 55: 515-26.

Brantl, V., Teschemacher, H., Henschen, A. & Lottspeich, F., 1979, Novel opioid peptides derived from casein (beta-casomorphins), Hoppe-Seyler’s Zeitschrift fur Physiologische Chemie 360:1211-6.

Brostoff, J., & Gamlin, L., 1989, The complete guide to food allergy and intolerance, Bloomsbury, London.

Chang, T. T., 1989, Domestication and the spread of the cultivated rices, in D.R. Harris and G.C. Hillman, eds, Foraging and farming: the evolution of plant exploitation, Unwin Hyman, London.

Claessen, H. J. M. & Skalnik P., eds, 1978, The early state, Mouton, The Hague.

Cohen, M. N., 1977, Population pressure and the origins of agriculture: an archaeological example from the coast of Peru, in Reed, C.A., ed., The origins of agriculture, Mouton, The Hague.

Cohen, M. N., 1989, Health and the rise of civilization, Yale University Press, New Haven.

Constantini, L., 1989, Plant exploitation at Grotta dell’Uzzo, Sicily: new evidence for the transition from Mesolithic to Neolithic subsistence in southern Europe, in Harris, D. R. & Hillman, G. C., eds, Foraging and farming: the evolution of plant exploitation, Unwin Hyman, London.

Dawkins, R., 1989, Darwinism and human purpose, in Durant, J. R., ed., Human origins, Clarendon Press, Oxford.

Dohan, F., 1966, Cereals and schizophrenia: data and hypothesis, Acta Psychiatrica Scandinavica 42:125-52.

Dohan, F., 1983, More on coeliac disease as a model of schizophrenia, Biological Psychiatry 18:561-4.

Dohan, F. & Grasberger, J., 1973, Relapsed schizophrenics: earlier discharge from the hospital after cereal-free, milk-free diet, American Journal of Psychiatry 130:685-8.

Dohan, F., Harper, E., Clark, M., Ratigue, R., & Zigos, V., 1984, Is schizophrenia rare if grain is rare? Biological Psychiatry 19: 385-99.

Eaton, S. B. & Konner, M., 1985, Paleolithic nutrition – a consideration of its nature and current implications, New England Journal of Medicine 312: 283-90.

Egger, J., 1988, Food allergy and the central nervous system, in Reinhardt, D. & Schmidt E., eds, Food allergy, Raven, New York.

Flannery, K. V., 1973, The origins of agriculture, Annual Review of Anthropology 2:271-310.

Fukudome, S., & Yoshikawa, M., 1992, Opioid peptides derived from wheat gluten: their isolation and characterization, FEBS Letters 296:107-11.

Gardner, M. L. G., 1985, Production of pharmacologically active peptides from foods in the gut. in Hunter, J. & Alun-Jones, V., eds, Food and the gut,Bailliere Tindall, London.

Gam, S. M. & Leonard, W. R., 1989, What did our ancestors eat? Nutritional Reviews 47:337 45.

Gordon, K. D., 1987, Evolutionary perspectives on human diet, in Johnston, F., ed, Nutritional Anthropology, Alan R. Liss, New York.

Greksch, G., Schweiger C., Matthies, H., 1981, Evidence for analgesic activity of beta-casomorphin in rats, Neuroscience Letters 27:325~8.

Harlan, J. R., 1986, Plant domestication: diffuse origins and diffusion, in Barigozzi, G., ed., The origin and domestication of cultivated plants, Elsevier, Amsterdam.

Harris, D. R., 1977, Alternative pathways towards agriculture, in Reed, C. A., ed., The origins of agriculture, Mouton, The Hague.

Harris, D. R. & Hillman, G. C., eds, 1989, Foraging and farming: the evolution of plant exploitation, Unwin Hyman, London.

Hayden, B., 1990, Nimrods, piscators, pluckers, and planters: the emergence of food production, Journal of Anthropological Archaeology 9:31-69.

Henry, D. O., 1989, From foraging to agriculture: the Levant at the end of the ice age, University of Pennsylvania Press, Philadelphia.

Heubner, F., Liebeman, K., Rubino, R. & Wall, J., 1984, Demonstration of high opioid-like activity in isolated peptides from wheat gluten hydrolysates, Peptides 5:1139-47.

Irons, W., 1990, Let’s make our perspective broader rather than narrower, Ethology and Sociobiology 11: 361-74

Johnson, A. W. & Earle, T., 1987, The evolution of human societies: from foraging group to agrarian state, Stanford University Press, Stanford.

Katz, S. H. & Voigt, M. M., 1986, Bread and beer: the early use of cereals in the human diet, Expedition 28:23-34.

Kay, R. F., 1985, Dental evidence for the diet of Australopithecus, Annual Review of Anthropology 14:315 41.

Kroker, G. F., 1987, Chronic candiosis and allergy, in Brostoff, J. & Challacombe, S.J., eds, Food allergy and intolerance, Bailliere Tindall, London.

Lee, R. B. & DeVore, I., 1968, Problems in the study of hunters and gatherers, in Lee, R.B. & DeVore, I., eds, Man the hunter, Aldine, Chicago.

Mycroft, F. J., Wei, E. T., Bernardin, J. E. & Kasarda, D. D., 1982, MlF-like sequences in milk and wheat proteins, New England Journal of Medicine301:895.

Mycroft, F. J., Bhargava, H. N. & Wei, E. T., 1987, Pharmacalogical activities of the MIF-1 analogues Pro-Leu-Gly, Tyr-Pro-Leu-Gly and pareptide,Peptides 8:1051-5.

Panksepp, J., Normansell, L., Siviy, S., Rossi, J. & Zolovick, A., 1984, Casomorphins reduce separation distress in chicks, Peptides 5:829-83.

Paroli, E., 1988, Opioid peptides from food (the exorphins), World review of nutrition and dietetics 55:58-97.

Pedersen, B., Knudsen, K. E. B. & Eggum, B. 0., 1989, Nutritive value of cereal products with emphasis on the effect of milling, World review of nutrition and dietetics 60:1-91.

Peters, C. R. & O’Brien, E. M., 1981, The early hominid plant-food niche: insights from an analysis of plant exploitation by Homo, Pan, and Papio in eastern and southern Africa, Current Anthropology 22:127-40.

Peuch, P., Albertini, H. & Serratrice, C., 1983, Tooth microwear and dietary patterns in early hominids from Laetoli, Hadar, and Olduvai, Journal ofHuman Evolution 12:721-9.

Pfeiffer, J. E., 1977, The emergence of society: a prehistory of the establishment, McGraw Hill, New York.

Pryor, F. L., 1986, The adoption of agriculture: some theoretical and empirical evidence, American Anthropologist 88:879-97.

Radcliffe, M. J., 1987, Diagnostic use of dietary regimes, in Brostoff, J. & Challacombe, S. J., eds, Food allergy and intolerance, Bailliere Tindall, London.

Ramabadran, K. & Bansinath, M., 1988, Opioid peptides from milk as a possible cause of Sudden Infant Death Syndrome, Medical Hypotheses 27:181-7.

Randolph, T. G., 1978, Specific adaptation, in Annals of Allergy 40:333-45

Redding, R., 1988, A general explanation of subsistence change from hunting and gathering to food production, Journal of Anthropological Archaeology7:56-97.

Reed, C. A., ed., 1977, The origins of agriculture, Mouton, The Hague.

Rindos, D., 1984, The origins of agriculture: an evolutionary perspective, Academic Press, Orlando.

Scadding, G. K. & Brostoff, J., 1988, The dietic treatment of food allergy, in Reinhardt, D. & Schmidt, E., eds, Food allergy, Raven, New York.

Simopoulos, A. P., 1990, Genetics and nutrition: or what your genes can tell you about nutrition, World review of nutrition and dietetics 63:25-34.

Sprague, D. E. & Milam, M. J., 1987, Concept of an environmental unit, in Brostoff, J. & .Challacombe, S. J., eds, Food allergy and intolerance, Bailliere Tindall, London.

Stark, B. L., 1986, Origins of food production in the New World, in Meltzer, D. J., Fowler, D. D. & Sabloff, J. A., eds, American archaeology past and future, Smithsonian Institute Press, Washington.

Susman, R. L., 1987, Pygmy chimpanzees and common chimpanzees: models for the behavioural ecology of the earliest hominids, in Kinzey, W. G., ed., The evolution of human behaviour: primate models, State University of New York Press, Albany.

Svedburg, J., De Haas, J., Leimenstoll, G., Paul, F. & Teschemacher, H., 1985, Demonstration of betacasomorphin immunoreactive materials in in-vitro digests of bovine milk and in small intestine contents after bovine milk ingestion in adult humans, Peptides 6:825-30.

Thatcher, J. P., 1987, The economic base for civilization in the New World, in Melko, M. & Scott, L. R., eds, The boundaries of civilizations in space andtime, University Press of America, Lanham.

Walker, A., 1981, Dietary hypotheses and human evolution, Philosophical Transactions of the Royal Society of London B292:57-64.

Washburn, L. & Lancaster, C. S., 1968, The evolution of hunting, in Lee, R. B. & DeVore, I., eds, Man the hunter, Aldine, Chicago.

Wittfogel, K., 1957, Oriental Despotism, Yale University Press, New Haven.

Wraith, D. G., 1987, Asthma, in Brostoff, J. & Challacombe, S. J., eds, Food allergy and intolerance, Bailliere Tindall, London.

Wright, H. E., 1977, Environmental changes and the origin of agriculture in the Near East, in Reed, C. A., ed, The origins of agriculture, Mouton, The Hague.

Zioudrou, C., Streaty, R. & Klee, W., 1979, Opioid peptides derived from food proteins: the exorphins Journal of Biological Chemistry 254:244S9.

Zohari, D., 1986, The origin and early spread of agriculture in the Old World, in Barigozzi, G., ed., The origin and domestication of cultivated plants,Elsevier, Amsterdam

 

via www.dis.unimelb.edu.au/staff/gwadley/ethology/ab-paper.html

US-Israel Alliance: Obama’s “Green Light” to Israel Attack on Gaza

US-Israel Alliance: Obama’s “Green Light” to Israel Attack on Gaza

A short interview broadcast by CNN late last week featuring two participants – a Palestinian in Gaza and an Israeli within range of the rocket attacks – did not follow the usual script.

For once, a media outlet dropped its role as gatekeeper, there to mediate and therefore impair our understanding of what is taking place between Israel and the Palestinians, and inadvertently became a simple window on real events.

The usual aim of such “balance” interviews relating to the Israeli-Palestinian conflict is twofold: to reassure the audience that both sides of the story are being presented fairly; and to dissipate potential outrage at the deaths of Palestinian civilians by giving equal time to the suffering of Israelis.

But the deeper function of such coverage in relation to Gaza, given the media’s assumption that Israeli bombs are simply a reaction to Hamas terror, is to redirect the audience’s anger exclusively towards Hamas. In this way, Hamas is made implicitly responsible for the suffering of both Israelis and Palestinians.

The dramatic conclusion to CNN’s interview appears, however, to have otherwise trumped normal journalistic considerations.

The pre-recorded interview via Skype opened with Mohammed Sulaiman in Gaza. From what looked like a cramped room, presumably serving as a bomb shelter, he spoke of how he was too afraid to step outside his home. Throughout the interview, we could hear the muffled sound of bombs exploding in the near-distance. Mohammed occasionally glanced nervously to his side.

The other interviewee, Nissim Nahoom, an Israeli official in Ashkelon, also spoke of his family’s terror, arguing that it was no different from that of Gazans. Except in one respect, he hastened to add: things were worse for Israelis because they had to live with the knowledge that Hamas rockets were intended to harm civilians, unlike the precision missiles and bombs Israel dropped on Gaza.

The interview returned to Mohammed. As he started to speak, the bombing grew much louder. He pressed on, saying he would not be silenced by what was taking place outside. The interviewer, Isha Sesay, interrupted – seemingly unsure of what she was hearing – to inquire about the noise.

Then, with an irony that Mohammed could not have appreciated as he spoke, he began to say he refused to be drawn into a comparison about whose suffering was worse when an enormous explosion threw him from his chair and severed the internet connection. Switching back to the studio, Sesay reassured viewers that Mohammed had not been hurt.

The bombs, however, spoke more eloquently than either Mohammed or Nissim.

If Mohammed had had more time, he might have been able to challenge Nissim’s point about Israelis’ greater fears as well as pointing to another important difference between his and his Israeli interlocutor’s respective plights.

The far greater accuracy of Israel’s weaponry in no way confers peace of mind. The fact is that a Palestinian civilian in Gaza is in far more danger of being killed or injured by one of Israel’s precision armaments than an Israeli is by one of the more primitive rockets being launched out of Gaza.

In Operation Cast Lead, Israel’s attack on Gaza in winter 2008-09, three Israelis were killed by rocket attacks, and six soldiers died in fighting. In Gaza, meanwhile, nearly 1,400 Palestinians were killed, of whom at least 1,000 were not involved in hostilities, according to the Israeli group B’Tselem. Many, if not most, of those civilians were killed by so-called precision bombs and missiles.

If Israelis like Nissim really believe they have to endure greater suffering because the Palestinians lack accurate weapons, then maybe they should start lobbying Washington to distribute its military hardware more equitably, so that the Palestinians can receive the same allocations of military aid and armaments as Israel.

Or alternatively, they could lobby their own government to allow Iran and Hizbullah to bring into Gaza more sophisticated technology than can currently be smuggled in via the tunnels.

The other difference is that, unlike Nissim and his family, most people in Gaza have nowhere else to flee. And the reason that they must live under the rain of bombs in one of the most densely populated areas on earth is because Israel – and to a lesser extent Egypt – has sealed the borders to create a prison for them.

Israel has denied Gaza a port, control of its airspace and the right of its inhabitants to move to the other Palestinian territory recognised by the Oslo accords, the West Bank. It is not, as Israel’s supporters allege, that Hamas is hiding among Palestinian civilians; rather, Israel has forced Palestinian civilians to live in a tiny strip of land that Israel turned into a war zone.

So who is chiefly to blame for the escalation that currently threatens the nearly two million inhabitants of Gaza? Though Hamas’ hands are not entirely clean, there are culprits far more responsible than the Palestinian militants.

First culprit: The state of Israel

The inciting cause of the latest confrontation between Israel and Hamas has little to do with the firing of rockets, whether by Hamas or the other Palestinian factions.

The conflict predates the rockets – and even the creation of Hamas – by decades. It is the legacy of Israel’s dispossession of Palestinians in 1948, forcing many of them from their homes in what is now Israel into the tiny Gaza Strip. That original injustice has been compounded by the occupation Israel has not only failed to end but has actually intensified in recent years with its relentless siege of the small strip of territory.

Israel has been progressively choking the life out of Gaza, destroying its economy, periodically wrecking its infrastructure, denying its inhabitants freedom of movement and leaving its population immiserated.

One only needs to look at the restrictions on Gazans’ access to their own sea. Here we are not considering their right to use their own coast to leave and enter their territory, simply their right to use their own waters to feed themselves. According to one provision of the Oslo accords, Gaza was given fishing rights up to 20 miles off its shore. Israel has slowly whittled that down to just three miles, with Israeli navy vessels firing on fishing boats even inside that paltry limit.

Palestinians in Gaza are entitled to struggle for their right to live and prosper. That struggle is a form of self-defence – not aggression – against occupation, oppression, colonialism and imperialism.

Second culprit: Binyamin Netanyahu and Ehud Barak

The Israeli prime minister and defence minister have taken a direct and personal hand, above and beyond Israel’s wider role in enforcing the occupation, in escalating the violence.

Israel and its supporters always make it their first priority when Israel launches a new war of aggression to obscure the timeline of events as a way to cloud responsibility. The media willingly regurgitates such efforts at misdirection.

In reality, Israel engineered a confrontation to provide the pretext for a “retaliatory” attack, just as it did four years earlier in Operation Cast Lead. Then Israel broke a six-month ceasefire agreed with Hamas by staging a raid into Gaza that killed six Hamas members.

This time, on 8 November, Israel achieved the same end by invading Gaza again, on this occasion following a two-week lull in tensions. A 13-year-old boy out playing football was killed by an Israeli bullet.

Tit-for-tat violence over the following days resulted in the injury of eight Israelis, including four soldiers, and the deaths of five Palestinian civilians, and the wounding of dozens more in Gaza.

On November 12, as part of efforts to calm things down, the Palestinian militant factions agreed a truce that held two days – until Israel broke it by assassinating Hamas military leader Ahmed Jabari. The rockets out of Gaza that followed these various Israeli provocations have been misrepresented as the casus belli.

But if Netanyahu and Barak are responsible for creating the immediate pretext for an attack on Gaza, they are also criminally negligent for failing to pursue an opportunity to secure a much longer truce with Hamas.

We now know, thanks to Israeli peace activist Gershon Baskin, that in the period leading up to Jabari’s execution Egypt had been working to secure a long-term truce between Israel and Hamas. Jabari was apparently eager to agree to it.

Baskin, who was intimately involved in the talks, was a credible conduit between Israel and Hamas because he had played a key role last year in getting Jabari to sign off on a prisoner exchange that led to the release of Israeli soldier Gilad Shalit. Baskin noted in the Haaretz newspaper that Jabari’s assassination “killed the possibility of achieving a truce and also the Egyptian mediators’ ability to function.”

The peace activist had already met Barak to alert him to the truce, but it seems the defence minister and Netanyahu had more pressing concerns than ending the tensions between Israel and Hamas.

What could have been more important than finding a mechanism for saving lives, on both the Palestinian and Israeli sides. Baskin offers a clue: “Those who made the decision must be judged by the voters, but to my regret they will get more votes because of this.”

It seems Israel’s general election, due in January, was uppermost in the minds of Netanyahu and Barak.

A lesson learnt by Israeli leaders over recent years, as Baskin notes, is that wars are vote-winners solely for the right wing. That should be clear to no one more than Netanyahu. He has twice before become prime minister on the back of wars waged by his more “moderate” political opponents as they faced elections.

Shimon Peres, a dove by no standard except a peculiar Israeli one, launched an attack on Lebanon, Operation Grapes of Wrath, that cost him the election in 1996. And centrists Ehud Olmert and Tzipi Livni again helped Netanyahu to victory by attacking Gaza in late 2008.

Israelis, it seems, prefer a leader who does not bother to wrap a velvet glove around his iron fist.

Netanyahu was already forging ahead in the polls before he minted Operation Pillar of Defence. But the electoral fortunes of Ehud Barak, sometimes described as Netanyahu’s political Siamese twin and a military mentor to Netanyahu from their commando days together, have been looking grim indeed.

Barak desperately needed a military rather than a political campaign to boost his standing and get his renegade Independence party across the electoral threshold and into the Israeli parliament. It seems Netanyahu, thinking he had little to lose himself from an operation in Gaza, may have been willing to oblige.

Third culprit: The Israeli army

Israel’s army has become addicted to two doctrines it calls the “deterrence principle” and its “qualitative military edge”. Both are fancy ways of saying that, like some mafia heavy, the Israeli army wants to be sure it alone can “whack” its enemies. Deterrence, in Israeli parlance, does not refer to a balance of fear but Israel’s exclusive right to use terror.

The amassing of rockets by Hamas, therefore, violates the Israeli army’s own sense of propriety, just as Hizbullah’s stockpiling does further north. Israel wants its neighbouring enemies to have no ability to resist its dictates.

Doubtless the army was only too ready to back Netanyahu and Barak’s electioneering if it also provided an opportunity to clean out some of Hamas’ rocket arsenal.

But there is another strategic reason why the Israeli army has been chomping at the bit to crack down on Hamas again.

Haaretz’s two chief military correspondents explained the logic of the army’s position last week, shortly after Israel killed Jabari. They reported: “For a long time now Israel has been pursuing a policy of containment in the Gaza Strip, limiting its response to the prolonged effort on the part of Hamas to dictate new rules of the game surrounding the fence, mainly in its attempt to prevent the entry of the IDF into the ‘perimeter,’ the strip of a few hundred meters wide to the west of the fence.”

In short, Hamas has angered Israeli commanders by refusing to sit quietly while the army treats large areas of Gaza as its playground and enters at will.

Israel has created what it terms a “buffer zone” inside the fence around Gaza, often up to a kilometre wide, that Palestinians cannot enter but the Israeli army can use as a gateway for launching its “incursions”. Remote-controlled guns mounted on Israeli watch-towers around Gaza can open fire on any Palestinian who is considered to have approached too close.

Three incidents shortly before Jabari’s extra-judicial execution illustrate the struggle for control over Gaza’s interior.

On November 4, the Israeli army shot dead a young Palestinian man inside Gaza as he was reported to have approached the fence. Palestinians say he was mentally unfit and that he could have been saved by medics had ambulances not been prevented from reaching him for several hours.

On November 8, as already noted, the Israeli army made an incursion into Gaza to attack Palestinian militants and in the process shot dead a boy playing football.

And on November 10, two days later, Palestinian fighters fired an anti-tank missile that destroyed a Jeep patrolling the perimeter fence around Gaza, wounding four soldiers.

As the Haaretz reporters note, Hamas appears to be trying to demonstrate that it has as much right to defend its side of the “border fence” as Israel does on the other side.

The army’s response to this display of native impertinence has been to inflict a savage form of collective punishment on Gaza to remind Hamas who is boss.

Fourth culprit: the White House

It is near-impossible to believe that Netanyahu decided to revive Israel’s policy of extra-judicial executions of Hamas leaders – and bystanders – without at least consulting the White House. Israel clearly also held off from beginning its escalation until after the US elections, restricting itself, as it did in Cast Lead, to the “downtime” in US politics between the elections and the presidential inauguration.

That was designed to avoid overly embarrassing the US president. A fair assumption must be that Barack Obama approved Israel’s operation in advance. Certainly he has provided unstinting backing since, despite the wildly optimistic scenarios painted by some analysts that he was likely to seek revenge on Netanyahu in his second term.

Also, it should be remembered that Israel’s belligerence towards Gaza, and the easing of domestic pressure on Israel to negotiate with Hamas or reach a ceasefire, has largely been made possible because Obama forced US taxpayers to massively subsidise Israel’s rocket interception system, Iron Dome, to the tune of hundreds of millions of dollars.

Iron Dome is being used to shoot down rockets out of Gaza that might otherwise have landed in built-up areas of Israel. Israel and the White House have therefore been able to sell US munificence on the interception of rockets as a humanitarian gesture.

But the reality is that Iron Dome has swung Israel’s cost-benefit calculus sharply in favour of greater aggression because it is has increased Israel’s sense of impunity. Whatever Hamas’ ability to smuggle into Gaza more sophisticated weaponry, Israel believes it can neutralise that threat using interception systems.

Far from being a humanitarian measure, Iron Dome has simply served to ensure that Gaza will continue to suffer a far larger burden of deaths and injuries in confrontations with Israel and that such confrontations will continue to occur regularly.

Here are the four main culprits. They should be held responsible for the deaths of Palestinians and Israelis in the days and, if Israel expands its operation, weeks ahead.

Jonathan Cook won the Martha Gellhorn Special Prize for Journalism. His latest books are “Israel and the Clash of Civilisations: Iraq, Iran and the Plan to Remake the Middle East” (Pluto Press) and “Disappearing Palestine: Israel’s Experiments in Human Despair” (Zed Books).  His new website is www.jonathan-cook.net.

MAP: CIA Interventions Since World War II

MAP: CIA Interventions Since World War II

invasions … bombings … overthrowing governments …
occupations … suppressing movements for social change …
assassinating political leaders … perverting elections …
manipulating labor unions … manufacturing “news” …
economic and political sanctions … death squads …
torture … biological warfare … depleted uranium …
drug trafficking … mercenaries …