
A little over a year and a half ago, my wife and I were in Cambodia, sitting cross-legged on the wooden bow of a weathered prop boat, chatting with a local guide as he took us through a floating village populated – he told us – by ethnic Vietnamese immigrants. The boat cut through a wide channel in a patch of inundated forest on the Tonle Sap, our wake fanning out lazily to disappear among the trees or slosh mildly against the wooden bases of floating houses.
I don’t recall the precise details, but somehow our conversation turned to taxonomy. Our guide wanted to know what, according to our view, counted as an animal. Was an ant an animal? What about a human? This initially struck me as something of an odd question, but through the course of our back-and-forth it became apparent that it was quite sensible when understood from his perspective. Though the Western system of taxonomic classification exudes a comfortable aura of familiarity to those who have been raised with it, it is far from the most obvious way to order the natural world. No doubt this is part of the reason it didn’t occur to anyone prior to the past three centuries of human history.
In the view of our guide, the relationships among living things are defined by principles of opposition and symmetry. There must be two of each kind in a category and the relationships among categories are defined by an array of similarities and differences. That dogs and monkeys and snakes count as animals seemed to him common sense. That ants and humans are also members of the same category struck him as a little more peculiar.
When it comes to ordering the world of birds and beasts, the primary difference between our guide and ourselves was that we happened to subscribe to a classification system that orders the living world according to relationships of descent with modification. We apply a Linaean classification system structured around the notion that all living things share a common origin, and that their relationships are defined by how recently they diverged from a shared ancestor in a nested hierarchy extending back to a successful batch of single-celled organisms that emerged from the primordial soup some 3.5 billion years ago. Though many of the features that seemed most salient in his classification system are irrelevant to our Darwinian framework, they nonetheless have their own internal logic.
Unfortunately, our conversation was far too short for anyone involved to build a comprehensive picture of the other’s worldview. Nevertheless, a few general points are obvious. Foremost, that our guide’s view differed so substantively from our own is surely not the product of any innate difference in cognitive capacity. Nor can it be attributed to any inherent difference between his most recent ancestors and our own.
Instead, his perspective is primarily a product of vast networks of highly contingent influences. A non-exhaustive list might include the distinct and ever changing traditional beliefs of whatever ethnic group (or groups) he might belong to, his parent’s interpretation thereof, the ideas of his friends, the recent political history of Cambodia, and so forth. It’s highly unlikely that his ideas about the world are identical to those of his friends and neighbors, but are nonetheless shaped by sampling a more widely shared cultural repertoire. The same can be said of my wife and I – the fact that we have learned about biological evolution is a consequence of the circumstances in which we were born and raised, and the distinct cultural trajectory pursued by a certain subset of Western Europeans following the invention of the printing press and the subsequent intellectual revolution of the Enlightenment.
A Brief History of Cultural Relativism
The view that human differences in knowledge, belief, or practice are usefully (if only partially) explained by the proximate influence of culture, and that these difference can be usefully illuminated through an understanding of the internal logic of the cultures that produce them, is the perspective offered by cultural relativism.
Unfortunately, cultural relativism is a widely misunderstood principle. The concept (though not the term itself) probably first emerged in the writings of the anthropologist Franz Boas in the closing decades of the 19th century. Since then, it has been a useful tool for ethnographers seeking to understand culturally motivated behavior. At the same time, it has also been a plague upon the larger enterprise of producing scientifically justifiable explanations for human behavior. Upon entering the popular vernacular, it has offered of a veneer of intellectual rigor and scholarly nuance to facile arguments about the plurality of knowledge and the primacy of subjective experience. It has granted succor to the idea that there are “other ways of knowing” as reliable and equally deserving of confidence as the scientific method. It has sired pleas for “trigger warnings”, outrage over “microaggressions”, and an exaggerated emphasis on the subjective experience of feeling offended as a suitable justification for curtailing speech.
In its most extreme incarnations, cultural relativism has been advanced as a prohibition on any and all cross-cultural evaluation of beliefs, practices, and knowledge-claims. Everything from traditional accounts of cosmic creation and religious explanations of natural phenomena to moral arguments about the nature of good and bad behavior is taken off the table, shielded from scrutiny by the notion that cultural differences are sufficiently deep to render the humans they divide mutually unintelligible. Superficially, this stance has often been made to look like a sophisticated embrace of uncertainty, but it is really nothing of the sort. Instead, it is a pedantic, dehumanizing, and pusillanimous retreat from the hard work of uncovering truth and defending modern ideas about universal human rights.
Cultural relativism can be decomposed into three components. The first is methodological cultural relativism. This is a tool used by anthropologists and ethnographers to understand the beliefs, customs, ideas, and behavior of people with cultures and histories different from their own. In many respects mundane, methodological cultural relativism has proven extremely useful, allowing researchers to strip away the often blinding baggage of personal history and cultural bias.
Sadly, methodological relativism has come to be associated with the parasitic vices of normative (or moral) and cognitive relativism. Too often, these have served to rob mundane cultural relativism of its methodological utility. Cognitive realism is what emerges when one looks upon the real problems inherent in the human quest for knowledge – the things that make rock-solid, axiomatic certainty so difficult to achieve – and withdraws from the search entirely. Similarly, normative relativism represents a withdrawal from the challenges that emerge from the fluid, turbulent ambiguity inherent in developing and applying coherent ethical systems.
Taken together, cognitive and normative relativism are the seeds for a system of intellectual and ethical obfuscation that will hence-forth be referred to as fundamentalist cultural relativism (FCR). Rather than a useful system of intellectual criticism and methodological prescriptions, FCR is the intellectual equivalent of a mountaineer collapsing at the base of gnarly, ferocious peak and declaring it impossible to summit before even giving it a serious attempt. Put more simply, it’s what happens when someone looks at a problem, notes that it’s hard, and immediately declares it intractable.
As a consequence, many serious academics have abandoned cultural relativism entirely. The physicist David Deutsch has cast it as a brand of “persistently popular irrationality.” Insofar as he is referring to the idea’s most common flavor, so thoroughly infused with cognitive and normative relativism and particularly pervasive among the social justice crowd, he isn’t far off the mark. Indeed, the extreme relativism that flourished in the humanities and social sciences in the 1980s and 1990s is nothing if not a persistently popular brand of irrationality. With the spread of the vapid school of vigorous intellectual masturbation and enthusiastic pedantry politely referred to as “postmodernism,” the notion that all knowledge-claims or moral positions are equally valid became dogma throughout many of the subdisciplines that fall within the scope of the humanities and social sciences. My own discipline, anthropology, is still recovering from the effects of this intellectual rot.
One of the chief problems with that the brand of cultural relativism (justifiably) decried by thinkers like Deutsch is that it represents an abuse of a more useful principle. The notion that the beliefs, customs, and behaviors of different cultural groups could be usefully analyzed and understood in their own terms was a radical advance over previous anthropological methods. When the earliest forms of the concept were first advanced, beliefs about the social – and even biological – supremacy of white Western men were scarcely given a second thought. White men in Western, industrialized societies occupied the pinnacle of a long chain of social and biological evolution.
According to these views, Western culture represented the purest, most complete manifestation of a natural order, the metric against which all other human cultures should be judged. The cultural differences uncovered in places like the equatorial rain forests of Papua New Guinea or the frigid expanses of the high Arctic were commonly viewed as deviations from perfection. Today, these notions are rightly considered signally ridiculous and repugnant by all but the most intransigent racists and xenophobes. But to many prominent and serious thinkers occupying important positions of academic and political authority in the late 19th and early 20th century, nothing could have been more obvious than their own innate superiority.
Eventually, ethnographers and anthropologists like Franz Boas came to recognize that the rampant ethnocentrism of their peers was translating into bad science, fundamentally hobbling efforts to actually understand the roots of human behavioral variation. The idea that “traditional” societies, such as those found practicing foraging or horticulturalist lifestyles in the rainforests of Brazil, were somehow frozen in time – primitive vestiges of points along the slow but inevitable march toward the industrialized modernity of centralized state authority and market economics – represented a roadblock to intellectual progress. Recognizing the faults in this perspective, some turn of the century anthropologists began to advocate analyzing cultures according to their own internal logic.
It’s worth noting that the kind of relativism espoused by Boas and his students was almost certainly more permissive that the version being advocated here. Nonetheless, as an analytical tool, cultural relativism represented a considerable improvement over the parochial, Western-centric worldview that had been distorting the thinking of early explorers and ethnographers. Rather than viewing non-Western cultures as primitive holdovers, serious anthropologists began to recognize that individual cultures – and the differences among them – could be fruitfully illuminated by seriously attempting to understand them in their own terms. Substantial utility could be found in recognizing that each culture is a product of a vast web of historical contingency.
Cultural Relativity as a Modern Research Tool
To understand why this is so, it’s useful to couch the argument in more modern terms. Human behavior, like that of other animals, is a product of a complex stew of influences. This includes a suite of genes – some highly conserved, others more recently selected as a result of the specific challenges that came along with being a bipedal, largely hairless, highly social primate living on the African savannah for the better share of the Pleistocene. It also includes a capacity to learn and usefully modify our behavior over the span of hours, days, weeks, months, or years – allowing us to respond to environmental variability much more rapidly than the relatively glacial pace at which natural selection sorts the genetic wheat from the chaff. But the inheritance of genetic information is ubiquitous across the living world, and learning is not particularly rare. Humanity’s peculiar spark is found in our unique capacity to create, accumulate, store, and transmit vast sums of non-genetic information – i.e. culture.
Culture itself is an evolving system, emerging as generations of individuals interact with one another while attempting to cope with the challenges of living in particular environments and trying to build stable social arrangements therein. Ideas about how best to live are dictated not only by hard practicality (e.g. what foods offer the best caloric returns relative to caloric energy spent in acquiring them), but the beliefs, customs, superstitions, and additional cultural bric-a-brac that tends to accumulate as a byproduct of human efforts to live in a world they don’t fully comprehend. The ability to accumulate, compress, retain, and manipulate all of that extra information symbolically, and transmit it across generations, is the primary feature that has allowed humans to not only survive, but actively flourish, in environments as diverse as the frozen expanses of Canadian Arctic to the tropical rain forests of Papua New Guinea. The capacity for culture transformed an African primate into a global species. Without understanding culture – and the processes that shape it – it is fundamentally impossible to produce a complete account of human behavioral variation.
Cultural relativism, then, is a useful lens for investigating one of the principle variables shaping human behavior. If you want to understand why people think and act the way they do, it is often useful to try your best to see things from their angle – to adopt what anthropologists call an “emic”, or “inside-looking-out” view of culture. This is not controversial. It is simply a specific incarnation of what philosopher James Woodward has called the manipulability conception of causal explanation, a counterfactual account of causal relationships that suggests one can find out why things are the way they are by imagining how they might be otherwise. People are the way they are, in part, because of the culture they experience. If you changed or removed that culture, they would be different. Thus, to understand why they are the way they are, you should understand the set of beliefs, customs, practices, ideas, rituals and what-not that are caught under the umbrella term, “culture”.
This is also an idea that has been hijacked and corrupted into something that, in its own ways, has come to prove almost as debilitating as 19th century Western ethnocentrism. Extreme relativists advocate the position that cultures can only be understood in their own terms, that outsiders have no justifiable basis for cross-cultural evaluation, that all culturally derived knowledge is equally true, and that all culturally derived moral positions are equally valid. Proponents of this school of cultural relativism argue not that the differences between my perspective and that of my guide are usefully illuminated by understanding the details of the differing cultural contexts from which they emerged, but that they are also equally true pictures of the way the world works.
Fundamentalist Cultural Relativism and Science
By completely abandoning the prospect of “etic” (or “outside-looking-in”) analysis, the proponents of fundamentalist relativism place themselves in a thorny situation. Extreme relativism implies a fairly strict allegiance to idealist epistemologies. This means that they subscribe to the notion that subjective experience is the absolute arbiter of reality – or, roughly, that our minds make the world, rather than our minds being a product of a world with an existence independent of and discoverable by human observers.
Under this view, the special utility of science as a knowledge gaining activity is implicitly (and often explicitly) denied. This represents a shift from the rather mundane observation that the process of scientific discovery is a cultural phenomenon to the much more drastic (and far less tenable) position that scientific knowledge – the product of that process – is culturally contingent as well. Force is only equivalent to the mass of an object multiplied by its acceleration in cultures with the concepts supplied by Newtonian physics, a molecule of water is only comprised of two hydrogen atoms covalently bonded to an oxygen atom is societies with chemists, and the electrical resistance of a conductive metal only decreases with lower temperatures in societies that have developed an understanding of electromagnetism.
Such a line of thinking may seem ridiculous – a hyperbolic straw-man set to topple in the slightest breeze – but is in fact an accurate representation of positions forwarded by serious and influential intellectuals in the postmodern movement. This has resulted in a disconcerting number of otherwise intelligent people (typically either heavily informed by or professionally engaged in the humanities and social sciences) taking fundamentalist cultural relativism seriously. Instead of placing special emphasis on the process of scientific discovery, progressive intellectuals too-often celebrate “other ways of knowing”, as if astrology or eastern spiritualism can produce a knowledge-claim half as worthy of confidence as rigorous empirical investigation and peer review.
To a degree, this breed of extreme relativism may seem harmless enough. After all, as astrophysicist and science-popularizer Neil deGrasse Tyson is fond of pointing out, the findings of science are true regardless of whether or not anyone buys into them. Most of the people who fall into the trap of fundamentalist CR aren’t likely to be engaged in the process of scientific discovery – they can watch from the sidelines and pooh-pooh the veracity of scientific claims without impinging on their actual veracity in any way, just as I could sit on the sidelines of a basketball game and deny that a player has successfully made a free-throw without affecting whether or not a new point has actually been scored.
This would be true if a tolerance for nonsense didn’t inevitably yield ugly results. Someone making important life choices based on the sign of the zodiac is foolish and risky, but generally innocuous. But what of an impotent Chinese man who thinks ground-up rhinoceros horn will give him a powerful erection? Or parents who treat their child’s meningitis with maple syrup? How about parents who believe in faith-healing, praying their hearts out as their child withers and dies in agony?
Fundamentalist Cultural Relativism and Morality
Dangerous consequences also emerge when the principles of FCR are applied to problems of moral and ethical evaluation. Here, identifying the deficits of extreme relativism requires slightly more nuance than is necessary to articulate its failings with regard to the problem of scientific truth. This is because the discovery or demonstration of moral absolutes is extraordinarily difficult. Moral systems emerge from culture and cultures change from place to place and evolve over time. As a result, ideas about right and wrong behavior change along with the cultural substrate in which they are embedded.
Consider, for example, the notion of individual human rights. Today, the idea that individuals within human societies have certain rights is frequently taken for granted, and the moral consequences of either supporting or infringing on those rights are often taken into account when debating the merits of law or political action on the local, federal, and international stage. Prior to the advent of modern notions of human rights, the idea that certain humans should have access to a certain range of social resources and granted a certain level of equal treatment as a de facto condition of their existence was an alien proposition. Rights instead flowed from monarchs or religious officials. But since the Enlightenment, the Western notion of human rights has continued to expand, becoming increasingly inclusive as people have come to identify and labor to eradicate previously unexamined strains of prejudice and bigotry. One day – probably very soon – they will expand to encompass certain non-human animals. Our ideas about human rights have changed over time, continually altering the standards by which we judge moral propriety.
Another example should drive the point down to bedrock. Nowhere is the temporal plasticity of moral strictures more clearly demonstrated than in religion, where interpretations of the dictates of moral propriety outlined in sacred texts are constantly renegotiated in light of secular changes. The precise phrasing varies with translation, but the content of the Bible possessed by modern Episcopalians is basically the same as the content of the one possessed by 17th century Puritans. Nevertheless, the two sects exhibit significant differences concerning what kinds of behavior are considered morally acceptable. The primary cause of these deviations is that modern Episcopalians have shifted their understanding of doctrine to accommodate wider cultural changes regarding the perception of what does and does not count as righteous behavior.
So clearly, moral prescriptions change with time and context. Given this, anyone advocating a cross-cultural evaluation of moral propriety might seem to be on shaky ground. Here, it becomes important to recognize that cultural relativism, as a methodological tool, says nothing about moral evaluation. Appropriately applied, cultural relativism supplies an observer with the perspective needed to see – for instance – why parents in certain sects of fundamentalist Christianity might deny their child life-saving medical care. It does prevent anyone from feeling or expressing moral revulsion at the underlying beliefs and the practices they engender. Nor does it prevent a society that places value on the preservation of human life from interceding on the child’s behalf and putting the parents in prison for criminal negligence or even homicide.
The same can be said of any of the horrors that flow from religious fundamentalism: honor killings, homophobic hate crimes, female genital mutilation, child brides, suicide bombings, the torture and murder of heretics in places like Saudi Arabia, and the full litany of offenses that trespass the bounds of the moral intuitions of people who value human life and equality. Cultural relativism should be deployed as a tool to understand why these things occur – why the confluence of certain social contexts and religious beliefs leads people to kill and mutilate one another. It says nothing about how we should react to these things. By positing cultural relativism as a prohibition on a moral evaluation, extreme relativists retreat from all responsibility for upholding modern liberal values like universal education, racial and gender equality, freedom from oppression, and access to basic healthcare.
Indeed, though the articulation of universal human values is a notoriously thorny problem, one can only deny their existence by adopting a fantastical view of the forces structuring human behavior. As a rather odious stew of idealism, FCR entails a rank denial of the existence of anything resembling human nature – replaced, presumably, with a Lockean conception of humans as infinitely malleable blank slates. In this view, human moral proscriptions are fashioned from the aether, their only worldly determinant the cultural milieu in which they arise.
But worldwide, humans create and enforce prohibitions on certain types of in-group killing. Likewise, a capacity to monitor social contracts and detect cheaters seems to be innate, offering a clear indication that humans universally appreciate fairness in social arrangements. An aversion to incest is similarly widespread. Though they are differently expressed in myriad taboos and moral prescriptions, these are strong contenders for universal moral preferences. They very likely have a basis in humanity’s evolved psychology, and therefore offer the crude foundations for a universal code of ethics.
On the other end of the spectrum, humans have innate predispositions toward misconduct that can be exacerbated by an exaggerated emphasis on a cultural as an unassailable fount of moral knowledge. For instance, humans have an ugly impulse toward tribalism. Allowing cultural boundaries to play a greater role in shaping values than our shared identity as humans not only grants tribalism succor, it is a natural consequence of taking absolute moral relativism seriously. It is also detrimental to the project of building and maintaining humanistic moral codes and the universalized standards of human thriving they entail. With its exaggerated celebration of boundless pluralism, FCR has the potential to prove inimical to the practical goals of building and maintaining the social institutions most amenable to human success. Consider the very project of building stable social institutions. While it is important to encourage diversity, it is also true that some degree of assimilation is critical to the formation and long-term stability of societies. When the locus of identity in a given society is a multitude of distinct religions, ethnic affiliations, or political subgroups, the resulting fragmentation is a recipe for long term instability and strife. Each group is bound to pursue clannish interests, guided by moral codes that may be both mutually exclusive and entirely divorced from the best interests of the collective.
This is precisely what has happened in the East Ramapo Central School district in New York, where Orthodox Jews seized the local school board and began cutting services in an attempt to alleviate their tax burden. Their goal was to avoid paying taxes on schools their children didn’t intend, but their myopic focus on religious and ethnic affiliation has lead them to neglect – and even harm – the wellbeing of their neighbors. Adherence to the moral proscriptions of an ancient faith, in concert with a very likely evolved predisposition to favor cultural familiars, has led them to place value on their identity as Orthodox Jews to the exclusion of their identity as human beings – and all the ethical imperatives that identity implies.
Yet even in the absence of any evolved, universal preferences to form the foundation of widely applicable moral prescriptions, it is impossible to advocate extreme forms of cultural relativism without abandoning modest claims about right and wrong. For example, it might be argued without a lot of protest that individual behaviors that encourage or contribute to the physical health or well-being of others are laudable. Conversely, those that can be shown to be directly or indirectly harmful to others are not. It’s good to feed your kids, bad to starve them. A parent who poisons the lungs of her offspring with second-hand smoke can be sensibly accused of engaging in some form of wrongdoing. But what of the men who have foisted upon their wives the bizarre and inarguably sexist tradition of wearing burqas? Not only is this practice reflective of the possessive, proprietary interests of a regressive patriarchy, it has also been linked to vitamin D deficiency and associated problems like rickets.
According to FCR, the latter claim offers no basis for ethical evaluation. Lacking any direct knowledge of what it is like to be a Muslim woman, I have no basis for suggesting that their adherence to the tenets of Islamic faith leads them to live a less healthy and fulfilling life than they could otherwise. Boiled down, extreme relativism argues that my desire to see all humans, everywhere, granted basic human rights is ill-founded. Moreover, it posits that my opinion that certain cultural practices or religious beliefs are inimical to that goal is bigoted. Instead of expressing an interest in the well-being of all humans, the proponents of FCR see this view as “Islamophobic”.
Cultural Relativism as Ethical Obstructionism
By suggesting that cultural differences are essentially unbridgeable and denying the possibility of either uncovering or negotiating a universal standard of human thriving, FCR has the curious consequence of more substantially “otherizing” (to borrow a rather obscurantist term from the social justice world) people from distinct cultural backgrounds. In one of its more popular incarnations, it argues that everyone is gifted with a “positionality” (yet another term of polite PC pedantry) that renders their ideas about right and wrong both externally unintelligible and permanently unassailable. It is to say, in effect, “your worldview is so different from my own that I can hardly justify feeling or expressing outrage when you either abuse or are abused by a member of your cultural subgroup.” At best, this is a recipe for social stagnation. At worst, it’s a way of abrogating centuries of moral progress – dispensing with hard-earned notions of human rights in favor of milquetoast ideas about cultural sensitivity.
FCR is also a direct progenitor of the modern strain of intellectual and moral sensitivity sweeping college campuses in the form of pleas for “trigger warnings” and concern over “microaggressions”. These violations of Enlightened, humanistic ethics seem superficial in comparison to some of more heinous transgressions countenanced by FCR, but they presage something sinister. Where once FCR might have precipitated innocuously nondescript rhetorical utterances along the lines of “who are we to judge?”, it now motivates an authoritarian push for the establishment of pristine thought-sanctuaries – places where ideas are vetted for the slightest hint of potential trespass against the ideals and preferences of cultural subgroups. In this regard, FCR has turned from a bastion for moral cowardice to a direct assault on civil liberties – and it is here that it most significantly earns a sobriquet typically reserved for regressive strains of religious intolerance and slavish adherence to ideology: fundamentalism.
Thomas Paine was absolutely correct when he observed, in The Rights of Man, that natural rights are irrevocable. Though it may be a recent invention, the introduction of the concept of “human rights” represents a monumental transition in our understanding of human ethics. Absent its cataclysmic obliteration from the realm of human thought, it will remain a critical component of our modern assessments of right and wrong. The secular notion of individual human rights is, inarguably, an improvement over previous moral codes, yet FCR – manifest in the mutual unintelligibility implied by overwrought concerns over “positionality” and an obsequious devotion to cultural sensitivity – would have us abandon that progress by asserting that people’s culturally derived beliefs about the will of Allah or the efficacy of vaccines are more important that anyone’s right to live a healthy, independent life. Brass tacks, there are good reasons to think some ways of living are better than others.
I would humbly submit that we shouldn’t throw the proverbial baby out with the bathwater. Cultural relativism, in its original, lighter form has proven immensely useful to anthropologists and ethnographers seeking to understand the proximate causes of locally expressed human differences. It was a substantial leap over the ham-fisted, Western-centric, deeply racialized theorizing of 19th century intellectuals. The underlying principle, that all human cultural systems deserve to be understood in their own terms, is useful as more than just a methodological tool for ethnographers. On the global stage, it has a role to play in the shaping of foreign policy and international affairs. On the more humble scale of individual lives, it has a role to play in helping neighbors understand one another.
It just needs to be deployed under the recognition that, inasmuch as it is useful tool for building human understanding, it is has little value as tool for evaluating knowledge claims and moral proscriptions. To the extent that cultural relativism, manifest in its fundamentalist form, is interpreted as endorsing the primacy of subjective experience in dictating the structure of reality or mandating abstinence from adopting or defending human rights on the grounds that those rights are a “Western construct”, it deserves all the ridicule it receives. Science is the best tool for gauging truth. Cultural practices that inhibit the achievement of universal human rights can be justifiably viewed as harmful and ought to be stridently opposed and vigorously critiqued. These assertions aren’t oppressive or marginalizing or bigoted. They’re true.