Today, there is a crisis of trust in science. Many people–including politicians and, yes, even presidents–publicly express doubts about the validity of scientific findings. Meanwhile, scientific institutions and journals express their concerns about the public’s increasing distrust in science. How is it possible that science, the products of which permeate our everyday lives, making them in many ways more comfortable, elicits such negative attitudes among a substantial part of the population? Understanding why people distrust science will go a long way towards understanding what needs to be done for people to take science seriously.

Political ideology is seen by many researchers as the main culprit of science skepticism. The sociologist Gordon Gauchat has shown that political conservatives in the United States have become more distrusting of science, a trend that started in the 1970s. And a swath of recent research conducted by social and political psychologists has consistently shown that climate-change skepticism in particular is typically found among those on the conservative side of the political spectrum. However, there is more to science skepticism than just political ideology.

The same research that has observed the effects of political ideology on attitudes towards climate change has also found that political ideology is not that predictive of skepticism about other controversial research topics. Work by the cognitive scientist Stephan Lewandowsky, as well as research led by the psychologist Sydney Scott, observed no relation between political ideology and attitudes toward genetic modification. Lewandowsky also found no clear relation between political conservatism and vaccine skepticism.

So there is more that underlies science skepticism than just political conservatism. But what? It is important to systematically map which factors do and do not contribute to science skepticism and science (dis)trust in order to provide more precise explanations for why a growing number of individuals reject the notion of anthropogenic climate change, or fear that eating genetically modified products is dangerous, or believe that vaccines cause autism.

My colleagues and I recently published a set of studies that investigated science trust and science skepticism. One of the take-home messages of our research is that it is crucial not to lump various forms of science skepticism together. And although we were certainly not the first to look beyond political ideology, we did note two important lacunae in the literature. First, religiosity has so far been curiously under-researched as a precursor to science skepticism, perhaps because political ideology commanded so much attention. Second, current research lacks a systematic investigation into various forms of skepticism, alongside more general measures of trust in science. We attempted to correct both oversights.

People can be skeptical or distrusting of science for different reasons, whether it is about one specific finding from one discipline (for example, ‘The climate is not warming, but I believe in evolution’), or about science in general (‘Science is just one of many opinions’). We identified four major predictors of science acceptance and science skepticism: political ideology; religiosity; morality; and knowledge about science. These variables tend to intercorrelate–in some cases quite strongly–which means that they are potentially confounded. To illustrate, an observed relation between political conservatism and trust in science might in reality be caused by another variable, for example, religiosity. When not measuring all constructs simultaneously, it is hard to properly assess what the predictive value of each of these is.

So, we investigated the heterogeneity of science skepticism among samples of North American participants (a large-scale cross-national study of science skepticism in Europe and beyond will follow). We provided participants with statements about climate change (eg, “Human CO2 emissions cause climate change”), genetic modification (eg, “GM of foods is a safe and reliable technology”), and vaccination (eg, “I believe that vaccines have negative side effects that outweigh the benefits of vaccination for children”). Participants could indicate to what extent they agreed or disagreed with these statements. We also measured participants’ general faith in science, and included a task in which they could indicate how much federal money should be spent on science, compared with various other domains. We assessed the impact of political ideology, religiosity, moral concerns and science knowledge (measured with a science literacy test, consisting of true or false items such as “All radioactivity is made by humans,” and “The center of the Earth is very hot”) on participants’ responses to these various measures.