When we target chemophobia, are we punching down?

Over at Pharyngula, Chris Clarke challenges those in the chemical know on their use of “dihydrogen monoxide” jokes. He writes:

Doing what I do for a living, I often find myself reading things on Facebook, Twitter, or those increasingly archaic sites called “blogs” in which the writer expresses concern about industrial effluent in our air, water, consumer products or food. Sometimes the concerns are well-founded, as in the example of pipeline breaks releasing volatile organic chemicals into your backyard. Sometimes, as in the case of concern over chemtrails or toxic vaccines, the concerns are ill-informed and spurious.

And often enough, the educational system in the United States being the way it’s been since the Reagan administration, those concerns are couched in terms that would not be used by a person with a solid grounding in science. People sometimes miss the point of dose-dependency, of acute versus chronic exposure, of the difference between parts per million and parts per trillion. Sometimes their unfamiliarity with the basic facts of chemistry causes them to make patently ridiculous alarmist statements and then double down on them when corrected.

And more times than I can count, if said statements are in a public venue like a comment thread, someone will pipe up by repeating a particular increasingly stale joke. Say it’s a discussion of contaminants in tap water allegedly stemming from hydraulic fracturing for natural gas extraction. Said wit will respond with something like:

“You know what else might be coming out of your tap? DIHYDROGEN MONOXIDE!”

Two hydrogens, one oxygen … what’s coming out of your tap here is water. Hilarious! Or perhaps not.

Clarke argues that those in the chemical know whip out the dihydrogen monoxide joke to have a laugh at the expense of someone who doesn’t have enough chemical knowledge to understand whether conditions they find alarming really ought to alarm them. However, how it usually goes down is that other chemically literate people in earshot laugh while the target of the joke ends up with no better chemical understanding of things.

Really, all the target of the joke learns is that the teller of the joke has knowledge and is willing to use it to make someone else look dumb.

Clarke explains:

Ignorance of science is an evil that for the most part is foisted upon the ignorant. The dihydrogen monoxide joke depends for its humor on ridiculing the victims of that state of affairs, while offering no solution (pun sort of intended) to the ignorance it mocks. It’s like the phrase “chemophobia.” It’s a clan marker for the Smarter Than You tribe.

The dihydrogen monoxide joke punches down, in other words. It mocks people for not having had access to a good education. And the fact that many of its practitioners use it in order to belittle utterly valid environmental concerns, in the style of (for instance) Penn Jillette, makes it all the worse — even if those concerns aren’t always expressed in phraseology a chemist would find beyond reproach, or with math that necessarily works out on close examination.

There’s a weird way in which punching down with the dihydrogen monoxide joke is the evil twin of the “deficit model” in science communication.

The deficit model assumes that the focus in science communication to audiences of non-scientists should be squarely on filling in gaps in their scientific knowledge, teaching people facts and theories that they didn’t already know, as if that is the main thing they must want from science. (It’s worth noting that the deficit model seems to assume a pretty unidirectional flow of information, from the science communicator to the non-scientist.)

The dihydrogen monoxide joke, used the way Clarke describes, identifies a gap in understanding and then, instead of trying to fill it, points and laughs. If the deficit model naïvely assumes that filling gaps in knowledge will make the public cool with science, this kind of deployment of the dihydrogen monoxide joke seems unlikely to provoke any warm feelings towards science or scientists from the person with a gappy understanding.

What’s more, this kind of joking misses an opportunity to engage with what they’re really worried about and why. Are they scared of chemicals per se? Of being at the mercy of others who have information about which chemicals can hurt us (and in which amounts) and/or who have more knowledge about or control of where those chemicals are in our environment? Do they not trust scientists at all, or are they primarily concerned about whether they can trust scientists in the employ of multinational corporations?

Do their concerns have more to do with the information and understanding our policymakers have with regard to chemicals in our world — particularly about whether these policymakers have enough to keep us relatively safe, or about whether they have the political will to do so?

Actually having a conversation and listening to what people are worried about could help. It might turn out that people with the relevant scientific knowledge to laugh at the dihydrogen monoxide joke and those without share a lot of the same concerns.

Andrew Bissette notes that there are instances where the dihydrogen monoxide joke isn’t punching down but punching up, where educated people who should know better use large platforms to take advantage of the ignorant. So perhaps it’s not the case that we need a permanent moratorium on the joke so much as more careful thought about what we hope to accomplish with it.

Let’s return to Chris Clarke’s claim that the term “chemophobia” is “a clan marker for the Smarter Than You tribe.”

Lots of chemists in the blogosphere regularly blog and tweet about chemophobia. If they took to relentlessly tagging as “chemophobe!” people who are lacking access to the body of knowledge and patterns of reasoning that define chemistry, I’d agree that it was the same kind of punching down as the use of the dihydrogen monoxide joke Clarke describes. To the extent that chemists are actually doing this to assert membership in the Smarter Than You tribe, I think it’s counterproductive and mean to boot, and we should cut it out.

But, knowing the folks I do who blog and tweet about chemophobia, I’m pretty sure their goal is not to maintain clear boundaries between The Smart and The Dumb. When they fire off a #chemophobia tweet, it’s almost like they’re sending up the Batsignal, rallying their chemical community to fight some kind of crime.

So what is it these chemists — the people who have access to the body of knowledge and patterns of reasoning that define chemistry — find problematic about the “chemophobia” of others? What do they hope to accomplish by pointing it out?

Part of where they’re coming from is probably grounded in good old fashioned deficit-model reasoning, but with more emphasis on helping others learn a bit of chemistry because it’s cool. There’s usually a conviction that the basics of the chemistry that expose the coolness are not beyond the grasp of adults of normal intelligence — if only we explain in accessibly enough. Ash Jogalekar suggests more concerted efforts in this direction, proposing a lobby for chemistry (not the chemical industry) that takes account of how people feel about chemistry and what they want to know. However it’s done, the impulse to expose the cool workings of a bit of the world to those who want to understand them should be offered as a kindness. Otherwise, we’re doing it wrong.

Another part of what moves the chemists I know who are concerned with chemophobia is that they don’t want people who are not at home with chemistry to get played. They don’t want them to be vulnerable to quack doctors, nor to merchants of doubt trying to undermine sound science to advance a particular economic or political end, nor to people trying to make a buck with misleading claims, nor to legitimately confused people who think they know much more than they really do.

People with chemical know-how could help address this kind of vulnerability, being partners to help sort out the reliable information from the bogus, the overblown risks from risks that ought to be taken seriously or investigated further.

But short of teaching the folks without access to the body of knowledge and patterns of reasoning that define chemistry everything they know to be their own experts (which is the deficit model again), providing this kind of help requires cultivating trust. It requires taking the people to whom your offering the help seriously, recognizing that gaps in their chemical understanding don’t make them unintelligent or of less value as human beings.

And laughing at the expense of the people who could use your help — using your superior chemical knowledge to punch down — seems unlikely to foster that trust.

Want good reasons to be a Creationist? You won’t find them here.

I don’t know why it surprises me when technology reporters turn out to be not only anti-science, but also deeply confused about what’s actually going on in scientific knowledge-building. Today’s reminder comes in Virginia Heffernan’s column, “Why I’m a creationist”.

There seems not to be much in the way of a coherent argument in support of Creationism in the column. As near as I can tell, Heffernan is down on science because:

  1. Science sometimes uses chains of inference that are long and complicated.
  2. Science has a hard time coming up with decisive answers to complicated questions (at least at a satisfyingly prompt rate).
  3. Science maybe provides some good reasons to worry about the environment, and she’d prefer not to worry about the environment.
  4. A scientist was mean to a religious person at some point. Some scientists just don’t seem like nice people.
  5. Science trades in hypotheses, and hypotheses aren’t facts — they could be false!
  6. Darwin based his whole theory on a tautology, “whatever survives survives”! [Nope!]
  7. Evolutionary psychology first claimed X, then claimed Y (which seems to directly contradict X), and neither of those claims seems to have especially rigorous empirical backing … so all of evolutionary theory must be wrong!
  8. Evolutionary theory just isn’t as compelling (at least to Heffernan) as a theory of human origins should be.

On item #5 there, if this is an issue for one’s acceptance of evolutionary theory, it’s also an issue for one’s acceptance knowledge claims from other areas of science.

This is something we can lay at the feet of the problem of induction. But, we can also notice that scientists deal quite sensibly with the problem of induction lurking in the background. Philosopher of science Heather Douglas explains this nicely in her book Science, Policy, and the Value-Free Ideal, where she describes what it means for scientists to accept a hypothesis.

To say P has been accepted is to say P belongs to the stock of established scientific knowledge, which means it satisfies criteria for standards of appraisal from within science (including what kind of empirical evidence there is for P, whether there is empirical evidence that supports not-P, etc.). Accepting P is saying that there is no reason to expect that P will be rejected after more research, and that only general inductive doubts render P uncertain.

That’s as certain as knowledge can get, at least without a divine guarantee. Needless to say, such a “guarantee” would present epistemic problems of its own.

As for Heffernan’s other reasons for preferring Creationism to science, I’m not sure I have much to say that I haven’t already said elsewhere about why they’re silly, but I invite you to mount your own critiques in the comments.

Professional communities, barriers to inclusion, and the value of a posse.

Last week, I wrote a post about an incident connected to a professional conference. A male conference-goer wrote a column attempting to offer praise for a panel featuring four female conference-goers but managed to package this praise in a way that reinforced sexist assumptions about the value women colleagues add to a professional community.

The women panelists communicated directly with the male commentator about his problematic framing. The male commentator seemed receptive to this feedback. I blogged about it as an example of why it’s important to respond to disrespect within professional communities, even if it’s not intended as disrespect, and despite the natural inclination to let it go. And my post was praised for offering a discussion of the issue that was calm, sensitive, and measured.

But honestly? I’m unconvinced that my calm, sensitive, measured discussion will do one whit of good to reduce the incidence of such casual sexism in the future, in the community of science journalist or in any other professional community. Perhaps there were some readers who, owing to the gentle tone, were willing to examine the impact of describing colleagues who are women primarily in terms of their looks, but if a less gentle tone would have put them off from considering the potential for harm to members of their professional communities, it’s hard to believe these readers would devote much energy to combatting these harms — whether or not they were being asked nicely to do so.

Sometimes someone has to really get your attention — in a way that shakes you up and makes you deeply uncomfortable — in order for you to pay attention going forward. Maybe feeling bad about the harm to someone else is a necessary first step to developing empathy.

And certainly, laying out the problem while protecting you from what it feels like to be one of the people struggling under the effects of that problem takes some effort. If going to all that trouble doesn’t actually leave enough of an impression to keep the problem from happening some more, what’s the point?

* * * * *

What does it take to create a diverse professional community? It requires more than an absence of explicit rules or standing practices that bar certain kinds of people from membership, more even that admitting lots of different kinds of people into the “pipeline” for that profession. If you’re in the community by virtue of your educational or employment status but you’re not actually part of the discussions that define your professional community, it may help the appearance of diversity, but not the reality of it.

The chilly climate women have been talking about in a variety of male-dominated professional communities is a real thing.

Being a real member of a professional community includes being able to participate fully in venues for getting your work and insights into the community’s discussions. These venues include journals and professional meetings, as well as panels or study sections that evaluate grant proposals. Early in one’s membership in a professional community, venues like graduate seminars and department symposia are also really important.

One problem here is that usually individuals without meaningful access to participation are also without the power in the community required to effectively address particular barriers to their access. Such individuals can point out the barriers, but they are less likely to be listened to than someone else in the community without those barriers.

Everyday sexism is just one such barrier.

This barrier can take a number of particular forms.

For the students on their way into a professional community, it’s a barrier to find out that senior members of the community who you expected would help train you and eventually take you seriously as a colleague are more inclined to sexualize you or full-on sexually harass you. It’s a barrier when you see people in your community minimize that behavior, whether offhandedly or with rather more deliberation.

It’s a barrier when members of your community focus on your looks rather than your intellectual contributions, or act like it’s cute or somehow surprising that someone like you could actually make an intellectual contribution. It’s a further barrier when other members of your community advise you to ignore tangible disrespect because surely it wasn’t intentional — especially when those other members of the community make no visible effort to help address the disrespect.

It’s a barrier when students don’t see people like themselves represented among the recognized knowledge-builders in the professional community as they are being taught the core knowledge expected of members of that community. It’s also a barrier when the more senior members of the professional community are subject to implicit biases in their expert evaluations of who’s cut out to be a full contributing member of the community.

Plenty of well-meaning folks in professional communities that have a hard time fully integrating women (among others) may be puzzled as to why this is so. If they don’t personally experience the barriers, they may not even realize that they’re there. Listening to lived experiences of their female colleagues might reveal some of the barriers — but listening also assumes that the community really takes its female members seriously as part of the community, when this is precisely the problem with which the women in the community are struggling.

* * * * *

Professional meetings can be challenging terrain for women in predominantly male professional communities. Such meetings are essential venues in which to present one’s work and get career credit for doing so. They are also crucially important for networking and building relationships with people who might become collaborators, who will be called on to evaluate one’s work, and who are the peers with whom one hopes to be engaged in productive discussions over the course of one’s career.

There is also a strong social component to these meetings, an imperative to have fun with one’s people — which is to say, in this context, the people with whom one shares a professional community. Part of this, I think, is related to how strongly people identify with their professional community: the connection is not just about what people in that community do but about who they are. They have taken on the values and goals of the professional community as their own. It’s not just a job, it’s a social identity.

For some people, the social component of professional meetings has a decidedly carnal flavor. Unfortunately, rejecting a pass from someone in your professional community, especially someone with more power in that community than you, can screw with your professional relationships within the community — even assuming that the person who made the pass accepts your “no” and moves on. In other cases, folks within the professional community may be perfectly aware of power gradients and willing to use them to get what they want, applying persistent unwanted attention that can essentially deprive the target of full participation in the conference. Given the importance professional conferences have, this is a significant professional harm.

Lest you imagine that this is a merely hypothetical worry, I assure you that it is not. If you ask around you may discover that some of the members of your professional community choose which conference sessions to attend in order to avoid their harassers. That is surely a constraint on how much one can get out of a professional meeting.

Recently a number of conferences and conventions have adopted policies against harassment, policies that are getting some use. Many of these are fan-oriented conventions or tech conferences, rather than the kind of research oriented, academically inclined professional meetings most of us university types attend. I know of at least one scientific professional society (the American Astronomical Society) that has adopted a harassment policy for its meetings and that seems generally to be moving in a good direction from the point of view of building an inclusive community. However, when I checked the websites of three professional societies to which I belong (American Chemical Society, American Philosophical Association, and Philosophy of Science Association), I could find no sign of anti-harassment policies for their conferences. This is disappointing, but not surprising to me.

The absence of anti-harassment policies doesn’t mean that there’s no harassment happening at the meetings of these professional societies, either.

And even if a professional community has anti-harassment policies in place for its meetings, this doesn’t remove the costs — especially on a relatively junior member of the community — associated with asking that the policies be enforced. Will a professional society be willing to caution a member of the program committee for the conference? To eject the most favored grad student of a luminary in the field — or, for that matter, a luminary — who violates the policy? Shining light on over-the-line behavior at conferences is a species of whistleblowing, and is likely to be received about as warmly as other forms.

* * * * *

Despite the challenges, I don’t think the prospects for building diverse and productive professional communities are dim. Progress is being made, even if most weeks the pace of progress is agonizingly slow.

But I think things could get better faster if people who take their professional communities for granted step up and become more active in maintaining them.

In much the same way that it is not science that is self-correcting but rather individual scientists who bother to engage critically with particular contributions to the ongoing scientific conversation and keep the community honest, a healthy professional community doesn’t take care of itself — at least, not without effort on the part of individual members of the community.

Professional communities require everyday maintenance. They require tending to keep their collective actions aligned with the values members of the community say they share.

People who work very hard to be part of a professional community despite systemic barriers are people committed enough to the values of the professional community to fight their way through a lot of crap. These are people who really care about the values you purport to care about as a member of the professional community, else why would they waste their time and effort fighting through the crap?

These are the kind of people you should want as colleagues, at least if you value what you say you value. Their contributions could be huge in accomplishing your community’s shared goals and ensuring your community a vibrant future.

Even more than policies that aim to address systemic barriers to their entry to the professional community, these people need a posse. They need others in the community who are unwilling to sacrifice their values — or the well-being of less powerful people who share those values — to take consistent stands against behaviors that create barriers and that undermine the shared work of the community.

These stands needn’t be huge heroic gestures. It could be as simple as reliably being that guy who asks for better gender balance in planning seminars, or who reacts to casual sexist banter with, “Dude, not cool!” It could take the form of asking about policies that might lessen barriers, and taking on some of the work involved in creating or implementing them.

It could be listening to your women colleagues when they describe what it has been like for them within your professional community and assuming the default position of believing them, rather than looking for possible ways they must have misunderstood their own experiences.

If you care about your professional community, in other words, the barriers to entry in the way of people who want badly to be part of that community because they believe fiercely in its values are your problem, too. Acting like it, and doing your part to address these barriers, is sharing the regular maintenance of the professional community you count on.

_____________
While this post is focused on barriers to full participation in professional communities that flow from gender bias, there are plenty of other types of bias that throw up similar barriers, and that could benefit from similar types of response from members of the professional communities not directly targeted by these biases.

Addressing (unintended) disrespect in your professional community.

I am a believer in the power of the professional conference. Getting people in the same room to share ideas, experiences, and challenges is one of the best ways to build a sense of community, to break down geographical and generational barriers, to energize people and remind them what they love about what they’re doing.

Sometimes, though, interactions flowing from a professional conference have a way of reinforcing barriers. Sometimes a member of the community makes an attempt to express appreciation of colleagues that actually has the effect of treating those colleagues like they’re not really part of the community after all.

Last week, the 8th World Conference of Science Journalists met in Helsinki, Finland. Upon his return from the conference, journalist Nicolás Luco posted a column reflecting on his experience there. (Here’s an English translation of the column by Wladimir Labeikovsky.) Luco’s piece suggests some of the excitement of finding connections with science journalists from other countries, as well as finding common ground with journalists entering the profession in a very different decade with a panoply of different technological tools:

If I hadn’t come, I wouldn’t have had that experience. I have submerged into an atmosphere where what I had seen as the future is already taken for granted. And yet, the fundamentals [e.g., that the story is what matters] remain.

It is, without a doubt, a description of a very positive personal experience.

However, Luco’s column is also a description of his experience of female colleagues at this conference framed primarily in terms of their physical attributes: shining blonde hair, limpid blue eyes, translucent complexions, apparent youth. His description of the panel of journalists using the tools of new media to practice the fundamentals of good journalism describes them as

four Americans: Rose, Lena, Kathleen and Erin (blonde), none older than 25

All of the other conference-goers who are identified by name are identified with surnames as well as given names. We do learn of the two women identified by their full names in the column that they are not blonde. It is left to the reader to imagine the hair color of Philip J. Hilts, the only male attendee mentioned by name.

I understand that Nicolás Luco was aiming to give a vivid visual description to draw his readers into his experience of being in Helsinki for this conference, and that this description was meant to convey a positive, optimistic mood about the future of science journalism.

But I also understand that these stylistic choices carry baggage that make it harder for Rose Eveleth, and Lena Groeger, and Kathleen Raven, and Erin Podolak, the journalists on the panel, to be taken seriously within this international community of science journalists.

Their surnames matter. In a field where they want their work to be recognized, disconnecting their bylines from the valuable insights they shared as part of a conference panel is not helpful.

Moreover, I am told that the journalistic convention is to identify adults by full name, and to identify people by first name alone only when those people are children.

Eveleth, Groeger, Raven, and Podolak are not children. They may seem relatively young to a journalist who came into the profession in the age of linotype (indeed, to the extent that he underestimated their ages, which range from 25 to 30), but they are professionals. Their ages should not be a barrier to treating them as if they are full members of the professional community of science journalists, but focusing unduly on their ages could well present such a barrier.

And, needless to say, their hair color should have no relevance at all in assessing whether they are skilled journalists with valuable insights to share.

As it happens, only days before the 8th World Conference of Science Journalists, Podolak wrote a blog post describing why she needs feminism. In that post, she wrote:

I’m a feminist for myself because yes, I want a fair shake, I want to be recognized for the value of my work and not whether or not my hair looks shiny that day. But, adding my voice to the other feminist voices out there is about more than just me. I’ve got it pretty good. I’m not trying to argue that I don’t. But I can support the women out there who are dealing with overt sexism, who are being attacked. I can try to be an ally. That to me is the real value of feminism, of standing together.

It is profoundly disheartening to take yourself to be accepted by your professional community, valued for the skills and ideas you bring to the table, only to discover that this is not how your presumptive colleagues actually see you. You would think that other journalists should be the ones most likely to appreciate the value of using new technologies to tell compelling stories. What a disappointment to find that their focus gets stuck on the surface. Who can tell whether the work has value if the hair of the journalist is shiny?

You will likely not be surprised that Eveleth, Groeger, Raven, and Podolak were frustrated at Nicolás Luco’s description of their panel, despite understanding that Luco was trying to be flattering. In an email to Luco the four sent in response to the column, they wrote:

Leading your story with a note about your attraction to blondes and then noting Erin’s hair color, is both inappropriate and, frankly, sexist. We were not there for anyone to ogle, and our physical appearance is completely irrelevant to the point of our panel. It is important for you to understand why were are upset about your tone in this piece. Women are constantly appraised for their looks, rather than their thoughts and skills, and in writing your story the way you did you are contributing to that sexism.

And, in a postscript to that email, Kathleen Raven noted:

I was under the impression that you wrote your article using hair color as a narrative tool to tie together your meetings with journalists. I appreciate this creativity, but I am worried that American women can perceive — as we have — the article as not fully respecting us as journalists in our own right.

What Eveleth, Groeger, Raven, and Podolak are up against is a larger society that values women more for their aesthetic appeal than their professional skills. That their own professional community repeats this pattern — presenting them as first young and pretty and only secondarily as good journalists — is a source of frustration. As Eveleth wrote to me:

Last I checked, being pretty has nothing to do with your skills at any kind of journalism. Having long blonde hair is not going to get Erin the story. Erin is going to get the story because she’s good at her job, because she’s got experience and passion, because she’s talented and tough and hard working. The same goes for Kathleen and Lena. 

The idea that it is not just okay, but actually complimentary to focus on a young woman’s (or really any aged woman’s) looks as leading part of her professional identity is wrong. The idea that it’s flattering to call out Erin’s hair and age before her skills is wrong. The idea that a woman’s professional skill set is made better if she is blonde and pretty is wrong. And the idea that someone who writes something like this should just be able to pass it off as “tongue in cheek” or “a cultural difference” is also wrong.

I should pause here to take note of another dimension of professional communities in this story. There is a strong pressure to get along with one’s professional colleagues, to get along rather than raising a fuss. Arguably this pressure is stronger on newer members of a professional community, and on members of that community with characteristics (e.g., of gender, race, disability, etc.) that are not well represented in the more established members of that professional community.

Practically, this pressure manifests itself as an inclination to let things go, to refrain from pointing out the little instances which devalue one’s professional identity or status as a valued member of the community. Most of the time it seems easier to sigh and say to oneself, “Well, he meant well,” or, “What can you expect from someone of that generation/cultural background?” than to point out the ways that the comments hurt. It feels like a tradeoff where you should swallow some individual hurt for the good of the community.

But accepting this tradeoff is accepting that your full membership in the community (and that of others like you) is less important. To the extent that you believe that you make a real contribution to the community, swallowing your individual hurt is dancing on the edge of accepting what is arguably a harm to the professional community as a whole by letting the hurtful behaviors pass unexamined.

Eveleth, Groeger, Raven, and Podolak had more respect than that for their professional community, and for Nicolás Luco as a professional colleague. They did not just sigh and roll their eyes. Rather, they emailed Luco to explain what the problem was.

In his reply to them (which I quote with his permission), Luco makes it clear that he did not intend to do harm to anyone, especially not to Eveleth, Groeger, Raven, and Podolak, with his column. Still, he also makes it clear that he may not fully grasp just what the problem is:

I write as a known voice who can write tongue in cheek and get away with it because I am willing to laugh at myself. 

I strive to make what I write entertaining. And maybe sneak in the more serious arguments.

Sorry about my misjudgment on your ages.  But the point is: you are generations apart.

I did not include your last names because they would interrupt the flow of reading and clog the line with surnames, an obstacle.

Finally, it is so much in U.S. culture to discard the looks vis a vis the brains when the looks, as President Clinton knows so well, can be a good hook into the brains.  And since this is a personal column, in the first person singular, I can tell how I personally react at good looks.  For example, Ms. Anne Glover, was extraordinarily beautiful and charming besides being bright and political, which helps, in front of the probable mean thoughts of envious uglier looking colleagues.

Thank you, I still prize the panel as the best and most important in the Conference.

Is there a way Nicolás Luco could have described his personal experience of the conference, and of this panel within the conference that he found particular valuable, in a way that was entertaining, even tongue-in-cheek, while avoiding the pitfalls of describing his female colleagues in terms that undercut their status in the professional community? I think so.

He might, for example, have talked about his own expectations that journalists who are generations apart would agree upon what makes good journalism good journalism. The way that these expectations were thwarted would surely be a good opportunity to laugh at oneself.

He might even have written about his own surprise that a young women he finds attractive contributed a valuable insight — using this as an opportunity to examine this expectation and whether it’s one he ought to be carrying around with him in his professional interactions. There’s even a line in his column that seems like it might provide a hook for this bit of self-examination:

Erin, the youngest and a cancer specialist, insists that decorations don’t matter: good journalism is good journalism, period. Makes me happy.

(Bold emphasis added.)

Extending the lesson about the content of the story mattering more than the packaging to a further lesson about the professional capabilities of the storyteller mattering more than one’s reaction to her superficial appearance — that could drive home some of the value of a conference like this.

Nicolás Luco wrote the column he wrote. Eveleth, Groeger, Raven, and Podolak took him seriously as a professional colleague who is presumptively concerned to strengthen their shared community. They asked him to consider the effect of his description on members of the professional community who stand where they do, to take responsibility as a writer for even the effects of his words that he had not intended or foreseen.

Engaging with colleagues when they hurt us without meaning to is not easy work, but it’s absolutely essential to the health of a professional community. I am hopeful that this engagement will continue productively.

The ethics of opting out of vaccination.

At my last visit to urgent care with one of my kids, the doctor who saw us mentioned that there is currently an epidemic of pertussis (whooping cough) in California, one that presents serious danger for the very young children (among others) hanging out in the waiting area. We double-checked that both my kids are current on their pertussis vaccinations (they are). I checked that I was current on my own pertussis vaccination back in December when I got my flu shot.

Sharing a world with vulnerable little kids, it’s just the responsible thing to do.

You’re already on the internet reading about science and health, so it will probably come as no surprise to you that California’s pertussis epidemic is a result of the downturn in vaccination in recent years, nor that this downturn has been driven in large part by parents worried that childhood vaccinations might lead to their kids getting autism, or asthma, or some other chronic disease. Never mind that study after study has failed to uncover evidence of such a link; these parents are weighing the risks and benefits (at least as they understand them) of vaccinating or opting out and trying to make the best decision they can for their children.

The problem is that the other children with which their children are sharing a world get ignored in the calculation.

Of course, parents are accountable to the kids they are raising. They have a duty to do what is best for them, as well as they can determine what that is. They probably also have a duty to put some effort into making a sensible determination of what’s best for their kids (which may involve seeking out expert advice, and evaluating who has the expertise to be offering trustworthy advice).


But parents and kids are also part of a community, and arguably they are accountable to other members of that community. I’d argue that members of a community may have an obligation to share relevant information with each other — and, to avoid spreading misinformation, not to represent themselves as experts when they are not. Moreover, when parents make choices with the potential to impact not only themselves and their kids but also other members of the community, they have a duty to do what is necessary to minimize bad impacts on others. Among other things, this might mean keeping your unvaccinated-by-choice kids isolated from kids who haven’t been vaccinated because of their age, because of compromised immune function, or because they are allergic to a vaccine ingredient. If you’re not willing to do your part for herd immunity, you need to take responsibility for staying out of the herd.

Otherwise, you are a free-rider on the sacrifices of the other members of the community, and you are breaking trust with them.

I know from experience that this claim upsets non-vaccinating parents a lot. They imagine that I am declaring them bad people, guilty of making a conscious choice to hurt others. I am not. However, I do think they are making a choice that has the potential to cause great harm to others. If I didn’t think that pointing out the potential consequences might be valuable to these non-vaccinating parents, at least in helping them understand more fully what they’re choosing, I wouldn’t bother.

So here, let’s take a careful look at my claim that vaccination refuseniks are free-riders.


First, what’s a free-rider?


In the simplest terms, a free-rider is someone who accepts a benefit without paying for it. The free-rider is able to partake of this benefit because others have assumed the costs necessary to bring it about. But if no one was willing to assume those costs (or indeed, in some cases, if there is not a critical mass of people assuming those costs), then that benefit would not be available, either.


Thus, when I claim that people who opt out of vaccination are free-riders on society, what I’m saying is that they are receiving benefits for which they haven’t paid their fair share — and that they receive these benefits only because other members of society have assumed the costs by being vaccinated.


Before we go any further, let’s acknowledge that people who choose to vaccinate and those who do not probably have very different understandings of the risks and benefits, and especially of their magnitudes and likelihoods. Ideally, we’d be starting this discussion about the ethics of opting out of vaccination with some agreement about what the likely outcomes are, what the unlikely outcomes are, what the unfortunate-but-tolerable outcomes are, and what the to-be-avoided-at-all-costs outcomes are.


That’s not likely to happen. People don’t even accept the same facts (regardless of scientific consensus), let alone the same weightings of them in decision making.


But ethical decision making is supposed to help us get along even in a world where people have different values and interests than our own. So, plausibly, we can talk about whether certain kinds of choices fit the pattern of free-riding even if we can’t come to agreement on probabilities and a hierarchy of really bad outcomes.


So, let’s say all the folks in my community are vaccinated against measles except me. Within this community (assuming I’m not wandering off to exotic and unvaccinated lands, and that people from exotic and unvaccinated lands don’t come wandering through), my chances of getting measles are extremely low. Indeed, they are as low as they are because everyone else in the community has been vaccinated against measles — none of my neighbors can serve as a host where the virus can hang out and then get transmitted to me. (By the way, the NIH has a nifty Disease Transmission Simulator that you can play around with to get a feel for how infectious diseases and populations whose members have differing levels of immunity interact.)


I get a benefit (freedom from measles) that I didn’t pay for. The other folks in my community who got the vaccine paid for it.


In fact, it usually doesn’t require that everyone else in the community be vaccinated against measles for me to be reasonably safe from it. Owing to “herd immunity,” measles is unlikely to run through the community if the people without immunity are relatively few and well interspersed with the vaccinated people. This is a good thing, since babies in the U.S. don’t get their first vaccination against measles until 12 months, and some people are unable to get vaccinated even if they’re willing to bear the cost (e.g., because they have compromised immune systems or are allergic to an ingredient of the vaccine). And, in other cases, people may get vaccinated but the vaccines might not be fully effective — if exposed, they might still get the disease. Herd immunity tends to protect these folks from the disease — at least as long as enough of the herd is vaccinated.


If too few members of the herd are vaccinated, even some of those who have borne the costs of being vaccinated (because even very good vaccines can’t deliver 100% protection to 100% of the people who get them), or who would bear those costs were they able (owing to their age or health or access to medical care), may miss out on the benefit. Too many free-riders can spoil things even for those who are paying their fair share.


A standard reply from non-vaccinating parents is that their unvaccinated kids are not free-riders on the vaccinated mass of society because they actually get diseases like chicken pox, pertussis, and measles (and are not counting on avoiding the other diseases against which people are routinely vaccinated). In other words, they argue, they didn’t pay the cost, but they didn’t get the benefit, either.


Does this argument work?


I’m not convinced that it does. First off, even though unvaccinated kids may get a number of diseases that their vaccinated neighbors do not, it is still unlikely that they will catch everything against which we routinely vaccinate. By opting out of vaccination but living in the midst of a herd that is mostly vaccinated, non-vaccinating parents significantly reduce the chances of their kids getting many diseases compared to what the chances would be if they lived in a completely unvaccinated herd. That statistical reduction in disease is a benefit, and the people who got vaccinated are the ones paying for it.


Now, one might reply that unvaccinated kids are actually incurring harm from their vaccinated neighbors, for example if they contract measles from a recently vaccinated kid shedding the live virus from the vaccine. However, the measles virus in the MMR vaccine is an attenuated virus — which is to say, it’s quite likely that unvaccinated kids contacting measles from vaccinated kids will have a milder bout of measles than they might have if they had been exposed to a full-strength measles virus out in the wild.
 A milder case of measles is a benefit, at least when the alternative is a severe case of measles. Again, it’s a benefit that is available because other people bore the cost of being vaccinated.


Indeed, even if they were to catch every single disease against which we vaccinate, unvaccinated kids would still reap further benefits by living in a society with a high vaccination rate. The fact that most members of society are vaccinated means that there is much less chance that epidemic diseases will shut down schools, industries, or government offices, much more chance that hospitals and medical offices will not be completely overwhelmed when outbreaks happen, much more chance that economic productivity will not be crippled and that people will be able to work and pay the taxes that support all manner of public services we take for granted.

The people who vaccinate are assuming the costs that bring us a largely epidemic-free way of life. Those who opt out of vaccinating are taking that benefit for free.


I understand that the decision not to vaccinate is often driven by concerns about what costs those who receive the vaccines might bear, and whether those costs might be worse than the benefits secured by vaccination. Set aside for the moment the issue of whether these concerns are well grounded in fact. Instead, let’s look at the parallel me might draw: 
If I vaccinate my kids, no matter what your views about the etiology of autism and asthma, you are not going to claim that my kids getting their shots raise your kids’ odds of getting autism or asthma. But if you don’t vaccinate your kids, even if I vaccinate mine, your decision does raise my kids’ chance of catching preventable infectious diseases. My decision to vaccinate doesn’t hurt you (and probably helps you in the ways discussed above). Your decision not to vaccinate could well hurt me.


The asymmetry of these choices is pretty unavoidable.


Here, it’s possible that a non-vaccinating parent might reply by saying that it ought to be possible for her to prioritize protecting her kids from whatever harms vaccination might bring to them without being accused of violating a social contract.


The herd immunity thing works for us because of an implicit social contract of sorts: those who are medically able to be vaccinated get vaccinated. Obviously, this is a social contract that views the potential harms of the diseases as more significant than the potential harms of vaccination. I would argue that under such a social contract, we as a society have an obligation to take care of those who end up paying a higher cost to achieve the shared benefit.


But if a significant number of people disagree, and think the potential harms of vaccination outweigh the potential harms of the diseases, shouldn’t they be able to opt out of this social contract?


The only way to do this without being a free-rider is to opt out of the herd altogether — or to ensure that your actions do not bring additional costs to the folks who are abiding by the social contract. If you’re planning on getting those diseases naturally, this would mean taking responsibility for keeping the germs contained and away from the herd (which, after all, contains members who are vulnerable owing to age, medical reasons they could not be vaccinated, or the chance of less than complete immunity from the vaccines). No work, no school, no supermarkets, no playgrounds, no municipal swimming pools, no doctor’s office waiting rooms, nothing while you might be able to transmit the germs. The whole time you’re able to transmit the germs, you need to isolate yourself from the members of society whose default assumption is vaccination. Otherwise, you endanger members of the herd who bore the costs of achieving herd immunity while reaping benefits (of generally disease-free work, school, supermarkets, playgrounds, municipal swimming pools, doctor’s office waiting rooms, and so forth, for which you opted out of paying your fair share).


Since you’ll generally be able to transmit these diseases before the first symptoms appear — even before you know for sure that you’re infected — you will not be able to take regular contact with the vaccinators for granted.


And if you’re traveling to someplace where the diseases whose vaccines you’re opting out of are endemic, you have a duty not to bring the germs back with you to the herd of vaccinators. Does this mean quarantining yourself for some minimum number of days before your return? It probably does. Would this be a terrible inconvenience for you? Probably so, but the 10-month-old who catches the measles you bring back might also be terrible inconvenienced. Or worse.

Here, I don’t think I’m alone in judging the harm of a vaccine refusenik giving an infant pertussis as worse than the harm in making a vaccine refusenik feel bad about violating a social contract.


An alternative, one which would admittedly required some serious logistical work, might be to join a geographically isolated herd of other people opting out of vaccination, and to commit to staying isolated from the vaccinated herd. Indeed, if the unvaccinated herd showed a lower incidence of asthma and autism after a few generations, perhaps the choices of the members of the non-vaccinating herd would be vindicated.


In the meantime, however, opting out of vaccines but sharing a society with those who get vaccinated is taking advantage of benefits that others have paid for and even threatening those benefits. Like it or not, that makes you a free-rider.
* * * * *
An earlier version of this essay originally appeared on my other blog.

Strategies to address questionable statistical practices.

If you have not yet read all you want to read about the wrongdoing of social psychologist Diederik Stapel, you may be interested in reading the 2012 Tilburg Report (PDF) on the matter. The full title of the English translation is “Flawed science: the fraudulent research practices of social psychologist Diederik Stapel” (in Dutch, “Falende wetenschap: De fruaduleuze onderzoekspraktijken van social-psycholoog Diederik Stapel”), and it’s 104 pages long, which might make it beach reading for the right kind of person.

If you’re not quite up to the whole report, Error Statistics Philosophy has a nice discussion of some of the highlights. In that post, D. G. Mayo writes:

The authors of the Report say they never anticipated giving a laundry list of “undesirable conduct” by which researchers can flout pretty obvious requirements for the responsible practice of science. It was an accidental byproduct of the investigation of one case (Diederik Stapel, social psychology) that they walked into a culture of “verification bias”. Maybe that’s why I find it so telling. It’s as if they could scarcely believe their ears when people they interviewed “defended the serious and less serious violations of proper scientific method with the words: that is what I have learned in practice; everyone in my research environment does the same, and so does everyone we talk to at international conferences” (Report 48). …

I would place techniques for ‘verification bias’ under the general umbrella of techniques for squelching stringent criticism and repressing severe tests. These gambits make it so easy to find apparent support for one’s pet theory or hypotheses, as to count as no evidence at all (see some from their list ). Any field that regularly proceeds this way I would call a pseudoscience, or non-science, following Popper. “Observations or experiments can be accepted as supporting a theory (or a hypothesis, or a scientific assertion) only if these observations or experiments are severe tests of the theory.”

You’d imagine this would raise the stakes pretty significantly for the researcher who could be teetering on the edge of verification bias: fall off that cliff and what you’re doing is no longer worthy of the name scientific knowledge-building.

Psychology, after all, is one of those fields given a hard time by people in “hard sciences,” which are popularly reckoned to be more objective, more revealing of actual structures and mechanisms in the world — more science-y. Fair or not, this might mean that psychologist have something to prove about their hardheadedness as researchers, about the stringency of their methods. Some peer pressure within the field to live up to such standards would obviously be a good thing — and certainly, it would be a better thing for the scientific respectability of psychology than an “everyone is doing it” excuse for less stringent methods.

Plus, isn’t psychology a field whose practitioners should have a grip on the various cognitive biases to which we humans fall prey? Shouldn’t psychologists understand better than most the wisdom of putting structures in place (whether embodied in methodology or in social interactions) to counteract those cognitive biases?

Remember that part of Stapel’s M.O. was keeping current with the social psychology literature so he could formulate hypotheses that fit very comfortably with researchers’ expectations of how the phenomena they studied behaved. Then, fabricating the expected results for his “investigations” of these hypotheses, Stapel caught peer reviewers being credulous rather than appropriately skeptical.

Short of trying to reproduce the experiments Stapel described themselves, how could peer reviewers avoid being fooled? Mayo has a suggestion:

Rather than report on believability, researchers need to report the properties of the methods they used: What was their capacity to have identified, avoided, admitted verification bias? The role of probability here would not be to quantify the degree of confidence or believability in a hypothesis, given the background theory or most intuitively plausible paradigms, but rather to check how severely probed or well-tested a hypothesis is– whether the assessment is formal, quasi-formal or informal. Was a good job done in scrutinizing flaws…or a terrible one?  Or was there just a bit of data massaging and cherry picking to support the desired conclusion? As a matter of routine, researchers should tell us.

I’m no social psychologist, but this strikes me as a good concrete step that could help peer reviewers make better evaluations — and that should help scientists who don’t want to fool themselves (let alone their scientific peers) to be clearer about what they really know and how well they really know it.

The continuum between outright fraud and “sloppy science”: inside the frauds of Diederik Stapel (part 5).

It’s time for one last look at the excellent article by Yudhijit Bhattacharjee in the New York Times Magazine (published April 26, 2013) on social psychologist and scientific fraudster Diederik Stapel. We’ve already examined strategy Stapel pursued to fabricate persuasive “results”, the particular harms Stapel’s misconduct did to the graduate students he was training, and the apprehensions of the students and colleagues who suspected fraud was afoot about the prospect of blowing the whistle on Stapel. To close, let’s look at some of the uncomfortable lessons the Stapel case has for his scientific community — and perhaps for other scientific communities as well.

Bhattacharjee writes:

At the end of November, the universities unveiled their final report at a joint news conference: Stapel had committed fraud in at least 55 of his papers, as well as in 10 Ph.D. dissertations written by his students. The students were not culpable, even though their work was now tarnished. The field of psychology was indicted, too, with a finding that Stapel’s fraud went undetected for so long because of “a general culture of careless, selective and uncritical handling of research and data.” If Stapel was solely to blame for making stuff up, the report stated, his peers, journal editors and reviewers of the field’s top journals were to blame for letting him get away with it. The committees identified several practices as “sloppy science” — misuse of statistics, ignoring of data that do not conform to a desired hypothesis and the pursuit of a compelling story no matter how scientifically unsupported it may be.

The adjective “sloppy” seems charitable. Several psychologists I spoke to admitted that each of these more common practices was as deliberate as any of Stapel’s wholesale fabrications. Each was a choice made by the scientist every time he or she came to a fork in the road of experimental research — one way pointing to the truth, however dull and unsatisfying, and the other beckoning the researcher toward a rosier and more notable result that could be patently false or only partly true. What may be most troubling about the research culture the committees describe in their report are the plentiful opportunities and incentives for fraud. “The cookie jar was on the table without a lid” is how Stapel put it to me once. Those who suspect a colleague of fraud may be inclined to keep mum because of the potential costs of whistle-blowing.

The key to why Stapel got away with his fabrications for so long lies in his keen understanding of the sociology of his field. “I didn’t do strange stuff, I never said let’s do an experiment to show that the earth is flat,” he said. “I always checked — this may be by a cunning manipulative mind — that the experiment was reasonable, that it followed from the research that had come before, that it was just this extra step that everybody was waiting for.” He always read the research literature extensively to generate his hypotheses. “So that it was believable and could be argued that this was the only logical thing you would find,” he said. “Everybody wants you to be novel and creative, but you also need to be truthful and likely. You need to be able to say that this is completely new and exciting, but it’s very likely given what we know so far.”

Fraud like Stapel’s — brazen and careless in hindsight — might represent a lesser threat to the integrity of science than the massaging of data and selective reporting of experiments. The young professor who backed the two student whistle-blowers told me that tweaking results — like stopping data collection once the results confirm a hypothesis — is a common practice. “I could certainly see that if you do it in more subtle ways, it’s more difficult to detect,” Ap Dijksterhuis, one of the Netherlands’ best known psychologists, told me. He added that the field was making a sustained effort to remedy the problems that have been brought to light by Stapel’s fraud.

(Bold emphasis added.)

If the writers of this report are correct, the field of psychology failed in multiple ways here. First, they were insufficiently skeptical — both of Stapel’s purported findings and of their own preconceptions — to nip Stapel’s fabrications in the bud. And, they were themselves routinely engaging in practices that were bound to mislead.

Maybe these practices don’t rise to the level of outright fabrication. However, neither do they rise to the level of rigorous and intellectually honest scientific methodology.

There could be a number of explanations for these questionable methodological choices.

Possibly some of the psychologists engaging in this “sloppy science” lack a good understanding of statistics or of what counts as a properly rigorous test of one’s hypothesis. Essentially, this is an explanation of faulty methodology on the basis of ignorance. However, it’s likely that this is culpable ignorance — that psychology researchers have a positive duty to learn what they ought to know about statistics and hypothesis testing, and to avail themselves of available resources to ensure that they aren’t ignorant in this particular way.

I don’t know if efforts to improve statistics education are a part of the “sustained effort to remedy the problems that have been brought to light by Stapel’s fraud,” but I think they should be.

Another explanation for the lax methodology decried by the report is alluded to in the quoted passage: perhaps psychology researchers let the strength of their own intuitions about what they were going to see in their research results drive their methodology. Perhaps they unconsciously drifted away from methodological rigor and toward cherry-picking and misuse of statistics and the like because they knew in their hearts what the “right” answer would be. Given this kind of conviction, of course they would reject methods that didn’t yield the “right” answer in favor of those that did.

Here, too, the explanation does not provide an excuse. The scientist’s brief is not to take strong intuitions as true, but to look for evidence — especially evidence that could demonstrate that the intuitions are wrong. A good scientist should be on the alert for instances where she is being fooled by her intuitions. Rigorous methodology is one of the tools at her disposal to avoid being fooled. Organized skepticism from her fellow scientists is another.

From here, the explanations drift into waters where the researchers are even more culpable for their sloppiness. If you understand how to test hypotheses properly, and if you’re alert enough to the seductive power of your intuitions, it seems like the other reason you might engage in “sloppy science” is to make your results look less ambiguous, more certain, more persuasive than they really are, either to your fellow scientists or to others (administrators evaluating your tenure or promotion case? the public?). Knowingly providing a misleading picture of how good your results are is lying. It may be a lie of a smaller magnitude than Diederik Stapel’s full-scale fabrications, but it’s still dishonest.

And of course, there are plenty of reasons scientists (like other human beings) might try to rationalize a little lie as being not that bad. Maybe you really needed more persuasive preliminary data than you got to land the grant without which you won’t be able to support graduate students. Maybe you needed to make your conclusions look stronger to satisfy the notoriously difficult peer reviewers at the journal to which you submitted your manuscript. Maybe you are on the verge of getting credit for a paradigm-shaking insight in your field (if only you can put up the empirical results to support it), or of beating a competing research group to the finish line for an important discovery (if only you can persuade your peers that the results you have establish that discovery).

But maybe all these excuses prioritize scientific scorekeeping to the detriment of scientific knowledge-building.

Science is supposed to be an activity aimed at building a reliable body of knowledge about the world. You can’t reconcile this with lying, whether to yourself or to your fellow scientists. This means that scientists who are committed to the task must refrain from the little lies, and that they must take serious conscious steps to ensure that they don’t lie to themselves. Anything else runs the risk of derailing the whole project.

C.K. Gunsalus on responsible — and prudent — whistleblowing.

In my last post, I considered why, despite good reasons to believe that social psychologist Diederik Stapel’s purported results were too good to be true, the scientific colleagues and students who were suspicious of his work were reluctant to pursue these suspicions. Questioning the integrity of a member of your professional community is hard, and blowing the whistle on misconduct and misbehavior can be downright dangerous.

In her excellent article “How to Blow the Whistle and Still Have a Career Afterwards”, C. K. Gunsalus describes some of the challenges that come from less than warm community attitudes towards members who point out wrongdoing:

[Whistleblowers pay a high price] due to our visceral cultural dislike of tattletales. While in theory we believe the wrong-doing should be reported, our feelings about practice are more ambivalent. …

Perhaps some of this ambivalence is rooted in fear of becoming oneself the target of maliciously motivated false charges filed by a disgruntled student or former colleague. While this concern is probably overblown, it seems not far from the surface in many discussions of scientific integrity. (p. 52)

I suspect that much of this is a matter of empathy — or, more precisely, of who it is within our professional community with whom we empathize. Maybe we have an easier time empathizing with the folks who seem to be trying to get along, rather than those who seem to be looking for trouble. Or maybe we have more empathy for our colleagues, with whom we share experiences and responsibilities and the expectation of longterm durable bonds, than we have for our students.

But perhaps distaste for a tattletale is more closely connected to our distaste for the labor involved in properly investigating allegations of wrongdoing and then, if wrongdoing is established, addressing it. It would certainly be easier to assume the charges are baseless, and sometimes disinclination to investigate takes the form of finding reasons not to believe the person raising the concerns.

Still, if the psychology of scientists cannot permit them to take allegations of misbehavior seriously, there is no plausible way for science to be self-correcting. Gunsalus writes:

[E]very story has at least two sides, and a problem often looks quite different when both are in hand than when only one perspective is in view. The knowledge that many charges are misplaced or result from misunderstandings reinforces ingrained hesitancies against encouraging charges without careful consideration.

On the other hand, serious problems do occur where the right and best thing for all is a thorough examination of the problem. In most instances, this examination cannot occur without someone calling the problem to attention. Early, thorough review of potential problems is in the interest of every research organization, and conduct that leads to it should be encouraged. (p. 53)

(Bold emphasis added.)

Gunsalus’s article (which you should read in full) takes account of negative attitudes towards whistleblowers despite the importance of rooting out misconduct and lays out a sensible strategy for bringing wrongdoing to light without losing your membership in your professional community. She lays out “rules for responsible whistleblowing”:

  1. Consider alternative explanations (especially that you may be wrong).
  2. In light of #1, ask questions, do not make charges.
  3. Figure out what documentation supports your concerns and where it is.
  4. Separate your personal and professional concerns.
  5. Assess your goals.
  6. Seek advice and listen to it.

and her “step-by-step procedures for responsible whistleblowing”:

  1. Review your concern with someone you trust.
  2. Listen to what that person tells you.
  3. Get a second opinion and take that seriously, too.
  4. If you decide to initiate formal proceedings, seek strength in numbers.
  5. Find the right place to file charges; study the procedures.
  6. Report your concerns.
  7. Ask questions; keep notes.
  8. Cultivate patience!

The focus is very much on moving beyond hunches to establish clear evidence — and on avoiding self-deception. The potential whistleblower must hope that those to whom he or she is bringing concerns are themselves as committed to looking at the available evidence and avoiding self-deception.

Sometimes this is the situation, as it seems to have been in the Stapel case. In other cases, though, whistleblowers have done everything Gunsalus recommends and still found themselves without the support of their community. This is not just a bad thing for the whistleblowers. It is also a bad thing for the scientific community and the reliability of the shared body of knowledge it tries to build.
_____
C. K. Gunsalus, “How to Blow the Whistle and Still Have a Career Afterwards,” Science and Engineering Ethics, 4(1) 1998, 51-64.

Reluctance to act on suspicions about fellow scientists: inside the frauds of Diederik Stapel (part 4).

It’s time for another post in which I chew on some tidbits from Yudhijit Bhattacharjee’s incredibly thought-provoking New York Times Magazine article (published April 26, 2013) on social psychologist and scientific fraudster Diederik Stapel. (You can also look at the tidbits I chewed on in part 1, part 2, and part 3.) This time I consider the question of why it was that, despite mounting clues that Stapel’s results were too good to be true, other scientists in Stapel’s orbit were reluctant to act on their suspicions that Stapel might be up to some sort of scientific misbehavior.

Let’s look at how Bhattacharjee sets the scene in the article:

[I]n the spring of 2010, a graduate student noticed anomalies in three experiments Stapel had run for him. When asked for the raw data, Stapel initially said he no longer had it. Later that year, shortly after Stapel became dean, the student mentioned his concerns to a young professor at the university gym. Each of them spoke to me but requested anonymity because they worried their careers would be damaged if they were identified.

The bold emphasis here (and in the quoted passages that follow) is mine. I find it striking that even now, when Stapel has essentially been fully discredited as a trustworthy scientist, these two members of the scientific community feel safer not being identified. It’s not entirely obvious to me if their worry is being identified as someone who was suspicious that fabrication was taking place but who said nothing to launch official inquiries — or whether they fear that being identified as someone who was suspicious of a fellow scientist could harm their standing in the scientific community.

If you dismiss that second possibility as totally implausible, read on:

The professor, who had been hired recently, began attending Stapel’s lab meetings. He was struck by how great the data looked, no matter the experiment. “I don’t know that I ever saw that a study failed, which is highly unusual,” he told me. “Even the best people, in my experience, have studies that fail constantly. Usually, half don’t work.”

The professor approached Stapel to team up on a research project, with the intent of getting a closer look at how he worked. “I wanted to kind of play around with one of these amazing data sets,” he told me. The two of them designed studies to test the premise that reminding people of the financial crisis makes them more likely to act generously.

In early February, Stapel claimed he had run the studies. “Everything worked really well,” the professor told me wryly. Stapel claimed there was a statistical relationship between awareness of the financial crisis and generosity. But when the professor looked at the data, he discovered inconsistencies confirming his suspicions that Stapel was engaging in fraud.

If one has suspicions about how reliable a fellow scientist’s results are, doing some empirical investigation seems like the right thing to do. Keeping an open mind and then examining the actual data might well show one’s suspicions to be unfounded.

Of course, that’s not what happened here. So, given a reason for doubt with stronger empirical support — not to mention the fact that scientists are trying to build a shared body of scientific knowledge (which means that unreliable papers in the literature can hurt the knowledge-building efforts of other scientists who trust that the work reported in that literature was done honestly), you would think the time was right for this professor to pass on what he had found to those at the university who could investigate further. Right?

The professor consulted a senior colleague in the United States, who told him he shouldn’t feel any obligation to report the matter.

For all the talk of science, and the scientific literature, being “self-correcting,” it’s hard to imagine the precise mechanism for such self-correction in a world where no scientist who is aware of likely scientific misconduct feels any obligation to report the matter.

But the person who alerted the young professor, along with another graduate student, refused to let it go. That spring, the other graduate student examined a number of data sets that Stapel had supplied to students and postdocs in recent years, many of which led to papers and dissertations. She found a host of anomalies, the smoking gun being a data set in which Stapel appeared to have done a copy-paste job, leaving two rows of data nearly identical to each other.

The two students decided to report the charges to the department head, Marcel Zeelenberg. But they worried that Zeelenberg, Stapel’s friend, might come to his defense. To sound him out, one of the students made up a scenario about a professor who committed academic fraud, and asked Zeelenberg what he thought about the situation, without telling him it was hypothetical. “They should hang him from the highest tree” if the allegations were true, was Zeelenberg’s response, according to the student.

Some might think these students were being excessively cautious, but the sad fact is that scientists faced with allegations of misconduct against a colleague — especially if they are brought by students — frequently side with their colleague and retaliate against those making the allegations. Students, after all, are new members of one’s professional community, so green one might not even think of them as really members. They are low status, they are learning how things work, they are judged likely to have misunderstood what they have seen. And, in contrast to one’s colleagues, students are transients. They are just passing through the training program, whereas you might hope to be with your colleagues for your whole professional life. In a case of dueling testimony, who are you more likely to believe?

Maybe the question should be whether your bias towards believing one over the other is strong enough to keep you from examining the available evidence to determine whether your trust is misplaced.

The students waited till the end of summer, when they would be at a conference with Zeelenberg in London. “We decided we should tell Marcel at the conference so that he couldn’t storm out and go to Diederik right away,” one of the students told me.

In London, the students met with Zeelenberg after dinner in the dorm where they were staying. As the night wore on, his initial skepticism turned into shock. It was nearly 3 when Zeelenberg finished his last beer and walked back to his room in a daze. In Tilburg that weekend, he confronted Stapel.

It might not be universally true, but at least some of the people who will lie about their scientific findings in a journal article will lie right to your face about whether they obtained those findings honestly. Yet lots of us think we can tell — at least with the people we know — whether they are being honest with us. This hunch can be just as wrong as the wrongest scientific hunch waiting for us to accumulate empirical evidence against it.

The students seeking Zeelenberg’s help in investigating Stapel’s misbehavior found a situation in which Zeelenberg would have to look at the empirical evidence first before he looked his colleague in the eye and asked him whether he was fabricating his results. They had already gotten him to say, at least in the abstract, that the kind of behavior they had reason to believe Stapel was committing was unacceptable in their scientific community. To make a conscious decision to ignore the empirical evidence would have meant Zeelenberg would have to see himself as displaying a kind of intellectual dishonesty — because if fabrication is harmful to science, it is harmful to science no matter who perpetrates it.

As it was, Zeelenberg likely had to make the painful concession that he had misjudged his colleague’s character and trustworthiness. But having wrong hunches is science is much less of a crime than clinging to those hunches in the face of mounting evidence against them.

Doing good science requires a delicate balance of trust and accountability. Scientists’ default position is to trust that other scientists are making honest efforts to build reliable scientific knowledge about the world, using empirical evidence and methods of inference that they display for the inspection (and critique) of their colleagues. Not to hold this default position means you have to build all your knowledge of the world yourself (which makes achieving anything like objective knowledge really hard). However, this trust is not unconditional, which is where the accountability comes is. Scientists recognize that they need to be transparent about what they did to build the knowledge — to be accountable when other scientists ask questions or disagree about conclusions — else that trust evaporates. When the evidence warrants it, distrusting a fellow scientist is not mean or uncollegial — it’s your duty. We need the help of other to build scientific knowledge, but if they insist that they ignore evidence of their scientific misbehavior, they’re not actually helping.

Scientific training and the Kobayashi Maru: inside the frauds of Diederik Stapel (part 3).

This post continues my discussion of issues raised in the article by Yudhijit Bhattacharjee in the New York Times Magazine (published April 26, 2013) on social psychologist and scientific fraudster Diederik Stapel. Part 1 looked at how expecting to find a particular kind of order in the universe may leave a scientific community more vulnerable to a fraudster claiming to have found results that display just that kind of order. Part 2 looked at some of the ways Stapel’s conduct did harm to the students he was supposed to be training to be scientists. Here, I want to point out another way that Stapel failed his students — ironically, by shielding them from failure.

Bhattacharjee writes:

[I]n the spring of 2010, a graduate student noticed anomalies in three experiments Stapel had run for him. When asked for the raw data, Stapel initially said he no longer had it. Later that year, shortly after Stapel became dean, the student mentioned his concerns to a young professor at the university gym. Each of them spoke to me but requested anonymity because they worried their careers would be damaged if they were identified.

The professor, who had been hired recently, began attending Stapel’s lab meetings. He was struck by how great the data looked, no matter the experiment. “I don’t know that I ever saw that a study failed, which is highly unusual,” he told me. “Even the best people, in my experience, have studies that fail constantly. Usually, half don’t work.”

In the next post, we’ll look at how this other professor’s curiosity about Stapel’s too-good-to-be-true results led to the unraveling of Stapel’s fraud. But I think it’s worth pausing here to say a bit more on how very odd a training environment Stapel’s research group provided for his students.

None of his studies failed. Since, as we saw in the last post, Stapel was also conducting (or, more accurately, claiming to conduct) his students’ studies, that means none of his students’ studies failed.

This is pretty much the opposite of every graduate student experience in an empirical field that I have heard described. Most studies fail. Getting to a 50% success rate with your empirical studies is a significant achievement.

Graduate students who are also Trekkies usually come to recognize that the travails of empirical studies are like a version of the Kobayashi Maru.

Introduced in Star Trek II: The Wrath of Khan, the Kobayashi Maru is a training simulation in which Star Fleet cadets are presented with a civilian ship in distress. Saving the civilians requires the cadet to violate treaty by entering the Neutral Zone (and in the simulation, this choice results in a Klingon attack and the boarding of the cadet’s ship). Honoring the treaty, on the other hand, means abandoning the civilians and their disabled ship in the Neutral Zone. The Kobayashi Maru is designed as a “no-win” scenario. The intent of the test is to discover how trainees face such a situation. Owing to James T. Kirk’s performance on the test, Wikipedia notes that some Trekkies also view the Kobayashi Maru as a problem whose solution depends on redefining the problem.

Scientific knowledge-building turns out to be packed with particular plans that cannot succeed at yielding the particular pieces of knowledge the scientists hope to discover. This is because scientists are formulating plans on the basis of what is already known to try to reveal what isn’t yet known — so knowing where to look, or what tools to use to do the looking, or what other features of the world are there to confound your ability to get clear information with those tools, is pretty hard.

Failed attempts happen. If they’re the sort of thing that will crush your spirit and leave you unable to shake it off and try it again, or to come up with a new strategy to try, then the life of a scientist will be a pretty hard life for you.

Grown-up scientists have studies fail all the time. Graduate students training to be scientists do, too. But graduate students also have mentors who are supposed to help them bounce back from failure — to figure out the most likely sources of failure, whether it’s worth trying the study again, whether a new approach would be better, whether some crucial piece of knowledge has been learned despite the failure of what was planned. Mentors give scientific trainees a set of strategies for responding to particular failures, and they also give reassurance that even good scientists fail.

Scientific knowledge is built by actual humans who don’t have perfect foresight about the features of the world as yet undiscovered, humans who don’t have perfectly precise instruments (or hands and eyes using those instruments), humans who sometimes mess up in executing their protocols. Yet the knowledge is built, and it frequently works pretty well.

In the context of scientific training, it strikes me as malpractice to send new scientists out into the world with the expectation that all of their studies should work, and without any experience grappling with studies that don’t work. Shielding his students from their Kobayashi Maru is just one more way Diederik Stapel cheated them out of a good scientific training.