The Gino data scandal in behavioural science and research misconduct

Allegations of fraud hit the behavioural sciences recently when a team of independent investigators published a series of articles detailing apparent data manipulation in more than four prominent papers in the field. Ironically, the papers described studies of morality and honesty, and so far, the accusations have landed at the feet of one author common to all these papers, Harvard University professor Francesca Gino.

Since the allegations were levelled, the papers have been retracted, but not without disagreement and controversy. While the university floated its own investigation into the claims before it placed Dr. Gino on administrative leave, she filed cases against the university and the authors of the original articles – researchers Leif Nelson, Joe Simmons, and Uri Simonsohn. Since then, with help from its peers, the trio has crowd-funded money to pay for its legal defence.

The rise of this scandal has spawned many questions – from the simpler one of Dr. Gino’s guilt to the more involved one of where it will leave the field of behavioural sciences itself. But underlying them all is an older, more familiar one: why does misconduct happen?

What are the effects of misconduct?

Outright fabrication, falsification, and plagiarism, plus some of their more benign variations constitute a tale almost as old as scientific inquiry itself. Beginning with the Piltdown Man in 1912 – a fraudulent attempt to fill in the missing link between primate and man – to more recent cases like that of Diederik Stapel, scientific misconduct has always been and continues to be around, to different degrees in different fields.

Even if one instance of misconduct is small in scope, it can have dire consequences for scientists and for the field – especially if those committing it are the field’s leaders. One way to identify leaders is by the extent to which their work has laid the foundation for that of others; this is considerable in Dr. Gino’s case.

Many other papers and findings free of misconduct that rely on the original but faulty work will also be brought into question, risking years of work.

Why do researchers commit misconduct?

There is some consensus that the leading contributors to misconduct today are the existing incentive structures for researchers and shortcomings in peer-reviews and replication studies.

Researchers have many incentives – including from grant-providers, editors, and academic institutions – to pursue more groundbreaking findings and results that support alternative hypotheses. Flashier results can elevate the researchers who obtain them to higher standing, make them and their employers more famous, and allow their funders to claim sufficient bang for the buck. But on the flip side, the size of the incentives may have encouraged many researchers to do work that is sloppy at best and outright manufactured at worst.

Some experts have backed the idea that incentive structures, manifesting as pressure to publish, affect researcher’s motivations. They also blame the low risk of detection by reviewers and research supervisors’ mentoring styles as probable motivators of misconduct. Some others have blamed cultural norms around criticism and the absence, or incompleteness, of policies at the national or at the institutional levels to penalise misconduct.

How should misconduct be dealt with?

One novel response to the challenges of dealing with misconduct is the Open Science Framework (OSF) to ensure scientific integrity. It promotes practices such as pre-registration (i.e. fixing a study’s hypotheses, methods, and analyses before it is conducted and agreeing to share the results, whatever they are) and making research data more accessible.

As such, OSF has tried to reduce the amount of misconduct by putting both researchers’ original intentions and the eventual data up for scrutiny. The team behind OSF has also launched the more ambitious ‘Systematizing Confidence in Open Research and Evidence’ (SCORE) project, which tries to make research more credible by developing automated tools to generate “rapid, scalable, and accurate confidence scores for research claims”.

This said, OSF still requires institutions and/or researchers to buy into abiding by it to be able to effectively eliminate misconduct. SCORE can work around this barrier but has its own drawbacks, such as a risk of uncritical use en masse to assess the ‘credibility’ of scientists – something that those developing SCORE have said isn’t its use case.

In addition, while there are methods at both small and large scales to handle fraud, they can be inconsistent across institutions. The result is for researchers who are willing to cooperate to still face significant ‘unofficial’ forms of punishment – or, as with the three researchers who reported concerns with the papers co-authored by Dr. Gino, for independent investigators to be at risk of facing expensive litigation.

What are the systematic causes of misconduct?

Less-novel ways to combat the incidence of misconduct include adequate funding and less pressure on researchers, support for replication studies (i.e. studies that check the results of other studies), and ‘detectives’ incentivised to check for fraud.

For example, setting aside a part of a grant sanctioned for a study for quality-control activities would go a long way to counter misconduct. Investigators could use these resources to make probes more thorough and also faster, which could help increase younger scientists’ confidence in the system. Similarly, providing financial aid for replication studies – such as in the form of cash rewards – could also help.

The ability of science to keep out misconduct and police itself partly comes down to the choices individual researchers make. Whether it’s the temptation to be a bit less rigorous when double-checking a result or the values they impart to one’s mentees, the willingness to stick to scientific norms regardless of the impact it has on one’s prospects ultimately decides how far misconduct spreads.

What is the role of scientific publishing?

This said, beyond research facilities and academia, the structure of scientific publishing is also implicated in the persistence of research misconduct. In particular, many journals – like grantors – prefer to publish sensational results and have been less than forthcoming to investigate or rectify signs of misconduct in published papers.

Recently, for example, Nature retracted a paper it had published last year after independent researchers reported that its data didn’t add up. But the journal hasn’t explained how it cleared the paper for publication in the first place.

What can, and must, scientists do?

Some scientists are doing the right thing. In the absence of similar institutional efforts, many of Dr. Gino’s co-authors have decided to examine work on which Dr. Gino had collaborated and provided the data, in order to separate ‘good’ papers from ‘bad’ instead of allowing all of them to be tarred with the same brush.

This said, scientists are aware of a much-needed rethink, especially by those who have power, regarding the methods and norms around science. The popular imagination of science is that it will always be rigorous and self-correcting, but this is naïve and unrealistic.

The contemporary scientific process needs to be enhanced with technology and incentives to make inquiries about scientific inquiry itself – and they should become standard practice, rather than requiring ‘special’ circumstances to kick in.

Abhishek V. is Research Assistant at the Department of Economics at Monk Prayogshala, Mumbai.

Leave a Reply

Your email address will not be published. Required fields are marked *

Back to top button