NEW YORK – A large group of researchers set out to repeat 100 experiments published by leading psychology journals to see how often they would get the same results.

The answer: Less than half the time.

That doesn’t mean all those unconfirmed studies were wrong. But it’s a stark reminder that a single study rarely provides definitive answers and why scientists often greet new findings by saying, "More research is needed."

"Any one study is not going to be the last word," said Brian Nosek, a psychology professor at the University of Virginia.

"Each individual study has some evidence. It contributes some information toward a conclusion. But the real conclusion, when you can say confidently that something is true or false, is based on an accumulation of evidence over many studies," said Nosek, who led the project.

And yes, he said at a news conference, "even this project itself is not â ¦ a definitive word about reproducibility."

The work was carried out by an international team of more than 300 people and appeared Thursday in the journal Science. The project focused on psychology because its organizers came from that field. Researchers worked with the authors of the original studies in setting up the replication attempts.

Only about 40 percent of those attempts produced the original results.

The effort focused on 100 experiments reported during 2008 in any of three major psychology journals: Psychological Science, the Journal of Personality and Social Psychology and the Journal of Experimental Psychology: Learning, Memory, and Cognition.

None of these experiments tested any treatments. They focused on basic research into how people think, remember, perceive their world and interact with others; for example, one explored why people are reluctant to tempt fate.

Studies with stronger statistical evidence for their conclusions were more likely to be replicated than others, as were those with findings that were judged to be less surprising.

When a study’s results were not replicated, there could be several explanations, Nosek said. The original study could be wrong. Or it could be right, and the repeat study overlooked a real effect just by chance. Or both studies could be correct, with conflicting conclusions because of differences in how they were carried out.

Project workers tried to minimize such differences, but matching an original study could be tricky. E.J. Masicampo of Wake Forest University in North Carolina, a co-author of the new study, said one of his own experiments was not confirmed by the project.

His original study required participants to make decisions that required significant mental effort. To create that situation, researchers asked undergrads to choose between off-campus apartments. But that task wasn’t as mentally challenging when it was tried again at a different campus, which evidently threw off the results of the experiment, he said.