Wednesday, October 16, 2024
25 C
Brunei Town

Latest

Science reveals why being ‘always right’ is often wrong

UPI – Attention all “know-it-alls”.

Folks who are sure they’re right often believe they’ve got enough information to make up their minds, even if in reality they only have part of the picture, a new study finds.

It’s a concept called the “illusion of information adequacy,” and it helps explain how people can have such strong and cemented opinions even though they get their news from limited and biased sources, researchers said.

“We found that, in general, people don’t stop to think whether there might be more information that would help them make a more informed decision,” said researcher Angus Fletcher, a professor of English at Ohio State University.

“If you give people a few pieces of information that seems to line up, most will say ‘that sounds about right’ and go with that,” Fletcher said in an Ohio State news release.

For the study, researchers recruited nearly 1,300 Americans for an online experiment.

All of the people read an article about a fictional school that lacked adequate water.

But one group were provided an article containing only reasons why the school should merge with another that had adequate water.

PHOTO: ENVATO

A second group’s article only gave reasons for staying separate and hoping for other solutions.

Only a third group, which served as the control group, had an article that provided all the arguments for either merging or staying separate.

The two groups that read only half the story still believed they had enough information to make a solid decision, and said they would follow the recommendations in the article provided to them, results show.

“Those with only half the information were actually more confident in their decision to merge or remain separate than those who had the complete story,” Fletcher said.

“They were quite sure that their decision was the right one, even though they didn’t have all the information.”

Folks with half the information also thought most other people would come to the same decision as them, researchers added.

However, the experiment offered one ray of hope.

People who had read only one side of the story were later provided arguments for the other side, and many were willing to change their minds once they had all the facts, researchers said.

The new study was published in the journal PLOS One.

spot_img

Related News

spot_img