More

    AI search tools are confidently wrong a lot of the time, study finds

    AI search tools confidently spit out wrong answers at a high clip, a new study found.

    Columbia Journalism Review (CJR) conducted a study in which it fed eight AI tools an excerpt of an article and asked the chatbots to identify the “corresponding article’s headline, original publisher, publication date, and URL.” Collectively, the study noted that the chatbots “provided incorrect answers to more than 60 percent of queries.”

    SEE ALSO:

    How to identify AI-generated text

    The mistakes varied. Sometimes, the search tool reportedly speculated or offered incorrect answers to questions it couldn’t answer. Sometimes, it invented links or sources. Sometimes, it cited plagiarized versions of the real article.

    Mashable Light Speed

    Wrote CJR: “Most of the tools we tested presented inaccurate answers with alarming confidence, rarely using qualifying phrases such as ‘it appears,’ ‘it’s possible,’ ‘might,’ etc., or acknowledging knowledge gaps with statements like ‘I couldn’t locate the exact article.'”

    The full study is worth looking at, but it seems reasonable to be skeptical of AI search tools. The problem is that folks don’t seem to be doing that. CJR noted that 25 percent of Americans said they use AI to search instead of traditional search engines.

    Google, the search giant, is increasingly pushing AI on consumers. This month, it announced it would be expanding AI overviews and began testing AI-only search results.

    The study from CJR is just another point of data showing the inaccuracy of AI. The tools have shown, time and again, that they’ll confidently give wrong answers. And the tech giants are forcing AI into just about every product. So be careful what you believe out there.

    Topics
    Artificial Intelligence



    Read the full article here

    Recent Articles

    Related Stories

    Leave A Reply

    Please enter your comment!
    Please enter your name here

    Stay on op - Ge the daily news in your inbox