Consolidated AI Thread: A Discussion For Everything AI

This reminds me when my school teachers would rage against Wikipedia and tell us not to use it at all.

The issue however was never Wikipedia. It was how my classmates used Wikipedia. We just copied and pasted whatever we saw there and didn’t check whether the information was correct.

Too often many users misuse AI. They make vague prompts, no follow ups and copy and paste whatever ChatGPT spits out. And they don’t check the sources. Lawyers have been caught being utterly lazy when they submitted court papers that cited non existent cases.

And we must be honest with ourselves. Humans use shortcuts, have biases and make mistakes too that leads us astray when doing work, this isn’t unique to AI. But humans have systems to check each others work. We cite people in the field, we do peer review, we have other humans contribute to our papers and they do checks too.

University know that students can be lazy even before AI. They spend money to have software that checks for plagiarism

If we can recognise that humans can be faulty and that we need systems in place to make sure our output is accurate then why expect AI to be error free 100% of the time.

It’s like citing an article in some journal and never being critical about the source you’re using. Like are you checking whether you’re presenting the arguments made in that paper correctly, what’s the paper’s methodology, have others responded to this paper and given their own perspective or criticism.

I think people expect so much from AI. The companies who made them are hyping them too much to the point that people are forgetting their still tools. And tools have their purpose but also weaknesses that one must be aware of. You don’t use a hammer to cut wood.

So if human output also needs to be verified then so too must we expect AI output to be verified.

My philosophy professor once complained that when he was writing his book, he was shocked to find that many academic papers he looked up would cite a source incorrectly or misrepresent/misquote stuff in the cited source.

He found it frustrating. Even in peer reviewed journals mistakes happen.

Just because something is written by a human it doesn’t mean we should trust it without checking it.

For me, when I ask AI to give me books to read on a certain topic, my follow up is always “check online and link the books and articles you’ve recommended”

This always gives me real books and articles I can use as a starting point in some topic/subject.

And speaking about hallucinations and the accuracy of AI, I remember how an AI model (Google Notebook) made a mistake when it summarized Madison v Marbury.

It stated that the Supreme Court found that Madison had acted illegally by denying to give Marbury his commission. This was because the Supreme Court found the Judicial Act of 1789 did not empower the Supreme Court to force Madison to give Marbury the commission.

To me, this summary looked accurate. Then I checked it online and found the AI made a subtle mistake. It was a small mistake that one could miss. The Supreme Court did find Madison acted illegally. But the Judicial Act did empower the Supreme Court to force Madison to give Marbury his commission however the Supreme Court found that parts of the Judicial Act that gave it that power to force Madison to send the commission was unconstitutional because it gave the Court powers beyond what the Constitution had stated.

It’s such a small mistake. And I think that’s why most hallucination are hard to find. AI is really good at finding stuff and writing some paper.

But the problem is that AI can make a small mistake that sounds reasonable and most people don’t double check because AI writing for the most part does pass the sniff test.

It can be 98% correct but the 2% is critically wrong. Terrence Tau made this observation about AI writing proofs. AI can write convincing proofs but then make a small mistake that a human might not have ever done.

This is why I think its important for people to check AI even if what it writes sounds reasonable at first glance. The same goes with human writing/output too.

1 Like

Because that’s what people do. A not insignificant amount of people assume that what a computer says is always true. It’s chilling when it happens in a bank or somesuch.

1 Like

Yeah too many people don’t want to check it because it takes effort and time. And the paper the AI has written usually “feels” right :grinning_face:.

1 Like

That’s a really complicated topic. I approach LLM tools from an academic perspective. In that context, they have a somewhat limited use case- hallucinations become more common the further you deviate from “normal” questions.

I asked about banking in the middle ages, and the results were deeply faulted and referred to non-existent sources that were impossible to verify.

That’s not a good use of my time. I outlined a use for them as a research tool above, where you need an answer to serve as a starting point. However, I wouldn’t trust anything an LLM produces to be factual because it’s not for that.

It’s an excellent tool for entertainment, and can serve to answer common knowledge questions about non-critical topics with reasonable accuracy.

What’s the use case I’m ignoring?

2 Likes

Because the AI industry has put a massive amount of money and time into convincing people that AI is perfect and can do everything for you. This isn’t just an issue with people, it’s an issue with how the technology is presented, propagated by people who stand to make a lot of money off of lying about its efficacy.

6 Likes

Anything non factual, and idea generation, translation, coding, research assistant, I’ve also seen a lot of usage for emotional support and therapy,

of course a toaster can’t help with math problems, but good thing that it’s not a toaster, it already has math solving capabilities far exceeding the average person.

The research assistant angle is questionable, imo, for reasons I’ve already outlined but- yeah for personal use LLM can certain generate ideas or carry them forward if prompted correctly.

I think their use as a therapeutic tool needs more study before being implemented. It seems risky with how easily humans anthropomorphize things.

I mean I don’t think it’s illegal to view the LLM as a human like entity, given the philosophical questions involved.

There’s some researchers already using it this way, even in the medical field, as a tool to help during research. Think he used o3 model though so your mileage may vary.

So far the impression I get is pretty positive, but it’s a huge use case for a lot of users, and since they keep using still, i can only say it’s a popular feature.

I mean I don’t think it’s illegal to view the LLM as a human like entity, given the philosophical questions involved.

Could you expand on this? I want to be positive I understand your exact meaning.