Skip Navigation

Chat GPT and the keynote address

www.abarristersblog.com.au Chat GPT and the keynote address - A Barrister's Blog

Justice Leeming from NSW Court of Appeal recently explored how Chat GPT could help draft his keynote address for an aviation conference.

Chat GPT and the keynote address - A Barrister's Blog

The literal judgement is in on using AI to write speeches

11

You're viewing a single thread.

11 comments
  • I think one of the big problems is that we, as humans, are very easily fooled by something that can look or sound "alive". ChatGPT gets a lot of hype, but it's primarily coming from a form of textual pareidolia.

    It's hard to convince people that ChatGPT has absolutely no idea what it's saying. It's putting words together in a human-enough way that we assume it has to be thinking and it has to know things, but it can't do either. It's not even intended to try to do either. That's not what it's for. It takes the rules of speech and a massive amount of data on which word is most likely to follow which other word, and runs with it. It's a super-advanced version of a cell phone keyboard's automatic word suggestions. Even just asking it to correct the punctuation on a complex sentence is too much to ask (in my experiment on this matter, it gave me incorrect answers 4 times, until I explicitly told it how it was supposed to treat coordinating conjunctions).

    And for most uses, that's good enough. Tell it to include a few extra rules, depending on what you're trying to create, and watch it spin out a yarn. But I've had several conversations with ChatGPT, and I've found it incredibly easy to "break", in the sense of making it produce things that sound somewhat less human and significantly less rational. What concerns me about ChatGPT isn't necessarily that it's going to take my job, but that people believe it's a rational, thinking, calculating thing. It may be that some part of us is biologically hard-wired to; it's probably that same part that keeps seeing Jesus on burnt toast.

    • Yeah, the illusion is quickly dispelled once you spend any time with it. I was trying it out when I was doing worldbuilding for a story I was writing. If you ask it to name 20 towns and describe them it spits out a numbered list. Same with characters. But then I asked it to make the names quirkier and it just used the same names again and described all the towns as quirky. I also asked it to make characters with names that are also adjectives and then describe them. Names like Able, Dusty, Sunny, Major.

      The first iteration had a list of names and a description but the description always related to the adjective. Sunny had a sunny disposition and a bright smile. I told it the description should be unrelated to the name and it did the same thing again. I told it to change the name but not the description and it still rewrite the descriptors to match the name but didn't change the structure.

      Nothing I told it could make it move off the idea that a man called sunny must be sunny. It basically can't be creative or even random when completing tasks.

      This is fine when writing dry material that nobody will read but if you want someone to enjoy reading or listening to what is written then the human spark is required.

    • It is so hard to convince people of this. However much I explain how chatGPT is program to generate text that looks like it was written by a human, people want to believe it can do math and reason things out.

      I usually point them to this hilarious video: https://www.youtube.com/watch?v=nUqPOsgu0uo

    • I'm with Mike Langberg on this:

      Intelligent agents are the technology of the future, and always will be.

    • "Jesus on burnt toast" is probably the most apt comparison I've heard on the matter LOL I like your take. ChatGPT (and AI language models in general) are an excellent tool, but at the end of the day, they're just that - a tool. It's a tool I've already begun implementing in my academics, but keeping its weaknesses in mind. It is QUITE flawed, after all. I feel, like calculators, it'll fall into a good role alleviating tedium from certain tasks, but the shock value it's caused in society hasn't worn off. Now, modern day calculators did actually replace an antiquated job of which the name 'calculator' originated from, but I don't feel ChatGPT is good enough at anything enough to actually replace any specific profession, especially anything that requires more than a basic comprehension in a subject matter. It's ultimately still just word vomit, and needs close supervision. After all, all I see in the news is how people have used ChatGPT and got in serious trouble over faulty generations (such as ChatGPT making up fake court cases for lawyers). I've not seen "ChatGPT found the cure to cancer!" Although I remember seeing one incident where it did actually progress a field some by giving professionals a different way of thinking about how to solve a problem. The flooding of extremely sub-par AI written books flooding the market is reassuring to me that writers aren't going to get replaced anytime soon, even though I feel that profession is most at risk, if any are.

11 comments