Another "Google AI" post.
Another "Google AI" post.
Another "Google AI" post.
A link to the mentioned BMJ publication.
https://www.bmj.com/content/363/bmj.k5094
That's why you should always have your backpack when you board a plane (along with your towel, of course).
Thanks! The OP is a mess (at least when using Boost).
At first I was at least impressed that it came up with such a hilarious idea, but then it of course turns out (just as with the pizza glue) it just stole it from somewhere else.
"AI" as we know them are fundamentally incapable of coming up with something new, everything it spits out is a combination of someone else's work.
The big problem with these AI is not that the technology is fundamentally flawed. It's that stupid humans just dump huge amounts of data at it without checking any of it for validity.
If you train a human on bullshit of course they're not going to know the difference between truth and lies, how could they possibly do that if all they're ever told is utter nonsense. I really would have thought better of people at Google, you'd have thought someone there would have had the intelligence to realize that not everything on the internet is solid gold.
This is why I've never really bothered that they were training on data from Reddit.
I thought this was fake, but I searched Google for "parachute effectiveness" and that satirical study is at the top, and literally every single link below it is a reference to that satirical study. I have scrolled for a good minute and found no genuine answer...
I have scrolled for a good minute and found no genuine answer...
Presumably because the effectiveness of parachutes is pretty self-evident and no one has done a formal study on the subject because why would they.
It's not like they're going to find that parachutes actually aren't needed and the humans just float gently to the ground on their own. Or that a sufficiently tall top hat can be just as effective.
Parachute effectiveness is a very reasonable thing to study, it's pretty important to know how one parachute design performs compared to other designs and the obvious baseline is no parachute. A lot of things which appear to be self-evident have been extensively studied, generally you don't want to just assume you know how something works.
Though throwing people out of a plane at altitude with no parachute probably isn't the most ethical way to study parachute effectiveness.
This is a good point I didn't think of. It's possible that the answer is so obvious, that the only articles made about it would be jokes.
Why waste money on parachutes when everyone has a backpack at home?!
&udm=14 🙏
Better yet, don't use Google for searching when possible.
Want to find stuff on the internet? Don't! Follow for more tips like these!
I just searched "do maple trees have a tap root", ai overview says yes. Literally all other reputable sources say no 🙄
Representative study participant jumping from aircraft with an empty backpack. This individual did not incur death or major injury upon impact with the ground
Hah! I just told someone the other day that LLMs trick people into seeing intelligence basically by cold reading! At last, sweet validation!
At least in the actual Gemini chat when asked about parachute study effectiveness correctly notes this study as satire.
Weird that it actually made note of that but didn't put it in the summary
It's because it has no idea what is important or isn't important. It's all just information to it and it all carries the same weight, whether it's "parachutes aren't any more effective than empty backpacks" or "this study is a satire making fun of other studies that extrapolate information carefully". Even though that second bit essentially negates the first bit.
Bet the authors weren't expecting their joke study to hit a second time like this, demonstrating that AI is just as bad at extrapolating since it extrapolated true information from false.
It's reckless to use these AIs in searches. If someone jokes about pretending to be a doctor and suggesting a stick of butter being the best treatment for a heart attack and that joke makes it into the training set, how would an AI have any idea that it doesn't have the same value as some doctors talking about patterns they've noticed in heart attack patients?
I can't believe this got released and this is still happening.
This is the revolution in search of RAG (retrieval augmented generation), where the top N results of a search get fed into an LLM and reprioritized.
The LLM doesn't understand it want to understand the current, it loads the content info it's content window and then queries the context window.
It's sensitive to initial prompts, base model biases, and whatever is in the article.
I'm surprised we don't see more prompt injection attacks on this feature.