On Exceptions
pixxelkick @ pixxelkick @lemmy.world Posts 4Comments 888Joined 2 yr. ago
Personally I don’t really trust the LLMs to synthesize disparate sources.
The #1 best use case for LLMs is using them as extremely powerful fuzzy searchers on very large datasets, so stuff like hunting down published papers on topics.
Dont actually use their output as the basis for reasoning, but use it to find the original articles.
For example, as a software dev, I use them often to search for the specific documentation for what I need. I then go look at the actual documentation, but the LLM is exceptionally fast at locating the document itself for me.
Basically, using them as a powerful resource to look up and find resources is key, and was why I was able to find documentation on the symptoms of my pet so fast. It would have taken me ages to find those esoteric published papers on my own, there's so much to sift through, especially when many papers cover huge amounts of info and what Im looking for is one small piece of info in that one paper.
But with an LLM I can trim down the search space instantly to a way way smaller set, and then go through that by hand. Thousands of papers turn into a couple in a matter of seconds.
The power server side for 5 minutes of chatgpt, vs the power burned browsing the internet to find the info on my own (which would take hours to manually sift through)
Thats the comparison.
Even though server side power consumption to run GPT is very high, its not so high that its more than hours and hours of a laptop usage
Not at tremendously less of a power cost anyways. My laptop draws 35W
5 minutes of GPT is genuinely less power consumption than several hours of my laptop being actively used to do the searching manually. Laptops burn non trivial amounts of power when in use. Anyone who has held a laptop on their lap can attest to the fact they aren't exactly running cold.
Hell even a whole day of using your mobile phone is non trivial in power consumption, they also use 8~10W or so.
Using GPT for dumb shit is arguably unethical, but only in the sense that baking cookies in the oven is. You gonna go and start yelling at people for making cookies? Cooking up one batch of cookies burns WAAAY more energy than fucking around with GPT. And yet I don't see people going around bashing people for using their ovens to cook things as a hobby.
There's no good argument against what I did, by all metrics it genuinely was the ethical choice.
...no that's not the summarization.
The summarization is:
if you reinforce your model via user feedback, via "likes" or "dislikes" or etc, such that you condition the model towards getting positive user feedback, it will start to lean towards just telling users whatever they want to hear in order to get those precious likes, cuz obviously you trained it to do that
They demo'd in the same paper other examples.
Basically, if you train it on likes, the model becomes duper sycophantic, laying it on super thick...
Which should sound familiar to you.
AI saved my pets life. You won't convince me it's 100% all bad and there's no "right" way to use it.
The way it is trained isnt intellectual theft imo.
It only becomes intellectual theft if it is used to generate something that then competes with and takes away profits from the original creators.
Thus the intellectual theft only kicks in at generation time, but the onus is still on the AI owners for not preventing it
However if I use AI to generate anything that doesn't "compete" with anyone, then "intellectual theft" doesn't matter.
For example, when I used it to assist with diagnosing a serious issue my pet was having 2 months ago that was stumping even our vet and it got the answer right, which surprised our vet when we asked them to check a very esoteric possibility (which they dubious checked and then they were shocked to find something there.
They asked us how on earth we managed to guess to check that place of all things, how could we have known. As a result we caught the issue very early when it was easy to treat and saved our pets life
It was a gallbladder infection, and her symptoms had like 20 other more likely causes individually.
But when I punched all her symptoms into GPT, everytime, it asserted if was likely the gallbladder. It had found some papers on other animals and mammals and how gallbladder infections cause that specific combo of symptoms rarely, and encouraged us to check it out.
If you think "intellectual theft" still applies here, despite it being used to save an animals life, then you are the asshole. No one "lost" profit or business to this, no one's intellectual property was infringed, and consuming the same amount of power it takes to cook 1 pizza in my oven to save my pets life is a pretty damn good trade, in my opinion.
So, yes. I think I used AI ethically there. Fight me.
Exceedingly false representation of the actual experiment.
They took Llama 3 and then trained it further on specific conditions (reinforcing it on "likes" / "thumbs up"s based on positive feedback from a simulated userbase)
And then after that the scientists found the new model (which you can't really call Llama 3 anymore, it's been trained further and it's behavior fundamentally altered) behaved like this when prior informed that the user was easily influenced by the model specifically
What is important to gather though, is the fact that when a model gets trained on the metrics of "likes", it starts to behave in a manner like this, telling the user whatever they want to hear... Which makes sense, the model is effectively getting trained to min/max positive feedback from users, rather than being trained on being right / correct
But to try and represent this as a "real" chatbot's behavior is definitely false, this was a model trained by scientists explicitly to test if this behavior happens under extreme conditioning.
You need to use a model trained for inpainting.
I would say your best bet is typically using inpainting.
Start with just getting all the figures framed/positioned well, such that they look about right in terms of stance and position in the frame.
At this stage only use descriptors that describe all 3 of them well.
Then inpaint only one of them, and add in all the descriptors just for that figure, and regenerate just that section with a bit higher noise level until she looks right.
Repeat for each one individually.
Eh, Im not conservative but I do think this is a valid bipartisan criticism.
If they pulled this stunt with a liberal leaning byelection Id consider it a greasy trick.
I have no issue with putting a cap on ballot names via some kind of reasonable system (IE if you get more than x candidates, then the top y candidates can only go on the ballot based on who has the most vouches or whatever)
That way you can avoid dozens and dozens of randos flooding the ballot to confuse people. It's a valid hole to patch in our system that both sides should be on board with fixing in a reasonable way.
Permanently Deleted
I was gonna ask how this thing would even have access to execute a command like this
But then I realized we are talking about a place that uses a tool like this in the first place so, yeah, makes sense I guess
If they are mandated, that's just as bad I agree.
At my company we have tonnes of in house Lunch and Learns (on paid time, non mandatory) that are effectively "I found this super useful thing and want others to know about it"
And I'll join these things, and see (person), who is on my team, in it too. Later I'll hat with them about it, or at least try, and they'll have zero clue wtf I'm talking about.
And it becomes obvious they just joined the meeting to give the illusion of caring, they prolly were afk the whole time. And I suspect this cuz they often do the same for our "in team" mandatory important meetings discussing critical stuff on the project.
Sorta just sounds like you can probably fire a few employees who don't give a fuck.
From experience, a lot of companies tend to be propped up by like 10% of their developers doing 90% of the work, maybe 50% of developers doing the last 10%, and then like 40% of developers either doing fuck all or actively harming the codebases and making more work for the other 60%.
And more often than not, these people are the ones sending stuff like "AI Note Takers" merely to give the illusion of existing.
In reality you have like three devs who actually care and do most of the work having a discussion and like 10 to 30 AFK participants with their cameras off who probably arent even listening or care.
And the thing is, it shows later. The same devs have zero clue wtf is going on, they have zero clue how to do their job, and they constantly get their code sent back or they plague up some codebase that doesnt have reviewers to catch them.
The AI note takers are just the new version of people showing up to meetings with their camera off and never speaking a word.
Except now they burn orders of magitude more power to do it, which is awful.
That's more like it, thank you!
Source? This is just some random picture, I'd prefer if stuff like this gets posted and shared with actual proof backing it up.
While this might be true, we should hold ourselves to a standard better than just upvoting what appears to literally just be a random image that anyone could have easily doctored, not even any kind of journalistic article or etc backing it.
I thought he already put tariffs on us? And then again?
As a Canadian, all I can say is "hey wait a minute, I've seen this one before!"
I'm shocked this sorta shit still happens in 2025, how did this come into being? Thus might be a rabbit hole I go down, who founded this program, who vetted it, etc
See documentation of business transactions for tax audits.
Yeah... no.
You can't compare taxes which involve transactions with the outside world, and are arguably the most important thing the government cares about, to the source code of some shitty mobile game that got made 5 years ago or whatever.
If you genuinely tried to make a law in your country that tech companies are legally required to preserve all their source code for games forever, do you know what would actually happen?
Your country's entire game industry would quickly dry up because that's an incredibly stupid thing to try and ask.
Companies aren't gonna sit and audit their developers git history commits for some mobile game or random steam release.
And, if you have any concept of how git or other forms of source control for games works, you'd also know that basic day to day operations would, potentially violate such a law, depending on interpretation.
And no company will wanna incur that risk so they will just avoid your country cuz it's law was written by someone with clearly zero understanding of how source control works.
Classic example of gamers demanding stupid stuff with zero clue about the actual implementation details of what they are asking for.
As long as the cost of losing business in the EU is higher than designing an EOL transition for games and hiring developers to actually do it, it’s in their best interests.
I hate to break it to you but the EU is not that strong of a market lol
People seriously underestimate the cost of this sort of thing, companies do NOT want to hand out copies of their proprietary software to the public.
The pretty much always have tonnes of important shit baked into it that still gets used in their newer software, so even if its old stuff, it still has bits and bobs in it that matter for their newer stuff they just put out.
But also just, in general, companies are not gonna be chill with people demanding they give them a copy of their backend software. It's just not gonna happen, and the EU is definitely the weaker of the 3 major markets. Companies are just gonna go "lol, now you don't get to play online I guess" instead.
If you designed it for that eventuality, yeah, it’s easy to do
This just goes back to the other issue:
If your country demands the game devs contort and twist their architecture to suit that country's demands, they just wont release it in your country at all.
Sorry but thems the breaks.
You'll have to get way more than the entirety of the EU on board with this to make any change. Youd have to get China and the US on board at the same time
If you target only one of them, that country will decline because it would just argue "you'd fuck up our industry and everyone would leave to [other country] for sales"
And good luck getting the EU, US, and China to all simultaneously agree to this sort of thing, lol.
The training costs effectively enter a "divide by infinity" argument given enough time.
While they continue to train models at this time, eventually you hit a point where a given model can be used in perpetuity.
Costs to train go down, whereas the usability of that model stretches on to effectively infinity.
So you hit a point where you have a one time energy cost to make the model, and an infinite timescale to use it on.