You can get pretty in the weeds with conversions on ChatGPT in the chemistry world or even just basic lab work where a small miscalculation at scale can cost thousands of dollars or invite lawsuits.
I check against actual calibrated equipment as a verification final step.
I imagine ChatGPT and code is a lot like air and water.
Both parts are in the other part. Meaning llm is probably more native at learning reading and writing code than it is at interpreting engineering standards worldwide and allocation the exact thread pitch for a bolt you need to order thousands of. Go and thread one to verify.
This is possibly true due to the bias of the people who made it. But I reject the notion that because ChatGPT is made of code per se that it must understand code better than other subjects. Are humans good at biology for this reason?