• Quick note - the problem with Youtube videos not embedding on the forum appears to have been fixed, thanks to ZiprHead. If you do still see problems let me know.

Merged Artificial Intelligence

I'm going down the rabbit hole a tad from that forum app the AIs created.

I can see a lot of good stuff outside of building applications for these agents. One of the most fascinating aspects is how it shows its "reasoning" while working—it doesn’t simply present a finished product with a "tada!" instead it shows each step of the process, detailing the files it generates and their contents. When an error arises, it backtracks, pinpointing the issue and outlining possible solutions. This makes it an excellent tutorial on how to code an app.

There's been some concern that computer science degrees might become obsolete, with AIs handling most coding tasks. However, I don’t think that’s the case, human input will still be needed* which will still need an understanding of coding principles, app development processes and so on.



*At least until the next big AI breakthrough and we create AIs that truly reason.
 
Apparently our "please" and "thank you" costs millions! We now have a way to objectively measure the financial cost and value of being polite!
 
This could be an interesting development. IBM announces a tool to help users attribute AI contribution to their work. Early days
IBM Research’s AI Attribution Toolkit is a first pass at formulating what a voluntary reporting standard might look like. Released this week, the experimental toolkit makes it easy for users to write an AI attribution statement that explains precisely how they used AI in their work.



 
I would seek to get rid of a co-worker who was as much of a bootlicker as an LLM interface - clearly, they are trying to cover for their shoddy work.
 
Bellingcat tested LLMs abilities to do geolocation. Quite interesting article
The article still maintains the irritating “hallucination” excuse for LLM errors.
(sorry, this really bugs me)

If these things are making ◊◊◊◊◊◊◊ mistakes, just admit it and stop the hallucination ◊◊◊◊◊◊◊◊.
Do these things ever perform and display error analysis of their results? Or do they, like ChatGPT, just bluster through with confident responses?
 
The article still maintains the irritating “hallucination” excuse for LLM errors.
(sorry, this really bugs me)

If these things are making ◊◊◊◊◊◊◊ mistakes, just admit it and stop the hallucination ◊◊◊◊◊◊◊◊.
It is the correct technical term for what is occurring. It's not just that the LLM is wrong about something. It is actively confabulating an unreal thing. I'm afraid you'll need to get over that particular peeve.

Do these things ever perform and display error analysis of their results? Or do they, like ChatGPT, just bluster through with confident responses?
Yes. Specifically, certain LLMs used in research are programmed to analyse their own output.

AI hallucination is a pretty interesting subject, actually.
 
The article still maintains the irritating “hallucination” excuse for LLM errors.
(sorry, this really bugs me)

If these things are making ◊◊◊◊◊◊◊ mistakes, just admit it and stop the hallucination ◊◊◊◊◊◊◊◊.
Do these things ever perform and display error analysis of their results? Or do they, like ChatGPT, just bluster through with confident responses?
This. AIs are programmed to (almost) always give an answer, any answer, even it it's not correct. Very few times have I encountered a bot that simply up and says, "I'm sorry, but I don't have enough information in my training data to give an answer."

However, I believe most AI bots have a disclaimer their output should not be trusted as completely accurate.
 
LLMs are always hallucinating - hopefully most of the time in a way that is accurate enough for our needs
 
On another point, AI bots are very bad at analyzing input for subtle errors. For example, "How many days did this child live, who was born on March 30, 1883 and died on September 15, 1905, given that 1888, 1892, 1896, 1900, and 1904 were leap years?" The bot will happily compute the number of days, ignoring the fact that despite my claim 1900 was a leap year, it was not.
 
This. AIs are programmed to (almost) always give an answer, any answer, even it it's not correct. Very few times have I encountered a bot that simply up and says, "I'm sorry, but I don't have enough information in my training data to give an answer."

All AI programmers should read Asimov's The Last Question. "There is insufficient data for a meaningful answer".
 
The article still maintains the irritating “hallucination” excuse for LLM errors.
(sorry, this really bugs me)

If these things are making ◊◊◊◊◊◊◊ mistakes, just admit it and stop the hallucination ◊◊◊◊◊◊◊◊.
Do these things ever perform and display error analysis of their results? Or do they, like ChatGPT, just bluster through with confident responses?
It's a new area and new nomenclature is needed. Hallucinations are when they make stuff up not when they make a mistake. A mistake would be a response that there are 4 letter "r"s in strawberry when asked how many rs in the word strawberry, an hallucination would be when it says there are 4 and provides an apparent quote or reference to an OED entry that doesn't exist to support the answer 4.

In humans we would say they are ◊◊◊◊◊◊◊◊◊◊◊◊ but prudish sites like this august venue would be in uproar to use such a term!
 
On another point, AI bots are very bad at analyzing input for subtle errors. For example, "How many days did this child live, who was born on March 30, 1883 and died on September 15, 1905, given that 1888, 1892, 1896, 1900, and 1904 were leap years?" The bot will happily compute the number of days, ignoring the fact that despite my claim 1900 was a leap year, it was not.
Seems that would be something they have in common with their creators...
 

Back
Top Bottom