
News updates on the development and ramifications of AI. Obvious header joke is obvious.
Idle thought: the premise "make an AI more factual by having it do a web search before answering" relies strongly on the assumption that arbitrary web pages almost never include malicious jailbreak instructions.
Or lies?
I mean, then it's a value judgment - if Bing serves the same fact to two people they may disagree whether it's a lie or not.
I'm probably weird, but I tend to think it's not very useful to talk about truth-vs-lies when it comes to generative AI. Like, by analogy to image generators, if stable diffusion draws me a picture of Einstein using a smartphone, is it lying? Considering that GPT does to tokens more or less what SD does to pixels, it strikes me as the same kind of question.
Don't confuse cultural ideas of truth with epistemological ideas. For regular people, arguing about "truth" is not the same as determining whether a statement is true or not.
For example, Trump lost the 2020 US presidential election. This is a true statement by objective standards. However, culturally - subjectively - it could be false. But that in itself would be a false judgment, again by objective standards.
It's not useful to hold up *cultural* truths as examples of truth, it's just confusing the issue. Our choices tend to be "We don't know enough to say", or "as best we can discover, this is true", or "as best we can discover, this is false". The difference between this and cultural truths is whether or not they are congruent with what is real, all the way down the stack. Cultural truths veer off at some point based on considerations other than shared realities; epistemological truth goes down the stack with verification at every step.
If you want AIs to be truly useful, their statements need to be based on solid epistemic evidence, rather than just accepting whatever they find as "true" because it is asserted on the Internet. If people disagree on a statement, then either it's an open statement - "We don't know enough to say" - or one of them, or both, are wrong.
Ultimately, AIs should be able to settle these questions *before* responding, which is going to take an interesting "fact verification" infrastructure. This will become necessary, because if you control what are taken to be facts by the general public, you control the thinking of your population. An uncontrolled AI knowledge base is an invitation to dictatorship, hacking, and many other problems. We are already seeing this.
If you want AIs to be truly useful, their statements need to be based on solid epistemic evidence, rather than just accepting whatever they find as "true" because it is asserted on the Internet. If people disagree on a statement, then either it's an open statement - "We don't know enough to say" - or one of them, or both, are wrong.
The real problem is that AI doesn't have another way to learn if something is "true" or not. Things written in books can be wrong. Things that people in general agree on can be wrong. Things on Wikipedia can be wrong! People make arguments like "vaccines cause autism", "the earth is flat", and "9/11 was an inside job" and they cite sources for all of these things. AI doesn't have access to all of these sources, and even if it did, it can't really make a judgement on whether their conclusions are "correct" or not. The best it could say is that "this contradicts something else". Pretty soon everything "contradicts something else" and AI has no idea whatsoever if even the most basic facts are true.
We would have to have the AI create an objective "database of facts" containing only things that are absolutely true - but then AI would have to be able to add more things to it as it "learns", and sooner or later "Humans are the worst thing that has ever happened to the planet" gets added to the list and there are multiple science fiction franchises that deal with that and pretty much all of them would be bad for us.
That's why I say that a "truth infrastructure" would be "interesting". I'm tempted to argue that this is simply a business opportunity - or one for scientists and philosophers. A new field of research might be created. Heck, it might be being created as we speak.
Obviously, science does a pretty good job at this, especially in teaching environments. Perhaps we need to train AIs at the university and post-graduate level. But I suspect that there has to a "subjectivity rating" that puts, for example, evidence about physical phenomenon at a higher level of confidence than, say, Marxist Literary Theory... And making that quiet hierarchy explicit will be an "interesting" task in the cultures of academia.
Also we need to be able to express pragmatics - degrees of truth, even likelihood of change over time. That's it's own challenge as well.
And then, culturally, in countries like the US, when Evolution shows up as more credible than anything based on religious reasoning, fur is gonna fly. At the speed of bullets and bombs, I suspect...
Robear, that premise is like saying we could train AIs to win the lottery if we just gave them a database of future lottery numbers. A fact database that only contains true facts isn't and can't be a thing, but if we had one we could just query it without getting an AI involved. (Also if you think it might be a good idea, note for the record that the person currently most likely to try making such a database is probably Elon Musk...)
Whereas, the point I was trying to make is that I don't think it's useful to talk about the truthiness of AI outputs in the first place. When someone complains that ChatGPT told them a lie, to me they're making the same category error as someone who complains that Stable Diffusion gave them a picture of something that doesn't actually exist. LLM training inputs aren't labeled for objective truth - it's simply not something AIs know how to maximize.
So why can't we train on something like Encyclopedia Brittanica, as a start? Perfect is the enemy of good enough, and right now we don't have anything like "good enough".
The idea that we can't accumulate knowledge of what is actually real flies in the face of the entire purpose of science...
So why can't we train on something like Encyclopedia Brittanica, as a start?"
Because scale. You need many orders of magnitude more input data to train an LLM that exists in all the encyclopedias. The "large" in LLM is an understatement.
So the problem is discernment. That's an interesting one. Again, truth evaluation would be important, as well as distinguishing conversational elements from subjects that require truth evaluation. (off-the-cuff observation, could be wildly off-base)
The idea that we can't accumulate knowledge of what is actually real...
I mean.. not only is that not what I said, it's not even a thing that anybody ever would say.
So why can't we train on something like Encyclopedia Brittanica, as a start? Perfect is the enemy of good enough, and right now we don't have anything like "good enough".
LLMs have already been trained on encyclopedias. What's being talked about here is doing a web search for additional text to add in as context to a given prompt, and doing that can certainly make an LLM more useful at question answering. But it can't solve the problem of the LLM saying untrue things - not least because you don't start out knowing which web pages say true things (or if you do, you don't need the LLM).
But more importantly, like I said being truthful just isn't a constraint that LLMs have a way to follow. An LLM trained solely on true statements will still say untrue things, for the same reason that an image generator trained solely on historical photographs will still make images of things that don't exist.
Anyone else feel the Hollywood writer strike is coming at the worst possible time? I support their cause, but I also imagine the studios will just double down on AI-generated scripts and expecting the actors to improvise the rest.
Anyone else feel the Hollywood writer strike is coming at the worst possible time? I support their cause, but I also imagine the studios will just double down on AI-generated scripts and expecting the actors to improvise the rest.
Probably better now than later when those scripts/AI get more advanced. It might be good to get some out there to fail.
Probably better now than later when those scripts/AI get more advanced. It might be good to get some out there to fail.
I mean what writer's room could compete with the pure gold of highly compelling scripts AI's are currently sh*tting out?
There's a Kokomo, IN but it's a pit.
As a kid watching music videos including that particular Beach Boys song in the very early 90's, I actually thought Kokomo was real. I mean the chorus rattles off real places (ooh I wanna take her to Bermuda, Bahamas...). I later attempted to find Kokomo using the paper Britannica encyclopaedia my parents bought us a few years later with no success.
Thanks for the link - I had heard the strike was more over pay and benefits but trying to set more rules for AI make sense.
All that being said, I could easily see execs signing off on that terrible Kokomo scene. I mean we are talking about an industry that not too long ago pitched having Julia Roberts play Harriet Tubman.
The strike does touch on the use of AI, but the writers are really trying to deal with the other technology/trend that f*cked them over: the massive shift to streaming.
Chris James from the youtube prank channel Not Even A Show has been leaning hard into AI voice replication for the last several months, calling politicians and media figures while playing pre-generated clips of people like Ben Shapiro, Sebastian Gorka, and Mike Huckabee. Way too many of the people of he calls never seem to realize it’s AI, and several of them recognized the voices and thought it’s the actual person.
Counterpoint:
https://www.fastcompany.com/90892235...
FC: On CNN recently, Hinton downplayed the concerns of Timnit Gebru—who Google fired in 2020 for refusing to withdraw a paper about AI’s harms on marginalized people—saying her ideas were not as “existentially serious” as his own. What do you make of that?MW: I think it’s stunning that someone would say that the harms [from AI] that are happening now—which are felt most acutely by people who have been historically minoritized: Black people, women, disabled people, precarious workers, et cetera—that those harms aren’t existential.
What I hear in that is, “Those aren’t existential to me. I have millions of dollars, I am invested in many, many AI startups, and none of this affects my existence. But what could affect my existence is if a sci-fi fantasy came to life and AI were actually super intelligent, and suddenly men like me would not be the most powerful entities in the world, and that would affect my business.”
FC: So, we shouldn’t be worried that AI will come to life and wipe out humanity?
MW: I don’t think there’s any evidence that large machine learning models—that rely on huge amounts of surveillance data and the concentrated computational infrastructure that only a handful of corporations control—have the spark of consciousness.
We can still unplug the servers, the data centers can flood as the climate encroaches, we can run out of the water to cool the data centers, the surveillance pipelines can melt as the climate becomes more erratic and less hospitable.
I think we need to dig into what is happening here, which is that, when faced with a system that presents itself as a listening, eager interlocutor that’s hearing us and responding to us, that we seem to fall into a kind of trance in relation to these systems, and almost counterfactually engage in some kind of wish fulfillment: thinking that they’re human, and there’s someone there listening to us. It’s like when you’re a kid, and you’re telling ghost stories, something with a lot of emotional weight, and suddenly everybody is terrified and reacting to it. And it becomes hard to disbelieve.
Pages