AI chatbot deception paper suggests that some people (and bots) aren't very persuasive.
See full article...
See full article...
The Stackexhange thread you link says there is no relationship. Still…If you have a better explanation, why do you not provide it?
And complexity
Do you understand the difference? Is there a difference?
I don't get it. You claim that I don't understand, but how am I wrong? Are you going to explain what it means to be Turing complete in a way that is truthful to your experience and knowledge?
Did the LLM get something wrong? If so, can you explain the reasoning for the error?
I'm sick of the bullshit. Be concise in your explanations, don't just tell me I'm stupid. That's not an arguable point, that's just your opinion.
Edit: You know what, I'll even provide a source:
https://cs.stackexchange.com/questi...to-do-with-turing-completeness-or-turing-mach
In other words, in order to pass a Turing test the machine needs to first be Turing complete.
Get it?
Your response?
Who says you can't make it an actual test? Next you're going to tell me I should let my cat out of the box he's in!!!The Turing test is really widely misunderstood. Turing never intended it to be used as an actual test. His imitation game was just a metaphor to illustrate a point about the nature of intelligence: it's defined by behavior. An intelligent agent is one that behaves intelligently. It doesn't matter what mechanism leads to the behavior. If a machine behaves exactly like a human, then by definition it's as intelligent as a human.
Somewhere along the way, people started misunderstanding it and thinking they were actually supposed to administer it as a real test. That leads them to think intelligence is defined as behaving exactly like a human. It isn't. A machine that behaves exactly like a human is clearly intelligent, but a machine that behaves very differently can also be intelligent. Or worse, as in this case, people think intelligence is defined as the ability to fool humans into thinking you're intelligent.
But is there really a cat in the box? And is it able to get out of the box, or is it dead?Who says you can't make it an actual test? Next you're going to tell me I should let my cat out of the box he's in!!!
That study had a huge (and deliberate) flaw introduced into it - the "actual humans" were selected from the generator's training dataset using another classification AI - for each AI image they picked a real image that the classifier placed closest to the AI image.In an unrelated study from November (Miller, et al), researchers found that people thought AI-generated images of humans looked more real than actual humans.
Your post makes little sense, though I think I catch your drift, with which I tend to agree. Transparency on training sources would be a good thing.That study had a huge (and deliberate) flaw introduced into it - the "actual humans" were selected from the generator's training dataset using another classification AI - for each AI image they picked a real image that the classifier placed closest to the AI image.
Which is bizarre, and they didn't publish the result for the obvious original version of the experiment where the actual humans would be selected on random from the training dataset. Neither a justification nor a measure of the effect of this bizarre procedure was given.
As typical they used images stolen off the social media, i.e. of unknown provenance rather than non-manipulated images taken in a standardized setting. It could just as well be that the training dataset included some heavily retouched images that were more frequently chosen as "matches" to the AI images.
I would imagine that if you used some writing style classification AI to pick "witnesses" who resembled ChatGPT the most, you could end up selecting the above mentioned trolls who deliberately pretend to be an AI.
It is rather hard to explain what was done in the [Miller, et al] paper, because what they done makes basically no sense.Your post makes little sense, though I think I catch your drift, with which I tend to agree. Transparency on training sources would be a good thing.
Emphasis mine.For each synthesized face, we collected a matching real face (in terms of gender, age, race, and overall appearance) from the underlying face database used in the StyleGAN2 learning stage. A standard convolutional neural network descriptor (15) was used to extract a low-dimensional, perceptually meaningful (16) representation of each synthetic face. The extracted representation for each synthetic face—a 4,096-D real-valued vector —was compared with all other facial representations in the data set of 70,000 real faces to find the most similar face. The real face with representation with minimal Euclidean distance to , and satisfying our qualitative selection criteria, is selected as the matching face. As with the synthetic faces, to reduce extraneous cues, we only included images 1) with a mostly uniform background, 2) with unobstructed faces (e.g., no hats or hands in front of face), 3) in focus and high resolution, and 4) with no obvious writing or logos on clothing. We visually inspected up to 50 of the best matched faces and selected the one that met the above criteria and was also matched in terms of overall face position, posture, and expression, and presence of glasses and jewelry. Shown in Fig. 4 are representative examples of these matched real and synthetic faces.
For brilliant mind, Turing seems confused on the topic of machines thinking. Read any of his papers on the subject, e.g. in Copeland, The Essential Alan Turing. One almost wonders if he is pulling everyone’s leg in these papers.
I think it is simply that at the time, several enabling factors of modern AI were utterly inconceivable to anyone, not even Turing.I don't think he was confused at all, more that based on his very limited exposure to computers he designed, the test was more a theoretical "conversation starter" about how far computers will take us.
The Stackexhange thread you link says there is no relationship. Still…
Turing complete is a description applied to a system that can simulate a Turing machine, which is a formalism developed by Alan Turing for studying computability and complexity theories (which are formal mathematics/computer science fields). It is related to Church’s lambda calculus. As it turns out, all modern digital computers are (aside from finiteness) Turing complete.
The Turing test is an informal “game” that Alan Turing proposed as a way to determine whether a machine is acting intelligently. It has nothing to do with actual machines, but rather with a theory or mind or intelligence - it was definitely not a formal test.
These are completely unrelated. In the grand scheme of things, the former is far more important than the latter. Feel free to check Wikipedia for more info on both - that’s a lot more solid than looking at Stackexchange for that sort of info.
For whatever it’s worth, Alan Turing was a giant in numerous fields and his name is attached to many things that were often unrelated (other than generally having a connection to mathematics or computer science). Again, feel free to check Wikipedia’s article on Turing.
I am not a behavioral psychologist. I do think the test setup has a deep flaw: the incentives it sets up may encourage humans to throw the test. (Case in point: the paper authors acknowledge trolling.)If the humans are scoring less than the high 90s that just shows that the test setup is insufficient.
And it may be impossible to set up the test with impartial judges now that there is so much public interest in AI.
While this test is clearly insufficient in the modern age to achieve the intent of Turing, and should be replaced, I would nevertheless like to propose a scoring improvement. Namely, filter the scoring so that you only use the scores of the people who correctly identified the humans.
It should be obvious that when the purpose of the test is to measure the ability of the AI to impersonate a human, if it is realistic for the AI to score "more human than human" then the test itself is a failure. Complete, absolute success would be equality between the groups. Filtering the scores in the way that I propose would achieve this correction.
This exchange really shows why it is a bad idea for us to be putting LLMs in everything, or at least in searches. Until this comment section, I didn't know anything about Turing machines or Turing completeness. Because these are complex topics, I still, basically, don't know anything. I could do a google search and find a website willing to explain these topics in more-or-less detail in a sort of end-user-y kind of way, the quality of which will depend on the website. Or I could ask an LLM to produce absolute rubbish but, because I don't know any better, and it is well-formatted, it sounds reasonable. And then I go on my merry way "knowing" something that is wrong, and I'll have no chance to correct, unless I say something stupid to someone who knows better, because it doesn't actually impact my life, and I don't have a framework to gauge the truth-value of the statement. In the mean time, I have the opportunity to infect other people with my mistaken understanding of these concepts, because most of the people I know don't really have a reason to know what Turing machines or Turing completeness are. This is the don't-trust-what-you-read-on-social-media problem on steroids.Yeah, what a bizarre thing to say.
It is equivalent to saying: "The game of Minecraft, and nearly all programming languages, among other things, are indistinguishable from a human being if you hold a conversation with them."
These, of course, all being examples of things which are Turing complete, but of dubious ability at the Turing test. Except insofar as a Turing complete system can technically run any program we might want to Turing test...might take awhile to run GPT 4 in Minecraft, though.
And yet students took it seriously at MIT. Which prompted Professor Weizenbaum to write Computer Power and Human Reason: From Judgment to Calculation.Ah, ELIZA. Used to run it through the Automatic Mouth synthesizer for party entertainment. It was like getting therapy from a Conehead.
Any useful computer language is Turing Complete. It's actually kind of hard to find a programming system that isn't Turing Complete, although we know that there is a hierarchy of computing models with increasing power (string grammars, push down automatons, etc.).Google is a thing you know. However just for you, a Turing test is designed to see if a machine can mimic human conversation. Turing Complete refers to an algorithm that can simulate a turing machine - which is a universal computer, not a human.
You can't know until you open the box. Until then, the cat is both alive and dead.But is there really a cat in the box? And is it able to get out of the box, or is it dead?
Respectfully, I think it worked out exactly the way he intended.You can't know until you open the box. Until then, the cat is both alive and dead.
---
Erwin Schrödinger actually proposed this experiment to show how ridiculous quantum mechanics seemed. It didn't work out the way he intended.
First: your premier model should be named Jigsaw Man.Two things: first, "More human than human" sounds like a good motto for an android company.
Second: if tricking humans into thinking you're human by trying to act like a human is a thing, what's it mean when some humans try to trick humans into thinking they're not human by trying not to act like a human?
The OG. Rickman was a good choice, but Moore created the character, and is still the one I hear in my head.Well, Stephen Moore, the voice of Marvin in the original radio and TV series did it pretty well, too...
I'm remembering an anecdote from a few years back where the bot that "won" a Turing "competition" (had the highest success rate at being judged human) passed itself off as an 8 year old Hungarian kid with limited English. So yeah, perfectly valid strategy.Perhaps playing 'too stupid to be a computer' might have worked.
Except that, by the "rules" of most attempts at creating a Turing Test, that is explicitly forbidden.I'm remembering an anecdote from a few years back where the bot that "won" a Turing "competition" (had the highest success rate at being judged human) passed itself off as an 8 year old Hungarian kid with limited English. So yeah, perfectly valid strategy.
The Imitation Game is just that -- a game. It's fun to play around with but fundamentally tells you little. You can ID most bots by asking them specific personal questions, claiming postulates that run directly contrary to physical reality, memory / recall tests, and just seeing if they'll object to outright gibberish.That said, I very much don't think that the Turing Test is actually a thing.
We have less face to face social interaction than ever, Bowling Alone etc. We are probably stiffer at communication than we were a few generations ago.It is remarkable that over a quarter of humans didn't successfully identity other humans!
A forced-choice test where an Interrogator interacts with an AI and a human and picks which one is more likely to be human would probably (hopefully!) have a much higher accuracy rate.
But LLMs are neither Turing complete nor are they intelligent! What are you arguing?The Stackexhange thread you link says there is no relationship. Still…
Turing complete is a description applied to a system that can simulate a Turing machine, which is a formalism developed by Alan Turing for studying computability and complexity theories (which are formal mathematics/computer science fields). It is related to Church’s lambda calculus. As it turns out, all modern digital computers are (aside from finiteness) Turing complete.
The Turing test is an informal “game” that Alan Turing proposed as a way to determine whether a machine is acting intelligently. It has nothing to do with actual machines, but rather with a theory or mind or intelligence - it was definitely not a formal test.
These are completely unrelated. In the grand scheme of things, the former is far more important than the latter. Feel free to check Wikipedia for more info on both - that’s a lot more solid than looking at Stackexchange for that sort of info.
For whatever it’s worth, Alan Turing was a giant in numerous fields and his name is attached to many things that were often unrelated (other than generally having a connection to mathematics or computer science). Again, feel free to check Wikipedia’s article on Turing.
Right, no, it is not.The other thing is that we can say that ChatGPT is not human-level intelligent
If the latest GPT is rules-based (digital rules in logic that's operating a neural network, via linking all the parameters in the digital neural network), then the human brain is also defacto rules-based (analog rules from all the linkages of synapses/neurons in the biological neural network).GPT is rules-based. There's more rules (and sure, there's data the model can access) but there are pre-defined rules, pre-defined algorithms. GPT is, in its fundamentals, identical to ELIZA.
Bottled knowledge it may (it's a good description of current models). It is currently being iterated to becoming more capable of creating more original outputs than earlier AI's. Example: An original joke on URL being Unusually Rowdy Lemurs, something that doesn't even exist on Google via a quoted search.My thinking on this is that ChatGPT and other large language models represent "bottled" human intelligence.
I want my kid to learn to do math. Not rely on a chatbot to do it for him. In my era, people railed against using calculators to help. Times evolve. Tools get better, but the human needs to know what the tool is producing.Right, no, it is not.
That being said, we can see that GPT4-Turbo and GPT3.5 (non-Turbo) has roughly the same number of parameters, yet GPT4-Turbo is more intelligent than GPT3.5 (free ChatGPT). And it does manage to line-item converse on some topics better than the average human of 50-percentile skill on the respective topic. And for very long conversations (especially with the new 128K context memory, about 500KB of text worth of short-term memory).
And you've been hearing about the much more intelligent Q*, the very scare that caused the OpenAI fracas. We can't deny the technological pace has, to put it mildly, been very torrid.
Bottled knowledge it may (it's a good description of current models). It is currently being iterated to becoming more capable of creating more original outputs than earlier AI's. Example: An original joke on URL being Unusually Rowdy Lemurs, something that doesn't even exist on Google via a quoted search.
Various amounts of bottled knowledge plus some of the improved algorithms I talked about in earlier pages of that other comments section, is going to be able to make a 99% Virtuso AGI (under DeepMind's new AGI definitions) by end of the decade in my knowledge.
AI's aren't being designed to be human-level intelligent anymore -- it is starting to line-item surpass, while still greatly inferior than others. By the time we optimized final skills to equal humans, the existing longtime AI skills will be literally ASI-league.
Q* does grade school math better than a 50%-ile human, from what is reported -- the very thing that caused almost 95% of the company to start preparing to move to Microsoft because Sam Altman got fired. They've got something seriously better of an AI that actually thinks/reasons/plans, already in the wings -- that caused the CEO to be fired. Then 95% of the company was getting ready to quit, to move to Microsoft (who offered them all jobs). Until Sam got rehired. Despite being a controversial figure (and all the copyright hoovering controversies), one can't deny that something seriously smarter has been birthed in the lab.
The path to AGI's are a complex mix of ANI, AGI, ASI rather than the yesteryear definitions. It's going to be more like talking to an alien who's hyperintelligent, moreso than 99th-percentile human on 99% of topics, and only fail on a tiny percent. This isn't going to be human-level intelligent. We're witnessing emerging AGI's that are both simultaneously inferior and superior to humans, but the number of skillsets that transfers over to the superior-side, rapidly increase. And that has been my exact experience with the paid AI's (especially the new ones with the new much larger short-term memories).
Averaging out, we're not really averaging to human-level intelligent, so I agree with you there. But it's more nuanced than that, especially witnessing what is happening with GPT 3.5 versus GPT-4 Turbo versus what we all keep hearing about Q*, the much more intelligent AI that caused the board to fire the founder of OpenAI.
I'm predicting that particular one (Q* and variants) will quickly be ranked as "Competent AGI" rather than "Emerging AGI", but not a "Expert AGI" nor "Virtuso AGI" under the new DeepMind AGI definitions (Page 6 of https://arxiv.org/pdf/2311.02462.pdf ) at least when narrowscoped to skills achievable in disembodied abilities (e.g. things that don't yet require a robot body). And even industry's working on that too.
Too many people make assumptions about how dumb AIs still are, based off free ChatGPT or the older version of paid ChatGPT, when the pace behind the scene has been rather crazy fast, and the November 6th version of GPT4 suddenly had literally something like ~15x to ~20x as much short-term memory ready for much longer conversations or even accurately meeting-minuting a multihours-long meeting successfully where it couldn't before November -- they're really improving the paid version of the AI subscription very fast.
On that item as a Gen X I agree.I want my kid to learn to do math. Not rely on a chatbot to do it for him. In my era, people railed against using calculators to help. Times evolve. Tools get better, but the human needs to know what the tool is producing.
There is no AI. There are decent LLMs. My kid's mom cannot read a map, and has fallen astray when on an international trip. I do not think cursive is on my son's agenda as he moves forward in life. I believe they now call it some version of italics.On that item as a Gen X I agree.
The loss of cursive handwriting too...
The loss of ability to read paper maps...
The loss of...
Nontheless, as a deaf individual and as of past special-needs student back in my day, AI has made a massive impact as being one of the best coding tutors (even for those times I don't want to use its code), as well as the best speech-to-text transcriber. I can now do captioned FaceTime thanks to Apple's adding a neural-based Live Caption feature in iOS 16+. Right Tool for Right Job for me.
If over a dozen people independently point out that you are wrong, and why, a sane person would think "hmm, maybe it's me".I'm going to say it again, LLMs need to be Turing complete in order to pass the Turing test!
I read an article a few years ago claiming the Rust type system is Turing-complete.Yeah, what a bizarre thing to say.
It is equivalent to saying: "The game of Minecraft, and nearly all programming languages, among other things, are indistinguishable from a human being if you hold a conversation with them."
These, of course, all being examples of things which are Turing complete, but of dubious ability at the Turing test. Except insofar as a Turing complete system can technically run any program we might want to Turing test...might take awhile to run GPT 4 in Minecraft, though.
But instead of thinking "maybe it's me", the earlyberd goes on the offensive... making me suspect it's just a troll.If over a dozen people independently point out that you are wrong, and why, a sane person would think "hmm, maybe it's me".
This person, allegedly, writes LLM policy for his/her company. Sleep tight!