Admit it: ‘Artificial general intelligence’ may already be obsolete, Expecting OpenAI’s GPT and other large language models to beat humans at thinking like a human might be missing the point.

Elon Musk filed a lawsuit in San Francisco’s Superior Court accusing OpenAI and its CEO, Sam Altman, of betraying the startup’s initial commitment to openness, the betterment of society, and lack of profit as a motive. Among other things, Musk’s 35-page complaint argues that OpenAI has violated its original deal to share its GPT large language models with Microsoft, which stated that the software giant would lose access to new LLMs once OpenAI had achieved AGI. According to the complaint, OpenAI reached that epoch-shifting moment a year ago with GPT-4, its most powerful model to date.

Musk—who cofounded OpenAI but left in 2018—is at least as entitled as anyone to come up with his own definition of AGI. His complaint describes it as “a general purpose artificial intelligence system—a machine having intelligence for a wide variety of tasks like a human.” That does sound like GPT-4 as I, a mere layperson, experience it in ChatGPT Plus.

But Musk’s declaration that the AGI era is already upon us is hardly the consensus among AI scientists. Even those who think it’s not far off predict arrival dates that are least a few years away. And GPT-4 falls well short of meeting OpenAI’s own explanation of the term: “A highly autonomous system that outperforms humans at most economically valuable work.”

Consider the evidence:

GPT-4 isn’t remotely autonomous; indeed, it does its best work when humans provide plenty of hand-holding in the form of detailed prompts. The world is still in the process of figuring out what tasks GPT-4 can do, and we frequently overrate its competence. That’s not even getting into the fact that OpenAI’s reference to “most economically valuable work” suggests that true AGI may involve not just software but also sophisticated robotics that don’t exist yet. To guess when OpenAI—or a rival such as Google, Anthropic, Meta, Mistral, or Perplexity—might reach AGI, as OpenAI defines it, is to expect that it’ll be an obvious moment in time. But OpenAI’s definition, like all the others, is squishy and difficult to put to a conclusive test. To riff on Supreme Court Justice Potter Stewart’s famous comment about pornography, maybe we’ll know it when we see it. At the moment, however, I’m convinced that obsessing over AGI’s existence or nonexistence is counterproductive.

The whole notion of AGI is predicated on the assumption that AI started out dumber than a human but could someday match or exceed our level of thinking. Already, though, generative AI is different than human intelligence—far closer to omniscient than any individual flesh-and-blood thinker, yet also preternaturally gullible and prone to blurring fact and fiction in ways that don’t map to common human frailties. That’s because it’s a predictive engine, trained to string together words without truly understanding them. If its present trajectory of simulated brilliance mixed with boneheadedness continues, it might wander off in a direction far afield from most definitions of AGI.

Even if the world lands on a new, more inclusive definition of AGI, it may be hard to prove whether a particular LLM has attained it. Musk’s lawsuit cites proof points of GPT-4’s reasoning power, such as its scoring in the 90th percentile on the Uniform Bar Exam for lawyers and the 99th percentile on the GRE Verbal Assessment. That it can do so is astounding. But acing tests is not synonymous with performing useful work. And even if it were, who gets to decide how many tests an LLM must pass before it’s achieved AGI rather than just bobbled somewhere in its vicinity?

For decades, the Turing Test—which a computer would pass by fooling a human into thinking that it, too, was human—was computer science’s beloved thought experiment for determining when AI had gotten real. Strangely enough, it’s useless as a tool for assessing today’s LLM-based chatbots. But not because they know too little to fake humanity convincingly, or can’t express it glibly enough—but because they betray their artificiality by being so good at churning out endless wordage on more topics than any human knows. AGI could end up in a similar predicament: a benchmark, devised by humans, that’s rendered obsolete by the technology it was meant to measure.

DID YOU HEAR THE ONE ABOUT THE “MAC CAR?” Last week, Apple’s long, expensive quest to build an autonomous EV entered its rearview-mirror phase—a sad fate my colleague Jared Newman blamed on the company’s sometimes counterproductive pursuit of perfection. Wondering what an Apple car would be like has been an obsession for techies since 2012, when news broke that Steve Jobs had toyed with getting into the automobile business even before there was an iPhone. Or maybe it started in 2008, when reports of a meeting between Steve Jobs and Volkswagen’s CEO led to wild speculation about an “iCar.”

Or how about 1998? According to Snopes, that’s when a joke involving cars designed by software companies began spreading like crabgrass across the internet, eventually evolving into an urban legend involving a Bill Gates keynote and a General Motors press release. Along with a Microsoft car that crashed twice a day and occasionally needed its engine replaced for no apparent reason, it mentioned a “Mac car” that “was powered by the sun, was reliable, five times as fast, twice as easy to drive—but would only run on 5% of the roads.”

KeenFlame,

I’m so confused by “it’s already here” “what even defines agi” etc. “what do we mean”

Like is it really that hard to understand the difference of a function you call and it returns something

And an autonomous entity

so hard to distinguish?

Fedizen,

There are ethical implications to GAI as well.

ruben,

“was powered by the sun, was reliable, five times as fast, twice as easy to drive—but would only run on 5% of the roads.”

Which ones? The ones that go downhill??

Rolando,

We are approaching another AI Winter. AI goes through hype cycles:

  • some flashy new capability captures everyone’s imaginations
  • Companies and researchers exaggerate the possibilities and people are led to believe that full AI is right around the corner.
  • Then with familiarity people realize the new capability isn’t really that revolutionary, and the term “AI” is distrusted again, for another decade or two.

It’s been that way since the 1950s. Read more about it: en.wikipedia.org/wiki/AI_winter

sukotai,
@sukotai@lemmy.world avatar

if everybody says ‘the earth is a square’, these A.I will say ‘the earth is a square’ : there is no intelligence, it’s just a summary of what everybody says. if one day, a machine is able to say : you’re wrong, the earth is a sphere, and here is the reason…, then i will say ’ ok, you’re a real A.I ’

cloudless,
@cloudless@feddit.uk avatar

In a way, human intelligence is like that.

People used to think earth used to be the centre of the universe, because everybody said so. Would you say that only Nicolaus Copernicus was intelligent?

KRAW,
@KRAW@linux.community avatar

How do you get AI to change its answer when one researcher discovers what was generally accepted as fact is no longer true?

cloudless,
@cloudless@feddit.uk avatar

They have to update the training data with the latest findings. Some AI models may use external sources to fetch the most current information.

TheChurn,

But the most current information does not mean it is the most correct information.

I could publish 100 papers on Arxiv claiming the Earth is, in fact, a cube - but that doesn't make it true even though it is more recent than the sphere claims.

Some mechanism must decide what is true and send that information to train the model - that act of deciding is where the actual intelligence in this process lives. Today that decision is made by humans, they curate the datasets used to train the model.

There's no intelligence in these current models.

cloudless,
@cloudless@feddit.uk avatar

Flat earthers have access to all the information yet they still decide that flat earth is true.

I am not saying that current AI is intelligent. I am just seeing similarity between how human and AI process information.

TheChurn,

Humans are intelligent animals, but humans are not only intelligent animals. We do not make decisions and choose which beliefs to hold based solely on sober analysis of facts.

That doesn't change the general point that a model given the vast corpus of human knowledge will prefer the most oft-repeated bits to the true bits, whereas we humans have muddled our way through to some modicum of understanding of the world around us by not doing that.

sukotai,
@sukotai@lemmy.world avatar

human behavior is like that : we repeat what we heard, so A.I is just copying the human behavior, it’s not a proof of ’ intelligence ’ for me. ( I’m in no way intelligent ). But may be these A.I will be able to have a ‘relationship intelligence’ : Know how to manipulate human behavior is probably a kind of intelligence.

dustyData,

That’s a problem with the technological comparison model of intelligence. We have dealt with it since the inception of calculators. Humans are not machines. Machines can emulate behaviors, but they are not and will never be like human entities. We have used them as metaphors for everything from mathematical thinking to memory and visual processing. But the truth of the matter is that, both neurological and phenomenologically speaking, we don’t think like machines, and we are not anything like them.

Humans don’t just repeat what we hear, just like artist don’t just mix and mash all the art they have seen in their lives the way stable diffusion and other image generators do. There’s a lot of things underlying the superficial process of stringing words together or composing a drawing or painting, happening in our brains that machines cannot do.

Pheonixdown,

Other people had the capability to do what Copernicus did, but lacked desire/resources. A LLM will never have the capability for a novel idea.

HaywardT,

An LLM may not. Will an AI?

Pheonixdown,

If by AI, you mean the things people are making today and calling AI, no, they’re all basically powerful regression algorithms. They can be strong tools for people to use to solve complex problems. Anything a program does will be based on what it was programmed to do, at best it will find novel things based on being programmed to look for novel things randomly and people will test and confirm those guesses. They already kind of do this for some medical purposes. Is trying an uncountably large number of randomized guesses and giving a probability for success based on historical data intelligent?

Could a true AI exist like we see in SciFi, maybe?

wizardbeard,
@wizardbeard@lemmy.dbzer0.com avatar

They also would lack the “desire” and resources to do so.

They can’t act of their own volition without input, and they can’t access systems they were not designed to interface with and data that they were not trained on or given through the input.

I think it’s preferable that way, given the immense overhyping of this technology that is ocurring, and the existing cases of misuse.

i_have_no_enemies,

i heard this is where q* learning comes to play, this algorithm will allow it to reason

mutant_zz,

It’s all bullshit marketing hype until we actually see it. There’s no reason to believe AI will advance better than linearly in the next 5-10 years.

i_have_no_enemies,

yes but the algorithm is real though

topinambour_rex,
@topinambour_rex@lemmy.world avatar

Which AI it would be : emotional, logical, spatial, etc ? Because there is not one intelligence in the humand mind, but several.

What will be very amazing, will be an AM, an artificial mind.

FaceDeer,
@FaceDeer@fedia.io avatar

Ah, irony. It's common for people to say "AI art generators are just collage machines, copying and pasting bits of existing images together, unable to generate anything novel." I guess there's no intelligence there either, they're just parroting each other.

scorpious,

It seems to me that the long experiment playing out may include simply waiting to see if there is a critical mass threshold to be reached (ie, of this LLM “simply repeating what everyone agrees on” idea) that allows the process to evolve into something closer to “thinking.”

I’m sure I don’t know enough about LLMs, but as others others here are pointing out, this seeming regurgitation of the already-known does seem to provide the foundation or potential for generating hypotheses and/or “new” ideas.

dustyData,

critical mass threshold to be reached

There isn’t. No amount of computational accumulation can result spontaneously into a mind. There’s not enough flexibility and malleability in the underlying processes (algorithms) that run LLMs. The process never changes therefore it cannot evolve into something other than what it already is. It’s like adding pixels to a monitor, no amount of pixels will ever spontaneously morph into reality. The switch from a 2D representation of a 3D world is not something that is possible.

General_Effort,

Funny how goals have evolved. From making a machine that is like a human to making one that is not.

GiveMemes,

Huh? I think you may have misinterpreted his comment. He’s looking for a machine that is like a human (capable of reason).

General_Effort,

If everybody says ‘the earth is a square’, he wants AI to come to a different conclusion.

GiveMemes,

Indeed. Because what everybody says is the data it’s trained on and has nothing to do with what people actually know/understand. Kinda like how I can say “the sky is orange”… crazy, right?

General_Effort,

The idea is that a “real” AI should be able to detect implied inconsistencies in the training data and point them out?

GiveMemes,

Indeed. It should be able to reason. Like a human. Not a hard idea to grasp tbh.

General_Effort,

What does “reason like a human” mean to you?

GiveMemes, (edited )

azquotes.com/…/quote-i-shall-not-today-attempt-fu…

I guess I should elaborate a bit. This is from a famous SC court case concerning ‘obscenity’ it’s almost impossible to provide any kind of definition concerning reason or thinking because it’s on the very edge of what we can ever really ‘know’. At the same time I know that if we train something on both the questions and the answers and make it really efficient at giving the right answers, it’s obviously not thinking, just indexing information. A great example is how AI can’t create new information without a seed of absolute randomness. Humans don’t have a random bone in their body.

A fun (though outdated) video series about the edge of the knowable:

youtube.com/playlist?list=PL3096540179B12F8D&si=S…

General_Effort,

You know, Alan Turing describes almost the same problem in 1950, though he talks about defining “thinking”. He was famously good at reasoning and proposed a solution.

GiveMemes,

That’s kinda the whole point of my comment is that things like Turing’s method completely fall apart under heavy scrutiny. Further, the Turing Test specifically tells you nothing about whether or not something IS thinking, just that it MAY be. Big difference.

I see you didn’t engage with the rest of my comment tho. Would you like to?

Just wanted to add this as it and stuff like it comes up pretty quickly when you research the turing test:

“On the other hand, there are several criticisms and limitations of the Turing Test as a measure of machine intelligence. Some of the main issues include:

The test focuses solely on the ability to mimic human-like behavior and communication, rather than on the underlying intelligence or consciousness of the machine.

The test is heavily dependent on the human evaluator’s subjective judgment, and may be influenced by factors such as the machine’s appearance or the human’s own biases.

The test does not take into account the possibility that a machine could be intelligent in ways that are fundamentally different from human intelligence.

The test does not consider the possibility of a machine deceiving the human evaluator, by providing pre-programmed or rehearsed responses rather than truly understanding the meaning of the questions.”

LLMs would fall into the last, as they train on the “answers” so to speak and just match them to the “question”.

General_Effort,

I see you didn’t engage with the rest of my comment tho. Would you like to?

I am not sure if I should. The topic is veering into the spiritual. To me, this is merely a matter of intellectual curiosity. But for many people it is a very emotional subject. I do not wish to cause emotional distress.

GiveMemes,

We both know this has nothing to do with ‘emotional distress’ and everything to do with your overly large ego being bruised by the fact that you’re wrong. It’s classic fallacious behavior to argue as you have and then not engage with the opposition. The only “emotionally distressed” one here is you, and it’s honestly really sad considering it’s an anonymous forum and nobody even knows that it’s you being stupid behind the screen. :/

General_Effort,

Huh. An emotional subject, indeed. I didn’t think merely pointing it out would be enough to trigger you. Sorry for causing you distress. I’m just not good at picking up emotional cues.

GiveMemes,

We can do this all day. The subject isn’t emotional for me at all. Perhaps you’re projecting your own insecurities about the debate onto me?

Like I really don’t understand why you won’t make a point and instead keep acting like an aloof teenager.

General_Effort,

If I were still a teenager, I would not have worried about causing anyone distress. I’ve had many exchanges with people about matters that touch on the religious or spiritual. I’ve come to understand some things. Some people, if they stop voicing the “right” opinion, they will be disowned by their families and shunned by their communities. Other people have specific ideas about life after death. To them, if anything contradicts these ideas, then it’s like they learn that their relatives are dead and they themselves will die soon. To me, all this is just interesting. It seems cruel to expose others to this kind of threat and emotional distress while I’m just sitting here all comfortable. I’m sure it took me way longer than others to understand that.

I don’t know what your situation is. You could have told me not to worry but instead responded rather emotionally. I don’t know what to make of that.

But you want a point. I guess I can do that.


We need to step back and ask how we know things. In science, it’s all about experiments. You try things out. It’s not quite as straight-forward as it seems but we don’t need the details. Another way to know things is a legal system. If you want to know whose property something is, science cannot help you. In case of doubt, you have to go to court and get a judgment. There are lots of other ways but we don’t need to bother.

Obscenity is not a matter for science. There is no experiment which can determine if something is or isn’t obscene. The courts decide and they use no uniform standard.

If reason is like obscenity, then it is for the courts to decide or the law-makers.

GiveMemes,

I really just don’t get why somebody would get emotional over an argument like this but to each their own I suppose. The reason for the emotionality of my reply is rather simply stated: I still don’t believe you had any intent to spare anybody ‘emotional distress’ and were trying to remain aloof and, honestly, rather cunty, by bringing up something literally everybody even mildly interested in AI knows all about as if it’s the end all be all of understanding the potential of thinking arising from a machine. On top of that, you purposefully haven’t engaged with any of the points directly refuting the things you’ve said. Honestly, some of the emotionality comes from when I remember being like you, thinking I knew everything, and whenever somebody would hold me to my words I’d do something along the lines of what you’re doing (engaging in argumentative discussion dishonestly in order to maintain the appearance of ‘winning’ when I really should have been learning more and changing my mind instead of bringing up the same tired pop-culture “smart people” bs.)

Anyway,

My point wasn’t about obscenity. It’s about the nebulousness of something like reason, and the Turing test isn’t scientific in the first place, so I’m really not sure where you got all this ‘science vs law’ bs from.

The point wasn’t that reason is like obscenity, but that I can clearly see, from the way that we train LLMs, that they aren’t reasoning in any form, rather using values that have been derived over time from the training data fed in and the ‘reward’ system used to get the right answers over time. An LLM is no more than a complicated calculator, controlled in many ways by the humans that train it, just as with any form of machine learning. Rather that I “know it when I see it”

I’ve read some studies on ‘game states’ which is the closest that ai scientists have come to anything resembling reason, but even in a model that played the relatively simple game of Othello, the metric they were testing the AI (which was trained on data of legal Othello boardstates) against to ‘prove’ that it was ‘thinking’ (creating game states) was that it was doing better at choosing legal moves than random chance. Another reason it might have been doing better than random chance? Oh yeah… the training data full of legal boardstates. And when the AI was trained on less data? Oh? Would you look at that? The margin by which it beats random chance falls drastically. Almost like the LLM has no fucking clue what’s going on and it’s just matching boardstates… indexing. It doesn’t understand the rules of Othello; it’s just matching piece placement locations with the legal boardstates it was trained on. A human trained on even a few hundred (vs thousands) of such boardstates could likely start to reason out the rules of the game quite easily.

I’m not even against AI or anything, but to call the machine learning that we have now anything close to true, thinking AI is just foolish talk.

General_Effort,

‘science vs law’

There is no versus. These are examples of how we know things. There are other ways of knowing. I chose these, because they were already brought up. You brought up obscenity as a matter of law, and I alluded to Turing.


The “Turing Test” comes from a scientific mindset. Methodology has evolved since then, and Turing was a mathematician; so perhaps not the best at designing experiments. It has features we would expect today: It is controlled and it is blinded. Today, we’d also want a sample size big enough to apply statistics.

We could apply this thinking to “obscenity”. For example, we take a bunch of images and have people rate them as obscene or not. This could be a way for sociologists to learn something about community standards. We could also correlate the results to the subjects’ cultural background, age, education and so on. One could also measure EG physiological arousal.

However, knowing statistically what community members consider obscene is not the same as knowing what is legally obscene (or religiously). If we define obscenity as something that is considered obscene by a certain percentage of a community, then such an experiment would give the answer to what is obscene.

Turing was interested in the question if machines can think. We can approach this experimentally. We let a machine perform a task that is agreed to require thinking. Humans perform the same task as a control. Then we look for differences. This is basically how a typical medical trial works.

Scientifically, the only value of such an experiment would be sociological. It could probe how people construe “thinking”. Learning the results of such an experiment, may change how people construe thinking, which is just how it goes in social science.

In practice, we get methodological problems. We get effects from unblinding, for example. People might form an opinion on which the machine is or the human, and then be guided by bias. When that happens, we can no longer make conclusions about “thinking”. In practice, the test always becomes a test of whether the machine can successfully pass as a human and not whether it can think. Ideally, we want to isolate a single variable. The only factor that should make a difference is whether thinking took place.

Philosophically, one can also see problems. The implied assumption is that “thinking” is a function. If a laptop is playing music, we could not be confident that it was streaming. It might be playing a file, have a radio receiver, … Some people might say that “thinking” requires some component unknown to science, like a soul. If a soulless entity (such as a machine or animal) were to perform the same task, they would just be computing or reacting to stimuli.


So, you’ve brought up a number of things. Saying that a LLM is just a complicated calculator might be saying that some (non-physical?) component is missing.

What the paragraph on Othello is saying is not quite clear to me. Training leading to better performance is consistent with reason?

I think some issues need to be examined a bit more closely. You are interested in whether machines can reason, right? Is that a question that can answered empirically, IE through data, facts, observations and experiment? There must be some observable difference between an entity or being using reason and one that does not.

Perhaps citizenship is a better analogy than obscenity. Citizenship is not a matter of science, yet a legal system can clearly establish the answer. It might be sufficient to inspect documentation. Establishing ethnicity is more difficult. In many cultures, ethnicity and citizenship are connected, but there often is no authoritative way to establish someone’s ethnicity. There even may be no consensus on which observable features are necessary or sufficient.

Basically, what are we looking for?

GiveMemes,

Isn’t this basically just what my comment about the edge of the knowable was and you snarkily replied with the Turing Test?

Like go watch one of the videos I linked if you haven’t. I think they’d be really interesting to you, especially the first one.

I agree with you tho. What are we looking for is the question to ask. By that same notion, I can say with certainty for myself that what we have doesn’t reason, but I can’t elaborate on what it might take to make up something that does. Just as with obscenity in that famous SC case.

To elaborate on the Othello point:

They tested the LLM with a probe and changed a board piece. They used this change and probed the resultant probability distribution to determine whether or not the AI would change its probability distribution to ‘prove’ that it was creating world representations of the board. The problem is, and this is what makes it kinda fallacious thinking by the study authors, that if you change the input data of course the output data is going to change. That’s just a result of training the AI on different legal boardstates, as the way that moves that are made will have a direct result on the placement of the pieces.

Furthermore, they showed that it outperformed random chance at predicting legal moves, but that’s just the way that training AI works. An LLM is better at predicting the next word than random chance as a result of its training.

If you don’t really get what I’m talking about here I recommend this video: m.youtube.com/watch?v=wjZofJX0v4M&vl=en

General_Effort,

Isn’t this basically just what my comment about the edge of the knowable was

No. Not even close.

We know what obscenity is. A court will tell you if something is obscene. End of story.

The problem with the SC quote is, that it is at odds with the rule of law. The meaning of the law must be known. It can’t be whatever some judge feels like. US courts use so-called tests to determine - with as much objectivity as possible - if something is meant by a statute or not. Currently, the “Miller test” is used for obscenity.

No true Scotsman is not an illustration of the edge of the knowable but of irrationality.

To elaborate on the Othello point:

I don’t see what point you are trying to make. A bit of googling leads to this: thegradient.pub/othello/

Is that what you are writing about? You are trying to show that the conclusions are unwarranted? What do you think that would imply?

GiveMemes, (edited )

The legal system has nothing to do with understanding and everything to do with arbitrarily assigned human bullshit (just like the turing test). While law tends to be rational, it’s notoriously shit as a way of understanding the universe. (Live in a fascist country? Well, the law’s the law). I really regret trying to use that quote as an example because you’ve ratcheted onto it like a bulldog and simply can’t let go.

Science is the only way by which we can advance our understanding of the universe. There are cases of unknowable questions in which people use philosophy or religion to try and fill the gap, but they still never actually know, just think.

That wasn’t the exact study I was referencing, but it is actually better at explaining some of the related concepts both in analogy and in their discussion (a discussion in which, they admit that what they think their findings indicate and what their findings actually indicate could be two different things.)

But, to conclude that somehow the multidimensional set of vectors is mapping the board out because when you change part of the input data, even counterfactual input data in which the computer hasn’t seen that move before as it’s illegal, the output data changes is another huge leap. Of course the data changes, as the patterns change, and the gpt has internalized the patterns in its training data, just as it internalizes syntax and rules of language.

I don’t think that it really has any meaningful impact if they were incorrect, but if they are correct it could mean that AI is somehow creating a representation of data within itself, which really also wouldn’t surprise me.

I guess I was more arguing against the guy trying to quote the study at me in the first place than the study itself, though I do have my issues with their analogy bc it’s simply clownish to compare a crow to a mathematical construct purposely created to internalize the rules and syntax of language.

Also that journal has a high schooler on the board of editorialists and has no name for itself… not exactly The Journal of Machine Learning Research lol

General_Effort,

I really regret trying to use that quote as an example

As an example of the unknowable? Perhaps you could elaborate what you feel to be unknowable.

Science is the only way by which we can advance our understanding of the universe.

I’m actually surprised to read that from you. It doesn’t really square with your fairly dismissive attitude toward empiricism.


Apparently, you are sure that GPTs can’t reason. However, you don’t know what reason means. So, IDK how you could possibly know whether anything or anyone is capable of reason.

GiveMemes,

Not an example of the unknowable. An example of knowing that something ‘is not’ without defining what ‘is’.

I have a dismissive attitude towards things like the Turing test because they’re only empirical insofar as they empirically record a subjective opinion.

Similarly, with the Othello study, my problem is not with their data, but what they attempt to extrapolate from it.

In the same way that I can’t define God, I can say with some certainty that you aren’t it. Could I be wrong? Potentially, in an incredibly, incredibly unlikely scenario. Am I willing to take that risk? Yes… and Occams Razor supports such.

General_Effort,

An example of knowing that something ‘is not’ without defining what ‘is’.

So can you define what reason is not?

I have a dismissive attitude towards things like the Turing test because they’re only empirical insofar as they empirically record a subjective opinion.

Then you have a dismissive attitude toward much of science.

Similarly, with the Othello study, my problem is not with their data, but what they attempt to extrapolate from it.

What do they attempt to extrapolate from it?

GiveMemes,

Poor word choice on my part. I can know that something is not without defining what is. See God example.

I definitely have a dismissive attitude towards social sciences. Not actual science tho.

For my criticisms of the Othello study please see my previous comment elaborating on them.

General_Effort,

Poor word choice on my part. I can know that something is not without defining what is. See God example.

So your opinions on reasoning are part of a religious belief?

I definitely have a dismissive attitude towards social sciences. Not actual science tho.

You don’t know what actual science is.

For my criticisms of the Othello study please see my previous comment elaborating on them.

I didn’t ask about your criticisms. I’m still trying to piece together where you are actually going with this.

GiveMemes,

If you could read with a 5th grader’s level of comprehension you would know by now. Feel free to go back. It’s all there. Just put it together.

Never said anything about a religious belief. People are open to hold them, even if irrational.

I’m a literal scientist so I don’t know how that could be possible. It seems like you’re the one that has a problem with extrapolating erroneously from data.

I will not be replying again.

General_Effort,

I’m a literal scientist

This is a very obvious lie.

zephr_c,

This just fundamentally doesn’t understand what artificial general intelligence means. It’s not a fancy way of saying “human but smarter”. That’s just wrong. It’s an artificial intelligence that’s good at a lot of different things. You know. General. Someday, if we live long enough, we will create an AI capable of figuring things out that it wasn’t designed for and we didn’t teach it. Maybe that will be tomorrow. Maybe it’s still centuries away. It’s actually really hard to figure out how long it will take us. Making a really good text generation algorithm doesn’t make the concept of learning more than one thing obsolete though. And predicting what text goes in a bar exam after being given a massive sample of bar exams isn’t the same thing as learning to be a lawyer.

Tech bros with more money than sense suing each other over terms they don’t understand is not a rational system to base research off of, and we should ignore them.

BananaTrifleViolin,

Yeah absolutely. Even AI as a term has become a crock of shit because it’s been latched on to by companies to market their products in the AI the equivalent of the dotcom boom.

Artificial Intelligence was once a sufficient smterm for Artificial General Intelligence. Now any old algorithm is being labelled AI to sell it.

But the terms don’t matter - the concept is sound but it’s further away than we probably expect because so much crap is being sold to make a quick buck.

Chat-GPT is basically beta software and it is practically useless because it’s inaccurate. You can’t use a tool in business, government or health are when it can be wrong and worse so confidently wrong. It’s an impressive tool but they still haven’t got that working well, let alone any further “advances”.

And blindly throwing data at LLMs and hoping to stumble on AGI is not going to work - crudely that is the approach of much of the cow boy outfits out there claiming to be innovating in AI. That includes big tech companies who have jumped on the bandwagon over the last 18 months.

FaceDeer,
@FaceDeer@fedia.io avatar

The term "Artificial Intelligence" is an umbrella term for a wide range of algorithms and techniques that has been in use by the scientific and engineering communities for over half a century. The term was brought into use by the Dartmouth workshop in 1956. It's perfectly applicable to LLMs and other similar generative algorithms being used today, and many less sophisticated ones as well. "Artificial general intelligence" is a subset of AI.

cloudless,
@cloudless@feddit.uk avatar

The article title makes no sense, and the article itself too.

No one is saying GPT has achieved AGI. Is it a strawman argument?

AGI could end up in a similar predicament: a benchmark, devised by humans, that’s rendered obsolete by the technology it was meant to measure.

Just because we can’t measure it doesn’t mean it is obsolete.

  • All
  • Subscribed
  • Moderated
  • Favorites
  • technology@lemmy.world
  • ngwrru68w68
  • rosin
  • GTA5RPClips
  • osvaldo12
  • love
  • Youngstown
  • slotface
  • khanakhh
  • everett
  • kavyap
  • mdbf
  • DreamBathrooms
  • thenastyranch
  • magazineikmin
  • anitta
  • InstantRegret
  • normalnudes
  • tacticalgear
  • cubers
  • ethstaker
  • modclub
  • cisconetworking
  • Durango
  • provamag3
  • tester
  • Leos
  • megavids
  • JUstTest
  • All magazines