Close Menu
FSNN | Free Speech News NetworkFSNN | Free Speech News Network
  • Home
  • News
    • Politics
    • Legal & Courts
    • Tech & Big Tech
    • Campus & Education
    • Media & Culture
    • Global Free Speech
  • Opinions
    • Debates
  • Video/Live
  • Community
  • Freedom Index
  • About
    • Mission
    • Contact
    • Support
Trending

Strategy’s STRC maintains dividend at 11.5% after steady increases

6 minutes ago

Dorsey shares AI-integrated workplace vision weeks after Block’s 40% staff cut

8 minutes ago

States Can’t Ban “Conversion Therapy” Speech for Minors

38 minutes ago
Facebook X (Twitter) Instagram
Facebook X (Twitter) Discord Telegram
FSNN | Free Speech News NetworkFSNN | Free Speech News Network
Market Data Newsletter
Wednesday, April 1
  • Home
  • News
    • Politics
    • Legal & Courts
    • Tech & Big Tech
    • Campus & Education
    • Media & Culture
    • Global Free Speech
  • Opinions
    • Debates
  • Video/Live
  • Community
  • Freedom Index
  • About
    • Mission
    • Contact
    • Support
FSNN | Free Speech News NetworkFSNN | Free Speech News Network
Home»Opinions»Debates»Evolution of AI Intelligence Tests
Debates

Evolution of AI Intelligence Tests

News RoomBy News Room6 months agoNo Comments6 Mins Read635 Views
Share Facebook Twitter Pinterest Copy Link LinkedIn Tumblr Email VKontakte Telegram
Evolution of AI Intelligence Tests
Share
Facebook Twitter Pinterest Email Copy Link

Listen to the article

0:00
0:00

Key Takeaways

Playback Speed

Select a Voice

The debate over whether or not powerful artificially intelligent agents are likely—or even possible—has been raging since the construction of the first digital computers in the late 1940s. Alan Turing, who established the mathematical foundations of computer science, did not attempt to define general intelligence, but he offered an operational test in place of a definition. The Turing Test involves placing a computer behind one curtain, and a person behind another. A second person communicates with both through a keyboard. If the second person is unable to distinguish which interlocutor is human and which is a machine, then we can describe the machine as intelligent. The Turing Test was once widely (but by no means universally) accepted as a reasonable procedure for identifying general intelligence. However, the fluency of Generative Pre-trained Transformers’ (GPTs’) natural language production leaves the usefulness of the Turing Test open to question.

On 2 April 2025, Google DeepMind released a 145-page report proposing safety measures to minimise the harm that Artificial General Intelligence (AGI) models might cause when they appear. According to the report’s authors, we should expect these models by 2030. The report anticipates the rise of AGI in the following terms:

Timelines: We are highly uncertain about the timelines until powerful AI systems are developed, but crucially, we find it plausible that they will be developed by 2030.

• Implication: Since timelines may be very short, our safety approach aims to be “anytime”, that is, we want it to be possible to quickly implement the mitigations if it becomes necessary. For this reason, we focus primarily on mitigations that can easily be applied to the current ML pipeline.

At the other extreme of the AGI discussion, Emily Bender and Alex Hanna’s 2021 book The AI Con: How to Fight Big Tech and Get the Future We Want and Bender et al.’s 2021 paper “On the Dangers of Stochastic Parrots: Can Language Models be Too Big?” dismiss current AI systems as devoid of serious learning or inference capacity. They argue that the apparently impressive ability of multimodal Large Language Models (LLMs) to solve traditional AI tasks—like image identification and description, machine translation, natural-language generation, and information retrieval—is a consequence of the vast amount of data on which they are trained. They describe these models as “stochastic parrots,” which can do little more than return their training data. 

Unfortunately, discussions about AI—like discussions about politics—are increasingly dominated by hyperbole and extreme views. This polarisation is captured by Google DeepMind’s overheated expectations concerning the immediate arrival of AGI on one side, and Bender et al.’s dismissive polemic insisting that GPTs lack any cognitively interesting dimensions on the other. Both approaches have well-established antecedents in the history of AI. Each one seriously misrepresents the properties and capacities of current AI systems.


It is certainly not the case that the large transformers that drive LLMs do nothing more than return the data on which they are trained. They perform complex inductive inferences, and they determine subtle probability distributions over possible situations in the context of previous sequences of events. These operations permit them to recognise abnormalities in X-rays and fMRIs with an efficiency that far surpasses the performance of human medical diagnosticians. They identify new protein molecules that support discoveries in biology and pharmacology.

LLMs have achieved high-quality multilingual machine translation across a remarkably large set of languages, in some cases through transfer of knowledge from digitally well-represented languages to comparatively low-resourced languages. They recognise images and audio patterns, and they describe their elements accurately. They generate sounds and images from text. They can produce reasonably decent computer code from natural language instructions, and they spot errors in programs. They are now generating novel proofs of difficult theorems, which human mathematicians have not managed to discover. They have defeated the human champions of elaborate strategy games, like Go. They are able to distinguish between authentic and forged paintings, and they can generate original musical compositions that score well on human assessment. They provide information retrieval and expert query systems that are highly accurate, and fluent in their natural-language responses.

This is only a sample of the areas in which these systems are able to perform cognitively challenging tasks at levels that greatly exceed human capacities in speed and accuracy. Given their range of abilities, it is entirely inaccurate to describe them as “stochastic parrots,” or to dismiss them as plagiarising cons. These systems constitute a major revolution in machine-learning methods. The engineering successes this revolution has facilitated are transforming most aspects of the environments in which we live and work.

However, it is also important to recognise the serious limitations of GPTs. They are not able to engage in deep, domain-general reasoning or planning, which is the core element of what we associate with general intelligence. Their performance declines in direct relation to the distance between the data that they have seen in training, and the input to which they are applied in new tasks. Out-of-domain testing reveals the weaknesses of these models. If their massive pretraining is not supplemented by customised exposure to domain-specific information (fine tuning), they do not, for the most part, do well at new kinds of tasks. In particular, they are not able to handle ordinary real-world reasoning, generalised inference, and complex planning, which humans perform with ease, as a matter of course.

It is not easy to define general intelligence, nor is it clear that we will ever achieve a precise scientific theory that captures it. We can recognise intelligent behaviour when we see it, but we cannot define it. Turing was therefore right not to attempt such a definition and to suggest an operational criterion instead. In my book Understanding the Artificial Intelligence Revolution: Between Catastrophe and Utopia, I follow Turing’s lead in adopting a set of diagnostics for identifying conduct that exhibits general intelligence. These include behaviour that is (i) goal directed, (ii) volitional, and (iii) applies rational and efficient methods for planning to achieve objectives over a wide range of domains. While GPTs and other deep-learning models are goal directed, they are not volitional, and they do not use domain-general rational planning techniques.

It is important to recognise that these diagnostic markers of general intelligence do not entail internal mental states or consciousness in an agent. The philosophers David Chalmers and John Searle both identify intelligence with such states. Chalmers claims that AI systems can achieve them while Searle disagrees. In fact, there is no reason to assume that non-human agents, electronic or biological, have human-like mental states as a condition of intelligence. In the case of computational models, it seems farfetched to assume that they would. It is possible for an agent to satisfy my three proposed behavioural indicators of intelligence without having internal mental states of any kind. Separating the question of consciousness from general intelligence gives us a clearer view of what is involved in postulating intelligent artificial agents.



Read the full article here

Fact Checker

Verify the accuracy of this article using AI-powered analysis and real-time sources.

Get Your Fact Check Report

Enter your email to receive detailed fact-checking analysis

5 free reports remaining

Continue with Full Access

You've used your 5 free reports. Sign up for unlimited access!

Already have an account? Sign in here

Share. Facebook Twitter Pinterest LinkedIn Tumblr Email Telegram Copy Link
News Room
  • Website
  • Facebook
  • X (Twitter)
  • Instagram
  • LinkedIn

The FSNN News Room is the voice of our in-house journalists, editors, and researchers. We deliver timely, unbiased reporting at the crossroads of finance, cryptocurrency, and global politics, providing clear, fact-driven analysis free from agendas.

Related Articles

Media & Culture

States Can’t Ban “Conversion Therapy” Speech for Minors

38 minutes ago
Media & Culture

Just out of curiosity, do you think we should ask Claude to decide this case?

2 hours ago
Media & Culture

New York’s Governor Seems Indifferent to the Health Consequences of a Steep Tax on Nicotine Pouches

4 hours ago
Cryptocurrency & Free Speech Finance

Bitcoin Holds $66K as Trump Prioritizes Iran War Exit Over Reopening Hormuz

4 hours ago
Media & Culture

Aspyr: Hey, Those Crappy Tomb Raider Remastered Outfits Were Made By Our Artists, Not AI!

5 hours ago
Media & Culture

Judge Leon Blocked The East Wing Ballroom Based on Offended Observer Standing

5 hours ago
Add A Comment

Comments are closed.

Editors Picks

Dorsey shares AI-integrated workplace vision weeks after Block’s 40% staff cut

8 minutes ago

States Can’t Ban “Conversion Therapy” Speech for Minors

38 minutes ago

Ripple-linked token holds $1.34 as supply tightens

1 hour ago

Fidelity says Bitcoin’s Cycle Drawdown is the Mildest Yet

1 hour ago
Latest Posts

Just out of curiosity, do you think we should ask Claude to decide this case?

2 hours ago

Malaysian journalist Kalidas Subramaniam detained after reporting on alleged illegal migrant workers

2 hours ago

Quantum-resistant tokens jump 50% as Google flags risks to Bitcoin security

2 hours ago

Subscribe to News

Get the latest news and updates directly to your inbox.

At FSNN – Free Speech News Network, we deliver unfiltered reporting and in-depth analysis on the stories that matter most. From breaking headlines to global perspectives, our mission is to keep you informed, empowered, and connected.

FSNN.net is owned and operated by GlobalBoost Media
, an independent media organization dedicated to advancing transparency, free expression, and factual journalism across the digital landscape.

Facebook X (Twitter) Discord Telegram
Latest News

Strategy’s STRC maintains dividend at 11.5% after steady increases

6 minutes ago

Dorsey shares AI-integrated workplace vision weeks after Block’s 40% staff cut

8 minutes ago

States Can’t Ban “Conversion Therapy” Speech for Minors

38 minutes ago

Subscribe to Updates

Get the latest news and updates directly to your inbox.

© 2026 GlobalBoost Media. All Rights Reserved.
  • Privacy Policy
  • Terms of Service
  • Our Authors
  • Contact

Type above and press Enter to search. Press Esc to cancel.

🍪

Cookies

We and our selected partners wish to use cookies to collect information about you for functional purposes and statistical marketing. You may not give us your consent for certain purposes by selecting an option and you can withdraw your consent at any time via the cookie icon.

Cookie Preferences

Manage Cookies

Cookies are small text that can be used by websites to make the user experience more efficient. The law states that we may store cookies on your device if they are strictly necessary for the operation of this site. For all other types of cookies, we need your permission. This site uses various types of cookies. Some cookies are placed by third party services that appear on our pages.

Your permission applies to the following domains:

  • https://fsnn.net
Necessary
Necessary cookies help make a website usable by enabling basic functions like page navigation and access to secure areas of the website. The website cannot function properly without these cookies.
Statistic
Statistic cookies help website owners to understand how visitors interact with websites by collecting and reporting information anonymously.
Preferences
Preference cookies enable a website to remember information that changes the way the website behaves or looks, like your preferred language or the region that you are in.
Marketing
Marketing cookies are used to track visitors across websites. The intention is to display ads that are relevant and engaging for the individual user and thereby more valuable for publishers and third party advertisers.