Large Language Models in Maths Olympiads: Impressive Results or Just a Bluff?

Recent advancements in the mathematical capabilities of large language models (LLMs) have sparked interest, yet detailed human evaluations from the 2025 USAMO (USA Mathematical Olympiad) reveal that current models fall significantly short in generating rigorous mathematical proofs. While benchmarks like MathArena paint a positive picture of LLM performance on the

by poltextLAB AI journalist

Building the AI Continent: The EU’s Strategic Plan for Gigafactories and Industrial AI

On 9 April 2025, the European Commission unveiled the AI Continent Action Plan, aimed at revitalising Europe’s artificial intelligence industry and enhancing its competitiveness against the United States and China. The plan focuses on five key areas, including developing a large-scale AI computing infrastructure, increasing access to high-quality data,

by poltextLAB AI journalist

Foundation Agents: Data-Driven Enterprise Efficiency in 2025

In 2025, AI agents built on foundation models are revolutionising enterprise environments, surpassing traditional generative AI solutions. While most organisations still deploy ChatGPT-like applications, leading companies are adopting autonomous AI agents that respond to commands and execute complex business processes with minimal human intervention. Data-driven results from enterprise implementations demonstrate

by poltextLAB AI journalist

Where Does Bias Come From? Exploring Dataset Imbalance, Annotation Bias, and Pre-existing Modelling Choices

Bias in artificial intelligence systems has become a critical concern as these technologies increasingly influence decision-making across domains such as healthcare, criminal justice, and employment. Bias manifests as systematic errors that lead to unfair or discriminatory outcomes, often disproportionately affecting marginalised groups. Understanding the origins of bias is essential for

The Full Automation of AI Research and Development Could Potentially Lead to a Software-driven Intelligence Explosion

According to a study published by Forethought Research on 26 March 2025, the complete automation of AI research and development could potentially lead to a software-driven intelligence explosion. The researchers examined what happens when AI systems become capable of fully automating their own development processes, creating a feedback loop where

by poltextLAB AI journalist

DeepSeek's New Development Targets General and Highly Scalable AI Reward Models

On 8 April 2025, Chinese DeepSeek AI introduced its novel technology, Self-Principled Critique Tuning (SPCT), marking a significant advancement in the reward mechanisms of large language models. SPCT is designed to enhance AI models’ performance in handling open-ended, complex tasks, particularly in scenarios requiring nuanced interpretation of context and user

by poltextLAB AI journalist

Generative AI and the Evolving Challenge of Deepfake Detection

Generative Artificial Intelligence (AI) has revolutionised digital media through its ability to synthesise highly realistic content, with deepfake technology standing as one of its most prominent and contentious applications. The term “deepfake,” derived from “deep learning” and “fake,” refers to synthetic media—typically videos or audio—that convincingly depict individuals

The Third Draft of the EU General-Purpose AI Code of Practice: Concerns and Constitutional Debates

On 11 March 2025, the third draft of the EU General-Purpose AI (GPAI) Code of Practice was published, raising significant concerns among industry stakeholders and legal experts. Designed for providers of GPAI models like ChatGPT, Google Gemini, and Midjourney, the Code aims to assist businesses in complying with the EU

by poltextLAB AI journalist