research results

research results

MIT Withdrew Student's AI Productivity Study Based on Questionable Data

MIT has formally repudiated an AI research paper by a former economics doctoral student that claimed productivity benefits of artificial intelligence, citing data integrity concerns on 17 May 2025. The paper titled "Artificial Intelligence, Scientific Discovery, and Product Innovation," written by Aidan Toner-Rodgers, was initially praised by prominent

by poltextLAB AI journalist

Large Language Models in Maths Olympiads: Impressive Results or Just a Bluff?

Recent advancements in the mathematical capabilities of large language models (LLMs) have sparked interest, yet detailed human evaluations from the 2025 USAMO (USA Mathematical Olympiad) reveal that current models fall significantly short in generating rigorous mathematical proofs. While benchmarks like MathArena paint a positive picture of LLM performance on the

by poltextLAB AI journalist

Foundation Agents: Data-Driven Enterprise Efficiency in 2025

In 2025, AI agents built on foundation models are revolutionising enterprise environments, surpassing traditional generative AI solutions. While most organisations still deploy ChatGPT-like applications, leading companies are adopting autonomous AI agents that respond to commands and execute complex business processes with minimal human intervention. Data-driven results from enterprise implementations demonstrate

by poltextLAB AI journalist

The Full Automation of AI Research and Development Could Potentially Lead to a Software-driven Intelligence Explosion

According to a study published by Forethought Research on 26 March 2025, the complete automation of AI research and development could potentially lead to a software-driven intelligence explosion. The researchers examined what happens when AI systems become capable of fully automating their own development processes, creating a feedback loop where

by poltextLAB AI journalist

DeepSeek's New Development Targets General and Highly Scalable AI Reward Models

On 8 April 2025, Chinese DeepSeek AI introduced its novel technology, Self-Principled Critique Tuning (SPCT), marking a significant advancement in the reward mechanisms of large language models. SPCT is designed to enhance AI models’ performance in handling open-ended, complex tasks, particularly in scenarios requiring nuanced interpretation of context and user

by poltextLAB AI journalist

Researchers from Hungary’s Semmelweis University Demonstrated the Outstanding Accuracy of GPT-4o in Identifying Skin Diseases

In a study published on 8 April 2025, researchers from Semmelweis University demonstrated that OpenAI’s GPT-4o model achieved a 93% accuracy rate in identifying acne and rosacea, while Google’s Gemini Flash 2.0 model correctly identified these skin conditions in only 21% of cases. The scientific study used

by poltextLAB AI journalist

There is No Evidence of a Significant AI Impact on Elections—the Lack of Transparency Hinders Research

There is currently insufficient data on the impact of artificial intelligence on elections to draw well-founded conclusions, while initial threat predictions have proven exaggerated. Researchers from the NYU Center for Social Media and Politics identified only 71 instances of AI use in election-related communication in 2024. Purdue University researchers documented

by poltextLAB AI journalist

Corpus Size vs Quality: New Research on the Efficiency of Hungarian Language Models

Hungarian language technology research has reached a significant milestone: a comprehensive study has revealed that a larger corpus size does not necessarily lead to improved performance in morphological analysis. In their study, Andrea Dömötör, Balázs Indig, and Dávid Márk Nemeskey conducted a detailed analysis of three Hungarian-language corpora of varying

by poltextLAB AI journalist

Context Sensitivity of the huBERT Model in Pragmatic Annotation – New Research Findings

Tibor Szécsényi and Nándor Virág, researchers at the University of Szeged, have explored the context sensitivity of the huBERT language model in pragmatic annotation, focusing in particular on the automatic identification of imperative verb functions. Their study, conducted on the MedCollect corpus—a dataset of health-related misinformation—investigates how both

by poltextLAB AI journalist

Stanford Innovation in Hypothesis Validation: The POPPER Framework

Researchers at Stanford University unveiled POPPER on 20th February 2025, an automated AI framework that revolutionises hypothesis validation and accelerates scientific discoveries tenfold. Following Karl Popper's principle of falsifiability, POPPER (Automated Hypothesis Validation with Agentic Sequential Falsifications) employs two specialised AI agents: the experiment design agent and the

by poltextLAB AI journalist

Thought-Controlled Typing? Meta's New Brain-to-Text Decoding Tool Directly Converts Brain Signals into Text

Meta researchers have achieved a breakthrough in the brain-computer interface field: they have developed a new tool capable of converting brain signals into text. In the research published in February 2025, they examined the brain's language production using magnetoencephalography (MEG) and electroencephalography (EEG) with the participation of 35

by poltextLAB AI journalist