OpenAI o3

Comparing the FutureHouse Platform’s Falcon Agent and OpenAI’s o3 for Literature Search on Machine Coding for the Comparative Agendas Project

Having previously explored the FutureHouse Platform’s agents in tasks such as identifying tailor-made laws and generating a literature review on legislative backsliding, we now directly compare its Falcon agent and OpenAI’s o3. Our aim was to assess their performance on a focused literature search task: compiling a ranked

Using GenAI for Critical Discourse Analysis: OpenAI o3 vs. Fairclough on Tony Blair’s 1996 Stakeholder Economy Speech

Can a generative AI model perform critical discourse analysis at a level comparable to a trained scholar? This case study puts OpenAI’s o3 model to the test by asking it to analyse Tony Blair’s 1996 “stakeholder economy” speech—a foundational text in New Labour rhetoric. The model'

Can a Philosopher Be Simulated? Testing AI Argumentation on Nozick’s Experience Machine

Can a language model reason like a philosopher—not just mimic their tone, but follow and evaluate arguments with analytic precision? To explore this, we designed a two-step prompt experiment using Robert Nozick’s famous “Experience Machine” thought experiment. First, we instructed OpenAI’s o3 model to adopt the persona

Structured Explanation and Critical Reading of Hegel’s Philosophy of Right Preface: A Persona Pattern Prompt Tested on OpenAI o3

This case study examines the capabilities of OpenAI’s o3 model in interpreting and critically analysing classical philosophical texts. As of today, it is the most advanced language model released by OpenAI in terms of reasoning performance, and is particularly well-suited to delivering deeper comprehension and structured explanation of complex