prompt library

prompt library

Prompt-Driven Video Analysis of Animal Behaviour Using Gemini 2.5 Pro in Google AI Studio and via API

Can state-of-the-art multimodal models analyse animal behaviour directly from video footage? In this study, we tested Google’s Gemini 2.5 Pro — both in AI Studio and via its API — to assess whether it can produce structured ethological descriptions based purely on short animal-related videos. By applying a consistent prompt

Prompt-Based Hamster Ovary Cell Segmentation with OpenAI o4-mini-high on Microscopy Images

Recent advancements in multimodal language models have opened new avenues for analysing scientific image data using natural language instructions. In this post, we explore the capabilities of OpenAI’s o4-mini-high model for performing cell segmentation tasks on microscopy images through prompt-based interaction. Rather than relying on traditional computer vision techniques

Zero-Shot Distal Fibula Fracture Detection: Gemini 2.5 Flash Delivers Spot-On Results on German Radiology Reports

Identifying specific clinical findings in unstructured medical texts is a common challenge in healthcare data science. In this post, we benchmark Google’s Gemini 2.5 Flash language model on a zero-shot classification task: detecting the presence or absence of distal fibula fractures in real-world German radiology reports. Without any

Prompt-Based Disease Mention Extraction with DeepSeek-V3: A Biomedical NER Case Study on a Structured NCBI Test Set

Prompt-based methods are becoming increasingly relevant in biomedical text mining, offering flexible ways to perform tasks such as named entity recognition without explicit model training. In this case study, we assess the performance of DeepSeek-V3 on a structured disease mention extraction task using a curated subset of the NCBI Disease

Benchmarking GenAI Models for Penguin Species Prediction: Grok 3, DeepSeek-V3, and Qwen2.5-Max Delivered Top Results

How well can today’s leading GenAI models classify real-world biodiversity data—without bespoke code or traditional machine learning pipelines? In this study, we benchmarked a range of large language models on the task of predicting penguin species from tabular ecological measurements, including both numerical and categorical variables. Using a

Argania Detection from Sentinel-2 Spectral Data: DeepSeek-V3 Excels with Prompt-Based Labelling of Structured Data

How far can today’s large language models go in scientific data analysis—without bespoke coding or deep learning pipelines? In this experiment, we explore the ability of DeepSeek-V3 to perform pixel-level detection of Argania trees (i.e., binary classification for each pixel) using only tabular Sentinel-2 spectral data. By

Linking Headlines to Article Bodies for Stance Detection: A Structured Pre-processing Workflow Using GPT-4o

Working with real-world text data often means dealing with structures that are not yet analysis-ready. In our case, the dataset included headlines and full article texts stored separately, across two different tables. The only link between them was a shared identifier field: Body ID. Before we could begin any further

Automating Plant Disease Detection at Scale: From Prompt Limitations to a High-Accuracy API Workflow with GPT-4o

Image-based classification is increasingly used across biology, ecology, and agriculture—from identifying animal species to detecting plant diseases. One common use case is the analysis of leaf images to distinguish between healthy and diseased plants. In this post, we compare two approaches to classifying strawberry leaves as either fresh (healthy)

No-Code Transformation of the NCBI Disease Corpus into a Structured CSV

Working with biomedical corpora often requires programming skills, specialised formats, and time-consuming preprocessing. But what if you could transform a complex annotated dataset—like the NCBI Disease Corpus—into a structured, analysis-ready CSV using nothing more than a single, well-designed prompt? In this post, we demonstrate how a no-code, GenAI-powered

Comparing the FutureHouse Platform’s Falcon Agent and OpenAI’s o3 for Literature Search on Machine Coding for the Comparative Agendas Project

Having previously explored the FutureHouse Platform’s agents in tasks such as identifying tailor-made laws and generating a literature review on legislative backsliding, we now directly compare its Falcon agent and OpenAI’s o3. Our aim was to assess their performance on a focused literature search task: compiling a ranked