Driftline

Independent research lab building evaluation tools for structural reliability in generative images.

Driftline studies structural correctness in generative image systems and builds tools to make visual failures easier to measure, compare, and understand.

Current work focuses on the gap between visual plausibility and structural correctness in diffusion-based image models, using controlled probes, scored image subsets, and evaluation workflows designed to support repeatable review.

Driftline Evaluator

Driftline Evaluator measures structural correctness in generative images by classifying recurring failure patterns and comparing how prompts or workflows actually perform.

Driftline Evaluator is a visual evaluation system for generative images. It is designed to measure structural correctness, classify recurring failure patterns, and compare how prompts or workflows perform under repeatable conditions. Instead of focusing only on style or realism, it looks at whether an image holds together as a believable structure.

Who it’s for

Research

A scored observational note examining chair-generation failures under minimal prompt conditions. In a broader baseline run of approximately 7,000 generated images, 280 chair-class outputs were identified. A manually reviewed illustrative subset of 58 chairs was then scored using Driftline’s Structural Validity Score (SVS), yielding an even split between structurally acceptable and structurally failed outputs.
A structured hand-family comparison testing whether constrained prompt variations improve structural correctness in diffusion-generated hands. Prompt branches including pose cues, numeric wording, and semantic-styling language were reviewed under a locked rubric designed to separate visual plausibility from anatomical correctness.

GitHub

Research repositories, experiment materials, and related project work are available on GitHub.

github.com/driftline-ai/driftline-research

Contact

research@driftline-us.ai