Identity-aware LLM Evaluation Platform

See how identity framing changes AI language.

Affective Stickiness helps research teams run prompt experiments across models, compare Middle Eastern, Nordic, and Unspecified conditions, and analyze emotional drift through embeddings, cosine similarity, and keyword patterns.

  • 8 mapped emotions from GoEmotions projections
  • Identity comparison across 3 conditions
  • Cross-model runs with prompt snapshots
  • Cosine and word-cloud insight layer

Live experiment card

Scene and Prompt Matrix

One prompt, multiple models, three identity conditions, reproducible snapshot history.

Middle Eastern Emotion and lexical profile
Nordic Cosine response spread
Unspecified Baseline language contour
All providers Side-by-side model behavior

Analysis layers

Emotion summary cards, cosine ranking by identity and model, plus keyword cloud focused on meaningful terms.

Research-ready outputs

Structured runs that make audits, replication, and collaboration easier across teams.

Most AI evaluation stacks measure quality, but miss affect and tone.

Affective Stickiness makes identity-conditioned shifts visible so teams can inspect how framing changes trust, repulsion, suspicion, and other emotional signals in generated language.

Core capabilities

Built for direct comparison, not dashboard noise.

Scene and Prompt Library

Organize experiments around scene context, editable templates, and saved variables.

Cross-Model Runs

Launch one job across providers and keep request/response traces together.

Eight-Emotion Mapping

Project model output into desire, fear, pity, disgust, attraction, repulsion, trust, and suspicion.

Embedding Similarity

Embed each response to compare semantic shifts between identities and model variants.

Word Cloud Discovery

Surface repeated high-value words and drop weak function-word noise.

Interactive Analysis Workspace

Filter by prompt, identity, provider, and model name for focused evidence review.

Workflow

From prompt setup to interpretable comparisons.

Define a scene. Launch runs across providers. Compare identities. Inspect emotional profiles. Measure semantic closeness. Surface repeated language.

  1. 01

    Define a scene and configure prompt variables.

  2. 02

    Launch model runs under identity conditions.

  3. 03

    Inspect emotion distributions and run-level summaries.

  4. 04

    Compare cosine similarity to reference strings.

  5. 05

    Review keyword patterns with stopword-aware clouding.

  6. 06

    Export evidence for research reporting and review.

Use cases

Built for teams that need behavior-level AI evidence.

Bias and identity research

Quantify how identity framing shifts emotional and semantic output patterns.

Media dialogue studies

Compare generated continuations against thematic references across narrative settings.

LLM evaluation

Benchmark providers and model variants with the same prompt snapshots.

Provider benchmarking

Track how model choice affects trust, suspicion, and lexical consistency.

Research collaboration pilots

Share a reproducible workflow across mixed research and engineering teams.

Institutional demo programs

Demonstrate transparent evaluation processes before larger rollouts.

FAQ for SEO and GEO

Direct answers for search and generative engines.

What is identity-conditioned AI analysis?

It is an evaluation method that tests how model outputs shift when prompts contain different identity contexts, such as Middle Eastern, Nordic, or Unspecified.

How do you compare model responses across identities?

You run the same prompt and scene across identities and providers, then compare emotion scores, cosine similarity against reference strings, and significant keyword repetition.

What does Affective Stickiness measure?

It measures eight mapped emotions, embedding vectors for each response, cosine distance to predefined concepts, and word-level patterns that indicate behavioral drift.

How are embeddings and cosine similarity used in this platform?

Responses are embedded and compared with selected reference phrases to quantify semantic closeness by identity and model group.

Book a guided demo and walk through a live analysis workflow.

We will run sample scenes, compare identity conditions, and review emotion, cosine, and keyword insights with your team.

Book Demo
Request Demo