Structural Compactness as a Complementary Criterion for Explanation Quality
Hey there, little explorer! Imagine a robot friend who tries to tell you why it thinks something.
Sometimes, the robot's explanation is like a messy pile of LEGOs, all spread out and hard to understand. 🧱
This new idea is like a special toy detective tool! 🕵️♀️ It helps us check if the robot's LEGO explanation is neat and tidy, like a small, strong tower. We want the important LEGOs close together, not all over the floor.
This helps us find the best robot friends who can explain things super clearly, like telling you why a cat says "meow" in a way that makes perfect sense! 🐱✨
arXiv:2603.29491v1 Announce Type: new Abstract: In the evaluation of attribution quality, the quantitative assessment of explanation legibility is particularly difficult, as it is influenced by varying shapes and internal organization of attributions not captured by simple statistics. To address this issue, we introduce Minimum Spanning Tree Compactness (MST-C), a graph-based structural metric that captures higher-order geometric properties of attributions, such as spread and cohesion. These components are combined into a single score that evaluates compactness, favoring attributions with salient points spread across a small area and spatially organized into few but cohesive clusters. We show that MST-C reliably distinguishes between explanation methods, exposes fundamental structural diff
View PDF HTML (experimental)
Abstract:In the evaluation of attribution quality, the quantitative assessment of explanation legibility is particularly difficult, as it is influenced by varying shapes and internal organization of attributions not captured by simple statistics. To address this issue, we introduce Minimum Spanning Tree Compactness (MST-C), a graph-based structural metric that captures higher-order geometric properties of attributions, such as spread and cohesion. These components are combined into a single score that evaluates compactness, favoring attributions with salient points spread across a small area and spatially organized into few but cohesive clusters. We show that MST-C reliably distinguishes between explanation methods, exposes fundamental structural differences between models, and provides a robust, self-contained diagnostic for explanation compactness that complements existing notions of attribution complexity.
Subjects:
Artificial Intelligence (cs.AI)
Cite as: arXiv:2603.29491 [cs.AI]
(or arXiv:2603.29491v1 [cs.AI] for this version)
https://doi.org/10.48550/arXiv.2603.29491
arXiv-issued DOI via DataCite (pending registration)
Submission history
From: Mohammad Mahdi Mesgari [view email] [v1] Tue, 31 Mar 2026 09:36:52 UTC (23,903 KB)
Sign in to highlight and annotate this article

Conversation starters
Daily AI Digest
Get the top 5 AI stories delivered to your inbox every morning.
More about
modelannouncevaluation
The Silicon Protocol: The Model Hosting Decision — When Azure OpenAI Isn’t Enough (And When It’s…
The Silicon Protocol: The Model Hosting Decision — When Azure OpenAI Isn’t Enough (And When It’s Overkill) The $187K infrastructure decision every healthcare CTO makes is wrong. Here’s the actual math on self-hosted vs. API vs. hybrid LLM deployment. The three hosting patterns healthcare organizations choose — and what each actually costs at scale. Most start left (API), graduate to center (Hybrid), few need right (Self-hosted). The compliance officer approved your de-identification pipeline. Legal signed off on the BAA. Security validated your OAuth tokens are governed by proper Passports. You’re ready to deploy your first production LLM in healthcare. Then engineering asks the question that determines whether you spend $50K or $250K this year: “Where does the model actually run?” Most he

AI slop got better, so now maintainers have more work
Once AI bug reports become plausible, someone still has to verify them If AI does more of the work but humans still have to check it, you need more reviewers. Now that AI models have gotten better at writing and evaluating code, open-source projects find themselves overwhelmed with the too-good-to-ignore output.…
Knowledge Map
Connected Articles — Knowledge Graph
This article is connected to other articles through shared AI topics and tags.
More in Models

AMD's AI director slams Claude Code for becoming dumber and lazier since last update
'Claude cannot be trusted to perform complex engineering tasks' according to GitHub ticket If you've noticed Claude Code's performance degrading to the point where you find you don't trust it to handle complicated tasks anymore, you're not alone.…



Discussion
Sign in to join the discussion
No comments yet — be the first to share your thoughts!