<?xml version="1.0" encoding="utf-8" standalone="yes"?>
<rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom">
  <channel>
    <title>Metric on ViCoS Lab</title>
    <link>/tags/metric/</link>
    <description>Recent content in Metric on ViCoS Lab</description>
    <generator>Hugo</generator>
    <language>en-us</language>
    <atom:link href="/tags/metric/index.xml" rel="self" type="application/rss+xml" />
    <item>
      <title>Introducing DIAD: A Novel Metric for Assessing the Difficulty of Anomaly Detection Problems</title>
      <link>/publications/pahor2025diad/</link>
      <pubDate>Mon, 01 Jan 0001 00:00:00 +0000</pubDate>
      <guid>/publications/pahor2025diad/</guid>
      <description>&lt;p&gt;Assessing the complexity of anomaly detection tasks is essential for benchmarking datasets and models as well as for understanding the problem domains. While numerous anomaly detection methods have been developed, there remains a need for a simple, learning-free metric to estimate the difficulty of a given anomaly detection task. In this paper, we introduce DIAD (Difficulty Index for Anomaly Detection), a lightweight metric designed to quantify task difficulty without requiring model training or inference. DIAD builds upon and extends the recently proposed AD3 metric by incorporating both the saliency of anomalies and the heterogeneity of normal appearance across the dataset. We evaluate DIAD on five widely used visual anomaly detection datasets and compare its scores with the observed performance of three state-of-the-art detection models. Results show that DIAD correlates more consistently with model performance than AD3, offering a practical and interpretable tool for assessing the complexity of anomaly detection problems.&lt;/p&gt;</description>
    </item>
  </channel>
</rss>
