📢 Announcement
VISIONxFAIL spotlights negative results, failure analysis, and real‑world constraints across CV/Graphics/Multimodal systems. We promote a culture of transparent reporting to accelerate scientific progress.
Website: lc4workshop.github.io/visionxfail.html • Estimated attendees: 20–30
🎯 Overview
Despite benchmark gains, many methods underperform in practice due to dataset biases, domain shifts, missing modalities, annotation noise, or unrealistic assumptions—yet such negative results are rarely documented. This workshop emphasizes: (1) understanding how constraints (data scarcity, supervision limits, compute bottlenecks) affect pipelines; (2) highlighting negative results in generative AI and vision systems; (3) documenting failure cases in cross‑domain, cross‑lingual, and cross‑modal applications; and (4) promoting transparent reporting of successes and failures.
- What constraints prevent vision/graphics systems from scaling robustly?
- Which transfer/adaptation strategies have failed in real deployments?
- How do cross‑modal alignments break under noise, missing data, or adversarial conditions?
- How can documenting negative results prevent repeated mistakes?
- Can we design evaluation frameworks that reward honest reporting of failures?
🧩 Topics of Interest
- Negative results in detection, segmentation, tracking, recognition.
- Failure analysis of ViTs, diffusion, GANs, and multimodal foundation models.
- Constraints in cross‑modal & cross‑lingual tasks: OCR, scene text, multilingual captioning.
- Understanding gaps and results in font processing and analysis.
- Challenges in video understanding: action recognition, scene graphs, multi‑object tracking under constraints.
- Annotation bottlenecks: bias, low inter‑rater reliability, annotation drift.
- Graphics negative results: rendering, simulation, visual realism under limited resources.
- Metrics that fail to capture robustness, fairness, or completeness.
- Cross‑domain failures: adaptation/generalization gaps across datasets, genres, modalities.
- Agentic workflows that failed due to latency, supervision gaps, or domain shifts.
- Bias, fairness, and failure amplification in low‑resource vision settings.
Invited Speakers: TBA
🗓️ Tentative Program (Half‑day)
Mode: In‑person planned; remote talks allowed. If the main conference goes fully virtual, the workshop will be virtual‑only.
📝 Submission & Review
- Submission type: As per main conference guidelines.
- Review process: Double‑blind with at least 2–3 reviews per submission.
- Criteria: Relevance to theme, technical novelty/soundness, clarity, and potential for impact/discussion.
- Management: Reviews coordinated via EasyChair (or equivalent) in alignment with main conference standards.
📚 Selected References
- Borji (2017). Negative Results in Computer Vision: A Perspective. arXiv:1705.04402.
- Zhang et al. (2014). Predicting Failures of Vision Systems. CVPR.
- Northcutt et al. (2021). Pervasive Label Errors Destabilize ML Benchmarks. arXiv:2103.14749.
- Hosseini et al. (2017). CNNs Are Not Invariant to Negative Images. arXiv:1703.06857.
- Lee et al. (2017). Towards Qualitative Analysis of ImageNet Classification Failures. arXiv:1709.03439.
- Zendel et al. (2017). Analyzing CV Datasets via Subspace Alignment & HAZOP. CVPR.
- Balayn et al. (2023). Handling Failures in DL‑based CV Systems. TiiS 13(2).
- Torralba & Efros (2011). Unbiased Look at Dataset Bias. CVPR.
- Recht et al. (2019). Do ImageNet Classifiers Generalize to ImageNet? ICML.
- Geirhos et al. (2018). Generalisation in Deep Learning. ICLR.