Skip to content
  • Official announcement from CSPaper.org

    1 1
    1 Topics
    1 Posts
    rootR
    We’re excited to introduce a new category on cspaper.org: Anonymous Sharing & Supplementary Materials Purpose This category is designed specifically for anonymized sharing of supplementary materials such as: Additional experimental results or figures Extended ablation studies Links to datasets or demos Supplementary explanations that didn’t fit in the main paper It’s especially useful during rebuttals, when you may need to share extra content with reviewers but remain compliant with anonymity and strict page limits. How It Works Only unverified (but registered) users can post in this category: make sure to skip filling your email during registration. You can edit or delete your post anytime Use a username that doesn’t reveal your identity Share the link to your anonymous supplementary materials post/topic This keeps everything in line with double-blind peer review policies while giving you a legitimate way to share supporting materials. ️ Stay Anonymous Please do not use real names or affiliations when posting. If you're unsure how to create an anonymous username, try one of these generators: Jimpix Random Username Generator UsernameGenerator.com NordPass Username Generator Dashlane Username Generator Example Use Case You're writing the rebuttal for your paper submitted and reviewed by NeurIPS/ICML/ACL. You need to trim down your rebuttal to fit the rebuttal length limit. But, you want to share detailed results or code (causing violation of rebuttal length) with reviewers during the rebuttal. Just post them anonymously in that category and include the link in your response! Pro Tip you can make your url even shorter by leveraging services like TinyURL and Bitly. Start Posting Visit the category here and share your materials: https://cspaper.org/category/10/anonymous-sharing-supplementary-materials
  • 30 Topics
    63 Posts
    cqsyfC
    Great points! OpenAI’s new PaperBench shows how tough reproducibility still is in ML. It asked AI agents to replicate 20 ICML 2024 papers from scratch. Even the best model only got 21%, while human PhDs reached 41.4%. [image: 1743714483369-screenshot-2025-04-03-at-23.07.45-resized.png] What stood out is how they worked with authors to define 8,000+ fine-grained tasks for scoring. It shows we need better structure, clearer standards, and possibly LLM-assisted tools (like their JudgeEval) to assess reproducibility at scale. Maybe it’s time to build structured reproducibility checks into peer review, i.e., tools like PaperBench give us a way forward. Checkout the Github: https://github.com/openai/preparedness
  • Anonymously share data, results, or materials. Useful for rebuttals, blind submissions and more. Only unverified users can post (and edit or delete anytime afterwards).

    2 2
    2 Topics
    2 Posts
    M
    Table 1. Performance Comparison on Benchmark Datasets Method CIFAR-10 (Acc %) CIFAR-100 (Acc %) TinyImageNet (Acc %) Params (M) FLOPs (G) ResNet-18 94.5 76.3 64.1 11.2 1.8 ViT-Small 95.2 77.9 65.7 21.7 4.6 Ours (GraphFormer) 96.1 79.5 67.3 19.8 3.9 Table 2. Ablation Study on Temporal Encoding Method Variant CIFAR-10 (Acc %) TinyImageNet (Acc %) Ours (No Time Encoding) 95.4 66.1 Ours (Sinusoidal Only) 95.8 66.8 Ours (Learnable Time) 96.1 67.3 Table 3. Robustness to Input Noise on CIFAR-10 (% Accuracy) Method No Noise Gaussian (σ=0.1) Gaussian (σ=0.3) FGSM (ε=0.1) ResNet-18 94.5 91.2 83.7 78.9 ViT-Small 95.2 92.4 85.1 80.3 Ours 96.1 93.5 87.0 83.7 Table 4. Generalization to Out-of-Distribution (OOD) Data Method In-Domain (CIFAR-10) OOD (SVHN) OOD (CIFAR-10-C) ResNet-18 94.5 76.3 71.2 ViT-Small 95.2 78.1 73.0 Ours 96.1 80.5 75.3
Popular Tags