|
1 | 1 | --- |
2 | | -title: ML Reproducibility Challenge 2023 |
| 2 | +title: ML Reproducibility Challenge |
3 | 3 | type: book # Do not modify. |
4 | 4 | toc: false |
5 | 5 | headless: true |
6 | 6 | --- |
7 | 7 |
|
8 | | -Welcome to the ML Reproducibility Challenge 2023 (**MLRC 2023**). This is the |
9 | | -seventh edition of the event |
| 8 | +Welcome to the home of ML Reproducibility Challenge. This is an annual event for |
| 9 | +providing a space for research into reproducibility of Machine Learning |
| 10 | +literature. |
10 | 11 | ([v1](https://www.cs.mcgill.ca/~jpineau/ICLR2018-ReproducibilityChallenge.html), |
11 | 12 | [v2](https://www.cs.mcgill.ca/~jpineau/ICLR2019-ReproducibilityChallenge.html), |
12 | 13 | [v3](https://reproducibility-challenge.github.io/neurips2019/), |
13 | 14 | [v4](https://paperswithcode.com/rc2020), |
14 | 15 | [v5](https://paperswithcode.com/rc2021), |
15 | | -[v6](https://paperswithcode.com/rc2022)). The primary goal of this event is to |
16 | | -encourage the publishing and sharing of scientific results that are reliable and |
17 | | -reproducible. In support of this, the objective of this challenge is to |
18 | | -investigate reproducibility of papers accepted for publication at top |
19 | | -conferences by inviting members of the community at large to select a paper, and |
20 | | -verify the empirical results and claims in the paper by reproducing the |
21 | | -computational experiments, either via a new implementation or using code/data or |
22 | | -other information provided by the authors. |
23 | | - |
24 | | -## Final decisions for MLRC 2023 |
25 | | - |
26 | | -We are now releasing the final list of decisions for MLRC 2023. This list |
27 | | -includes the previous partial list published on July 5th, 2024. We have given |
28 | | -additional time to TMLR to complete the reviews, however it is unfortunate that |
29 | | -few papers are still awaiting a decision due to unresponsive Action Editors from |
30 | | -TMLR. As we need to wrap up this edition, we are proceeding with the final list |
31 | | -of 22 accepted papers. Congratulations to all! |
32 | | - |
33 | | -- Ana-Maria Vasilcoiu, Batu Helvacioğlu, Thies Kersten, Thijs Stessen; |
34 | | - _GNNInterpreter: A probabilistic generative model-level explanation for Graph |
35 | | - Neural Networks_, [OpenReview](https://openreview.net/forum?id=8cYcR23WUo) |
36 | | -- Miklos Hamar, Matey Krastev, Kristiyan Hristov, David Beglou; _Explaining |
37 | | - Temporal Graph Models through an Explorer-Navigator Framework_, |
38 | | - [OpenReview](https://openreview.net/forum?id=FI1XvwpchC) |
39 | | -- Clio Feng, Colin Bot, Bart den Boef, Bart Aaldering; _Reproducibility Study of |
40 | | - "Explaining RL Decisions with Trajectories"_, |
41 | | - [OpenReview](https://openreview.net/forum?id=JQoWmeNaC2) |
42 | | -- Ethan Harvey, Mikhail Petrov, Michael C. Hughes; _Transfer Learning with |
43 | | - Informative Priors: Simple Baselines Better than Previously Reported_, |
44 | | - [OpenReview](https://openreview.net/forum?id=BbvSU02jLg) |
45 | | -- Gijs de Jong,Macha Meijer,Derck W.E. Prinzhorn,Harold Ruiter; _Reproducibility |
46 | | - study of FairAC_, [OpenReview](https://openreview.net/forum?id=ccDi5jtSF7) |
47 | | -- Nesta Midavaine, Gregory Hok Tjoan Go, Diego Canez, Ioana Simion, Satchit |
48 | | - Chatterji; _On the Reproducibility of Post-Hoc Concept Bottleneck Models_; |
49 | | - [OpenReview](https://openreview.net/forum?id=8UfhCZjOV7) |
50 | | -- Jiapeng Fan, Paulius Skaigiris, Luke Cadigan, Sebastian Uriel Arias; |
51 | | - _Reproducibility Study of "Learning Perturbations to Explain Time Series |
52 | | - Predictions"_, [OpenReview](https://openreview.net/forum?id=fCNqD2IuoD) |
53 | | -- Karim Ahmed Abdel Sadek, Matteo Nulli, Joan Velja, Jort Vincenti; _Explaining |
54 | | - RL Decisions with Trajectories’: A Reproducibility Study_, |
55 | | - [OpenReview](https://openreview.net/forum?id=QdeBbK5CSh) |
56 | | -- Markus Semmler, Miguel de Benito Delgado; _Classwise-Shapley values for data |
57 | | - valuation_ [OpenReview](https://openreview.net/forum?id=srFEYJkqD7) |
58 | | -- Daniel Gallo Fernández, Răzvan-Andrei Matișan, Alejandro Monroy Muñoz, Janusz |
59 | | - Partyka; _Reproducibility Study of "ITI-GEN: Inclusive Text-to-Image |
60 | | - Generation"_ [OpenReview](https://openreview.net/forum?id=d3Vj360Wi2) |
61 | | -- Kacper Bartosik, Eren Kocadag, Vincent Loos, Lucas Ponticelli; |
62 | | - _Reproducibility study of "Robust Fair Clustering: A Novel Fairness Attack and |
63 | | - Defense Framework"_, [OpenReview](https://openreview.net/forum?id=Xu1sEPhjqH) |
64 | | -- Barath Chandran C; _CUDA: Curriculum of Data Augmentation for Long‐Tailed |
65 | | - Recognition_, [OpenReview](https://openreview.net/forum?id=Wm6d44I8St) |
66 | | -- Christina Isaicu, Jesse Wonnink, Andreas Berentzen, Helia Ghasemi; |
67 | | - _Reproducibility Study of “Explaining Temporal Graph Models Through an |
68 | | - Explorer-Navigator Framework"_, |
69 | | - [OpenReview](https://openreview.net/forum?id=9M2XqvH2SB) |
70 | | -- Iason Skylitsis, Zheng Feng, Idries Nasim, Camille Niessink; _Reproducibility |
71 | | - Study of "Robust Fair Clustering: A Novel Fairness Attack and Defense |
72 | | - Framework"_, [OpenReview](https://openreview.net/forum?id=H1hLNjwrGy) |
73 | | -- Fatemeh Nourilenjan Nokabadi, Jean-Francois Lalonde, Christian Gagné; |
74 | | - _Reproducibility Study on Adversarial Attacks Against Robust Transformer |
75 | | - Trackers_, [OpenReview](https://openreview.net/forum?id=FEEKR0Vl9s) |
76 | | -- Luan Fletcher, Robert van der Klis, Martin Sedlacek, Stefan Vasilev, Christos |
77 | | - Athanasiadis; _Reproducibility study of “LICO: Explainable Models with |
78 | | - Language-Image Consistency"_, |
79 | | - [OpenReview](https://openreview.net/forum?id=Mf1H8X5DVb) |
80 | | -- Wouter Bant, Ádám Divák, Jasper Eppink, Floris Six Dijkstra; _On the |
81 | | - Reproducibility of: "Learning Perturbations to Explain Time Series |
82 | | - Predictions"_, [OpenReview](https://openreview.net/forum?id=nPZgtpfgIx) |
83 | | -- Berkay Chakar,Amina Izbassar,Mina Janićijević,Jakub Tomaszewski; |
84 | | - _Reproducibility Study: Equal Improvability: A New Fairness Notion Considering |
85 | | - the Long-Term Impact_, |
86 | | - [OpenReview](https://openreview.net/forum?id=Yj8fUQGXXL) |
87 | | -- Oliver Bentham, Nathan Stringham, Ana Marasović; _Chain-of-Thought |
88 | | - Unfaithfulness as Disguised Accuracy_, |
89 | | - [OpenReview](https://openreview.net/forum?id=ydcrP55u2e) |
90 | | -- Shivank Garg, Manyana Tiwari; _Unmasking the Veil: An Investigation into |
91 | | - Concept Ablation for Privacy and Copyright Protection in Images_ |
92 | | - [OpenReview](https://openreview.net/forum?id=TYYApLzjaQ) |
93 | | -- Adrian Sauter, Milan Miletić, Ryan Ott, Rohith Saai Pemmasani Prabakaran; |
94 | | - _Studying How to Efficiently and Effectively Guide Models with Explanations” - |
95 | | - A Reproducibility Study_, |
96 | | - [OpenReview](https://openreview.net/forum?id=9ZzASCVhDF) |
97 | | -- Thijmen Nijdam, Taiki Papandreou-Lazos, Jurgen de Heus, Juell Sprott; |
98 | | - _Reproducibility Study Of Learning Fair Graph Representations Via Automated |
99 | | - Data Augmentations_, [OpenReview](https://openreview.net/forum?id=4WiqHopXQX) |
100 | | - |
101 | | -If you are an author of the below mentioned papers and have not |
102 | | -[submitted the form](https://forms.gle/JJ28rLwBSxMriyE89) with the camera ready |
103 | | -items, please consider doing so at the earliest. We will reach out to the |
104 | | -accepted authors soon with the next steps. We will also announce the best paper |
105 | | -awards and share details on the logistics of NeurIPS poster session in the |
106 | | -coming weeks. |
107 | | - |
108 | | -**Update, Sept 13th, 2024**: A couple of papers received acceptance status post |
109 | | -our final date of MLRC 2023 acceptance. We have now incorporated them too in the |
110 | | -final list. |
111 | | - |
112 | | -## An update on decisions |
113 | | - |
114 | | -_July 5th, 2024_ |
115 | | - |
116 | | -We initially communicated to have all decisions of MLRC 2023 out by 31st of |
117 | | -May, 2024. Unfortunately, several submissions are still under review at TMLR, |
118 | | -and we are waiting for the final decisions to trickle in. Overall, MLRC 2023 had |
119 | | -46 valid submissions, out of which we have recieved decisions on 61% of them. We |
120 | | -are in touch with TMLR to expedite the process of decisions for the remaining |
121 | | -submissions - we expect all decisions to come in by the next couple of weeks. |
122 | | - |
123 | | -Until then, we are happy to announce the (partial) list of accepted papers. |
124 | | -Congratulations to all :tada:! If you are an author of the below mentioned |
125 | | -papers and have not [submitted the form](https://forms.gle/JJ28rLwBSxMriyE89) |
126 | | -with the camera ready items, please consider doing so at the earliest. We will |
127 | | -reach out to the accepted authors soon with the next steps. |
128 | | - |
129 | | -_(partial paper list removed as we release the final list above)_ |
130 | | - |
131 | | -## [Deprecated] Call For Papers |
132 | | - |
133 | | -We invite contributions from academics, practitioners and industry researchers |
134 | | -of the ML community to submit novel and insightful reproducibility studies. |
135 | | -Please read our [blog post](/blog/announcing_mlrc2023/) regarding our |
136 | | -retrospectives of running the challenge and the future roadmap. We are happy to |
137 | | -announce the formal partnership with |
138 | | -[Transactions of Machine Learning Research (TMLR)](https://jmlr.org/tmlr/) |
139 | | -journal. The challenge goes live on **October 23, 2023**. |
140 | | - |
141 | | -We recommend you choose any paper(s) published in the 2023 calendar year from |
142 | | -the top conferences and journals ([NeurIPS](https://neurips.cc/), |
143 | | -[ICML](https://icml.cc/), [ICLR](https://iclr.cc/), |
144 | | -[ACL](https://2023.aclweb.org/), [EMNLP](https://2023.emnlp.org/), |
145 | | -[ICCV](https://iccv2023.thecvf.com/), |
146 | | -[CVPR](https://cvpr2023.thecvf.com/Conferences/2023), |
147 | | -[TMLR](https://jmlr.org/tmlr/), [JMLR](https://jmlr.org/), |
148 | | -[TACL](https://transacl.org/index.php/tacl)) to run your reproducibility study |
149 | | -on. |
150 | | - |
151 | | -{{< figure src="uploads/mlrc.drawio.svg" class="mlrc_dark" >}} |
152 | | - |
153 | | -{{< figure src="uploads/mlrc.light.drawio.svg" class="mlrc_light" >}} |
154 | | - |
155 | | -In order for your paper to be submitted and presented at MLRC 2023, it first |
156 | | -needs to be **accepted and published** at TMLR. While TMLR aims to follow a |
157 | | -2-months timeline to complete the review process of its regular submissions, |
158 | | -this timeline is not guaranteed. If you haven’t already, we therefore recommend |
159 | | -submitting your original paper to TMLR by **February 16th, 2024**, that is a |
160 | | -little over 3 months in advance of the MLRC publication announcement date. |
161 | | - |
162 | | -## Key Dates |
163 | | - |
164 | | -- Challenge goes live: October 23, 2023 |
165 | | -- Deadline to share your **intent to submit** a TMLR paper to MLRC: **February |
166 | | - 16th, 2024** at the following form: **https://forms.gle/JJ28rLwBSxMriyE89**. |
167 | | - This form requires that you provide a link to your TMLR submission. Once it |
168 | | - gets accepted (if it isn’t already), you should then update the same form with |
169 | | - your paper camera ready details. Your accepted TMLR paper will finally undergo |
170 | | - a light AC review to verify MLRC compatibility. |
171 | | -- We aim to announce the accepted papers by ~~**May 31st, 2024**~~ **July 17th, |
172 | | - 2024**, pending decisions of all papers. |
173 | | - |
174 | | -## Contact Information |
175 | | - |
176 | | -- For query regarding MLRC 2023, mail us at: |
177 | | - |
178 | | -- For general queries, media, sponsorship, partnership requests, mail us at |
179 | | - |
| 16 | +[v6](https://paperswithcode.com/rc2022), [v7](/proceedings/mlrc2023/)). The |
| 17 | +primary goal of this event is to encourage the publishing and sharing of |
| 18 | +scientific results that are reliable and reproducible. In support of this, the |
| 19 | +objective of this challenge is to investigate reproducibility of papers accepted |
| 20 | +for publication at top conferences by inviting members of the community at large |
| 21 | +to select a paper, and verify the empirical results and claims in the paper by |
| 22 | +reproducing the computational experiments, either via a new implementation or |
| 23 | +using code/data or other information provided by the authors. |
| 24 | + |
| 25 | +{{% callout note %}} |
| 26 | + |
| 27 | +- :mortar_board: [MLRC 2023](/proceedings/mlrc2023/) papers featured in |
| 28 | + [NeurIPS 2024 Poster Sessions](https://neurips.cc/), Dec 10-15, 2024 at |
| 29 | + Vancouver, Canada. If you are attending NeurIPS, do |
| 30 | + [drop by to the posters](/proceedings/) to say hi! |
| 31 | +- Next iteration of MLRC will be **MLRC2025**, and it will be **in-person** - a |
| 32 | + one-day conference! Announcement will be made very soon, stay tuned! |
| 33 | + |
| 34 | +{{% /callout %}} |
| 35 | + |
| 36 | +{{< tweet user="hugo_larochelle" id="1819465878641262862" >}} |
| 37 | + |
| 38 | +<a href="https://twitter.com/x?ref_src=twsrc%5Etfw" class="twitter-follow-button" data-show-count="false">Follow |
| 39 | +@x</a><script async src="https://platform.twitter.com/widgets.js" charset="utf-8"></script> |
0 commit comments