100 likes | 226 Vues
The NoEBenchmark initiative highlights key challenges faced in evaluating 3D angular orientation in technology. Although the community embraced the challenge, it was not born from it, leading to issues such as a black box evaluation mechanism and lack of standardized interchange formats. Future benchmarks should involve well-defined tasks and datasets proposed by the community, transparent evaluation methods, and a competitive yet healthy environment. By learning from past experiences, we can promote continual improvement and innovation in the field.
E N D
Network of Excellence Benchmark Carlos Oscar Sorzano Techn. Director I2PC Natl. Center Biotechnology (CSIC)
NoEBenchmark • Therewas a definedtask. • The communityacceptedthechallenge. • The result of thechallengecouldhaveserved as a startingpointtoimprove3D angular orientation.
NoEBenchmark • The challengedidnotarisefromthecommunity: althoughthecommunityacceptedit.
NoEBenchmark • Evaluationwas a black box: theevaluation Figure of Meritwasunknown a priori.
NoEBenchmark • Therewasnota clearinterchangestandard: therewas no way of testingthatthesubmittedresultswhereactually in thecorrectformat, coordinatesystem, Euler signs, …
NoEBenchmark • Therewas a single dataset: therewas no exploration of differentpackagesperformingbetterunderdifferentcircumstances.
NoEBenchmark • Egos weretoohigh
NoEBenchmark • Itwas a single shot. 1891: 10.8s 1906: 10.6s 1911: 10.5s 1924: 10.4s : 10.3s 1936: 10.2s … 2009: 9.58s
For futurebenchmarks • Welldefinedtask. • Welldefinedinterchangeformat. • Taskproposedbythecommunity. • Differentdatasets. • Transparentevaluation with pre-test datasets. • Acceptance of competition as a healthywaytoimprove. • Possibilitytowithdrawfromcompetition. • Competitionsshould be repeatedover time.