What are the Next Measurable Challenges in AI?
Building systems that can integrate learning, reasoning and optimization has long been a dream for artificial intelligence. One of the major challenges, within this context, is certainly to evaluate novel ideas and frameworks on appropriate benchmarks. Too often, in fact, the tasks and the datasets that are considered and proposed for experimental evaluation are tailored to some algorithms or methodologies,
and limited to ad-hoc scenarios and application domains. More in general, they lack an open and wider perspective to test the considered approaches across a variety of different tasks and under different conditions, making experimental comparisons hard to obtain.
Can we define a set of requirements for a challenge/benchmark that goes beyond those currently available?
Can we do it with the goal of having a benchmark that meets these requirements and can still be implemented in a
reasonable time? Possibly building on top of existing ones?
13:00-13:15 Doors open
13:15-13:30 Introduction & Expectations – Luc de Raedt
13:30-14:00 Invited Talk: Lessons Learned at NeurIPS 2021 Datasets and Benchmarks – Joaquin Vanschoren
14:00-14:15 Presentation Datasets/Systems Tables – Marco Lippi
14:15-15:30 Discussion on Tables – Working groups
15:45-16:45 Panel on Limitations of Existing Benchmarks and New Challenges – Andrea Passerini
16:45-18:00 Discussion on Panel – Working groups
18:00-18:15 What’s Next? – Luc de Raedt