Activity: Participating in or organising an event › Workshop/course/forum › Industry
Big Data has become ubiquitous in modern society, but drawing insights from it remains a challenge due to its unprecedented degrees of heterogeneity, often compounded by inadequate experimental design. The past decade has seen considerable developments with big data algorithms, but significant challenges remain for the area’s theoretical underpinning.
The aim of this workshop is to gather experts who develop theory and methodology for big data sets; i.e. scientists who construct new algorithms, but also develop theoretical understanding as to the analysis techniques that are optimal or preferable in different sampling scenarios. The workshop will feature research into computational and statistical efficiency trade-offs, high-dimensional dependency structures (such as spatiotemporal models), as well as high-dimensional estimation and learning, and privacy-preserving algorithms.