servers in Discovery Building

Large data sets require statistical analysis, but some data may come in varied forms (text, audio, video, sensor data, etc.). To cope with data structure variations, optimization explores integrating statistical processing techniques with data processing systems to make such systems easer to build, maintain and deploy.

Mass digitization of printed media into plain test is changing the types of data that companies and academic institutions manage. Scanned images are converted to plain text by conversion software, but often the software is error-prone. Any query of the digitized media may miss information that may lead to a poor results or applications. Staccato software improves the digitization process.


Laura Albert is an affiliate of the optimization group at WID […]


There is an arms race to perform increasingly sophisticated data […]


Jellyfish is an algorithm for solving data-processing problems with matrix-valued […]

Active learning methods automatically adapt data collection by selecting the […]