CeDAR/UCD4IDS First Joint Conference

CeDAR UCD4IDS Conference Flyer

Event Date

Location
Mathematical Sciences Building, MSB 1147

CeDAR and UCD4IDS will hold their first joint conference (in person) on Friday, December 2. Please see details below. If you plan to attend this conference, please register by Tuesday, November 22.

This conference in the in-person format will showcase the research of some members of CeDAR and UCD4IDS and provide more discussions and opportunities for further collaborations among different disciplines. This event will involve faculty members, postdocs, and graduate students. 

Registration here.

Conference Schedule

10:00-10:10 am -- Introduction/Updates

10:10-10:55 am -- Speaker 1: Patrice Koehl - A physicist's view on partial 3D shape comparison

11:00-11:45 am -- Speaker 2: Alex Wein - Computational Complexity of Tensor Decomposition

11:45-13:15 pm -- Lunch

13:15-14:00 pm -- Speaker 3: Laura Marcu - Fluorescence Lifetime Imaging in Surgical Oncology

14:05-14:50 pm -- Speaker 4: Xueheng Shi - Changepoint Analysis in Time Series Data: Past and Present

14:50-15:05 pm -- Coffee Break 

15:05-15:50 pm -- Speaker 5: Stefan Schonsheck - Representation of Non-Euclidean Domains and Analysis of Signals Thereon

15:55-16:40 pm -- Speaker 6: James Sharpnack - Lessons in domain adaptation and an unsupervised hunt for gravitational lenses

16:40-16:55 pm -- Summary/Handing out questions

17:00-18:30 pm -- Reception


Speakers, Titles & Abstract

Patrice Koehl, Professor of Department of Computer Science, Founding Director of Data Science Initiative

Title: A physicist's view on partial 3D shape comparison

Abstract: Scientists have access to a wide range of digital sensors that allow them to report at multiple time and length scales on the subjects of their studies. Finding efficient algorithms to describe and compare the shapes included in those reports has become a central problem in data science. Those algorithms have gained from developments in computational geometry and in machine learning. In this talk I will present another source of support to further improve those algorithms. Using techniques from statistical physics, I show that we can define a possibly partial correspondence between 3D shapes, with a cost associated with this correspondence that serves as a measure of the similarity of the shapes. I will illustrate the effectiveness of this approach on synthetic data as well as on real anatomical data. 

 

Alex Wein, Assistant Professor of Mathematics

Title: Computational Complexity of Tensor Decomposition

Abstract: Tensor decomposition is an important subroutine in numerous algorithms for data analysis, including topic modeling, community detection, clustering, and dictionary learning. We consider a simple model for tensor decomposition: suppose we are given a random rank-r order-3 tensor---that is, an n-by-n-by-n array of numbers that is the sum of r random rank-1 terms---and our goal is to recover the individual rank-1 terms. In principle this decomposition task is possible when r < cn^2 for a constant c, but all known polynomial-time algorithms require r << n^{3/2}. Is this a fundamental barrier for efficient algorithms? In recent years, the computational complexity of various high-dimensional statistical tasks has been resolved in restricted-but-powerful models of computation such as statistical queries, sum-of-squares, or low-degree polynomials. However, tensor decomposition has remained elusive, largely because its hardness is not explained by a "planted versus null" testing problem. We show the first formal hardness for average-case tensor decomposition: when r >> n^{3/2}, the decomposition task is hard for algorithms that can be expressed as low-degree polynomials in the tensor entries.

 

Laura Marcu, Professor of Biomedical Engineering, Professor of Department of Neurological Surgery, Director of National Center for Biophotonic Technologies

Title: Fluorescence Lifetime Imaging in Surgical Oncology

Abstract: This presentation reviews the development of clinically-compatible fluorescence lifetime imaging (FLIM) technology and applications in surgical oncology. Emphasis is placed on the integration of FLIM in surgical workflow and the potential of this approach to improve surgical decision-making during trans-oral robotic surgery (TORS) and neurosurgical procedures. Clinical outcomes and results will be discussed. We demonstrate the straightforward coupling of FLIM apparatus with the da Vinci surgical platform and the neuronavigation system. Also, we show innovative methods for real-time dynamic augmentation of imaging parameters on the surgical field of view as seen on the da Vinci console and surgical microscope. Current results demonstrate the utility of FLIM-derived parameters detecting tissue biochemical and metabolic characteristics to distinguish oral and oropharyngeal cancer in real-time from surrounding normal tissue in patients in-situ during TORS as well as to sense infiltrative brain cancer at the resection margins. Our findings suggest that label-free FLIM-based tissue assessment, characterized by simple, fast and flexible data acquisition and visualization, could find applications in a variety of surgical procedures.

 

Xueheng Shi, Assistant Professor of Department of Statistics at University of Nebraska-Lincoln

Title: Changepoint Analysis in Time Series Data: Past and Present

Abstract: Abrupt structural changes (changepoints) arise in many scenarios, for example, mean/trend shifts in time series, coefficient changes in the regressions. Changepoint analysis plays an important role in modelling and prediction of time series, and has vast applications in finance, climatology, signal processing and so on. This talk reviews prominent algorithms (Binary Segmentation/Wild Binary Segmentation and Pruned Exact Linear Time (PELT)) to detect mean shifts in time series. However, these methods require IID model errors while the time series are often autocorrelated (serial dependence). Changepoint analysis under serial dependence is a well-known difficult problem. We propose a gradient-descent dynamic programming algorithm to find the changepoints in time series data.

This research is joint work with Dr. Gallagher (Clemson University), Dr. Killick (Lancaster University, UK) and Dr. Lund (UCSC).

 

Stefan Schonsheck, Krener Assistant Professor of Mathematics

Title: Representation of Non-Euclidean Domains and Analysis of Signals Thereon

Abstract: Nonlinear dimensionality reduction models have made tremendous impacts on image and signal representation learning. However, almost all extant methods assume that real-valued Euclidean representations are rich enough faithfully preserve the data. However, flat geometry is too simplistic to meaningfully reflect the topological structure of many classes of data with cyclic structures. In the first half of this talk, we develop several methods for accurately representing data sampled from several non-euclidean domains: graphs, manifolds, varifolds, and simplicial complexes. A common theme in this section is to look for approaches that sidestep the curse of dimensionality by defining models that rely on the problem's intrinsic dimension rather than the dimension in which the observation is made. In the second part of this talk, we develop strategies for analyzing signals in these domains, intrinsically solving approximation, regression, and classification problems. This analysis is based on developing generalizations of classical techniques such as multiscale bases and overcomplete transforms, as well as neural network-based machine learning techniques.

 

James Sharpnack, Senior Applied Scientist in Amazon, Adjunct Associate Professor at UC Davis in the Statistics Department

Title: Lessons in domain adaptation and an unsupervised hunt for gravitational lenses

Abstract: Distribution shift occurs when the training and test distributions differ in a prediction task, and adapting our predictors to this setting is called domain adaptation.  In this talk, we will take a tour of the primary domain adaptation methods in classification under the lens of an unprecedented, comprehensive benchmark study and an application to astrophysics.  Our benchmark study and accompanying toolkit, RLSbench, consists of 11 vision datasets spanning > 200 distribution shift pairs with varying class proportions.  These involve both shifts in image (X) distribution and also label (Y) distribution, which is an often overlooked setting called relaxed label shift.  In the second half of this talk, we will apply domain adaptation methods to detecting gravitational lenses in a new astronomical survey (Deep Lens Survey), without the use of known lenses to train with.  Instead, we train on simulated lenses and use domain adaptation to generalize to real lenses.  We find that with a combination of MixMatch and smart data augmentations, we can dramatically improve model performance (precision at high recall).  This study has identified 9 Grade-A lenses, two of which have been spectroscopically confirmed, and 13 Grade-B lenses.