AAII Seminar: November 20, 2019

Title: Hidden Physics Models: Machine Learning of Non-Linear Partial Differential Equations

Speaker: Maziar Raissi (Brown University)

Abstract: A grand challenge with great opportunities is to develop a coherent framework that enables blending conservation laws, physical principles, and/or phenomenological behaviors expressed by differential equations with the vast data sets available in many fields of engineering, science, and technology. At the intersection of probabilistic machine learning, deep learning, and scientific computations, this work is pursuing the overall vision to establish promising new directions for harnessing the long-standing developments of classical methods in applied mathematics and mathematical physics to design learning machines with the ability to operate in complex domains without requiring large quantities of data. To materialize this vision, this work is exploring two complementary directions: (1) designing data-efficient learning machines capable of leveraging the underlying laws of physics, expressed by time dependent and non-linear differential equations, to extract patterns from high-dimensional data generated from experiments, and (2) designing novel numerical algorithms that can seamlessly blend equations and noisy multi-fidelity data, infer latent quantities of interest (e.g., the solution to a differential equation), and naturally quantify uncertainty in computations. The latter is aligned in spirit with the emerging field of probabilistic numerics.

@UCSC: E2-215 at 12pm PT

Online:  https://ucsc.zoom.us/j/468337241

AAII Seminar: November 6, 2019 at 12pm

Title:  TRIMAP: LARGE-SCALE DIMENSIONALITY REDUCTION USING TRIPLETS
Speakers: Ehsan Amid and Manfred Warmuth (UCSC / visiting Google Brain)

Abstract:  We introduce “TriMap”; a dimensionality reduction technique based on triplet constraints that preserves the global accuracy of the data better than the other commonly used methods such as t-SNE, LargeVis, and UMAP. To quantify the global accuracy, we introduce a score which roughly reflects the relative placement of the clusters rather than the individual points. We empirically show the excellent performance of TriMap on a large variety of datasets in terms of the quality of the embedding as well as the runtime. On our performance benchmarks, TriMap easily scales to millions of points without depleting the memory and clearly outperforms t-SNE, LargeVis, and UMAP in terms of runtime.

@UCSC:  E2-215  at 12pm PT

Online: https://ucsc.zoom.us/j/120429746

AAII Seminar: October 30, 2019

When/where: E2-215 at 12pm

Presenter: Oskar Elek
Title: Monte Carlo Physarum Machine: Unconventional AI for astronomy and beyond

Abstract: Slime mold (Physarum Polycephallum) is a freak of the natural world that – out of decaying forest debris – builds near-optimal transport networks. We leverage a custom Monte-Carlo simulation of this organism to approximate such transport networks: both in terms of structure and their density (likelihood) in 3D space. I will discuss our analysis of the Cosmic Web using this hybrid model, as well as future prospects in training and adapting it to several other domains.

With luck this Zoom link will work.

AAII Seminar: October 23, 2019

When/where: E2-215 at 12pm

Presenter: Reuben Harry Cohn-Gordon (Stanford University)
Title: Bayesian Models of Pragmatics for Natural Language

Abstract: Emerging from work in Bayesian cognitive science and game theory, probabilistic models of pragmatic reasoning have been successful at modelling human inferences about linguistic meaning, on the basis of their interlocutor’s intentions and knowledge. However, they have largely been applied to idealized domains. Meanwhile, natural language processing systems have made significant progress at open-domain tasks requiring language understanding, but often struggle to behave in human-like ways. I describe work combining these approaches, with the aim of obtaining interpretable models of pragmatic reasoning for natural language.

AAII Seminar: October 16, 2020

Where:  CfAO (we return to E2-215 next week)

When: 12-1pm

What: Or own David Reiman (Physics) will describe his internship over the Summer at Deep Mind.  Others are encouraged to the fill the remainder of the time by describing their own AI-related activities over the summer.

We will have a formal seminar on October 23, 2020 when we return to E2-215.  And it rumored that pizza will appear.

AAII Slack

Anyone at UCSC can join the AAII Slack with this link:

https://join.slack.com/t/appliedaiinstitute/signup

Come on in.

AAII Seminar: October 9, 2019

Happy new academic year!

We will celebrate with our first ‘seminar’.  As the standard E2-215 room is needed for an interview next week, we will meet in the CfAO atrium (on the ground floor of this building on Science Hill).  We will brainstorm ideas and speakers for the coming year, discuss new papers, and introduce ourselves to one another.

If you have been before, bring a new friend.  If you are new, then welcome!

AAII Seminar: June 5, 2019

This week we will have 2 presentations from Computational Media:

Speaker 1:
Mahika Dubey, Graduate Student in Computational Media, UCSC Creative Coding Lab
https://www.mahikadubey.com/about
Title:
Data Brushes: Interactive Neural Style Transfer for Data Art
Abstract:
We introduce in-browser applications for the application of data art-based style-transfer brushes onto an image, inviting casual creators and other non-technical users to interact with deep convolutional neural networks to co-create custom artworks. In addition to enabling a novel creative workflow, the process of interactively modifying an image via multiple style transfer neural networks reveals meaningful features encoded within the networks, and provides insight into the effect particular networks have on different images, or even different regions of an image, such as border artifacts and spatial stability or instability. Our data brushes provide new perspectives on the creative use of neural style transfer for data art and enhance intuition for the expressive range of particular style transfer features.
Speaker 2:
Oskar Elek, Postdoctoral Researcher in Computational Media, UCSC Creative Coding Lab
https://cgg.mff.cuni.cz/~oskar/
Title:
Learning Patterns in Sample Distributions for Monte Carlo Variance Reduction
Abstract:
This ongoing project investigates the prediction of unknown distributions from stochastic, sparse samples. A relevant problem in many domains (especially forecasting), we address it from the perspective of stochastic Monte Carlo integration in physically based image synthesis (rendering). Because the sample distributions obtained in rendering are complex, chaotic, and do not conform to known statistical distributions, simple estimation methods usually yield results with high amounts of variance. To tackle this issue, we systematically study these sample distributions to understand common patterns and archetypes, and propose to use deep neural networks to learn them. I will present our current results, with an open discussion centered on the main challenge: How can we use the knowledge of characteristic sample patterns to bootstrap the network and get better predictions?

AAII Seminar: May 29, 2019

When/where:  E2-215 at 12pm

Presenter: Jaehoon Lee (Google Brain)

Title:  Everything you wanted to know about batch size (in neural net training) but were afraid to ask

Abstract: Recent hardware developments have made unprecedented amounts of data parallelism available for accelerating neural network training. Among the simplest ways to harness next-generation accelerators is to increase the batch size in standard mini-batch neural network training algorithms. In this work, we aim to experimentally characterize the effects of increasing the batch size on training time, as measured in the number of steps necessary to reach a goal out-of-sample error. Eventually, increasing the batch size will no longer reduce the number of training steps required, but the exact relationship between the batch size and how many training steps are necessary is of critical importance to practitioners, researchers, and hardware designers alike. We study how this relationship varies with the training algorithm, model, and data set and find extremely large variation between workloads. Along the way, we reconcile disagreements in the literature on whether batchsize affects model quality. Finally, we discuss the implications of our results for efforts to train neural networks much faster in the future.

Reference: https://arxiv.org/abs/1811.03600

 

Skip to toolbar