- 30
- 35 405
JumpTrading ELLIS UCL CSML Seminar Series
เข้าร่วมเมื่อ 12 ธ.ค. 2020
Machine Learning weekly seminar series at UCL, sponsored by Jump Trading and ELLIS.
The Centre for Computational Statistics and Machine Learning (CSML) spans three departments at UCL: Computer Science, Statistical Science, and the Gatsby Unit.
The Centre for Computational Statistics and Machine Learning (CSML) spans three departments at UCL: Computer Science, Statistical Science, and the Gatsby Unit.
Causal foundations for safe AGI - Tom Everitt (DeepMind)
Slides: www.dropbox.com/s/l797ktr8w3n3yp4/20220826_Tom_Everitt_Causal_Alignment.pdf?dl=0
Abstract:
With great power comes great responsibility. Human-level+ artificial general intelligence (AGI) may become humanity’s best friend or worst enemy, depending on whether we manage to align its behavior with human interests or not. To overcome this challenge, we must identify the potential pitfalls and develop effective mitigation strategies. In this talk, I’ll argue that (Pearlian) causality offers a useful formal framework for reasoning about AI risk, and describe some of our recent work on this topic. In particular, I’ll cover causal definitions of incentives, agents, side effects, generalization, and preference manipulation, and discuss how techniques like recursion, interpretability, impact measures, incentive design, and path-specific effects can combine to address AGI risks.
Bio:
Tom Everitt is a senior researcher at DeepMind, leading a small team on causal approaches to AGI safety. He holds a PhD from Australian National University, where he wrote the first PhD thesis fully focused on AGI safety under the supervision of Prof. Marcus Hutter.
Abstract:
With great power comes great responsibility. Human-level+ artificial general intelligence (AGI) may become humanity’s best friend or worst enemy, depending on whether we manage to align its behavior with human interests or not. To overcome this challenge, we must identify the potential pitfalls and develop effective mitigation strategies. In this talk, I’ll argue that (Pearlian) causality offers a useful formal framework for reasoning about AI risk, and describe some of our recent work on this topic. In particular, I’ll cover causal definitions of incentives, agents, side effects, generalization, and preference manipulation, and discuss how techniques like recursion, interpretability, impact measures, incentive design, and path-specific effects can combine to address AGI risks.
Bio:
Tom Everitt is a senior researcher at DeepMind, leading a small team on causal approaches to AGI safety. He holds a PhD from Australian National University, where he wrote the first PhD thesis fully focused on AGI safety under the supervision of Prof. Marcus Hutter.
มุมมอง: 527
วีดีโอ
Maurice Weiler - Equivariant and Coordinate Independent Convolutional Networks
มุมมอง 9202 ปีที่แล้ว
Speaker: Maurice Weiler (University of Amsterdam) Title: Equivariant and Coordinate Independent Convolutional Networks Abstract: The classical convolutional network architecture can be derived solely from the requirement for translational equivariance. Steerable CNNs generalize this idea to affine symmetry groups, resulting in network architectures that are equivariant under additional symmetri...
Rebecca Lewis - Inference in High-Dimensional Logistic Regression Models with Separated Data
มุมมอง 4072 ปีที่แล้ว
Speaker: Rebecca Lewis (Imperial College London) Title: Inference in High-Dimensional Logistic Regression Models with Separated Data Abstract: Existence of the maximum likelihood estimate of logistic regression coefficients requires that the observed sequence of covariate and response values are not linearly separable. Even when the maximum likelihood estimator exists, it can suffer from consid...
Badr-Eddine Chérief-Abdellatif - Robust Estimation via Maximum Mean Discrepancy
มุมมอง 3592 ปีที่แล้ว
Speaker: Badr-Eddine Chérief-Abdellatif (University of Oxford) Title: Robust Estimation via Maximum Mean Discrepancy Paper link: proceedings.mlr.press/v118/cherief-abdellatif20a/cherief-abdellatif20a.pdf Abstract: In this talk, we will study the properties of a minimum distance estimator based on the Maximum Mean Discrepancy (MMD). We will show that this estimator is universal in the i.i.d. set...
Harita Dellaporta - Robust Bayesian Inference for Simulator-Based Models via MMD Posterior Bootstrap
มุมมอง 2232 ปีที่แล้ว
Speaker: Harita Dellaporta (University of Warwick) Title: Robust Bayesian Inference for Simulator-Based Models via MMD Posterior Bootstrap Abstract: Simulator-based models are models for which the likelihood is intractable but simulation of synthetic data is possible. They are often used to describe complex real-world phenomena, and as such can often be misspecified in practice. In this talk, I...
Alexander Terenin - Non-Euclidean Matérn Gaussian Processes
มุมมอง 3292 ปีที่แล้ว
Speaker: Alexander Terenin (Cambridge University) Title: Non-Euclidean Matérn Gaussian Processes Abstract: In recent years, the machine learning community has become increasingly interested in learning in settings where data lives in non-Euclidean spaces, for instance in applications to physics and engineering, or other settings where it is important that symmetries are enforced. In this talk, ...
Siu Lun Chau - Explaining Kernel Methods with RKHS-SHAP
มุมมอง 4972 ปีที่แล้ว
Speaker: Siu Lun Chau (University of Oxford) Title: Explaining Kernel Methods with RKHS-SHAP Abstract: Feature attribution for kernel methods is often heuristic and not individualised for each prediction. To address this, we turn to the concept of Shapley values, a coalition game theoretical framework that has previously been applied to different machine learning model interpretation tasks. By ...
Manfred K. Warmuth - The blessing and the curse of the multiplicative updates
มุมมอง 3242 ปีที่แล้ว
Speaker: Manfred K. Warmuth, Google Brain (formerly University of California at Santa Cruz) Title: The blessing and the curse of the multiplicative updates - discusses connections between in evolution and the multiplicative updates of online learning Abstract: Multiplicative updates multiply the parameters by nonnegative factors. These updates are motivated by a Maximum Entropy Principle and th...
Julien Mairal - Lucas-Kanade Reloaded: End-to-End Super-Resolution from Raw Image Bursts
มุมมอง 2402 ปีที่แล้ว
Abstract: This presentation addresses the problem of reconstructing a high-resolution image from multiple lower-resolution snapshots captured from slightly different viewpoints in space and time. Key challenges for solving this super-resolution problem include (i) aligning the input pictures with sub-pixel accuracy, (ii) handling raw (noisy) images for maximal faithfulness to native camera data...
Stefano Ermon - Utilitarian Information Theory
มุมมอง 8982 ปีที่แล้ว
Abstract: Shannon’s information theory, which lies at the foundation of AI and machine learning, provides a conceptual framework to characterize information in a mathematically rigorous sense. However, important computational aspects are not considered, as it does not account for how much information can actually be used by a computationally bounded decision maker. This limits its utility in se...
Emtiyaz Khan - The Bayesian Learning Rule for Adaptive AI
มุมมอง 6702 ปีที่แล้ว
Paper: The Bayesian Learning Rule, (Preprint) M.E. Khan, H. Rue arxiv.org/abs/2107.04562 Abstract: Humans and animals have a natural ability to autonomously learn and quickly adapt to their surroundings. How can we design AI systems that do the same? In this talk, I will present Bayesian principles to bridge such gaps between humans and AI. I will show that a wide-variety of machine-learning al...
Alexandre Gramfort - Machine learning without human supervision on neuroscience signals
มุมมอง 4872 ปีที่แล้ว
Abstract: The revolution of artificial intelligence over the last decade has been made possible by statistical machine learning, and in particular by supervised learning where algorithms are given the labels associated with each observation. Although very efficient, this approach faces several difficulties in a neuroscience and more broadly in a medical context: one needs enough labels, one nee...
Hossein Mobahi: Sharpness-Aware Minimization (SAM): Current Method and Future Directions
มุมมอง 3.8K2 ปีที่แล้ว
Slides: www.dropbox.com/s/66wet9ps2a6i5ey/Hossein_Mobahi_SAM_CSML_Talk.pdf?dl=0 TITLE: Sharpness-Aware Minimization (SAM): Current Method and Future Directions ABSTRACT: In today's heavily overparameterized models, the value of the training loss provides few guarantees on model generalization ability. Indeed, optimizing only the training loss value, as is commonly done, can easily lead to subop...
Stéphanie Allassonnière - Geometry-Aware Variational Autoencoders for Medical Data Augmentation
มุมมอง 6192 ปีที่แล้ว
Abstract: In this presentation, we propose a new method to perform data augmentation in a reliable way in the High Dimensional Low Sample Size (HDLSS) setting using a geometry-based variational autoencoder. Our approach combines a proper latent space modeling of the VAE seen as a Riemannian manifold with a new generation scheme which produces more meaningful samples especially in the context of...
Richard Samworth - Optimal Subgroup Selection
มุมมอง 2912 ปีที่แล้ว
Abstract: In clinical trials and other applications, we often see regions of the feature space that appear to exhibit interesting behaviour, but it is unclear whether these observed phenomena are reflected at the population level. Focusing on a regression setting, we consider the subgroup selection challenge of identifying a region of the feature space on which the regression function exceeds a...
Makoto Yamada - Selective Inference with Kernels
มุมมอง 3052 ปีที่แล้ว
Makoto Yamada - Selective Inference with Kernels
Mark Herbster - Online Multitask Learning with Long-Term Memory
มุมมอง 1812 ปีที่แล้ว
Mark Herbster - Online Multitask Learning with Long-Term Memory
Contrastive Self-Supervised Learning and Potential Limitations - Dr Ting Chen from Google Brain
มุมมอง 2.8K2 ปีที่แล้ว
Contrastive Self-Supervised Learning and Potential Limitations - Dr Ting Chen from Google Brain
Nicholas Bishop - Strategic Least Squares Regression with Verified Training Data
มุมมอง 883 ปีที่แล้ว
Nicholas Bishop - Strategic Least Squares Regression with Verified Training Data
Petar Veličković - A Tale of Three Implicit Planners and the XLVIN agent
มุมมอง 8853 ปีที่แล้ว
Petar Veličković - A Tale of Three Implicit Planners and the XLVIN agent
Greg Yang - Feature Learning in Infinite-Width Neural Networks
มุมมอง 2K3 ปีที่แล้ว
Greg Yang - Feature Learning in Infinite-Width Neural Networks
Chelsea Finn - Principles for Tackling Distribution Shift: Pessimism, Adaptation, and Anticipation
มุมมอง 2.1K3 ปีที่แล้ว
Chelsea Finn - Principles for Tackling Distribution Shift: Pessimism, Adaptation, and Anticipation
David Duvenaud - Latent Stochastic Differential Equations: An Unexplored Model Class
มุมมอง 3K3 ปีที่แล้ว
David Duvenaud - Latent Stochastic Differential Equations: An Unexplored Model Class
Mihaela van der Schaar --- Why medicine is creating exciting new frontiers for machine learning
มุมมอง 4003 ปีที่แล้ว
Mihaela van der Schaar Why medicine is creating exciting new frontiers for machine learning
Guido Montúfar: Implicit bias of gradient descent for MSE regression with wide neural networks
มุมมอง 3413 ปีที่แล้ว
Guido Montúfar: Implicit bias of gradient descent for MSE regression with wide neural networks
Marta Garnelo - Meta-Learning and Neural Processes
มุมมอง 5K3 ปีที่แล้ว
Marta Garnelo - Meta-Learning and Neural Processes
Jakob Foerster - Zero-Shot (Human-AI) Coordination (in Hanabi) and Ridge Rider
มุมมอง 8873 ปีที่แล้ว
Jakob Foerster - Zero-Shot (Human-AI) Coordination (in Hanabi) and Ridge Rider
Sergey Levine - Data-Driven Reinforcement Learning: Deriving Common Sense from Past Experience
มุมมอง 3.6K3 ปีที่แล้ว
Sergey Levine - Data-Driven Reinforcement Learning: Deriving Common Sense from Past Experience
Gresele & Fissore- Relative gradient optimization of the Jacobian term in unsupervised deep learning
มุมมอง 2253 ปีที่แล้ว
Gresele & Fissore- Relative gradient optimization of the Jacobian term in unsupervised deep learning
Jonathan Frankle - The Lottery Ticket Hypothesis: On Sparse, Trainable Neural Networks
มุมมอง 2.9K3 ปีที่แล้ว
Jonathan Frankle - The Lottery Ticket Hypothesis: On Sparse, Trainable Neural Networks
When you say, it can train to full accuracy in the same number of steps - isn't that sort of untrue? If it is the same number of steps (N) to train the reduced network as the big network, wouldn't we have 2N at the end of finding and training the winning ticket?
Awesome idea.
Thank you for the insightful presentation
For the initial conditions that work, have anybody look at how much wiggle room you have. Is there an epsilon-neighborhood of the initial state you can safely start from, and how small is epsilon?
Did you find anything on this?
Thanks for a great talk
The world of NODE is so small, I can always meet you guys
thanks
Fascinating!
Hi~ Is there a video about ASAM? I wanna learn more about ASAM
Very helpful. A clear explanation of the idea behind the algorithm.
But 0.9 is actually not the only unique answer to the lever problem, there is a unique 1.0 lever directly opposite to the .9 lever. This is stable to the symmetries implied by the problem statement where both players are shown the same set of levers, as well as being robust to different reflections and rotations being presented to different players (though not arbitrary permutations). So assuming whoever I am paired with is also optimizing we would both earn 1. This strategy doesn't work if there are an odd number of buttons to begin with.
what computing power was used to achieve such results ?
I suggest 0.75x speed. Even though my ears can hear Marta clearly but my brain just doesn't catch up with the ideas.
Covers these papers: AWAC: arxiv.org/abs/2006.09359 MOPO: arxiv.org/abs/2005.13239 CQL: arxiv.org/abs/2006.04779 COG: arxiv.org/abs/2010.14500