This document summarizes a presentation about variational autoencoders (VAEs) presented at the ICLR 2016 conference. The document discusses 5 VAE-related papers presented at ICLR 2016, including Importance Weighted Autoencoders, The Variational Fair Autoencoder, Generating Images from Captions with Attention, Variational Gaussian Process, and Variationally Auto-Encoded Deep Gaussian Processes. It also provides background on variational inference and VAEs, explaining how VAEs use neural networks to model probability distributions and maximize a lower bound on the log likelihood.
This document discusses various methods for calculating Wasserstein distance between probability distributions, including:
- Sliced Wasserstein distance, which projects distributions onto lower-dimensional spaces to enable efficient 1D optimal transport calculations.
- Max-sliced Wasserstein distance, which focuses sampling on the most informative projection directions.
- Generalized sliced Wasserstein distance, which uses more flexible projection functions than simple slicing, like the Radon transform.
- Augmented sliced Wasserstein distance, which applies a learned transformation to distributions before projecting, allowing more expressive matching between distributions.
These sliced/generalized Wasserstein distances have been used as loss functions for generative models with promising
This document summarizes a presentation about variational autoencoders (VAEs) presented at the ICLR 2016 conference. The document discusses 5 VAE-related papers presented at ICLR 2016, including Importance Weighted Autoencoders, The Variational Fair Autoencoder, Generating Images from Captions with Attention, Variational Gaussian Process, and Variationally Auto-Encoded Deep Gaussian Processes. It also provides background on variational inference and VAEs, explaining how VAEs use neural networks to model probability distributions and maximize a lower bound on the log likelihood.
This document discusses various methods for calculating Wasserstein distance between probability distributions, including:
- Sliced Wasserstein distance, which projects distributions onto lower-dimensional spaces to enable efficient 1D optimal transport calculations.
- Max-sliced Wasserstein distance, which focuses sampling on the most informative projection directions.
- Generalized sliced Wasserstein distance, which uses more flexible projection functions than simple slicing, like the Radon transform.
- Augmented sliced Wasserstein distance, which applies a learned transformation to distributions before projecting, allowing more expressive matching between distributions.
These sliced/generalized Wasserstein distances have been used as loss functions for generative models with promising
The document introduces several approaches to semi-supervised learning, including self-training, multi-view algorithms like co-training, generative models using EM, S3VMs which extend SVMs to incorporate unlabeled data, and graph-based algorithms. Semi-supervised learning can make use of large amounts of unlabeled data together with smaller amounts of labeled data to build accurate predictive models in domains where labeling data is expensive.
This document summarizes an internship project using deep reinforcement learning to develop an agent that can automatically park a car simulator. The agent takes input from virtual cameras mounted on the car and uses a DQN network to learn which actions to take to reach a parking goal. Several agent configurations were tested, with the three-camera subjective view agent showing the most success after modifications to the reward function and task difficulty via curriculum learning. While the agent could sometimes learn to park, the learning was not always stable, indicating further refinement is needed to the deep RL approach for this automatic parking task.
1) The document discusses using wearable sensors to measure electrodermal activity (EDA) for autistic individuals to help understand their emotions and stress levels.
2) EDA can indicate sympathetic nervous system arousal which may not match outward appearances. Measuring EDA daily over long periods provides a better understanding of baseline levels.
3) A wearable EDA sensor was designed for comfort during long-term, everyday use to gain insights into how social interactions impact physiological states in autistic individuals.
1. The document discusses probabilistic modeling and variational inference. It introduces concepts like Bayes' rule, marginalization, and conditioning.
2. An equation for the evidence lower bound is derived, which decomposes the log likelihood of data into the Kullback-Leibler divergence between an approximate and true posterior plus an expected log likelihood term.
3. Variational autoencoders are discussed, where the approximate posterior is parameterized by a neural network and optimized to maximize the evidence lower bound. Latent variables are modeled as Gaussian distributions.
PFN福田圭祐による東大大学院「融合情報学特別講義Ⅲ」(2022年10月19日)の講義資料です。
・Introduction to Preferred Networks
・Our developments to date
・Our research & platform
・Simulation ✕ AI
47. GAN 敵対的⽣生成モデル
z
x = G(z)
x
次の⼿手順でxを⽣生成する
(1) z 〜~ U(0, I)でサンプリングする
(2) x = G(z)を計算する
最後にサンプリング
がないことに注意p(z)がGaussianでなく
⼀一様分布Uを使うのも特徴
⾼高次元の⼀一様分布の場合
隅が離離れた表現を扱える
48. GAN 敵対的⽣生成モデルの学習
l 偽物かを判定するD(x)を⽤用意
̶— 本物なら1, 偽物なら0を返す
l Dは上式を最⼤大化するように学習し
Gは最⼩小化するように学習する
̶— この学習はうまく進めば
∫p(z)G(z)dz=P(x), D(x)=1/2という
均衡解にたどり着ける
z
x'
x = G(z)
{1(本物), 0(偽物)}
y = D(x)
x
64. 参考⽂文献
l [Lin+ 16] “Why does deep and cheap learning work so well?”, H. W. Lin, M.
Tegmark
l [Vinnikov+ 14] “K-means Recovers ICA Filters when Independent
Components are Sparse”, ICML 2014, A. Vinnikov, S. S.-Shwartz
l [Kingma+ 13] ”Auto-encoding Variational Bayes”, D. P. Kingma, M. Welling
l [Kingma+ 14] “Semi-supervised Learning with Deep Generative Models”, D.
P. Kingma, D. J. Rezende, S. Mohamed, M. Welling
l [Burda+ 15] “Importance weighted autoencoders”, Y. Burda, R. Grosse, R.
Salakhutdinov
l [Maaloe+ 16] ”Auxiliary Deep Generative Models”, L. Maaloe, c. K.
Sonderby, S. K. Sonderby, O. Winther
l [Goodfellow+ 14] “Gerative Adversarial Networks”, I. J. Goodfellow and et.
al.
65. l [Salimans+ 16] ”Improved Techniques for Training GANs”, T. Salimans, I.
Goodfellow, W. Zaremba, V. Cheung, A. Radford, X. Chen
l [Oord+ 16a] “Pixcel Reucurrent Neural Network”, A. Oord. et al.
l [Oord+ 16b] “Conditional Image Generation with PixelCNN Decoders”, A.
Oord et al.
l [Oord+ 16c] “WaveNet: A Generative Model for Raw Audito”, A. Oord et al.
l [Kim+ 16] “Deep Directed Generative Models with Energy-based Probability
estimation”, T. Kim, Y. Bengio
l [Li+ 15] “Generative Moment Matching Network”, Y. Li, K. Swersky, R.
Zemel
l [Dahl+ 14] “Multi-task Neural Networks for QSAR Predictions”, G. E. Dahl, N.
Jaitly, R. alakhutdinov
l [Lee+ 16] “DeepTarget: End-to-end Learning Framework for microRNA
Target Prediction using Deep Recurrent Neural Networks”, B. Leett, J. Baek,
S. Park, S. Yoon