1. WIT: Wikipedia-based Image Text Dataset for Multimodal Multilingual Machine Learning
Krishna Srinivasan, Karthik Raman, Jiecao Chen, Michael Bendersky, Marc Najork
The milestone improvements brought about by deep representation learning and pre-training techniques have led to large performance gains across downstream NLP, IR and Vision tasks. Multimodal modeling techniques aim to leverage large high-quality visio-linguistic datasets for learning complementary information (across image and text modalities). In this paper, we introduce the Wikipedia-based Image Text (WIT) Dataset (https://github.com/google-research-datasets/wit) to better facilitate multimodal, multilingual learning. WIT is composed of a curated set of 37.6 million entity rich image-text examples with 11.5 million unique images across 108 Wikipedia languages. Its size enables WIT to be used as a pretraining dataset for multimodal models, as we show when applied to downstream tasks such as image-text retrieval. WIT has four main and unique advantages. First, WIT is the largest multimodal dataset by the number of image-text examples by 3x (at the time of writing). Second, WIT is massively multilingual (first of its kind) with coverage over 100+ languages (each of which has at least 12K examples) and provides cross-lingual texts for many images. Third, WIT represents a more diverse set of concepts and real world entities relative to what previous datasets cover. Lastly, WIT provides a very challenging real-world test set, as we empirically illustrate using an image-text retrieval task as an example.
WIT: Wikipedia-based Image Text Dataset for Multimodal
— AK (@ak92501) March 3, 2021
Multilingual Machine Learning
pdf: https://t.co/fblyzH2hGe
abs: https://t.co/tVgBdfOnQ5
github: https://t.co/NNkF3oheok pic.twitter.com/nnFUaPJaYU
2. Contrastive Explanations for Model Interpretability
Alon Jacovi, Swabha Swayamdipta, Shauli Ravfogel, Yanai Elazar, Yejin Choi, Yoav Goldberg
Contrastive explanations clarify why an event occurred in contrast to another. They are more inherently intuitive to humans to both produce and comprehend. We propose a methodology to produce contrastive explanations for classification models by modifying the representation to disregard non-contrastive information, and modifying model behavior to only be based on contrastive reasoning. Our method is based on projecting model representation to a latent space that captures only the features that are useful (to the model) to differentiate two potential decisions. We demonstrate the value of contrastive explanations by analyzing two different scenarios, using both high-level abstract concept attribution and low-level input token/span attribution, on two widely used text classification tasks. Specifically, we produce explanations for answering: for which label, and against which alternative label, is some aspect of the input useful? And which aspects of the input are useful for and against particular decisions? Overall, our findings shed light on the ability of label-contrastive explanations to provide a more accurate and finer-grained interpretability of a model’s decision.
Check out our new paper 😊 (w\ @swabhz @ravfogel @yanaiela @YejinChoinka @yoavgo )
— Alon Jacovi (@alon_jacovi) March 3, 2021
Contrastive Explanations for Model Interpretabilityhttps://t.co/StaGM5q5qN
This paper is about explaining classifier decisions contrastively against alternative decisions. pic.twitter.com/ZRa0Z1xUDc
3. Mixture of Volumetric Primitives for Efficient Neural Rendering
Stephen Lombardi, Tomas Simon, Gabriel Schwartz, Michael Zollhoefer, Yaser Sheikh, Jason Saragih
Real-time rendering and animation of humans is a core function in games, movies, and telepresence applications. Existing methods have a number of drawbacks we aim to address with our work. Triangle meshes have difficulty modeling thin structures like hair, volumetric representations like Neural Volumes are too low-resolution given a reasonable memory budget, and high-resolution implicit representations like Neural Radiance Fields are too slow for use in real-time applications. We present Mixture of Volumetric Primitives (MVP), a representation for rendering dynamic 3D content that combines the completeness of volumetric representations with the efficiency of primitive-based rendering, e.g., point-based or mesh-based methods. Our approach achieves this by leveraging spatially shared computation with a deconvolutional architecture and by minimizing computation in empty regions of space with volumetric primitives that can move to cover only occupied regions. Our parameterization supports the integration of correspondence and tracking constraints, while being robust to areas where classical tracking fails, such as around thin or translucent structures and areas with large topological variability. MVP is a hybrid that generalizes both volumetric and primitive-based representations. Through a series of extensive experiments we demonstrate that it inherits the strengths of each, while avoiding many of their limitations. We also compare our approach to several state-of-the-art methods and demonstrate that MVP produces superior results in terms of quality and runtime performance.
Mixture of Volumetric Primitives for Efficient Neural Rendering
— AK (@ak92501) March 3, 2021
pdf: https://t.co/N3L5PLVWGb
abs: https://t.co/TpbAOLSJZk pic.twitter.com/PRURQky6Pl
4. Predicting Video with VQVAE
Jacob Walker, Ali Razavi, Aäron van den Oord
In recent years, the task of video prediction-forecasting future video given past video frames-has attracted attention in the research community. In this paper we propose a novel approach to this problem with Vector Quantized Variational AutoEncoders (VQ-VAE). With VQ-VAE we compress high-resolution videos into a hierarchical set of multi-scale discrete latent variables. Compared to pixels, this compressed latent space has dramatically reduced dimensionality, allowing us to apply scalable autoregressive generative models to predict video. In contrast to previous work that has largely emphasized highly constrained datasets, we focus on very diverse, large-scale datasets such as Kinetics-600. We predict video at a higher resolution on unconstrained videos, 256x256, than any other previous method to our knowledge. We further validate our approach against prior work via a crowdsourced human evaluation.
Predicting Video with VQVAE
— AK (@ak92501) March 3, 2021
pdf: https://t.co/E6Ij9rCOlL
abs: https://t.co/joGZcCUWlC pic.twitter.com/1NQMfrZlxh
5. Categorical Foundations of Gradient-Based Learning
G.S.H. Cruttwell, Bruno Gavranović, Neil Ghani, Paul Wilson, Fabio Zanasi
We propose a categorical foundation of gradient-based machine learning algorithms in terms of lenses, parametrised maps, and reverse derivative categories. This foundation provides a powerful explanatory and unifying framework: it encompasses a variety of gradient descent algorithms such as ADAM, AdaGrad, and Nesterov momentum, as well as a variety of loss functions such as as MSE and Softmax cross-entropy, shedding new light on their similarities and differences. Our approach also generalises beyond neural networks (modelled in categories of smooth maps), accounting for other structures relevant to gradient-based learning such as boolean circuits. Finally, we also develop a novel implementation of gradient-based learning in Python, informed by the principles introduced by our framework.
Our paper on "Categorical Foundations of Gradient-Based Learning" is out! (https://t.co/aREoHErf0L), It's accompanied by a very short blog post (https://t.co/qdEbl63zuH) describing some of the main ideas, as well as a video with the presentation https://t.co/gGi6vF5YXk
— Bruno Gavranović (@bgavran3) March 4, 2021
Categorical foundations of gradient-based learning, by Geoffrey Cruttwell, @bgavran3, @Anarchia45, @statusfailed and Fabio Zanasi
— julesh (@_julesh_) March 3, 2021
I believe this is quite an important paper for applied category theory!https://t.co/36QAUodnaw pic.twitter.com/cZ0LWTaDJq
Cruttwell et al "Categorical Foundations of Gradient-Based Learning"https://t.co/E9SEOEHFtx
— theHigherGeometer (@HigherGeometer) March 3, 2021
should be some juicy goodies in there. Also, an implementation (in Python)!
6. Reinforcement Learning for Adaptive Mesh Refinement
Jiachen Yang, Tarik Dzanic, Brenden Petersen, Jun Kudo, Ketan Mittal, Vladimir Tomov, Jean-Sylvain Camier, Tuo Zhao, Hongyuan Zha, Tzanio Kolev, Robert Anderson, Daniel Faissol
Large-scale finite element simulations of complex physical systems governed by partial differential equations crucially depend on adaptive mesh refinement (AMR) to allocate computational budget to regions where higher resolution is required. Existing scalable AMR methods make heuristic refinement decisions based on instantaneous error estimation and thus do not aim for long-term optimality over an entire simulation. We propose a novel formulation of AMR as a Markov decision process and apply deep reinforcement learning (RL) to train refinement policies directly from simulation. AMR poses a new problem for RL in that both the state dimension and available action set changes at every step, which we solve by proposing new policy architectures with differing generality and inductive bias. The model sizes of these policy architectures are independent of the mesh size and hence scale to arbitrarily large and complex simulations. We demonstrate in comprehensive experiments on static function estimation and the advection of different fields that RL policies can be competitive with a widely-used error estimator and generalize to larger, more complex, and unseen test problems.
Reinforcement Learning for Adaptive Mesh Refinement
— AK (@ak92501) March 3, 2021
pdf: https://t.co/6pxJhHmh4I
abs: https://t.co/wEX7r6G1rQ
project page: https://t.co/Xk4NdmqszT pic.twitter.com/VygqoYLwHP
7. PyQUBO: Python Library for Mapping Combinatorial Optimization Problems to QUBO Form
Mashiyat Zaman, Kotaro Tanahashi, Shu Tanaka
We present PyQUBO, an open-source, Python library for constructing quadratic unconstrained binary optimizations (QUBOs) from the objective functions and the constraints of optimization problems. PyQUBO enables users to prepare QUBOs or Ising models for various combinatorial optimization problems with ease thanks to the abstraction of expressions and the extensibility of the program. QUBOs and Ising models formulated using PyQUBO are solvable by Ising machines, including quantum annealing machines. We introduce the features of PyQUBO with applications in the number partitioning problem, knapsack problem, graph coloring problem, and integer factorization using a binary multiplier. Moreover, we demonstrate how PyQUBO can be applied to production-scale problems through integration with quantum annealing machines. Through its flexibility and ease of use, PyQUBO has the potential to make quantum annealing a more practical tool among researchers.
arXivに論文プレプリントをアップロードしました!
— Shu Tanaka (@tnksh) March 3, 2021
量子アニーリング等イジングマシンを利用するためのOSS「PYQUBO」(棚橋さんが数年前から開発。2018年9月より公開)に関する論文
PyQUBO: Python Library for Mapping Combinatorial Optimization Problems to QUBO Formhttps://t.co/Wte56CqAO6
8. Kernel Interpolation for Scalable Online Gaussian Processes
Samuel Stanton, Wesley J. Maddox, Ian Delbridge, Andrew Gordon Wilson
Gaussian processes (GPs) provide a gold standard for performance in online settings, such as sample-efficient control and black box optimization, where we need to update a posterior distribution as we acquire data in a sequential fashion. However, updating a GP posterior to accommodate even a single new observation after having observed points incurs at least computations in the exact setting. We show how to use structured kernel interpolation to efficiently recycle computations for constant-time online updates with respect to the number of points , while retaining exact inference. We demonstrate the promise of our approach in a range of online regression and classification settings, Bayesian optimization, and active sampling to reduce error in malaria incidence forecasting. Code is available at https://github.com/wjmaddox/online_gp.
In our new #AISTATS2021 paper, "Kernel Interpolation for Scalable Online Gaussian Processes", we show how to do O(1) streaming Bayesian updates, while retaining exact inference! https://t.co/i4TeqoSVz3
— Andrew Gordon Wilson (@andrewgwils) March 4, 2021
with @samscub, W. Maddox, @DelbridgeIan. 1/6 pic.twitter.com/YyWQdhClJL
9. Median Optimal Treatment Regimes
Liu Leqi, Edward H. Kennedy
Optimal treatment regimes are personalized policies for making a treatment decision based on subject characteristics, with the policy chosen to maximize some value. It is common to aim to maximize the mean outcome in the population, via a regime assigning treatment only to those whose mean outcome is higher under treatment versus control. However, the mean can be an unstable measure of centrality, resulting in imprecise statistical procedures, as well as unfair decisions that can be overly influenced by a small fraction of subjects. In this work, we propose a new median optimal treatment regime that instead treats individuals whose conditional median is higher under treatment. This ensures that optimal decisions for individuals from the same group are not overly influenced either by (i) a small fraction of the group (unlike the mean criterion), or (ii) unrelated subjects from different groups (unlike marginal median/quantile criteria). We introduce a new measure of value, the Average Conditional Median Effect (ACME), which summarizes across-group median treatment outcomes of a policy, and which the optimal median treatment regime maximizes. After developing key motivating examples that distinguish median optimal treatment regimes from mean and marginal median optimal treatment regimes, we give a nonparametric efficiency bound for estimating the ACME of a policy, and propose a new doubly robust-style estimator that achieves the efficiency bound under weak conditions. Finite-sample properties of the estimator are explored via numerical simulations and the proposed algorithm is illustrated using data from a randomized clinical trial in patients with HIV.
Cool new paper by @leqi_liu!
— Edward Kennedy (@edwardhkennedy) March 3, 2021
Mean optimal trt rules aren't robust: sensitive to small % w/ extreme outcomes
Marginal median opt rules are unfair in different way: my treatment depends on *your* outcomes
We study fair+robust median optimal trt rules:https://t.co/oXWwRKYnSh pic.twitter.com/ieQoLS8GYX
10. Image-to-image Translation via Hierarchical Style Disentanglement
Xinyang Li, Shengchuan Zhang, Jie Hu, Liujuan Cao, Xiaopeng Hong, Xudong Mao, Feiyue Huang, Yongjian Wu, Rongrong Ji
Recently, image-to-image translation has made significant progress in achieving both multi-label (\ie, translation conditioned on different labels) and multi-style (\ie, generation with diverse styles) tasks. However, due to the unexplored independence and exclusiveness in the labels, existing endeavors are defeated by involving uncontrolled manipulations to the translation results. In this paper, we propose Hierarchical Style Disentanglement (HiSD) to address this issue. Specifically, we organize the labels into a hierarchical tree structure, in which independent tags, exclusive attributes, and disentangled styles are allocated from top to bottom. Correspondingly, a new translation process is designed to adapt the above structure, in which the styles are identified for controllable translations. Both qualitative and quantitative results on the CelebA-HQ dataset verify the ability of the proposed HiSD. We hope our method will serve as a solid baseline and provide fresh insights with the hierarchically organized annotations for future research in image-to-image translation. The code has been released at https://github.com/imlixinyang/HiSD.
Image-to-image Translation via Hierarchical Style Disentanglement
— AK (@ak92501) March 3, 2021
pdf: https://t.co/WgqJRgGnrY
abs: https://t.co/QLeuF76wlL pic.twitter.com/uJx0kUF66w
11. An Analysis of Distributed Systems Syllabi With a Focus on Performance-Related Topics
Cristina L. Abad, Alexandru Iosup, Edwin F. Boza, Eduardo Ortiz-Holguin
We analyze a dataset of 51 current (2019-2020) Distributed Systems syllabi from top Computer Science programs, focusing on finding the prevalence and context in which topics related to performance are being taught in these courses. We also study the scale of the infrastructure mentioned in DS courses, from small client-server systems to cloud-scale, peer-to-peer, global-scale systems. We make eight main findings, covering goals such as performance, and scalability and its variant elasticity; activities such as performance benchmarking and monitoring; eight selected performance-enhancing techniques (replication, caching, sharding, load balancing, scheduling, streaming, migrating, and offloading); and control issues such as trade-offs that include performance and performance variability.
As members of @spec_perf Research Group, we also took a look at Distributed Systems syllabi focusing on performance-related topics. Paper to be presented at Workshop on Education & Practice of Performance Eng. at @ICPEconf: https://t.co/hVstqKujWz cc @AIosup @edwinboza @leortyz https://t.co/hPmJnjcRL1 pic.twitter.com/nV21sub2VH
— Cristina L. Abad (@cabad3) March 3, 2021
12. Patch-NetVLAD: Multi-Scale Fusion of Locally-Global Descriptors for Place Recognition
Stephen Hausler, Sourav Garg, Ming Xu, Michael Milford, Tobias Fischer
Visual Place Recognition is a challenging task for robotics and autonomous systems, which must deal with the twin problems of appearance and viewpoint change in an always changing world. This paper introduces Patch-NetVLAD, which provides a novel formulation for combining the advantages of both local and global descriptor methods by deriving patch-level features from NetVLAD residuals. Unlike the fixed spatial neighborhood regime of existing local keypoint features, our method enables aggregation and matching of deep-learned local features defined over the feature-space grid. We further introduce a multi-scale fusion of patch features that have complementary scales (i.e. patch sizes) via an integral feature space and show that the fused features are highly invariant to both condition (season, structure, and illumination) and viewpoint (translation and rotation) changes. Patch-NetVLAD outperforms both global and local feature descriptor-based methods with comparable compute, achieving state-of-the-art visual place recognition results on a range of challenging real-world datasets, including winning the Facebook Mapillary Visual Place Recognition Challenge at ECCV2020. It is also adaptable to user requirements, with a speed-optimised version operating over an order of magnitude faster than the state-of-the-art. By combining superior performance with improved computational efficiency in a configurable framework, Patch-NetVLAD is well suited to enhance both stand-alone place recognition capabilities and the overall performance of SLAM systems.
Incredibly happy that our @CVPR #CVPR2021 submission "Patch-NetVLAD: Multi-Scale Fusion of Locally-Global Descriptors for Place Recognition" has been accepted - preprint: https://t.co/hifE1u9P31
— Tobias Fischer (@TobiasRobotics) March 4, 2021
Great teamwork by @Dalek25 @sourav_garg_ Ming Xu and @maththrills @QUTRobotics. 1/n pic.twitter.com/NW6DU0xX9K
13. A practical tutorial on Variational Bayes
Minh-Ngoc Tran, Trong-Nghia Nguyen, Viet-Hung Dao
This tutorial gives a quick introduction to Variational Bayes (VB), also called Variational Inference or Variational Approximation, from a practical point of view. The paper covers a range of commonly used VB methods and an attempt is made to keep the materials accessible to the wide community of data analysis practitioners. The aim is that the reader can quickly derive and implement their first VB algorithm for Bayesian inference with their data analysis problem. An end-user software package in Matlab together with the documentation can be found at https://vbayeslab.github.io/VBLabDocs/
A practical tutorial on Variational Bayes. (arXiv:2103.01327v1 [https://t.co/uFsa4flLGi]) https://t.co/4OpQ53KQFq
— Stat.ML Papers (@StatMLPapers) March 3, 2021
14. Listen, Read, and Identify: Multimodal Singing Language Identification
Keunwoo Choi, Yuxuan Wang
We propose a multimodal singing language classification model that uses both audio content and textual metadata. LRID-Net, the proposed model, takes an audio signal and a language probability vector estimated from the metadata and outputs the probabilities of the ten target languages. Optionally, LRID-Net is facilitated with modality dropouts to handle a missing modality. In the experiment, we trained several LRID-Nets with varying modality dropout configuration and test them with various combinations of input modalities. The experiment results demonstrate that using multimodal input improves the performance. The results also suggest that adopting modality dropout does not degrade performance of the model when there are full modality inputs while enabling the model to handle missing modality cases to some extent.
https://t.co/frxIYHmuQs Long time no first-authoring! Listen, Read, and Identify network (LRID-Net) identifies singing language by reading the metadata (title, album, artist) and listening to the audio.
— Keunwoo Choi (@keunwoochoi) March 3, 2021
15. A Deep Emulator for Secondary Motion of 3D Characters
Mianlun Zheng, Yi Zhou, Duygu Ceylan, Jernej Barbic
Fast and light-weight methods for animating 3D characters are desirable in various applications such as computer games. We present a learning-based approach to enhance skinning-based animations of 3D characters with vivid secondary motion effects. We design a neural network that encodes each local patch of a character simulation mesh where the edges implicitly encode the internal forces between the neighboring vertices. The network emulates the ordinary differential equations of the character dynamics, predicting new vertex positions from the current accelerations, velocities and positions. Being a local method, our network is independent of the mesh topology and generalizes to arbitrarily shaped 3D character meshes at test time. We further represent per-vertex constraints and material properties such as stiffness, enabling us to easily adjust the dynamics in different parts of the mesh. We evaluate our method on various character meshes and complex motion sequences. Our method can be over 30 times more efficient than ground-truth physically based simulation, and outperforms alternative solutions that provide fast approximations.