1. Coordinate Independent Convolutional Networks — Isometry and Gauge Equivariant Convolutions on Riemannian Manifolds
Maurice Weiler, Patrick Forré, Erik Verlinde, Max Welling
- retweets: 16856, favorites: 4 (06/15/2021 07:49:25)
- links: abs | pdf
- cs.LG | cs.CG | cs.CV | stat.ML
Motivated by the vast success of deep convolutional networks, there is a great interest in generalizing convolutions to non-Euclidean manifolds. A major complication in comparison to flat spaces is that it is unclear in which alignment a convolution kernel should be applied on a manifold. The underlying reason for this ambiguity is that general manifolds do not come with a canonical choice of reference frames (gauge). Kernels and features therefore have to be expressed relative to arbitrary coordinates. We argue that the particular choice of coordinatization should not affect a network’s inference — it should be coordinate independent. A simultaneous demand for coordinate independence and weight sharing is shown to result in a requirement on the network to be equivariant under local gauge transformations (changes of local reference frames). The ambiguity of reference frames depends thereby on the G-structure of the manifold, such that the necessary level of gauge equivariance is prescribed by the corresponding structure group G. Coordinate independent convolutions are proven to be equivariant w.r.t. those isometries that are symmetries of the G-structure. The resulting theory is formulated in a coordinate free fashion in terms of fiber bundles. To exemplify the design of coordinate independent convolutions, we implement a convolutional network on the M”obius strip. The generality of our differential geometric formulation of convolutional networks is demonstrated by an extensive literature review which explains a large number of Euclidean CNNs, spherical CNNs and CNNs on general surfaces as specific instances of coordinate independent convolutions.
Happy to announce our work on Coordinate Independent Convolutional Networks.
— Maurice Weiler (@maurice_weiler) June 14, 2021
It develops a theory of CNNs on Riemannian manifolds and clarifies the interplay of the kernels' local gauge equivariance and the networks' global isometry equivariance.https://t.co/Vb0KHlTJ7S
[1/N] pic.twitter.com/TU3FusKk7M
2. LocoProp: Enhancing BackProp via Local Loss Optimization
Ehsan Amid, Rohan Anil, Manfred K. Warmuth
We study a local loss construction approach for optimizing neural networks. We start by motivating the problem as minimizing a squared loss between the pre-activations of each layer and a local target, plus a regularizer term on the weights. The targets are chosen so that the first gradient descent step on the local objectives recovers vanilla BackProp, while the exact solution to each problem results in a preconditioned gradient update. We improve the local loss construction by forming a Bregman divergence in each layer tailored to the transfer function which keeps the local problem convex w.r.t. the weights. The generalized local problem is again solved iteratively by taking small gradient descent steps on the weights, for which the first step recovers BackProp. We run several ablations and show that our construction consistently improves convergence, reducing the gap between first-order and second-order methods.
Excited to share our progress on faster training of deep neural networks!
— Rohan Anil 🏠🤯 (@_arohan_) June 14, 2021
With the amazing Ehsan Amid (@esiamid) & Manfred K. Warmuth
Achieves convergence close to second-order methods e.g. Shampoo & K-FAC while significantly faster in walltime.
🧵👇https://t.co/KwZroZIiUU pic.twitter.com/CDYSdkRElB
3. Graph Neural Networks for Natural Language Processing: A Survey
Lingfei Wu, Yu Chen, Kai Shen, Xiaojie Guo, Hanning Gao, Shucheng Li, Jian Pei, Bo Long
Deep learning has become the dominant approach in coping with various tasks in Natural LanguageProcessing (NLP). Although text inputs are typically represented as a sequence of tokens, there isa rich variety of NLP problems that can be best expressed with a graph structure. As a result, thereis a surge of interests in developing new deep learning techniques on graphs for a large numberof NLP tasks. In this survey, we present a comprehensive overview onGraph Neural Networks(GNNs) for Natural Language Processing. We propose a new taxonomy of GNNs for NLP, whichsystematically organizes existing research of GNNs for NLP along three axes: graph construction,graph representation learning, and graph based encoder-decoder models. We further introducea large number of NLP applications that are exploiting the power of GNNs and summarize thecorresponding benchmark datasets, evaluation metrics, and open-source codes. Finally, we discussvarious outstanding challenges for making the full use of GNNs for NLP as well as future researchdirections. To the best of our knowledge, this is the first comprehensive overview of Graph NeuralNetworks for Natural Language Processing.
🔥 Graph Neural Networks for Natural Language Processing: A Survey
— elvis (@omarsar0) June 14, 2021
Lots of NLP tasks and applications can benefit from deep learning techniques on graphs.
Learn about these in one of the most comprehensive overviews of GNNs for NLP I've seen.https://t.co/EsjWHxhKsN pic.twitter.com/Ut17E5hlpZ
4. Scaling Vision with Sparse Mixture of Experts
Carlos Riquelme, Joan Puigcerver, Basil Mustafa, Maxim Neumann, Rodolphe Jenatton, André Susano Pinto, Daniel Keysers, Neil Houlsby
Sparsely-gated Mixture of Experts networks (MoEs) have demonstrated excellent scalability in Natural Language Processing. In Computer Vision, however, almost all performant networks are “dense”, that is, every input is processed by every parameter. We present a Vision MoE (V-MoE), a sparse version of the Vision Transformer, that is scalable and competitive with the largest dense networks. When applied to image recognition, V-MoE matches the performance of state-of-the-art networks, while requiring as little as half of the compute at inference time. Further, we propose an extension to the routing algorithm that can prioritize subsets of each input across the entire batch, leading to adaptive per-image compute. This allows V-MoE to trade-off performance and compute smoothly at test-time. Finally, we demonstrate the potential of V-MoE to scale vision models, and train a 15B parameter model that attains 90.35% on ImageNet.
Check out what we’ve been working on for the last months: We decouple the model size and the compute cost in a Vision Transformer backbone by using Sparse MoE layers. These have been popularised in NLP, and they are fantastic for Vision too! https://t.co/AdDwJKvkBk pic.twitter.com/apg9sLYirY
— Joan Puigcerver (@joapuipe) June 14, 2021
Scaling Vision with Sparse Mixture of Experts
— Aran Komatsuzaki (@arankomatsuzaki) June 14, 2021
Trains a 15B Vision Transformer with MoE that attains 90.35% on ImageNet.https://t.co/UitlGeymrv pic.twitter.com/t3B3Mtcu1z
[3/3] Towards big vision
— Neil Houlsby (@neilhoulsby) June 14, 2021
While dense models are still the norm, sparse MoE layers can work well too!
Large Vision-MoEs (15B params) can be trained to high performance relatively efficiently, and can even prioritize amongst patches (see duck).https://t.co/0WHyJTlGG5
... pic.twitter.com/INsCIiSmIi
Scaling Vision with Sparse Mixture of Experts
— AK (@ak92501) June 14, 2021
pdf: https://t.co/GoTcuIGldr
abs: https://t.co/ZurZ2FhZnm
demonstrate the potential of V-MoE to scale vision models, and train a 15B parameter model that attains 90.35% on ImageNet pic.twitter.com/nOOyWPK9w4
Can we scale deep vision models to billions of parameters? Yes! By only activating the relevant parts of the network for each input. We present the Vision Mixture of Experts & train a 15B-parameter model with 24 routers; transfers to ImageNet w. 90.35% acc https://t.co/QzgLQfKGZt
— Carlos Riquelme (@rikelhood) June 14, 2021
5. SimSwap: An Efficient Framework For High Fidelity Face Swapping
Renwang Chen, Xuanhong Chen, Bingbing Ni, Yanhao Ge
We propose an efficient framework, called Simple Swap (SimSwap), aiming for generalized and high fidelity face swapping. In contrast to previous approaches that either lack the ability to generalize to arbitrary identity or fail to preserve attributes like facial expression and gaze direction, our framework is capable of transferring the identity of an arbitrary source face into an arbitrary target face while preserving the attributes of the target face. We overcome the above defects in the following two ways. First, we present the ID Injection Module (IIM) which transfers the identity information of the source face into the target face at feature level. By using this module, we extend the architecture of an identity-specific face swapping algorithm to a framework for arbitrary face swapping. Second, we propose the Weak Feature Matching Loss which efficiently helps our framework to preserve the facial attributes in an implicit way. Extensive experiments on wild faces demonstrate that our SimSwap is able to achieve competitive identity performance while preserving attributes better than previous state-of-the-art methods. The code is already available on github: https://github.com/neuralchen/SimSwap.
SimSwap: An Efficient Framework For High Fidelity
— AK (@ak92501) June 14, 2021
Face Swapping
pdf: https://t.co/l2aWTrM1CP
abs: https://t.co/ZSuDnRLUuF
github: https://t.co/deYKr8rhLY pic.twitter.com/cBuaXySkd9
6. MlTr: Multi-label Classification with Transformer
Xing Cheng, Hezheng Lin, Xiangyu Wu, Fan Yang, Dong Shen, Zhongyuan Wang, Nian Shi, Honglin Liu
The task of multi-label image classification is to recognize all the object labels presented in an image. Though advancing for years, small objects, similar objects and objects with high conditional probability are still the main bottlenecks of previous convolutional neural network(CNN) based models, limited by convolutional kernels’ representational capacity. Recent vision transformer networks utilize the self-attention mechanism to extract the feature of pixel granularity, which expresses richer local semantic information, while is insufficient for mining global spatial dependence. In this paper, we point out the three crucial problems that CNN-based methods encounter and explore the possibility of conducting specific transformer modules to settle them. We put forward a Multi-label Transformer architecture(MlTr) constructed with windows partitioning, in-window pixel attention, cross-window attention, particularly improving the performance of multi-label image classification tasks. The proposed MlTr shows state-of-the-art results on various prevalent multi-label datasets such as MS-COCO, Pascal-VOC, and NUS-WIDE with 88.5%, 95.8%, and 65.5% respectively. The code will be available soon at https://github.com/starmemda/MlTr/
MlTr: Multi-label Classification with Transformer
— AK (@ak92501) June 14, 2021
pdf: https://t.co/wqvd89AtJq
abs: https://t.co/H2n64N5OGa pic.twitter.com/yXFM5caLMK
7. Dynamic Language Models for Continuously Evolving Content
Spurthi Amba Hombaiah, Tao Chen, Mingyang Zhang, Michael Bendersky, Marc Najork
The content on the web is in a constant state of flux. New entities, issues, and ideas continuously emerge, while the semantics of the existing conversation topics gradually shift. In recent years, pre-trained language models like BERT greatly improved the state-of-the-art for a large spectrum of content understanding tasks. Therefore, in this paper, we aim to study how these language models can be adapted to better handle continuously evolving web content. In our study, we first analyze the evolution of 2013 - 2019 Twitter data, and unequivocally confirm that a BERT model trained on past tweets would heavily deteriorate when directly applied to data from later years. Then, we investigate two possible sources of the deterioration: the semantic shift of existing tokens and the sub-optimal or failed understanding of new tokens. To this end, we both explore two different vocabulary composition methods, as well as propose three sampling methods which help in efficient incremental training for BERT-like models. Compared to a new model trained from scratch offline, our incremental training (a) reduces the training costs, (b) achieves better performance on evolving content, and (c) is suitable for online deployment. The superiority of our methods is validated using two downstream tasks. We demonstrate significant improvements when incrementally evolving the model from a particular base year, on the task of Country Hashtag Prediction, as well as on the OffensEval 2019 task.
Dynamic Language Models for Continuously Evolving Content
— AK (@ak92501) June 14, 2021
pdf: https://t.co/dlkofC1MZ9
abs: https://t.co/Ibed9GQSMJ pic.twitter.com/W6HoCer8ta
8. Part-aware Panoptic Segmentation
Daan de Geus, Panagiotis Meletis, Chenyang Lu, Xiaoxiao Wen, Gijs Dubbelman
In this work, we introduce the new scene understanding task of Part-aware Panoptic Segmentation (PPS), which aims to understand a scene at multiple levels of abstraction, and unifies the tasks of scene parsing and part parsing. For this novel task, we provide consistent annotations on two commonly used datasets: Cityscapes and Pascal VOC. Moreover, we present a single metric to evaluate PPS, called Part-aware Panoptic Quality (PartPQ). For this new task, using the metric and annotations, we set multiple baselines by merging results of existing state-of-the-art methods for panoptic segmentation and part segmentation. Finally, we conduct several experiments that evaluate the importance of the different levels of abstraction in this single task.
Part-aware Panoptic Segmentation
— AK (@ak92501) June 14, 2021
pdf: https://t.co/519l92lKsm
abs: https://t.co/tm2UsWNnlz
github: https://t.co/uGYVU0PPmZ pic.twitter.com/Fu1xRslPPP
9. Conditional Variational Autoencoder with Adversarial Learning for End-to-End Text-to-Speech
Jaehyeon Kim, Jungil Kong, Juhee Son
Several recent end-to-end text-to-speech (TTS) models enabling single-stage training and parallel sampling have been proposed, but their sample quality does not match that of two-stage TTS systems. In this work, we present a parallel end-to-end TTS method that generates more natural sounding audio than current two-stage models. Our method adopts variational inference augmented with normalizing flows and an adversarial training process, which improves the expressive power of generative modeling. We also propose a stochastic duration predictor to synthesize speech with diverse rhythms from input text. With the uncertainty modeling over latent variables and the stochastic duration predictor, our method expresses the natural one-to-many relationship in which a text input can be spoken in multiple ways with different pitches and rhythms. A subjective human evaluation (mean opinion score, or MOS) on the LJ Speech, a single speaker dataset, shows that our method outperforms the best publicly available TTS systems and achieves a MOS comparable to ground truth.
Conditional Variational Autoencoder with Adversarial Learning for End-to-End Text-to-Speech 🗣️
— AK (@ak92501) June 14, 2021
pdf: https://t.co/7YqvteNvDv
abs: https://t.co/uf8lt52BYd
github: https://t.co/GnZOHaS7o5
project page: https://t.co/ZQPcUhuOZ3 pic.twitter.com/ViCgjeYXDZ
LJ Speech: Game over.https://t.co/yFGqLEM1ZZ pic.twitter.com/DwFLzhhvoy
— Seung-won Park (@veydpz_public) June 14, 2021
10. View Generalization for Single Image Textured 3D Models
Anand Bhattad, Aysegul Dundar, Guilin Liu, Andrew Tao, Bryan Catanzaro
Humans can easily infer the underlying 3D geometry and texture of an object only from a single 2D image. Current computer vision methods can do this, too, but suffer from view generalization problems - the models inferred tend to make poor predictions of appearance in novel views. As for generalization problems in machine learning, the difficulty is balancing single-view accuracy (cf. training error; bias) with novel view accuracy (cf. test error; variance). We describe a class of models whose geometric rigidity is easily controlled to manage this tradeoff. We describe a cycle consistency loss that improves view generalization (roughly, a model from a generated view should predict the original view well). View generalization of textures requires that models share texture information, so a car seen from the back still has headlights because other cars have headlights. We describe a cycle consistency loss that encourages model textures to be aligned, so as to encourage sharing. We compare our method against the state-of-the-art method and show both qualitative and quantitative improvements.
View Generalization for Single Image Textured 3D Models
— AK (@ak92501) June 14, 2021
pdf: https://t.co/cHbMfNofTH
abs: https://t.co/0a0YVyb8Qt
a new 3D inference pipeline, principally designed to display good view generalization pic.twitter.com/kzbNdqT1Pz
11. Scalable Variational Gaussian Processes via Harmonic Kernel Decomposition
Shengyang Sun, Jiaxin Shi, Andrew Gordon Wilson, Roger Grosse
We introduce a new scalable variational Gaussian process approximation which provides a high fidelity approximation while retaining general applicability. We propose the harmonic kernel decomposition (HKD), which uses Fourier series to decompose a kernel as a sum of orthogonal kernels. Our variational approximation exploits this orthogonality to enable a large number of inducing points at a low computational cost. We demonstrate that, on a range of regression and classification problems, our approach can exploit input space symmetries such as translations and reflections, and it significantly outperforms standard variational methods in scalability and accuracy. Notably, our approach achieves state-of-the-art results on CIFAR-10 among pure GP models.
Excited to share our HVGP @ICML2021
— Shengyang Sun (@ssydasheng) June 14, 2021
:
1) discrete Fourier transform meets kernels
2) decomposing a kernel into an orthogonal sum of kernels
3) a scalable SVGP to use more inducing points
4) being applicable to common kernels: RBF, Matern, poly, ...https://t.co/3WZElhCOdt pic.twitter.com/6v2LHnIPXg
12. Preferential Temporal Difference Learning
Nishanth Anand, Doina Precup
Temporal-Difference (TD) learning is a general and very useful tool for estimating the value function of a given policy, which in turn is required to find good policies. Generally speaking, TD learning updates states whenever they are visited. When the agent lands in a state, its value can be used to compute the TD-error, which is then propagated to other states. However, it may be interesting, when computing updates, to take into account other information than whether a state is visited or not. For example, some states might be more important than others (such as states which are frequently seen in a successful trajectory). Or, some states might have unreliable value estimates (for example, due to partial observability or lack of data), making their values less desirable as targets. We propose an approach to re-weighting states used in TD updates, both when they are the input and when they provide the target for the update. We prove that our approach converges with linear function approximation and illustrate its desirable empirical behaviour compared to other TD-style methods.
Did you know that Temporal Difference (TD) Learning has a major flaw?
— Nishanth Anand (@itsNVA7) June 14, 2021
Happy to announce Preferential Temporal Difference Learning (accepted at the @icmlconf) that fixes these!
Work supervised by Doina Precup. Link: https://t.co/j0UPOUHs5k
More in the thread below. (1/6)
13. Catch-A-Waveform: Learning to Generate Audio from a Single Short Example
Gal Greshler, Tamar Rott Shaham, Tomer Michaeli
Models for audio generation are typically trained on hours of recordings. Here, we illustrate that capturing the essence of an audio source is typically possible from as little as a few tens of seconds from a single training signal. Specifically, we present a GAN-based generative model that can be trained on one short audio signal from any domain (e.g. speech, music, etc.) and does not require pre-training or any other form of external supervision. Once trained, our model can generate random samples of arbitrary duration that maintain semantic similarity to the training waveform, yet exhibit new compositions of its audio primitives. This enables a long line of interesting applications, including generating new jazz improvisations or new a-cappella rap variants based on a single short example, producing coherent modifications to famous songs (e.g. adding a new verse to a Beatles song based solely on the original recording), filling-in of missing parts (inpainting), extending the bandwidth of a speech signal (super-resolution), and enhancing old recordings without access to any clean training example. We show that in all cases, no more than 20 seconds of training audio commonly suffice for our model to achieve state-of-the-art results. This is despite its complete lack of prior knowledge about the nature of audio signals in general.
Catch-A-Waveform: Learning to Generate Audio
— AK (@ak92501) June 14, 2021
from a Single Short Example
pdf: https://t.co/hXZ7oPhBI1
abs: https://t.co/TwIkJGvbHi
project page: https://t.co/KWNr0YIU4o
no more than 20 seconds of training audio commonly suffice for model to achieve sota results pic.twitter.com/N2oZxsz2kh
14. Zero-Shot Controlled Generation with Encoder-Decoder Transformers
Devamanyu Hazarika, Mahdi Namazifar, Dilek Hakkani-Tür
Controlling neural network-based models for natural language generation (NLG) has broad applications in numerous areas such as machine translation, document summarization, and dialog systems. Approaches that enable such control in a zero-shot manner would be of great importance as, among other reasons, they remove the need for additional annotated data and training. In this work, we propose novel approaches for controlling encoder-decoder transformer-based NLG models in a zero-shot manner. This is done by introducing three control knobs; namely, attention biasing, decoder mixing, and context augmentation, that are applied to these models at generation time. These knobs control the generation process by directly manipulating trained NLG models (e.g., biasing cross-attention layers) to realize the desired attributes in the generated outputs. We show that not only are these NLG models robust to such manipulations, but also their behavior could be controlled without an impact on their generation performance. These results, to the best of our knowledge, are the first of their kind. Through these control knobs, we also investigate the role of transformer decoder’s self-attention module and show strong evidence that its primary role is maintaining fluency of sentences generated by these models. Based on this hypothesis, we show that alternative architectures for transformer decoders could be viable options. We also study how this hypothesis could lead to more efficient ways for training encoder-decoder transformer models.
Zero-Shot Controlled Generation with Encoder-Decoder Transformers
— AK (@ak92501) June 14, 2021
pdf: https://t.co/Ak2pHvh9J5
abs: https://t.co/9npeMy0kHa
can achieve zero-shot controllability for models that are orders of magnitude smaller than GPT-3 (e.g., BART-base) pic.twitter.com/O3yL5PzJgw
15. Recovery of Meteorites Using an Autonomous Drone and Machine Learning
Robert I. Citron, Peter Jenniskens, Christopher Watkins, Sravanthi Sinha, Amar Shah, Chedy Raissi, Hadrien Devillepoix, Jim Albers
- retweets: 73, favorites: 27 (06/15/2021 07:49:28)
- links: abs | pdf
- astro-ph.EP | cs.CV | cs.LG
The recovery of freshly fallen meteorites from tracked and triangulated meteors is critical to determining their source asteroid families. However, locating meteorite fragments in strewn fields remains a challenge with very few meteorites being recovered from the meteors triangulated in past and ongoing meteor camera networks. We examined if locating meteorites can be automated using machine learning and an autonomous drone. Drones can be programmed to fly a grid search pattern and take systematic pictures of the ground over a large survey area. Those images can be analyzed using a machine learning classifier to identify meteorites in the field among many other features. Here, we describe a proof-of-concept meteorite classifier that deploys off-line a combination of different convolution neural networks to recognize meteorites from images taken by drones in the field. The system was implemented in a conceptual drone setup and tested in the suspected strewn field of a recent meteorite fall near Walker Lake, Nevada.
Researchers have taught a drone to recognize and retrieve meteorites autonomously - https://t.co/7sZVEGUOLX pic.twitter.com/ux8UlTBoPG
— Fraser Cain (@fcain) June 14, 2021
16. Where to Encode: A Performance Analysis of x86 and Arm-based Amazon EC2 Instance
Roland Mathá, Dragi Kimovski, Anatoliy Zabrovskiy, Christian Timmerer, Radu Prodan
Video streaming became an undivided part of the Internet. To efficiently utilize the limited network bandwidth it is essential to encode the video content. However, encoding is a computationally intensive task, involving high-performance resources provided by private infrastructures or public clouds. Public clouds, such as Amazon EC2, provide a large portfolio of services and instances optimized for specific purposes and budgets. The majority of Amazon instances use x86 processors, such as Intel Xeon or AMD EPYC. However, following the recent trends in computer architecture, Amazon introduced Arm-based instances that promise up to 40% better cost-performance ratio than comparable x86 instances for specific workloads. We evaluate in this paper the video encoding performance of x86 and Arm instances of four instance families using the latest FFmpeg version and two video codecs. We examine the impact of the encoding parameters, such as different presets and bitrates, on the time and cost for encoding. Our experiments reveal that Arm instances show high time and cost-saving potential of up to 33.63% for specific bitrates and presets, especially for the x264 codec. However, the x86 instances are more general and achieve low encoding times, regardless of the codec.
In this paper, researchers evaluated the video encoding performance of x86 and ARM instances of four Amazon EC2 instance families using the latest FFmpeg version and two video codecs.https://t.co/GcaqVNLy7Q pic.twitter.com/gtrVoW8QIC
— Underfox (@Underfox3) June 14, 2021
17. Going Beyond Linear Transformers with Recurrent Fast Weight Programmers
Kazuki Irie, Imanol Schlag, Róbert Csordás, Jürgen Schmidhuber
Transformers with linearised attention (“linear Transformers”) have demonstrated the practical scalability and effectiveness of outer product-based Fast Weight Programmers (FWPs) from the ’90s. However, the original FWP formulation is more general than the one of linear Transformers: a slow neural network (NN) continually reprograms the weights of a fast NN with arbitrary NN architectures. In existing linear Transformers, both NNs are feedforward and consist of a single layer. Here we explore new variations by adding recurrence to the slow and fast nets. We evaluate our novel recurrent FWPs (RFWPs) on two synthetic algorithmic tasks (code execution and sequential ListOps), Wikitext-103 language models, and on the Atari 2600 2D game environment. Our models exhibit properties of Transformers and RNNs. In the reinforcement learning setting, we report large improvements over LSTM in several Atari games. Our code is public.
Going Beyond Linear Transformers with Recurrent Fast Weight Programmers
— AK (@ak92501) June 14, 2021
pdf: https://t.co/oS6D1IQTRG
abs: https://t.co/LqKRq9kBrM
github: https://t.co/KGewO2s85A
various new linear Transformer variants with recurrent connections pic.twitter.com/d4geCESa9z
18. Probabilistic, Structure-Aware Algorithms for Improved Variety, Accuracy, and Coverage of AMR Alignments
Austin Blodgett, Nathan Schneider
We present algorithms for aligning components of Abstract Meaning Representation (AMR) graphs to spans in English sentences. We leverage unsupervised learning in combination with heuristics, taking the best of both worlds from previous AMR aligners. Our unsupervised models, however, are more sensitive to graph substructures, without requiring a separate syntactic parse. Our approach covers a wider variety of AMR substructures than previously considered, achieves higher coverage of nodes and edges, and does so with higher accuracy. We will release our LEAMR datasets and aligner for use in research on AMR parsing, generation, and evaluation.