diff --git "a/data/de/train.jsonl" "b/data/de/train.jsonl" new file mode 100644--- /dev/null +++ "b/data/de/train.jsonl" @@ -0,0 +1,1992 @@ +{"source": "Due to the success of deep learning to solving a variety of challenging machine learning tasks, there is a rising interest in understanding loss functions for training neural networks from a theoretical aspect.Particularly, the properties of critical points and the landscape around them are of importance to determine the convergence performance of optimization algorithms.In this paper, we provide a necessary and sufficient characterization of the analytical forms for the critical points (as well as global minimizers) of the square loss functions for linear neural networks.We show that the analytical forms of the critical points characterize the values of the corresponding loss functions as well as the necessary and sufficient conditions to achieve global minimum.Furthermore, we exploit the analytical forms of the critical points to characterize the landscape properties for the loss functions of linear neural networks and shallow ReLU networks.One particular conclusion is that: While the loss function of linear networks has no spurious local minimum, the loss function of one-hidden-layer nonlinear networks with ReLU activation function does have local minimum that is not global minimum.", "target": ["Wir bieten notwendige und ausreichend analytische Formulare für die kritischen Punkte der quadratischen Verlustfunktionen für verschiedene neuronale Netze und nutzen die analytischen Formulare, um die Landschaftseigenschaften für die Verlustfunktionen dieser neuronalen Netzwerke zu charakterisieren."]} +{"source": "The backpropagation (BP) algorithm is often thought to be biologically implausible in the brain.One of the main reasons is that BP requires symmetric weight matrices in the feedforward and feedback pathways.To address this “weight transport problem” (Grossberg, 1987), two biologically-plausible algorithms, proposed by Liao et al. (2016) and Lillicrap et al. (2016), relax BP’s weight symmetry requirements and demonstrate comparable learning capabilities to that of BP on small datasets.However, a recent study by Bartunov et al. (2018) finds that although feedback alignment (FA) and some variants of target-propagation (TP) perform well on MNIST and CIFAR, they perform significantly worse than BP on ImageNet.Here, we additionally evaluate the sign-symmetry (SS) algorithm (Liao et al., 2016), which differs from both BP and FA in that the feedback and feedforward weights do not share magnitudes but share signs.We examined the performance of sign-symmetry and feedback alignment on ImageNet and MS COCO datasets using different network architectures (ResNet-18 and AlexNet for ImageNet; RetinaNet for MS COCO).Surprisingly, networks trained with sign-symmetry can attain classification performance approaching that of BP-trained networks.These results complement the study by Bartunov et al. (2018) and establish a new benchmark for future biologically-plausible learning algorithms on more difficult datasets and more complex architectures.", "target": ["Biologisch plausible Lernalgorithmen, insbesondere Zeichensymmetrie, funktionieren gut auf ImageNet"]} +{"source": "We introduce the 2-simplicial Transformer, an extension of the Transformer which includes a form of higher-dimensional attention generalising the dot-product attention, and uses this attention to update entity representations with tensor products of value vectors.We show that this architecture is a useful inductive bias for logical reasoning in the context of deep reinforcement learning.", "target": ["Wir führen den 2-simplical Transformer ein und zeigen, dass diese Architektur ein nützlicher induktiver Bias für logisches Argumentation im Kontext von Deep Reinforcement Learning."]} +{"source": "We present Tensor-Train RNN (TT-RNN), a novel family of neural sequence architectures for multivariate forecasting in environments with nonlinear dynamics.Long-term forecasting in such systems is highly challenging, since there exist long-term temporal dependencies, higher-order correlations and sensitivity to error propagation.Our proposed tensor recurrent architecture addresses these issues by learning the nonlinear dynamics directly using higher order moments and high-order state transition functions.Furthermore, we decompose the higher-order structure using the tensor-train (TT) decomposition to reduce the number of parameters while preserving the model performance.We theoretically establish the approximation properties of Tensor-Train RNNs for general sequence inputs, and such guarantees are not available for usual RNNs.We also demonstrate significant long-term prediction improvements over general RNN and LSTM architectures on a range of simulated environments with nonlinear dynamics, as well on real-world climate and traffic data.", "target": ["Genaue Vorhersagen über sehr lange Zeithorizonte mit Tensor-Train-RNNs"]} +{"source": "Recent efforts on combining deep models with probabilistic graphical models are promising in providing flexible models that are also easy to interpret.We propose a variational message-passing algorithm for variational inference in such models.We make three contributions.First, we propose structured inference networks that incorporate the structure of the graphical model in the inference network of variational auto-encoders (VAE).Second, we establish conditions under which such inference networks enable fast amortized inference similar to VAE.Finally, we derive a variational message passing algorithm to perform efficient natural-gradient inference while retaining the efficiency of the amortized inference.By simultaneously enabling structured, amortized, and natural-gradient inference for deep structured models, our method simplifies and generalizes existing methods.", "target": ["Wir schlagen einen variationalen Message-Passing-Algorithmus für Modelle vor, die sowohl das tiefe Modell als auch das probabilistische grafische Modell enthalten."]} +{"source": "Modern deep neural networks have a large amount of weights, which make them difficult to deploy on computation constrained devices such as mobile phones.One common approach to reduce the model size and computational cost is to use low-rank factorization to approximate a weight matrix.However, performing standard low-rank factorization with a small rank can hurt the model expressiveness and significantly decrease the performance.In this work, we propose to use a mixture of multiple low-rank factorizations to model a large weight matrix, and the mixture coefficients are computed dynamically depending on its input.We demonstrate the effectiveness of the proposed approach on both language modeling and image classification tasks.Experiments show that our method not only improves the computation efficiency but also maintains (sometimes outperforms) its accuracy compared with the full-rank counterparts.", "target": ["Eine einfache Modifikation der Low-Rank-Faktorisierung, die die Leistung (sowohl bei Bild- als auch bei Sprachaufgaben) verbessert und gleichzeitig kompakt ist."]} +{"source": "Deep learning training accesses vast amounts of data at high velocity, posing challenges for datasets retrieved over commodity networks and storage devices.We introduce a way to dynamically reduce the overhead of fetching and transporting training data with a method we term Progressive Compressed Records (PCRs).PCRs deviate from previous formats by leveraging progressive compression to split each training example into multiple examples of increasingly higher fidelity, without adding to the total data size.Training examples of similar fidelity are grouped together, which reduces both the system overhead and data bandwidth needed to train a model.We show that models can be trained on aggressively compressed representations of the training data and still retain high accuracy, and that PCRs can enable a 2x speedup on average over baseline formats using JPEG compression.Our results hold across deep learning architectures for a wide range of datasets: ImageNet, HAM10000, Stanford Cars, and CelebA-HQ.", "target": ["Wir schlagen ein einfaches, allgemeines und speichersparendes Datenformat vor, um das Training von Deep Learning zu beschleunigen, indem die Sample Fidelity dynamisch zur Trainingszeit ausgewählt werden kann."]} +{"source": "It is fundamental and challenging to train robust and accurate Deep Neural Networks (DNNs) when semantically abnormal examples exist.Although great progress has been made, there is still one crucial research question which is not thoroughly explored yet: What training examples should be focused and how much more should they be emphasised to achieve robust learning?In this work, we study this question and propose gradient rescaling (GR) to solve it.GR modifies the magnitude of logit vector’s gradient to emphasise on relatively easier training data points when noise becomes more severe, which functions as explicit emphasis regularisation to improve the generalisation performance of DNNs.Apart from regularisation, we connect GR to examples weighting and designing robust loss functions.We empirically demonstrate that GR is highly anomaly-robust and outperforms the state-of-the-art by a large margin, e.g., increasing 7% on CIFAR100 with 40% noisy labels.It is also significantly superior to standard regularisers in both clean and abnormal settings.Furthermore, we present comprehensive ablation studies to explore the behaviours of GR under different cases, which is informative for applying GR in real-world scenarios.", "target": ["Robustes diskriminatives Repräsentationslernen mittels Gradienten-Reskalierung: Eine Perspektive der Regulierung von Schwerpunkten"]} +{"source": "Generative Adversarial Networks (GANs) have achieved remarkable results in the task of generating realistic natural images.In most applications, GAN models share two aspects in common.On the one hand, GANs training involves solving a challenging saddle point optimization problem, interpreted as an adversarial game between a generator and a discriminator functions.On the other hand, the generator and the discriminator are parametrized in terms of deep convolutional neural networks.The goal of this paper is to disentangle the contribution of these two factors to the success of GANs.In particular, we introduce Generative Latent Optimization (GLO), a framework to train deep convolutional generators without using discriminators, thus avoiding the instability of adversarial optimization problems.Throughout a variety of experiments, we show that GLO enjoys many of the desirable properties of GANs: learning from large data, synthesizing visually-appealing samples, interpolating meaningfully between samples, and performing linear arithmetic with noise vectors.", "target": ["Sind GANs erfolgreich, weil sie nachteilig trainiert werden oder weil sie ConvNets verwenden? Wir zeigen, dass ein ConvNet-Generator, der mit einem einfachen Rekonstruktionsverlust und erlernbaren Rauschvektoren trainiert wird, viele der wünschenswerten Eigenschaften eines GANs aufweist."]} +{"source": "In this paper, we propose a novel kind of kernel, random forest kernel, to enhance the empirical performance of MMD GAN.Different from common forests with deterministic routings, a probabilistic routing variant is used in our innovated random-forest kernel, which is possible to merge with the CNN frameworks.Our proposed random-forest kernel has the following advantages: From the perspective of random forest, the output of GAN discriminator can be viewed as feature inputs to the forest, where each tree gets access to merely a fraction of the features, and thus the entire forest benefits from ensemble learning.In the aspect of kernel method, random-forest kernel is proved to be characteristic, and therefore suitable for the MMD structure.Besides, being an asymmetric kernel, our random-forest kernel is much more flexible, in terms of capturing the differences between distributions.Sharing the advantages of CNN, kernel method, and ensemble learning, our random-forest kernel based MMD GAN obtains desirable empirical performances on CIFAR-10, CelebA and LSUN bedroom data sets.Furthermore, for the sake of completeness, we also put forward comprehensive theoretical analysis to support our experimental results.", "target": ["MMD GANs mit einem neuen Random-Forest-Kernel ausrüsten."]} +{"source": "Reinforcement learning in an actor-critic setting relies on accurate value estimates of the critic.However, the combination of function approximation, temporal difference (TD) learning and off-policy training can lead to an overestimating value function.A solution is to use Clipped Double Q-learning (CDQ), which is used in the TD3 algorithm and computes the minimum of two critics in the TD-target. We show that CDQ induces an underestimation bias and propose a new algorithm that accounts for this by using a weighted average of the target from CDQ and the target coming from a single critic.The weighting parameter is adjusted during training such that the value estimates match the actual discounted return on the most recent episodes and by that it balances over- and underestimation.Empirically, we obtain more accurate value estimates and demonstrate state of the art results on several OpenAI gym tasks.", "target": ["Eine Methode für genauere kritische Schätzungen beim Reinforcement Learning."]} +{"source": "We introduce a systematic framework for quantifying the robustness of classifiers to naturally occurring perturbations of images found in videos.As part of this framework, we construct ImageNet-Vid-Robust, a human-expert--reviewed dataset of 22,668 images grouped into 1,145 sets of perceptually similar images derived from frames in the ImageNet Video Object Detection dataset.We evaluate a diverse array of classifiers trained on ImageNet, including models trained for robustness, and show a median classification accuracy drop of 16\\%.Additionally, we evaluate the Faster R-CNN and R-FCN models for detection, and show that natural perturbations induce both classification as well as localization errors, leading to a median drop in detection mAP of 14 points.Our analysis shows that natural perturbations in the real world are heavily problematic for current CNNs, posing a significant challenge to their deployment in safety-critical environments that require reliable, low-latency predictions.", "target": ["Wir stellen einen systematischen Rahmen für die Quantifizierung der Robustheit von Klassifikatoren gegenüber natürlich auftretenden Störungen von Bildern in Videos vor."]} +{"source": "Structured tabular data is the most commonly used form of data in industry according to a Kaggle ML and DS Survey.Gradient Boosting Trees, Support Vector Machine, Random Forest, and Logistic Regression are typically used for classification tasks on tabular data.The recent work of Super Characters method using two-dimensional word embedding achieved state-of-the-art results in text classification tasks, showcasing the promise of this new approach.In this paper, we propose the SuperTML method, which borrows the idea of Super Characters method and two-dimensional embedding to address the problem of classification on tabular data.For each input of tabular data, the features are first projected into two-dimensional embedding like an image, and then this image is fed into fine-tuned ImageNet CNN models for classification.Experimental results have shown that the proposed SuperTML method have achieved state-of-the-art results on both large and small datasets.", "target": ["Deep Learning für strukturierte tabellarische Daten im maschinellen Lernen unter Verwendung eines vortrainierten CNN-Modell aus ImageNet."]} +{"source": "Learning rich representations from predictive learning without labels has been a longstanding challenge in the field of machine learning.Generative pre-training has so far not been as successful as contrastive methods in modeling representations of raw images.In this paper, we propose a neural architecture for self-supervised representation learning on raw images called the PatchFormer which learns to model spatial dependencies across patches in a raw image.Our method learns to model the conditional probability distribution of missing patches given the context of surrounding patches.We evaluate the utility of the learned representations by fine-tuning the pre-trained model on low data-regime classification tasks.Specifically, we benchmark our model on semi-supervised ImageNet classification which has become a popular benchmark recently for semi-supervised and self-supervised learning methods.Our model is able to achieve 30.3% and 65.5% top-1 accuracies when trained only using 1% and 10% of the labels on ImageNet showing the promise for generative pre-training methods.", "target": ["Die Dekodierung von Pixeln kann immer noch für das Repräsentationslernen auf Bildern funktionieren"]} +{"source": "Adaptive regularization methods pre-multiply a descent direction by a preconditioning matrix.Due to the large number of parameters of machine learning problems, full-matrix preconditioning methods are prohibitively expensive.We show how to modify full-matrix adaptive regularization in order to make it practical and effective.We also provide novel theoretical analysisfor adaptive regularization in non-convex optimization settings.The core of our algorithm, termed GGT, consists of efficient inverse computation of square roots of low-rank matrices.Our preliminary experiments underscore improved convergence rate of GGT across a variety of synthetic tasks and standard deep learning benchmarks.", "target": ["Schnelle, wirklich skalierbare Vollmatrix AdaGrad / Adam, mit Theorie für adaptive stochastische nicht konvexe Optimierung"]} +{"source": "Dialogue systems require a great deal of different but complementary expertise to assist, inform, and entertain humans.For example, different domains (e.g., restaurant reservation, train ticket booking) of goal-oriented dialogue systems can be viewed as different skills, and so does ordinary chatting abilities of chit-chat dialogue systems.In this paper, we propose to learn a dialogue system that independently parameterizes different dialogue skills, and learns to select and combine each of them through Attention over Parameters (AoP).The experimental results show that this approach achieves competitive performance on a combined dataset of MultiWOZ (Budzianowski et al., 2018), In-Car Assistant (Eric et al.,2017), and Persona-Chat (Zhang et al., 2018).Finally, we demonstrate that each dialogue skill is effectively learned and can be combined with other skills to produce selective responses.", "target": ["In dieser Arbeit schlagen wir vor, ein Dialogsystem zu erlernen, das unabhängig voneinander verschiedene Dialogfähigkeiten parametrisiert, und lernt jede von ihnen durch Attention over Parameters (AoP) auszuwählen und zu kombinieren."]} +{"source": "Model distillation aims to distill the knowledge of a complex model into a simpler one.In this paper, we consider an alternative formulation called dataset distillation: we keep the model fixed and instead attempt to distill the knowledge from a large training dataset into a small one.The idea is to synthesize a small number of data points that do not need to come from the correct data distribution, but will, when given to the learning algorithm as training data, approximate the model trained on the original data.For example, we show that it is possible to compress 60,000 MNIST training images into just 10 synthetic distilled images (one per class) and achieve close to the original performance, given a fixed network initialization.We evaluate our method in various initialization settings. Experiments on multiple datasets, MNIST, CIFAR10, PASCAL-VOC, and CUB-200, demonstrate the ad-vantage of our approach compared to alternative methods. Finally, we include a real-world application of dataset distillation to the continual learning setting: we show that storing distilled images as episodic memory of previous tasks can alleviate forgetting more effectively than real images.", "target": ["Wir schlagen vor, einen großen Datensatz in einen kleinen Satz synthetischer Daten zu destillieren, mit denen Netze nahe an der ursprünglichen Leistung trainiert werden können."]} +{"source": "We relate the minimax game of generative adversarial networks (GANs) to finding the saddle points of the Lagrangian function for a convex optimization problem, where the discriminator outputs and the distribution of generator outputs play the roles of primal variables and dual variables, respectively.This formulation shows the connection between the standard GAN training process and the primal-dual subgradient methods for convex optimization.The inherent connection does not only provide a theoretical convergence proof for training GANs in the function space, but also inspires a novel objective function for training.The modified objective function forces the distribution of generator outputs to be updated along the direction according to the primal-dual subgradient methods.A toy example shows that the proposed method is able to resolve mode collapse, which in this case cannot be avoided by the standard GAN or Wasserstein GAN.Experiments on both Gaussian mixture synthetic data and real-world image datasets demonstrate the performance of the proposed method on generating diverse samples.", "target": ["Wir schlagen eine primär-duale Subgradienten-Methode für das Training von GANs vor, und diese Methode mildert den Modus-Kollaps effektiv."]} +{"source": "Specifying reward functions is difficult, which motivates the area of reward inference: learning rewards from human behavior.The starting assumption in the area is that human behavior is optimal given the desired reward function, but in reality people have many different forms of irrationality, from noise to myopia to risk aversion and beyond.This fact seems like it will be strictly harmful to reward inference: it is already hard to infer the reward from rational behavior, and noise and systematic biases make actions have less direct of a relationship to the reward.Our insight in this work is that, contrary to expectations, irrationality can actually help rather than hinder reward inference.For some types and amounts of irrationality, the expert now produces more varied policies compared to rational behavior, which help disambiguate among different reward parameters -- those that otherwise correspond to the same rational behavior.We put this to the test in a systematic analysis of the effect of irrationality on reward inference.We start by covering the space of irrationalities as deviations from the Bellman update, simulate expert behavior, and measure the accuracy of inference to contrast the different types and study the gains and losses.We provide a mutual information-based analysis of our findings, and wrap up by discussing the need to accurately model irrationality, as well as to what extent we might expect (or be able to train) real people to exhibit helpful irrationalities when teaching rewards to learners.", "target": ["Wir stellen fest, dass die Irrationalität eines Experten-Vorführers dem Lernenden helfen kann, seine Präferenzen zu erkennen."]} +{"source": "Natural Language Processing models lack a unified approach to robustness testing.In this paper we introduce WildNLP - a framework for testing model stability in a natural setting where text corruptions such as keyboard errors or misspelling occur.We compare robustness of models from 4 popular NLP tasks: Q&A, NLI, NER and Sentiment Analysis by testing their performance on aspects introduced in the framework.In particular, we focus on a comparison between recent state-of-the- art text representations and non-contextualized word embeddings.In order to improve robust- ness, we perform adversarial training on se- lected aspects and check its transferability to the improvement of models with various cor- ruption types.We find that the high perfor- mance of models does not ensure sufficient robustness, although modern embedding tech- niques help to improve it.We release cor- rupted datasets and code for WildNLP frame- work for the community.", "target": ["Wir vergleichen die Robustheit von Modellen aus 4 beliebten NLP-Aufgaben: Q&A, NLI, NER und Sentiment Analysis, indem wir ihre Leistung bei gestörten Eingaben testen."]} +{"source": "Training generative models like Generative Adversarial Network (GAN) is challenging for noisy data.A novel curriculum learning algorithm pertaining to clustering is proposed to address this issue in this paper.The curriculum construction is based on the centrality of underlying clusters in data points. The data points of high centrality takes priority of being fed into generative models during training.To make our algorithm scalable to large-scale data, the active set is devised, in the sense that every round of training proceeds only on an active subset containing a small fraction of already trained data and the incremental data of lower centrality.Moreover, the geometric analysis is presented to interpret the necessity of cluster curriculum for generative models.The experiments on cat and human-face data validate that our algorithm is able to learn the optimal generative models (e.g. ProGAN) with respect to specified quality metrics for noisy data.An interesting finding is that the optimal cluster curriculum is closely related to the critical point of the geometric percolation process formulated in the paper.", "target": ["Ein neuartiger cluster-basierter Algorithmus des Curriculum-Lernens wird vorgeschlagen, um das robuste Training generativer Modelle zu lösen."]} +{"source": "Backdoor attacks aim to manipulate a subset of training data by injecting adversarial triggers such that machine learning models trained on the tampered dataset will make arbitrarily (targeted) incorrect prediction on the testset with the same trigger embedded.While federated learning (FL) is capable of aggregating information provided by different parties for training a better model, its distributed learning methodology and inherently heterogeneous data distribution across parties may bring new vulnerabilities.In addition to recent centralized backdoor attacks on FL where each party embeds the same global trigger during training, we propose the distributed backdoor attack (DBA) --- a novel threat assessment framework developed by fully exploiting the distributed nature of FL.DBA decomposes a global trigger pattern into separate local patterns and embed them into the training set of different adversarial parties respectively.Compared to standard centralized backdoors, we show that DBA is substantially more persistent and stealthy against FL on diverse datasets such as finance and image data.We conduct extensive experiments to show that the attack success rate of DBA is significantly higher than centralized backdoors under different settings.Moreover, we find that distributed attacks are indeed more insidious, as DBA can evade two state-of-the-art robust FL algorithms against centralized backdoors.We also provide explanations for the effectiveness of DBA via feature visual interpretation and feature importance ranking.To further explore the properties of DBA, we test the attack performance by varying different trigger factors, including local trigger variations (size, gap, and location), scaling factor in FL, data distribution, and poison ratio and interval.Our proposed DBA and thorough evaluation results shed lights on characterizing the robustness of FL.", "target": ["Wir haben einen neuartigen verteilten Backdoor-Angriff auf föderiertes Lernen vorgeschlagen und zeigen, dass er im Vergleich zu zentralisierten Standardangriffen nicht nur effektiver ist, sondern auch schwieriger durch bestehende robuste FL-Methoden abgewehrt werden kann"]} +{"source": "Graph networks have recently attracted considerable interest, and in particular in the context of semi-supervised learning.These methods typically work by generating node representations that are propagated throughout a given weighted graph.Here we argue that for semi-supervised learning, it is more natural to consider propagating labels in the graph instead.Towards this end, we propose a differentiable neural version of the classic Label Propagation (LP) algorithm.This formulation can be used for learning edge weights, unlike other methods where weights are set heuristically.Starting from a layer implementing a single iteration of LP, we proceed by adding several important non-linear steps that significantly enhance the label-propagating mechanism.Experiments in two distinct settings demonstrate the utility of our approach.", "target": ["Neuronales Netz für graphenbasiertes halbüberwachtes Lernen; greift die Klassiker wieder auf und propagiert *Labels* statt Merkmalsdarstellungen"]} +{"source": "Neural architecture search (NAS) has made rapid progress incomputervision,wherebynewstate-of-the-artresultshave beenachievedinaseriesoftaskswithautomaticallysearched neural network (NN) architectures.In contrast, NAS has not made comparable advances in natural language understanding (NLU).Corresponding to encoder-aggregator meta architecture of typical neural networks models for NLU tasks (Gong et al. 2018), we re-define the search space, by splittingitinto twoparts:encodersearchspace,andaggregator search space.Encoder search space contains basic operations such as convolutions, RNNs, multi-head attention and its sparse variants, star-transformers.Dynamic routing is included in the aggregator search space, along with max (avg) pooling and self-attention pooling.Our search algorithm is then fulfilled via DARTS, a differentiable neural architecture search framework.We progressively reduce the search space every few epochs, which further reduces the search time and resource costs.Experiments on five benchmark data-sets show that, the new neural networks we generate can achieve performances comparable to the state-of-the-art models that does not involve language model pre-training.", "target": ["Neuronale Architektur Suche für eine Reihe von Aufgaben zum Verstehen natürlicher Sprache. Entwurf des Suchraums für NLU-Aufgaben. Anwendung differenzierbarer Architektursuche zur Entdeckung neuer Modelle"]} +{"source": "Network embedding (NE) methods aim to learn low-dimensional representations of network nodes as vectors, typically in Euclidean space.These representations are then used for a variety of downstream prediction tasks.Link prediction is one of the most popular choices for assessing the performance of NE methods.However, the complexity of link prediction requires a carefully designed evaluation pipeline to provide consistent, reproducible and comparable results.We argue this has not been considered sufficiently in recent works.The main goal of this paper is to overcome difficulties associated with evaluation pipelines and reproducibility of results.We introduce EvalNE, an evaluation framework to transparently assess and compare the performance of NE methods on link prediction.EvalNE provides automation and abstraction for tasks such as hyper-parameter tuning, model validation, edge sampling, computation of edge embeddings and model validation.The framework integrates efficient procedures for edge and non-edge sampling and can be used to easily evaluate any off-the-shelf embedding method.The framework is freely available as a Python toolbox.Finally, demonstrating the usefulness of EvalNE in practice, we conduct an empirical study in which we try to replicate and analyse experimental sections of several influential papers.", "target": ["In dieser Arbeit stellen wir EvalNE vor, eine Python-Toolbox zur automatisierten Bewertung von Netzwerkeinbettungsmethoden zur Link-Vorhersage und zur Gewährleistung der Reproduzierbarkeit der Ergebnisse."]} +{"source": "Deep learning models can be efficiently optimized via stochastic gradient descent, but there is little theoretical evidence to support this.A key question in optimization is to understand when the optimization landscape of a neural network is amenable to gradient-based optimization.We focus on a simple neural network two-layer ReLU network with two hidden units, and show that all local minimizers are global.This combined with recent work of Lee et al. (2017); Lee et al. (2016) show that gradient descent converges to the global minimizer.", "target": ["Wiederherstellungsgarantie des stochastischen Gradientenabstiegs mit zufälliger Initialisierung für das Lernen eines zweischichtigen neuronalen Netzes mit zwei versteckten Knoten, Gewichten mit Einheitsnorm, ReLU-Aktivierungsfunktionen und Gaußschen Eingaben."]} +{"source": "Dropout is a simple yet effective technique to improve generalization performance and prevent overfitting in deep neural networks (DNNs).In this paper, we discuss three novel observations about dropout to better understand the generalization of DNNs with rectified linear unit (ReLU) activations: 1) dropout is a smoothing technique that encourages each local linear model of a DNN to be trained on data points from nearby regions; 2) a constant dropout rate can result in effective neural-deactivation rates that are significantly different for layers with different fractions of activated neurons; and 3) the rescaling factor of dropout causes an inconsistency to occur between the normalization during training and testing conditions when batch normalization is also used. The above leads to three simple but nontrivial improvements to dropout resulting in our proposed method \"Jumpout.\"Jumpout samples the dropout rate using a monotone decreasing distribution (such as the right part of a truncated Gaussian), so the local linear model at each data point is trained, with high probability, to work better for data points from nearby than from more distant regions.Instead of tuning a dropout rate for each layer and applying it to all samples, jumpout moreover adaptively normalizes the dropout rate at each layer and every training sample/batch, so the effective dropout rate applied to the activated neurons are kept the same.Moreover, we rescale the outputs of jumpout for a better trade-off that keeps both the variance and mean of neurons more consistent between training and test phases, which mitigates the incompatibility between dropout and batch normalization.Compared to the original dropout, jumpout shows significantly improved performance on CIFAR10, CIFAR100, Fashion- MNIST, STL10, SVHN, ImageNet-1k, etc., while introducing negligible additional memory and computation costs.", "target": ["Jumpout wendet drei einfache, aber effektive Modifikationen des Dropouts an, die auf neuen Erkenntnissen über die Generalisierungsleistung von DNN mit ReLU in lokalen Regionen basieren."]} +{"source": "Concerns about interpretability, computational resources, and principled inductive priors have motivated efforts to engineer sparse neural models for NLP tasks.If sparsity is important for NLP, might well-trained neural models naturally become roughly sparse?Using the Taxi-Euclidean norm to measure sparsity, we find that frequent input words are associated with concentrated or sparse activations, while frequent target words are associated with dispersed activations but concentrated gradients.We find that gradients associated with function words are more concentrated than the gradients of content words, even controlling for word frequency.", "target": ["Wir untersuchen die natürliche Entstehung von Sparsity in den Aktivierungen und Gradienten für einige Layer eines dichten LSTM-Sprachmodells während des Trainings."]} +{"source": "The integration of a Knowledge Base (KB) into a neural dialogue agent is one of the key challenges in Conversational AI.Memory networks has proven to be effective to encode KB information into an external memory to thus generate more fluent and informed responses.Unfortunately, such memory becomes full of latent representations during training, so the most common strategy is to overwrite old memory entries randomly. In this paper, we question this approach and provide experimental evidence showing that conventional memory networks generate many redundant latent vectors resulting in overfitting and the need for larger memories.We introduce memory dropout as an automatic technique that encourages diversity in the latent space by1) Aging redundant memories to increase their probability of being overwritten during training2) Sampling new memories that summarize the knowledge acquired by redundant memories.This technique allows us to incorporate Knowledge Bases to achieve state-of-the-art dialogue generation in the Stanford Multi-Turn Dialogue dataset.Considering the same architecture, its use provides an improvement of +2.2 BLEU points for the automatic generation of responses and an increase of +8.1% in the recognition of named entities.", "target": ["Herkömmliche Speichernetzwerke erzeugen viele redundante latente Vektoren, was zu einer Überanpassung und dem Bedarf größerer Speicherkapazitäten führt. Wir führen Memory Dropout als eine automatische Technik ein, die die Vielfalt im latenten Raum fördert."]} +{"source": "Su-Boyd-Candes (2014) made a connection between Nesterov's method and an ordinary differential equation (ODE). We show if a Hessian damping term is added to the ODE from Su-Boyd-Candes (2014), then Nesterov's method arises as a straightforward discretization of the modified ODE.Analogously, in the strongly convex case, a Hessian damping term is added to Polyak's ODE, which is then discretized to yield Nesterov's method for strongly convex functions. Despite the Hessian term, both second order ODEs can be represented as first order systems.Established Liapunov analysis is used to recover the accelerated rates of convergence in both continuous and discrete time. Moreover, the Liapunov analysis can be extended to the case of stochastic gradients which allows the full gradient case to be considered as a special case of the stochastic case. The result is a unified approach to convex acceleration in both continuous and discrete time and in both the stochastic and full gradient cases.", "target": ["Wir leiten die Nesterov-Methode als eine direkte Diskretisierung einer ODE ab, die sich von der Su-Boyd-Candes-Methode unterscheidet, und beweisen die Beschleunigung im stochastischen Fall"]} +{"source": "We propose learning to transfer learn (L2TL) to improve transfer learning on a target dataset by judicious extraction of information from a source dataset.L2TL considers joint optimization of vastly-shared weights between models for source and target tasks, and employs adaptive weights for scaling of constituent losses.The adaptation of the weights is based on reinforcement learning, guided with a performance metric on the target validation set.We demonstrate state-of-the-art performance of L2TL given fixed models, consistently outperforming fine-tuning baselines on various datasets.In the regimes of small-scale target datasets and significant label mismatch between source and target datasets, L2TL outperforms previous work by an even larger margin.", "target": ["Wir schlagen Learning to Transfer-Learn (L2TL) vor, um das Transfer-Lernen auf einem Zieldatensatz durch gezielte Extraktion von Informationen aus einem Quelldatensatz zu verbessern."]} +{"source": "In many partially observable scenarios, Reinforcement Learning (RL) agents must rely on long-term memory in order to learn an optimal policy.We demonstrate that using techniques from NLP and supervised learning fails at RL tasks due to stochasticity from the environment and from exploration.Utilizing our insights on the limitations of traditional memory methods in RL, we propose AMRL, a class of models that can learn better policies with greater sample efficiency and are resilient to noisy inputs.Specifically, our models use a standard memory module to summarize short-term context, and then aggregate all prior states from the standard model without respect to order.We show that this provides advantages both in terms of gradient decay and signal-to-noise ratio over time.Evaluating in Minecraft and maze environments that test long-term memory, we find that our model improves average return by 19% over a baseline that has the same number of parameters and by 9% over a stronger baseline that has far more parameters.", "target": ["In Deep RL können ordnungsinvariante Funktionen in Verbindung mit Standardspeichermodulen verwendet werden, um den Gradientenabfall und die Widerstandsfähigkeit gegen Rauschen zu verbessern."]} +{"source": "Optimization on manifold has been widely used in machine learning, to handle optimization problems with constraint.Most previous works focus on the case with a single manifold.However, in practice it is quite common that the optimization problem involves more than one constraints, (each constraint corresponding to one manifold).It is not clear in general how to optimize on multiple manifolds effectively and provably especially when the intersection of multiple manifolds is not a manifold or cannot be easily calculated.We propose a unified algorithm framework to handle the optimization on multiple manifolds.Specifically, we integrate information from multiple manifolds and move along an ensemble direction by viewing the information from each manifold as a drift and adding them together.We prove the convergence properties of the proposed algorithms.We also apply the algorithms into training neural network with batch normalization layers and achieve preferable empirical results.", "target": ["In dieser Arbeit wird ein Algorithmus zur Behandlung von Optimierungsproblemen mit mehreren Nebenbedingungen unter Berücksichtigung der Vielfältigkeit vorgestellt."]} +{"source": "It has long been assumed that high dimensional continuous control problems cannot be solved effectively by discretizing individual dimensions of the action space due to the exponentially large number of bins over which policies would have to be learned.In this paper, we draw inspiration from the recent success of sequence-to-sequence models for structured prediction problems to develop policies over discretized spaces.Central to this method is the realization that complex functions over high dimensional spaces can be modeled by neural networks that predict one dimension at a time.Specifically, we show how Q-values and policies over continuous spaces can be modeled using a next step prediction model over discretized dimensions.With this parameterization, it is possible to both leverage the compositional structure of action spaces during learning, as well as compute maxima over action spaces (approximately).On a simple example task we demonstrate empirically that our method can perform global search, which effectively gets around the local optimization issues that plague DDPG.We apply the technique to off-policy (Q-learning) methods and show that our method can achieve the state-of-the-art for off-policy methods on several continuous control tasks.", "target": ["Eine Methode zum Q-Learning auf kontinuierlichen Aktionsräumen durch Vorhersage einer Sequenz diskretisierter 1-D Aktionen."]} +{"source": "Model-based reinforcement learning (MBRL) aims to learn a dynamic model to reduce the number of interactions with real-world environments.However, due to estimation error, rollouts in the learned model, especially those of long horizon, fail to match the ones in real-world environments.This mismatching has seriously impacted the sample complexity of MBRL.The phenomenon can be attributed to the fact that previous works employ supervised learning to learn the one-step transition models, which has inherent difficulty ensuring the matching of distributions from multi-step rollouts.Based on the claim, we propose to learn the synthesized model by matching the distributions of multi-step rollouts sampled from the synthesized model and the real ones via WGAN.We theoretically show that matching the two can minimize the difference of cumulative rewards between the real transition and the learned one.Our experiments also show that the proposed model imitation method outperforms the state-of-the-art in terms of sample complexity and average return.", "target": ["Unsere Methode beinhaltet WGAN, um eine Anpassung der Belegungsmessungen für das transition leraning zu erreichen."]} +{"source": "Batch Normalization (BN) and its variants have seen widespread adoption in the deep learning community because they improve the training of deep neural networks.Discussions of why this normalization works so well remain unsettled. We make explicit the relationship between ordinary least squares and partial derivatives computed when back-propagating through BN.We recast the back-propagation of BN as a least squares fit, which zero-centers and decorrelates partial derivatives from normalized activations.This view, which we term {\\em gradient-least-squares}, is an extensible and arithmetically accurate description of BN.To further explore this perspective, we motivate, interpret, and evaluate two adjustments to BN.", "target": ["Die Gaußsche Normalisierung führt während der Backpropagation eine Anpassung nach dem Prinzip der kleinsten Quadrate durch, die die partiellen Ableitungen der normalisierten Aktivierungen in die Nulllinie bringt und dekorreliert."]} +{"source": "Batch Normalization (BN) has become a cornerstone of deep learning across diverse architectures, appearing to help optimization as well as generalization.While the idea makes intuitive sense, theoretical analysis of its effectiveness has been lacking.Here theoretical support is provided for one of its conjectured properties, namely, the ability to allow gradient descent to succeed with less tuning of learning rates.It is shown that even if we fix the learning rate of scale-invariant parameters (e.g., weights of each layer with BN) to a constant (say, 0.3), gradient descent still approaches a stationary point (i.e., a solution where gradient is zero) in the rate of T^{−1/2} in T iterations, asymptotically matching the best bound for gradient descent with well-tuned learning rates.A similar result with convergence rate T^{−1/4} is also shown for stochastic gradient descent.", "target": ["Wir geben eine theoretische Analyse der Fähigkeit der Batch-Normalisierung zur automatischen Abstimmung der Lernraten im Kontext der Suche nach stationären Punkten für ein Deep Learning Ziel."]} +{"source": "Generative models of natural images have progressed towards high fidelity samples by the strong leveraging of scale.We attempt to carry this success to the field of video modeling by showing that large Generative Adversarial Networks trained on the complex Kinetics-600 dataset are able to produce video samples of substantially higher complexity and fidelity than previous work. Our proposed model, Dual Video Discriminator GAN (DVD-GAN), scales to longer and higher resolution videos by leveraging a computationally efficient decomposition of its discriminator.We evaluate on the related tasks of video synthesis and video prediction, and achieve new state-of-the-art Fréchet Inception Distance for prediction for Kinetics-600, as well as state-of-the-art Inception Score for synthesis on the UCF-101 dataset, alongside establishing a strong baseline for synthesis on Kinetics-600.", "target": ["Wir schlagen DVD-GAN vor, ein großes generatives Videomodell, das bei mehreren Aufgaben den neusten Stand darstellt und hochkomplexe Videos produziert, wenn es auf großen realen Datensätzen trainiert wird."]} +{"source": "Understanding procedural language requires anticipating the causal effects of actions, even when they are not explicitly stated.In this work, we introduce Neural Process Networks to understand procedural text through (neural) simulation of action dynamics. Our model complements existing memory architectures with dynamic entity tracking by explicitly modeling actions as state transformers.The model updates the states of the entities by executing learned action operators.Empirical results demonstrate that our proposed model can reason about the unstated causal effects of actions, allowing it to provide more accurate contextual information for understanding and generating procedural text, all while offering more interpretable internal representations than existing alternatives.", "target": ["Wir schlagen eine neue wiederkehrende Speicherarchitektur vor, die Zustandsänderungen von Entitäten im Sinne des gesunden Menschenverstands verfolgen kann, indem sie die kausalen Auswirkungen von Aktionen simuliert."]} +{"source": "There has been a recent trend in training neural networks to replace data structures that have been crafted by hand, with an aim for faster execution, better accuracy, or greater compression. In this setting, a neural data structure is instantiated by training a network over many epochs of its inputs until convergence.In many applications this expensive initialization is not practical, for example streaming algorithms --- where inputs are ephemeral and can only be inspected a small number of times. In this paper we explore the learning of approximate set membership over a stream of data in one-shot via meta-learning.We propose a novel memory architecture, the Neural Bloom Filter, which we show to be more compressive than Bloom Filters and several existing memory-augmented neural networks in scenarios of skewed data or structured sets.", "target": ["Wir untersuchen die Raumeffizienz von speichererweiterten neuronalen Netzen beim Lernen von Mengenzugehörigkeit."]} +{"source": "We leverage recent insights from second-order optimisation for neural networks to construct a Kronecker factored Laplace approximation to the posterior over the weights of a trained network.Our approximation requires no modification of the training procedure, enabling practitioners to estimate the uncertainty of their models currently used in production without having to retrain them.We extensively compare our method to using Dropout and a diagonal Laplace approximation for estimating the uncertainty of a network.We demonstrate that our Kronecker factored method leads to better uncertainty estimates on out-of-distribution data and is more robust to simple adversarial attacks.Our approach only requires calculating two square curvature factor matrices for each layer.Their size is equal to the respective square of the input and output size of the layer, making the method efficient both computationally and in terms of memory usage.We illustrate its scalability by applying it to a state-of-the-art convolutional network architecture.", "target": ["Wir konstruieren eine Kronecker-faktorisierte Laplace-Approximation für neuronale Netze, die zu einer effizienten Matrixnormalverteilung über die Gewichte führt."]} +{"source": "Spectral embedding is a popular technique for the representation of graph data.Several regularization techniques have been proposed to improve the quality of the embedding with respect to downstream tasks like clustering.In this paper, we explain on a simple block model the impact of the complete graph regularization, whereby a constant is added to all entries of the adjacency matrix.Specifically, we show that the regularization forces the spectral embedding to focus on the largest blocks, making the representation less sensitive to noise or outliers.We illustrate these results on both on both synthetic and real data, showing how regularization improves standard clustering scores.", "target": ["Die Graphenregulierung zwingt die spektrale Einbettung dazu, sich auf die größten Cluster zu konzentrieren, wodurch die Darstellung weniger empfindlich gegenüber Rauschen ist."]} +{"source": "The exposure bias problem refers to the training-inference discrepancy caused by teacher forcing in maximum likelihood estimation (MLE) training for auto-regressive neural network language models (LM).It has been regarded as a central problem for natural language generation (NLG) model training.Although a lot of algorithms have been proposed to avoid teacher forcing and therefore to alleviate exposure bias, there is little work showing how serious the exposure bias problem is.In this work, we first identify the auto-recovery ability of MLE-trained LM, which casts doubt on the seriousness of exposure bias.We then develop a precise, quantifiable definition for exposure bias.However, according to our measurements in controlled experiments, there's only around 3% performance gain when the training-inference discrepancy is completely removed.Our results suggest the exposure bias problem could be much less serious than it is currently assumed to be.", "target": ["Wir zeigen, dass der Expositionsbias viel weniger gravierend sein könnte, als er derzeit für das MLE-LM-Training angenommen wird."]} +{"source": "The ability of algorithms to evolve or learn (compositional) communication protocols has traditionally been studied in the language evolution literature through the use of emergent communication tasks.Here we scale up this research by using contemporary deep learning methods and by training reinforcement-learning neural network agents on referential communication games.We extend previous work, in which agents were trained in symbolic environments, by developing agents which are able to learn from raw pixel data, a more challenging and realistic input representation.We find that the degree of structure found in the input data affects the nature of the emerged protocols, and thereby corroborate the hypothesis that structured compositional language is most likely to emerge when agents perceive the world as being structured.", "target": ["Eine kontrollierte Studie über die Rolle von Umgebungen in Bezug auf Eigenschaften in aufkommenden Kommunikationsprotokollen."]} +{"source": "For understanding generic documents, information like font sizes, column layout, and generally the positioning of words may carry semantic information that is crucial for solving a downstream document intelligence task.Our novel BERTgrid, which is based on Chargrid by Katti et al. (2018), represents a document as a grid of contextualized word piece embedding vectors, thereby making its spatial structure and semantics accessible to the processing neural network.The contextualized embedding vectors are retrieved from a BERT language model.We use BERTgrid in combination with a fully convolutional network on a semantic instance segmentation task for extracting fields from invoices.We demonstrate its performance on tabulated line item and document header field extraction.", "target": ["Rasterbasierte Dokumentendarstellung mit kontextualisierten Einbettungsvektoren für Dokumente mit 2D-Layout"]} +{"source": "Deep reinforcement learning (RL) policies are known to be vulnerable to adversarial perturbations to their observations, similar to adversarial examples for classifiers.However, an attacker is not usually able to directly modify another agent's observations.This might lead one to wonder: is it possible to attack an RL agent simply by choosing an adversarial policy acting in a multi-agent environment so as to create natural observations that are adversarial?We demonstrate the existence of adversarial policies in zero-sum games between simulated humanoid robots with proprioceptive observations, against state-of-the-art victims trained via self-play to be robust to opponents.The adversarial policies reliably win against the victims but generate seemingly random and uncoordinated behavior.We find that these policies are more successful in high-dimensional environments, and induce substantially different activations in the victim policy network than when the victim plays against a normal opponent.Videos are available at https://attackingrl.github.io.", "target": ["Deep-RL-Methoden können von anderen Akteueren angegriffen werden, die Aktionen durchführen, um natürliche Beobachtungen zu erzeugen, die gegnerisch sind."]} +{"source": "GloVe and Skip-gram word embedding methods learn word vectors by decomposing a denoised matrix of word co-occurrences into a product of low-rank matrices.In this work, we propose an iterative algorithm for computing word vectors based on modeling word co-occurrence matrices with Generalized Low Rank Models.Our algorithm generalizes both Skip-gram and GloVe as well as giving rise to other embedding methods based on the specified co-occurrence matrix, distribution of co-occurences, and the number of iterations in the iterative algorithm.For example, using a Tweedie distribution with one iteration results in GloVe and using a Multinomial distribution with full-convergence mode results in Skip-gram.Experimental results demonstrate that multiple iterations of our algorithm improves results over the GloVe method on the Google word analogy similarity task.", "target": ["Wir stellen einen neuen iterativen Algorithmus vor, der auf verallgemeinerten Low-Rank-Modellen zur Berechnung und Interpretation von Worteinbettungsmodellen basiert."]} +{"source": "Deterministic models are approximations of reality that are often easier to build and interpret than stochastic alternatives. Unfortunately, as nature is capricious, observational data can never be fully explained by deterministic models in practice. Observation and process noise need to be added to adapt deterministic models to behave stochastically, such that they are capable of explaining and extrapolating from noisy data.Adding process noise to deterministic simulators can induce a failure in the simulator resulting in no return value for certain inputs -- a property we describe as ``brittle.''We investigate and address the wasted computation that arises from these failures, and the effect of such failures on downstream inference tasks.We show that performing inference in this space can be viewed as rejection sampling, and train a conditional normalizing flow as a proposal over noise values such that there is a low probability that the simulator crashes, increasing computational efficiency and inference fidelity for a fixed sample budget when used as the proposal in an approximate inference algorithm.", "target": ["Wir lernen einen bedingten autoregressiven Ablauf, um Störungen vorzuschlagen, die kein Simulatorversagen hervorrufen, was die Inferenzleistung verbessert."]} +{"source": "Multi-hop question answering requires models to gather information from different parts of a text to answer a question.Most current approaches learn to address this task in an end-to-end way with neural networks, without maintaining an explicit representation of the reasoning process.We propose a method to extract a discrete reasoning chain over the text, which consists of a series of sentences leading to the answer.We then feed the extracted chains to a BERT-based QA model to do final answer prediction.Critically, we do not rely on gold annotated chains or ``supporting facts:'' at training time, we derive pseudogold reasoning chains using heuristics based on named entity recognition and coreference resolution.Nor do we rely on these annotations at test time, as our model learns to extract chains from raw text alone. We test our approach on two recently proposed large multi-hop question answering datasets: WikiHop and HotpotQA, and achieve state-of-art performance on WikiHop and strong performance on HotpotQA.Our analysis shows the properties of chains that are crucial for high performance: in particular, modeling extraction sequentially is important, as is dealing with each candidate sentence in a context-aware way.Furthermore, human evaluation shows that our extracted chains allow humans to give answers with high confidence, indicating that these are a strong intermediate abstraction for this task.", "target": ["Wir verbessern die Beantwortung von Fragen, die Multi-Hop-Argumentation erfordern, indem wir eine Zwischenkette von Sätzen extrahieren."]} +{"source": "Normalizing constant (also called partition function, Bayesian evidence, or marginal likelihood) is one of the central goals of Bayesian inference, yet most of the existing methods are both expensive and inaccurate.Here we develop a new approach, starting from posterior samples obtained with a standard Markov Chain Monte Carlo (MCMC).We apply a novel Normalizing Flow (NF) approach to obtain an analytic density estimator from these samples, followed by Optimal Bridge Sampling (OBS) to obtain the normalizing constant.We compare our method which we call Gaussianized Bridge Sampling (GBS) to existing methods such as Nested Sampling (NS) and Annealed Importance Sampling (AIS) on several examples, showing our method is both significantly faster and substantially more accurate than these methods, and comes with a reliable error estimation.", "target": ["Wir entwickeln eine neue Methode zur Schätzung von Normalisierungskonstanten (Bayes'sche Evidenz) unter Verwendung von Optimal Bridge Sampling und eines neuartigen Normalisierungsflusses, der nachweislich die bestehenden Methoden in Bezug auf Genauigkeit und Rechenzeit übertrifft."]} +{"source": "We present a large-scale empirical study of catastrophic forgetting (CF) in modern Deep Neural Network (DNN) models that perform sequential (or: incremental) learning.A new experimental protocol is proposed that takes into account typical constraints encountered in application scenarios.As the investigation is empirical, we evaluate CF behavior on the hitherto largest number of visual classification datasets, from each of which we construct a representative number of Sequential Learning Tasks (SLTs) in close alignment to previous works on CF.Our results clearly indicate that there is no model that avoids CF for all investigated datasets and SLTs under application conditions.We conclude with a discussion of potential solutions and workarounds to CF, notably for the EWC and IMM models.", "target": ["Wir überprüfen DNN-Modelle auf katastrophales Vergessen anhand eines neuen Bewertungsschemas, das typische Anwendungsbedingungen widerspiegelt, mit überraschenden Ergebnissen."]} +{"source": "Federated Learning (FL) refers to learning a high quality global model based on decentralized data storage, without ever copying the raw data.A natural scenario arises with data created on mobile phones by the activity of their users.Given the typical data heterogeneity in such situations, it is natural to ask how can the global model be personalized for every such device, individually.In this work, we point out that the setting of Model Agnostic Meta Learning (MAML), where one optimizes for a fast, gradient-based, few-shot adaptation to a heterogeneous distribution of tasks, has a number of similarities with the objective of personalization for FL.We present FL as a natural source of practical applications for MAML algorithms, and make the following observations.1) The popular FL algorithm, Federated Averaging, can be interpreted as a meta learning algorithm.2) Careful fine-tuning can yield a global model with higher accuracy, which is at the same time easier to personalize.However, solely optimizing for the global model accuracy yields a weaker personalization result.3) A model trained using a standard datacenter optimization method is much harder to personalize, compared to one trained using Federated Averaging, supporting the first claim.These results raise new questions for FL, MAML, and broader ML research.", "target": ["Federated Averaging ist bereits ein Meta-Learning-Algorithmus, während in Rechenzentren trainierte Methoden wesentlich schwieriger zu personalisieren sind."]} +{"source": "Memorization of data in deep neural networks has become a subject of significant research interest. In this paper, we link memorization of images in deep convolutional autoencoders to downsampling through strided convolution. To analyze this mechanism in a simpler setting, we train linear convolutional autoencoders and show that linear combinations of training data are stored as eigenvectors in the linear operator corresponding to the network when downsampling is used. On the other hand, networks without downsampling do not memorize training data. We provide further evidence that the same effect happens in nonlinear networks. Moreover, downsampling in nonlinear networks causes the model to not only memorize just linear combinations of images, but individual training images. Since convolutional autoencoder components are building blocks of deep convolutional networks, we envision that our findings will shed light on the important phenomenon of memorization in over-parameterized deep networks.", "target": ["Wir identifizieren Downsampling als einen Mechanismus zur Erinnerung in Convolutional-Autoencodern."]} +{"source": "Reinforcement learning provides a powerful and general framework for decisionmaking and control, but its application in practice is often hindered by the needfor extensive feature and reward engineering.Deep reinforcement learning methodscan remove the need for explicit engineering of policy or value features, butstill require a manually specified reward function.Inverse reinforcement learningholds the promise of automatic reward acquisition, but has proven exceptionallydifficult to apply to large, high-dimensional problems with unknown dynamics.Inthis work, we propose AIRL, a practical and scalable inverse reinforcement learningalgorithm based on an adversarial reward learning formulation that is competitivewith direct imitation learning algorithms.Additionally, we show that AIRL isable to recover portable reward functions that are robust to changes in dynamics,enabling us to learn policies even under significant variation in the environmentseen during training.", "target": ["Wir schlagen einen Algorithmus zum inversen Reinforcement Learning vor, der in der Lage ist, Belohnungsfunktionen zu erlernen, die auf neue, ungesehene Umgebungen übertragen werden können."]} +{"source": "We consider two questions at the heart of machine learning; how can we predict if a minimum will generalize to the test set, and why does stochastic gradient descent find minima that generalize well?Our work responds to \\citet{zhang2016understanding}, who showed deep neural networks can easily memorize randomly labeled training data, despite generalizing well on real labels of the same inputs.We show that the same phenomenon occurs in small linear models.These observations are explained by the Bayesian evidence, which penalizes sharp minima but is invariant to model parameterization.We also demonstrate that, when one holds the learning rate fixed, there is an optimum batch size which maximizes the test set accuracy.We propose that the noise introduced by small mini-batches drives the parameters towards minima whose evidence is large.Interpreting stochastic gradient descent as a stochastic differential equation, we identify the ``noise scale\" $g = \\epsilon (\\frac{N}{B} - 1) \\approx \\epsilon N/B$, where $\\epsilon$ is the learning rate, $N$ the training set size and $B$ the batch size.Consequently the optimum batch size is proportional to both the learning rate and the size of the training set, $B_{opt} \\propto \\epsilon N$.We verify these predictions empirically.", "target": ["Die Generalisierung ist stark mit der Bayes'schen Evidenz korreliert, und das Gradientenrauschen minimiert die SGD, deren Evidenz groß ist."]} +{"source": "In the industrial field, the positron annihilation is not affected by complex environment, and the gamma-ray photon penetration is strong, so the nondestructive detection of industrial parts can be realized.Due to the poor image quality caused by gamma-ray photon scattering, attenuation and short sampling time in positron process, we propose the idea of combining deep learning to generate positron images with good quality and clear details by adversarial nets.The structure of the paper is as follows: firstly, we encode to get the hidden vectors of medical CT images based on transfer Learning, and use PCA to extract positron image features.Secondly, we construct a positron image memory based on attention mechanism as a whole input to the adversarial nets which uses medical hidden variables as a query.Finally, we train the whole model jointly and update the input parameters until convergence.Experiments have proved the possibility of generating rare positron images for industrial non-destructive testing using countermeasure networks, and good imaging results have been achieved.", "target": ["Adversarial Nets, Aufmerksamkeitsmechanismus, Positronenbilder, Datenknappheit"]} +{"source": "We revisit the Recurrent Attention Model (RAM, Mnih et al. (2014)), a recurrent neural network for visual attention, from an active information sampling perspective. We borrow ideas from neuroscience research on the role of active information sampling in the context of visual attention and gaze (Gottlieb, 2018), where the author suggested three types of motives for active information sampling strategies.We find the original RAM model only implements one of them.We identify three key weakness of the original RAM and provide a simple solution by adding two extra terms on the objective function.The modified RAM1) achieves faster convergence,2) allows dynamic decision making per sample without loss of accuracy, and3) generalizes much better on longer sequence of glimpses which is not trained for, compared with the original RAM.", "target": ["Inspiriert von der neurowissenschaftlichen Forschung, lösen von drei Hauptschwächen des viel zitierten Recurrent Attention Model, indem Sie einfach zwei Terme zur Zielfunktion hinzufügen."]} +{"source": "Graph Neural Networks (GNNs) for prediction tasks like node classification or edge prediction have received increasing attention in recent machine learning from graphically structured data.However, a large quantity of labeled graphs is difficult to obtain, which significantly limit the true success of GNNs.Although active learning has been widely studied for addressing label-sparse issues with other data types like text, images, etc., how to make it effective over graphs is an open question for research. In this paper, we present the investigation on active learning with GNNs for node classification tasks. Specifically, we propose a new method, which uses node feature propagation followed by K-Medoids clustering of the nodes for instance selection in active learning.With a theoretical bound analysis we justify the design choice of our approach.In our experiments on four benchmark dataset, the proposed method outperforms other representative baseline methods consistently and significantly.", "target": ["Diese Arbeit stellt einen Clustering-basierten aktiven Lernalgorithmus für Graphen vor."]} +{"source": "Continuous Normalizing Flows (CNFs) have emerged as promising deep generative models for a wide range of tasks thanks to their invertibility and exact likelihood estimation.However, conditioning CNFs on signals of interest for conditional image generation and downstream predictive tasks is inefficient due to the high-dimensional latent code generated by the model, which needs to be of the same size as the input data.In this paper, we propose InfoCNF, an efficient conditional CNF that partitions the latent space into a class-specific supervised code and an unsupervised code that shared among all classes for efficient use of labeled information.Since the partitioning strategy (slightly) increases the number of function evaluations (NFEs), InfoCNF also employs gating networks to learn the error tolerances of its ordinary differential equation (ODE) solvers for better speed and performance.We show empirically that InfoCNF improves the test accuracy over the baseline while yielding comparable likelihood scores and reducing the NFEs on CIFAR10.Furthermore, applying the same partitioning strategy in InfoCNF on time-series data helps improve extrapolation performance.", "target": ["Wir schlagen die InfoCNF vor, eine effiziente bedingte CNF, die Gating-Netzwerke einsetzt, um die Fehlertoleranzen der ODE-Solver zu lernen"]} +{"source": "A central goal of unsupervised learning is to acquire representations from unlabeled data or experience that can be used for more effective learning of downstream tasks from modest amounts of labeled data.Many prior unsupervised learning works aim to do so by developing proxy objectives based on reconstruction, disentanglement, prediction, and other metrics.Instead, we develop an unsupervised meta-learning method that explicitly optimizes for the ability to learn a variety of tasks from small amounts of data.To do so, we construct tasks from unlabeled data in an automatic way and run meta-learning over the constructed tasks.Surprisingly, we find that, when integrated with meta-learning, relatively simple task construction mechanisms, such as clustering embeddings, lead to good performance on a variety of downstream, human-specified tasks.Our experiments across four image datasets indicate that our unsupervised meta-learning approach acquires a learning algorithm without any labeled data that is applicable to a wide range of downstream classification tasks, improving upon the embedding learned by four prior unsupervised learning methods.", "target": ["Eine unbeaufsichtigte Lernmethode, die Meta-Lernen verwendet, um effizientes Lernen von nachgelagerten Bildklassifizierungsaufgaben zu ermöglichen, und dabei die modernsten Methoden übertrifft."]} +{"source": "Domain transfer is a exciting and challenging branch of machine learning because models must learn to smoothly transfer between domains, preserving local variations and capturing many aspects of variation without labels. However, most successful applications to date require the two domains to be closely related (ex. image-to-image, video-video), utilizing similar or shared networks to transform domain specific properties like texture, coloring, and line shapes. Here, we demonstrate that it is possible to transfer across modalities (ex. image-to-audio) by first abstracting the data with latent generative models and then learning transformations between latent spaces. We find that a simple variational autoencoder is able to learn a shared latent space to bridge between two generative models in an unsupervised fashion, and even between different types of models (ex. variational autoencoder and a generative adversarial network). We can further impose desired semantic alignment of attributes with a linear classifier in the shared latent space. The proposed variation autoencoder enables preserving both locality and semantic alignment through the transfer process, as shown in the qualitative and quantitative evaluations.Finally, the hierarchical structure decouples the cost of training the base generative models and semantic alignments, enabling computationally efficient and data efficient retraining of personalized mapping functions.", "target": ["Bedingte VAE auf latenten Räumen von vortrainierten generativen Modellen, die den Transfer zwischen drastisch unterschiedlichen Domänen unter Beibehaltung der Lokalität und der semantischen Ausrichtung ermöglicht."]} +{"source": "We propose Adversarial Inductive Transfer Learning (AITL), a method for addressing discrepancies in input and output spaces between source and target domains.AITL utilizes adversarial domain adaptation and multi-task learning to address these discrepancies.Our motivating application is pharmacogenomics where the goal is to predict drug response in patients using their genomic information.The challenge is that clinical data (i.e. patients) with drug response outcome is very limited, creating a need for transfer learning to bridge the gap between large pre-clinical pharmacogenomics datasets (e.g. cancer cell lines) and clinical datasets.Discrepancies exist between1) the genomic data of pre-clinical and clinical datasets (the input space), and2) the different measures of the drug response (the output space).To the best of our knowledge, AITL is the first adversarial inductive transfer learning method to address both input and output discrepancies.Experimental results indicate that AITL outperforms state-of-the-art pharmacogenomics and transfer learning baselines and may guide precision oncology more accurately.", "target": ["Eine neuartige Methode des induktiven Transfer-Lernens, die adversariales Lernen und Multi-Task-Lernen einsetzt, um die Diskrepanz zwischen Input- und Output-Raum zu beheben"]} +{"source": "Named entity recognition (NER) and relation extraction (RE) are two important tasks in information extraction and retrieval (IE & IR).Recent work has demonstrated that it is beneficial to learn these tasks jointly, which avoids the propagation of error inherent in pipeline-based systems and improves performance.However, state-of-the-art joint models typically rely on external natural language processing (NLP) tools, such as dependency parsers, limiting their usefulness to domains (e.g. news) where those tools perform well.The few neural, end-to-end models that have been proposed are trained almost completely from scratch.In this paper, we propose a neural, end-to-end model for jointly extracting entities and their relations which does not rely on external NLP tools and which integrates a large, pre-trained language model.Because the bulk of our model's parameters are pre-trained and we eschew recurrence for self-attention, our model is fast to train.On 5 datasets across 3 domains, our model matches or exceeds state-of-the-art performance, sometimes by a large margin.", "target": ["Eine neue, leistungsstarke Architektur für durchgängige Named Entity Recognition und Relationsextraktionen, die schnell zu trainieren ist."]} +{"source": "In this work we explore a straightforward variational Bayes scheme for Recurrent Neural Networks.Firstly, we show that a simple adaptation of truncated backpropagation through time can yield good quality uncertainty estimates and superior regularisation at only a small extra computational cost during training, also reducing the amount of parameters by 80\\%.Secondly, we demonstrate how a novel kind of posterior approximation yields further improvements to the performance of Bayesian RNNs.We incorporate local gradient information into the approximate posterior to sharpen it around the current batch statistics.We show how this technique is not exclusive to recurrent neural networks and can be applied more widely to train Bayesian neural networks.We also empirically demonstrate how Bayesian RNNs are superior to traditional RNNs on a language modelling benchmark and an image captioning task, as well as showing how each of these methods improve our model over a variety of other schemes for training them.We also introduce a new benchmark for studying uncertainty for language models so future methods can be easily compared.", "target": ["Variationales Bayes-Schema für Recurrent Neural Networks"]} +{"source": "Over the passage of time Unmanned Autonomous Vehicles (UAVs), especiallyAutonomous flying drones grabbed a lot of attention in Artificial Intelligence.Since electronic technology is getting smaller, cheaper and more efficient, hugeadvancement in the study of UAVs has been observed recently.From monitoringfloods, discerning the spread of algae in water bodies to detecting forest trail, theirapplication is far and wide.Our work is mainly focused on autonomous flyingdrones where we establish a case study towards efficiency, robustness and accuracyof UAVs where we showed our results well supported through experiments.We provide details of the software and hardware architecture used in the study.Wefurther discuss about our implementation algorithms and present experiments thatprovide a comparison between three different state-of-the-art algorithms namelyTrailNet, InceptionResnet and MobileNet in terms of accuracy, robustness, powerconsumption and inference time.In our study, we have shown that MobileNet hasproduced better results with very less computational requirement and power consumption.We have also reported the challenges we have faced during our workas well as a brief discussion on our future work to improve safety features andperformance.", "target": ["Fallstudie über optimale Deep Learning Modelle für UAVs"]} +{"source": "Music relies heavily on repetition to build structure and meaning. Self-reference occurs on multiple timescales, from motifs to phrases to reusing of entire sections of music, such as in pieces with ABA structure. The Transformer (Vaswani et al., 2017), a sequence model based on self-attention, has achieved compelling results in many generation tasks that require maintaining long-range coherence.This suggests that self-attention might also be well-suited to modeling music.In musical composition and performance, however, relative timing is critically important. Existing approaches for representing relative positional information in the Transformer modulate attention based on pairwise distance (Shaw et al., 2018). This is impractical for long sequences such as musical compositions since their memory complexity is quadratic in the sequence length. We propose an algorithm that reduces the intermediate memory requirements to linear in the sequence length.This enables us to demonstrate that a Transformer with our modified relative attention mechanism can generate minute-long (thousands of steps) compositions with compelling structure, generate continuations that coherently elaborate on a given motif, and in a seq2seq setup generate accompaniments conditioned on melodies. We evaluate the Transformer with our relative attention mechanism on two datasets, JSB Chorales and Piano-e-competition, and obtain state-of-the-art results on the latter.", "target": ["Wir zeigen die erste erfolgreiche Verwendung von Transformers bei der Erzeugung von Musik, die eine langfristige Struktur aufweist."]} +{"source": "Sequential decision problems for real-world applications often need to be solved in real-time, requiring algorithms to perform well with a restricted computational budget.Width-based lookaheads have shown state-of-the-art performance in classical planning problems as well as over the Atari games with tight budgets.In this work we investigate width-based lookaheads over Stochastic Shortest paths (SSP).We analyse why width-based algorithms perform poorly over SSP problems, and overcome these pitfalls proposing a method to estimate costs-to-go.We formalize width-based lookaheads as an instance of the rollout algorithm, give a definition of width for SSP problems and explain its sample complexity.Our experimental results over a variety of SSP benchmarks show the algorithm to outperform other state-of-the-art rollout algorithms such as UCT and RTDP.", "target": ["Wir schlagen einen neuen Monte-Carlo-Baum-Such- / Rollout-Algorithmus vor, der auf Breiten-basierte Suche stützt, um eine Lookahead zu konstruieren."]} +{"source": "Deep Neural Networks (DNNs) are known for excellent performance in supervised tasks such as classification.Convolutional Neural Networks (CNNs), in particular, can learn effective features and build high-level representations that can be used forclassification, but also for querying and nearest neighbor search.However, CNNs have also been shown to suffer from a performance drop when the distribution of the data changes from training to test data.In this paper we analyze the internalrepresentations of CNNs and observe that the representations of unseen data in each class, spread more (with higher variance) in the embedding space of the CNN compared to representations of the training data.More importantly, this difference is more extreme if the unseen data comes from a shifted distribution.Based on this observation, we objectively evaluate the degree of representation’s variance in each class by applying eigenvalue decomposition on the within-class covariance of the internal representations of CNNs and observe the same behaviour.This can be problematic as larger variances might lead to mis-classification if the sample crosses the decision boundary of its class.We apply nearest neighbor classification on the representations and empirically show that the embeddings with the high variance actually have significantly worse KNN classification performances, although this could not be foreseen from their end-to-end classification results.To tackle this problem, we propose Deep Within-Class Covariance Analysis (DWCCA), a deep neural network layer that significantly reduces the within-class covariance of a DNN’s representation, improving performance on unseen test data from a shifted distribution.We empirically evaluate DWCCA on two datasets for Acoustic Scene Classification (DCASE2016 and DCASE2017).We demonstrate that not only does DWCCA significantly improve the network’s internal representation, italso increases the end-to-end classification accuracy, especially when the test set exhibits a slight distribution shift.By adding DWCCA to a VGG neural network, we achieve around 6 percentage points improvement in the case of a distributionmismatch.", "target": ["Wir schlagen eine neuartige Deep Neural Network Schicht zur Normalisierung der klasseninternen Kovarianz einer internen Repräsentation in einem neuronalen Netzwerk vor, die zu einer deutlichen Verbesserung der Generalisierung der gelernten Repräsentationen führt."]} +{"source": "Generative models have proven to be an outstanding tool for representing high-dimensional probability distributions and generating realistic looking images.A fundamental characteristic of generative models is their ability to produce multi-modal outputs.However, while training, they are often susceptible to mode collapse, which means that the model is limited in mapping the input noise to only a few modes of the true data distribution.In this paper, we draw inspiration from Determinantal Point Process (DPP) to devise a generative model that alleviates mode collapse while producing higher quality samples.DPP is an elegant probabilistic measure used to model negative correlations within a subset and hence quantify its diversity.We use DPP kernel to model the diversity in real data as well as in synthetic data.Then, we devise a generation penalty term that encourages the generator to synthesize data with a similar diversity to real data.In contrast to previous state-of-the-art generative models that tend to use additional trainable parameters or complex training paradigms, our method does not change the original training scheme.Embedded in an adversarial training and variational autoencoder, our Generative DPP approach shows a consistent resistance to mode-collapse on a wide-variety of synthetic data and natural image datasets including MNIST, CIFAR10, and CelebA, while outperforming state-of-the-art methods for data-efficiency, convergence-time, and generation quality.Our code will be made publicly available.", "target": ["Die Hinzufügung eines Diversitätskriteriums, das von DPP inspiriert ist, in das GAN-Ziel vermeidet den Moduszusammenbruch und führt zu besseren Generationen."]} +{"source": "Despite existing work on ensuring generalization of neural networks in terms of scale sensitive complexity measures, such as norms, margin and sharpness, these complexity measures do not offer an explanation of why neural networks generalize better with over-parametrization.In this work we suggest a novel complexity measure based on unit-wise capacities resulting in a tighter generalization bound for two layer ReLU networks.Our capacity bound correlates with the behavior of test error with increasing network sizes (within the range reported in the experiments), and could partly explain the improvement in generalization with over-parametrization.We further present a matching lower bound for the Rademacher complexity that improves over previous capacity lower bounds for neural networks.", "target": ["Wir schlagen eine Generalisierungsgrenze vor, die die Verbesserung der Generalisierungen durch Überparametrisierung teilweise erklären könnte."]} +{"source": "We introduce three generic point cloud processing blocks that improve both accuracy and memory consumption of multiple state-of-the-art networks, thus allowing to design deeper and more accurate networks.The novel processing blocks that facilitate efficient information flow are a convolution-type operation block for point sets that blends neighborhood information in a memory-efficient manner; a multi-resolution point cloud processing block; and a crosslink block that efficiently shares information across low- and high-resolution processing branches.Combining these blocks, we design significantly wider and deeper architectures.We extensively evaluate the proposed architectures on multiple point segmentation benchmarks (ShapeNetPart, ScanNet, PartNet) and report systematic improvements in terms of both accuracy and memory consumption by using our generic modules in conjunction with multiple recent architectures (PointNet++, DGCNN, SpiderCNN, PointCNN).We report a 9.7% increase in IoU on the PartNet dataset, which is the most complex, while decreasing memory footprint by 57%.", "target": ["Wir stellen drei generische Cloud Verarbeitungsblöcke vor, die sowohl die Genauigkeit als auch den Speicherverbrauch mehrerer moderner Netzwerke verbessern und somit die Entwicklung tieferer und genauerer Netzwerke ermöglichen."]} +{"source": "End-to-end acoustic-to-word speech recognition models have recently gained popularity because they are easy to train, scale well to large amounts of training data, and do not require a lexicon.In addition, word models may also be easier to integrate with downstream tasks such as spoken language understanding, because inference (search) is much simplified compared to phoneme, character or any other sort of sub-word units.In this paper, we describe methods to construct contextual acoustic word embeddings directly from a supervised sequence-to-sequence acoustic-to-word speech recognition model using the learned attention distribution.On a suite of 16 standard sentence evaluation tasks, our embeddings show competitive performance against a word2vec model trained on the speech transcriptions.In addition, we evaluate these embeddings on a spoken language understanding task and observe that our embeddings match the performance of text-based embeddings in a pipeline of first performing speech recognition and then constructing word embeddings from transcriptions.", "target": ["Methoden zum Erlernen kontextueller akustischer Worteinbettungen aus einem Ende-zu-Ende-Spracherkennungsmodell, die mit textbasierten Worteinbettungen konkurrieren können."]} +{"source": "Unsupervised monocular depth estimation has made great progress after deeplearning is involved.Training with binocular stereo images is considered as agood option as the data can be easily obtained.However, the depth or disparityprediction results show poor performance for the object boundaries.The mainreason is related to the handling of occlusion areas during the training.In this paper,we propose a novel method to overcome this issue.Exploiting disparity mapsproperty, we generate an occlusion mask to block the back-propagation of the occlusionareas during image warping.We also design new networks with flippedstereo images to induce the networks to learn occluded boundaries.It shows thatour method achieves clearer boundaries and better evaluation results on KITTIdriving dataset and Virtual KITTI dataset.", "target": ["In dieser Arbeit wird eine Maskenmethode vorgeschlagen, die die bisherigen unscharfen Ergebnisse der unüberwachten monokularen Tiefenschätzung, die durch Verdeckung verursacht werden, löst."]} +{"source": "Graph classification is currently dominated by graph kernels, which, while powerful, suffer some significant limitations.Convolutional Neural Networks (CNNs) offer a very appealing alternative.However, processing graphs with CNNs is not trivial.To address this challenge, many sophisticated extensions of CNNs have recently been proposed.In this paper, we reverse the problem: rather than proposing yet another graph CNN model, we introduce a novel way to represent graphs as multi-channel image-like structures that allows them to be handled by vanilla 2D CNNs.Despite its simplicity, our method proves very competitive to state-of-the-art graph kernels and graph CNNs, and outperforms them by a wide margin on some datasets.It is also preferable to graph kernels in terms of time complexity.Code and data are publicly available.", "target": ["Wir stellen eine neuartige Methode zur Darstellung von Graphen als mehrkanalige bildähnliche Strukturen vor, die es ermöglicht, sie mit einfachen 2D CNNs zu verarbeiten."]} +{"source": "The key attribute that drives the unprecedented success of modern Recurrent Neural Networks (RNNs) on learning tasks which involve sequential data, is their ever-improving ability to model intricate long-term temporal dependencies.However, a well established measure of RNNs' long-term memory capacity is lacking, and thus formal understanding of their ability to correlate data throughout time is limited.Though depth efficiency in convolutional networks is well established by now, it does not suffice in order to account for the success of deep RNNs on inputs of varying lengths, and the need to address their 'time-series expressive power' arises.In this paper, we analyze the effect of depth on the ability of recurrent networks to express correlations ranging over long time-scales.To meet the above need, we introduce a measure of the information flow across time that can be supported by the network, referred to as the Start-End separation rank.Essentially, this measure reflects the distance of the function realized by the recurrent network from a function that models no interaction whatsoever between the beginning and end of the input sequence.We prove that deep recurrent networks support Start-End separation ranks which are exponentially higher than those supported by their shallow counterparts.Moreover, we show that the ability of deep recurrent networks to correlate different parts of the input sequence increases exponentially as the input sequence extends, while that of vanilla shallow recurrent networks does not adapt to the sequence length at all.Thus, we establish that depth brings forth an overwhelming advantage in the ability of recurrent networks to model long-term dependencies, and provide an exemplar of quantifying this key attribute which may be readily extended to other RNN architectures of interest, e.g. variants of LSTM networks.We obtain our results by considering a class of recurrent networks referred to as Recurrent Arithmetic Circuits (RACs), which merge the hidden state with the input via the Multiplicative Integration operation.", "target": ["Wir schlagen ein Maß für das Langzeitgedächtnis vor und beweisen, dass tiefe Recurrent Networks viel besser geeignet sind, langfristige zeitliche Abhängigkeiten zu modellieren als untiefe Netze."]} +{"source": "Holistically exploring the perceptual and neural representations underlying animal communication has traditionally been very difficult because of the complexity of the underlying signal.We present here a novel set of techniques to project entire communicative repertoires into low dimensional spaces that can be systematically sampled from, exploring the relationship between perceptual representations, neural representations, and the latent representational spaces learned by machine learning algorithms.We showcase this method in one ongoing experiment studying sequential and temporal maintenance of context in songbird neural and perceptual representations of syllables.We further discuss how studying the neural mechanisms underlying the maintenance of the long-range information content present in birdsong can inform and be informed by machine sequence modeling.", "target": ["Wir vergleichen Wahrnehmungs-, neuronale und modellierte Darstellungen der tierischen Kommunikation mit maschinellem Lernen, Verhalten und Physiologie."]} +{"source": "The information bottleneck principle (Shwartz-Ziv & Tishby, 2017) suggests that SGD-based training of deep neural networks results in optimally compressed hidden layers, from an information theoretic perspective.However, this claim was established on toy data.The goal of the work we present here is to test these claims in a realistic setting using a larger and deeper convolutional architecture, a ResNet model.We trained PixelCNN++ models as inverse representation decoders to measure the mutual information between hidden layers of a ResNet and input image data, when trained for (1) classification and (2) autoencoding.We find that two stages of learning happen for both training regimes, and that compression does occur, even for an autoencoder.Sampling images by conditioning on hidden layers’ activations offers an intuitive visualisation to understand what a ResNets learns to forget.", "target": ["Das Prinzip des Informationsengpasses, angewandt auf ResNets, unter Verwendung von PixelCNN++ Modellen zur Dekodierung der gegenseitigen Information und zur bedingten Erzeugung von Bildern zur Informationsdarstellung"]} +{"source": "We study the problem of safe adaptation: given a model trained on a variety of past experiences for some task, can this model learn to perform that task in a new situation while avoiding catastrophic failure?This problem setting occurs frequently in real-world reinforcement learning scenarios such as a vehicle adapting to drive in a new city, or a robotic drone adapting a policy trained only in simulation.While learning without catastrophic failures is exceptionally difficult, prior experience can allow us to learn models that make this much easier.These models might not directly transfer to new settings, but can enable cautious adaptation that is substantially safer than na\\\"{i}ve adaptation as well as learning from scratch.Building on this intuition, we propose risk-averse domain adaptation (RADA).RADA works in two steps: it first trains probabilistic model-based RL agents in a population of source domains to gain experience and capture epistemic uncertainty about the environment dynamics.Then, when dropped into a new environment, it employs a pessimistic exploration policy, selecting actions that have the best worst-case performance as forecasted by the probabilistic model.We show that this simple maximin policy accelerates domain adaptation in a safety-critical driving environment with varying vehicle sizes.We compare our approach against other approaches for adapting to new environments, including meta-reinforcement learning.", "target": ["Die Anpassung eines RL-Agenten in einer Zielumgebung mit unbekannter Dynamik ist schnell und sicher, wenn wir vorherige Erfahrungen in einer Vielzahl von Umgebungen übertragen und dann risikoaverse Aktionen während der Anpassung auswählen."]} +{"source": "We propose the Neuro-Symbolic Concept Learner (NS-CL), a model that learns visual concepts, words, and semantic parsing of sentences without explicit supervision on any of them; instead, our model learns by simply looking at images and reading paired questions and answers.Our model builds an object-based scene representation and translates sentences into executable, symbolic programs.To bridge the learning of two modules, we use a neuro-symbolic reasoning module that executes these programs on the latent scene representation.Analogical to human concept learning, the perception module learns visual concepts based on the language description of the object being referred to.Meanwhile, the learned visual concepts facilitate learning new words and parsing new sentences.We use curriculum learning to guide the searching over the large compositional space of images and language.Extensive experiments demonstrate the accuracy and efficiency of our model on learning visual concepts, word representations, and semantic parsing of sentences.Further, our method allows easy generalization to new object attributes, compositions, language concepts, scenes and questions, and even new program domains.It also empowers applications including visual question answering and bidirectional image-text retrieval.", "target": ["Wir schlagen den Neuro-Symbolic Concept Learner (NS-CL) vor, ein Modell, das visuelle Konzepte, Wörter und das semantische Parsing von Sätzen ohne explizite Überwachung erlernt."]} +{"source": "Bayesian inference offers a theoretically grounded and general way to train neural networks and can potentially give calibrated uncertainty.However, it is challenging to specify a meaningful and tractable prior over the network parameters, and deal with the weight correlations in the posterior.To this end, this paper introduces two innovations:(i) a Gaussian process-based hierarchical model for the network parameters based on recently introduced unit embeddings that can flexibly encode weight structures, and(ii) input-dependent contextual variables for the weight prior that can provide convenient ways to regularize the function space being modeled by the network through the use of kernels. We show these models provide desirable test-time uncertainty estimates, demonstrate cases of modeling inductive biases for neural networks with kernels and demonstrate competitive predictive performance on an active learning benchmark.", "target": ["Wir führen einen Gauß'schen Prozessprior über Gewichte in einem neuronalen Netz ein und erforschen seine Fähigkeit, eingabeabhängige Gewichte zu modellieren, mit Vorteilen für verschiedene Aufgaben, einschließlich Unsicherheitsabschätzung und Generalisierung in einer Umgebung mit geringer Stichprobe."]} +{"source": "We perform an in-depth investigation of the suitability of self-attention models for character-level neural machine translation.We test the standard transformer model, as well as a novel variant in which the encoder block combines information from nearby characters using convolution.We perform extensive experiments on WMT and UN datasets, testing both bilingual and multilingual translation to English using up to three input languages (French, Spanish, and Chinese).Our transformer variant consistently outperforms the standard transformer at the character-level and converges faster while learning more robust character-level alignments.", "target": ["Wir führen eine eingehende Untersuchung der Eignung von Selg-Attention Modellen für die neuronale maschinelle Übersetzung auf Zeichenebene durch."]} +{"source": "The field of medical diagnostics contains a wealth of challenges which closely resemble classical machine learning problems; practical constraints, however, complicate the translation of these endpoints naively into classical architectures.Many tasks in radiology, for example, are largely problems of multi-label classification wherein medical images are interpreted to indicate multiple present or suspected pathologies.Clinical settings drive the necessity for high accuracy simultaneously across a multitude of pathological outcomes and greatly limit the utility of tools which consider only a subset.This issue is exacerbated by a general scarcity of training data and maximizes the need to extract clinically relevant features from available samples -- ideally without the use of pre-trained models which may carry forward undesirable biases from tangentially related tasks.We present and evaluate a partial solution to these constraints in using LSTMs to leverage interdependencies among target labels in predicting 14 pathologic patterns from chest x-rays and establish state of the art results on the largest publicly available chest x-ray dataset from the NIH without pre-training.Furthermore, we propose and discuss alternative evaluation metrics and their relevance in clinical practice.", "target": ["Wir präsentieren die modernsten Ergebnisse der Verwendung neuronaler Netze zur Diagnose von Röntgenaufnahmen des Brustkorbs"]} +{"source": "Semmelhack et al. (2014) have achieved high classification accuracy in distinguishing swim bouts of zebrafish using a Support Vector Machine (SVM).Convolutional Neural Networks (CNNs) have reached superior performance in various image recognition tasks over SVMs, but these powerful networks remain a black box.Reaching better transparency helps to build trust in their classifications and makes learned features interpretable to experts.Using a recently developed technique called Deep Taylor Decomposition, we generated heatmaps to highlight input regions of high relevance for predictions.We find that our CNN makes predictions by analyzing the steadiness of the tail's trunk, which markedly differs from the manually extracted features used by Semmelhack et al. (2014).We further uncovered that the network paid attention to experimental artifacts.Removing these artifacts ensured the validity of predictions.After correction, our best CNN beats the SVM by 6.12%, achieving a classification accuracy of 96.32%.Our work thus demonstrates the utility of AI explainability for CNNs.", "target": ["Wir demonstrieren den Nutzen einer neuen KI-Erklärungstechnik, indem wir die gelernten Merkmale eines CNN, das auf die binäre Klassifizierung von Zebrafischbewegungen trainiert wurde, visualisieren."]} +{"source": "When communicating, humans rely on internally-consistent language representations.That is, as speakers, we expect listeners to behave the same way we do when we listen.This work proposes several methods for encouraging such internal consistency in dialog agents in an emergent communication setting.We consider two hypotheses about the effect of internal-consistency constraints:1) that they improve agents’ ability to refer to unseen referents, and2) that they improve agents’ ability to generalize across communicative roles (e.g. performing as a speaker de- spite only being trained as a listener).While we do not find evidence in favor of the former, our results show significant support for the latter.", "target": ["Interne Konsistenzbeschränkungen verbessern die Fähigkeit von Agenten, emergente Protokolle zu entwickeln, die sich über kommunikative Rollen hinweg verallgemeinern lassen."]} +{"source": "Neural networks (NNs) are able to perform tasks that rely on compositional structure even though they lack obvious mechanisms for representing this structure.To analyze the internal representations that enable such success, we propose ROLE, a technique that detects whether these representations implicitly encode symbolic structure.ROLE learns to approximate the representations of a target encoder E by learning a symbolic constituent structure and an embedding of that structure into E’s representational vector space.The constituents of the approximating symbol structure are defined by structural positions — roles — that can be filled by symbols.We show that when E is constructed to explicitly embed a particular type of structure (e.g., string or tree), ROLE successfully extracts the ground-truth roles defining that structure.We then analyze a seq2seq network trained to perform a more complex compositional task (SCAN), where there is no ground truth role scheme available.For this model, ROLE successfully discovers an interpretable symbolic structure that the model implicitly uses to perform the SCAN task, providing a comprehensive account of the link between the representations and the behavior of a notoriously hard-to-interpret type of model.We verify the causal importance of the discovered symbolic structure by showing that, when we systematically manipulate hidden embeddings based on this symbolic structure, the model’s output is also changed in the way predicted by our analysis.Finally, we use ROLE to explore whether popular sentence embedding models are capturing compositional structure and find evidence that they are not; we conclude by discussing how insights from ROLE can be used to impart new inductive biases that will improve the compositional abilities of such models.", "target": ["Wir stellen eine neue Analysetechnik vor, die interpretierbare kompositorische Strukturen in bekanntermaßen schwer zu interpretierenden Recurrent Neural Networks erkennt."]} +{"source": "The vertebrate visual system is hierarchically organized to process visual information in successive stages.Neural representations vary drastically across the first stages of visual processing: at the output of the retina, ganglion cell receptive fields (RFs) exhibit a clear antagonistic center-surround structure, whereas in the primary visual cortex (V1), typical RFs are sharply tuned to a precise orientation.There is currently no unified theory explaining these differences in representations across layers.Here, using a deep convolutional neural network trained on image recognition as a model of the visual system, we show that such differences in representation can emerge as a direct consequence of different neural resource constraints on the retinal and cortical networks, and for the first time we find a single model from which both geometries spontaneously emerge at the appropriate stages of visual processing.The key constraint is a reduced number of neurons at the retinal output, consistent with the anatomy of the optic nerve as a stringent bottleneck.Second, we find that, for simple downstream cortical networks, visual representations at the retinal output emerge as nonlinear and lossy feature detectors, whereas they emerge as linear and faithful encoders of the visual scene for more complex cortical networks.This result predicts that the retinas of small vertebrates (e.g. salamander, frog) should perform sophisticated nonlinear computations, extracting features directly relevant to behavior, whereas retinas of large animals such as primates should mostly encode the visual scene linearly and respond to a much broader range of stimuli.These predictions could reconcile the two seemingly incompatible views of the retina as either performing feature extraction or efficient coding of natural scenes, by suggesting that all vertebrates lie on a spectrum between these two objectives, depending on the degree of neural resources allocated to their visual system.", "target": ["Wir haben die neuronalen Repräsentationen biologischer visueller Systeme reproduziert, indem wir deren neuronale Ressourcenbeschränkungen in einem tiefen Convolutional Modell simuliert haben."]} +{"source": "While it has not yet been proven, empirical evidence suggests that model generalization is related to local properties of the optima which can be described via the Hessian.We connect model generalization with the local property of a solution under the PAC-Bayes paradigm.In particular, we prove that model generalization ability is related to the Hessian, the higher-order \"smoothness\" terms characterized by the Lipschitz constant of the Hessian, and the scales of the parameters.Guided by the proof, we propose a metric to score the generalization capability of the model, as well as an algorithm that optimizes the perturbed model accordingly.", "target": ["Eine Theorie, die die Hessian der Lösung und die Verallgemeinerungsfähigkeit des Modells verbindet"]} +{"source": "Unsupervised learning is about capturing dependencies between variables and is driven by the contrast between the probable vs improbable configurations of these variables, often either via a generative model which only samples probable ones or with an energy function (unnormalized log-density) which is low for probable ones and high for improbable ones.Here we consider learning both an energy function and an efficient approximate sampling mechanism for the corresponding distribution.Whereas the critic (or discriminator) in generative adversarial networks (GANs) learns to separate data and generator samples, introducing an entropy maximization regularizer on the generator can turn the interpretation of the critic into an energy function, which separates the training distribution from everything else, and thus can be used for tasks like anomaly or novelty detection. This paper is motivated by the older idea of sampling in latent space rather than data space because running a Monte-Carlo Markov Chain (MCMC) in latent space has been found to be easier and more efficient, and because a GAN-like generator can convert latent space samples to data space samples.For this purpose, we show how a Markov chain can be run in latent space whose samples can be mapped to data space, producing better samples.These samples are also used for the negative phase gradient required to estimate the log-likelihood gradient of the data space energy function.To maximize entropy at the output of the generator, we take advantage of recently introduced neural estimators of mutual information.We find that in addition to producing a useful scoring function for anomaly detection, the resulting approach produces sharp samples (like GANs) while covering the modes well, leading to high Inception and Fréchet scores.", "target": ["Wir haben die Entropiemaximierung in GANs eingeführt, was zu einer Neuinterpretation der Kritik als Energiefunktion führt."]} +{"source": "Neural Style Transfer has become a popular technique forgenerating images of distinct artistic styles using convolutional neural networks.Thisrecent success in image style transfer has raised the question ofwhether similar methods can be leveraged to alter the “style” of musicalaudio.In this work, we attempt long time-scale high-quality audio transferand texture synthesis in the time-domain that captures harmonic,rhythmic, and timbral elements related to musical style, using examples thatmay have different lengths and musical keys.We demonstrate the abilityto use randomly initialized convolutional neural networks to transferthese aspects of musical style from one piece onto another using 3different representations of audio: the log-magnitude of the Short TimeFourier Transform (STFT), the Mel spectrogram, and the Constant-Q Transformspectrogram.We propose using these representations as a way ofgenerating and modifying perceptually significant characteristics ofmusical audio content.We demonstrate each representation'sshortcomings and advantages over others by carefully designingneural network structures that complement the nature of musical audio.Finally, we show that the mostcompelling “style” transfer examples make use of an ensemble of theserepresentations to help capture the varying desired characteristics ofaudio signals.", "target": ["Wir stellen einen langzeit skalierten Algorithmus zur Übertragung von musikalischen Audiostilen vor, der Audio im Zeitbereich synthetisiert, aber Zeit-Frequenz-Darstellungen von Audio verwendet."]} +{"source": "To communicate with new partners in new contexts, humans rapidly form new linguistic conventions.Recent language models trained with deep neural networks are able to comprehend and produce the existing conventions present in their training data, but are not able to flexibly and interactively adapt those conventions on the fly as humans do.We introduce a repeated reference task as a benchmark for models of adaptation in communication and propose a regularized continual learning framework that allows an artificial agent initialized with a generic language model to more accurately and efficiently understand their partner over time.We evaluate this framework through simulations on COCO and in real-time reference game experiments with human partners.", "target": ["Wir schlagen eine wiederholte Referenz-Benchmark-Aufgabe und einen regulierten, kontinuierlichen Lernansatz für die adaptive Kommunikation mit Menschen in unbekannten Domänen vor"]} +{"source": "Traditional set prediction models can struggle with simple datasets due to an issue we call the responsibility problem.We introduce a pooling method for sets of feature vectors based on sorting features across elements of the set.This can be used to construct a permutation-equivariant auto-encoder that avoids this responsibility problem.On a toy dataset of polygons and a set version of MNIST, we show that such an auto-encoder produces considerably better reconstructions and representations.Replacing the pooling function in existing set encoders with FSPool improves accuracy and convergence speed on a variety of datasets.", "target": ["Sortieren im Kodierer und Rückgängigmachen der Sortierung im Dekodierer zur Vermeidung von Verantwortungsproblemen bei eingestellten Auto-Enkodierern"]} +{"source": "We present a method for policy learning to navigate indoor environments.We adopt a hierarchical policy approach, where two agents are trained to work in cohesion with one another to perform a complex navigation task.A Planner agent operates at a higher level and proposes sub-goals for an Executor agent.The Executor reports an embedding summary back to the Planner as additional side information at the end of its series of operations for the Planner's next sub-goal proposal.The end goal is generated by the environment and exposed to the Planner which then decides which set of sub-goals to propose to the Executor.We show that this Planner-Executor setup drastically increases the sample efficiency of our method over traditional single agent approaches, effectively mitigating the difficulty accompanying long series of actions with a sparse reward signal.On the challenging Habitat environment which requires navigating various realistic indoor environments, we demonstrate that our approach offers a significant improvement over prior work for navigation.", "target": ["Wir präsentieren einen hierarchischen Lernrahmen für die Navigation in einer enthaltenen Lernumgebung"]} +{"source": "Saliency methods aim to explain the predictions of deep neural networks.These methods lack reliability when the explanation is sensitive to factors that do not contribute to the model prediction.We use a simple and common pre-processing step ---adding a mean shift to the input data--- to show that a transformation with no effect on the model can cause numerous methods to incorrectly attribute.We define input invariance as the requirement that a saliency method mirror the sensitivity of the model with respect to transformations of the input.We show, through several examples, that saliency methods that do not satisfy a input invariance property are unreliable and can lead to misleading and inaccurate attribution.", "target": ["Attribution kann manchmal irreführend sein"]} +{"source": "Large Transformer models routinely achieve state-of-the-art results ona number of tasks but training these models can be prohibitively costly,especially on long sequences.We introduce two techniques to improvethe efficiency of Transformers.For one, we replace dot-product attentionby one that uses locality-sensitive hashing, changing its complexityfrom O(L^2) to O(L), where L is the length of the sequence.Furthermore, we use reversible residual layers instead of the standardresiduals, which allows storing activations only once in the trainingprocess instead of N times, where N is the number of layers.The resulting model, the Reformer, performs on par with Transformer modelswhile being much more memory-efficient and much faster on long sequences.", "target": ["Effizienter Transformer mit ortsabhängigem Hashing und reversiblen Schichten"]} +{"source": "Obtaining policies that can generalise to new environments in reinforcement learning is challenging.In this work, we demonstrate that language understanding via a reading policy learner is a promising vehicle for generalisation to new environments.We propose a grounded policy learning problem, Read to Fight Monsters (RTFM), in which the agent must jointly reason over a language goal, relevant dynamics described in a document, and environment observations.We procedurally generate environment dynamics and corresponding language descriptions of the dynamics, such that agents must read to understand new environment dynamics instead of memorising any particular information.In addition, we propose txt2π, a model that captures three-way interactions between the goal, document, and observations.On RTFM, txt2π generalises to new environments with dynamics not seen during training via reading.Furthermore, our model outperforms baselines such as FiLM and language-conditioned CNNs on RTFM.Through curriculum learning, txt2π produces policies that excel on complex RTFM tasks requiring several reasoning and coreference steps.", "target": ["Wir zeigen, dass Sprachverständnis durch Lesen ein vielversprechender Weg ist, um Regeln zu lernen, die sich auf neue Umgebungen übertragen lassen."]} +{"source": "An open question in the Deep Learning community is why neural networks trained with Gradient Descent generalize well on real datasets even though they are capable of fitting random data.We propose an approach to answering this question based on a hypothesis about the dynamics of gradient descent that we call Coherent Gradients: Gradients from similar examples are similar and so the overall gradient is stronger in certain directions where these reinforce each other.Thus changes to the network parameters during training are biased towards those that (locally) simultaneously benefit many examples when such similarity exists.We support this hypothesis with heuristic arguments and perturbative experiments and outline how this can explain several common empirical observations about Deep Learning.Furthermore, our analysis is not just descriptive, but prescriptive.It suggests a natural modification to gradient descent that can greatly reduce overfitting.", "target": ["Wir stellen eine Hypothese auf, warum Gradient-Descent verallgemeinert, und zwar auf der Grundlage der Interaktion zwischen den Gradienten pro Beispiel."]} +{"source": "Recent advances in deep learning have shown promising results in many low-level vision tasks.However, solving the single-image-based view synthesis is still an open problem.In particular, the generation of new images at parallel camera views given a single input image is of great interest, as it enables 3D visualization of the 2D input scenery.We propose a novel network architecture to perform stereoscopic view synthesis at arbitrary camera positions along the X-axis, or Deep 3D Pan, with \"t-shaped\" adaptive kernels equipped with globally and locally adaptive dilations.Our proposed network architecture, the monster-net, is devised with a novel t-shaped adaptive kernel with globally and locally adaptive dilation, which can efficiently incorporate global camera shift into and handle local 3D geometries of the target image's pixels for the synthesis of naturally looking 3D panned views when a 2-D input image is given.Extensive experiments were performed on the KITTI, CityScapes and our VXXLXX_STEREO indoors dataset to prove the efficacy of our method.Our monster-net significantly outperforms the state-of-the-art method, SOTA, by a large margin in all metrics of RMSE, PSNR, and SSIM.Our proposed monster-net is capable of reconstructing more reliable image structures in synthesized images with coherent geometry.Moreover, the disparity information that can be extracted from the \"t-shaped\" kernel is much more reliable than that of the SOTA for the unsupervised monocular depth estimation task, confirming the effectiveness of our method.", "target": ["Neue Architektur für die Synthese stereoskopischer Ansichten bei beliebigen Kameraverschiebungen unter Verwendung adaptiver t-förmiger Kernel mit adaptiven Dilatationen."]} +{"source": "Deep Neutral Networks(DNNs) require huge GPU memory when training on modern image/video databases.Unfortunately, the GPU memory as a hardware resource is always finite, which limits the image resolution, batch size, and learning rate that could be used for better DNN performance.In this paper, we propose a novel training approach, called Re-forwarding, that substantially reduces memory usage in training.Our approach automatically finds a subset of vertices in a DNN computation graph, and stores tensors only at these vertices during the first forward.During backward, extra local forwards (called the Re-forwarding process) are conducted to compute the missing tensors between the subset of vertices.The total memory cost becomes the sum of (1) the memory cost at the subset of vertices and (2) the maximum memory cost among local re-forwards.Re-forwarding trades training time overheads for memory and does not compromise any performance in testing.We propose theories and algorithms that achieve the optimal memory solutions for DNNs with either linear or arbitrary computation graphs.Experiments show that Re-forwarding cuts down up-to 80% of training memory on popular DNNs such as Alexnet, VGG, ResNet, Densenet and Inception net.", "target": ["In diesem Beitrag werden eine grundlegende Theorie und optimale Algorithmen für das DNN-Training vorgeschlagen, die den Trainingsspeicher für gängige DNNs um bis zu 80% reduzieren."]} +{"source": "Compression is a key step to deploy large neural networks on resource-constrained platforms.As a popular compression technique, quantization constrains the number of distinct weight values and thus reducing the number of bits required to represent and store each weight.In this paper, we study the representation power of quantized neural networks.First, we prove the universal approximability of quantized ReLU networks on a wide class of functions.Then we provide upper bounds on the number of weights and the memory size for a given approximation error bound and the bit-width of weights for function-independent and function-dependent structures.Our results reveal that, to attain an approximation error bound of $\\epsilon$, the number of weights needed by a quantized network is no more than $\\mathcal{O}\\left(\\log^5(1/\\epsilon)\\right)$ times that of an unquantized network.This overhead is of much lower order than the lower bound of the number of weights needed for the error bound, supporting the empirical success of various quantization techniques.To the best of our knowledge, this is the first in-depth study on the complexity bounds of quantized neural networks.", "target": ["Dieses Papier beweist die universelle Approximierbarkeit von quantisierten neuronalen ReLU-Netzen und legt die Komplexitätsgrenze bei beliebigem Fehler dar."]} +{"source": "Reinforcement learning (RL) with value-based methods (e.g., Q-learning) has shown success in a variety of domains such asgames and recommender systems (RSs).When the action space is finite, these algorithms implicitly finds a policy by learning the optimal value function, which are often very efficient. However, one major challenge of extending Q-learning to tackle continuous-action RL problems is that obtaining optimal Bellman backup requires solving a continuous action-maximization (max-Q) problem.While it is common to restrict the parameterization of the Q-function to be concave in actions to simplify the max-Q problem, such a restriction might lead to performance degradation.Alternatively, when the Q-function is parameterized with a generic feed-forward neural network (NN), the max-Q problem can be NP-hard.In this work, we propose the CAQL method which minimizes the Bellman residual using Q-learning with one of several plug-and-play action optimizers.In particular, leveraging the strides of optimization theories in deep NN, we show that max-Q problem can be solved optimally with mixed-integer programming (MIP)---when the Q-function has sufficient representation power, this MIP-based optimization induces better policies and is more robust than counterparts, e.g., CEM or GA, that approximate the max-Q solution.To speed up training of CAQL, we develop three techniques, namely(i) dynamic tolerance,(ii) dual filtering, and(iii) clustering.To speed up inference of CAQL, we introduce the action function that concurrently learns the optimal policy.To demonstrate the efficiency of CAQL we compare it with state-of-the-art RL algorithms on benchmark continuous control problems that have different degrees of action constraints and show that CAQL significantly outperforms policy-based methods in heavily constrained environments.", "target": ["Ein allgemeiner Rahmen für wertbasiertes Reinforcement Learning für kontinuierliche Kontrolle"]} +{"source": "Generative Adversarial Networks (GANs) have proven to be a powerful framework for learning to draw samples from complex distributions.However, GANs are also notoriously difficult to train, with mode collapse and oscillations a common problem.We hypothesize that this is at least in part due to the evolution of the generator distribution and the catastrophic forgetting tendency of neural networks, which leads to the discriminator losing the ability to remember synthesized samples from previous instantiations of the generator.Recognizing this, our contributions are twofold.First, we show that GAN training makes for a more interesting and realistic benchmark for continual learning methods evaluation than some of the more canonical datasets.Second, we propose leveraging continual learning techniques to augment the discriminator, preserving its ability to recognize previous generator samples.We show that the resulting methods add only a light amount of computation, involve minimal changes to the model, and result in better overall performance on the examined image and text generation tasks.", "target": ["Generative Adversarial Network Training ist ein kontinuierliches Lernproblem."]} +{"source": "Many problems with large-scale labeled training data have been impressively solved by deep learning.However, Unseen Class Categorization (UCC) with minimal information provided about target classes is the most commonly encountered setting in industry, which remains a challenging research problem in machine learning.Previous approaches to UCC either fail to generate a powerful discriminative feature extractor or fail to learn a flexible classifier that can be easily adapted to unseen classes.In this paper, we propose to address these issues through network reparameterization, \\textit{i.e.}, reparametrizing the learnable weights of a network as a function of other variables, by which we decouple the feature extraction part and the classification part of a deep classification model to suit the special setting of UCC, securing both strong discriminability and excellent adaptability.Extensive experiments for UCC on several widely-used benchmark datasets in the settings of zero-shot and few-shot learning demonstrate that, our method with network reparameterization achieves state-of-the-art performance.", "target": ["Ein vereinheitlichter Rahmen für Few-shot Learning und Zero-shot Learning auf der Grundlage einer Reparametrisierung des Netzes"]} +{"source": "Proteins are ubiquitous molecules whose function in biological processes is determined by their 3D structure.Experimental identification of a protein's structure can be time-consuming, prohibitively expensive, and not always possible. Alternatively, protein folding can be modeled using computational methods, which however are not guaranteed to always produce optimal results.GraphQA is a graph-based method to estimate the quality of protein models, that possesses favorable properties such as representation learning, explicit modeling of both sequential and 3D structure, geometric invariance and computational efficiency. In this work, we demonstrate significant improvements of the state-of-the-art for both hand-engineered and representation-learning approaches, as well as carefully evaluating the individual contributions of GraphQA.", "target": ["GraphQA ist eine graphenbasierte Methode zur Bewertung der Qualität von Proteinen, die den Stand der Technik sowohl bei handwerklichen als auch bei repräsentationsbasierten Ansätzen verbessert."]} +{"source": "We study the problem of training machine learning models incrementally using active learning with access to imperfect or noisy oracles.We specifically consider the setting of batch active learning, in which multiple samples are selected as opposed to a single sample as in classical settings so as to reduce the training overhead.Our approach bridges between uniform randomness and score based importance sampling of clusters when selecting a batch of new samples.Experiments onbenchmark image classification datasets (MNIST, SVHN, and CIFAR10) shows improvement over existing active learning strategies.We introduce an extra denoising layer to deep networks to make active learning robust to label noises and show significant improvements.", "target": ["Wir befassen uns mit aktivem Lernen in einer Batch-Umgebung mit verrauschten Orakeln und verwenden Modellunsicherheit, um die Entscheidungsqualität des aktiven Lernalgorithmus während der Erfassung zu kodieren."]} +{"source": "Artistic style transfer is the problem of synthesizing an image with content similar to a given image and style similar to another.Although recent feed-forward neural networks can generate stylized images in real-time, these models produce a single stylization given a pair of style/content images, and the user doesn't have control over the synthesized output.Moreover, the style transfer depends on the hyper-parameters of the model with varying ``optimum\" for different input images.Therefore, if the stylized output is not appealing to the user, she/he has to try multiple models or retrain one with different hyper-parameters to get a favorite stylization.In this paper, we address these issues by proposing a novel method which allows adjustment of crucial hyper-parameters, after the training and in real-time, through a set of manually adjustable parameters.These parameters enable the user to modify the synthesized outputs from the same pair of style/content images, in search of a favorite stylized image.Our quantitative and qualitative experiments indicate how adjusting these parameters is comparable to retraining the model with different hyper-parameters.We also demonstrate how these parameters can be randomized to generate results which are diverse but still very similar in style and content.", "target": ["Stochastische Übertragung mit einstellbaren Merkmalen."]} +{"source": "Recent work has shown that deep reinforcement-learning agents can learn to follow language-like instructions from infrequent environment rewards.However, this places on environment designers the onus of designing language-conditional reward functions which may not be easily or tractably implemented as the complexity of the environment and the language scales.To overcome this limitation, we present a framework within which instruction-conditional RL agents are trained using rewards obtained not from the environment, but from reward models which are jointly trained from expert examples. As reward models improve, they learn to accurately reward agents for completing tasks for environment configurations---and for instructions---not present amongst the expert data.This framework effectively separates the representation of what instructions require from how they can be executed.In a simple grid world, it enables an agent to learn a range of commands requiring interaction with blocks and understanding of spatial relations and underspecified abstract arrangements.We further show the method allows our agent to adapt to changes in the environment without requiring new expert examples.", "target": ["Wir schlagen AGILE vor, ein Framework für das Training von Bearbeitern zur Ausführung von Anweisungen anhand von Beispielen für entsprechende Zielzustände."]} +{"source": "We present Multitask Soft Option Learning (MSOL), a hierarchical multi-task framework based on Planning-as-Inference.MSOL extends the concept of Options, using separate variational posteriors for each task, regularized by a shared prior.The learned soft-options are temporally extended, allowing a higher-level master policy to train faster on new tasks by making decisions with lower frequency.Additionally, MSOL allows fine-tuning of soft-options for new tasks without unlearning previously useful behavior, and avoids problems with local minima in multitask training.We demonstrate empirically that MSOL significantly outperforms both hierarchical and flat transfer-learning baselines in challenging multi-task environments.", "target": ["In Hierarchical RL führen wir den Begriff einer \"weichen\", d.h. anpassungsfähigen Option ein und zeigen, dass dies das Lernen in Multitasking-Umgebungen erleichtert."]} +{"source": "We propose an algorithm, guided variational autoencoder (Guided-VAE), that is able to learn a controllable generative model by performing latent representation disentanglement learning.The learning objective is achieved by providing signal to the latent encoding/embedding in VAE without changing its main backbone architecture, hence retaining the desirable properties of the VAE.We design an unsupervised and a supervised strategy in Guided-VAE and observe enhanced modeling and controlling capability over the vanilla VAE.In the unsupervised strategy, we guide the VAE learning by introducing a lightweight decoder that learns latent geometric transformation and principal components; in the supervised strategy, we use an adversarial excitation and inhibition mechanism to encourage the disentanglement of the latent variables.Guided-VAE enjoys its transparency and simplicity for the general representation learning task, as well as disentanglement learning.On a number of experiments for representation learning, improved synthesis/sampling, better disentanglement for classification, and reduced classification errors in meta learning have been observed.", "target": ["Erlernen eines kontrollierbaren generativen Modells durch Entflechtungslernen der latenten Repräsentation."]} +{"source": "Neural language models (NLMs) are generative, and they model the distribution of grammatical sentences.Trained on huge corpus, NLMs are pushing the limit of modeling accuracy.Besides, they have also been applied to supervised learning tasks that decode text, e.g., automatic speech recognition (ASR).By re-scoring the n-best list, NLM can select grammatically more correct candidate among the list, and significantly reduce word/char error rate.However, the generative nature of NLM may not guarantee a discrimination between “good” and “bad” (in a task-specific sense) sentences, resulting in suboptimal performance.This work proposes an approach to adapt a generative NLM to a discriminative one.Different from the commonly used maximum likelihood objective, the proposed method aims at enlarging the margin between the “good” and “bad” sentences.It is trained end-to-end and can be widely applied to tasks that involve the re-scoring of the decoded text.Significant gains are observed in both ASR and statistical machine translation (SMT) tasks.", "target": ["Verbessern des Sprachmodells für überwachte Lernaufgaben"]} +{"source": "Conventionally, convolutional neural networks (CNNs) process different images with the same set of filters.However, the variations in images pose a challenge to this fashion.In this paper, we propose to generate sample-specific filters for convolutional layers in the forward pass.Since the filters are generated on-the-fly, the model becomes more flexible and can better fit the training data compared to traditional CNNs.In order to obtain sample-specific features, we extract the intermediate feature maps from an autoencoder.As filters are usually high dimensional, we propose to learn a set of coefficients instead of a set of filters.These coefficients are used to linearly combine the base filters from a filter repository to generate the final filters for a CNN.The proposed method is evaluated on MNIST, MTFL and CIFAR10 datasets.Experiment results demonstrate that the classification accuracy of the baseline model can be improved by using the proposed filter generation method.", "target": ["Dynamische Erzeugung von Filtern, die vom Eingangsbild abhängig sind, für CNNs in jedem Vorwärtsdurchlauf"]} +{"source": "We propose a new anytime neural network which allows partial evaluation by subnetworks with different widths as well as depths.Compared to conventional anytime networks only with the depth controllability, the increased architectural diversity leads to higher resource utilization and consequent performance improvement under various and dynamic resource budgets.We highlight architectural features to make our scheme feasible as well as efficient, and show its effectiveness in image classification tasks.", "target": ["Wir schlagen ein neues, jederzeit einsetzbares neuronales Netz vor, das eine Teilauswertung durch Teilnetze mit unterschiedlicher Breite und Tiefe ermöglicht."]} +{"source": "We propose a new model for making generalizable and diverse retrosynthetic reaction predictions.Given a target compound, the task is to predict the likely chemical reactants to produce the target.This generative task can be framed as a sequence-to-sequence problem by using the SMILES representations of the molecules.Building on top of the popular Transformer architecture, we propose two novel pre-training methods that construct relevant auxiliary tasks (plausible reactions) for our problem.Furthermore, we incorporate a discrete latent variable model into the architecture to encourage the model to produce a diverse set of alternative predictions.On the 50k subset of reaction examples from the United States patent literature (USPTO-50k) benchmark dataset, our model greatly improves performance over the baseline, while also generating predictions that are more diverse.", "target": ["Wir schlagen ein neues Modell vor, um verallgemeinerbare und vielfältige retrosynthetische Reaktionsvorhersagen zu machen."]} +{"source": "Unsupervised learning of disentangled representations is an open problem in machine learning.The Disentanglement-PyTorch library is developed to facilitate research, implementation, and testing of new variational algorithms.In this modular library, neural architectures, dimensionality of the latent space, and the training algorithms are fully decoupled, allowing for independent and consistent experiments across variational methods.The library handles the training scheduling, logging, and visualizations of reconstructions and latent space traversals.It also evaluates the encodings based on various disentanglement metrics.The library, so far, includes implementations of the following unsupervised algorithms VAE, Beta-VAE, Factor-VAE, DIP-I-VAE, DIP-II-VAE, Info-VAE, and Beta-TCVAE, as well as conditional approaches such as CVAE and IFCVAE.The library is compatible with the Disentanglement Challenge of NeurIPS 2019, hosted on AICrowd and was used to compete in the first and second stages of the challenge, where it was ranked among the best few participants.", "target": ["Disentanglement-PyTorch ist eine Bibliothek für das Lernen variabler Darstellungen"]} +{"source": "Trust region methods, such as TRPO, are often used to stabilize policy optimization algorithms in reinforcement learning (RL).While current trust region strategies are effective for continuous control, they typically require a large amount of on-policy interaction with the environment.To address this problem, we propose an off-policy trust region method, Trust-PCL, which exploits an observation that the optimal policy and state values of a maximum reward objective with a relative-entropy regularizer satisfy a set of multi-step pathwise consistencies along any path.The introduction of relative entropy regularization allows Trust-PCL to maintain optimization stability while exploiting off-policy data to improve sample efficiency.When evaluated on a number of continuous control tasks, Trust-PCL significantly improves the solution quality and sample efficiency of TRPO.", "target": ["Wir erweitern die jüngsten Erkenntnisse im Zusammenhang mit der Softmax-Beständigkeit, um modernste Ergebnisse in der kontinuierlichen Steuerung zu erzielen."]} +{"source": "Deep reinforcement learning has achieved many recent successes, but our understanding of its strengths and limitations is hampered by the lack of rich environments in which we can fully characterize optimal behavior, and correspondingly diagnose individual actions against such a characterization. Here we consider a family of combinatorial games, arising from work of Erdos, Selfridge, and Spencer, and we propose their use as environments for evaluating and comparing different approaches to reinforcement learning.These games have a number of appealing features: they are challenging for current learning approaches, but they form(i) a low-dimensional, simply parametrized environment where(ii) there is a linear closed form solution for optimal behavior from any state, and(iii) the difficulty of the game can be tuned by changing environment parameters in an interpretable way.We use these Erdos-Selfridge-Spencer games not only to compare different algorithms, but also to compare approaches based on supervised and reinforcement learning, to analyze the power of multi-agent approaches in improving performance, and to evaluate generalization to environments outside the training set.", "target": ["Wir passen eine Familie von kombinatorischen Spielen mit abstimmbarer Schwierigkeit und einem optimalen Regelwerk an, die als lineares Netzwerk ausgedrückt werden kann. Wir entwickeln sie als eine reichhaltige Umgebung für das Reinforcement Learning, zeigen Kontraste in der Leistung mit dem überwachten Lernen und analysieren das Multiagentenlernen und die Verallgemeinerung."]} +{"source": "Adoption of deep learning in safety-critical systems raise the need for understanding what deep neural networks do not understand.Several methodologies to estimate model uncertainty have been proposed, but these methodologies constrain either how the neural network is trained or constructed.We present Outlier Detection In Neural networks (ODIN), an assumption-free method for detecting outlier observations during prediction, based on principles widely used in manufacturing process monitoring.By using a linear approximation of the hidden layer manifold, we add prediction-time outlier detection to models after training without altering architecture or training.We demonstrate that ODIN efficiently detect outliers during prediction on Fashion-MNIST, ImageNet-synsets and speech command recognition.", "target": ["Ein zusätzliches Verfahren für Deep Learning zur Erkennung von Ausreißern während der Vorhersagezeit"]} +{"source": "This work introduces a simple network for producing character aware word embeddings.Position agnostic and position aware character embeddings are combined to produce an embedding vector for each word.The learned word representations are shown to be very sparse and facilitate improved results on language modeling tasks, despite using markedly fewer parameters, and without the need to apply dropout.A final experiment suggests that weight sharing contributes to sparsity, increases performance, and prevents overfitting.", "target": ["Eine vollständig vernetzte Architektur wird verwendet, um Worteinbettungen aus Zeichendarstellungen zu erzeugen, übertrifft traditionelle Einbettungen und bietet Einblicke in Spärlichkeit und Dropout."]} +{"source": "Neural networks with low-precision weights and activations offer compellingefficiency advantages over their full-precision equivalents.The two mostfrequently discussed benefits of quantization are reduced memory consumption,and a faster forward pass when implemented with efficient bitwiseoperations.We propose a third benefit of very low-precision neural networks:improved robustness against some adversarial attacks, and in the worst case,performance that is on par with full-precision models.We focus on the verylow-precision case where weights and activations are both quantized to $\\pm$1,and note that stochastically quantizing weights in just one layer can sharplyreduce the impact of iterative attacks.We observe that non-scaled binary neuralnetworks exhibit a similar effect to the original \\emph{defensive distillation}procedure that led to \\emph{gradient masking}, and a false notion of security.We address this by conducting both black-box and white-box experiments withbinary models that do not artificially mask gradients.", "target": ["Wir führen gegnerische Angriffe gegen binarisierte neuronale Netze durch und zeigen, dass wir die Auswirkungen der stärksten Angriffe reduzieren, während wir eine vergleichbare Genauigkeit in einer Blackbox-Umgebung beibehalten."]} +{"source": "Unsupervised bilingual dictionary induction (UBDI) is useful for unsupervised machine translation and for cross-lingual transfer of models into low-resource languages.One approach to UBDI is to align word vector spaces in different languages using Generative adversarial networks (GANs) with linear generators, achieving state-of-the-art performance for several language pairs.For some pairs, however, GAN-based induction is unstable or completely fails to align the vector spaces.We focus on cases where linear transformations provably exist, but the performance of GAN-based UBDI depends heavily on the model initialization.We show that the instability depends on the shape and density of the vector sets, but not on noise; it is the result of local optima, but neither over-parameterization nor changing the batch size or the learning rate consistently reduces instability.Nevertheless, we can stabilize GAN-based UBDI through best-of-N model selection, based on an unsupervised stopping criterion.", "target": ["Eine empirische Untersuchung des GAN-basierten Alignments von Wortvektorräumen mit Schwerpunkt auf Fällen, in denen lineare Transformationen nachweislich existieren, das Training aber instabil ist."]} +{"source": "Owing to the ubiquity of computer software, software vulnerability detection (SVD) has become an important problem in the software industry and in the field of computer security.One of the most crucial issues in SVD is coping with the scarcity of labeled vulnerabilities in projects that require the laborious manual labeling of code by software security experts.One possible way to address is to employ deep domain adaptation which has recently witnessed enormous success in transferring learning from structural labeled to unlabeled data sources.The general idea is to map both source and target data into a joint feature space and close the discrepancy gap of those data in this joint feature space.Generative adversarial network (GAN) is a technique that attempts to bridge the discrepancy gap and also emerges as a building block to develop deep domain adaptation approaches with state-of-the-art performance.However, deep domain adaptation approaches using the GAN principle to close the discrepancy gap are subject to the mode collapsing problem that negatively impacts the predictive performance.Our aim in this paper is to propose Dual Generator-Discriminator Deep Code Domain Adaptation Network (Dual-GD-DDAN) for tackling the problem of transfer learning from labeled to unlabeled software projects in the context of SVD in order to resolve the mode collapsing problem faced in previous approaches.The experimental results on real-world software projects show that our proposed method outperforms state-of-the-art baselines by a wide margin.", "target": ["Unser Ziel in diesem Papier ist es, einen neuen Ansatz für das Problem des Transfer-Lernens von gelabelten auf nicht gelabelte Software-Projekten im Kontext von SVD vorzuschlagen, um das Problem des Modus-Kollabierens zu lösen, mit dem bisherige Ansätze konfrontiert sind."]} +{"source": "Representation learning is one of the foundations of Deep Learning and allowed important improvements on several Machine Learning tasks, such as Neural Machine Translation, Question Answering and Speech Recognition.Recent works have proposed new methods for learning representations for nodes and edges in graphs.Several of these methods are based on the SkipGram algorithm, and they usually process a large number of multi-hop neighbors in order to produce the context from which node representations are learned.In this paper, we propose an effective and also efficient method for generating node embeddings in graphs that employs a restricted number of permutations over the immediate neighborhood of a node as context to generate its representation, thus ego-centric representations.We present a thorough evaluation showing that our method outperforms state-of-the-art methods in six different datasets related to the problems of link prediction and node classification, being one to three orders of magnitude faster than baselines when generating node embeddings for very large graphs.", "target": ["Eine schnellere Methode zur Erzeugung von Knoteneinbettungen, die eine Reihe von Permutationen über die unmittelbare Nachbarschaft eines Knotens als Kontext verwendet, um seine Darstellung zu erzeugen."]} +{"source": "Orthogonal recurrent neural networks address the vanishing gradient problem by parameterizing the recurrent connections using an orthogonal matrix.This class of models is particularly effective to solve tasks that require the memorization of long sequences.We propose an alternative solution based on explicit memorization using linear autoencoders for sequences.We show how a recently proposed recurrent architecture, the Linear Memory Network, composed of a nonlinear feedforward layer and a separate linear recurrence, can be used to solve hard memorization tasks.We propose an initialization schema that sets the weights of a recurrent architecture to approximate a linear autoencoder of the input sequences, which can be found with a closed-form solution.The initialization schema can be easily adapted to any recurrent architecture. We argue that this approach is superior to a random orthogonal initialization due to the autoencoder, which allows the memorization of long sequences even before training.The empirical analysis show that our approach achieves competitive results against alternative orthogonal models, and the LSTM, on sequential MNIST, permuted MNIST and TIMIT.", "target": ["Wir zeigen, wie man rekurrente Architekturen mit der geschlossenen Formlösung eines linearen Autoencoders für Sequenzen initialisiert. Wir zeigen die Vorteile dieses Ansatzes im Vergleich zu orthogonalen RNNs."]} +{"source": "This paper improves upon the line of research that formulates named entity recognition (NER) as a sequence-labeling problem.We use so-called black-box long short-term memory (LSTM) encoders to achieve state-of-the-art results while providing insightful understanding of what the auto-regressive model learns with a parallel self-attention mechanism.Specifically, we decouple the sequence-labeling problem of NER into entity chunking, e.g., Barack_B Obama_E was_O elected_O, and entity typing, e.g., Barack_PERSON Obama_PERSON was_NONE elected_NONE, and analyze how the model learns to, or has difficulties in, capturing text patterns for each of the subtasks.The insights we gain then lead us to explore a more sophisticated deep cross-Bi-LSTM encoder, which proves better at capturing global interactions given both empirical results and a theoretical justification.", "target": ["Wir bieten ein aufschlussreiches Verständnis der Sequenzkennzeichnung von NER und schlagen die Verwendung von zwei Arten von Kreuzstrukturen vor, die beide theoretische und empirische Verbesserungen bringen."]} +{"source": "Knowledge Graph Embedding (KGE) is the task of jointly learning entity and relation embeddings for a given knowledge graph.Existing methods for learning KGEs can be seen as a two-stage process where(a) entities and relations in the knowledge graph are represented using some linear algebraic structures (embeddings), and(b) a scoring function is defined that evaluates the strength of a relation that holds between two entities using the corresponding relation and entity embeddings.Unfortunately, prior proposals for the scoring functions in the first step have been heuristically motivated, and it is unclear as to how the scoring functions in KGEs relate to the generation process of the underlying knowledge graph.To address this issue, we propose a generative account of the KGE learning task.Specifically, given a knowledge graph represented by a set of relational triples (h, R, t), where the semantic relation R holds between the two entities h (head) and t (tail), we extend the random walk model (Arora et al., 2016a) of word embeddings to KGE.We derive a theoretical relationship between the joint probability p(h, R, t) and the embeddings of h, R and t.Moreover, we show that marginal loss minimisation, a popular objective used by much prior work in KGE, follows naturally from the log-likelihood ratio maximisation under the probabilities estimated from the KGEs according to our theoretical relationship.We propose a learning objective motivated by the theoretical analysis to learn KGEs from a given knowledge graph.The KGEs learnt by our proposed method obtain state-of-the-art performance on FB15K237 and WN18RR benchmark datasets, providing empirical evidence in support of the theory.", "target": ["Wir stellen ein theoretisch bewährtes generatives Modell der Wissensgrapheneinbettung vor."]} +{"source": "Currently the only techniques for sharing governance of a deep learning model are homomorphic encryption and secure multiparty computation.Unfortunately, neither of these techniques is applicable to the training of large neural networks due to their large computational and communication overheads.As a scalable technique for shared model governance, we propose splitting deep learning model between multiple parties.This paper empirically investigates the security guarantee of this technique, which is introduced as the problem of model completion: Given the entire training data set or an environment simulator, and a subset of the parameters of a trained deep learning model, how much training is required to recover the model’s original performance? We define a metric for evaluating the hardness of the model completion problem and study it empirically in both supervised learning on ImageNet and reinforcement learning on Atari and DeepMind Lab.Our experiments show that (1) the model completion problem is harder in reinforcement learning than in supervised learning because of the unavailability of the trained agent’s trajectories, and (2) its hardness depends not primarily on the number of parameters of the missing part, but more so on their type and location. Our results suggest that model splitting might be a feasible technique for shared model governance in some settings where training is very expensive.", "target": ["Wir untersuchen empirisch, wie schwierig es ist, fehlende Teile von trainierten Modellen wiederherzustellen"]} +{"source": "This paper proposes variational domain adaptation, a unified, scalable, simple framework for learning multiple distributions through variational inference.Unlike the existing methods on domain transfer through deep generative models, such as StarGAN (Choi et al., 2017) and UFDN (Liu et al., 2018), the variational domain adaptation has three advantages.Firstly, the samples from the target are not required.Instead, the framework requires one known source as a prior $p(x)$ and binary discriminators, $p(\\mathcal{D}_i|x)$, discriminating the target domain $\\mathcal{D}_i$ from others.Consequently, the framework regards a target as a posterior that can be explicitly formulated through the Bayesian inference, $p(x|\\mathcal{D}_i) \\propto p(\\mathcal{D}_i|x)p(x)$, as exhibited by a further proposed model of dual variational autoencoder (DualVAE).Secondly, the framework is scablable to large-scale domains.As well as VAE encodes a sample $x$ as a mode on a latent space: $\\mu(x) \\in \\mathcal{Z}$, DualVAE encodes a domain $\\mathcal{D}_i$ as a mode on the dual latent space $\\mu^*(\\mathcal{D}_i) \\in \\mathcal{Z}^*$, named domain embedding.It reformulates the posterior with a natural paring $\\langle, \\rangle: \\mathcal{Z} \\times \\mathcal{Z}^* \\rightarrow \\Real$, which can be expanded to uncountable infinite domains such as continuous domains as well as interpolation.Thirdly, DualVAE fastly converges without sophisticated automatic/manual hyperparameter search in comparison to GANs as it requires only one additional parameter to VAE.Through the numerical experiment, we demonstrate the three benefits with multi-domain image generation task on CelebA with up to 60 domains, and exhibits that DualVAE records the state-of-the-art performance outperforming StarGAN and UFDN.", "target": ["In diesem Papier wird die variable Domänenanpassung vorgeschlagen, ein einheitliches, skalierbares und einfaches Rahmenwerk für das Lernen mehrerer Verteilungen durch Variationsinferenz."]} +{"source": "We propose a new method to train neural networks based on a novel combination of adversarial training and provable defenses.The key idea is to model training as a procedure which includes both, the verifier and the adversary.In every iteration, the verifier aims to certify the network using convex relaxation while the adversary tries to find inputs inside that convex relaxation which cause verification to fail.We experimentally show that this training method is promising and achieves the best of both worlds – it produces a model with state-of-the-art accuracy (74.8%) and certified robustness (55.9%) on the challenging CIFAR-10 dataset with a 2/255 L-infinity perturbation.This is a significant improvement over the currently known best results of 68.3% accuracy and 53.9% certified robustness, achieved using a 5 times larger network than our work.", "target": ["Wir schlagen eine neuartige Kombination aus gegnerischem Training und beweisbaren Verteidigungsmaßnahmen vor, die ein Modell mit modernster Genauigkeit und zertifizierter Robustheit auf CIFAR-10 hervorbringt."]} +{"source": "Learning tasks on source code (i.e., formal languages) have been considered recently, but most work has tried to transfer natural language methods and does not capitalize on the unique opportunities offered by code's known syntax.For example, long-range dependencies induced by using the same variable or function in distant locations are often not considered.We propose to use graphs to represent both the syntactic and semantic structure of code and use graph-based deep learning methods to learn to reason over program structures.In this work, we present how to construct graphs from source code and how to scale Gated Graph Neural Networks training to such large graphs.We evaluate our method on two tasks: VarNaming, in which a network attempts to predict the name of a variable given its usage, and VarMisuse, in which the network learns to reason about selecting the correct variable that should be used at a given program location.Our comparison to methods that use less structured program representations shows the advantages of modeling known structure, and suggests that our models learn to infer meaningful names and to solve the VarMisuse task in many cases.Additionally, our testing showed that VarMisuse identifies a number of bugs in mature open-source projects.", "target": ["Programme haben eine Struktur, die als Graphen dargestellt werden kann, und neuronale Graphen-Netzwerke können lernen, Fehler in solchen Graphen zu finden."]} +{"source": "Overfitting is an ubiquitous problem in neural network training and usually mitigated using a holdout data set.Here we challenge this rationale and investigate criteria for overfitting without using a holdout data set.Specifically, we train a model for a fixed number of epochs multiple times with varying fractions of randomized labels and for a range of regularization strengths. A properly trained model should not be able to attain an accuracy greater than the fraction of properly labeled data points.Otherwise the model overfits. We introduce two criteria for detecting overfitting and one to detect underfitting.We analyze early stopping, the regularization factor, and network depth.In safety critical applications we are interested in models and parameter settings which perform well and are not likely to overfit.The methods of this paper allow characterizing and identifying such models.", "target": ["Wir führen mehrere Kriterien zur Erkennung von Überanpassung ein und analysieren sie."]} +{"source": "Partially observable Markov decision processes (POMDPs) are a widely-used framework to model decision-making with uncertainty about the environment and under stochastic outcome.In conventional POMDP models, the observations that the agent receives originate from fixed known distribution.However, in a variety of real-world scenarios the agent has an active role in its perception by selecting which observations to receive.Due to combinatorial nature of such selection process, it is computationally intractable to integrate the perception decision with the planning decision.To prevent such expansion of the action space, we propose a greedy strategy for observation selection that aims to minimize the uncertainty in state. We develop a novel point-based value iteration algorithm that incorporates the greedy strategy to achieve near-optimal uncertainty reduction for sampled belief points.This in turn enables the solver to efficiently approximate the reachable subspace of belief simplex by essentially separating computations related to perception from planning.Lastly, we implement the proposed solver and demonstrate its performance and computational advantage in a range of robotic scenarios where the robot simultaneously performs active perception and planning.", "target": ["Wir entwickeln einen punktbasierten Wert-Iterationslöser für POMDPs mit aktiven Wahrnehmungs- und Planungsaufgaben."]} +{"source": "Deep neural networks, in particular convolutional neural networks, have become highly effective tools for compressing images and solving inverse problems including denoising, inpainting, and reconstruction from few and noisy measurements.This success can be attributed in part to their ability to represent and generate natural images well.Contrary to classical tools such as wavelets, image-generating deep neural networks have a large number of parameters---typically a multiple of their output dimension---and need to be trained on large datasets. In this paper, we propose an untrained simple image model, called the deep decoder, which is a deep neural network that can generate natural images from very few weight parameters.The deep decoder has a simple architecture with no convolutions and fewer weight parameters than the output dimensionality.This underparameterization enables the deep decoder to compress images into a concise set of network weights, which we show is on par with wavelet-based thresholding.Further, underparameterization provides a barrier to overfitting, allowing the deep decoder to have state-of-the-art performance for denoising.The deep decoder is simple in the sense that each layer has an identical structure that consists of only one upsampling unit, pixel-wise linear combination of channels, ReLU activation, and channelwise normalization.This simplicity makes the network amenable to theoretical analysis, and it sheds light on the aspects of neural networks that enable them to form effective signal representations.", "target": ["Wir stellen ein unterparametrisiertes, nicht convolutional, und einfaches tiefes neuronales Netzwerk vor, das ohne Training natürliche Bilder effektiv darstellen und Bildverarbeitungsaufgaben wie Kompression und Entrauschung konkurrenzfähig lösen kann."]} +{"source": "In this paper we investigate the family of functions representable by deep neural networks (DNN) with rectified linear units (ReLU).We give an algorithm to train a ReLU DNN with one hidden layer to {\\em global optimality} with runtime polynomial in the data size albeit exponential in the input dimension.Further, we improve on the known lower bounds on size (from exponential to super exponential) for approximating a ReLU deep net function by a shallower ReLU net.Our gap theorems hold for smoothly parametrized families of ``hard'' functions, contrary to countable, discrete families known in the literature. An example consequence of our gap theorems is the following: for every natural number $k$ there exists a function representable by a ReLU DNN with $k^2$ hidden layers and total size $k^3$, such that any ReLU DNN with at most $k$ hidden layers will require at least $\\frac12k^{k+1}-1$ total nodes.Finally, for the family of $\\R^n\\to \\R$ DNNs with ReLU activations, we show a new lowerbound on the number of affine pieces, which is larger than previous constructions in certain regimes of the network architecture and most distinctively our lowerbound is demonstrated by an explicit construction of a \\emph{smoothly parameterized} family of functions attaining this scaling.Our construction utilizes the theory of zonotopes from polyhedral theory.", "target": ["Dieses Papier 1) charakterisiert Funktionen, die durch ReLU-DNNs dargestellt werden können, 2) untersucht formal den Nutzen der Tiefe in solchen Architekturen, 3) gibt einen Algorithmus zur Implementierung empirischer Risikominimierung zur globalen Optimalität für zweischichtige ReLU-Netze."]} +{"source": "The backpropagation of error algorithm (BP) is often said to be impossible to implement in a real brain.The recent success of deep networks in machine learning and AI, however, has inspired a number of proposals for understanding how the brain might learn across multiple layers, and hence how it might implement or approximate BP.As of yet, none of these proposals have been rigorously evaluated on tasks where BP-guided deep learning has proved critical, or in architectures more structured than simple fully-connected networks.Here we present the first results on scaling up a biologically motivated model of deep learning to datasets which need deep networks with appropriate architectures to achieve good performance.We present results on CIFAR-10 and ImageNet. For CIFAR-10 we show that our algorithm, a straightforward, weight-transport-free variant of difference target-propagation (DTP) modified to remove backpropagation from the penultimate layer, is competitive with BP in training deep networks with locally defined receptive fields that have untied weights. For ImageNet we find that both DTP and our algorithm perform significantly worse than BP, opening questions about whether different architectures or algorithms are required to scale these approaches.Our results and implementation details help establish baselines for biologically motivated deep learning schemes going forward.", "target": ["Benchmarks für biologisch plausible Lernalgorithmen auf komplexen Datensätzen und Architekturen"]} +{"source": "Deep neural networks (DNNs) usually contain millions, maybe billions, of parameters/weights, making both storage and computation very expensive.This has motivated a large body of work to reduce the complexity of the neural network by using sparsity-inducing regularizers. Another well-known approach for controlling the complexity of DNNs is parameter sharing/tying, where certain sets of weights are forced to share a common value.Some forms of weight sharing are hard-wired to express certain in- variances, with a notable example being the shift-invariance of convolutional layers.However, there may be other groups of weights that may be tied together during the learning process, thus further re- ducing the complexity of the network.In this paper, we adopt a recently proposed sparsity-inducing regularizer, named GrOWL (group ordered weighted l1), which encourages sparsity and, simulta- neously, learns which groups of parameters should share a common value.GrOWL has been proven effective in linear regression, being able to identify and cope with strongly correlated covariates.Unlike standard sparsity-inducing regularizers (e.g., l1 a.k.a.Lasso), GrOWL not only eliminates unimportant neurons by setting all the corresponding weights to zero, but also explicitly identifies strongly correlated neurons by tying the corresponding weights to a common value.This ability of GrOWL motivates the following two-stage procedure: (i) use GrOWL regularization in the training process to simultaneously identify significant neurons and groups of parameter that should be tied together; (ii) retrain the network, enforcing the structure that was unveiled in the previous phase, i.e., keeping only the significant neurons and enforcing the learned tying structure.We evaluate the proposed approach on several benchmark datasets, showing that it can dramatically compress the network with slight or even no loss on generalization performance.", "target": ["Wir haben vorgeschlagen, den neuen GrOWL-Regulierer für die gleichzeitige Spärlichkeit und Bindung von Parametern beim DNN-Lernen zu verwenden."]} +{"source": "Weight-sharing—the simultaneous optimization of multiple neural networks using the same parameters—has emerged as a key component of state-of-the-art neural architecture search.However, its success is poorly understood and often found to be surprising.We argue that, rather than just being an optimization trick, the weight-sharing approach is induced by the relaxation of a structured hypothesis space, and introduces new algorithmic and theoretical challenges as well as applications beyond neural architecture search.Algorithmically, we show how the geometry of ERM for weight-sharing requires greater care when designing gradient- based minimization methods and apply tools from non-convex non-Euclidean optimization to give general-purpose algorithms that adapt to the underlying structure.We further analyze the learning-theoretic behavior of the bilevel optimization solved by practical weight-sharing methods.Next, using kernel configuration and NLP feature selection as case studies, we demonstrate how weight-sharing applies to the architecture search generalization of NAS and effectively optimizes the resulting bilevel objective.Finally, we use our optimization analysis to develop a simple exponentiated gradient method for NAS that aligns with the underlying optimization geometry and matches state-of-the-art approaches on CIFAR-10.", "target": ["Eine Analyse der Lern- und Optimierungsstrukturen der Architektursuche in neuronalen Netzen und darüber hinaus."]} +{"source": "Deep latent variable models have seen recent success in many data domains.Lossless compression is an application of these models which, despite having the potential to be highly useful, has yet to be implemented in a practical manner.We present '`Bits Back with ANS' (BB-ANS), a scheme to perform lossless compression with latent variable models at a near optimal rate.We demonstrate this scheme by using it to compress the MNIST dataset with a variational auto-encoder model (VAE), achieving compression rates superior to standard methods with only a simple VAE.Given that the scheme is highly amenable to parallelization, we conclude that with a sufficiently high quality generative model this scheme could be used to achieve substantial improvements in compression rate with acceptable running time.We make our implementation available open source at https://github.com/bits-back/bits-back .", "target": ["Wir führen eine verlustfreie Komprimierung großer Bilddatensätze mit einer VAE durch und übertreffen damit bestehende Kompressionsalgorithmen."]} +{"source": "Hyperparameter tuning is arguably the most important ingredient for obtaining state of art performance in deep networks. We focus on hyperparameters that are related to the optimization algorithm, e.g. learning rates, which have a large impact on the training speed and the resulting accuracy.Typically, fixed learning rate schedules are employed during training.We propose Hyperdyn a dynamic hyperparameter optimization method that selects new learning rates on the fly at the end of each epoch.Our explore-exploit framework combines Bayesian optimization (BO) with a rejection strategy, based on a simple probabilistic wait and watch test. We obtain state of art accuracy results on CIFAR and Imagenet datasets, but with significantly faster training, when compared with the best manually tuned networks.", "target": ["Auf Bayes'scher Optimierung basierende Online-Hyperparameter-Optimierung."]} +{"source": "Multi-hop text-based question-answering is a current challenge in machine comprehension. This task requires to sequentially integrate facts from multiple passages to answer complex natural language questions.In this paper, we propose a novel architecture, called the Latent Question Reformulation Network (LQR-net), a multi-hop and parallel attentive network designed for question-answering tasks that require reasoning capabilities.LQR-net is composed of an association of \\textbf{reading modules} and \\textbf{reformulation modules}.The purpose of the reading module is to produce a question-aware representation of the document.From this document representation, the reformulation module extracts essential elements to calculate an updated representation of the question.This updated question is then passed to the following hop.We evaluate our architecture on the \\hotpotqa question-answering dataset designed to assess multi-hop reasoning capabilities.Our model achieves competitive results on the public leaderboard and outperforms the best current \\textit{published} models in terms of Exact Match (EM) and $F_1$ score.Finally, we show that an analysis of the sequential reformulations can provide interpretable reasoning paths.", "target": ["In diesem Papier schlagen wir das Latent Question Reformulation Network (LQR-net) vor, ein Multi-Hop- und paralleles, aufmerksames Netzwerk, das für die Beantwortung von Fragen entwickelt wurde, die logische Fähigkeiten erfordern."]} +{"source": "We propose a method to automatically compute the importance of features at every observation in time series, by simulating counterfactual trajectories given previous observations.We define the importance of each observation as the change in the model output caused by replacing the observation with a generated one.Our method can be applied to arbitrarily complex time series models.We compare the generated feature importance to existing methods like sensitivity analyses, feature occlusion, and other explanation baselines to show that our approach generates more precise explanations and is less sensitive to noise in the input signals.", "target": ["Erklärung von multivariaten Zeitreihenmodellen durch Auffinden wichtiger Beobachtungen in der Zeit mit Hilfe von kontrafaktischen Daten"]} +{"source": "This paper addresses unsupervised domain adaptation, the setting where labeled training data is available on a source domain, but the goal is to have good performance on a target domain with only unlabeled data.Like much of previous work, we seek to align the learned representations of the source and target domains while preserving discriminability.The way we accomplish alignment is by learning to perform auxiliary self-supervised task(s) on both domains simultaneously. Each self-supervised task brings the two domains closer together along the direction relevant to that task.Training this jointly with the main task classifier on the source domain is shown to successfully generalize to the unlabeled target domain. The presented objective is straightforward to implement and easy to optimize.We achieve state-of-the-art results on four out of seven standard benchmarks, and competitive results on segmentation adaptation.We also demonstrate that our method composes well with another popular pixel-level adaptation method.", "target": ["Wir verwenden Selbstüberwachung für beide Domänen, um sie für die unbeaufsichtigte Domänenanpassung abzugleichen."]} +{"source": "We introduce simple, efficient algorithms for computing a MinHash of a probability distribution, suitable for both sparse and dense data, with equivalent running times to the state of the art for both cases.The collision probability of these algorithms is a new measure of the similarity of positive vectors which we investigate in detail.We describe the sense in which this collision probability is optimal for any Locality Sensitive Hash based on sampling.We argue that this similarity measure is more useful for probability distributions than the similarity pursued by other algorithms for weighted MinHash, and is the natural generalization of the Jaccard index.", "target": ["Das Minimum einer Menge exponentiell verteilter Hashes hat eine sehr nützliche Kollisionswahrscheinlichkeit, die den Jaccard-Index auf Wahrscheinlichkeitsverteilungen verallgemeinert."]} +{"source": "Recently, progress has been made towards improving relational reasoning in machine learning field.Among existing models, graph neural networks (GNNs) is one of the most effective approaches for multi-hop relational reasoning.In fact, multi-hop relational reasoning is indispensable in many natural language processing tasks such as relation extraction.In this paper, we propose to generate the parameters of graph neural networks (GP-GNNs) according to natural language sentences, which enables GNNs to process relational reasoning on unstructured text inputs.We verify GP-GNNs in relation extraction from text.Experimental results on a human-annotated dataset and two distantly supervised datasets show that our model achieves significant improvements compared to the baselines.We also perform a qualitative analysis to demonstrate that our model could discover more accurate relations by multi-hop relational reasoning.", "target": ["Ein graphisches neuronales Netzmodell mit Parametern, die aus natürlichen Sprachen generiert werden, das Multi-Hop-Reasoning durchführen kann."]} +{"source": "Off-Policy Actor-Critic (Off-PAC) methods have proven successful in a variety of continuous control tasks.Normally, the critic’s action-value function is updated using temporal-difference, and the critic in turn provides a loss for the actor that trains it to take actions with higher expected return.In this paper, we introduce a novel and flexible meta-critic that observes the learning process and meta-learns an additional loss for the actor that accelerates and improves actor-critic learning.Compared to the vanilla critic, the meta-critic network is explicitly trained to accelerate the learning process; and compared to existing meta-learning algorithms, meta-critic is rapidly learned online for a single task, rather than slowly over a family of tasks.Crucially, our meta-critic framework is designed for off-policy based learners, which currently provide state-of-the-art reinforcement learning sample efficiency.We demonstrate that online meta-critic learning leads to improvements in a variety of continuous control environments when combined with contemporary Off-PAC methods DDPG, TD3 and the state-of-the-art SAC.", "target": ["Wir stellen Meta-Critic vor, ein Hilfskritikmodul für Off-Policy-Actor-Critic-Methoden, das online während des Lernens einer einzelnen Aufgabe erlernt werden kann."]} +{"source": "Modern neural networks are highly overparameterized, with capacity to substantially overfit to training data.Nevertheless, these networks often generalize well in practice.It has also been observed that trained networks can often be ``compressed to much smaller representations.The purpose of this paper is to connect these two empirical observations.Our main technical result is a generalization bound for compressed networks based on the compressed size that, combined with off-the-shelf compression algorithms, leads to state-of-the-art generalization guarantees.In particular, we provide the first non-vacuous generalization guarantees for realistic architectures applied to the ImageNet classification problem.Additionally, we show that compressibility of models that tend to overfit is limited.Empirical results show that an increase in overfitting increases the number of bits required to describe a trained network.", "target": ["Durch die Kombination einer originalen PAC-Bayes-Schranke und einer handelsüblichen Komprimierungsmethode für neuronale Netze erhalten wir nicht-vakante Generalisierungsgrenzen für tiefe neuronale Netze auf ImageNet-Skala."]} +{"source": "Adversarial examples can be defined as inputs to a model which induce a mistake -- where the model output is different than that of an oracle, perhaps in surprising or malicious ways.Original models of adversarial attacks are primarily studied in the context of classification and computer vision tasks.While several attacks have been proposed in natural language processing (NLP) settings, they often vary in defining the parameters of an attack and what a successful attack would look like.The goal of this work is to propose a unifying model of adversarial examples suitable for NLP tasks in both generative and classification settings.We define the notion of adversarial gain: based in control theory, it is a measure of the change in the output of a system relative to the perturbation of the input (caused by the so-called adversary) presented to the learner.This definition, as we show, can be used under different feature spaces and distance conditions to determine attack or defense effectiveness across different intuitive manifolds.This notion of adversarial gain not only provides a useful way for evaluating adversaries and defenses, but can act as a building block for future work in robustness under adversaries due to its rooted nature in stability and manifold theory.", "target": ["Wir schlagen ein alternatives Maß zur Bestimmung der Effektivität von gegnerischen Angriffen in NLP-Modellen vor, das auf einem Abstandsmaß basiert, ähnlich dem inkrementellen L2-Gewinn in der Kontrolltheorie."]} +{"source": "We propose a Warped Residual Network (WarpNet) using a parallelizable warp operator for forward and backward propagation to distant layers that trains faster than the original residual neural network.We apply a perturbation theory on residual networks and decouple the interactions between residual units.The resulting warp operator is a first order approximation of the output over multiple layers.The first order perturbation theory exhibits properties such as binomial path lengths and exponential gradient scaling found experimentally by Veit et al (2016). We demonstrate through an extensive performance study that the proposed network achieves comparable predictive performance to the original residual network with the same number of parameters, while achieving a significant speed-up on the total training time.As WarpNet performs model parallelism in residual network training in which weights are distributed over different GPUs, it offers speed-up and capability to train larger networks compared to original residual networks.", "target": ["Wir schlagen das Warped Residual Network vor, das einen parallelisierbaren Warp-Operator für die Vorwärts- und Rückwärtspropagation zu entfernten Schichten verwendet, der schneller trainiert als das ursprüngliche Residual Neural Network."]} +{"source": "A plethora of methods attempting to explain predictions of black-box models have been proposed by the Explainable Artificial Intelligence (XAI) community.Yet, measuring the quality of the generated explanations is largely unexplored, making quantitative comparisons non-trivial.In this work, we propose a suite of multifaceted metrics that enables us to objectively compare explainers based on the correctness, consistency, as well as the confidence of the generated explanations.These metrics are computationally inexpensive, do not require model-retraining and can be used across different data modalities.We evaluate them on common explainers such as Grad-CAM, SmoothGrad, LIME and Integrated Gradients.Our experiments show that the proposed metrics reflect qualitative observations reported in earlier works.", "target": ["Wir schlagen eine Reihe von Metriken vor, die die gewünschten Eigenschaften von Erklärungsalgorithmen erfassen und verwenden sie, um solche Methoden objektiv zu vergleichen und zu bewerten"]} +{"source": "Neural networks are known to produce unexpected results on inputs that are far from the training distribution.One approach to tackle this problem is to detect the samples on which the trained network can not answer reliably.ODIN is a recently proposed method for out-of-distribution detection that does not modify the trained network and achieves good performance for various image classification tasks.In this paper we adapt ODIN for sentence classification and word tagging tasks.We show that the scores produced by ODIN can be used as a confidence measure for the predictions on both in-distribution and out-of-distribution datasets.", "target": ["Eine neue Methode zur Erkennung von Verteilungsfehlern hilft dabei, die Zuverlässigkeit von RNN-Vorhersagen für einige NLP-Aufgaben zu messen"]} +{"source": "Some recent work has shown separation between the expressive power of depth-2 and depth-3 neural networks.These separation results are shown by constructing functions and input distributions, so that the function is well-approximable by a depth-3 neural network of polynomial size but it cannot be well-approximated under the chosen input distribution by any depth-2 neural network of polynomial size.These results are not robust and require carefully chosen functions as well as input distributions.We show a similar separation between the expressive power of depth-2 and depth-3 sigmoidal neural networks over a large class of input distributions, as long as the weights are polynomially bounded.While doing so, we also show that depth-2 sigmoidal neural networks with small width and small weights can be well-approximated by low-degree multivariate polynomials.", "target": ["Tiefe-2-vs-3-Trennung für sigmoidale neuronale Netze über allgemeine Verteilungen"]} +{"source": "The smallest eigenvectors of the graph Laplacian are well-known to provide a succinct representation of the geometry of a weighted graph.In reinforcement learning (RL), where the weighted graph may be interpreted as the state transition process induced by a behavior policy acting on the environment, approximating the eigenvectors of the Laplacian provides a promising approach to state representation learning.However, existing methods for performing this approximation are ill-suited in general RL settings for two main reasons: First, they are computationally expensive, often requiring operations on large matrices.Second, these methods lack adequate justification beyond simple, tabular, finite-state settings.In this paper, we present a fully general and scalable method for approximating the eigenvectors of the Laplacian in a model-free RL context.We systematically evaluate our approach and empirically show that it generalizes beyond the tabular, finite-state setting.Even in tabular, finite-state settings, its ability to approximate the eigenvectors outperforms previous proposals.Finally, we show the potential benefits of using a Laplacian representation learned using our method in goal-achieving RL tasks, providing evidence that our technique can be used to significantly improve the performance of an RL agent.", "target": ["Wir schlagen eine skalierbare Methode zur Annäherung der Eigenvektoren des Laplacian im Kontext des Reinforcement Learning vor und zeigen, dass die erlernten Darstellungen die Leistung eines RL-Agenten verbessern können."]} +{"source": "Our work offers a new method for domain translation from semantic label mapsand Computer Graphic (CG) simulation edge map images to photo-realistic im-ages.We train a Generative Adversarial Network (GAN) in a conditional way togenerate a photo-realistic version of a given CG scene.Existing architectures ofGANs still lack the photo-realism capabilities needed to train DNNs for computervision tasks, we address this issue by embedding edge maps, and training it in anadversarial mode.We also offer an extension to our model that uses our GANarchitecture to create visually appealing and temporally coherent videos.", "target": ["Simulation von realen Bild-Übersetzungen und Video-Erzeugung"]} +{"source": "Deep neural networks are widely used in various domains, but the prohibitive computational complexity prevents their deployment on mobile devices.Numerous model compression algorithms have been proposed, however, it is often difficult and time-consuming to choose proper hyper-parameters to obtain an efficient compressed model.In this paper, we propose an automated framework for model compression and acceleration, namely PocketFlow.This is an easy-to-use toolkit that integrates a series of model compression algorithms and embeds a hyper-parameter optimization module to automatically search for the optimal combination of hyper-parameters.Furthermore, the compressed model can be converted into the TensorFlow Lite format and easily deployed on mobile devices to speed-up the inference.PocketFlow is now open-source and publicly available at https://github.com/Tencent/PocketFlow.", "target": ["Wir schlagen PocketFlow vor, ein automatisiertes Framework zur Modellkomprimierung und -beschleunigung, um den Einsatz von Deep Learning-Modellen auf mobilen Geräten zu erleichtern."]} +{"source": "Generative models provide a way to model structure in complex distributions and have been shown to be useful for many tasks of practical interest.However, current techniques for training generative models require access to fully-observed samples.In many settings, it is expensive or even impossible to obtain fully-observed samples, but economical to obtain partial, noisy observations.We consider the task of learning an implicit generative model given only lossy measurements of samples from the distribution of interest.We show that the true underlying distribution can be provably recovered even in the presence of per-sample information loss for a class of measurement models.Based on this, we propose a new method of training Generative Adversarial Networks (GANs) which we call AmbientGAN.On three benchmark datasets, and for various measurement models, we demonstrate substantial qualitative and quantitative improvements.Generative models trained with our method can obtain $2$-$4$x higher inception scores than the baselines.", "target": ["Wie man GANs aus verrauschten, verzerrten, partiellen Beobachtungen lernt"]} +{"source": "Random Matrix Theory (RMT) is applied to analyze the weight matrices of Deep Neural Networks (DNNs), including both production quality, pre-trained models such as AlexNet and Inception, and smaller models trained from scratch, such as LeNet5 and a miniature-AlexNet. Empirical and theoretical results clearly indicate that the empirical spectral density (ESD) of DNN layer matrices displays signatures of traditionally-regularized statistical models, even in the absence of exogenously specifying traditional forms of regularization, such as Dropout or Weight Norm constraints. Building on recent results in RMT, most notably its extension to Universality classes of Heavy-Tailed matrices, we develop a theory to identify 5+1 Phases of Training, corresponding to increasing amounts of Implicit Self-Regularization. For smaller and/or older DNNs, this Implicit Self-Regularization is like traditional Tikhonov regularization, in that there is a \"size scale\" separating signal from noise. For state-of-the-art DNNs, however, we identify a novel form of Heavy-Tailed Self-Regularization, similar to the self-organization seen in the statistical physics of disordered systems. This implicit Self-Regularization can depend strongly on the many knobs of the training process. By exploiting the generalization gap phenomena, we demonstrate that we can cause a small model to exhibit all 5+1 phases of training simply by changing the batch size.", "target": ["Siehe die Zusammenfassung. (Für die Überarbeitung ist die Arbeit identisch, mit Ausnahme eines 59-seitigen Zusatzmaterials, das als eigenständige Version des technischen Berichts des Papiers dienen kann)."]} +{"source": "We introduce an attention mechanism to improve feature extraction for deep active learning (AL) in the semi-supervised setting.The proposed attention mechanism is based on recent methods to visually explain predictions made by DNNs.We apply the proposed explanation-based attention to MNIST and SVHN classification.The conducted experiments show accuracy improvements for the original and class-imbalanced datasets with the same number of training examples and faster long-tail convergence compared to uncertainty-based methods.", "target": ["Wir führen einen Aufmerksamkeitsmechanismus ein, um die Merkmalsextraktion für tiefes aktives Lernen (AL) in der halbüberwachten Umgebung zu verbessern."]} +{"source": "We apply canonical forms of gradient complexes (barcodes) to explore neural networks loss surfaces.We present an algorithm for calculations of the objective function's barcodes of minima. Our experiments confirm two principal observations: (1) the barcodes of minima are located in a small lower part of the range of values of objective function and (2) increase of the neural network's depth brings down the minima's barcodes.This has natural implications for the neural network learning and the ability to generalize.", "target": ["Wir wenden kanonische Formen von Gradientenkomplexen (Barcodes) an, um die Verlustflächen neuronaler Netze zu untersuchen."]} +{"source": "New types of compute hardware in development and entering the market hold the promise of revolutionizing deep learning in a manner as profound as GPUs.However, existing software frameworks and training algorithms for deep learning have yet to evolve to fully leverage the capability of the new wave of silicon.In particular, models that exploit structured input via complex and instance-dependent control flow are difficult to accelerate using existing algorithms and hardware that typically rely on minibatching.We present an asynchronous model-parallel (AMP) training algorithm that is specifically motivated by training on networks of interconnected devices.Through an implementation on multi-core CPUs, we show that AMP training converges to the same accuracy as conventional synchronous training algorithms in a similar number of epochs, but utilizes the available hardware more efficiently, even for small minibatch sizes, resulting in shorter overall training times.Our framework opens the door for scaling up a new class of deep learning models that cannot be efficiently trained today.", "target": ["Verwendung asynchroner Gradientenaktualisierungen zur Beschleunigung des dynamischen Trainings neuronaler Netze"]} +{"source": "In cooperative multi-agent reinforcement learning (MARL), how to design a suitable reward signal to accelerate learning and stabilize convergence is a critical problem.The global reward signal assigns the same global reward to all agents without distinguishing their contributions, while the local reward signal provides different local rewards to each agent based solely on individual behavior.Both of the two reward assignment approaches have some shortcomings: the former might encourage lazy agents, while the latter might produce selfish agents.In this paper, we study reward design problem in cooperative MARL based on packet routing environments.Firstly, we show that the above two reward signals are prone to produce suboptimal policies.Then, inspired by some observations and considerations, we design some mixed reward signals, which are off-the-shelf to learn better policies.Finally, we turn the mixed reward signals into the adaptive counterparts, which achieve best results in our experiments.Other reward signals are also discussed in this paper.As reward design is a very fundamental problem in RL and especially in MARL, we hope that MARL researchers can rethink the rewards used in their systems.", "target": ["Wir untersuchen das Problem der Belohnungsgestaltung in kooperativen MARL auf der Grundlage von Paket-Routing-Umgebungen. Die experimentellen Ergebnisse erinnern uns daran, die Belohnungen sorgfältig zu gestalten, da sie wirklich wichtig sind, um das Verhalten der Agenten zu steuern."]} +{"source": "Recent advances have illustrated that it is often possible to learn to solve linear inverse problems in imaging using training data that can outperform more traditional regularized least squares solutions.Along these lines, we present some extensions of the Neumann network, a recently introduced end-to-end learned architecture inspired by a truncated Neumann series expansion of the solution map to a regularized least squares problem.Here we summarize the Neumann network approach, and show that it has a form compatible with the optimal reconstruction function for a given inverse problem.We also investigate an extension of the Neumann network that incorporates a more sample efficient patch-based regularization approach.", "target": ["Neumann-Netzwerke sind ein durchgängiger, stichprobeneffizienter Lernansatz zur Lösung linearer inverser Probleme in Abbildungen, der mit dem MSE-Optimum-Ansatz kompatibel ist und eine Erweiterung auf patch-basiertes Lernen zulässt."]} +{"source": "End-to-end task-oriented dialogue is challenging since knowledge bases are usually large, dynamic and hard to incorporate into a learning framework.We propose the global-to-local memory pointer (GLMP) networks to address this issue.In our model, a global memory encoder and a local memory decoder are proposed to share external knowledge.The encoder encodes dialogue history, modifies global contextual representation, and generates a global memory pointer.The decoder first generates a sketch response with unfilled slots.Next, it passes the global memory pointer to filter the external knowledge for relevant information, then instantiates the slots via the local memory pointers.We empirically show that our model can improve copy accuracy and mitigate the common out-of-vocabulary problem.As a result, GLMP is able to improve over the previous state-of-the-art models in both simulated bAbI Dialogue dataset and human-human Stanford Multi-domain Dialogue dataset on automatic and human evaluation.", "target": ["GLMP: Globale Memory Encoder (Kontext-RNN, globaler Zeiger) und lokale Memory Decoder (Skizzen-RNN, lokaler Zeiger), die externes Wissen (MemNN) gemeinsam nutzen, werden vorgeschlagen, um die Antwortgenerierung im aufgabenorientierten Dialog zu stärken."]} +{"source": "The checkerboard phenomenon is one of the well-known visual artifacts in the computer vision field.The origins and solutions of checkerboard artifacts in the pixel space have been studied for a long time, but their effects on the gradient space have rarely been investigated.In this paper, we revisit the checkerboard artifacts in the gradient space which turn out to be the weak point of a network architecture.We explore image-agnostic property of gradient checkerboard artifacts and propose a simple yet effective defense method by utilizing the artifacts.We introduce our defense module, dubbed Artificial Checkerboard Enhancer (ACE), which induces adversarial attacks on designated pixels.This enables the model to deflect attacks by shifting only a single pixel in the image with a remarkable defense rate.We provide extensive experiments to support the effectiveness of our work for various attack scenarios using state-of-the-art attack methods.Furthermore, we show that ACE is even applicable to large-scale datasets including ImageNet dataset and can be easily transferred to various pretrained networks.", "target": ["Wir schlagen ein neuartiges Aritificial Checkerboard Enhancer (ACE)-Modul vor, das Angriffe auf einen vorgegebenen Pixelraum lenkt und diesen mit einer einfachen Auffüllungsoperation erfolgreich verteidigt."]} +{"source": "Min-max formulations have attracted great attention in the ML community due to the rise of deep generative models and adversarial methods, and understanding the dynamics of (stochastic) gradient algorithms for solving such formulations has been a grand challenge.As a first step, we restrict to bilinear zero-sum games and give a systematic analysis of popular gradient updates, for both simultaneous and alternating versions.We provide exact conditions for their convergence and find the optimal parameter setup and convergence rates.In particular, our results offer formal evidence that alternating updates converge \"better\" than simultaneous ones.", "target": ["Wir analysieren systematisch das Konvergenzverhalten populärer Gradientenalgorithmen zur Lösung bilinearer Spiele, sowohl mit simultanen als auch mit alternierenden Updates."]} +{"source": "Most approaches in generalized zero-shot learning rely on cross-modal mapping between an image feature space and a class embedding space or on generating artificial image features.However, learning a shared cross-modal embedding by aligning the latent spaces of modality-specific autoencoders is shown to be promising in (generalized) zero-shot learning.While following the same direction, we also take artificial feature generation one step further and propose a model where a shared latent space of image features and class embeddings is learned by aligned variational autoencoders, for the purpose of generating latent features to train a softmax classifier.We evaluate our learned latent features on conventional benchmark datasets and establish a new state of the art on generalized zero-shot as well as on few-shot learning.Moreover, our results on ImageNet with various zero-shot splits show that our latent features generalize well in large-scale settings.", "target": ["Wir verwenden VAEs, um eine gemeinsame latente Raumeinbettung zwischen Bildmerkmalen und -attributen zu erlernen, und erreichen dadurch modernste Ergebnisse beim generalisierten Zero-Shot-Lernen."]} +{"source": "Intuitively, image classification should profit from using spatial information.Recent work, however, suggests that this might be overrated in standard CNNs.In this paper, we are pushing the envelope and aim to further investigate the reliance on and necessity of spatial information.We propose and analyze three methods, namely Shuffle Conv, GAP+FC and 1x1 Conv, that destroy spatial information during both training and testing phases.We extensively evaluate these methods on several object recognition datasets (CIFAR100, Small-ImageNet, ImageNet) with a wide range of CNN architectures (VGG16, ResNet50, ResNet152, MobileNet, SqueezeNet).Interestingly, we consistently observe that spatial information can be completely deleted from a significant number of layers with no or only small performance drops.", "target": ["Räumliche Informationen auf den letzten Ebenen sind für eine gute Klassifizierungsgenauigkeit nicht erforderlich."]} +{"source": "Disentangling underlying generative factors of a data distribution is important for interpretability and generalizable representations.In this paper, we introduce two novel disentangling methods.Our first method, Unlabeled Disentangling GAN (UD-GAN, unsupervised), decomposes the latent noise by generating similar/dissimilar image pairs and it learns a distance metric on these pairs with siamese networks and a contrastive loss.This pairwise approach provides consistent representations for similar data points.Our second method (UD-GAN-G, weakly supervised) modifies the UD-GAN with user-defined guidance functions, which restrict the information that goes into the siamese networks.This constraint helps UD-GAN-G to focus on the desired semantic variations in the data.We show that both our methods outperform existing unsupervised approaches in quantitative metrics that measure semantic accuracy of the learned representations.In addition, we illustrate that simple guidance functions we use in UD-GAN-G allow us to directly capture the desired variations in the data.", "target": ["Wir verwenden Siamesische Netze, um den Generierungsprozess in GANs ohne gelabelte Daten zu steuern und zu entwirren."]} +{"source": "We present Predicted Variables, an approach to making machine learning (ML) a first class citizen in programming languages.There is a growing divide in approaches to building systems: using human experts (e.g. programming) on the one hand, and using behavior learned from data (e.g. ML) on the other hand.PVars aim to make using ML in programming easier by hybridizing the two.We leverage the existing concept of variables and create a new type, a predicted variable.PVars are akin to native variables with one important distinction: PVars determine their value using ML when evaluated.We describe PVars and their interface, how they can be used in programming, and demonstrate the feasibility of our approach on three algorithmic problems: binary search, QuickSort, and caches.We show experimentally that PVars are able to improve over the commonly used heuristics and lead to a better performance than the original algorithms.As opposed to previous work applying ML to algorithmic problems, PVars have the advantage that they can be used within the existing frameworks and do not require the existing domain knowledge to be replaced.PVars allow for a seamless integration of ML into existing systems and algorithms.Our PVars implementation currently relies on standard Reinforcement Learning (RL) methods.To learn faster, PVars use the heuristic function, which they are replacing, as an initial function.We show that PVars quickly pick up the behavior of the initial function and then improve performance beyond that without ever performing substantially worse -- allowing for a safe deployment in critical applications.", "target": ["Wir stellen Predicted Variables vor, einen Ansatz, der maschinelles Lernen in Programmiersprachen zu einem First Class Citizen macht."]} +{"source": "Much recent research has been devoted to video prediction and generation, but mostly for short-scale time horizons.The hierarchical video prediction method by Villegas et al. (2017) is an example of a state of the art method for long term video prediction. However, their method has limited applicability in practical settings as it requires a ground truth pose (e.g., poses of joints of a human) at training time. This paper presents a long term hierarchical video prediction model that does not have such a restriction.We show that the network learns its own higher-level structure (e.g., pose equivalent hidden variables) that works better in cases where the ground truth pose does not fully capture all of the information needed to predict the next frame. This method gives sharper results than other video prediction methods which do not require a ground truth pose, and its efficiency is shown on the Humans 3.6M and Robot Pushing datasets.", "target": ["Wir zeigen Wege auf, wie man ein hierarchisches Videovorhersagemodell trainieren kann, ohne Posenmarkierungen zu benötigen."]} +{"source": "Combining information from different sensory modalities to execute goal directed actions is a key aspect of human intelligence.Specifically, human agents are very easily able to translate the task communicated in one sensory domain (say vision) into a representation that enables them to complete this task when they can only sense their environment using a separate sensory modality (say touch).In order to build agents with similar capabilities, in this work we consider the problem of a retrieving a target object from a drawer.The agent is provided with an image of a previously unseen object and it explores objects in the drawer using only tactile sensing to retrieve the object that was shown in the image without receiving any visual feedback.Success at this task requires close integration of visual and tactile sensing.We present a method for performing this task in a simulated environment using an anthropomorphic hand.We hope that future research in the direction of combining sensory signals for acting will find the object retrieval from a drawer to be a useful benchmark problem", "target": ["In dieser Arbeit untersuchen wir das Problem des Lernens von Repräsentationen zur Identifizierung neuer Objekte durch die Erkundung von Objekten mit Hilfe des Tastsinns. Der springende Punkt dabei ist, dass die Abfrage im Bildbereich erfolgt."]} +{"source": "Locality sensitive hashing schemes such as \\simhash provide compact representations of multisets from which similarity can be estimated.However, in certain applications, we need to estimate the similarity of dynamically changing sets. In this case, we need the representation to be a homomorphism so that the hash of unions and differences of sets can be computed directly from the hashes of operands. We propose two representations that have this property for cosine similarity (an extension of \\simhash and angle-preserving random projections), and make substantial progress on a third representation for Jaccard similarity (an extension of \\minhash).We employ these hashes to compress the sufficient statistics of a conditional random field (CRF) coreference model and study how this compression affects our ability to compute similarities as entities are split and merged during inference.\\cut{We study these hashes in a conditional random field (CRF) hierarchical coreference model in order to compute the similarity of entities as they are merged and split during inference.}We also provide novel statistical analysis of \\simhash to help justify it as an estimator inside a CRF, showing that the bias and variance reduce quickly with the number of bits.On a problem of author coreference, we find that our \\simhash scheme allows scaling the hierarchical coreference algorithm by an order of magnitude without degrading its statistical performance or the model's coreference accuracy, as long as we employ at least 128 or 256 bits. Angle-preserving random projections further improve the coreference quality, potentially allowing even fewer dimensions to be used.", "target": ["Wir verwenden lineare homomorphe Komprimierungsverfahren, um die hinreichenden Statistiken eines bedingten Zufallsfeldmodells der Koreferenz darzustellen, was uns erlaubt, die Inferenz zu skalieren und die Geschwindigkeit um eine Größenordnung zu verbessern."]} +{"source": "Motivated by applications to unsupervised learning, we consider the problem of measuring mutual information.Recent analysis has shown that naive kNN estimators of mutual information have serious statistical limitations motivating more refined methods.In this paper we prove that serious statistical limitations are inherent to any measurement method.More specifically, we show that any distribution-free high-confidence lower bound on mutual information cannot be larger than $O(\\ln N)$ where $N$ is the size of the data sample.We also analyze the Donsker-Varadhan lower bound on KL divergence in particular and show that, when simple statistical considerations are taken into account, this bound can never produce a high-confidence value larger than $\\ln N$.While large high-confidence lower bounds are impossible, in practice one can use estimators without formal guarantees.We suggest expressing mutual information as a difference of entropies and using cross entropy as an entropy estimator. We observe that, although cross entropy is only an upper bound on entropy, cross-entropy estimates converge to the true cross entropy at the rate of $1/\\sqrt{N}$.", "target": ["Wir geben eine theoretische Analyse der Messung und Optimierung der gegenseitigen Information."]} +{"source": "In this paper, we propose a neural network framework called neuron hierarchical network (NHN), that evolves beyond the hierarchy in layers, and concentrates on the hierarchy of neurons.We observe mass redundancy in the weights of both handcrafted and randomly searched architectures.Inspired by the development of human brains, we prune low-sensitivity neurons in the model and add new neurons to the graph, and the relation between individual neurons are emphasized and the existence of layers weakened.We propose a process to discover the best base model by random architecture search, and discover the best locations and connections of the added neurons by evolutionary search.Experiment results show that the NHN achieves higher test accuracy on Cifar-10 than state-of-the-art handcrafted and randomly searched architectures, while requiring much fewer parameters and less searching time.", "target": ["Durch das Aufbrechen der Schichtenhierarchie schlagen wir einen dreistufigen Ansatz für die Konstruktion von Neuronen-Hierarchienetzen vor, die NAS, SMASH und die hierarchische Darstellung mit weniger Parametern und kürzerer Suchzeit übertreffen."]} +{"source": "Simulation is a useful tool in situations where training data for machine learning models is costly to annotate or even hard to acquire.In this work, we propose a reinforcement learning-based method for automatically adjusting the parameters of any (non-differentiable) simulator, thereby controlling the distribution of synthesized data in order to maximize the accuracy of a model trained on that data.In contrast to prior art that hand-crafts these simulation parameters or adjusts only parts of the available parameters, our approach fully controls the simulator with the actual underlying goal of maximizing accuracy, rather than mimicking the real data distribution or randomly generating a large volume of data.We find that our approach(i) quickly converges to the optimal simulation parameters in controlled experiments and(ii) can indeed discover good sets of parameters for an image rendering simulator in actual computer vision applications.", "target": ["Wir schlagen einen Algorithmus vor, der automatisch die Parameter einer Simulationsmaschine anpasst, um Trainingsdaten für ein neuronales Netz zu erzeugen, so dass die Validierungsgenauigkeit maximiert wird."]} +{"source": "Modelling statistical relationships beyond the conditional mean is crucial in many settings.Conditional density estimation (CDE) aims to learn the full conditional probability density from data.Though highly expressive, neural network based CDE models can suffer from severe over-fitting when trained with the maximum likelihood objective.Due to the inherent structure of such models, classical regularization approaches in the parameter space are rendered ineffective.To address this issue, we develop a model-agnostic noise regularization method for CDE that adds random perturbations to the data during training.We demonstrate that the proposed approach corresponds to a smoothness regularization and prove its asymptotic consistency.In our experiments, noise regularization significantly and consistently outperforms other regularization methods across seven data sets and three CDE models.The effectiveness of noise regularization makes neural network based CDE the preferable method over previous non- and semi-parametric approaches, even when training data is scarce.", "target": ["Ein modellagnostisches Regularisierungsschema für die auf neuronalen Netzen basierende Schätzung der bedingten Dichte."]} +{"source": "Unsupervised representation learning holds the promise of exploiting large amount of available unlabeled data to learn general representations.A promising technique for unsupervised learning is the framework of Variational Auto-encoders (VAEs).However, unsupervised representations learned by VAEs are significantly outperformed by those learned by supervising for recognition.Our hypothesis is that to learn useful representations for recognition the model needs to be encouraged to learn about repeating and consistent patterns in data.Drawing inspiration from the mid-level representation discovery work, we propose PatchVAE, that reasons about images at patch level.Our key contribution is a bottleneck formulation in a VAE framework that encourages mid-level style representations.Our experiments demonstrate that representations learned by our method perform much better on the recognition tasks compared to those learned by vanilla VAEs.", "target": ["Eine patch-basierte Bottleneck-Formulierung in einem VAE-Rahmen, die unüberwachte Repräsentationen lernt, die besser für die visuelle Erkennung geeignet sind."]} +{"source": "Vanishing and exploding gradients are two of the main obstacles in training deep neural networks, especially in capturing long range dependencies in recurrent neural networks (RNNs).In this paper, we present an efficient parametrization of the transition matrix of an RNN that allows us to stabilize the gradients that arise in its training.Specifically, we parameterize the transition matrix by its singular value decomposition (SVD), which allows us to explicitly track and control its singular values.We attain efficiency by using tools that are common in numerical linear algebra, namely Householder reflectors for representing the orthogonal matrices that arise in the SVD.By explicitly controlling the singular values, our proposed svdRNN method allows us to easily solve the exploding gradient problem and we observe that it empirically solves the vanishing gradient issue to a large extent.We note that the SVD parameterization can be used for any rectangular weight matrix, hence it can be easily extended to any deep neural network, such as a multi-layer perceptron.Theoretically, we demonstrate that our parameterization does not lose any expressive power, and show how it potentially makes the optimization process easier.Our extensive experimental results also demonstrate that the proposed framework converges faster, and has good generalization, especially when the depth is large.", "target": ["Um das Problem des Verschwindens/Explodierens des Gradienten zu lösen, schlagen wir eine effiziente Parametrisierung der Übergangsmatrix von RNN vor, die keine Ausdruckskraft verliert, schneller konvergiert und eine gute Generalisierung aufweist."]} +{"source": "Recent image style transferring methods achieved arbitrary stylization with input content and style images.To transfer the style of an arbitrary image to a content image, these methods used a feed-forward network with a lowest-scaled feature transformer or a cascade of the networks with a feature transformer of a corresponding scale.However, their approaches did not consider either multi-scaled style in their single-scale feature transformer or dependency between the transformed feature statistics across the cascade networks.This shortcoming resulted in generating partially and inexactly transferred style in the generated images.To overcome this limitation of partial style transfer, we propose a total style transferring method which transfers multi-scaled feature statistics through a single feed-forward process.First, our method transforms multi-scaled feature maps of a content image into those of a target style image by considering both inter-channel correlations in each single scaled feature map and inter-scale correlations between multi-scaled feature maps.Second, each transformed feature map is inserted into the decoder layer of the corresponding scale using skip-connection.Finally, the skip-connected multi-scaled feature maps are decoded into a stylized image through our trained decoder network.", "target": ["Ein Papier, das eine Methode zur Umwandlung des Bildstils mithilfe von tiefen neuronalen Netzen vorschlägt."]} +{"source": "Beyond understanding what is being discussed, human communication requires an awareness of what someone is feeling.One challenge for dialogue agents is recognizing feelings in the conversation partner and replying accordingly, a key communicative skill that is trivial for humans.Research in this area is made difficult by the paucity of suitable publicly available datasets both for emotion and dialogues.This work proposes a new task for empathetic dialogue generation and EmpatheticDialogues, a dataset of 25k conversations grounded in emotional situations to facilitate training and evaluating dialogue systems.Our experiments indicate that dialogue models that use our dataset are perceived to be more empathetic by human evaluators, while improving on other metrics as well (e.g. perceived relevance of responses, BLEU scores), compared to models merely trained on large-scale Internet conversation data.We also present empirical comparisons of several ways to improve the performance of a given model by leveraging existing models or datasets without requiring lengthy re-training of the full model.", "target": ["Wir verbessern bestehende Dialogsysteme zur Beantwortung von Menschen, die persönliche Geschichten erzählen, indem wir Darstellungen zur Vorhersage von Emotionen einbeziehen und außerdem einen neuen Benchmark und Datensatz für einfühlsame Dialoge veröffentlichen."]} +{"source": "Granger causality is a widely-used criterion for analyzing interactions in large-scale networks.As most physical interactions are inherently nonlinear, we consider the problem of inferring the existence of pairwise Granger causality between nonlinearly interacting stochastic processes from their time series measurements.Our proposed approach relies on modeling the embedded nonlinearities in the measurements using a component-wise time series prediction model based on Statistical Recurrent Units (SRUs).We make a case that the network topology of Granger causal relations is directly inferrable from a structured sparse estimate of the internal parameters of the SRU networks trained to predict the processes’ time series measurements.We propose a variant of SRU, called economy-SRU, which, by design has considerably fewer trainable parameters, and therefore less prone to overfitting.The economy-SRU computes a low-dimensional sketch of its high-dimensional hidden state in the form of random projections to generate the feedback for its recurrent processing.Additionally, the internal weight parameters of the economy-SRU are strategically regularized in a group-wise manner to facilitate the proposed network in extracting meaningful predictive features that are highly time-localized to mimic real-world causal events.Extensive experiments are carried out to demonstrate that the proposed economy-SRU based time series prediction model outperforms the MLP, LSTM and attention-gated CNN-based time series models considered previously for inferring Granger causality.", "target": ["Eine neue rekurrente neuronale Netzarchitektur zur Erkennung paarweiser Granger-Kausalität zwischen nichtlinear interagierenden Zeitreihen."]} +{"source": "Graph convolutional networks (GCNs) are powerful deep neural networks for graph-structured data.However, GCN computes nodes' representation recursively from their neighbors, making the receptive field size grow exponentially with the number of layers. Previous attempts on reducing the receptive field size by subsampling neighbors do not have any convergence guarantee, and their receptive field size per node is still in the order of hundreds.In this paper, we develop a preprocessing strategy and two control variate based algorithms to further reduce the receptive field size.Our algorithms are guaranteed to converge to GCN's local optimum regardless of the neighbor sampling size.Empirical results show that our algorithms have a similar convergence speed per epoch with the exact algorithm even using only two neighbors per node.The time consumption of our algorithm on the Reddit dataset is only one fifth of previous neighbor sampling algorithms.", "target": ["Ein auf Kontrollvariablen basierender stochastischer Trainingsalgorithmus für graphische Faltungsnetzwerke, bei dem das rezeptive Feld nur zwei Nachbarn pro Knoten sein kann."]} +{"source": "Low bit-width integer weights and activations are very important for efficient inference, especially with respect to lower power consumption.We propose to apply Monte Carlo methods and importance sampling to sparsify and quantize pre-trained neural networks without any retraining.We obtain sparse, low bit-width integer representations that approximate the full precision weights and activations.The precision, sparsity, and complexity are easily configurable by the amount of sampling performed.Our approach, called Monte Carlo Quantization (MCQ), is linear in both time and space, while the resulting quantized sparse networks show minimal accuracy loss compared to the original full-precision networks.Our method either outperforms or achieves results competitive with methods that do require additional training on a variety of challenging tasks.", "target": ["Monte-Carlo-Methoden zur Quantisierung von vortrainierten Modellen ohne zusätzliches Training."]} +{"source": "We propose the Information Maximization Autoencoder (IMAE), an information theoretic approach to simultaneously learn continuous and discrete representations in an unsupervised setting.Unlike the Variational Autoencoder framework, IMAE starts from a stochastic encoder that seeks to map each input data to a hybrid discrete and continuous representation with the objective of maximizing the mutual information between the data and their representations.A decoder is included to approximate the posterior distribution of the data given their representations, where a high fidelity approximation can be achieved by leveraging the informative representations. We show that the proposed objective is theoretically valid and provides a principled framework for understanding the tradeoffs regarding informativeness of each representation factor, disentanglement of representations, and decoding quality.", "target": ["Informationstheoretischer Ansatz für das unüberwachte Lernen eines Hybrids aus diskreten und kontinuierlichen Repräsentationen,"]} +{"source": "Learning rules for neural networks necessarily include some form of regularization.Most regularization techniques are conceptualized and implemented in the space of parameters.However, it is also possible to regularize in the space of functions.Here, we propose to measure networks in an $L^2$ Hilbert space, and test a learning rule that regularizes the distance a network can travel through $L^2$-space each update. This approach is inspired by the slow movement of gradient descent through parameter space as well as by the natural gradient, which can be derived from a regularization term upon functional change.The resulting learning rule, which we call Hilbert-constrained gradient descent (HCGD), is thus closely related to the natural gradient but regularizes a different and more calculable metric over the space of functions.Experiments show that the HCGD is efficient and leads to considerably better generalization.", "target": ["Es ist wichtig, die Optimierung im Funktionsraum und nicht nur im Parameterraum zu betrachten. Wir führen eine Lernregel ein, die den im Funktionsraum zurückgelegten Weg reduziert, genau wie SGD den im Parameterraum zurückgelegten Weg begrenzt."]} +{"source": "Stochastic gradient descent (SGD), which dates back to the 1950s, is one of the most popular and effective approaches for performing stochastic optimization.Research on SGD resurged recently in machine learning for optimizing convex loss functions and training nonconvex deep neural networks.The theory assumes that one can easily compute an unbiased gradient estimator, which is usually the case due to the sample average nature of empirical risk minimization.There exist, however, many scenarios (e.g., graphs) where an unbiased estimator may be as expensive to compute as the full gradient because training examples are interconnected.Recently, Chen et al. (2018) proposed using a consistent gradient estimator as an economic alternative.Encouraged by empirical success, we show, in a general setting, that consistent estimators result in the same convergence behavior as do unbiased ones.Our analysis covers strongly convex, convex, and nonconvex objectives.We verify the results with illustrative experiments on synthetic and real-world data.This work opens several new research directions, including the development of more efficient SGD updates with consistent estimators and the design of efficient training algorithms for large-scale graphs.", "target": ["Konvergenztheorie für verzerrte (aber konsistente) Gradientenschätzer in der stochastischen Optimierung und Anwendung auf Graph Convolutional Networks"]} +{"source": "We consider the problem of uncertainty estimation in the context of (non-Bayesian) deep neural classification.In this context, all known methods are based on extracting uncertainty signals from a trained network optimized to solve the classification problem at hand.We demonstrate that such techniques tend to introduce biased estimates for instances whose predictions are supposed to be highly confident.We argue that this deficiency is an artifact of the dynamics of training with SGD-like optimizers, and it has some properties similar to overfitting.Based on this observation, we develop an uncertainty estimation algorithm that selectively estimates the uncertainty of highly confident points, using earlier snapshots of the trained model, before their estimates are jittered (and way before they are ready for actual classification).We present extensive experiments indicating that the proposed algorithm provides uncertainty estimates that are consistently better than all known methods.", "target": ["Wir verwenden Schnappschüsse aus dem Trainingsprozess, um jede Methode zur Unsicherheitsabschätzung eines DNN-Klassifikators zu verbessern."]} +{"source": "Existing public face image datasets are strongly biased toward Caucasian faces, and other races (e.g., Latino) are significantly underrepresented.The models trained from such datasets suffer from inconsistent classification accuracy, which limits the applicability of face analytic systems to non-White race groups.To mitigate the race bias problem in these datasets, we constructed a novel face image dataset containing 108,501 images which is balanced on race.We define 7 race groups: White, Black, Indian, East Asian, Southeast Asian, Middle Eastern, and Latino.Images were collected from the YFCC-100M Flickr dataset and labeled with race, gender, and age groups.Evaluations were performed on existing face attribute datasets as well as novel image datasets to measure the generalization performance.We find that the model trained from our dataset is substantially more accurate on novel datasets and the accuracy is consistent across race and gender groups.We also compare several commercial computer vision APIs and report their balanced accuracy across gender, race, and age groups.", "target": ["Ein neuer Gesichtsbilddatensatz für ausgewogene Rasse, Geschlecht und Alter, der für die Messung und Abschwächung von Vorurteilen verwendet werden kann"]} +{"source": "Dramatic advances in generative models have resulted in near photographic quality for artificially rendered faces, animals and other objects in the natural world.In spite of such advances, a higher level understanding of vision and imagery does not arise from exhaustively modeling an object, but instead identifying higher-level attributes that best summarize the aspects of an object. In this work we attempt to model the drawing process of fonts by building sequential generative models of vector graphics. This model has the benefit of providing a scale-invariant representation for imagery whose latent representation may be systematically manipulated and exploited to perform style propagation.We demonstrate these results on a large dataset of fonts and highlight how such a model captures the statistical dependencies and richness of this dataset.We envision that our model can find use as a tool for designers to facilitate font design.", "target": ["Wir versuchen, den Zeichenprozess von Schriftarten zu modellieren, indem wir sequentielle generative Modelle von Vektorgrafiken (SVGs), einer hochstrukturierten Darstellung von Schriftzeichen, erstellen."]} +{"source": "What can we learn about the functional organization of cortical microcircuits from large-scale recordings of neural activity? To obtain an explicit and interpretable model of time-dependent functional connections between neurons and to establish the dynamics of the cortical information flow, we develop 'dynamic neural relational inference' (dNRI).We study both synthetic and real-world neural spiking data and demonstrate that the developed method is able to uncover the dynamic relations between neurons more reliably than existing baselines.", "target": ["Wir entwickeln eine \"dynamische neuronale relationale Inferenz\", ein Variations-Auto-Codierer-Modell, das die verborgenen dynamischen Beziehungen zwischen Neuronen explizit und interpretierbar darstellen kann."]} +{"source": "DeePa is a deep learning framework that explores parallelism in all parallelizable dimensions to accelerate the training process of convolutional neural networks.DeePa optimizes parallelism at the granularity of each individual layer in the network.We present an elimination-based algorithm that finds an optimal parallelism configuration for every layer.Our evaluation shows that DeePa achieves up to 6.5× speedup compared to state-of-the-art deep learning frameworks and reduces data transfers by up to 23×.", "target": ["Unseres Wissens nach ist DeePa das erste Deep-Learning-Framework, das die Parallelität von CNNs in allen parallelisierbaren Dimensionen auf der Granularität der einzelnen Schichten kontrolliert und optimiert."]} +{"source": "One can substitute each neuron in any neural network with a kernel machine and obtain a counterpart powered by kernel machines.The new network inherits the expressive power and architecture of the original but works in a more intuitive way since each node enjoys the simple interpretation as a hyperplane (in a reproducing kernel Hilbert space).Further, using the kernel multilayer perceptron as an example, we prove that in classification, an optimal representation that minimizes the risk of the network can be characterized for each hidden layer.This result removes the need of backpropagation in learning the model and can be generalized to any feedforward kernel network.Moreover, unlike backpropagation, which turns models into black boxes, the optimal hidden representation enjoys an intuitive geometric interpretation, making the dynamics of learning in a deep kernel network simple to understand.Empirical results are provided to validate our theory.", "target": ["Wir kombinieren die Kernel-Methode mit konnektiven Modellen und zeigen, dass die resultierenden tiefen Architekturen schichtweise trainiert werden können und eine transparentere Lerndynamik aufweisen."]} +{"source": "Many notions of fairness may be expressed as linear constraints, and the resulting constrained objective is often optimized by transforming the problem into its Lagrangian dual with additive linear penalties.In non-convex settings, the resulting problem may be difficult to solve as the Lagrangian is not guaranteed to have a deterministic saddle-point equilibrium. In this paper, we propose to modify the linear penalties to second-order ones, and we argue that this results in a more practical training procedure in non-convex, large-data settings.For one, the use of second-order penalties allows training the penalized objective with a fixed value of the penalty coefficient, thus avoiding the instability and potential lack of convergence associated with two-player min-max games.Secondly, we derive a method for efficiently computing the gradients associated with the second-order penalties in stochastic mini-batch settings.Our resulting algorithm performs well empirically, learning an appropriately fair classifier on a number of standard benchmarks.", "target": ["Wir schlagen eine Methode zur stochastischen Optimierung von Strafen zweiter Ordnung vor und zeigen, wie diese für das Training von fairnessbewussten Klassifikatoren angewendet werden kann."]} +{"source": "Training methods for deep networks are primarily variants on stochastic gradient descent. Techniques that use (approximate) second-order information are rarely used because of the computational cost and noise associated with those approaches in deep learning contexts. However, in this paper, we show how feedforward deep networks exhibit a low-rank derivative structure. This low-rank structure makes it possible to use second-order information without needing approximations and without incurring a significantly greater computational cost than gradient descent. To demonstrate this capability, we implement Cubic Regularization (CR) on a feedforward deep network with stochastic gradient descent and two of its variants. There, we use CR to calculate learning rates on a per-iteration basis while training on the MNIST and CIFAR-10 datasets. CR proved particularly successful in escaping plateau regions of the objective function. We also found that this approach requires less problem-specific information (e.g. an optimal initial learning rate) than other first-order methods in order to perform well.", "target": ["Wir zeigen, dass die Ableitungen von Deep-Learning-Netzwerken eine Low-Rank-Struktur haben, und diese Struktur ermöglicht es uns, Ableitungsinformationen zweiter Ordnung zu verwenden, um die Lernraten adaptiv und auf eine rechnerisch machbare Weise zu berechnen."]} +{"source": "The worst-case training principle that minimizes the maximal adversarial loss, also known as adversarial training (AT), has shown to be a state-of-the-art approach for enhancing adversarial robustness against norm-ball bounded input perturbations.Nonetheless, min-max optimization beyond the purpose of AT has not been rigorously explored in the research of adversarial attack and defense.In particular, given a set of risk sources (domains), minimizing the maximal loss induced from the domain set can be reformulated as a general min-max problem that is different from AT.Examples of this general formulation include attacking model ensembles, devising universal perturbation under multiple inputs or data transformations, and generalized AT over different types of attack models.We show that these problems can be solved under a unified and theoretically principled min-max optimization framework. We also show that the self-adjusted domain weights learned from our method provides a means to explain the difficulty level of attack and defense over multiple domains.Extensive experiments show that our approach leads to substantial performance improvement over the conventional averaging strategy.", "target": ["Ein einheitlicher Min-Max-Optimierungsrahmen für gegnerische Angriffe und Verteidigung"]} +{"source": "Most deep learning models rely on expressive high-dimensional representations to achieve good performance on tasks such as classification.However, the high dimensionality of these representations makes them difficult to interpret and prone to over-fitting.We propose a simple, intuitive and scalable dimension reduction framework that takes into account the soft probabilistic interpretation of standard deep models for classification.When applying our framework to visualization, our representations more accurately reflect inter-class distances than standard visualization techniques such as t-SNE.We show experimentally that our framework improves generalization performance to unseen categories in zero-shot learning.We also provide a finite sample error upper bound guarantee for the method.", "target": ["Dimensionalitätsreduktion für Fälle, in denen Beispiele als weiche Wahrscheinlichkeitsverteilungen dargestellt werden können"]} +{"source": "Intrinsically motivated goal exploration algorithms enable machines to discover repertoires of policies that produce a diversity of effects in complex environments.These exploration algorithms have been shown to allow real world robots to acquire skills such as tool use in high-dimensional continuous state and action spaces.However, they have so far assumed that self-generated goals are sampled in a specifically engineered feature space, limiting their autonomy.In this work, we propose an approach using deep representation learning algorithms to learn an adequate goal space.This is a developmental 2-stage approach: first, in a perceptual learning stage, deep learning algorithms use passive raw sensor observations of world changes to learn a corresponding latent space; then goal exploration happens in a second stage by sampling goals in this latent space.We present experiments with a simulated robot arm interacting with an object, and we show that exploration algorithms using such learned representations can closely match, and even sometimes improve, the performance obtained using engineered representations.", "target": ["Wir schlagen eine neuartige Architektur für intrinsisch motivierte Zielexploration mit unbeaufsichtigtem Lernen von Zielraumrepräsentationen vor und bewerten, wie verschiedene Implementierungen die Entdeckung einer Vielfalt von Strategien ermöglichen."]} +{"source": "One of the main challenges of deep learning methods is the choice of an appropriate training strategy.In particular, additional steps, such as unsupervised pre-training, have been shown to greatly improve the performances of deep structures.In this article, we propose an extra training step, called post-training, which only optimizes the last layer of the network.We show that this procedure can be analyzed in the context of kernel theory, with the first layers computing an embedding of the data and the last layer a statistical model to solve the task based on this embedding.This step makes sure that the embedding, or representation, of the data is used in the best possible way for the considered task.This idea is then tested on multiple architectures with various data sets, showing that it consistently provides a boost in performance.", "target": ["Wir schlagen einen zusätzlichen Trainingsschritt vor, das so genannte Post-Training, bei dem die optimalen Gewichte für die letzte Schicht des Netzes berechnet werden."]} +{"source": "Natural language processing (NLP) models often require a massive number of parameters for word embeddings, resulting in a large storage or memory footprint.Deploying neural NLP models to mobile devices requires compressing the word embeddings without any significant sacrifices in performance.For this purpose, we propose to construct the embeddings with few basis vectors.For each word, the composition of basis vectors is determined by a hash code.To maximize the compression rate, we adopt the multi-codebook quantization approach instead of binary coding scheme.Each code is composed of multiple discrete numbers, such as (3, 2, 1, 8), where the value of each component is limited to a fixed range.We propose to directly learn the discrete codes in an end-to-end neural network by applying the Gumbel-softmax trick.Experiments show the compression rate achieves 98% in a sentiment analysis task and 94% ~ 99% in machine translation tasks without performance loss.In both tasks, the proposed method can improve the model performance by slightly lowering the compression rate.Compared to other approaches such as character-level segmentation, the proposed method is language-independent and does not require modifications to the network architecture.", "target": ["Komprimierung der Worteinbettungen um über 94% ohne Leistungseinbußen."]} +{"source": "It is important to detect anomalous inputs when deploying machine learning systems.The use of larger and more complex inputs in deep learning magnifies the difficulty of distinguishing between anomalous and in-distribution examples.At the same time, diverse image and text data are available in enormous quantities.We propose leveraging these data to improve deep anomaly detection by training anomaly detectors against an auxiliary dataset of outliers, an approach we call Outlier Exposure (OE).This enables anomaly detectors to generalize and detect unseen anomalies.In extensive experiments on natural language processing and small- and large-scale vision tasks, we find that Outlier Exposure significantly improves detection performance.We also observe that cutting-edge generative models trained on CIFAR-10 may assign higher likelihoods to SVHN images than to CIFAR-10 images; we use OE to mitigate this issue.We also analyze the flexibility and robustness of Outlier Exposure, and identify characteristics of the auxiliary dataset that improve performance.", "target": ["OE lehrt Anomalie-Detektoren, Heuristiken für die Erkennung ungesehener Anomalien zu erlernen; Experimente sind in der Klassifizierung, Dichte-Schätzung und Kalibrierung in NLP und Vision-Einstellungen; wir tunen nicht auf Test-Verteilung Proben, im Gegensatz zu früheren Arbeiten"]} +{"source": "While generative neural networks can learn to transform a specific input dataset into a specific target dataset, they require having just such a paired set of input/output datasets.For instance, to fool the discriminator, a generative adversarial network (GAN) exclusively trained to transform images of black-haired *men* to blond-haired *men* would need to change gender-related characteristics as well as hair color when given images of black-haired *women* as input.This is problematic, as often it is possible to obtain *a* pair of (source, target) distributions but then have a second source distribution where the target distribution is unknown.The computational challenge is that generative models are good at generation within the manifold of the data that they are trained on.However, generating new samples outside of the manifold or extrapolating \"out-of-sample\" is a much harder problem that has been less well studied.To address this, we introduce a technique called *neuron editing* that learns how neurons encode an edit for a particular transformation in a latent space.We use an autoencoder to decompose the variation within the dataset into activations of different neurons and generate transformed data by defining an editing transformation on those neurons.By performing the transformation in a latent trained space, we encode fairly complex and non-linear transformations to the data with much simpler distribution shifts to the neuron's activations.Our technique is general and works on a wide variety of data domains and applications.We first demonstrate it on image transformations and then move to our two main biological applications: removal of batch artifacts representing unwanted noise and modeling the effect of drug treatments to predict synergy between drugs.", "target": ["Verfahren zum Lernen einer Transformation zwischen einem Paar von Quell-/Zieldatensätzen und deren Anwendung auf einen separaten Quelldatensatz, für den es keinen Zieldatensatz gibt"]} +{"source": "In this paper, we propose to combine imitation and reinforcement learning via the idea of reward shaping using an oracle.We study the effectiveness of the near- optimal cost-to-go oracle on the planning horizon and demonstrate that the cost- to-go oracle shortens the learner’s planning horizon as function of its accuracy: a globally optimal oracle can shorten the planning horizon to one, leading to a one- step greedy Markov Decision Process which is much easier to optimize, while an oracle that is far away from the optimality requires planning over a longer horizon to achieve near-optimal performance.Hence our new insight bridges the gap and interpolates between imitation learning and reinforcement learning.Motivated by the above mentioned insights, we propose Truncated HORizon Policy Search (THOR), a method that focuses on searching for policies that maximize the total reshaped reward over a finite planning horizon when the oracle is sub-optimal.We experimentally demonstrate that a gradient-based implementation of THOR can achieve superior performance compared to RL baselines and IL baselines even when the oracle is sub-optimal.", "target": ["Kombination von Imitationslernen und Reinforcement Learning, um zu lernen, den Experten zu übertreffen"]} +{"source": "Recently, Generative Adversarial Networks (GANs) have emerged as a popular alternative for modeling complex high dimensional distributions.Most of the existing works implicitly assume that the clean samples from the target distribution are easily available.However, in many applications, this assumption is violated.In this paper, we consider the observation setting in which the samples from a target distribution are given by the superposition of two structured components, and leverage GANs for learning of the structure of the components.We propose a novel framework, demixing-GAN, which learns the distribution of two components at the same time.Through extensive numerical experiments, we demonstrate that the proposed framework can generate clean samples from unknown distributions, which further can be used in demixing of the unseen test images.", "target": ["Ein unüberwachter Lernansatz zur Trennung von zwei strukturierten Signalen aus ihrer Überlagerung"]} +{"source": "Normalising Flows (NFs) are a class of likelihood-based generative models that have recently gained popularity.They are based on the idea of transforming a simple density into that of the data.We seek to better understand this class of models, and how they compare to previously proposed techniques for generative modeling and unsupervised representation learning.For this purpose we reinterpret NFs in the framework of Variational Autoencoders (VAEs), and present a new form of VAE that generalises normalising flows.The new generalised model also reveals a close connection to denoising autoencoders, and we therefore call our model the Variational Denoising Autoencoder (VDAE).Using our unified model, we systematically examine the model space between flows, variational autoencoders, and denoising autoencoders, in a set of preliminary experiments on the MNIST handwritten digits.The experiments shed light on the modeling assumptions implicit in these models, and they suggest multiple new directions for future research in this space.", "target": ["Wir untersuchen die Beziehung zwischen normalisierenden Flows und Variations- und Denoising-Autoencodern und schlagen ein neues Modell vor, das diese verallgemeinert."]} +{"source": "We investigate methods to efficiently learn diverse strategies in reinforcement learning for a generative structured prediction problem: query reformulation.In the proposed framework an agent consists of multiple specialized sub-agents and a meta-agent that learns to aggregate the answers from sub-agents to produce a final answer.Sub-agents are trained on disjoint partitions of the training data, while the meta-agent is trained on the full training set.Our method makes learning faster, because it is highly parallelizable, and has better generalization performance than strong baselines, such asan ensemble of agents trained on the full data.We evaluate on the tasks of document retrieval and question answering.Theimproved performance seems due to the increased diversity of reformulation strategies.This suggests that multi-agent, hierarchical approaches might play an important role in structured prediction tasks of this kind.However, we also find that it is not obvious how to characterize diversity in this context, and a first attempt based on clustering did not produce good results.Furthermore, reinforcement learning for the reformulation task is hard in high-performance regimes.At best, it only marginally improves over the state of the art, which highlights the complexity of training models in this framework for end-to-end language understanding problems.", "target": ["Wir verwenden Reinforcement Learning für die Umformulierung von Anfragen bei zwei Aufgaben und stellen überraschenderweise fest, dass beim Training mehrerer Agenten die Vielfalt der Umformulierungen wichtiger ist als die Spezialisierung."]} +{"source": "We study continuous action reinforcement learning problems in which it is crucial that the agent interacts with the environment only through safe policies, i.e.,~policies that keep the agent in desirable situations, both during training and at convergence.We formulate these problems as {\\em constrained} Markov decision processes (CMDPs) and present safe policy optimization algorithms that are based on a Lyapunov approach to solve them.Our algorithms can use any standard policy gradient (PG) method, such as deep deterministic policy gradient (DDPG) or proximal policy optimization (PPO), to train a neural network policy, while guaranteeing near-constraint satisfaction for every policy update by projecting either the policy parameter or the selected action onto the set of feasible solutions induced by the state-dependent linearized Lyapunov constraints.Compared to the existing constrained PG algorithms, ours are more data efficient as they are able to utilize both on-policy and off-policy data.Moreover, our action-projection algorithm often leads to less conservative policy updates and allows for natural integration into an end-to-end PG training pipeline.We evaluate our algorithms and compare them with the state-of-the-art baselines on several simulated (MuJoCo) tasks, as well as a real-world robot obstacle-avoidance problem, demonstrating their effectiveness in terms of balancing performance and constraint satisfaction.", "target": ["Ein allgemeiner Rahmen für die Einbeziehung langfristiger Sicherheitseinschränkungen in regelbasiertes Reinforcement Learning"]} +{"source": "Generative networks are known to be difficult to assess.Recent works on generative models, especially on generative adversarial networks, produce nice samples of varied categories of images.But the validation of their quality is highly dependent on the method used.A good generator should generate data which contain meaningful and varied information and that fit the distribution of a dataset.This paper presents a new method to assess a generator.Our approach is based on training a classifier with a mixture of real and generated samples.We train a generative model over a labeled training set, then we use this generative model to sample new data points that we mix with the original training data.This mixture of real and generated data is thus used to train a classifier which is afterwards tested on a given labeled test dataset.We compare this result with the score of the same classifier trained on the real training data mixed with noise.By computing the classifier's accuracy with different ratios of samples from both distributions (real and generated) we are able to estimate if the generator successfully fits and is able to generalize the distribution of the dataset.Our experiments compare the result of different generators from the VAE and GAN framework on MNIST and fashion MNIST dataset.", "target": ["Bewertung von generativen Netzen durch ihre Fähigkeit zur Datenerweiterung bei diskriminativen Modellen."]} +{"source": "We propose Automating Science Journalism (ASJ), the process of producing a press release from a scientific paper, as a novel task that can serve as a new benchmark for neural abstractive summarization.ASJ is a challenging task as it requires long source texts to be summarized to long target texts, while also paraphrasing complex scientific concepts to be understood by the general audience.For this purpose, we introduce a specialized dataset for ASJ that contains scientific papers and their press releases from Science Daily.While state-of-the-art sequence-to-sequence (seq2seq) models could easily generate convincing press releases for ASJ, these are generally nonfactual and deviate from the source.To address this issue, we improve seq2seq generation via transfer learning by co-training with new targets:(i) scientific abstracts of sources and(ii) partitioned press releases.We further design a measure for factuality that scores how pertinent to the scientific papers the press releases under our seq2seq models are.Our quantitative and qualitative evaluation shows sizable improvements over a strong baseline, suggesting that the proposed framework could improve seq2seq summarization beyond ASJ.", "target": ["Neu: Anwendung der seq2seq-Modellierung zur Automatisierung des Wissenschaftsjournalismus; hoch abstrakter Datensatz; Transfer-Learning-Tricks; automatisches Bewertungsmaß."]} +{"source": "The interpretability of an AI agent's behavior is of utmost importance for effective human-AI interaction.To this end, there has been increasing interest in characterizing and generating interpretable behavior of the agent.An alternative approach to guarantee that the agent generates interpretable behavior would be to design the agent's environment such that uninterpretable behaviors are either prohibitively expensive or unavailable to the agent.To date, there has been work under the umbrella of goal or plan recognition design exploring this notion of environment redesign in some specific instances of interpretable of behavior.In this position paper, we scope the landscape of interpretable behavior and environment redesign in all its different flavors.Specifically, we focus on three specific types of interpretable behaviors -- explicability, legibility, and predictability -- and present a general framework for the problem of environment design that can be instantiated to achieve each of the three interpretable behaviors.We also discuss how specific instantiations of this framework correspond to prior works on environment design and identify exciting opportunities for future work.", "target": ["Wir stellen einen Ansatz vor, um die Umgebung so umzugestalten, dass uninterpretierbares Agentenverhalten minimiert oder eliminiert wird."]} +{"source": "Inference models, which replace an optimization-based inference procedure with a learned model, have been fundamental in advancing Bayesian deep learning, the most notable example being variational auto-encoders (VAEs).In this paper, we propose iterative inference models, which learn how to optimize a variational lower bound through repeatedly encoding gradients.Our approach generalizes VAEs under certain conditions, and by viewing VAEs in the context of iterative inference, we provide further insight into several recent empirical findings.We demonstrate the inference optimization capabilities of iterative inference models, explore unique aspects of these models, and show that they outperform standard inference models on typical benchmark data sets.", "target": ["Wir schlagen eine neue Klasse von Inferenzmodellen vor, die iterativ Gradienten kodieren, um ungefähre nachträgliche Verteilungen zu schätzen."]} +{"source": "In artificial neural networks trained with gradient descent, the weights used for processing stimuli are also used during backward passes to calculate gradients.For the real brain to approximate gradients, gradient information would have to be propagated separately, such that one set of synaptic weights is used for processing and another set is used for backward passes.This produces the so-called \"weight transport problem\" for biological models of learning, where the backward weights used to calculate gradients need to mirror the forward weights used to process stimuli.This weight transport problem has been considered so hard that popular proposals for biological learning assume that the backward weights are simply random, as in the feedback alignment algorithm.However, such random weights do not appear to work well for large networks.Here we show how the discontinuity introduced in a spiking system can lead to a solution to this problem.The resulting algorithm is a special case of an estimator used for causal inference in econometrics, regression discontinuity design.We show empirically that this algorithm rapidly makes the backward weights approximate the forward weights.As the backward weights become correct, this improves learning performance over feedback alignment on tasks such as Fashion-MNIST and CIFAR-10.Our results demonstrate that a simple learning rule in a spiking network can allow neurons to produce the right backward connections and thus solve the weight transport problem.", "target": ["Wir stellen eine Lernregel für Rückkopplungsgewichte in einem spikenden neuronalen Netz vor, die das Problem des Gewichtstransports angeht."]} +{"source": "Variational inference (VI) methods and especially variational autoencoders (VAEs) specify scalable generative models that enjoy an intuitive connection to manifold learning --- with many default priors the posterior/likelihood pair $q(z|x)$/$p(x|z)$ can be viewed as an approximate homeomorphism (and its inverse) between the data manifold and a latent Euclidean space.However, these approximations are well-documented to become degenerate in training.Unless the subjective prior is carefully chosen, the topologies of the prior and data distributions often will not match.Conversely, diffusion maps (DM) automatically \\textit{infer} the data topology and enjoy a rigorous connection to manifold learning, but do not scale easily or provide the inverse homeomorphism.In this paper, we propose \\textbf{a)} a principled measure for recognizing the mismatch between data and latent distributions and \\textbf{b)} a method that combines the advantages of variational inference and diffusion maps to learn a homeomorphic generative model.The measure, the \\textit{locally bi-Lipschitz property}, is a sufficient condition for a homeomorphism and easy to compute and interpret.The method, the \\textit{variational diffusion autoencoder} (VDAE), is a novel generative algorithm that first infers the topology of the data distribution, then models a diffusion random walk over the data.To achieve efficient computation in VDAEs, we use stochastic versions of both variational inference and manifold learning optimization.We prove approximation theoretic results for the dimension dependence of VDAEs, and that locally isotropic sampling in the latent space results in a random walk over the reconstructed manifold.Finally, we demonstrate the utility of our method on various real and synthetic datasets, and show that it exhibits performance superior to other generative models.", "target": ["Wir kombinieren Variationsinferenz und vielfältiges Lernen (insbesondere VAEs und Diffusionskarten), um ein generatives Modell zu erstellen, das auf einem zufälligen Diffusions-Walk auf einer Datenvielfalt basiert; wir erzeugen Beispiele, indem wir aus der stationären Verteilung des Walks ziehen."]} +{"source": "While deep learning and deep reinforcement learning systems have demonstrated impressive results in domains such as image classification, game playing, and robotic control, data efficiency remains a major challenge, particularly as these algorithms learn individual tasks from scratch.Multi-task learning has emerged as a promising approach for sharing structure across multiple tasks to enable more efficient learning.However, the multi-task setting presents a number of optimization challenges, making it difficult to realize large efficiency gains compared to learning tasks independently.The reasons why multi-task learning is so challenging compared to single task learning are not fully understood.Motivated by the insight that gradient interference causes optimization challenges, we develop a simple and general approach for avoiding interference between gradients from different tasks, by altering the gradients through a technique we refer to as “gradient surgery”.We propose a form of gradient surgery that projects the gradient of a task onto the normal plane of the gradient of any other task that has a conflicting gradient.On a series of challenging multi-task supervised and multi-task reinforcement learning problems, we find that this approach leads to substantial gains in efficiency and performance. Further, it can be effectively combined with previously-proposed multi-task architectures for enhanced performance in a model-agnostic way.", "target": ["Wir entwickeln einen einfachen und allgemeinen Ansatz zur Vermeidung von Interferenzen zwischen Gradienten aus verschiedenen Aufgaben, der die Leistung von Multi-Task-Lernen sowohl im Bereich des überwachten als auch des Reinforcement Learnings verbessert."]} +{"source": "In this paper we propose to view the acceptance rate of the Metropolis-Hastings algorithm as a universal objective for learning to sample from target distribution -- given either as a set of samples or in the form of unnormalized density.This point of view unifies the goals of such approaches as Markov Chain Monte Carlo (MCMC), Generative Adversarial Networks (GANs), variational inference.To reveal the connection we derive the lower bound on the acceptance rate and treat it as the objective for learning explicit and implicit samplers.The form of the lower bound allows for doubly stochastic gradient optimization in case the target distribution factorizes (i.e. over data points).We empirically validate our approach on Bayesian inference for neural networks and generative models for images.", "target": ["Lernen, zu prüfen, indem die Akzeptanzrate des Metropolis-Hastings-Algorithmus nach unten begrenzt wird"]} +{"source": "This paper proposes a self-supervised learning approach for video features that results in significantly improved performance on downstream tasks (such as video classification, captioning and segmentation) compared to existing methods.Our method extends the BERT model for text sequences to the case of sequences of real-valued feature vectors, by replacing the softmax loss with noise contrastive estimation (NCE).We also show how to learn representations from sequences of visual features and sequences of words derived from ASR (automatic speech recognition), and show that such cross-modal training (when possible) helps even more.", "target": ["Verallgemeinerter BERT für kontinuierliche und cross-modale Eingaben; modernste selbstüberwachte Videodarstellungen."]} +{"source": "We present a generic dynamic architecture that employs a problem specific differentiable forking mechanism to leverage discrete logical information about the problem data structure.We adapt and apply our model to CLEVR Visual Question Answering, giving rise to the DDRprog architecture; compared to previous approaches, our model achieves higher accuracy in half as many epochs with five times fewer learnable parameters.Our model directly models underlying question logic using a recurrent controller that jointly predicts and executes functional neural modules; it explicitly forks subprocesses to handle logical branching.While FiLM and other competitive models are static architectures with less supervision, we argue that inclusion of program labels enables learning of higher level logical operations -- our architecture achieves particularly high performance on questions requiring counting and integer comparison. We further demonstrate the generality of our approach though DDRstack -- an application of our method to reverse Polish notation expression evaluation in which the inclusion of a stack assumption allows our approach to generalize to long expressions, significantly outperforming an LSTM with ten times as many learnable parameters.", "target": ["Eine generische dynamische Architektur, die einen problemspezifischen differenzierbaren Forking-Mechanismus zur Kodierung harter Datenstrukturannahmen einsetzt. Angewandt auf CLEVR VQA und Ausdrucksbewertung."]} +{"source": "We propose Support-guided Adversarial Imitation Learning (SAIL), a generic imitation learning framework that unifies support estimation of the expert policy with the family of Adversarial Imitation Learning (AIL) algorithms.SAIL addresses two important challenges of AIL, including the implicit reward bias and potential training instability.We also show that SAIL is at least as efficient as standard AIL.In an extensive evaluation, we demonstrate that the proposed method effectively handles the reward bias and achieves better performance and training stability than other baseline methods on a wide range of benchmark control tasks.", "target": ["Wir vereinen die Support-Schätzung mit der Familie der Adversarial Imitation Learning Algorithmen zu Support-guided Adversarial Imitation Learning, einem robusteren und stabileren Imitations-Lernsystem."]} +{"source": "We consider the task of few shot link prediction, where the goal is to predict missing edges across multiple graphs using only a small sample of known edges.We show that current link prediction methods are generally ill-equipped to handle this task---as they cannot effectively transfer knowledge between graphs in a multi-graph setting and are unable to effectively learn from very sparse data.To address this challenge, we introduce a new gradient-based meta learning framework, Meta-Graph, that leverages higher-order gradients along with a learned graph signature function that conditionally generates a graph neural network initialization.Using a novel set of few shot link prediction benchmarks, we show that Meta-Graph enables not only fast adaptation but also better final convergence and can effectively learn using only a small sample of true edges.", "target": ["Wir wenden gradientenbasiertes Meta-Lernen auf den Bereich der Graphen an und führen eine neue graphen-spezifische Transferfunktion ein, um den Prozess weiter zu beschleunigen."]} +{"source": "Generative neural networks map a standard, possibly distribution to a complex high-dimensional distribution, which represents the real world data set.However, a determinate input distribution as well as a specific architecture of neural networks may impose limitations on capturing the diversity in the high dimensional target space.To resolve this difficulty, we propose a training framework that greedily produce a series of generative adversarial networks that incrementally capture the diversity of the target space.We show theoretically and empirically that our training algorithm converges to the theoretically optimal distribution, the projection of the real distribution onto the convex hull of the network's distribution space.", "target": ["Wir schlagen eine neue Methode zum inkrementellen Training eines gemischt generativen Modells vor, um die Informationsprojektion der realen Datenverteilung zu approximieren."]} +{"source": "Generative priors have become highly effective in solving inverse problems including denoising, inpainting, and reconstruction from few and noisy measurements.With a generative model we can represent an image with a much lower dimensional latent codes.In the context of compressive sensing, if the unknown image belongs to the range of a pretrained generative network, then we can recover the image by estimating the underlying compact latent code from the available measurements.However, recent studies revealed that even untrained deep neural networks can work as a prior for recovering natural images.These approaches update the network weights keeping latent codes fixed to reconstruct the target image from the given measurements.In this paper, we optimize over network weights and latent codes to use untrained generative network as prior for video compressive sensing problem.We show that by optimizing over latent code, we can additionally get concise representation of the frames which retain the structural similarity of the video frames.We also apply low-rank constraint on the latent codes to represent the video sequences in even lower dimensional latent space.We empirically show that our proposed methods provide better or comparable accuracy and low computational complexity compared to the existing methods.", "target": ["Wiederherstellung von Videos aus komprimierten Messungen durch Erlernen einer niedrigdimensionalen Darstellung (mit niedrigem Rang) direkt aus den Messungen beim Training eines tiefen Generators."]} +{"source": "Magnitude-based pruning is one of the simplest methods for pruning neural networks.Despite its simplicity, magnitude-based pruning and its variants demonstrated remarkable performances for pruning modern architectures.Based on the observation that the magnitude-based pruning indeed minimizes the Frobenius distortion of a linear operator corresponding to a single layer, we develop a simple pruning method, coined lookahead pruning, by extending the single layer optimization to a multi-layer optimization.Our experimental results demonstrate that the proposed method consistently outperforms the magnitude pruning on various networks including VGG and ResNet, particularly in the high-sparsity regime.", "target": ["Wir untersuchen eine mehrschichtige Verallgemeinerung des magnitudenbasierten Prunings."]} +{"source": "Recent literature has demonstrated promising results on the training of Generative Adversarial Networks by employing a set of discriminators, as opposed to the traditional game involving one generator against a single adversary.Those methods perform single-objective optimization on some simple consolidation of the losses, e.g. an average.In this work, we revisit the multiple-discriminator approach by framing the simultaneous minimization of losses provided by different models as a multi-objective optimization problem.Specifically, we evaluate the performance of multiple gradient descent and the hypervolume maximization algorithm on a number of different datasets.Moreover, we argue that the previously proposed methods and hypervolume maximization can all be seen as variations of multiple gradient descent in which the update direction computation can be done efficiently.Our results indicate that hypervolume maximization presents a better compromise between sample quality and diversity, and computational cost than previous methods.", "target": ["Wir führen die Hypervolumen-Maximierung für das Training von GANs mit mehreren Diskriminatoren ein und zeigen Leistungsverbesserungen in Bezug auf die Qualität und Vielfalt der Beispiele."]} +{"source": "Designing of search space is a critical problem for neural architecture search (NAS) algorithms.We propose a fine-grained search space comprised of atomic blocks, a minimal search unit much smaller than the ones used in recent NAS algorithms.This search space facilitates direct selection of channel numbers and kernel sizes in convolutions.In addition, we propose a resource-aware architecture search algorithm which dynamically selects atomic blocks during training.The algorithm is further accelerated by a dynamic network shrinkage technique.Instead of a search-and-retrain two-stage paradigm, our method can simultaneously search and train the target architecture in an end-to-end manner. Our method achieves state-of-the-art performance under several FLOPS configurations on ImageNet with a negligible searching cost.We open our entire codebase at: https://github.com/meijieru/AtomNAS.", "target": ["Ein neuer Stand der Technik auf Imagenet für mobile Umgebungen"]} +{"source": "We introduce Lyceum, a high-performance computational ecosystem for robotlearning. Lyceum is built on top of the Julia programming language and theMuJoCo physics simulator, combining the ease-of-use of a high-level program-ming language with the performance of native C. Lyceum is up to 10-20Xfaster compared to other popular abstractions like OpenAI’sGymand Deep-Mind’sdm-control. This substantially reduces training time for various re-inforcement learning algorithms; and is also fast enough to support real-timemodel predictive control with physics simulators. Lyceum has a straightfor-ward API and supports parallel computation across multiple cores or machines.The code base, tutorials, and demonstration videos can be found at: https://sites.google.com/view/lyceum-anon.", "target": ["Ein Rahmenwerk für Hochleistungs-Robotiksimulation und Algorithmenentwicklung."]} +{"source": "There is a strong incentive to develop versatile learning techniques that can transfer the knowledge of class-separability from a labeled source domain to an unlabeled target domain in the presence of a domain-shift.Existing domain adaptation (DA) approaches are not equipped for practical DA scenarios as a result of their reliance on the knowledge of source-target label-set relationship (e.g. Closed-set, Open-set or Partial DA).Furthermore, almost all the prior unsupervised DA works require coexistence of source and target samples even during deployment, making them unsuitable for incremental, real-time adaptation.Devoid of such highly impractical assumptions, we propose a novel two-stage learning process.Initially, in the procurement-stage, the objective is to equip the model for future source-free deployment, assuming no prior knowledge of the upcoming category-gap and domain-shift.To achieve this, we enhance the model’s ability to reject out-of-source distribution samples by leveraging the available source data, in a novel generative classifier framework.Subsequently, in the deployment-stage, the objective is to design a unified adaptation algorithm capable of operating across a wide range of category-gaps, with no access to the previously seen source samples.To achieve this, in contrast to the usage of complex adversarial training regimes, we define a simple yet effective source-free adaptation objective by utilizing a novel instance-level weighing mechanism, named as Source Similarity Metric (SSM).A thorough evaluation shows the practical usability of the proposed learning framework with superior DA performance even over state-of-the-art source-dependent approaches.", "target": ["Ein neuartiges Paradigma der unüberwachten Domänenanpassung - Durchführung der Anpassung ohne Zugriff auf die Quelldaten (\"quellenfrei\") und ohne Annahmen über die Quelle-Ziel-Kategorienlücke (\"universell\")."]} +{"source": "One of the long-standing challenges in Artificial Intelligence for learning goal-directed behavior is to build a single agent which can solve multiple tasks.Recent progress in multi-task learning for goal-directed sequential problems has been in the form of distillation based learning wherein a student network learns from multiple task-specific expert networks by mimicking the task-specific policies of the expert networks.While such approaches offer a promising solution to the multi-task learning problem, they require supervision from large expert networks which require extensive data and computation time for training.In this work, we propose an efficient multi-task learning framework which solves multiple goal-directed tasks in an on-line setup without the need for expert supervision.Our work uses active learning principles to achieve multi-task learning by sampling the harder tasks more than the easier ones.We propose three distinct models under our active sampling framework.An adaptive method with extremely competitive multi-tasking performance.A UCB-based meta-learner which casts the problem of picking the next task to train on as a multi-armed bandit problem.A meta-learning method that casts the next-task picking problem as a full Reinforcement Learning problem and uses actor-critic methods for optimizing the multi-tasking performance directly.We demonstrate results in the Atari 2600 domain on seven multi-tasking instances: three 6-task instances, one 8-task instance, two 12-task instances and one 21-task instance.", "target": ["Lässt man einen Meta-Learner entscheiden, welche Aufgabe ein Agent in einer Multitasking-Umgebung trainieren soll, verbessert sich die Multitasking-Fähigkeit erheblich"]} +{"source": "Numerous machine reading comprehension (MRC) datasets often involve manual annotation, requiring enormous human effort, and hence the size of the dataset remains significantly smaller than the size of the data available for unsupervised learning.Recently, researchers proposed a model for generating synthetic question-and-answer data from large corpora such as Wikipedia.This model is utilized to generate synthetic data for training an MRC model before fine-tuning it using the original MRC dataset.This technique shows better performance than other general pre-training techniques such as language modeling, because the characteristics of the generated data are similar to those of the downstream MRC data.However, it is difficult to have high-quality synthetic data comparable to human-annotated MRC datasets.To address this issue, we propose Answer-containing Sentence Generation (ASGen), a novel pre-training method for generating synthetic data involving two advanced techniques, (1) dynamically determining K answers and (2) pre-training the question generator on the answer-containing sentence generation task.We evaluate the question generation capability of our method by comparing the BLEU score with existing methods and test our method by fine-tuning the MRC model on the downstream MRC data after training on synthetic data.Experimental results show that our approach outperforms existing generation methods and increases the performance of the state-of-the-art MRC models across a range of MRC datasets such as SQuAD-v1.1, SQuAD-v2.0, KorQuAD and QUASAR-T without any architectural modifications to the original MRC model.", "target": ["Wir schlagen Answer-containing Sentence Generation (ASGen) vor, eine neuartige Pre-Training-Methode zur Erzeugung synthetischer Daten für das maschinelle Leseverstehen."]} +{"source": "Deep neural networks (DNNs) dominate current research in machine learning.Due to massive GPU parallelization DNN training is no longer a bottleneck, and large models with many parameters and high computational effort lead common benchmark tables.In contrast, embedded devices have a very limited capability.As a result, both model size and inference time must be significantly reduced if DNNs are to achieve suitable performance on embedded devices.We propose a soft quantization approach to train DNNs that can be evaluated using pure fixed-point arithmetic.By exploiting the bit-shift mechanism, we derive fixed-point quantization constraints for all important components, including batch normalization and ReLU.Compared to floating-point arithmetic, fixed-point calculations significantly reduce computational effort whereas low-bit representations immediately decrease memory costs.We evaluate our approach with different architectures on common benchmark data sets and compare with recent quantization approaches.We achieve new state of the art performance using 4-bit fixed-point models with an error rate of 4.98% on CIFAR-10.", "target": ["Soft-Quantisierungs-Ansatz zum Erlernen reiner Festkomma-Darstellungen tiefer neuronaler Netze"]} +{"source": "We present a new approach and a novel architecture, termed WSNet, for learning compact and efficient deep neural networks.Existing approaches conventionally learn full model parameters independently and then compress them via \\emph{ad hoc} processing such as model pruning or filter factorization.Alternatively, WSNet proposes learning model parameters by sampling from a compact set of learnable parameters, which naturally enforces {parameter sharing} throughout the learning process.We demonstrate that such a novel weight sampling approach (and induced WSNet) promotes both weights and computation sharing favorably.By employing this method, we can more efficiently learn much smaller networks with competitive performance compared to baseline networks with equal numbers of convolution filters.Specifically, we consider learning compact and efficient 1D convolutional neural networks for audio classification.Extensive experiments on multiple audio classification datasets verify the effectiveness of WSNet.Combined with weight quantization, the resulted models are up to \\textbf{180$\\times$} smaller and theoretically up to \\textbf{16$\\times$} faster than the well-established baselines, without noticeable performance drop.", "target": ["Wir präsentieren eine neuartige Netzwerkarchitektur für das Lernen kompakter und effizienter tiefer neuronaler Netze"]} +{"source": "Recent work in adversarial machine learning started to focus on the visual perception in autonomous driving and studied Adversarial Examples (AEs) for object detection models.However, in such visual perception pipeline the detected objects must also be tracked, in a process called Multiple Object Tracking (MOT), to build the moving trajectories of surrounding obstacles.Since MOT is designed to be robust against errors in object detection, it poses a general challenge to existing attack techniques that blindly target objection detection: we find that a success rate of over 98% is needed for them to actually affect the tracking results, a requirement that no existing attack technique can satisfy.In this paper, we are the first to study adversarial machine learning attacks against the complete visual perception pipeline in autonomous driving, and discover a novel attack technique, tracker hijacking, that can effectively fool MOT using AEs on object detection.Using our technique, successful AEs on as few as one single frame can move an existing object in to or out of the headway of an autonomous vehicle to cause potential safety hazards.We perform evaluation using the Berkeley Deep Drive dataset and find that on average when 3 frames are attacked, our attack can have a nearly 100% success rate while attacks that blindly target object detection only have up to 25%.", "target": ["Wir untersuchen zum ersten Mal die Angriffe auf die Mechanismen des Multiple Object Tracking durch maschinelles Lernen."]} +{"source": "Self-supervised learning (SlfSL), aiming at learning feature representations through ingeniously designed pretext tasks without human annotation, has achieved compelling progress in the past few years.Very recently, SlfSL has also been identified as a promising solution for semi-supervised learning (SemSL) since it offers a new paradigm to utilize unlabeled data.This work further explores this direction by proposing a new framework to seamlessly couple SlfSL with SemSL.Our insight is that the prediction target in SemSL can be modeled as the latent factor in the predictor for the SlfSL target.Marginalizing over the latent factor naturally derives a new formulation which marries the prediction targets of these two learning processes.By implementing this framework through a simple-but-effective SlfSL approach -- rotation angle prediction, we create a new SemSL approach called Conditional Rotation Angle Prediction (CRAP).Specifically, CRAP is featured by adopting a module which predicts the image rotation angle \\textbf{conditioned on the candidate image class}.Through experimental evaluation, we show that CRAP achieves superior performance over the other existing ways of combining SlfSL and SemSL.Moreover, the proposed SemSL framework is highly extendable.By augmenting CRAP with a simple SemSL technique and a modification of the rotation angle prediction task, our method has already achieved the state-of-the-art SemSL performance.", "target": ["Kopplung des halb-überwachten Lernens mit dem selbst-überwachten Lernen und explizite Modellierung der selbst-überwachten Aufgabe unter Berücksichtigung der halb-überwachten Aufgabe"]} +{"source": "Ranking is a central task in machine learning and information retrieval.In this task, it is especially important to present the user with a slate of items that is appealing as a whole.This in turn requires taking into account interactions between items, since intuitively, placing an item on the slate affects the decision of which other items should be chosen alongside it.In this work, we propose a sequence-to-sequence model for ranking called seq2slate.At each step, the model predicts the next item to place on the slate given the items already chosen.The recurrent nature of the model allows complex dependencies between items to be captured directly in a flexible and scalable way.We show how to learn the model end-to-end from weak supervision in the form of easily obtained click-through data.We further demonstrate the usefulness of our approach in experiments on standard ranking benchmarks as well as in a real-world recommendation system.", "target": ["Eine Zeigernetzwerk-Architektur für die Neubewertung von Artikeln, die aus Klickprotokollen gelernt wurde."]} +{"source": "While momentum-based methods, in conjunction with the stochastic gradient descent, are widely used when training machine learning models, there is little theoretical understanding on the generalization error of such methods.In practice, the momentum parameter is often chosen in a heuristic fashion with little theoretical guidance.In this work, we use the framework of algorithmic stability to provide an upper-bound on the generalization error for the class of strongly convex loss functions, under mild technical assumptions.Our bound decays to zero inversely with the size of the training set, and increases as the momentum parameter is increased.We also develop an upper-bound on the expected true risk, in terms of the number of training steps, the size of the training set, and the momentum parameter.", "target": ["Stochastische Gradientenmethode mit Momentum generalisiert."]} +{"source": "Imitation learning from demonstrations usually relies on learning a policy from trajectories of optimal states and actions.However, in real life expert demonstrations, often the action information is missing and only state trajectories are available.We present a model-based imitation learning method that can learn environment-specific optimal actions only from expert state trajectories.Our proposed method starts with a model-free reinforcement learning algorithm with a heuristic reward signal to sample environment dynamics, which is then used to train the state-transition probability.Subsequently, we learn the optimal actions from expert state trajectories by supervised learning, while back-propagating the error gradients through the modeled environment dynamics.Experimental evaluations show that our proposed method successfully achieves performance similar to (state, action) trajectory-based traditional imitation learning methods even in the absence of action information, with much fewer iterations compared to conventional model-free reinforcement learning methods.We also demonstrate that our method can learn to act from only video demonstrations of expert agent for simple games and can learn to achieve desired performance in less number of iterations.", "target": ["Lernen, einen Experten zu imitieren, wenn es keine optimalen Aktionen gibt, indem man ein dynamisches Modell lernt, während man die Umgebung erkundet."]} +{"source": "Recent research has proposed the lottery ticket hypothesis, suggesting that for a deep neural network, there exist trainable sub-networks performing equally or better than the original model with commensurate training steps.While this discovery is insightful, finding proper sub-networks requires iterative training and pruning.The high cost incurred limits the applications of the lottery ticket hypothesis.We show there exists a subset of the aforementioned sub-networks that converge significantly faster during the training process and thus can mitigate the cost issue.We conduct extensive experiments to show such sub-networks consistently exist across various model structures for a restrictive setting of hyperparameters (e.g., carefully selected learning rate, pruning ratio, and model capacity). As a practical application of our findings, we demonstrate that such sub-networks can help in cutting down the total time of adversarial training, a standard approach to improve robustness, by up to 49% on CIFAR-10 to achieve the state-of-the-art robustness.", "target": ["Wir zeigen die Möglichkeit des Pruning, um ein kleines Teilnetz mit deutlich höherer Konvergenzrate als das vollständige Modell zu finden."]} +{"source": "Disentangled representations, where the higher level data generative factors are reflected in disjoint latent dimensions, offer several benefits such as ease of deriving invariant representations, transferability to other tasks, interpretability, etc.We consider the problem of unsupervised learning of disentangled representations from large pool of unlabeled observations, and propose a variational inference based approach to infer disentangled latent factors.We introduce a regularizer on the expectation of the approximate posterior over observed data that encourages the disentanglement.We also propose a new disentanglement metric which is better aligned with the qualitative disentanglement observed in the decoder's output.We empirically observe significant improvement over existing methods in terms of both disentanglement and data likelihood (reconstruction quality).", "target": ["Wir schlagen einen auf Variationsinferenz basierenden Ansatz vor, um die Inferenz von entwirrten Latenzen zu fördern. Außerdem schlagen wir eine neue Metrik zur Quantifizierung der Entflechtung vor."]} +{"source": "In multiagent systems (MASs), each agent makes individual decisions but all of them contribute globally to the system evolution.Learning in MASs is difficult since each agent's selection of actions must take place in the presence of other co-learning agents.Moreover, the environmental stochasticity and uncertainties increase exponentially with the increase in the number of agents.Previous works borrow various multiagent coordination mechanisms into deep learning architecture to facilitate multiagent coordination.However, none of them explicitly consider action semantics between agents that different actions have different influences on other agents.In this paper, we propose a novel network architecture, named Action Semantics Network (ASN), that explicitly represents such action semantics between agents.ASN characterizes different actions' influence on other agents using neural networks based on the action semantics between them.ASN can be easily combined with existing deep reinforcement learning (DRL) algorithms to boost their performance.Experimental results on StarCraft II micromanagement and Neural MMO show ASN significantly improves the performance of state-of-the-art DRL approaches compared with several network architectures.", "target": ["Die von uns vorgeschlagene ASN charakterisiert den Einfluss verschiedener Aktionen auf andere Agenten mit Hilfe neuronaler Netze, die auf der Aktionssemantik zwischen ihnen basieren."]} +{"source": "Spiking neural networks are being investigated both as biologically plausible models of neural computation and also as a potentially more efficient type of neural network.While convolutional spiking neural networks have been demonstrated to achieve near state-of-the-art performance, only one solution has been proposed to convert gated recurrent neural networks, so far.Recurrent neural networks in the form of networks of gating memory cells have been central in state-of-the-art solutions in problem domains that involve sequence recognition or generation.Here, we design an analog gated LSTM cell where its neurons can be substituted for efficient stochastic spiking neurons.These adaptive spiking neurons implement an adaptive form of sigma-delta coding to convert internally computed analog activation values to spike-trains.For such neurons, we approximate the effective activation function, which resembles a sigmoid.We show how analog neurons with such activation functions can be used to create an analog LSTM cell; networks of these cells can then be trained with standard backpropagation.We train these LSTM networks on a noisy and noiseless version of the original sequence prediction task from Hochreiter & Schmidhuber (1997), and also on a noisy and noiseless version of a classical working memory reinforcement learning task, the T-Maze.Substituting the analog neurons for corresponding adaptive spiking neurons, we then show that almost all resulting spiking neural network equivalents correctly compute the original tasks.", "target": ["Wir demonstrieren ein gated recurrent asynchronous spiking neural Network, das einer LSTM-Einheit entspricht."]} +{"source": "CNNs are widely successful in recognizing human actions in videos, albeit with a great cost of computation.This cost is significantly higher in the case of long-range actions, where a video can span up to a few minutes, on average.The goal of this paper is to reduce the computational cost of these CNNs, without sacrificing their performance.We propose VideoEpitoma, a neural network architecture comprising two modules: a timestamp selector and a video classifier.Given a long-range video of thousands of timesteps, the selector learns to choose only a few but most representative timesteps for the video.This selector resides on top of a lightweight CNN such as MobileNet and uses a novel gating module to take a binary decision: consider or discard a video timestep.This decision is conditioned on both the timestep-level feature and the video-level consensus.A heavyweight CNN model such as I3D takes the selected frames as input and performs video classification.Using off-the-shelf video classifiers, VideoEpitoma reduces the computation by up to 50\\% without compromising the accuracy.In addition, we show that if trained end-to-end, the selector learns to make better choices to the benefit of the classifier, despite the selector and the classifier residing on two different CNNs.Finally, we report state-of-the-art results on two datasets for long-range action recognition: Charades and Breakfast Actions, with much-reduced computation.In particular, we match the accuracy of I3D by using less than half of the computation.", "target": ["Effiziente Videoklassifizierung mit einem Frame-basierten Conditional-Gating-Modul zur Auswahl der dominantesten Frames, gefolgt von einer zeitlichen Modellierung und einem Klassifikator."]} +{"source": "Human annotation for syntactic parsing is expensive, and large resources are available only for a fraction of languages.A question we ask is whether one can leverage abundant unlabeled texts to improve syntactic parsers, beyond just using the texts to obtain more generalisable lexical features (i.e. beyond word embeddings).To this end, we propose a novel latent-variable generative model for semi-supervised syntactic dependency parsing.As exact inference is intractable, we introduce a differentiable relaxation to obtain approximate samples and compute gradients with respect to the parser parameters.Our method (Differentiable Perturb-and-Parse) relies on differentiable dynamic programming over stochastically perturbed edge scores.We demonstrate effectiveness of our approach with experiments on English, French and Swedish.", "target": ["Differenzierbare dynamische Programmierung über gestörte Eingangsgewichte mit Anwendung auf semi-supervised VAE"]} +{"source": "DeConvNet, Guided BackProp, LRP, were invented to better understand deep neural networks.We show that these methods do not produce the theoretically correct explanation for a linear model.Yet they are used on multi-layer networks with millions of parameters.This is a cause for concern since linear models are simple neural networks.We argue that explanation methods for neural nets should work reliably in the limit of simplicity, the linear models.Based on our analysis of linear models we propose a generalization that yields two explanation techniques (PatternNet and PatternAttribution) that are theoretically sound for linear models and produce improved explanations for deep networks.", "target": ["Ohne Lernen ist es unmöglich, die Entscheidungen eines maschinellen Lernmodells zu erklären."]} +{"source": "Graph neural networks have shown promising results on representing and analyzing diverse graph-structured data such as social, citation, and protein interaction networks.Existing approaches commonly suffer from the oversmoothing issue, regardless of whether policies are edge-based or node-based for neighborhood aggregation.Most methods also focus on transductive scenarios for fixed graphs, leading to poor generalization performance for unseen graphs.To address these issues, we propose a new graph neural network model that considers both edge-based neighborhood relationships and node-based entity features, i.e. Graph Entities with Step Mixture via random walk (GESM).GESM employs a mixture of various steps through random walk to alleviate the oversmoothing problem and attention to use node information explicitly.These two mechanisms allow for a weighted neighborhood aggregation which considers the properties of entities and relations.With intensive experiments, we show that the proposed GESM achieves state-of-the-art or comparable performances on four benchmark graph datasets comprising transductive and inductive learning tasks.Furthermore, we empirically demonstrate the significance of considering global information.The source code will be publicly available in the near future.", "target": ["Einfaches und effektives neuronales Graphennetz mit einer Mischung aus Zufallsschritten und Aufmerksamkeit"]} +{"source": "Basis pursuit is a compressed sensing optimization in which the l1-norm is minimized subject to model error constraints.Here we use a deep neural network prior instead of l1-regularization.Using known noise statistics, we jointly learn the prior and reconstruct images without access to ground-truth data.During training, we use alternating minimization across an unrolled iterative network and jointly solve for the neural network weights and training set image reconstructions.At inference, we fix the weights and pass the measurements through the network.We compare reconstruction performance between unsupervised and supervised (i.e. with ground-truth) methods.We hypothesize this technique could be used to learn reconstruction when ground-truth data are unavailable, such as in high-resolution dynamic MRI.", "target": ["Wir stellen eine unbeaufsichtigte Deep Learning-Rekonstruktion für inverse Bildgebungsprobleme vor, die neuronale Netze mit modellbasierten Einschränkungen kombiniert."]} +{"source": "Deep learning approaches usually require a large amount of labeled data to generalize.However, humans can learn a new concept only by a few samples.One of the high cogntition human capablities is to learn several concepts at the same time.In this paper, we address the task of classifying multiple objects by seeing only a few samples from each category.To the best of authors' knowledge, there is no dataset specially designed for few-shot multiclass classification.We design a task of mutli-object few class classification and an environment for easy creating controllable datasets for this task.We demonstrate that the proposed dataset is sound using a method which is an extension of prototypical networks.", "target": ["Wir stellen eine diagnostische Aufgabe vor, die eine Variation des \"few-shot learning\" ist und stellen einen Datensatz dafür vor."]} +{"source": "We present a new approach to defining a sequence loss function to train a summarizer by using a secondary encoder-decoder as a loss function, alleviating a shortcoming of word level training for sequence outputs.The technique is based on the intuition that if a summary is a good one, it should contain the most essential information from the original article, and therefore should itself be a good input sequence, in lieu of the original, from which a summary can be generated.We present experimental results where we apply this additional loss function to a general abstractive summarizer on a news summarization dataset.The result is an improvement in the ROUGE metric and an especially large improvement in human evaluations, suggesting enhanced performance that is competitive with specialized state-of-the-art models.", "target": ["Wir stellen die Verwendung eines sekundären Kodierers/Dekodierers als Verlustfunktion vor, um das Training eines Zusammenfassers zu unterstützen."]} +{"source": "Existing unsupervised video-to-video translation methods fail to produce translated videos which are frame-wise realistic, semantic information preserving and video-level consistent.In this work, we propose a novel unsupervised video-to-video translation model.Our model decomposes the style and the content, uses specialized encoder-decoder structure and propagates the inter-frame information through bidirectional recurrent neural network (RNN) units.The style-content decomposition mechanism enables us to achieve long-term style-consistent video translation results as well as provides us with a good interface for modality flexible translation.In addition, by changing the input frames and style codes incorporated in our translation, we propose a video interpolation loss, which captures temporal information within the sequence to train our building blocks in a self-supervised manner.Our model can produce photo-realistic, spatio-temporal consistent translated videos in a multimodal way.Subjective and objective experimental results validate the superiority of our model over the existing methods.", "target": ["Ein zeitlich konsistentes und modalitätsflexibles unüberwachtes Video-zu-Video-Übersetzungssystem, das auf selbstüberwachte Weise trainiert wird."]} +{"source": "Providing transparency of AI planning systems is crucial for their success in practical applications.In order to create a transparent system, a user must be able to query it for explanations about its outputs.We argue that a key underlying principle for this is the use of causality within a planning model, and that argumentation frameworks provide an intuitive representation of such causality.In this paper, we discuss how argumentation can aid in extracting causalities in plans and models, and how they can create explanations from them.", "target": ["Argumentationsrahmen werden verwendet, um die Kausalität von Plänen/Modellen darzustellen, die für Erklärungen herangezogen werden sollen."]} +{"source": "Point clouds, as a form of Lagrangian representation, allow for powerful and flexible applications in a large number of computational disciplines.We propose a novel deep-learning method to learn stable and temporally coherent feature spaces for points clouds that change over time.We identify a set of inherent problems with these approaches: without knowledge of the time dimension, the inferred solutions can exhibit strong flickering, and easy solutions to suppress this flickering can result in undesirable local minima that manifest themselves as halo structures.We propose a novel temporal loss function that takes into account higher time derivatives of the point positions, and encourages mingling, i.e., to prevent the aforementioned halos.We combine these techniques in a super-resolution method with a truncation approach to flexibly adapt the size of the generated positions.We show that our method works for large, deforming point sets from different sources to demonstrate the flexibility of our approach.", "target": ["Wir schlagen einen generativen neuronalen Netzwerkansatz für zeitlich kohärente Punktansammlungen vor."]} +{"source": "We study the problem of generating adversarial examples in a black-box setting in which only loss-oracle access to a model is available.We introduce a framework that conceptually unifies much of the existing work on black-box attacks, and demonstrate that the current state-of-the-art methods are optimal in a natural sense.Despite this optimality, we show how to improve black-box attacks by bringing a new element into the problem: gradient priors.We give a bandit optimization-based algorithm that allows us to seamlessly integrate any such priors, and we explicitly identify and incorporate two examples.The resulting methods use two to four times fewer queries and fail two to five times less than the current state-of-the-art.The code for reproducing our work is available at https://git.io/fAjOJ.", "target": ["Wir stellen eine vereinheitlichende Sichtweise auf Black-Box-Angriffe als Gradientenschätzungsproblem vor und präsentieren dann einen Rahmen (basierend auf Bandits-Optimierung), um Prioritäten in die Gradientenschätzung zu integrieren, was zu einer signifikanten Leistungssteigerung führt."]} +{"source": "Collecting high-quality, large scale datasets typically requires significant resources.The aim of the present work is to improve the label efficiency of large neural networks operating on audio data through multitask learning with self-supervised tasks on unlabeled data.To this end, we trained an end-to-end audio feature extractor based on WaveNet that feeds into simple, yet versatile task-specific neural networks.We describe three self-supervised learning tasks that can operate on any large, unlabeled audio corpus.We demonstrate that, in a scenario with limited labeled training data, one can significantly improve the performance of a supervised classification task by simultaneously training it with these additional self-supervised tasks.We show that one can improve performance on a diverse sound events classification task by nearly 6\\% when jointly trained with up to three distinct self-supervised tasks.This improvement scales with the number of additional auxiliary tasks as well as the amount of unsupervised data.We also show that incorporating data augmentation into our multitask setting leads to even further gains in performance.", "target": ["Verbesserung der Kennzeichnungseffizienz durch Multitasking-Lernen bei auditiven Daten"]} +{"source": "Information need of humans is essentially multimodal in nature, enabling maximum exploitation of situated context.We introduce a dataset for sequential procedural (how-to) text generation from images in cooking domain.The dataset consists of 16,441 cooking recipes with 160,479 photos associated with different steps.We setup a baseline motivated by the best performing model in terms of human evaluation for the Visual Story Telling (ViST) task.In addition, we introduce two models to incorporate high level structure learnt by a Finite State Machine (FSM) in neural sequential generation process by: (1) Scaffolding Structure in Decoder (SSiD) (2) Scaffolding Structure in Loss (SSiL).These models show an improvement in empirical as well as human evaluation.Our best performing model (SSiL) achieves a METEOR score of 0.31, which is an improvement of 0.6 over the baseline model.We also conducted human evaluation of the generated grounded recipes, which reveal that 61% found that our proposed (SSiL) model is better than the baseline model in terms of overall recipes, and 72.5% preferred our model in terms of coherence and structure.We also discuss analysis of the output highlighting key important NLP issues for prospective directions.", "target": ["In diesem Beitrag werden zwei Techniken vorgestellt, die eine Struktur auf hoher Ebene in die Erzeugung von prozeduralem Text aus einer Bildsequenz einbeziehen."]} +{"source": "Implicit models, which allow for the generation of samples but not for point-wise evaluation of probabilities, are omnipresent in real-world problems tackled by machine learning and a hot topic of current research.Some examples include data simulators that are widely used in engineering and scientific research, generative adversarial networks (GANs) for image synthesis, and hot-off-the-press approximate inference techniques relying on implicit distributions.The majority of existing approaches to learning implicit models rely on approximating the intractable distribution or optimisation objective for gradient-based optimisation, which is liable to produce inaccurate updates and thus poor models.This paper alleviates the need for such approximations by proposing the \\emph{Stein gradient estimator}, which directly estimates the score function of the implicitly defined distribution.The efficacy of the proposed estimator is empirically demonstrated by examples that include meta-learning for approximate inference and entropy regularised GANs that provide improved sample diversity.", "target": ["Wir haben einen neuartigen Gradientenschätzer nach der Stein-Methode eingeführt und mit anderen Methoden zum Erlernen impliziter Modelle für ungefähre Schlussfolgerungen und Bilderzeugung verglichen."]} +{"source": "Noise injection is a fundamental tool for data augmentation, and yet there is no widely accepted procedure to incorporate it with learning frameworks.This study analyzes the effects of adding or applying different noise models of varying magnitudes to Convolutional Neural Network (CNN) architectures.Noise models that are distributed with different density functions are given common magnitude levels via Structural Similarity (SSIM) metric in order to create an appropriate ground for comparison.The basic results are conforming with the most of the common notions in machine learning, and also introduces some novel heuristics and recommendations on noise injection.The new approaches will provide better understanding on optimal learning procedures for image classification.", "target": ["Ideale Methode, um während des CNN-Trainings Rauschen in die Eingangsdaten zu injizieren"]} +{"source": "State-of-the-art Unsupervised Domain Adaptation (UDA) methods learn transferable features by minimizing the feature distribution discrepancy between the source and target domains.Different from these methods which do not model the feature distributions explicitly, in this paper, we explore explicit feature distribution modeling for UDA.In particular, we propose Distribution Matching Prototypical Network (DMPN) to model the deep features from each domain as Gaussian mixture distributions.With explicit feature distribution modeling, we can easily measure the discrepancy between the two domains.In DMPN, we propose two new domain discrepancy losses with probabilistic interpretations.The first one minimizes the distances between the corresponding Gaussian component means of the source and target data.The second one minimizes the pseudo negative log likelihood of generating the target features from source feature distribution.To learn both discriminative and domain invariant features, DMPN is trained by minimizing the classification loss on the labeled source data and the domain discrepancy losses together.Extensive experiments are conducted over two UDA tasks.Our approach yields a large margin in the Digits Image transfer task over state-of-the-art approaches.More remarkably, DMPN obtains a mean accuracy of 81.4% on VisDA 2017 dataset.The hyper-parameter sensitivity analysis shows that our approach is robust w.r.t hyper-parameter changes.", "target": ["Wir schlagen vor, tiefe Merkmalsverteilungen von Quell- und Zieldaten explizit als Gaußsche Mischverteilungen für die unüberwachte Domänenanpassung (UDA) zu modellieren und erzielen bei mehreren UDA-Aufgaben bessere Ergebnisse als State-of-the-Art-Methoden."]} +{"source": "Efficiently learning to solve tasks in complex environments is a key challenge for reinforcement learning (RL) agents. We propose to decompose a complex environment using a task-agnostic world graphs, an abstraction that accelerates learning by enabling agents to focus exploration on a subspace of the environment.The nodes of a world graph are important waypoint states and edges represent feasible traversals between them. Our framework has two learning phases: 1) identifying world graph nodes and edges by training a binary recurrent variational auto-encoder (VAE) on trajectory data and2) a hierarchical RL framework that leverages structural and connectivity knowledge from the learned world graph to bias exploration towards task-relevant waypoints and regions.We show that our approach significantly accelerates RL on a suite of challenging 2D grid world tasks: compared to baselines, world graph integration doubles achieved rewards on simpler tasks, e.g. MultiGoal, and manages to solve more challenging tasks, e.g. Door-Key, where baselines fail.", "target": ["Wir erlernen eine aufgabenunabhängige Weltgraphenabstraktion der Umgebung und zeigen, wie deren Verwendung für die strukturierte Exploration die nachgelagerte aufgabenspezifische RL erheblich beschleunigen kann."]} +{"source": "We introduce the notion of property signatures, a representation for programs andprogram specifications meant for consumption by machine learning algorithms.Given a function with input type τ_in and output type τ_out, a property is a functionof type: (τ_in, τ_out) → Bool that (informally) describes some simple propertyof the function under consideration.For instance, if τ_in and τ_out are both listsof the same type, one property might ask ‘is the input list the same length as theoutput list?’.If we have a list of such properties, we can evaluate them all for ourfunction to get a list of outputs that we will call the property signature.Crucially,we can ‘guess’ the property signature for a function given only a set of input/outputpairs meant to specify that function.We discuss several potential applications ofproperty signatures and show experimentally that they can be used to improveover a baseline synthesizer so that it emits twice as many programs in less thanone-tenth of the time.", "target": ["Wir stellen ein Computerprogramm durch eine Reihe von einfacheren Programmen dar und verwenden diese Darstellung zur Verbesserung von Programmsynthesetechniken."]} +{"source": "Social dilemmas are situations where individuals face a temptation to increase their payoffs at a cost to total welfare.Building artificially intelligent agents that achieve good outcomes in these situations is important because many real world interactions include a tension between selfish interests and the welfare of others.We show how to modify modern reinforcement learning methods to construct agents that act in ways that are simple to understand, nice (begin by cooperating), provokable (try to avoid being exploited), and forgiving (try to return to mutual cooperation).We show both theoretically and experimentally that such agents can maintain cooperation in Markov social dilemmas.Our construction does not require training methods beyond a modification of self-play, thus if an environment is such that good strategies can be constructed in the zero-sum case (eg. Atari) then we can construct agents that solve social dilemmas in this environment.", "target": ["Wie können wir künstliche Agenten entwickeln, die soziale Dilemmasituationen lösen (Situationen, in denen Individuen versucht sind, ihren Nutzen auf Kosten des Gesamtwohls zu erhöhen)?"]} +{"source": "The reparameterization trick has become one of the most useful tools in the field of variational inference.However, the reparameterization trick is based on the standardization transformation which restricts the scope of application of this method to distributions that have tractable inverse cumulative distribution functions or are expressible as deterministic transformations of such distributions.In this paper, we generalized the reparameterization trick by allowing a general transformation.Unlike other similar works, we develop the generalized transformation-based gradient model formally and rigorously.We discover that the proposed model is a special case of control variate indicating that the proposed model can combine the advantages of CV and generalized reparameterization.Based on the proposed gradient model, we propose a new polynomial-based gradient estimator which has better theoretical performance than the reparameterization trick under certain condition and can be applied to a larger class of variational distributions.In studies of synthetic and real data, we show that our proposed gradient estimator has a significantly lower gradient variance than other state-of-the-art methods thus enabling a faster inference procedure.", "target": ["Wir schlagen ein neuartiges verallgemeinertes transformationsbasiertes Gradientenmodell vor und entwickeln einen polynomialen Gradientenschätzer, der auf diesem Modell basiert."]} +{"source": "The fault diagnosis in a modern communication system is traditionally supposed to be difficult, or even impractical for a purely data-driven machine learning approach, for it is a humanmade system of intensive knowledge.A few labeled raw packet streams extracted from fault archive can hardly be sufficient to deduce the intricate logic of underlying protocols.In this paper, we supplement these limited samples with two inexhaustible data sources: the unlabeled records probed from a system in service, and the labeled data simulated in an emulation environment.To transfer their inherent knowledge to the target domain, we construct a directed information flow graph, whose nodes are neural network components consisting of two generators, three discriminators and one classifier, and whose every forward path represents a pair of adversarial optimization goals, in accord with the semi-supervised and transfer learning demands.The multi-headed network can be trained in an alternative approach, at each iteration of which we select one target to update the weights along the path upstream, and refresh the residual layer-wisely to all outputs downstream.The actual results show that it can achieve comparable accuracy on classifying Transmission Control Protocol (TCP) streams without deliberate expert features.The solution has relieved operation engineers from massive works of understanding and maintaining rules, and provided a quick solution independent of specific protocols.", "target": ["Semi-überwachtes und Transfer-Lernen zur Klassifizierung von Paketflüssen über ein System kooperativer oder adversarischer neuronaler Blöcke"]} +{"source": "Our work addresses two important issues with recurrent neural networks: (1) they are over-parameterized, and (2) the recurrent weight matrix is ill-conditioned.The former increases the sample complexity of learning and the training time.The latter causes the vanishing and exploding gradient problem.We present a flexible recurrent neural network model called Kronecker Recurrent Units (KRU).KRU achieves parameter efficiency in RNNs through a Kronecker factored recurrent matrix.It overcomes the ill-conditioning of the recurrent matrix by enforcing soft unitary constraints on the factors.Thanks to the small dimensionality of the factors, maintaining these constraints is computationally efficient.Our experimental results on seven standard data-sets reveal that KRU can reduce the number of parameters by three orders of magnitude in the recurrent weight matrix compared to the existing recurrent models, without trading the statistical performance.These results in particular show that while there are advantages in having a high dimensional recurrent space, the capacity of the recurrent part of the model can be dramatically reduced.", "target": ["Unsere Arbeit stellt eine Kronecker-Faktorisierung von rekurrenten Gewichtsmatrizen für parameter-effiziente und gut konditionierte Recurrent Neural Networks vor."]} +{"source": "This paper studies the undesired phenomena of over-sensitivity of representations learned by deep networks to semantically-irrelevant changes in data.We identify a cause for this shortcoming in the classical Variational Auto-encoder (VAE) objective, the evidence lower bound (ELBO).We show that the ELBO fails to control the behaviour of the encoder out of the support of the empirical data distribution and this behaviour of the VAE can lead to extreme errors in the learned representation.This is a key hurdle in the effective use of representations for data-efficient learning and transfer.To address this problem, we propose to augment the data with specifications that enforce insensitivity of the representation with respect to families of transformations.To incorporate these specifications, we propose a regularization method that is based on a selection mechanism that creates a fictive data point by explicitly perturbing an observed true data point.For certain choices of parameters, our formulation naturally leads to the minimization of the entropy regularized Wasserstein distance between representations.We illustrate our approach on standard datasets and experimentally show that significant improvements in the downstream adversarial accuracy can be achieved by learning robust representations completely in an unsupervised manner, without a reference to a particular downstream task and without a costly supervised adversarial training procedure.", "target": ["Wir schlagen eine Methode zur Berechnung von adversarial robusten Repräsentationen in einer völlig unbeaufsichtigten Weise vor."]} +{"source": "We propose a novel score-based approach to learning a directed acyclic graph (DAG) from observational data.We adapt a recently proposed continuous constrained optimization formulation to allow for nonlinear relationships between variables using neural networks.This extension allows to model complex interactions while being more global in its search compared to other greedy approaches.In addition to comparing our method to existing continuous optimization methods, we provide missing empirical comparisons to nonlinear greedy search methods.On both synthetic and real-world data sets, this new method outperforms current continuous methods on most tasks while being competitive with existing greedy search methods on important metrics for causal inference.", "target": ["Wir schlagen einen neuen, auf Punkten basierenden Ansatz für strukturelles/kausales Lernen vor, der neuronale Netze und eine neue kontinuierliche, eingeschränkte Formulierung für dieses Problem nutzt"]} +{"source": "We study the problem of designing provably optimal adversarial noise algorithms that induce misclassification in settings where a learner aggregates decisions from multiple classifiers.Given the demonstrated vulnerability of state-of-the-art models to adversarial examples, recent efforts within the field of robust machine learning have focused on the use of ensemble classifiers as a way of boosting the robustness of individual models.In this paper, we design provably optimal attacks against a set of classifiers.We demonstrate how this problem can be framed as finding strategies at equilibrium in a two player, zero sum game between a learner and an adversary and consequently illustrate the need for randomization in adversarial attacks.The main technical challenge we consider is the design of best response oracles that can be implemented in a Multiplicative Weight Updates framework to find equilibrium strategies in the zero-sum game.We develop a series of scalable noise generation algorithms for deep neural networks, and show that it outperforms state-of-the-art attacks on various image classification tasks.Although there are generally no guarantees for deep learning, we show this is a well-principled approach in that it is provably optimal for linear classifiers.The main insight is a geometric characterization of the decision space that reduces the problem of designing best response oracles to minimizing a quadratic function over a set of convex polytopes.", "target": ["Das Papier analysiert das Problem des Entwurfs von Angriffen gegen mehrere Klassifizierer und stellt Algorithmen vor, die für lineare Klassifizierer optimal sind und die modernste Ergebnisse für Deep Learning liefern."]} +{"source": "Multiagent systems where the agents interact among themselves and with an stochastic environment can be formalized as stochastic games.We study a subclass of these games, named Markov potential games (MPGs), that appear often in economic and engineering applications when the agents share some common resource.We consider MPGs with continuous state-action variables, coupled constraints and nonconvex rewards.Previous analysis followed a variational approach that is only valid for very simple cases (convex rewards, invertible dynamics, and no coupled constraints); or considered deterministic dynamics and provided open-loop (OL) analysis, studying strategies that consist in predefined action sequences, which are not optimal for stochastic environments.We present a closed-loop (CL) analysis for MPGs and consider parametric policies that depend on the current state and where agents adapt to stochastic transitions.We provide easily verifiable, sufficient and necessary conditions for a stochastic game to be an MPG, even for complex parametric functions (e.g., deep neural networks); and show that a closed-loop Nash equilibrium (NE) can be found (or at least approximated) by solving a related optimal control problem (OCP).This is useful since solving an OCP---which is a single-objective problem---is usually much simpler than solving the original set of coupled OCPs that form the game---which is a multiobjective control problem.This is a considerable improvement over the previously standard approach for the CL analysis of MPGs, which gives no approximate solution if no NE belongs to the chosen parametric family, and which is practical only for simple parametric forms.We illustrate the theoretical contributions with an example by applying our approach to a noncooperative communications engineering game.We then solve the game with a deep reinforcement learning algorithm that learns policies that closely approximates an exact variational NE of the game.", "target": ["Wir präsentieren eine allgemeine Closed-Loop-Analyse für Markov-Potential-Spiele und zeigen, dass Deep Reinforcement Learning zum Erlernen eines approximativen Closed-Loop-Nash-Gleichgewichts verwendet werden kann."]} +{"source": "We make the following striking observation: fully convolutional VAE models trained on 32x32 ImageNet can generalize well, not just to 64x64 but also to far larger photographs, with no changes to the model.We use this property, applying fully convolutional models to lossless compression, demonstrating a method to scale the VAE-based 'Bits-Back with ANS' algorithm for lossless compression to large color photographs, and achieving state of the art for compression of full size ImageNet images.We release Craystack, an open source library for convenient prototyping of lossless compression using probabilistic models, along with full implementations of all of our compression results.", "target": ["Wir erhöhen die verlustfreie Komprimierung mit latenten Variablen und übertreffen damit bestehende Ansätze für ImageNet-Bilder in voller Größe."]} +{"source": "State of the art computer vision models have been shown to be vulnerable to small adversarial perturbations of the input.In other words, most images in the data distribution are both correctly classified by the model and are very close to a visually similar misclassified image.Despite substantial research interest, the cause of the phenomenon is still poorly understood and remains unsolved.We hypothesize that this counter intuitive behavior is a naturally occurring result of the high dimensional geometry of the data manifold.As a first step towards exploring this hypothesis, we study a simple synthetic dataset of classifying between two concentric high dimensional spheres.For this dataset we show a fundamental tradeoff between the amount of test error and the average distance to nearest error.In particular, we prove that any model which misclassifies a small constant fraction of a sphere will be vulnerable to adversarial perturbations of size $O(1/\\sqrt{d})$.Surprisingly, when we train several different architectures on this dataset, all of their error sets naturally approach this theoretical bound.As a result of the theory, the vulnerability of neural networks to small adversarial perturbations is a logical consequence of the amount of test error observed.We hope that our theoretical analysis of this very simple case will point the way forward to explore how the geometry of complex real-world data sets leads to adversarial examples.", "target": ["Wir stellen die Hypothese auf, dass die Anfälligkeit von Bildmodellen für kleine Störungen eine natürliche Folge der hochdimensionalen Geometrie der Datenvielfalt ist. Wir untersuchen und beweisen diese Hypothese theoretisch für einen einfachen synthetischen Datensatz."]} +{"source": "It has been established that diverse behaviors spanning the controllable subspace of a Markov decision process can be trained by rewarding a policy for being distinguishable from other policies.However, one limitation of this formulation is the difficulty to generalize beyond the finite set of behaviors being explicitly learned, as may be needed in subsequent tasks.Successor features provide an appealing solution to this generalization problem, but require defining the reward function as linear in some grounded feature space.In this paper, we show that these two techniques can be combined, and that each method solves the other's primary limitation.To do so we introduce Variational Intrinsic Successor FeatuRes (VISR), a novel algorithm which learns controllable features that can be leveraged to provide enhanced generalization and fast task inference through the successor features framework.We empirically validate VISR on the full Atari suite, in a novel setup wherein the rewards are only exposed briefly after a long unsupervised phase.Achieving human-level performance on 12 games and beating all baselines, we believe VISR represents a step towards agents that rapidly learn from limited feedback.", "target": ["Wir stellen Variational Intrinsic Successor FeatuRes (VISR) vor, einen neuartigen Algorithmus, der kontrollierbare Merkmale erlernt, die genutzt werden können, um schnelle Aufgabeninferenz durch den Rahmen der Nachfolgermerkmale zu ermöglichen."]} +{"source": "Predicting structured outputs such as semantic segmentation relies on expensive per-pixel annotations to learn strong supervised models like convolutional neural networks.However, these models trained on one data domain may not generalize well to other domains unequipped with annotations for model finetuning.To avoid the labor-intensive process of annotation, we develop a domain adaptation method to adapt the source data to the unlabeled target domain.To this end, we propose to learn discriminative feature representations of patches based on label histograms in the source domain, through the construction of a disentangled space.With such representations as guidance, we then use an adversarial learning scheme to push the feature representations in target patches to the closer distributions in source ones.In addition, we show that our framework can integrate a global alignment process with the proposed patch-level alignment and achieve state-of-the-art performance on semantic segmentation.Extensive ablation studies and experiments are conducted on numerous benchmark datasets with various settings, such as synthetic-to-real and cross-city scenarios.", "target": ["Eine Methode zur Anpassung von strukturierten Ausgaben durch das Erlernen von diskriminierenden Merkmalsdarstellungen auf Patch-Ebene"]} +{"source": "Many machine learning algorithms are vulnerable to almost imperceptible perturbations of their inputs.So far it was unclear how much risk adversarial perturbations carry for the safety of real-world machine learning applications because most methods used to generate such perturbations rely either on detailed model information (gradient-based attacks) or on confidence scores such as class probabilities (score-based attacks), neither of which are available in most real-world scenarios.In many such cases one currently needs to retreat to transfer-based attacks which rely on cumbersome substitute models, need access to the training data and can be defended against.Here we emphasise the importance of attacks which solely rely on the final model decision.Such decision-based attacks are (1) applicable to real-world black-box models such as autonomous cars, (2) need less knowledge and are easier to apply than transfer-based attacks and (3) are more robust to simple defences than gradient- or score-based attacks.Previous attacks in this category were limited to simple models or simple datasets.Here we introduce the Boundary Attack, a decision-based attack that starts from a large adversarial perturbation and then seeks to reduce the perturbation while staying adversarial.The attack is conceptually simple, requires close to no hyperparameter tuning, does not rely on substitute models and is competitive with the best gradient-based attacks in standard computer vision tasks like ImageNet.We apply the attack on two black-box algorithms from Clarifai.com.The Boundary Attack in particular and the class of decision-based attacks in general open new avenues to study the robustness of machine learning models and raise new questions regarding the safety of deployed machine learning systems.An implementation of the attack is available as part of Foolbox (https://github.com/bethgelab/foolbox).", "target": ["Ein neuartiger feinflicher Angriff, der reale Black-Box-Modelle des maschinellen Lernens ohne Übertragung direkt angreifen kann."]} +{"source": "We demonstrate that it is possible to train large recurrent language models with user-level differential privacy guarantees with only a negligible cost in predictive accuracy. Our work builds on recent advances in the training of deep networks on user-partitioned data and privacy accounting for stochastic gradient descent.In particular, we add user-level privacy protection to the federated averaging algorithm, which makes large step updates from user-level data.Our work demonstrates that given a dataset with a sufficiently large number of users (a requirement easily met by even small internet-scale datasets), achieving differential privacy comes at the cost of increased computation, rather than in decreased utility as in most prior work.We find that our private LSTM language models are quantitatively and qualitatively similar to un-noised models when trained on a large dataset.", "target": ["Differenzierter Datenschutz auf Benutzerebene für Recurrent Neural Network-Sprachmodelle ist mit einem ausreichend großen Datensatz möglich."]} +{"source": "Convolutional neural networks (CNNs) are commonly trained using a fixed spatial image size predetermined for a given model.Although trained on images of a specific size, it is well established that CNNs can be used to evaluate a wide range of image sizes at test time, by adjusting the size of intermediate feature maps. In this work, we describe and evaluate a novel mixed-size training regime that mixes several image sizes at training time.We demonstrate that models trained using our method are more resilient to image size changes and generalize well even on small images.This allows faster inference by using smaller images at test time.For instance, we receive a 76.43% top-1 accuracy using ResNet50 with an image size of 160, which matches the accuracy of the baseline model with 2x fewer computations.Furthermore, for a given image size used at test time, we show this method can be exploited either to accelerate training or the final test accuracy.For example, we are able to reach a 79.27% accuracy with a model evaluated at a 288 spatial size for a relative improvement of 14% over the baseline.", "target": ["Das Training von Convnets mit gemischten Bildgrößen kann die Ergebnisse bei der Auswertung über mehrere Größen hinweg verbessern"]} +{"source": "We propose a simple technique for encouraging generative RNNs to plan ahead.We train a ``backward'' recurrent network to generate a given sequence in reverse order, and we encourage states of the forward model to predict cotemporal states of the backward model.The backward network is used only during training, and plays no role during sampling or inference.We hypothesize that our approach eases modeling of long-term dependencies by implicitly forcing the forward states to hold information about the longer-term future (as contained in the backward states).We show empirically that our approach achieves 9% relative improvement for a speech recognition task, and achieves significant improvement on a COCO caption generation task.", "target": ["In diesem Beitrag wird eine Methode zur Ausbildung generativer Recurrent Networks vorgestellt, die eine Vorausplanung ermöglicht. Wir lassen ein zweites RNN in umgekehrter Richtung laufen und machen eine weiche Einschränkung zwischen zeitgleichen Vorwärts- und Rückwärtszuständen."]} +{"source": "Deep generative models seek to recover the process with which the observed data was generated.They may be used to synthesize new samples or to subsequently extract representations.Successful approaches in the domain of images are driven by several core inductive biases.However, a bias to account for the compositional way in which humans structure a visual scene in terms of objects has frequently been overlooked.In this work we propose to structure the generator of a GAN to consider objects and their relations explicitly, and generate images by means of composition.This provides a way to efficiently learn a more accurate generative model of real-world images, and serves as an initial step towards learning corresponding object representations.We evaluate our approach on several multi-object image datasets, and find that the generator learns to identify and disentangle information corresponding to different objects at a representational level.A human study reveals that the resulting generative model is better at generating images that are more faithful to the reference distribution.", "target": ["Wir schlagen vor, den Generator eines GAN so zu strukturieren, dass er Objekte und ihre Beziehungen explizit berücksichtigt und Bilder durch Komposition erzeugt"]} +{"source": "Current literature in machine learning holds that unaligned, self-interested agents do not learn to use an emergent communication channel.We introduce a new sender-receiver game to study emergent communication for this spectrum of partially-competitive scenarios and put special care into evaluation.We find that communication can indeed emerge in partially-competitive scenarios, and we discover three things that are tied to improving it.First, that selfish communication is proportional to cooperation, and it naturally occurs for situations that are more cooperative than competitive.Second, that stability and performance are improved by using LOLA (Foerster et al, 2018), especially in more competitive scenarios.And third, that discrete protocols lend themselves better to learning cooperative communication than continuous ones.", "target": ["Wir schaffen es, die Kommunikation mit egoistischen Agenten hervorzubringen, im Gegensatz zur aktuellen Ansicht in ML"]} +{"source": "Deep neural networks (DNNs) typically have enough capacity to fit random data by brute force even when conventional data-dependent regularizations focusing on the geometry of the features are imposed.We find out that the reason for this is the inconsistency between the enforced geometry and the standard softmax cross entropy loss.To resolve this, we propose a new framework for data-dependent DNN regularization, the Geometrically-Regularized-Self-Validating neural Networks (GRSVNet).During training, the geometry enforced on one batch of features is simultaneously validated on a separate batch using a validation loss consistent with the geometry.We study a particular case of GRSVNet, the Orthogonal-Low-rank Embedding (OLE)-GRSVNet, which is capable of producing highly discriminative features residing in orthogonal low-rank subspaces.Numerical experiments show that OLE-GRSVNet outperforms DNNs with conventional regularization when trained on real data.More importantly, unlike conventional DNNs, OLE-GRSVNet refuses to memorize random data or random labels, suggesting it only learns intrinsic patterns by reducing the memorizing capacity of the baseline DNN.", "target": ["Wir schlagen einen neuen Rahmen für datenabhängige DNN-Regularisierung vor, der verhindern kann, dass DNNs sich zufälligen Daten oder zufälligen Bezeichnungen zu sehr anpassen."]} +{"source": "End-to-end automatic speech recognition (ASR) commonly transcribes audio signals into sequences of characters while its performance is evaluated by measuring the word-error rate (WER).This suggests that predicting sequences of words directly may be helpful instead.However, training with word-level supervision can be more difficult due to the sparsity of examples per label class.In this paper we analyze an end-to-end ASR model that combines a word-and-character representation in a multi-task learning (MTL) framework.We show that it improves on the WER and study how the word-level model can benefit from character-level supervision by analyzing the learned inductive preference bias of each model component empirically.We find that by adding character-level supervision, the MTL model interpolates between recognizing more frequent words (preferred by the word-level model) and shorter words (preferred by the character-level model).", "target": ["Multitasking-Lernen verbessert die Spracherkennung auf Wort- und Zeichenebene durch Interpolation der Präferenzvorlieben seiner Komponenten: Frequenz- und Wortlängenpräferenz."]} +{"source": "Discretizing floating-point vectors is a fundamental step of modern indexing methods.State-of-the-art techniques learn parameters of the quantizers on training data for optimal performance, thus adapting quantizers to the data.In this work, we propose to reverse this paradigm and adapt the data to the quantizer: we train a neural net whose last layers form a fixed parameter-free quantizer, such as pre-defined points of a sphere.As a proxy objective, we design and train a neural network that favors uniformity in the spherical latent space, while preserving the neighborhood structure after the mapping. For this purpose, we propose a new regularizer derived from the Kozachenko-Leonenko differential entropy estimator and combine it with a locality-aware triplet loss. Experiments show that our end-to-end approach outperforms most learned quantization methods, and is competitive with the state of the art on widely adopted benchmarks.Further more, we show that training without the quantization step results in almost no difference in accuracy, but yields a generic catalyser that can be applied with any subsequent quantization technique.", "target": ["Wir lernen ein neuronales Netz, das die Eingabeverteilung vereinheitlicht, was zu einer konkurrenzfähigen Indexierungsleistung im hochdimensionalen Raum führt"]} +{"source": "Temporal difference (TD) learning is a popular algorithm for policy evaluation in reinforcement learning, but the vanilla TD can substantially suffer from the inherent optimization variance.A variance reduced TD (VRTD) algorithm was proposed by Korda and La (2015), which applies the variance reduction technique directly to the online TD learning with Markovian samples.In this work, we first point out the technical errors in the analysis of VRTD in Korda and La (2015), and then provide a mathematically solid analysis of the non-asymptotic convergence of VRTD and its variance reduction performance.We show that VRTD is guaranteed to converge to a neighborhood of the fixed-point solution of TD at a linear convergence rate.Furthermore, the variance error (for both i.i.d. and Markovian sampling) and the bias error (for Markovian sampling) of VRTD are significantly reduced by the batch size of variance reduction in comparison to those of vanilla TD.", "target": ["Dieses Papier bietet eine rigorose Studie des varianzreduzierten TD-Lernens und charakterisiert seinen Vorteil gegenüber dem einfachen TD-Lernen"]} +{"source": "We tackle unsupervised domain adaptation by accounting for the fact that different domains may need to be processed differently to arrive to a common feature representation effective for recognition.To this end, we introduce a deep learning framework where each domain undergoes a different sequence of operations, allowing some, possibly more complex, domains to go through more computations than others.This contrasts with state-of-the-art domain adaptation techniques that force all domains to be processed with the same series of operations, even when using multi-stream architectures whose parameters are not shared.As evidenced by our experiments, the greater flexibility of our method translates to higher accuracy.Furthermore, it allows us to handle any number of domains simultaneously.", "target": ["Ein Multiflow-Netzwerk ist eine dynamische Architektur für die Domänenanpassung, die potenziell unterschiedliche Berechnungsgraphen pro Domäne erlernt, um sie auf eine gemeinsame Darstellung abzubilden, in der Inferenzen auf domänenunabhängige Weise durchgeführt werden können."]} +{"source": "The practical usage of reinforcement learning agents is often bottlenecked by the duration of training time.To accelerate training, practitioners often turn to distributed reinforcement learning architectures to parallelize and accelerate the training process.However, modern methods for scalable reinforcement learning (RL) often tradeoff between the throughput of samples that an RL agent can learn from (sample throughput) and the quality of learning from each sample (sample efficiency).In these scalable RL architectures, as one increases sample throughput (i.e. increasing parallelization in IMPALA (Espeholt et al., 2018)), sample efficiency drops significantly.To address this, we propose a new distributed reinforcement learning algorithm, IMPACT.IMPACT extends PPO with three changes: a target network for stabilizing the surrogate objective, a circular buffer, and truncated importance sampling.In discrete action-space environments, we show that IMPACT attains higher reward and, simultaneously, achieves up to 30% decrease in training wall-time than that of IMPALA.For continuous control environments, IMPACT trains faster than existing scalable agents while preserving the sample efficiency of synchronous PPO.", "target": ["IMPACT hilft RL-Agenten, schneller zu trainieren, indem es die Trainingszeit an der Wall-Clock verkürzt und gleichzeitig die Effizienz der Beispiele erhöht."]} +{"source": "In this paper, we show that a simple coloring scheme can improve, both theoretically and empirically, the expressive power of Message Passing Neural Networks (MPNNs).More specifically, we introduce a graph neural network called Colored Local Iterative Procedure (CLIP) that uses colors to disambiguate identical node attributes, and show that this representation is a universal approximator of continuous functions on graphs with node attributes.Our method relies on separability, a key topological characteristic that allows to extend well-chosen neural networks into universal representations.Finally, we show experimentally that CLIP is capable of capturing structural characteristics that traditional MPNNs fail to distinguish, while being state-of-the-art on benchmark graph classification datasets.", "target": ["In diesem Papier wird ein Färbungsschema für die Knoten-Disambiguierung in graphischen neuronalen Netzen vorgestellt, das auf Trennbarkeit basiert und sich als universelle MPNN-Erweiterung erweist."]} +{"source": "In this paper we present a method for algorithmic melody generation using a generative adversarial network without recurrent components.Music generation has been successfully done using recurrent neural networks, where the model learns sequence information that can help create authentic sounding melodies. Here, we use DCGAN architecture with dilated convolutions and towers to capture sequential information as spatial image information, and learn long-range dependencies in fixed-length melody forms such as Irish traditional reel.", "target": ["Wenn wir Melodien als Bilder mit semantischen Einheiten darstellen, können wir sie mit einem DCGAN ohne rekurrente Komponenten erzeugen."]} +{"source": "Neural machine translation (NMT) systems have reached state of the art performance in translating text and widely deployed. Yet little is understood about how these systems function or break. Here we show that NMT systems are susceptible to producing highly pathological translations that are completely untethered from the source material, which we term hallucinations. Such pathological translations are problematic because they are are deeply disturbing of user trust and easy to find. We describe a method t generate hallucinations and show that many common variations of the NMT architecture are susceptible to them.We study a variety of approaches to reduce the frequency of hallucinations, including data augmentation, dynamical systems and regularization techniques and show that data augmentation significantly reduces hallucination frequency.Finally, we analyze networks that produce hallucinations and show signatures of hallucinations in the attention matrix and in the stability measures of the decoder.", "target": ["Wir analysieren und stellen das Phänomen der \"Halluzinationen\" in der NMT vor, auch falsche Übersetzungen ohne Bezug zum Quelltext, und schlagen Methoden vor, um deren Häufigkeit zu reduzieren."]} +{"source": "For AI systems to garner widespread public acceptance, we must develop methods capable of explaining the decisions of black-box models such as neural networks.In this work, we identify two issues of current explanatory methods.First, we show that two prevalent perspectives on explanations—feature-additivity and feature-selection—lead to fundamentally different instance-wise explanations.In the literature, explainers from different perspectives are currently being directly compared, despite their distinct explanation goals.The second issue is that current post-hoc explainers have only been thoroughly validated on simple models, such as linear regression, and, when applied to real-world neural networks, explainers are commonly evaluated under the assumption that the learned models behave reasonably.However, neural networks often rely on unreasonable correlations, even when producing correct decisions.We introduce a verification framework for explanatory methods under the feature-selection perspective.Our framework is based on a non-trivial neural network architecture trained on a real-world task, and for which we are able to provide guarantees on its inner workings.We validate the efficacy of our evaluation by showing the failure modes of current explainers.We aim for this framework to provide a publicly available,1 off-the-shelf evaluation when the feature-selection perspective on explanations is needed.", "target": ["Ein Bewertungsrahmen auf der Grundlage eines realen neuronalen Netzes für post-hoc-Erklärungsmethoden"]} +{"source": "Planning in high-dimensional space remains a challenging problem, even with recent advances in algorithms and computational power.We are inspired by efference copy and sensory reafference theory from neuroscience. Our aim is to allow agents to form mental models of their environments for planning. The cerebellum is emulated with a two-stream, fully connected, predictor network.The network receives as inputs the efference as well as the features of the current state.Building on insights gained from knowledge distillation methods, we choose as our features the outputs of a pre-trained network, yielding a compressed representation of the current state. The representation is chosen such that it allows for fast search using classical graph search algorithms.We display the effectiveness of our approach on a viewpoint-matching task using a modified best-first search algorithm.", "target": ["Wir stellen eine von den Neurowissenschaften inspirierte Methode vor, die auf neuronalen Netzen für die Suche im latenten Raum basiert"]} +{"source": "Experimental evidence indicates that simple models outperform complex deep networks on many unsupervised similarity tasks.Introducing the concept of an optimal representation space, we provide a simple theoretical resolution to this apparent paradox.In addition, we present a straightforward procedure that, without any retraining or architectural modifications, allows deep recurrent models to perform equally well (and sometimes better) when compared to shallow models.To validate our analysis, we conduct a set of consistent empirical evaluations and introduce several new sentence embedding models in the process.Even though this work is presented within the context of natural language processing, the insights are readily applicable to other domains that rely on distributed representations for transfer tasks.", "target": ["Durch die Einführung des Begriffs eines optimalen Repräsentationsraums liefern wir ein theoretisches Argument und eine experimentelle Validierung, dass ein unbeaufsichtigtes Modell für Sätze sowohl bei überwachten Ähnlichkeitsaufgaben als auch bei unbeaufsichtigten Transferaufgaben gut abschneiden kann."]} +{"source": "Cloze test is widely adopted in language exams to evaluate students' language proficiency.In this paper, we propose the first large-scale human-designed cloze test dataset CLOTH in which the questions were used in middle-school and high-school language exams.With the missing blanks carefully created by teachers and candidate choices purposely designed to be confusing, CLOTH requires a deeper language understanding and a wider attention span than previous automatically generated cloze datasets.We show humans outperform dedicated designed baseline models by a significant margin, even when the model is trained on sufficiently large external data.We investigate the source of the performance gap, trace model deficiencies to some distinct properties of CLOTH, and identify the limited ability of comprehending a long-term context to be the key bottleneck.In addition, we find that human-designed data leads to a larger gap between the model's performance and human performance when compared to automatically generated data.", "target": ["Ein Lückentext-Datensatz, der von Lehrern zur Bewertung von Sprachkenntnissen entwickelt wurde"]} +{"source": "Recent work suggests goal-driven training of neural networks can be used to model neural activity in the brain.While response properties of neurons in artificial neural networks bear similarities to those in the brain, the network architectures are often constrained to be different.Here we ask if a neural network can recover both neural representations and, if the architecture is unconstrained and optimized, also the anatomical properties of neural circuits.We demonstrate this in a system where the connectivity and the functional organization have been characterized, namely, the head direction circuit of the rodent and fruit fly.We trained recurrent neural networks (RNNs) to estimate head direction through integration of angular velocity.We found that the two distinct classes of neurons observed in the head direction system, the Ring neurons and the Shifter neurons, emerged naturally in artificial neural networks as a result of training.Furthermore, connectivity analysis and in-silico neurophysiology revealed structural and mechanistic similarities between artificial networks and the head direction system.Overall, our results show that optimization of RNNs in a goal-driven task can recapitulate the structure and function of biological circuits, suggesting that artificial neural networks can be used to study the brain at the level of both neural activity and anatomical organization.", "target": ["Künstliche neuronale Netze, die mit Gradientenabstieg trainiert werden, sind in der Lage, sowohl realistische neuronale Aktivität als auch die anatomische Organisation eines biologischen Kreises zu rekapitulieren."]} +{"source": "Convolutional Neural Networks (CNNs) are computationally intensive, which limits their application on mobile devices.Their energy is dominated by the number of multiplies needed to perform the convolutions.Winograd’s minimal filtering algorithm (Lavin, 2015) and network pruning (Han et al., 2015) can reduce the operation count, but these two methods cannot be straightforwardly combined — applying the Winograd transform fills in the sparsity in both the weights and the activations.We propose two modifications to Winograd-based CNNs to enable these methods to exploit sparsity.First, we move the ReLU operation into the Winograd domain to increase the sparsity of the transformed activations.Second, we prune the weights in the Winograd domain to exploit static weight sparsity.For models on CIFAR-10, CIFAR-100 and ImageNet datasets, our method reduces the number of multiplications by 10.4x, 6.8x and 10.8x respectively with loss of accuracy less than 0.1%, outperforming previous baselines by 2.0x-3.0x.We also show that moving ReLU to the Winograd domain allows more aggressive pruning.", "target": ["Prune und ReLU im Winograd-Bereich für effiziente Convolutional Neural Networks"]} +{"source": "In this paper we present a novel optimization algorithm called Advanced Neuroevolution.The aim for this algorithm is to train deep neural networks, and eventually act as an alternative to Stochastic Gradient Descent (SGD) and its variants as needed.We evaluated our algorithm on the MNIST dataset, as well as on several global optimization problems such as the Ackley function.We find the algorithm performing relatively well for both cases, overtaking other global optimization algorithms such as Particle Swarm Optimization (PSO) and Evolution Strategies (ES).", "target": ["Ein neuer Algorithmus zum Trainieren tiefer neuronaler Netze. Getestet an Optimierungsfunktionen und MNIST."]} +{"source": "Stochastic neural net weights are used in a variety of contexts, including regularization, Bayesian neural nets, exploration in reinforcement learning, and evolution strategies.Unfortunately, due to the large number of weights, all the examples in a mini-batch typically share the same weight perturbation, thereby limiting the variance reduction effect of large mini-batches.We introduce flipout, an efficient method for decorrelating the gradients within a mini-batch by implicitly sampling pseudo-independent weight perturbations for each example.Empirically, flipout achieves the ideal linear variance reduction for fully connected networks, convolutional networks, and RNNs.We find significant speedups in training neural networks with multiplicative Gaussian perturbations.We show that flipout is effective at regularizing LSTMs, and outperforms previous methods.Flipout also enables us to vectorize evolution strategies: in our experiments, a single GPU with flipout can handle the same throughput as at least 40 CPU cores using existing methods, equivalent to a factor-of-4 cost reduction on Amazon Web Services.", "target": ["Wir stellen Flipout vor, eine effiziente Methode zur Dekorrelation der Gradienten, die durch stochastische neuronale Netzgewichte innerhalb eines Mini-Batches berechnet werden, indem implizit pseudo-unabhängige Gewichtsstörungen für jedes Beispiel gesammelt werden."]} +{"source": "Deep generative models such as Variational AutoEncoder (VAE) and Generative Adversarial Network (GAN) play an increasingly important role in machine learning and computer vision.However, there are two fundamental issues hindering their real-world applications: the difficulty of conducting variational inference in VAE and the functional absence of encoding real-world samples in GAN.In this paper, we propose a novel algorithm named Latently Invertible Autoencoder (LIA) to address the above two issues in one framework.An invertible network and its inverse mapping are symmetrically embedded in the latent space of VAE.Thus the partial encoder first transforms the input into feature vectors and then the distribution of these feature vectors is reshaped to fit a prior by the invertible network.The decoder proceeds in the reverse order of the encoder's composite mappings.A two-stage stochasticity-free training scheme is designed to train LIA via adversarial learning, in the sense that the decoder of LIA is first trained as a standard GAN with the invertible network and then the partial encoder is learned from an autoencoder by detaching the invertible network from LIA. Experiments conducted on the FFHQ face dataset and three LSUN datasets validate the effectiveness of LIA for inference and generation.", "target": ["Ein neues Modell eines Latently Invertible Autoencoders wird vorgeschlagen, um das Problem der variationalen Inferenz in VAE mit dem invertible Netzwerk und zweistufige adversarial Ausbildung zu lösen."]} +{"source": "Neural programs are highly accurate and structured policies that perform algorithmic tasks by controlling the behavior of a computation mechanism.Despite the potential to increase the interpretability and the compositionality of the behavior of artificial agents, it remains difficult to learn from demonstrations neural networks that represent computer programs.The main challenges that set algorithmic domains apart from other imitation learning domains are the need for high accuracy, the involvement of specific structures of data, and the extremely limited observability.To address these challenges, we propose to model programs as Parametrized Hierarchical Procedures (PHPs).A PHP is a sequence of conditional operations, using a program counter along with the observation to select between taking an elementary action, invoking another PHP as a sub-procedure, and returning to the caller.We develop an algorithm for training PHPs from a set of supervisor demonstrations, only some of which are annotated with the internal call structure, and apply it to efficient level-wise training of multi-level PHPs.We show in two benchmarks, NanoCraft and long-hand addition, that PHPs can learn neural programs more accurately from smaller amounts of both annotated and unannotated demonstrations.", "target": ["Wir stellen das PHP-Modell für die hierarchische Darstellung neuronaler Programme und einen Algorithmus zum Lernen von PHPs aus einer Mischung von starker und schwacher Überwachung vor."]} +{"source": "Deep Reinforcement Learning has managed to achieve state-of-the-art results in learning control policies directly from raw pixels.However, despite its remarkable success, it fails to generalize, a fundamental component required in a stable Artificial Intelligence system.Using the Atari game Breakout, we demonstrate the difficulty of a trained agent in adjusting to simple modifications in the raw image, ones that a human could adapt to trivially.In transfer learning, the goal is to use the knowledge gained from the source task to make the training of the target task faster and better.We show that using various forms of fine-tuning, a common method for transfer learning, is not effective for adapting to such small visual changes.In fact, it is often easier to re-train the agent from scratch than to fine-tune a trained agent.We suggest that in some cases transfer learning can be improved by adding a dedicated component whose goal is to learn to visually map between the known domain and the new one.Concretely, we use Unaligned Generative Adversarial Networks (GANs) to create a mapping function to translate images in the target task to corresponding images in the source task.These mapping functions allow us to transform between various variations of the Breakout game, as well as between different levels of a Nintendo game, Road Fighter.We show that learning this mapping is substantially more efficient than re-training.A visualization of a trained agent playing Breakout and Road Fighter, with and without the GAN transfer, can be seen in \\url{https://streamable.com/msgtm} and \\url{https://streamable.com/5e2ka}.", "target": ["Wir schlagen eine Methode zum Wissenstransfer zwischen verwandten RL-Aufgaben unter Verwendung visueller Mappings vor und demonstrieren ihre Effektivität an visuellen Varianten des Atari Breakout-Spiels und verschiedenen Levels von Road Fighter, einem Nintendo-Autofahrspiel."]} +{"source": "A commonplace belief in the machine learning community is that using adaptive gradient methods hurts generalization.We re-examine this belief both theoretically and experimentally, in light of insights and trends from recent years.We revisit some previous oft-cited experiments and theoretical accounts in more depth, and provide a new set of experiments in larger-scale, state-of-the-art settings.We conclude that with proper tuning, the improved training performance of adaptive optimizers does not in general carry an overfitting penalty, especially in contemporary deep learning.Finally, we synthesize a ``user's guide'' to adaptive optimizers, including some proposed modifications to AdaGrad to mitigate some of its empirical shortcomings.", "target": ["Richtig angewandte adaptive Gradientenmethoden sind nicht mit einem Generalisierungsnachteil verbunden."]} +{"source": "The ability to generalize quickly from few observations is crucial for intelligent systems.In this paper we introduce APL, an algorithm that approximates probability distributions by remembering the most surprising observations it has encountered.These past observations are recalled from an external memory module and processed by a decoder network that can combine information from different memory slots to generalize beyond direct recall.We show this algorithm can perform as well as state of the art baselines on few-shot classification benchmarks with a smaller memory footprint. In addition, its memory compression allows it to scale to thousands of unknown labels. Finally, we introduce a meta-learning reasoning task which is more challenging than direct classification.In this setting, APL is able to generalize with fewer than one example per class via deductive reasoning.", "target": ["Wir stellen ein Modell vor, das ausgehend von wenigen Beobachtungen schnell verallgemeinert, indem es überraschende Informationen speichert und zu jedem Zeitpunkt die wichtigsten Daten berücksichtigt."]} +{"source": "Recent advances in recurrent neural nets (RNNs) have shown much promise in many applications in natural language processing.For most of these tasks, such as sentiment analysis of customer reviews, a recurrent neural net model parses the entire review before forming a decision.We argue that reading the entire input is not always necessary in practice, since a lot of reviews are often easy to classify, i.e., a decision can be formed after reading some crucial sentences or words in the provided text.In this paper, we present an approach of fast reading for text classification.Inspired by several well-known human reading techniques, our approach implements an intelligent recurrent agent which evaluates the importance of the current snippet in order to decide whether to make a prediction, or to skip some texts, or to re-read part of the sentence.Our agent uses an RNN module to encode information from the past and the current tokens, and applies a policy module to form decisions.With an end-to-end training algorithm based on policy gradient, we train and test our agent on several text classification datasets and achieve both higher efficiency and better accuracy compared to previous approaches.", "target": ["Wir entwickeln einen durchgängig trainierbaren Ansatz für das Überfliegen, Wiederlesen und frühzeitige Beenden von Klassifikationsaufgaben."]} +{"source": "Reinforcement learning agents need to explore their unknown environments to solve the tasks given to them.The Bayes optimal solution to exploration is intractable for complex environments, and while several exploration methods have been proposed as approximations, it remains unclear what underlying objective is being optimized by existing exploration methods, or how they can be altered to incorporate prior knowledge about the task.Moreover, it is unclear how to acquire a single exploration strategy that will be useful for solving multiple downstream tasks.We address these shortcomings by learning a single exploration policy that can quickly solve a suite of downstream tasks in a multi-task setting, amortizing the cost of learning to explore.We recast exploration as a problem of State Marginal Matching (SMM), where we aim to learn a policy for which the state marginal distribution matches a given target state distribution, which can incorporate prior knowledge about the task.We optimize the objective by reducing it to a two-player, zero-sum game between a state density model and a parametric policy.Our theoretical analysis of this approach suggests that prior exploration methods do not learn a policy that does distribution matching, but acquire a replay buffer that performs distribution matching, an observation that potentially explains these prior methods' success in single-task settings.On both simulated and real-world tasks, we demonstrate that our algorithm explores faster and adapts more quickly than prior methods.", "target": ["Wir betrachten Exploration in RL als ein Problem der Anpassung einer Randverteilung über Zustände."]} +{"source": "The effectiveness of Convolutional Neural Networks stems in large part from their ability to exploit the translation invariance that is inherent in many learning problems.Recently, it was shown that CNNs can exploit other invariances, such as rotation invariance, by using group convolutions instead of planar convolutions.However, for reasons of performance and ease of implementation, it has been necessary to limit the group convolution to transformations that can be applied to the filters without interpolation.Thus, for images with square pixels, only integer translations, rotations by multiples of 90 degrees, and reflections are admissible.Whereas the square tiling provides a 4-fold rotational symmetry, a hexagonal tiling of the plane has a 6-fold rotational symmetry.In this paper we show how one can efficiently implement planar convolution and group convolution over hexagonal lattices, by re-using existing highly optimized convolution routines.We find that, due to the reduced anisotropy of hexagonal filters, planar HexaConv provides better accuracy than planar convolution with square filters, given a fixed parameter budget.Furthermore, we find that the increased degree of symmetry of the hexagonal grid increases the effectiveness of group convolutions, by allowing for more parameter sharing.We show that our method significantly outperforms conventional CNNs on the AID aerial scene classification dataset, even outperforming ImageNet pre-trained models.", "target": ["Wir stellen G-HexaConv vor, ein gruppenäquivariantes Convolutional Neural Network auf hexagonalen Gittern."]} +{"source": "Deep Convolutional Neural Networks (CNNs) have been repeatedly shown to perform well on image classification tasks, successfully recognizing a broad array of objects when given sufficient training data.Methods for object localization, however, are still in need of substantial improvement.Common approaches to this problem involve the use of a sliding window, sometimes at multiple scales, providing input to a deep CNN trained to classify the contents of the window.In general, these approaches are time consuming, requiring many classification calculations.In this paper, we offer a fundamentally different approach to the localization of recognized objects in images.Our method is predicated on the idea that a deep CNN capable of recognizing an object must implicitly contain knowledge about object location in its connection weights.We provide a simple method to interpret classifier weights in the context of individual classified images.This method involves the calculation of the derivative of network generated activation patterns, such as the activation of output class label units, with regard to each in- put pixel, performing a sensitivity analysis that identifies the pixels that, in a local sense, have the greatest influence on internal representations and object recognition.These derivatives can be efficiently computed using a single backward pass through the deep CNN classifier, producing a sensitivity map of the image.We demonstrate that a simple linear mapping can be learned from sensitivity maps to bounding box coordinates, localizing the recognized object.Our experimental results, using real-world data sets for which ground truth localization information is known, reveal competitive accuracy from our fast technique.", "target": ["Vorschlag eines neuartigen Ansatzes zur Objektlokalisierung (Erkennung) auf der Grundlage der Interpretation des tiefen CNN unter Verwendung der internen Darstellung und der Gedanken des Netzwerks"]} +{"source": "We present trellis networks, a new architecture for sequence modeling.On the one hand, a trellis network is a temporal convolutional network with special structure, characterized by weight tying across depth and direct injection of the input into deep layers.On the other hand, we show that truncated recurrent networks are equivalent to trellis networks with special sparsity structure in their weight matrices.Thus trellis networks with general weight matrices generalize truncated recurrent networks.We leverage these connections to design high-performing trellis networks that absorb structural and algorithmic elements from both recurrent and convolutional models.Experiments demonstrate that trellis networks outperform the current state of the art methods on a variety of challenging benchmarks, including word-level language modeling and character-level language modeling tasks, and stress tests designed to evaluate long-term memory retention.The code is available at https://github.com/locuslab/trellisnet .", "target": ["Trellis-Netzwerke sind eine neue Architektur zur Sequenzmodellierung, die rekurrente und konvolutionale Modelle miteinander verbindet und einen neuen Stand der Technik für die Sprachmodellierung auf Wort- und Zeichenebene darstellt."]} +{"source": "We propose an end-to-end framework for training domain specific models (DSMs) to obtain both high accuracy and computational efficiency for object detection tasks.DSMs are trained with distillation and focus on achieving high accuracy at a limited domain (e.g. fixed view of an intersection).We argue that DSMs can capture essential features well even with a small model size, enabling higher accuracy and efficiency than traditional techniques. In addition, we improve the training efficiency by reducing the dataset size by culling easy to classify images from the training set.For the limited domain, we observed that compact DSMs significantly surpass the accuracy of COCO trained models of the same size.By training on a compact dataset, we show that with an accuracy drop of only 3.6%, the training time can be reduced by 93%.", "target": ["Eine hohe Objekterkennungsgenauigkeit kann durch das Training von domänenspezifischen kompakten Modellen erreicht werden, und das Training kann sehr kurz sein."]} +{"source": "We compare the model-free reinforcement learning with the model-based approaches through the lens of the expressive power of neural networks for policies, $Q$-functions, and dynamics. We show, theoretically and empirically, that even for one-dimensional continuous state space, there are many MDPs whose optimal $Q$-functions and policies are much more complex than the dynamics.We hypothesize many real-world MDPs also have a similar property.For these MDPs, model-based planning is a favorable algorithm, because the resulting policies can approximate the optimal policy significantly better than a neural network parameterization can, and model-free or model-based policy optimization rely on policy parameterization.Motivated by the theory, we apply a simple multi-step model-based bootstrapping planner (BOOTS) to bootstrap a weak $Q$-function into a stronger policy.Empirical results show that applying BOOTS on top of model-based or model-free policy optimization algorithms at the test time improves the performance on MuJoCo benchmark tasks.", "target": ["Wir vergleichen tiefe modellbasierte und modellfreie RL-Algorithmen, indem wir die Approximierbarkeit von $Q$-Funktionen, Richtlinien und Dynamik durch neuronale Netze untersuchen."]} +{"source": "Common-sense physical reasoning is an essential ingredient for any intelligent agent operating in the real-world.For example, it can be used to simulate the environment, or to infer the state of parts of the world that are currently unobserved.In order to match real-world conditions this causal knowledge must be learned without access to supervised data.To address this problem we present a novel method that learns to discover objects and model their physical interactions from raw visual images in a purely unsupervised fashion.It incorporates prior knowledge about the compositional nature of human perception to factor interactions between object-pairs and learn efficiently.On videos of bouncing balls we show the superior modelling capabilities of our method compared to other unsupervised neural approaches that do not incorporate such prior knowledge.We demonstrate its ability to handle occlusion and show that it can extrapolate learned knowledge to scenes with different numbers of objects.", "target": ["Wir stellen einen neuartigen Ansatz für physikalische Schlussfolgerungen mit gesundem Menschenverstand vor, der lernt, Objekte zu erkennen und ihre physikalischen Interaktionen aus visuellen Rohbildern zu modellieren, und zwar auf rein unüberwachte Weise"]} +{"source": "The idea that neural networks may exhibit a bias towards simplicity has a long history.Simplicity bias provides a way to quantify this intuition. It predicts, for a broad class of input-output maps which can describe many systems in science and engineering, that simple outputs are exponentially more likely to occur upon uniform random sampling of inputs than complex outputs are. This simplicity bias behaviour has been observed for systems ranging from the RNA sequence to secondary structure map, to systems of coupled differential equations, to models of plant growth. Deep neural networks can be viewed as a mapping from the space of parameters (the weights) to the space of functions (how inputs get transformed to outputs by the network). We show that this parameter-function map obeys the necessary conditions for simplicity bias, and numerically show that it is hugely biased towards functions with low descriptional complexity. We also demonstrate a Zipf like power-law probability-rank relation. A bias towards simplicity may help explain why neural nets generalize so well.", "target": ["Bei vielen einfachen Input-Output-Zuordnungen ist eine sehr starke Tendenz zu einfachen Outputs zu beobachten. Die Parameter-Funktions Zuordnung von tiefen Netzen ist auf die gleiche Weise verzerrt."]} +{"source": "Imitation Learning (IL) is an appealing approach to learn desirable autonomous behavior.However, directing IL to achieve arbitrary goals is difficult.In contrast, planning-based algorithms use dynamics models and reward functions to achieve goals.Yet, reward functions that evoke desirable behavior are often difficult to specify.In this paper, we propose \"Imitative Models\" to combine the benefits of IL and goal-directed planning.Imitative Models are probabilistic predictive models of desirable behavior able to plan interpretable expert-like trajectories to achieve specified goals.We derive families of flexible goal objectives, including constrained goal regions, unconstrained goal sets, and energy-based goals.We show that our method can use these objectives to successfully direct behavior.Our method substantially outperforms six IL approaches and a planning-based approach in a dynamic simulated autonomous driving task, and is efficiently learned from expert demonstrations without online data collection. We also show our approach is robust to poorly-specified goals, such as goals on the wrong side of the road.", "target": ["In diesem Beitrag schlagen wir Imitative Modelle vor, um die Vorteile von IL und zielgerichteter Planung zu kombinieren: probabilistische Vorhersagemodelle für wünschenswertes Verhalten, die in der Lage sind, interpretierbare expertenähnliche Trajektorien zu planen, um bestimmte Ziele zu erreichen."]} +{"source": "Learning communication via deep reinforcement learning has recently been shown to be an effective way to solve cooperative multi-agent tasks.However, learning which communicated information is beneficial for each agent's decision-making remains a challenging task.In order to address this problem, we introduce a fully differentiable framework for communication and reasoning, enabling agents to solve cooperative tasks in partially-observable environments.The framework is designed to facilitate explicit reasoning between agents, through a novel memory-based attention network that can learn selectively from its past memories.The model communicates through a series of reasoning steps that decompose each agent's intentions into learned representations that are used first to compute the relevance of communicated information, and second to extract information from memories given newly received information.By selectively interacting with new information, the model effectively learns a communication protocol directly, in an end-to-end manner.We empirically demonstrate the strength of our model in cooperative multi-agent tasks, where inter-agent communication and reasoning over prior information substantially improves performance compared to baselines.", "target": ["Neuartige Architektur eines speicherbasierten Aufmerksamkeitsmechanismus für Multi-Agenten-Kommunikation."]} +{"source": "In this paper, we introduce Symplectic ODE-Net (SymODEN), a deep learning framework which can infer the dynamics of a physical system from observed state trajectories.To achieve better generalization with fewer training samples, SymODEN incorporates appropriate inductive bias by designing the associated computation graph in a physics-informed manner.In particular, we enforce Hamiltonian dynamics with control to learn the underlying dynamics in a transparent way which can then be leveraged to draw insight about relevant physical aspects of the system, such as mass and potential energy.In addition, we propose a parametrization which can enforce this Hamiltonian formalism even when the generalized coordinate data is embedded in a high-dimensional space or we can only access velocity data instead of generalized momentum.This framework, by offering interpretable, physically-consistent models for physical systems, opens up new possibilities for synthesizing model-based control strategies.", "target": ["Diese Arbeit erzwingt eine Hamilton-Dynamik mit Steuerung, um Systemmodelle aus eingebetteten Positions- und Geschwindigkeitsdaten zu lernen, und nutzt diese physikalisch konsistente Dynamik, um eine modellbasierte Steuerung durch Energieformung zu synthetisieren."]} +{"source": "Federated learning, where a global model is trained by iterative parameter averaging of locally-computed updates, is a promising approach for distributed training of deep networks; it provides high communication-efficiency and privacy-preservability, which allows to fit well into decentralized data environments, e.g., mobile-cloud ecosystems.However, despite the advantages, the federated learning-based methods still have a challenge in dealing with non-IID training data of local devices (i.e., learners).In this regard, we study the effects of a variety of hyperparametric conditions under the non-IID environments, to answer important concerns in practical implementations:(i) We first investigate parameter divergence of local updates to explain performance degradation from non-IID data.The origin of the parameter divergence is also found both empirically and theoretically.(ii) We then revisit the effects of optimizers, network depth/width, and regularization techniques; our observations show that the well-known advantages of the hyperparameter optimization strategies could rather yield diminishing returns with non-IID data.(iii) We finally provide the reasons of the failure cases in a categorized way, mainly based on metrics of the parameter divergence.", "target": ["Wir untersuchen die internen Gründe für unsere Beobachtungen, nämlich die abnehmenden Auswirkungen der bekannten Hyperparameter-Optimierungsmethoden auf das föderierte Lernen aus dezentralen Nicht-IID-Daten."]} +{"source": "Deep learning models are known to be vulnerable to adversarial examples.A practical adversarial attack should require as little as possible knowledge of attacked models T. Current substitute attacks need pre-trained models to generate adversarial examples and their attack success rates heavily rely on the transferability of adversarial examples.Current score-based and decision-based attacks require lots of queries for the T. In this study, we propose a novel adversarial imitation attack.First, it produces a replica of the T by a two-player game like the generative adversarial networks (GANs).The objective of the generative model G is to generate examples which lead D returning different outputs with T. The objective of the discriminative model D is to output the same labels with T under the same inputs.Then, the adversarial examples generated by D are utilized to fool the T. Compared with the current substitute attacks, imitation attack can use less training data to produce a replica of T and improve the transferability of adversarial examples.Experiments demonstrate that our imitation attack requires less training data than the black-box substitute attacks, but achieves an attack success rate close to the white-box attack on unseen data with no query.", "target": ["Ein neuartiger Nachahmungsangriff zur Täuschung von Modellen des maschinellen Lernens."]} +{"source": "Stochastic Gradient Descent (SGD) methods using randomly selected batches are widely-used to train neural network (NN) models.Performing design exploration to find the best NN for a particular task often requires extensive training with different models on a large dataset, which is very computationally expensive.The most straightforward method to accelerate this computation is to distribute the batch of SGD over multiple processors.However, large batch training often times leads to degradation in accuracy, poor generalization, and even poor robustness to adversarial attacks. Existing solutions for large batch training either do not work or require massive hyper-parameter tuning.To address this issue, we propose a novel large batch training method which combines recent results in adversarial training (to regularize against ``sharp minima'') and second order optimization (to use curvature information to change batch size adaptively during training).We extensively evaluate our method on Cifar-10/100, SVHN, TinyImageNet, and ImageNet datasets, using multiple NNs, including residual networks as well as compressed networks such as SqueezeNext. Our new approach exceeds the performance of the existing solutions in terms of both accuracy and the number of SGD iterations (up to 1\\% and $3\\times$, respectively).We emphasize that this is achieved without any additional hyper-parameter tuning to tailor our method to any of these experiments.", "target": ["Training mit großen Batches unter Verwendung von adversarialem Training und Informationen zweiter Ordnung"]} +{"source": "Inspired by neurophysiological discoveries of navigation cells in the mammalianbrain, we introduce the first deep neural network architecture for modeling EgocentricSpatial Memory (ESM).It learns to estimate the pose of the agent andprogressively construct top-down 2D global maps from egocentric views in a spatiallyextended environment.During the exploration, our proposed ESM networkmodel updates belief of the global map based on local observations using a recurrentneural network.It also augments the local mapping with a novel externalmemory to encode and store latent representations of the visited places based ontheir corresponding locations in the egocentric coordinate.This enables the agentsto perform loop closure and mapping correction.This work contributes in thefollowing aspects: first, our proposed ESM network provides an accurate mappingability which is vitally important for embodied agents to navigate to goal locations.In the experiments, we demonstrate the functionalities of the ESM network inrandom walks in complicated 3D mazes by comparing with several competitivebaselines and state-of-the-art Simultaneous Localization and Mapping (SLAM)algorithms.Secondly, we faithfully hypothesize the functionality and the workingmechanism of navigation cells in the brain.Comprehensive analysis of our modelsuggests the essential role of individual modules in our proposed architecture anddemonstrates efficiency of communications among these modules.We hope thiswork would advance research in the collaboration and communications over bothfields of computer science and computational neuroscience.", "target": ["erstes tiefes neuronales Netz zur Modellierung des egozentrischen räumlichen Gedächtnisses, inspiriert durch neurophysiologische Entdeckungen von Navigationszellen im Säugetiergehirn"]} +{"source": "We show that if the usual training loss is augmented by a Lipschitz regularization term, then the networks generalize. We prove generalization by first establishing a stronger convergence result, along with a rate of convergence. A second result resolves a question posed in Zhang et al. (2016): how can a model distinguish between the case of clean labels, and randomized labels? Our answer is that Lipschitz regularization using the Lipschitz constant of the clean data makes this distinction. In this case, the model learns a different function which we hypothesize correctly fails to learn the dirty labels.", "target": ["Wir beweisen die Generalisierung von DNNs durch Hinzufügen eines Lipschitz-Regularisierungsterms zum Trainingsverlust. Wir lösen eine Frage, die in Zhang et al. (2016) gestellt wurde."]} +{"source": "For fast and energy-efficient deployment of trained deep neural networks on resource-constrained embedded hardware, each learned weight parameter should ideally be represented and stored using a single bit. Error-rates usually increase when this requirement is imposed.Here, we report large improvements in error rates on multiple datasets, for deep convolutional neural networks deployed with 1-bit-per-weight.Using wide residual networks as our main baseline, our approach simplifies existing methods that binarize weights by applying the sign function in training; we apply scaling factors for each layer with constant unlearned values equal to the layer-specific standard deviations used for initialization.For CIFAR-10, CIFAR-100 and ImageNet, and models with 1-bit-per-weight requiring less than 10 MB of parameter memory, we achieve error rates of 3.9%, 18.5% and 26.0% / 8.5% (Top-1 / Top-5) respectively.We also considered MNIST, SVHN and ImageNet32, achieving 1-bit-per-weight test results of 0.27%, 1.9%, and 41.3% / 19.1% respectively.For CIFAR, our error rates halve previously reported values, and are within about 1% of our error-rates for the same network with full-precision weights.For networks that overfit, we also show significant improvements in error rate by not learning batch normalization scale and offset parameters.This applies to both full precision and 1-bit-per-weight networks.Using a warm-restart learning-rate schedule, we found that training for 1-bit-per-weight is just as fast as full-precision networks, with better accuracy than standard schedules, and achieved about 98%-99% of peak performance in just 62 training epochs for CIFAR-10/100.For full training code and trained models in MATLAB, Keras and PyTorch see https://github.com/McDonnell-Lab/1-bit-per-weight/ .", "target": ["Wir trainieren breite Residualnetzwerke, die sofort eingesetzt werden können, indem wir nur ein einziges Bit für jedes Convolutional Gewicht verwenden, und zwar mit deutlich besserer Genauigkeit als bisherige Methoden."]} +{"source": "This paper presents a system for immersive visualization of Non-Euclidean spaces using real-time ray tracing.It exploits the capabilities of the new generation of GPU’s based on the NVIDIA’s Turing architecture in order to develop new methods for intuitive exploration of landscapes featuring non-trivial geometry and topology in virtual reality.", "target": ["Immersive Visualisierung der klassischen nicht-euklidischen Räume mit Echtzeit-Raytracing."]} +{"source": "We propose the set autoencoder, a model for unsupervised representation learning for sets of elements.It is closely related to sequence-to-sequence models, which learn fixed-sized latent representations for sequences, and have been applied to a number of challenging supervised sequence tasks such as machine translation, as well as unsupervised representation learning for sequences.In contrast to sequences, sets are permutation invariant.The proposed set autoencoder considers this fact, both with respect to the input as well as the output of the model.On the input side, we adapt a recently-introduced recurrent neural architecture using a content-based attention mechanism.On the output side, we use a stable marriage algorithm to align predictions to labels in the learning phase.We train the model on synthetic data sets of point clouds and show that the learned representations change smoothly with translations in the inputs, preserve distances in the inputs, and that the set size is represented directly.We apply the model to supervised tasks on the point clouds using the fixed-size latent representation.For a number of difficult classification problems, the results are better than those of a model that does not consider the permutation invariance.Especially for small training sets, the set-aware model benefits from unsupervised pretraining.", "target": ["Wir schlagen den Set-Autoencoder vor, ein Modell zum unbeaufsichtigten Lernen von Repräsentationen für Mengen von Elementen."]} +{"source": "Deep reinforcement learning algorithms can learn complex behavioral skills, but real-world application of these methods requires a considerable amount of experience to be collected by the agent.In practical settings, such as robotics, this involves repeatedly attempting a task, resetting the environment between each attempt.However, not all tasks are easily or automatically reversible.In practice, this learning process requires considerable human intervention.In this work, we propose an autonomous method for safe and efficient reinforcement learning that simultaneously learns a forward and backward policy, with the backward policy resetting the environment for a subsequent attempt.By learning a value function for the backward policy, we can automatically determine when the forward policy is about to enter a non-reversible state, providing for uncertainty-aware safety aborts.Our experiments illustrate that proper use of the backward policy can greatly reduce the number of manual resets required to learn a task and can reduce the number of unsafe actions that lead to non-reversible states.", "target": ["Wir schlagen eine autonome Methode für sicheres und effizientes Reinforcement Learning vor, die gleichzeitig eine Vorwärts- und eine Rückwärtsstrategie erlernt, wobei die Rückwärtsstrategie die Umgebung für einen nachfolgenden Versuch zurücksetzt."]} +{"source": "It has been argued that current machine learning models do not have commonsense, and therefore must be hard-coded with prior knowledge (Marcus, 2018).Here we show surprising evidence that language models can already learn to capture certain common sense knowledge.Our key observation is that a language model can compute the probability of any statement, and this probability can be used to evaluate the truthfulness of that statement. On the Winograd Schema Challenge (Levesque et al., 2011), language models are 11% higher in accuracy than previous state-of-the-art supervised methods.Language models can also be fine-tuned for the task of Mining Commonsense Knowledge on ConceptNet to achieve an F1 score of 0.912 and 0.824, outperforming previous best results (Jastrzebskiet al., 2018). Further analysis demonstrates that language models can discover unique features of Winograd Schema contexts that decide the correct answers without explicit supervision.", "target": ["Wir präsentieren Beweise dafür, dass LMs den gesunden Menschenverstand erfassen, mit Spitzenergebnissen sowohl bei der Winograd Schema Challenge als auch beim Commonsense Knowledge Mining."]} +{"source": "In many real-world learning scenarios, features are only acquirable at a cost constrained under a budget.In this paper, we propose a novel approach for cost-sensitive feature acquisition at the prediction-time.The suggested method acquires features incrementally based on a context-aware feature-value function.We formulate the problem in the reinforcement learning paradigm, and introduce a reward function based on the utility of each feature.Specifically, MC dropout sampling is used to measure expected variations of the model uncertainty which is used as a feature-value function.Furthermore, we suggest sharing representations between the class predictor and value function estimator networks.The suggested approach is completely online and is readily applicable to stream learning setups.The solution is evaluated on three different datasets including the well-known MNIST dataset as a benchmark as well as two cost-sensitive datasets: Yahoo Learning to Rank and a dataset in the medical domain for diabetes classification.According to the results, the proposed method is able to efficiently acquire features and make accurate predictions.", "target": ["Ein Online-Algorithmus für die kostenbewusste Erfassung und Vorhersage von Merkmalen"]} +{"source": "This paper revisits the problem of sequence modeling using convolutional architectures. Although both convolutional and recurrent architectures havealong history in sequence prediction, the current \"default\" mindset in much ofthe deep learning community is that generic sequence modeling is best handledusing recurrent networks. The goal of this paper is to question this assumption. Specifically, we consider a simple generic temporal convolution network (TCN),which adopts features from modern ConvNet architectures such as a dilations and residual connections. We show that on a variety of sequence modeling tasks,including many frequently used as benchmarks for evaluating recurrent networks,the TCN outperforms baseline RNN methods (LSTMs, GRUs, and vanilla RNNs) andsometimes even highly specialized approaches. We further show that thepotential \"infinite memory\" advantage that RNNs have over TCNs is largelyabsent in practice: TCNs indeed exhibit longer effective history sizes than their recurrent counterparts. As a whole, we argue that it may be time to (re)consider ConvNets as the default \"go to\" architecture for sequence modeling.", "target": ["Wir argumentieren, dass Covolutional Networks als Standardausgangspunkt für Sequenzmodellierungsaufgaben betrachtet werden sollten."]} +{"source": "Deep neural networks work well at approximating complicated functions when provided with data and trained by gradient descent methods.At the same time, there is a vast amount of existing functions that programmatically solve different tasks in a precise manner eliminating the need for training.In many cases, it is possible to decompose a task to a series of functions, of which for some we may prefer to use a neural network to learn the functionality, while for others the preferred method would be to use existing black-box functions.We propose a method for end-to-end training of a base neural network that integrates calls to existing black-box functions.We do so by approximating the black-box functionality with a differentiable neural network in a way that drives the base network to comply with the black-box function interface during the end-to-end optimization process.At inference time, we replace the differentiable estimator with its external black-box non-differentiable counterpart such that the base network output matches the input arguments of the black-box function.Using this ``Estimate and Replace'' paradigm, we train a neural network, end to end, to compute the input to black-box functionality while eliminating the need for intermediate labels.We show that by leveraging the existing precise black-box function during inference, the integrated model generalizes better than a fully differentiable model, and learns more efficiently compared to RL-based methods.", "target": ["Training von DNNs, um Black-Box-Funktionen mit Zwischenetiketten zu verbinden, unter Verwendung eines Schätzer-Teilnetzes, das nach dem Training durch die Black-Box ersetzt werden kann"]} +{"source": "This paper proposes a new actor-critic-style algorithm called Dual Actor-Critic or Dual-AC. It is derived in a principled way from the Lagrangian dual form of the Bellman optimality equation, which can be viewed as a two-player game between the actor and a critic-like function, which is named as dual critic. Compared to its actor-critic relatives, Dual-AC has the desired property that the actor and dual critic are updated cooperatively to optimize the same objective function, providing a more transparent way for learning the critic that is directly related to the objective function of the actor.We then provide a concrete algorithm that can effectively solve the minimax optimization problem, using techniques of multi-step bootstrapping, path regularization, and stochastic dual ascent algorithm.We demonstrate that the proposed algorithm achieves the state-of-the-art performances across several benchmarks.", "target": ["Wir schlagen einen dualen akteurskritischen Algorithmus vor, der auf prinzipielle Weise aus der dualen Lagrangeschen Form der Bellman-Optimalitätsgleichung abgeleitet ist. Der Algorithmus erreicht in mehreren Benchmarks die beste Leistung."]} +{"source": "Training a model to perform a task typically requires a large amount of data from the domains in which the task will be applied.However, it is often the case that data are abundant in some domains but scarce in others.Domain adaptation deals with the challenge of adapting a model trained from a data-rich source domain to perform well in a data-poor target domain.In general, this requires learning plausible mappings between domains.CycleGAN is a powerful framework that efficiently learns to map inputs from one domain to another using adversarial training and a cycle-consistency constraint.However, the conventional approach of enforcing cycle-consistency via reconstruction may be overly restrictive in cases where one or more domains have limited training data.In this paper, we propose an augmented cyclic adversarial learning model that enforces the cycle-consistency constraint via an external task specific model, which encourages the preservation of task-relevant content as opposed to exact reconstruction.This task specific model both relaxes the cycle-consistency constraint and complements the role of the discriminator during training, serving as an augmented information source for learning the mapping.We explore adaptation in speech and visual domains in low resource in supervised setting.In speech domains, we adopt a speech recognition model from each domain as the task specific model.Our approach improves absolute performance of speech recognition by 2% for female speakers in the TIMIT dataset, where the majority of training samples are from male voices.In low-resource visual domain adaptation, the results show that our approach improves absolute performance by 14% and 4% when adapting SVHN to MNIST and vice versa, respectively, which outperforms unsupervised domain adaptation methods that require high-resource unlabeled target domain.", "target": ["Eine robuste Domänenanpassung durch den Einsatz eines aufgabenspezifischen Verlustes beim zyklischen adversen Lernen"]} +{"source": "The success of popular algorithms for deep reinforcement learning, such as policy-gradients and Q-learning, relies heavily on the availability of an informative reward signal at each timestep of the sequential decision-making process.When rewards are only sparsely available during an episode, or a rewarding feedback is provided only after episode termination, these algorithms perform sub-optimally due to the difficultly in credit assignment.Alternatively, trajectory-based policy optimization methods, such as cross-entropy method and evolution strategies, do not require per-timestep rewards, but have been found to suffer from high sample complexity by completing forgoing the temporal nature of the problem.Improving the efficiency of RL algorithms in real-world problems with sparse or episodic rewards is therefore a pressing need.In this work, we introduce a self-imitation learning algorithm that exploits and explores well in the sparse and episodic reward settings.We view each policy as a state-action visitation distribution and formulate policy optimization as a divergence minimization problem.We show that with Jensen-Shannon divergence, this divergence minimization problem can be reduced into a policy-gradient algorithm with shaped rewards learned from experience replays.Experimental results indicate that our algorithm works comparable to existing algorithms in environments with dense rewards, and significantly better in environments with sparse and episodic rewards.We then discuss limitations of self-imitation learning, and propose to solve them by using Stein variational policy gradient descent with the Jensen-Shannon kernel to learn multiple diverse policies.We demonstrate its effectiveness on a challenging variant of continuous-control MuJoCo locomotion tasks.", "target": ["Optimierung der Politik durch Verwendung vergangener guter Rollouts des Agenten; Lernen von geformten Belohnungen über Divergenzminimierung; SVPG mit JS-Kernel für populationsbasierte Exploration."]} +{"source": "We study the precise mechanisms which allow autoencoders to encode and decode a simple geometric shape, the disk.In this carefully controlled setting, we are able to describe the specific form of the optimal solution to the minimisation problem of the training step.We show that the autoencoder indeed approximates this solution during training.Secondly, we identify a clear failure in the generalisation capacity of the autoencoder, namely its inability to interpolate data.Finally, we explore several regularisation schemes to resolve the generalisation problem.Given the great attention that has been recently given to the generative capacity of neural networks, we believe that studying in depth simple geometric cases sheds some light on the generation process and can provide a minimal requirement experimental setup for more complex architectures.", "target": ["Wir untersuchen die Funktionsweise von Autoencodern in einer einfachen Umgebung und empfehlen neue Strategien für deren Regularisierung, um eine bessere Generalisierung mit latenter Interpolation für die Bildsynthese zu erreichen."]} +{"source": "We present a simple idea that allows to record a speaker in a given language and synthesize their voice in other languages that they may not even know.These techniques open a wide range of potential applications such as cross-language communication, language learning or automatic video dubbing.We call this general problem multi-language speaker-conditioned speech synthesis and we present a simple but strong baseline for it.Our model architecture is similar to the encoder-decoder Char2Wav model or Tacotron.The main difference is that, instead of conditioning on characters or phonemes that are specific to a given language, we condition on a shared phonetic representation that is universal to all languages.This cross-language phonetic representation of text allows to synthesize speech in any language while preserving the vocal characteristics of the original speaker.Furthermore, we show that fine-tuning the weights of our model allows us to extend our results to speakers outside of the training dataset.", "target": ["Wir stellen eine einfache Idee vor, die es ermöglicht, einen Sprecher in einer bestimmten Sprache aufzunehmen und seine Stimme in anderen Sprachen zu synthetisieren, die er vielleicht nicht einmal kennt."]} +{"source": "The goal of imitation learning (IL) is to enable a learner to imitate expert behavior given expert demonstrations.Recently, generative adversarial imitation learning (GAIL) has shown significant progress on IL for complex continuous tasks.However, GAIL and its extensions require a large number of environment interactions during training.In real-world environments, the more an IL method requires the learner to interact with the environment for better imitation, the more training time it requires, and the more damage it causes to the environments and the learner itself.We believe that IL algorithms could be more applicable to real-world problems if the number of interactions could be reduced. In this paper, we propose a model-free IL algorithm for continuous control.Our algorithm is made up mainly three changes to the existing adversarial imitation learning (AIL) methods –(a) adopting off-policy actor-critic (Off-PAC) algorithm to optimize the learner policy,(b) estimating the state-action value using off-policy samples without learning reward functions, and(c) representing the stochastic policy function so that its outputs are bounded.Experimental results show that our algorithm achieves competitive results with GAIL while significantly reducing the environment interactions.", "target": ["In dieser Arbeit haben wir einen modellfreien, richtlinienunabhängigen IL-Algorithmus für kontinuierliche Kontrolle vorgeschlagen. Experimentelle Ergebnisse zeigten, dass unser Algorithmus wettbewerbsfähige Ergebnisse mit GAIL erzielt und gleichzeitig die Interaktionen mit der Umgebung deutlich reduziert."]} +{"source": "The dominant approach to unsupervised \"style transfer'' in text is based on the idea of learning a latent representation, which is independent of the attributes specifying its \"style''.In this paper, we show that this condition is not necessary and is not always met in practice, even with domain adversarial training that explicitly aims at learning such disentangled representations.We thus propose a new model that controls several factors of variation in textual data where this condition on disentanglement is replaced with a simpler mechanism based on back-translation.Our method allows control over multiple attributes, like gender, sentiment, product type, etc., and a more fine-grained control on the trade-off between content preservation and change of style with a pooling operator in the latent space.Our experiments demonstrate that the fully entangled model produces better generations, even when tested on new and more challenging benchmarks comprising reviews with multiple sentences and multiple attributes.", "target": ["Ein System zum Umschreiben von Text, der von mehreren kontrollierbaren Attributen abhängt"]} +{"source": "Vanilla RNN with ReLU activation have a simple structure that is amenable to systematic dynamical systems analysis and interpretation, but they suffer from the exploding vs. vanishing gradients problem.Recent attempts to retain this simplicity while alleviating the gradient problem are based on proper initialization schemes or orthogonality/unitary constraints on the RNN’s recurrency matrix, which, however, comes with limitations to its expressive power with regards to dynamical systems phenomena like chaos or multi-stability.Here, we instead suggest a regularization scheme that pushes part of the RNN’s latent subspace toward a line attractor configuration that enables long short-term memory and arbitrarily slow time scales.We show that our approach excels on a number of benchmarks like the sequential MNIST or multiplication problems, and enables reconstruction of dynamical systems which harbor widely different time scales.", "target": ["Wir entwickeln einen neuen Optimierungsansatz für einfache ReLU-basierte RNN, der ein langes Kurzzeitgedächtnis und die Identifikation beliebiger nichtlinearer dynamischer Systeme mit sehr unterschiedlichen Zeitskalen ermöglicht."]} +{"source": "In the field of Generative Adversarial Networks (GANs), how to design a stable training strategy remains an open problem.Wasserstein GANs have largely promoted the stability over the original GANs by introducing Wasserstein distance, but still remain unstable and are prone to a variety of failure modes.In this paper, we present a general framework named Wasserstein-Bounded GAN (WBGAN), which improves a large family of WGAN-based approaches by simply adding an upper-bound constraint to the Wasserstein term.Furthermore, we show that WBGAN can reasonably measure the difference of distributions which almost have no intersection.Experiments demonstrate that WBGAN can stabilize as well as accelerate convergence in the training processes of a series of WGAN-based variants.", "target": ["Vorschlag eines verbesserten Rahmens für WGANs und Nachweis seiner besseren Leistung in Theorie und Praxis."]} +{"source": "Generative models such as Variational Auto Encoders (VAEs) and Generative Adversarial Networks (GANs) are typically trained for a fixed prior distribution in the latent space, such as uniform or Gaussian.After a trained model is obtained, one can sample the Generator in various forms for exploration and understanding, such as interpolating between two samples, sampling in the vicinity of a sample or exploring differences between a pair of samples applied to a third sample.In this paper, we show that the latent space operations used in the literature so far induce a distribution mismatch between the resulting outputs and the prior distribution the model was trained on.To address this, we propose to use distribution matching transport maps to ensure that such latent space operations preserve the prior distribution, while minimally modifying the original operation. Our experimental results validate that the proposed operations give higher quality samples compared to the original operations.", "target": ["Operationen im latenten Raum des GAN können zu einer Ungleichverteilung im Vergleich zur Trainingsverteilung führen."]} +{"source": "Neural program embeddings have shown much promise recently for a variety of program analysis tasks, including program synthesis, program repair, code completion, and fault localization.However, most existing program embeddings are based on syntactic features of programs, such as token sequences or abstract syntax trees.Unlike images and text, a program has well-defined semantics that can be difficult to capture by only considering its syntax (i.e. syntactically similar programs can exhibit vastly different run-time behavior), which makes syntax-based program embeddings fundamentally limited.We propose a novel semantic program embedding that is learned from program execution traces.Our key insight is that program states expressed as sequential tuples of live variable values not only capture program semantics more precisely, but also offer a more natural fit for Recurrent Neural Networks to model.We evaluate different syntactic and semantic program embeddings on the task of classifying the types of errors that students make in their submissions to an introductory programming class and on the CodeHunt education platform.Our evaluation results show that the semantic program embeddings significantly outperform the syntactic program embeddings based on token sequences and abstract syntax trees.In addition, we augment a search-based program repair system with predictions made from our semantic embedding and demonstrate significantly improved search efficiency.", "target": ["Ein neuer Weg zum Erlernen der semantischen Programmeinbettung"]} +{"source": "In this paper, we propose a generalization of the BN algorithm, diminishing batch normalization (DBN), where we update the BN parameters in a diminishing moving average way.Batch normalization (BN) is very effective in accelerating the convergence of a neural network training phase that it has become a common practice. Our proposed DBN algorithm remains the overall structure of the original BN algorithm while introduces a weighted averaging update to some trainable parameters. We provide an analysis of the convergence of the DBN algorithm that converges to a stationary point with respect to trainable parameters.Our analysis can be easily generalized for original BN algorithm by setting some parameters to constant.To the best knowledge of authors, this analysis is the first of its kind for convergence with Batch Normalization introduced.We analyze a two-layer model with arbitrary activation function. The primary challenge of the analysis is the fact that some parameters are updated by gradient while others are not. The convergence analysis applies to any activation function that satisfies our common assumptions.For the analysis, we also show the sufficient and necessary conditions for the stepsizes and diminishing weights to ensure the convergence. In the numerical experiments, we use more complex models with more layers and ReLU activation.We observe that DBN outperforms the original BN algorithm on Imagenet, MNIST, NI and CIFAR-10 datasets with reasonable complex FNN and CNN models.", "target": ["Wir schlagen eine Erweiterung der Batch-Normalisierung vor, zeigen eine erstmalige Konvergenzanalyse für diese Erweiterung und zeigen in numerischen Experimenten, dass sie eine bessere Leistung hat als die ursprüngliche Batch-Normalisierung."]} +{"source": "Generative models such as Variational Auto Encoders (VAEs) and Generative Adversarial Networks (GANs) are typically trained for a fixed prior distribution in the latent space, such as uniform or Gaussian.After a trained model is obtained, one can sample the Generator in various forms for exploration and understanding, such as interpolating between two samples, sampling in the vicinity of a sample or exploring differences between a pair of samples applied to a third sample.However, the latent space operations commonly used in the literature so far induce a distribution mismatch between the resulting outputs and the prior distribution the model was trained on.Previous works have attempted to reduce this mismatch with heuristic modification to the operations or by changing the latent distribution and re-training models.In this paper, we propose a framework for modifying the latent space operations such that the distribution mismatch is fully eliminated.Our approach is based on optimal transport maps, which adapt the latent space operations such that they fully match the prior distribution, while minimally modifying the original operation.Our matched operations are readily obtained for the commonly used operations and distributions and require no adjustment to the training procedure.", "target": ["Wir schlagen einen Rahmen für die Modifizierung der latenten Raumoperationen vor, so dass die Verteilungsfehlanpassung zwischen den resultierenden Ausgaben und der vorherigen Verteilung, auf der das generative Modell trainiert wurde, vollständig beseitigt wird."]} +{"source": "The problem of building a coherent and non-monotonous conversational agent with proper discourse and coverage is still an area of open research.Current architectures only take care of semantic and contextual information for a given query and fail to completely account for syntactic and external knowledge which are crucial for generating responses in a chit-chat system.To overcome this problem, we propose an end to end multi-stream deep learning architecture which learns unified embeddings for query-response pairs by leveraging contextual information from memory networks and syntactic information by incorporating Graph Convolution Networks (GCN) over their dependency parse.A stream of this network also utilizes transfer learning by pre-training a bidirectional transformer to extract semantic representation for each input sentence and incorporates external knowledge through the neighbourhood of the entities from a Knowledge Base (KB).We benchmark these embeddings on next sentence prediction task and significantly improve upon the existing techniques.Furthermore, we use AMUSED to represent query and responses along with its context to develop a retrieval based conversational agent which has been validated by expert linguists to have comprehensive engagement with humans.", "target": ["Dieses Papier bietet einen Multi-Stream-End-to-End-Ansatz, um einheitliche Einbettungen für Frage-Antwort-Paare in Dialogsystemen zu lernen, indem kontextuelle, syntaktische, semantische und externe Informationen zusammen genutzt werden."]} +{"source": "We examine techniques for combining generalized policies with search algorithms to exploit the strengths and overcome the weaknesses of each when solving probabilistic planning problems.The Action Schema Network (ASNet) is a recent contribution to planning that uses deep learning and neural networks to learn generalized policies for probabilistic planning problems.ASNets are well suited to problems where local knowledge of the environment can be exploited to improve performance, but may fail to generalize to problems they were not trained on.Monte-Carlo Tree Search (MCTS) is a forward-chaining state space search algorithm for optimal decision making which performs simulations to incrementally build a search tree and estimate the values of each state.Although MCTS can achieve state-of-the-art results when paired with domain-specific knowledge, without this knowledge, MCTS requires a large number of simulations in order to obtain reliable estimates in the search tree.By combining ASNets with MCTS, we are able to improve the capability of an ASNet to generalize beyond the distribution of problems it was trained on, as well as enhance the navigation of the search space by MCTS.", "target": ["Techniken zur Kombination verallgemeinerter Strategien mit Suchalgorithmen, um die Stärken und Schwächen beider Verfahren bei der Lösung probabilistischer Planungsprobleme zu nutzen"]} +{"source": "Modern deep neural networks can achieve high accuracy when the training distribution and test distribution are identically distributed, but this assumption is frequently violated in practice.When the train and test distributions are mismatched, accuracy can plummet.Currently there are few techniques that improve robustness to unforeseen data shifts encountered during deployment.In this work, we propose a technique to improve the robustness and uncertainty estimates of image classifiers.We propose AugMix, a data processing technique that is simple to implement, adds limited computational overhead, and helps models withstand unforeseen corruptions.AugMix significantly improves robustness and uncertainty measures on challenging image classification benchmarks, closing the gap between previous methods and the best possible performance in some cases by more than half.", "target": ["Wir erreichen den neuesten Stand der Technik in Bezug auf die Robustheit gegenüber Datenverschiebungen und behalten die Kalibrierung auch dann bei, wenn die Genauigkeit abnimmt."]} +{"source": "Automatic Piano Fingering is a hard task which computers can learn using data.As data collection is hard and expensive, we propose to automate this process by automatically extracting fingerings from public videos and MIDI files, using computer-vision techniques.Running this process on 90 videos results in the largest dataset for piano fingering with more than 150K notes.We show that when running a previously proposed model for automatic piano fingering on our dataset and then fine-tuning it on manually labeled piano fingering data, we achieve state-of-the-art results.In addition to the fingering extraction method, we also introduce a novel method for transferring deep-learning computer-vision models to work on out-of-domain data, by fine-tuning it on out-of-domain augmentation proposed by a Generative Adversarial Network (GAN).For demonstration, we anonymously release a visualization of the output of our process for a single video on https://youtu.be/Gfs1UWQhr5Q", "target": ["Wir extrahieren automatisch Informationen über den Fingersatz aus Videos von Klavierspielen, um sie in automatischen Modellen zur Vorhersage des Fingersatzes zu verwenden."]} +{"source": "Domain adaptation refers to the problem of leveraging labeled data in a source domain to learn an accurate model in a target domain where labels are scarce or unavailable.A recent approach for finding a common representation of the two domains is via domain adversarial training (Ganin & Lempitsky, 2015), which attempts to induce a feature extractor that matches the source and target feature distributions in some feature space.However, domain adversarial training faces two critical limitations:1) if the feature extraction function has high-capacity, then feature distribution matching is a weak constraint,2) in non-conservative domain adaptation (where no single classifier can perform well in both the source and target domains), training the model to do well on the source domain hurts performance on the target domain.In this paper, we address these issues through the lens of the cluster assumption, i.e., decision boundaries should not cross high-density data regions.We propose two novel and related models:1) the Virtual Adversarial Domain Adaptation (VADA) model, which combines domain adversarial training with a penalty term that punishes the violation the cluster assumption;2) the Decision-boundary Iterative Refinement Training with a Teacher (DIRT-T) model, which takes the VADA model as initialization and employs natural gradient steps to further minimize the cluster assumption violation.Extensive empirical results demonstrate that the combination of these two models significantly improve the state-of-the-art performance on the digit, traffic sign, and Wi-Fi recognition domain adaptation benchmarks.", "target": ["SOTA auf unüberwachter Domänenanpassung durch Ausnutzung der Cluster-Annahme."]} +{"source": "In this paper, we propose Continuous Graph Flow, a generative continuous flow based method that aims to model complex distributions of graph-structured data. Once learned, the model can be applied to an arbitrary graph, defining a probability density over the random variables represented by the graph.It is formulated as an ordinary differential equation system with shared and reusable functions that operate over the graphs. This leads to a new type of neural graph message passing scheme that performs continuous message passing over time.This class of models offers several advantages: a flexible representation that can generalize to variable data dimensions; ability to model dependencies in complex data distributions; reversible and memory-efficient; and exact and efficient computation of the likelihood of the data.We demonstrate the effectiveness of our model on a diverse set of generation tasks across different domains: graph generation, image puzzle generation, and layout generation from scene graphs.Our proposed model achieves significantly better performance compared to state-of-the-art models.", "target": ["Generative Graphenmodelle auf der Grundlage der Verallgemeinerung der Nachrichtenübermittlung auf kontinuierliche Zeit unter Verwendung gewöhnlicher Differentialgleichungen"]} +{"source": "The practical successes of deep neural networks have not been matched by theoretical progress that satisfyingly explains their behavior.In this work, we study the information bottleneck (IB) theory of deep learning, which makes three specific claims: first, that deep networks undergo two distinct phases consisting of an initial fitting phase and a subsequent compression phase; second, that the compression phase is causally related to the excellent generalization performance of deep networks; and third, that the compression phase occurs due to the diffusion-like behavior of stochastic gradient descent.Here we show that none of these claims hold true in the general case.Through a combination of analytical results and simulation, we demonstrate that the information plane trajectory is predominantly a function of the neural nonlinearity employed: double-sided saturating nonlinearities like tanh yield a compression phase as neural activations enter the saturation regime, but linear activation functions and single-sided saturating nonlinearities like the widely used ReLU in fact do not.Moreover, we find that there is no evident causal connection between compression and generalization: networks that do not compress are still capable of generalization, and vice versa.Next, we show that the compression phase, when it exists, does not arise from stochasticity in training by demonstrating that we can replicate the IB findings using full batch gradient descent rather than stochastic gradient descent.Finally, we show that when an input domain consists of a subset of task-relevant and task-irrelevant information, hidden representations do compress the task-irrelevant information, although the overall information about the input may monotonically increase with training time, and that this compression happens concurrently with the fitting process rather than during a subsequent compression period.", "target": ["Wir zeigen, dass mehrere Behauptungen der Theorie des Informationsengpasses beim Deep Learning im allgemeinen Fall nicht zutreffen."]} +{"source": "Over the past four years, neural networks have been proven vulnerable to adversarial images: targeted but imperceptible image perturbations lead to drastically different predictions.We show that adversarial vulnerability increases with the gradients of the training objective when viewed as a function of the inputs.For most current network architectures, we prove that the L1-norm of these gradients grows as the square root of the input size.These nets therefore become increasingly vulnerable with growing image size.Our proofs rely on the network’s weight distribution at initialization, but extensive experiments confirm that our conclusions still hold after usual training.", "target": ["Neuronale Netze haben von Haus aus große Gradienten; das macht sie anfällig für negative Einflüsse."]} +{"source": "Effective performance of neural networks depends critically on effective tuning of optimization hyperparameters, especially learning rates (and schedules thereof).We present Amortized Proximal Optimization (APO), which takes the perspective that each optimization step should approximately minimize a proximal objective (similar to the ones used to motivate natural gradient and trust region policy optimization).Optimization hyperparameters are adapted to best minimize the proximal objective after one weight update.We show that an idealized version of APO (where an oracle minimizes the proximal objective exactly) achieves global convergence to stationary point and locally second-order convergence to global optimum for neural networks.APO incurs minimal computational overhead.We experiment with using APO to adapt a variety of optimization hyperparameters online during training, including (possibly layer-specific) learning rates, damping coefficients, and gradient variance exponents.For a variety of network architectures and optimization algorithms (including SGD, RMSprop, and K-FAC), we show that with minimal tuning, APO performs competitively with carefully tuned optimizers.", "target": ["Wir stellen die Amortisierte Proximale Optimierung (APO) vor, eine Methode zur Online-Anpassung einer Vielzahl von Optimierungshyperparametern während des Trainings, einschließlich Lernraten, Dämpfungskoeffizienten und Varianzexponenten des Gradienten."]} +{"source": "Dense word vectors have proven their values in many downstream NLP tasks over the past few years.However, the dimensions of such embeddings are not easily interpretable.Out of the d-dimensions in a word vector, we would not be able to understand what high or low values mean.Previous approaches addressing this issue have mainly focused on either training sparse/non-negative constrained word embeddings, or post-processing standard pre-trained word embeddings.On the other hand, we analyze conventional word embeddings trained with Singular Value Decomposition, and reveal similar interpretability.We use a novel eigenvector analysis method inspired from Random Matrix Theory and show that semantically coherent groups not only form in the row space, but also the column space.This allows us to view individual word vector dimensions as human-interpretable semantic features.", "target": ["Wir zeigen, dass die hervorstechenden Dimensionen von Wortvektoren als semantische Merkmale interpretiert werden können, ohne dass irgendwelche Einschränkungen oder Nachbearbeitungen erforderlich sind."]} +{"source": "Neural networks in the brain and in neuromorphic chips confer systems with the ability to perform multiple cognitive tasks.However, both kinds of networks experience a wide range of physical perturbations, ranging from damage to edges of the network to complete node deletions, that ultimately could lead to network failure.A critical question is to understand how the computational properties of neural networks change in response to node-damage and whether there exist strategies to repair these networks in order to compensate for performance degradation.Here, we study the damage-response characteristics of two classes of neural networks, namely multilayer perceptrons (MLPs) and convolutional neural networks (CNNs) trained to classify images from MNIST and CIFAR-10 datasets respectively.We also propose a new framework to discover efficient repair strategies to rescue damaged neural networks.The framework involves defining damage and repair operators for dynamically traversing the neural networks loss landscape, with the goal of mapping its salient geometric features.Using this strategy, we discover features that resemble path-connected attractor sets in the loss landscape.We also identify that a dynamic recovery scheme, where networks are constantly damaged and repaired, produces a group of networks resilient to damage as it can be quickly rescued.Broadly, our work shows that we can design fault-tolerant networks by applying on-line retraining consistently during damage for real-time applications in biology and machine learning.", "target": ["Strategie zur Reparatur beschädigter neuronaler Netze"]} +{"source": "Automatic question generation from paragraphs is an important and challenging problem, particularly due to the long context from paragraphs.In this paper, we propose and study two hierarchical models for the task of question generation from paragraphs.Specifically, we propose(a) a novel hierarchical BiLSTM model with selective attention and(b) a novel hierarchical Transformer architecture, both of which learn hierarchical representations of paragraphs. We model a paragraph in terms of its constituent sentences, and a sentence in terms of its constituent words.While the introduction of the attention mechanism benefits the hierarchical BiLSTM model, the hierarchical Transformer, with its inherent attention and positional encoding mechanisms also performs better than flat transformer model.We conducted empirical evaluation on the widely used SQuAD and MS MARCO datasets using standard metrics. The results demonstrate the overall effectiveness of the hierarchical models over their flat counterparts. Qualitatively, our hierarchical models are able to generate fluent and relevant questions.", "target": ["Automatische Fragengenerierung aus Absätzen unter Verwendung hierarchischer Modelle"]} +{"source": "Many deployed learned models are black boxes: given input, returns output.Internal information about the model, such as the architecture, optimisation procedure, or training data, is not disclosed explicitly as it might contain proprietary information or make the system more vulnerable.This work shows that such attributes of neural networks can be exposed from a sequence of queries.This has multiple implications.On the one hand, our work exposes the vulnerability of black-box neural networks to different types of attacks -- we show that the revealed internal information helps generate more effective adversarial examples against the black box model.On the other hand, this technique can be used for better protection of private content from automatic recognition models using adversarial examples.Our paper suggests that it is actually hard to draw a line between white box and black box models.", "target": ["Die Abfrage eines neuronalen Black-Box-Netzes offenbart viele Informationen über dieses Netz; wir schlagen neuartige \"Metamodelle\" vor, um effektiv Informationen aus einer Black-Box zu extrahieren."]} +{"source": "Inverse reinforcement learning (IRL) is used to infer the reward function from the actions of an expert running a Markov Decision Process (MDP).A novel approach using variational inference for learning the reward function is proposed in this research.Using this technique, the intractable posterior distribution of the continuous latent variable (the reward function in this case) is analytically approximated to appear to be as close to the prior belief while trying to reconstruct the future state conditioned on the current state and action.The reward function is derived using a well-known deep generative model known as Conditional Variational Auto-encoder (CVAE) with Wasserstein loss function, thus referred to as Conditional Wasserstein Auto-encoder-IRL (CWAE-IRL), which can be analyzed as a combination of the backward and forward inference.This can then form an efficient alternative to the previous approaches to IRL while having no knowledge of the system dynamics of the agent.Experimental results on standard benchmarks such as objectworld and pendulum show that the proposed algorithm can effectively learn the latent reward function in complex, high-dimensional environments.", "target": ["Verwendung eines überwachten latenten Variablenmodellierungsrahmens zur Bestimmung der Belohnung bei einer Aufgabe des inversen Reinforcement Learnings"]} +{"source": "The travelling salesman problem (TSP) is a well-known combinatorial optimization problem with a variety of real-life applications.We tackle TSP by incorporating machine learning methodology and leveraging the variable neighborhood search strategy.More precisely, the search process is considered as a Markov decision process (MDP), where a 2-opt local search is used to search within a small neighborhood, while a Monte Carlo tree search (MCTS) method (which iterates through simulation, selection and back-propagation steps), is used to sample a number of targeted actions within an enlarged neighborhood.This new paradigm clearly distinguishes itself from the existing machine learning (ML) based paradigms for solving the TSP, which either uses an end-to-end ML model, or simply applies traditional techniques after ML for post optimization.Experiments based on two public data sets show that, our approach clearly dominates all the existing learning based TSP algorithms in terms of performance, demonstrating its high potential on the TSP.More importantly, as a general framework without complicated hand-crafted rules, it can be readily extended to many other combinatorial optimization problems.", "target": ["Diese Arbeit kombiniert die Monte-Carlo-Baumsuche mit der lokalen 2-Opt-Suche in einem variablen Nachbarschaftsmodus, um das TSP effektiv zu lösen."]} +{"source": "Significant strides have been made toward designing better generative models in recent years.Despite this progress, however, state-of-the-art approaches are still largely unable to capture complex global structure in data.For example, images of buildings typically contain spatial patterns such as windows repeating at regular intervals; state-of-the-art generative methods can’t easily reproduce these structures.We propose to address this problem by incorporating programs representing global structure into the generative model—e.g., a 2D for-loop may represent a configuration of windows.Furthermore, we propose a framework for learning these models by leveraging program synthesis to generate training data.On both synthetic and real-world data, we demonstrate that our approach is substantially better than the state-of-the-art at both generating and completing images that contain global structure.", "target": ["Anwendung der Programmsynthese auf die Aufgaben der Bildvervollständigung und -erzeugung innerhalb eines Deep Learning Frameworks"]} +{"source": "Learning control policies in robotic tasks requires a large number of interactions due to small learning rates, bounds on the updates or unknown constraints.In contrast humans can infer protective and safe solutions after a single failure or unexpected observation. In order to reach similar performance, we developed a hierarchical Bayesian optimization algorithm that replicates the cognitive inference and memorization process for avoiding failures in motor control tasks.A Gaussian Process implements the modeling and the sampling of the acquisition function.This enables rapid learning with large learning rates while a mental replay phase ensures that policy regions that led to failures are inhibited during the sampling process. The features of the hierarchical Bayesian optimization method are evaluated in a simulated and physiological humanoid postural balancing task.We quantitatively compare the human learning performance to our learning approach by evaluating the deviations of the center of mass during training.Our results show that we can reproduce the efficient learning of human subjects in postural control tasks which provides a testable model for future physiological motor control tasks.In these postural control tasks, our method outperforms standard Bayesian Optimization in the number of interactions to solve the task, in the computational demands and in the frequency of observed failures.", "target": ["In diesem Beitrag wird ein Berechnungsmodell für die effiziente Anpassung der menschlichen Haltungskontrolle vorgestellt, das auf hierarchischen Erfassungsfunktionen mit bekannten Merkmalen basiert."]} +{"source": "Deep reinforcement learning has achieved great success in many previously difficult reinforcement learning tasks, yet recent studies show that deep RL agents are also unavoidably susceptible to adversarial perturbations, similar to deep neural networks in classification tasks.Prior works mostly focus on model-free adversarial attacks and agents with discrete actions.In this work, we study the problem of continuous control agents in deep RL with adversarial attacks and propose the first two-step algorithm based on learned model dynamics.Extensive experiments on various MuJoCo domains (Cartpole, Fish, Walker, Humanoid) demonstrate that our proposed framework is much more effective and efficient than model-free based attacks baselines in degrading agent performance as well as driving agents to unsafe states.", "target": ["Wir untersuchen das Problem der kontinuierlichen Kontrolle von Agenten in tiefen RL mit gegnerischen Angriffen und schlagen einen zweistufigen Algorithmus vor, der auf gelernter Modelldynamik basiert."]} +{"source": "A leading hypothesis for the surprising generalization of neural networks is that the dynamics of gradient descent bias the model towards simple solutions, by searching through the solution space in an incremental order of complexity.We formally define the notion of incremental learning dynamics and derive the conditions on depth and initialization for which this phenomenon arises in deep linear models.Our main theoretical contribution is a dynamical depth separation result, proving that while shallow models can exhibit incremental learning dynamics, they require the initialization to be exponentially small for these dynamics to present themselves.However, once the model becomes deeper, the dependence becomes polynomial and incremental learning can arise in more natural settings.We complement our theoretical findings by experimenting with deep matrix sensing, quadratic neural networks and with binary classification using diagonal and convolutional linear networks, showing all of these models exhibit incremental learning.", "target": ["Wir untersuchen die Spärlichkeit verursachende Verzerrung von tiefen Modellen, die durch ihre Lerndynamik verursacht wird."]} +{"source": "Generative modeling of high dimensional data like images is a notoriously difficult and ill-defined problem.In particular, how to evaluate a learned generative model is unclear.In this paper, we argue that *adversarial learning*, pioneered with generative adversarial networks (GANs), provides an interesting framework to implicitly define more meaningful task losses for unsupervised tasks, such as for generating \"visually realistic\" images.By relating GANs and structured prediction under the framework of statistical decision theory, we put into light links between recent advances in structured prediction theory and the choice of the divergence in GANs.We argue that the insights about the notions of \"hard\" and \"easy\" to learn losses can be analogously extended to adversarial divergences.We also discuss the attractive properties of parametric adversarial divergences for generative modeling, and perform experiments to show the importance of choosing a divergence that reflects the final task.", "target": ["Parametrische adversarische Divergenzen definieren implizit aussagekräftigere Aufgabenverluste für die generative Modellierung. Wir ziehen Parallelen zur strukturierten Vorhersage, um die Eigenschaften dieser Divergenzen und ihre Fähigkeit zur Codierung der interessierenden Aufgabe zu untersuchen."]} +{"source": "Experimental reproducibility and replicability are critical topics in machine learning.Authors have often raised concerns about their lack in scientific publications to improve the quality of the field.Recently, the graph representation learning field has attracted the attention of a wide research community, which resulted in a large stream of works.As such, several Graph Neural Network models have been developed to effectively tackle graph classification.However, experimental procedures often lack rigorousness and are hardly reproducible.Motivated by this, we provide an overview of common practices that should be avoided to fairly compare with the state of the art.To counter this troubling trend, we ran more than 47000 experiments in a controlled and uniform framework to re-evaluate five popular models across nine common benchmarks.Moreover, by comparing GNNs with structure-agnostic baselines we provide convincing evidence that, on some datasets, structural information has not been exploited yet.We believe that this work can contribute to the development of the graph learning field, by providing a much needed grounding for rigorous evaluations of graph classification models.", "target": ["Wir bieten einen rigorosen Vergleich verschiedener Graph Neural Networks für die Graphklassifizierung."]} +{"source": "Data augmentation (DA) is fundamental against overfitting in large convolutional neural networks, especially with a limited training dataset.In images, DA is usually based on heuristic transformations, like geometric or color transformations.Instead of using predefined transformations, our work learns data augmentation directly from the training data by learning to transform images with an encoder-decoder architecture combined with a spatial transformer network.The transformed images still belong to the same class, but are new, more complex samples for the classifier.Our experiments show that our approach is better than previous generative data augmentation methods, and comparable to predefined transformation methods when training an image classifier.", "target": ["Automatisches Lernen von Datenerweiterungen unter Verwendung einer GAN-basierten Architektur zur Verbesserung eines Bildklassifizierers"]} +{"source": "We consider the problem of information compression from high dimensional data.Where many studies consider the problem of compression by non-invertible trans- formations, we emphasize the importance of invertible compression.We introduce new class of likelihood-based auto encoders with pseudo bijective architecture, which we call Pseudo Invertible Encoders.We provide the theoretical explanation of their principles.We evaluate Gaussian Pseudo Invertible Encoder on MNIST, where our model outperform WAE and VAE in sharpness of the generated images.", "target": ["Neue Klasse von Autoencodern mit pseudo-invertierbarer Architektur"]} +{"source": "We exploit a recently derived inversion scheme for arbitrary deep neural networks to develop a new semi-supervised learning framework that applies to a wide range of systems and problems. The approach reaches current state-of-the-art methods on MNIST and provides reasonable performances on SVHN and CIFAR10.Through the introduced method, residual networks are for the first time applied to semi-supervised tasks.Experiments with one-dimensional signals highlight the generality of the method.Importantly, our approach is simple, efficient, and requires no change in the deep network architecture.", "target": ["Wir nutzen ein Inversionsschema für beliebige tiefe neuronale Netze, um einen neuen halbüberwachten Lernrahmen zu entwickeln, der auf viele Topologien anwendbar ist."]} +{"source": "Deep learning has become a widely used tool in many computational and classification problems. Nevertheless obtaining and labeling data, which is needed for strong results, is often expensive or even not possible. In this paper three different algorithmic approaches to deal with limited access to data are evaluated and compared to each other. We show the drawbacks and benefits of each method. One successful approach, especially in one- or few-shot learning tasks, is the use of external data during the classification task. Another successful approach, which achieves state of the art results in semi-supervised learning (SSL) benchmarks, is consistency regularization.Especially virtual adversarial training (VAT) has shown strong results and will be investigated in this paper. The aim of consistency regularization is to force the network not to change the output, when the input or the network itself is perturbed.Generative adversarial networks (GANs) have also shown strong empirical results. In many approaches the GAN architecture is used in order to create additional data and therefor to increase the generalization capability of the classification network.Furthermore we consider the use of unlabeled data for further performance improvement. The use of unlabeled data is investigated both for GANs and VAT.", "target": ["Vergleich von Siamesischen Neuronalen Netzen, GANs und VAT für das Few-Shot Learning."]} +{"source": "This paper introduces a new neural structure called FusionNet, which extends existing attention approaches from three perspectives.First, it puts forward a novel concept of \"History of Word\" to characterize attention information from the lowest word-level embedding up to the highest semantic-level representation.Second, it identifies an attention scoring function that better utilizes the \"history of word\" concept.Third, it proposes a fully-aware multi-level attention mechanism to capture the complete information in one text (such as a question) and exploit it in its counterpart (such as context or passage) layer by layer.We apply FusionNet to the Stanford Question Answering Dataset (SQuAD) and it achieves the first position for both single and ensemble model on the official SQuAD leaderboard at the time of writing (Oct. 4th, 2017).Meanwhile, we verify the generalization of FusionNet with two adversarial SQuAD datasets and it sets up the new state-of-the-art on both datasets: on AddSent, FusionNet increases the best F1 metric from 46.6% to 51.4%; on AddOneSent, FusionNet boosts the best F1 metric from 56.0% to 60.7%.", "target": ["Wir schlagen eine leichtgewichtige Verbesserung der Aufmerksamkeit und eine neuronale Architektur, FusionNet, vor, um SotA auf SQuAD und adversarial SQuAD zu erreichen."]} +{"source": "We propose a novel hierarchical generative model with a simple Markovian structure and a corresponding inference model.Both the generative and inference model are trained using the adversarial learning paradigm.We demonstrate that the hierarchical structure supports the learning of progressively more abstract representations as well as providing semantically meaningful reconstructions with different levels of fidelity.Furthermore, we show that minimizing the Jensen-Shanon divergence between the generative and inference network is enough to minimize the reconstruction error. The resulting semantically meaningful hierarchical latent structure discovery is exemplified on the CelebA dataset. There, we show that the features learned by our model in an unsupervised way outperform the best handcrafted features. Furthermore, the extracted features remain competitive when compared to several recent deep supervised approaches on an attribute prediction task on CelebA.Finally, we leverage the model's inference network to achieve state-of-the-art performance on a semi-supervised variant of the MNIST digit classification task.", "target": ["Adversarial trainiertes hierarchisches generatives Modell mit robuster und semantisch gelernter latenter Repräsentation."]} +{"source": "Conservation laws are considered to be fundamental laws of nature.It has broad application in many fields including physics, chemistry, biology, geology, and engineering.Solving the differential equations associated with conservation laws is a major branch in computational mathematics.Recent success of machine learning, especially deep learning, in areas such as computer vision and natural language processing, has attracted a lot of attention from the community of computational mathematics and inspired many intriguing works in combining machine learning with traditional methods.In this paper, we are the first to explore the possibility and benefit of solving nonlinear conservation laws using deep reinforcement learning.As a proof of concept, we focus on 1-dimensional scalar conservation laws.We deploy the machinery of deep reinforcement learning to train a policy network that can decide on how the numerical solutions should be approximated in a sequential and spatial-temporal adaptive manner.We will show that the problem of solving conservation laws can be naturally viewed as a sequential decision making process and the numerical schemes learned in such a way can easily enforce long-term accuracy. Furthermore, the learned policy network is carefully designed to determine a good local discrete approximation based on the current state of the solution, which essentially makes the proposed method a meta-learning approach.In other words, the proposed method is capable of learning how to discretize for a given situation mimicking human experts.Finally, we will provide details on how the policy network is trained, how well it performs compared with some state-of-the-art numerical solvers such as WENO schemes, and how well it generalizes.Our code is released anomynously at \\url{https://github.com/qwerlanksdf/L2D}.", "target": ["Wir stellen fest, dass numerische PDE-Löser als Markov-Desicion-Prozesse betrachtet werden können, und schlagen vor, Reinforcement Learning zur Lösung skalarer 1D-Erhaltungsgesetze einzusetzen"]} +{"source": "We present a neural rendering architecture that helps variational autoencoders (VAEs) learn disentangled representations.Instead of the deconvolutional network typically used in the decoder of VAEs, we tile (broadcast) the latent vector across space, concatenate fixed X- and Y-“coordinate” channels, and apply a fully convolutional network with 1x1 stride.This provides an architectural prior for dissociating positional from non-positional features in the latent space, yet without providing any explicit supervision to this effect.We show that this architecture, which we term the Spatial Broadcast decoder, improves disentangling, reconstruction accuracy, and generalization to held-out regions in data space. We show the Spatial Broadcast Decoder is complementary to state-of-the-art (SOTA) disentangling techniques and when incorporated improves their performance.", "target": ["Wir stellen eine neuronale Rendering-Architektur vor, die VAEs dabei hilft, entkoppelte latente Repräsentationen zu lernen."]} +{"source": "Similar to humans and animals, deep artificial neural networks exhibit critical periods during which a temporary stimulus deficit can impair the development of a skill.The extent of the impairment depends on the onset and length of the deficit window, as in animal models, and on the size of the neural network.Deficits that do not affect low-level statistics, such as vertical flipping of the images, have no lasting effect on performance and can be overcome with further training. To better understand this phenomenon, we use the Fisher Information of the weights to measure the effective connectivity between layers of a network during training. Counterintuitively, information rises rapidly in the early phases of training, and then decreases, preventing redistribution of information resources in a phenomenon we refer to as a loss of \"Information Plasticity\". Our analysis suggests that the first few epochs are critical for the creation of strong connections that are optimal relative to the input data distribution.Once such strong connections are created, they do not appear to change during additional training.These findings suggest that the initial learning transient, under-scrutinized compared to asymptotic behavior, plays a key role in determining the outcome of the training process.Our findings, combined with recent theoretical results in the literature, also suggest that forgetting (decrease of information in the weights) is critical to achieving invariance and disentanglement in representation learning.Finally, critical periods are not restricted to biological systems, but can emerge naturally in learning systems, whether biological or artificial, due to fundamental constrains arising from learning dynamics and information processing.", "target": ["Sensorische Defizite in frühen Trainingsphasen können sowohl in künstlichen als auch in neuronalen Netzen zu irreversiblen Leistungseinbußen führen, was auf Informationsphänomene als gemeinsame Ursache hindeutet und auf die Bedeutung des anfänglichen Einschwingens und Vergessens hinweist."]} +{"source": "We propose a method to incrementally learn an embedding space over the domain of network architectures, to enable the careful selection of architectures for evaluation during compressed architecture search.Given a teacher network, we search for a compressed network architecture by using Bayesian Optimization (BO) with a kernel function defined over our proposed embedding space to select architectures for evaluation.We demonstrate that our search algorithm can significantly outperform various baseline methods, such as random search and reinforcement learning (Ashok et al., 2018).The compressed architectures found by our method are also better than the state-of-the-art manually-designed compact architecture ShuffleNet (Zhang et al., 2018).We also demonstrate that the learned embedding space can be transferred to new settings for architecture search, such as a larger teacher network or a teacher network in a different architecture family, without any training.", "target": ["Wir schlagen eine Methode zum inkrementellen Lernen eines Einbettungsraums über die Domäne der Netzwerkarchitekturen vor, um eine sorgfältige Auswahl von Architekturen für die Evaluierung während der komprimierten Architektursuche zu ermöglichen."]} +{"source": "Reinforcement Learning (RL) problem can be solved in two different ways - the Value function-based approach and the policy optimization-based approach - to eventually arrive at an optimal policy for the given environment.One of the recent breakthroughs in reinforcement learning is the use of deep neural networks as function approximators to approximate the value function or q-function in a reinforcement learning scheme.This has led to results with agents automatically learning how to play games like alpha-go showing better-than-human performance.Deep Q-learning networks (DQN) and Deep Deterministic Policy Gradient (DDPG) are two such methods that have shown state-of-the-art results in recent times.Among the many variants of RL, an important class of problems is where the state and action spaces are continuous --- autonomous robots, autonomous vehicles, optimal control are all examples of such problems that can lend themselves naturally to reinforcement based algorithms, and have continuous state and action spaces.In this paper, we adapt and combine approaches such as DQN and DDPG in novel ways to outperform the earlier results for continuous state and action space problems. We believe these results are a valuable addition to the fast-growing body of results on Reinforcement Learning, more so for continuous state and action space problems.", "target": ["Verbesserung der Leistung eines RL-Agenten in einem kontinuierlichen Handlungs- und Zustandsraum durch priorisierte Erfahrungswiedergabe und Parameterrauschen."]} +{"source": "Natural Language Inference (NLI) task requires an agent to determine the logical relationship between a natural language premise and a natural language hypothesis.We introduce Interactive Inference Network (IIN), a novel class of neural network architectures that is able to achieve high-level understanding of the sentence pair by hierarchically extracting semantic features from interaction space.We show that an interaction tensor (attention weight) contains semantic information to solve natural language inference, and a denser interaction tensor contains richer semantic information.One instance of such architecture, Densely Interactive Inference Network (DIIN), demonstrates the state-of-the-art performance on large scale NLI copora and large-scale NLI alike corpus.It's noteworthy that DIIN achieve a greater than 20% error reduction on the challenging Multi-Genre NLI (MultiNLI) dataset with respect to the strongest published system.", "target": ["zeigen, dass mehrkanalige Aufmerksamkeitsgewichte semantische Merkmale enthalten, um die Aufgabe der Inferenz natürlicher Sprache zu lösen."]} +{"source": "Determinantal Point Processes (DPPs) provide an elegant and versatile way to sample sets of items that balance the point-wise quality with the set-wise diversity of selected items.For this reason, they have gained prominence in many machine learning applications that rely on subset selection.However, sampling from a DPP over a ground set of size N is a costly operation, requiring in general an O(N^3) preprocessing cost and an O(Nk^3) sampling cost for subsets of size k. We approach this problem by introducing DppNets: generative deep models that produce DPP-like samples for arbitrary ground sets. We develop an inhibitive attention mechanism based on transformer networks that captures a notion of dissimilarity between feature vectors. We show theoretically that such an approximation is sensible as it maintains the guarantees of inhibition or dissimilarity that makes DPP so powerful and unique. Empirically, we demonstrate that samples from our model receive high likelihood under the more expensive DPP alternative.", "target": ["Wir approximieren Determinantal Point Processes mit neuronalen Netzen und begründen unser Modell theoretisch und empirisch."]} +{"source": "This paper introduces a network architecture to solve the structure-from-motion (SfM) problem via feature-metric bundle adjustment (BA), which explicitly enforces multi-view geometry constraints in the form of feature-metric error.The whole pipeline is differentiable, so that the network can learn suitable features that make the BA problem more tractable.Furthermore, this work introduces a novel depth parameterization to recover dense per-pixel depth.The network first generates several basis depth maps according to the input image, and optimizes the final depth as a linear combination of these basis depth maps via feature-metric BA.The basis depth maps generator is also learned via end-to-end training.The whole system nicely combines domain knowledge (i.e. hard-coded multi-view geometry constraints) and deep learning (i.e. feature learning and basis depth maps learning) to address the challenging dense SfM problem.Experiments on large scale real data prove the success of the proposed method.", "target": ["In diesem Beitrag wird eine Netzwerkarchitektur zur Lösung des Structure-from-Motion (SfM)-Problems mittels Feature-Bundle-Anpassung (BA) vorgestellt."]} +{"source": "Temporal Difference Learning with function approximation is known to be unstable.Previous work like \\citet{sutton2009fast} and \\citet{sutton2009convergent} has presented alternative objectives that are stable to minimize.However, in practice, TD-learning with neural networks requires various tricks like using a target network that updates slowly \\citep{mnih2015human}.In this work we propose a constraint on the TD update that minimizes change to the target values.This constraint can be applied to the gradients of any TD objective, and can be easily applied to nonlinear function approximation.We validate this update by applying our technique to deep Q-learning, and training without a target network.We also show that adding this constraint on Baird's counterexample keeps Q-learning from diverging.", "target": ["Wir zeigen, dass das Hinzufügen einer Beschränkung zu TD-Updates das Lernen stabilisiert und Deep Q-learning ohne ein Zielnetzwerk ermöglicht"]} +{"source": "We propose DuoRC, a novel dataset for Reading Comprehension (RC) that motivates several new challenges for neural approaches in language understanding beyond those offered by existing RC datasets.DuoRC contains 186,089 unique question-answer pairs created from a collection of 7680 pairs of movie plots where each pair in the collection reflects two versions of the same movie - one from Wikipedia and the other from IMDb - written by two different authors.We asked crowdsourced workers to create questions from one version of the plot and a different set of workers to extract or synthesize corresponding answers from the other version.This unique characteristic of DuoRC where questions and answers are created from different versions of a document narrating the same underlying story, ensures by design, that there is very little lexical overlap between the questions created from one version and the segments containing the answer in the other version.Further, since the two versions have different level of plot detail, narration style, vocabulary, etc., answering questions from the second version requires deeper language understanding and incorporating background knowledge not available in the given text.Additionally, the narrative style of passages arising from movie plots (as opposed to typical descriptive passages in existing datasets) exhibits the need to perform complex reasoning over events across multiple sentences.Indeed, we observe that state-of-the-art neural RC models which have achieved near human performance on the SQuAD dataset, even when coupled with traditional NLP techniques to address the challenges presented in DuoRC exhibit very poor performance (F1 score of 37.42% on DuoRC v/s 86% on SQuAD dataset).This opens up several interesting research avenues wherein DuoRC could complement other Reading Comprehension style datasets to explore novel neural approaches for studying language understanding.", "target": ["Wir schlagen DuoRC vor, einen neuartigen Datensatz für das Leseverständnis (RC), der 186.089 von Menschen erzeugte QA-Paare enthält, die aus einer Sammlung von 7680 Paaren paralleler Filmhandlungen erstellt wurden, und führen eine RC-Aufgabe ein, bei der eine Version der Handlung gelesen und Fragen beantwortet werden müssen, die aus der anderen Version erstellt wurden; dies erfordert komplexes logisches Denken und ein tieferes Sprachverständnis, um die geringe lexikalische Überschneidung zwischen der Handlung und der Frage zu überwinden."]} +{"source": "We consider the problem of weakly supervised structured prediction (SP) with reinforcement learning (RL) – for example, given a database table and a question, perform a sequence of computation actions on the table, which generates a response and receives a binary success-failure reward. This line of research has been successful by leveraging RL to directly optimizes the desired metrics of the SP tasks – for example, the accuracy in question answering or BLEU score in machine translation. However, different from the common RL settings, the environment dynamics is deterministic in SP, which hasn’t been fully utilized by the model-freeRL methods that are usually applied.Since SP models usually have full access to the environment dynamics, we propose to apply model-based RL methods, which rely on planning as a primary model component.We demonstrate the effectiveness of planning-based SP with a Neural Program Planner (NPP), which, given a set of candidate programs from a pretrained search policy, decides which program is the most promising considering all the information generated from executing these programs.We evaluate NPP on weakly supervised program synthesis from natural language(semantic parsing) by stacked learning a planning module based on pretrained search policies.On the WIKITABLEQUESTIONS benchmark, NPP achieves a new state-of-the-art of 47.2% accuracy.", "target": ["Eine modellbasierte Planungskomponente verbessert RL-basiertes semantisches Parsing auf WikiTableQuestions."]} +{"source": "Deep learning algorithms achieve high classification accuracy at the expense of significant computation cost.To address this cost, a number of quantization schemeshave been proposed - but most of these techniques focused on quantizing weights, which are relatively smaller in size compared to activations.This paper proposes a novel quantization scheme for activations during training - that enables neural networks to work well with ultra low precision weights and activations without any significant accuracy degradation. This technique, PArameterized Clipping acTi-vation (PACT), uses an activation clipping parameter α that is optimized duringtraining to find the right quantization scale.PACT allows quantizing activations toarbitrary bit precisions, while achieving much better accuracy relative to publishedstate-of-the-art quantization schemes.We show, for the first time, that both weights and activations can be quantized to 4-bits of precision while still achieving accuracy comparable to full precision networks across a range of popular models and datasets.We also show that exploiting these reduced-precision computational units in hardware can enable a super-linear improvement in inferencing performance dueto a significant reduction in the area of accelerator compute engines coupled with the ability to retain the quantized model and activation data in on-chip memories.", "target": ["Eine neue Art der Quantisierung der Aktivierung eines tiefen neuronalen Netzes durch parametrisiertes Clipping, das die Quantisierungsskala durch stochastischen Gradientenabstieg optimiert."]} +{"source": "Pruning neural network parameters is often viewed as a means to compress models, but pruning has also been motivated by the desire to prevent overfitting.This motivation is particularly relevant given the perhaps surprising observation that a wide variety of pruning approaches increase test accuracy despite sometimes massive reductions in parameter counts.To better understand this phenomenon, we analyze the behavior of pruning over the course of training, finding that pruning's effect on generalization relies more on the instability it generates (defined as the drops in test accuracy immediately following pruning) than on the final size of the pruned model.We demonstrate that even the pruning of unimportant parameters can lead to such instability, and show similarities between pruning and regularizing by injecting noise, suggesting a mechanism for pruning-based generalization improvements that is compatible with the strong generalization recently observed in over-parameterized networks.", "target": ["Wir zeigen, dass Pruning-Methoden, die eine größere Instabilität in den Verlust einbringen, auch zu einer verbesserten Generalisierung führen, und untersuchen die Mechanismen, die diesem Effekt zugrunde liegen."]} +{"source": "Neural networks trained with backpropagation, the standard algorithm of deep learning which uses weight transport, are easily fooled by existing gradient-based adversarial attacks.This class of attacks are based on certain small perturbations of the inputs to make networks misclassify them.We show that less biologically implausible deep neural networks trained with feedback alignment, which do not use weight transport, can be harder to fool, providing actual robustness.Tested on MNIST, deep neural networks trained without weight transport (1) have an adversarial accuracy of 98% compared to 0.03% for neural networks trained with backpropagation and (2) generate non-transferable adversarial examples.However, this gap decreases on CIFAR-10 but is still significant particularly for small perturbation magnitude less than 1 ⁄ 2.", "target": ["Weniger biologisch unplausible tiefe neuronale Netze, die ohne Gewichtstransport trainiert werden, können schwerer zu täuschen sein."]} +{"source": "Chemical reactions can be described as the stepwise redistribution of electrons in molecules.As such, reactions are often depicted using \"arrow-pushing\" diagrams which show this movement as a sequence of arrows.We propose an electron path prediction model (ELECTRO) to learn these sequences directly from raw reaction data.Instead of predicting product molecules directly from reactant molecules in one shot, learning a model of electron movement has the benefits of(a) being easy for chemists to interpret,(b) incorporating constraints of chemistry, such as balanced atom counts before and after the reaction, and(c) naturally encoding the sparsity of chemical reactions, which usually involve changes in only a small number of atoms in the reactants.We design a method to extract approximate reaction paths from any dataset of atom-mapped reaction SMILES strings.Our model achieves excellent performance on an important subset of the USPTO reaction dataset, comparing favorably to the strongest baselines.Furthermore, we show that our model recovers a basic knowledge of chemistry without being explicitly trained to do so.", "target": ["Ein generatives Modell zur Reaktionsvorhersage, das die mechanistischen Elektronenschritte einer Reaktion direkt aus den Rohdaten der Reaktion lernt."]} +{"source": "When machine learning models are used for high-stakes decisions, they should predict accurately, fairly, and responsibly.To fulfill these three requirements, a model must be able to output a reject option (i.e. say \"``I Don't Know\") when it is not qualified to make a prediction.In this work, we propose learning to defer, a method by which a model can defer judgment to a downstream decision-maker such as a human user.We show that learning to defer generalizes the rejection learning framework in two ways: by considering the effect of other agents in the decision-making process, and by allowing for optimization of complex objectives.We propose a learning algorithm which accounts for potential biases held by decision-makerslater in a pipeline.Experiments on real-world datasets demonstrate that learningto defer can make a model not only more accurate but also less biased.Even whenoperated by highly biased users, we show thatdeferring models can still greatly improve the fairness of the entire pipeline.", "target": ["Die Möglichkeit, \"Ich weiß es nicht\" zu sagen, kann die Fairness eines Klassifizierers verbessern, ohne dass die Genauigkeit zu sehr darunter leidet, und diese Verbesserung wird noch effektiver, wenn der Klassifizierer Einblick in die nachgelagerte Entscheidungsfindung hat."]} +{"source": "Hierarchical Task Networks (HTN) generate plans using a decomposition process guided by extra domain knowledge to guide search towards a planning task.While many HTN planners can make calls to external processes (e.g. to a simulator interface) during the decomposition process, this is a computationally expensive process, so planner implementations often use such calls in an ad-hoc way using very specialized domain knowledge to limit the number of calls.Conversely, the few classical planners that are capable of using external calls (often called semantic attachments) during planning do so in much more limited ways by generating a fixed number of ground operators at problem grounding time.In this paper we develop the notion of semantic attachments for HTN planning using semi co-routines, allowing such procedurally defined predicates to link the planning process to custom unifications outside of the planner.The resulting planner can then use such co-routines as part of its backtracking mechanism to search through parallel dimensions of the state-space (e.g. through numeric variables).We show empirically that our planner outperforms the state-of-the-art numeric planners in a number of domains using minimal extra domain knowledge.", "target": ["Ein Ansatz zur Durchführung der HTN-Planung unter Verwendung externer Prozeduren zur Auswertung von Prädikaten zur Laufzeit (semantische Anhänge)."]} +{"source": "Recent literature suggests that averaged word vectors followed by simple post-processing outperform many deep learning methods on semantic textual similarity tasks.Furthermore, when averaged word vectors are trained supervised on large corpora of paraphrases, they achieve state-of-the-art results on standard STS benchmarks.Inspired by these insights, we push the limits of word embeddings even further.We propose a novel fuzzy bag-of-words (FBoW) representation for text that contains all the words in the vocabulary simultaneously but with different degrees of membership, which are derived from similarities between word vectors.We show that max-pooled word vectors are only a special case of fuzzy BoW and should be compared via fuzzy Jaccard index rather than cosine similarity.Finally, we propose DynaMax, a completely unsupervised and non-parametric similarity measure that dynamically extracts and max-pools good features depending on the sentence pair.This method is both efficient and easy to implement, yet outperforms current baselines on STS tasks by a large margin and is even competitive with supervised word vectors trained to directly optimise cosine similarity.", "target": ["Max-gepoolte Wortvektoren mit Fuzzy-Jacard-Set-Ähnlichkeit sind eine äußerst wettbewerbsfähige Grundlage für semantische Ähnlichkeit; wir schlagen eine einfache dynamische Variante vor, die noch besser abschneidet."]} +{"source": "State-of-the-art results in imitation learning are currently held by adversarial methods that iteratively estimate the divergence between student and expert policies and then minimize this divergence to bring the imitation policy closer to expert behavior.Analogous techniques for imitation learning from observations alone (without expert action labels), however, have not enjoyed the same ubiquitous successes. Recent work in adversarial methods for generative models has shown that the measure used to judge the discrepancy between real and synthetic samples is an algorithmic design choice, and that different choices can result in significant differences in model performance.Choices including Wasserstein distance and various $f$-divergences have already been explored in the adversarial networks literature, while more recently the latter class has been investigated for imitation learning.Unfortunately, we find that in practice this existing imitation-learning framework for using $f$-divergences suffers from numerical instabilities stemming from the combination of function approximation and policy-gradient reinforcement learning.In this work, we alleviate these challenges and offer a reparameterization of adversarial imitation learning as $f$-divergence minimization before further extending the framework to handle the problem of imitation from observations only.Empirically, we demonstrate that our design choices for coupling imitation learning and $f$-divergences are critical to recovering successful imitation policies.Moreover, we find that with the appropriate choice of $f$-divergence, we can obtain imitation-from-observation algorithms that outperform baseline approaches and more closely match expert performance in continous-control tasks with low-dimensional observation spaces.With high-dimensional observations, we still observe a significant gap with and without action labels, offering an interesting avenue for future work.", "target": ["Das übergeordnete Ziel dieser Arbeit ist es, durch die Verwendung von f-Divergenzen eine stichprobeneffiziente Nachahmung von Expertendemonstrationen zu ermöglichen, sowohl mit als auch ohne die Bereitstellung von Experten-Handlungsbezeichnungen."]} +{"source": "Momentary fluctuations in attention (perceptual accuracy) correlate with neural activity fluctuations in primate visual areas.Yet, the link between such momentary neural fluctuations and attention state remains to be shown in the human brain.We investigate this link using a real-time cognitive brain machine interface (cBMI) based on steady state visually evoked potentials (SSVEPs): occipital EEG potentials evoked by rhythmically flashing stimuli.Tracking momentary fluctuations in SSVEP power, in real-time, we presented stimuli time-locked to when this power reached (predetermined) high or low thresholds.We observed a significant increase in discrimination accuracy (d') when stimuli were triggered during high (versus low) SSVEP power epochs, at the location cued for attention.Our results indicate a direct link between attention’s effects on perceptual accuracy and and neural gain in EEG-SSVEP power, in the human brain.", "target": ["Mit einem kognitiven Brain-Machine-Interface zeigen wir einen direkten Zusammenhang zwischen Aufmerksamkeitseffekten auf die Wahrnehmungsgenauigkeit und der neuronalen Verstärkung der EEG-SSVEP-Leistung im menschlichen Gehirn."]} +{"source": "Recent studies have demonstrated the vulnerability of deep convolutional neural networks against adversarial examples.Inspired by the observation that the intrinsic dimension of image data is much smaller than its pixel space dimension and the vulnerability of neural networks grows with the input dimension, we propose to embed high-dimensional input images into a low-dimensional space to perform classification.However, arbitrarily projecting the input images to a low-dimensional space without regularization will not improve the robustness of deep neural networks.We propose a new framework, Embedding Regularized Classifier (ER-Classifier), which improves the adversarial robustness of the classifier through embedding regularization.Experimental results on several benchmark datasets show that, our proposed framework achieves state-of-the-art performance against strong adversarial attack methods.", "target": ["Ein allgemeiner und einfach zu verwendender Rahmen, der die Widerstandsfähigkeit von tiefen Klassifikationsmodellen durch Einbettungsregularisierung verbessert."]} +{"source": "We investigate task clustering for deep learning-based multi-task and few-shot learning in the settings with large numbers of diverse tasks.Our method measures task similarities using cross-task transfer performance matrix.Although this matrix provides us critical information regarding similarities between tasks, the uncertain task-pairs, i.e., the ones with extremely asymmetric transfer scores, may collectively mislead clustering algorithms to output an inaccurate task-partition.Moreover, when the number of tasks is large, generating the full transfer performance matrix can be very time consuming.To overcome these limitations, we propose a novel task clustering algorithm to estimate the similarity matrix based on the theory of matrix completion.The proposed algorithm can work on partially-observed similarity matrices based on only sampled task-pairs with reliable scores, ensuring its efficiency and robustness.Our theoretical analysis shows that under mild assumptions, the reconstructed matrix perfectly matches the underlying “true” similarity matrix with an overwhelming probability.The final task partition is computed by applying an efficient spectral clustering algorithm to the recovered matrix.Our results show that the new task clustering method can discover task clusters that benefit both multi-task learning and few-shot learning setups for sentiment classification and dialog intent classification tasks.", "target": ["Wir schlagen einen auf Matrix-Vervollständigung basierenden Aufgaben-Clustering-Algorithmus für tiefes Multitasking und \"few-shot learning\" in Umgebungen mit einer großen Anzahl unterschiedlicher Aufgaben vor."]} +{"source": "The resemblance between the methods used in studying quantum-many body physics and in machine learning has drawn considerable attention.In particular, tensor networks (TNs) and deep learning architectures bear striking similarities to the extent that TNs can be used for machine learning.Previous results used one-dimensional TNs in image recognition, showing limited scalability and a request of high bond dimension.In this work, we train two-dimensional hierarchical TNs to solve image recognition problems, using a training algorithm derived from the multipartite entanglement renormalization ansatz (MERA).This approach overcomes scalability issues and implies novel mathematical connections among quantum many-body physics, quantum information theory, and machine learning.While keeping the TN unitary in the training phase, TN states can be defined, which optimally encodes each class of the images into a quantum many-body state.We study the quantum features of the TN states, including quantum entanglement and fidelity.We suggest these quantities could be novel properties that characterize the image classes, as well as the machine learning tasks.Our work could be further applied to identifying possible quantum properties of certain artificial intelligence methods.", "target": ["Dieser Ansatz überwindet Skalierbarkeitsprobleme und impliziert neuartige mathematische Verbindungen zwischen Quantenvielteilchenphysik, Quanteninformationstheorie und maschinellem Lernen."]} +{"source": "Predicting outcomes and planning interactions with the physical world are long-standing goals for machine learning.A variety of such tasks involves continuous physical systems, which can be described by partial differential equations (PDEs) with many degrees of freedom.Existing methods that aim to control the dynamics of such systems are typically limited to relatively short time frames or a small number of interaction parameters.We present a novel hierarchical predictor-corrector scheme which enables neural networks to learn to understand and control complex nonlinear physical systems over long time frames.We propose to split the problem into two distinct tasks: planning and control.To this end, we introduce a predictor network that plans optimal trajectories and a control network that infers the corresponding control parameters.Both stages are trained end-to-end using a differentiable PDE solver.We demonstrate that our method successfully develops an understanding of complex physical systems and learns to control them for tasks involving PDEs such as the incompressible Navier-Stokes equations.", "target": ["Wir trainieren eine Kombination von neuronalen Netzen, um optimale Flugbahnen für komplexe physikalische Systeme vorherzusagen."]} +{"source": "The ability of overparameterized deep networks to generalize well has been linked to the fact that stochastic gradient descent (SGD) finds solutions that lie in flat, wide minima in the training loss -- minima where the output of the network is resilient to small random noise added to its parameters. So far this observation has been used to provide generalization guarantees only for neural networks whose parameters are either \\textit{stochastic} or \\textit{compressed}.In this work, we present a general PAC-Bayesian framework that leverages this observation to provide a bound on the original network learned -- a network that is deterministic and uncompressed. What enables us to do this is a key novelty in our approach: our framework allows us to show that if on training data, the interactions between the weight matrices satisfy certain conditions that imply a wide training loss minimum, these conditions themselves {\\em generalize} to the interactions between the matrices on test data, thereby implying a wide test loss minimum.We then apply our general framework in a setup where we assume that the pre-activation values of the network are not too small (although we assume this only on the training data).In this setup, we provide a generalization guarantee for the original (deterministic, uncompressed) network, that does not scale with product of the spectral norms of the weight matrices -- a guarantee that would not have been possible with prior approaches.", "target": ["Wir bieten eine PAC-Bayes-basierte Generalisierungsgarantie für unkomprimierte, deterministische tiefe Netzwerke, indem wir die Rauschresistenz des Netzwerks auf den Trainingsdaten auf die Testdaten verallgemeinern."]} +{"source": "Training neural networks to be certifiably robust is critical to ensure their safety against adversarial attacks.However, it is currently very difficult to train a neural network that is both accurate and certifiably robust.In this work we take a step towards addressing this challenge.We prove that for every continuous function $f$, there exists a network $n$ such that:(i) $n$ approximates $f$ arbitrarily close, and(ii) simple interval bound propagation of a region $B$ through $n$ yields a result that is arbitrarily close to the optimal output of $f$ on $B$.Our result can be seen as a Universal Approximation Theorem for interval-certified ReLU networks.To the best of our knowledge, this is the first work to prove the existence of accurate, interval-certified networks.", "target": ["Wir beweisen, dass es für eine große Klasse von Funktionen f ein intervallzertifiziertes robustes Netzwerk gibt, das f bis zu einer beliebigen Genauigkeit approximiert."]} +{"source": "In this paper, we propose an efficient framework to accelerate convolutional neural networks.We utilize two types of acceleration methods: pruning and hints.Pruning can reduce model size by removing channels of layers.Hints can improve the performance of student model by transferring knowledge from teacher model.We demonstrate that pruning and hints are complementary to each other.On one hand, hints can benefit pruning by maintaining similar feature representations.On the other hand, the model pruned from teacher networks is a good initialization for student model, which increases the transferability between two networks.Our approach performs pruning stage and hints stage iteratively to further improve theperformance.Furthermore, we propose an algorithm to reconstruct the parameters of hints layer and make the pruned model more suitable for hints.Experiments were conducted on various tasks including classification and pose estimation.Results on CIFAR-10, ImageNet and COCO demonstrate the generalization and superiority of our framework.", "target": ["Dies ist eine Arbeit, die darauf abzielt, alle bestehenden Pruning- und Mimicmethoden zu verbessern."]} +{"source": "For typical sequence prediction problems such as language generation, maximum likelihood estimation (MLE) has commonly been adopted as it encourages the predicted sequence most consistent with the ground-truth sequence to have the highest probability of occurring.However, MLE focuses on once-to-all matching between the predicted sequence and gold-standard, consequently treating all incorrect predictions as being equally incorrect.We refer to this drawback as {\\it negative diversity ignorance} in this paper.Treating all incorrect predictions as equal unfairly downplays the nuance of these sequences' detailed token-wise structure.To counteract this, we augment the MLE loss by introducing an extra Kullback--Leibler divergence term derived by comparing a data-dependent Gaussian prior and the detailed training prediction.The proposed data-dependent Gaussian prior objective (D2GPo) is defined over a prior topological order of tokens and is poles apart from the data-independent Gaussian prior (L2 regularization) commonly adopted in smoothing the training of MLE.Experimental results show that the proposed method makes effective use of a more detailed prior in the data and has improved performance in typical language generation tasks, including supervised and unsupervised machine translation, text summarization, storytelling, and image captioning.", "target": ["Wir führen ein zusätzliches datenabhängiges Gauß'sches Prioritätsziel ein, um das derzeitige MLE-Training zu ergänzen, das dazu dient, das Vorwissen in den \"ground-truth\"-Daten zu erfassen."]} +{"source": "We propose an interactive classification approach for natural language queries.Instead of classifying given the natural language query only, we ask the user for additional information using a sequence of binary and multiple-choice questions.At each turn, we use a policy controller to decide if to present a question or pro-vide the user the final answer, and select the best question to ask by maximizing the system information gain.Our formulation enables bootstrapping the system without any interaction data, instead relying on non-interactive crowdsourcing an-notation tasks.Our evaluation shows the interaction helps the system increase its accuracy and handle ambiguous queries, while our approach effectively balances the number of questions and the final accuracy.", "target": ["Wir schlagen einen interaktiven Ansatz für die Klassifizierung von natürlichsprachlichen Anfragen vor, indem wir den Benutzer um zusätzliche Informationen bitten, unter der Nutzung des Informationgains und eines Reinforcement Learning Policy Controllers."]} +{"source": "Convolutional neural networks (CNNs) have achieved state of the art performance on recognizing and representing audio, images, videos and 3D volumes; that is, domains where the input can be characterized by a regular graph structure. However, generalizing CNNs to irregular domains like 3D meshes is challenging.Additionally, training data for 3D meshes is often limited.In this work, we generalize convolutional autoencoders to mesh surfaces.We perform spectral decomposition of meshes and apply convolutions directly in frequency space.In addition, we use max pooling and introduce upsampling within the network to represent meshes in a low dimensional space.We construct a complex dataset of 20,466 high resolution meshes with extreme facial expressions and encode it using our Convolutional Mesh Autoencoder.Despite limited training data, our method outperforms state-of-the-art PCA models of faces with 50% lower error, while using 75% fewer parameters.", "target": ["Convolutional Autoencoders verallgemeinert auf Mesh-Oberflächen für die Kodierung und Rekonstruktion von extremen 3D-Gesichtsausdrücken."]} +{"source": "Computing distances between examples is at the core of many learning algorithms for time series.Consequently, a great deal of work has gone into designing effective time series distance measures.We present Jiffy, a simple and scalable distance metric for multivariate time series.Our approach is to reframe the task as a representation learning problem---rather than design an elaborate distance function, we use a CNN to learn an embedding such that the Euclidean distance is effective.By aggressively max-pooling and downsampling, we are able to construct this embedding using a highly compact neural network.Experiments on a diverse set of multivariate time series datasets show that our approach consistently outperforms existing methods.", "target": ["Jiffy ist ein Convolutional Ansatz zum Erlernen einer Abstandsmetrik für multivariate Zeitreihen, der bestehende Methoden in Bezug auf die Klassifizierungsgenauigkeit der nächsten Nachbarn übertrifft."]} +{"source": "Prefrontal cortex (PFC) is a part of the brain which is responsible for behavior repertoire.Inspired by PFC functionality and connectivity, as well as human behavior formation process, we propose a novel modular architecture of neural networks with a Behavioral Module (BM) and corresponding end-to-end training strategy. This approach allows the efficient learning of behaviors and preferences representation.This property is particularly useful for user modeling (as for dialog agents) and recommendation tasks, as allows learning personalized representations of different user states. In the experiment with video games playing, the resultsshow that the proposed method allows separation of main task’s objectives andbehaviors between different BMs.The experiments also show network extendability through independent learning of new behavior patterns.Moreover, we demonstrate a strategy for an efficient transfer of newly learned BMs to unseen tasks.", "target": ["Eine erweiterbare modulare Architektur wird für die Entwicklung einer Vielzahl von Agentenverhaltensweisen im DQN vorgeschlagen."]} +{"source": "A major component of overfitting in model-free reinforcement learning (RL) involves the case where the agent may mistakenly correlate reward with certain spurious features from the observations generated by the Markov Decision Process (MDP).We provide a general framework for analyzing this scenario, which we use to design multiple synthetic benchmarks from only modifying the observation space of an MDP.When an agent overfits to different observation spaces even if the underlying MDP dynamics is fixed, we term this observational overfitting.Our experiments expose intriguing properties especially with regards to implicit regularization, and also corroborate results from previous works in RL generalization and supervised learning (SL).", "target": ["Wir isolieren einen Faktor der RL-Verallgemeinerung, indem wir den Fall analysieren, in dem der Agent nur auf die Beobachtungen überpasst. Wir zeigen, dass architektonische implizite Regularisierungen in diesem Regime auftreten."]} +{"source": "We propose a neural language model capable of unsupervised syntactic structure induction.The model leverages the structure information to form better semantic representations and better language modeling.Standard recurrent neural networks are limited by their structure and fail to efficiently use syntactic information.On the other hand, tree-structured recursive networks usually require additional structural supervision at the cost of human expert annotation.In this paper, We propose a novel neural language model, called the Parsing-Reading-Predict Networks (PRPN), that can simultaneously induce the syntactic structure from unannotated sentences and leverage the inferred structure to learn a better language model.In our model, the gradient can be directly back-propagated from the language model loss into the neural parsing network.Experiments show that the proposed model can discover the underlying syntactic structure and achieve state-of-the-art performance on word/character-level language model tasks.", "target": ["In diesem Papier schlagen wir ein neuartiges neuronales Sprachmodell vor, das so genannte Parsing-Reading-Predict Networks (PRPN), das gleichzeitig die syntaktische Struktur aus unannotierten Sätzen ableiten und die abgeleitete Struktur nutzen kann, um ein besseres Sprachmodell zu lernen."]} +{"source": "Unsupervised embedding learning aims to extract good representations from data without the use of human-annotated labels.Such techniques are apparently in the limelight because of the challenges in collecting massive-scale labels required for supervised learning.This paper proposes a comprehensive approach, called Super-AND, which is based on the Anchor Neighbourhood Discovery model.Multiple losses defined in Super-AND make similar samples gather even within a low-density space and keep features invariant against augmentation.As a result, our model outperforms existing approaches in various benchmark datasets and achieves an accuracy of 89.2% in CIFAR-10 with the Resnet18 backbone network, a 2.9% gain over the state-of-the-art.", "target": ["Wir haben einen umfassenden Ansatz für unüberwachtes Einbettungslernen auf der Grundlage des AND-Algorithmus vorgeschlagen."]} +{"source": "Analysis of histopathology slides is a critical step for many diagnoses, and in particular in oncology where it defines the gold standard.In the case of digital histopathological analysis, highly trained pathologists must review vast whole-slide-images of extreme digital resolution (100,000^2 pixels) across multiple zoom levels in order to locate abnormal regions of cells, or in some cases single cells, out of millions.The application of deep learning to this problem is hampered not only by small sample sizes, as typical datasets contain only a few hundred samples, but also by the generation of ground-truth localized annotations for training interpretable classification and segmentation models.We propose a method for disease available during training.Even without pixel-level annotations, we are able to demonstrate performance comparable with models trained with strong annotations on the Camelyon-16 lymph node metastases detection challenge.We accomplish this through the use of pre-trained deep convolutional networks, feature embedding, as well as learning via top instances and negative evidence, a multiple instance learning technique fromatp the field of semantic segmentation and object detection.", "target": ["Wir schlagen eine schwach überwachte Lernmethode für die Klassifizierung und Lokalisierung von Krebserkrankungen in extrem hochauflösenden histopathologischen Ganzbildaufnahmen vor, die nur bildweite Markierungen verwendet."]} +{"source": "Massively multi-label prediction/classification problems arise in environments like health-care or biology where it is useful to make very precise predictions.One challenge with massively multi-label problems is that there is often a long-tailed frequency distribution for the labels, resulting in few positive examples for the rare labels.We propose a solution to this problem by modifying the output layer of a neural network to create a Bayesian network of sigmoids which takes advantage of ontology relationships between the labels to help share information between the rare and the more common labels. We apply this method to the two massively multi-label tasks of disease prediction (ICD-9 codes) and protein function prediction (Gene Ontology terms) and obtain significant improvements in per-label AUROC and average precision.", "target": ["Wir schlagen eine neue Methode zur Verwendung von Ontologie-Informationen vor, um die Leistung bei massiven Mehr-Label-Prognosen/Klassifizierungsproblemen zu verbessern."]} +{"source": "Generative Adversarial Networks have made data generation possible in various use cases, but in case of complex, high-dimensional distributions it can be difficult to train them, because of convergence problems and the appearance of mode collapse.Sliced Wasserstein GANs and especially the application of the Max-Sliced Wasserstein distance made it possible to approximate Wasserstein distance during training in an efficient and stable way and helped ease convergence problems of these architectures.This method transforms sample assignment and distance calculation into sorting the one-dimensional projection of the samples, which results a sufficient approximation of the high-dimensional Wasserstein distance. In this paper we will demonstrate that the approximation of the Wasserstein distance by sorting the samples is not always the optimal approach and the greedy assignment of the real and fake samples can result faster convergence and better approximation of the original distribution.", "target": ["Zur Annäherung an die Wasserstein-Distanz wenden wir auf die projizierten Beispiele eine greedy Zuweisung an, anstatt zu sortieren"]} +{"source": "Lifelong machine learning focuses on adapting to novel tasks without forgetting the old tasks, whereas few-shot learning strives to learn a single task given a small amount of data.These two different research areas are crucial for artificial general intelligence, however, their existing studies have somehow assumed some impractical settings when training the models.For lifelong learning, the nature (or the quantity) of incoming tasks during inference time is assumed to be known at training time.As for few-shot learning, it is commonly assumed that a large number of tasks is available during training.Humans, on the other hand, can perform these learning tasks without regard to the aforementioned assumptions.Inspired by how the human brain works, we propose a novel model, called the Slow Thinking to Learn (STL), that makes sophisticated (and slightly slower) predictions by iteratively considering interactions between current and previously seen tasks at runtime.Having conducted experiments, the results empirically demonstrate the effectiveness of STL for more realistic lifelong and few-shot learning settings.", "target": ["In diesem Beitrag werden die Wechselwirkungen zwischen den Modellen für schnelles Lernen und langsame Vorhersage untersucht und es wird gezeigt, wie solche Wechselwirkungen die Fähigkeit von Maschinen verbessern können, die Probleme des Joint Lifelong Lernens und des Few-Shot Lernens zu lösen."]} +{"source": "Hypernetworks are meta neural networks that generate weights for a main neural network in an end-to-end differentiable manner.Despite extensive applications ranging from multi-task learning to Bayesian deep learning, the problem of optimizing hypernetworks has not been studied to date.We observe that classical weight initialization methods like Glorot & Bengio (2010) and He et al. (2015), when applied directly on a hypernet, fail to produce weights for the mainnet in the correct scale.We develop principled techniques for weight initialization in hypernets, and show that they lead to more stable mainnet weights, lower training loss, and faster convergence.", "target": ["Die erste prinzipielle Gewichtsinitialisierungsmethode für Hypernetze"]} +{"source": "For bidirectional joint image-text modeling, we develop variational hetero-encoder (VHE) randomized generative adversarial network (GAN), a versatile deep generative model that integrates a probabilistic text decoder, probabilistic image encoder, and GAN into a coherent end-to-end multi-modality learning framework.VHE randomized GAN (VHE-GAN) encodes an image to decode its associated text, and feeds the variational posterior as the source of randomness into the GAN image generator.We plug three off-the-shelf modules, including a deep topic model, a ladder-structured image encoder, and StackGAN++, into VHE-GAN, which already achieves competitive performance.This further motivates the development of VHE-raster-scan-GAN that generates photo-realistic images in not only a multi-scale low-to-high-resolution manner, but also a hierarchical-semantic coarse-to-fine fashion.By capturing and relating hierarchical semantic and visual concepts with end-to-end training, VHE-raster-scan-GAN achieves state-of-the-art performance in a wide variety of image-text multi-modality learning and generation tasks.", "target": ["Ein neuartiges Bayes'sches Deep-Learning-Framework, das hierarchische semantische und visuelle Konzepte erfasst und miteinander in Beziehung setzt und bei einer Vielzahl von Bild- und Textmodellierungs- und -generierungsaufgaben gute Ergebnisse erzielt."]} +{"source": "Current classical planners are very successful in finding (non-optimal) plans, even for large planning instances.To do so, most planners rely on a preprocessing stage that computes a grounded representation of the task.Whenever the grounded task is too big to be generated (i.e., whenever this preprocess fails) the instance cannot even be tackled by the actual planner.To address this issue, we introduce a partial grounding approach that grounds only a projection of the task, when complete grounding is not feasible.We propose a guiding mechanism that, for a given domain, identifies the parts of a task that are relevant to find a plan by using off-the-shelf machine learning methods.Our empirical evaluation attests that the approach is capable of solving planning instances that are too big to be fully grounded.", "target": ["Diese Arbeit führt partielles Grounding ein, um das Problem zu lösen, das auftritt, wenn der vollständige Grounding-Prozess, d. h. die Übersetzung einer PDDL-Eingabeaufgabe in eine geerdete Darstellung wie STRIPS, aufgrund von Speicher- oder Zeitbeschränkungen nicht durchführbar ist."]} +{"source": "In this paper, we introduce a novel method to interpret recurrent neural networks (RNNs), particularly long short-term memory networks (LSTMs) at the cellular level.We propose a systematic pipeline for interpreting individual hidden state dynamics within the network using response characterization methods.The ranked contribution of individual cells to the network's output is computed by analyzing a set of interpretable metrics of their decoupled step and sinusoidal responses.As a result, our method is able to uniquely identify neurons with insightful dynamics, quantify relationships between dynamical properties and test accuracy through ablation analysis, and interpret the impact of network capacity on a network's dynamical distribution.Finally, we demonstrate generalizability and scalability of our method by evaluating a series of different benchmark sequential datasets.", "target": ["Einführung der Antwort-Charakterisierungsmethode zur Interpretation der Zelldynamik in gelernten Long Short-Term Memory Netzwerken (LSTM)."]} +{"source": "Decades of research on the neural code underlying spatial navigation have revealed a diverse set of neural response properties.The Entorhinal Cortex (EC) of the mammalian brain contains a rich set of spatial correlates, including grid cells which encode space using tessellating patterns.However, the mechanisms and functional significance of these spatial representations remain largely mysterious.As a new way to understand these neural representations, we trained recurrent neural networks (RNNs) to perform navigation tasks in 2D arenas based on velocity inputs.Surprisingly, we find that grid-like spatial response patterns emerge in trained networks, along with units that exhibit other spatial correlates, including border cells and band-like cells.All these different functional types of neurons have been observed experimentally.The order of the emergence of grid-like and border cells is also consistent with observations from developmental studies.Together, our results suggest that grid cells, border cells and others as observed in EC may be a natural solution for representing space efficiently given the predominant recurrent connections in the neural circuits.", "target": ["Unseres Wissens ist dies die erste Studie, die zeigt, wie neuronale Repräsentationen des Raums, einschließlich gitterartiger Zellen und Grenzzellen, wie sie im Gehirn beobachtet werden, aus dem Training eines rekurrenten neuronalen Netzes zur Durchführung von Navigationsaufgaben hervorgehen können."]} +{"source": "Source separation for music is the task of isolating contributions, or stems, from different instruments recorded individually and arranged together to form a song.Such components include voice, bass, drums and any other accompaniments.While end-to-end models that directly generate the waveform are state-of-the-art in many audio synthesis problems, the best multi-instrument source separation models generate masks on the magnitude spectrum and achieve performances far above current end-to-end, waveform-to-waveform models.We present an in-depth analysis of a new architecture, which we will refer to as Demucs, based on a (transposed) convolutional autoencoder, with a bidirectional LSTM at the bottleneck layer and skip-connections as in U-Networks (Ronneberger et al., 2015).Compared to the state-of-the-art waveform-to-waveform model, Wave-U-Net (Stoller et al., 2018), the main features of our approach in addition of the bi-LSTM are the use of trans-posed convolution layers instead of upsampling-convolution blocks, the use of gated linear units, exponentially growing the number of channels with depth and a new careful initialization of the weights. Results on the MusDB dataset show that our architecture achieves a signal-to-distortion ratio (SDR) nearly 2.2 points higher than the best waveform-to-waveform competitor (from 3.2 to 5.4 SDR).This makes our model match the state-of-the-art performances on this dataset, bridging the performance gap between models that operate on the spectrogram and end-to-end approaches.", "target": ["Wir vergleichen die Leistung eines Spektrogramm-basierten Modells mit einem Modell, das durchgängig im Wellenformbereich trainiert wurde."]} +{"source": "Although challenging, strategy profile evaluation in large connected learner networks is crucial for enabling the next wave of machine learning applications.Recently, $\\alpha$-Rank, an evolutionary algorithm, has been proposed as a solution for ranking joint policy profiles in multi-agent systems.$\\alpha$-Rank claimed scalability through a polynomial time implementation with respect to the total number of pure strategy profiles.In this paper, we formally prove that such a claim is not grounded.In fact, we show that $\\alpha$-Rank exhibits an exponential complexity in number of agents, hindering its application beyond a small finite number of joint profiles.Realizing such a limitation, we contribute by proposing a scalable evaluation protocol that we title $\\alpha^{\\alpha}$-Rank.Our method combines evolutionary dynamics with stochastic optimization and double oracles for \\emph{truly} scalable ranking with linear (in number of agents) time and memory complexities.Our contributions allow us, for the first time, to conduct large-scale evaluation experiments of multi-agent systems, where we show successful results on large joint strategy profiles with sizes in the order of $\\mathcal{O}(2^{25})$ (i.e., $\\approx \\text{$33$ million strategies}$) -- a setting not evaluable using current techniques.", "target": ["Wir bieten eine skalierbare Lösung für die Multi-Agenten-Evaluierung mit linearer Zeit- und Speicherkomplexität in Bezug auf die Anzahl der Agenten"]} +{"source": "Deep neural networks are known to be annotation-hungry.Numerous efforts have been devoted to reducing the annotation cost when learning with deep networks.Two prominent directions include learning with noisy labels and semi-supervised learning by exploiting unlabeled data.In this work, we propose DivideMix, a novel framework for learning with noisy labels by leveraging semi-supervised learning techniques.In particular, DivideMix models the per-sample loss distribution with a mixture model to dynamically divide the training data into a labeled set with clean samples and an unlabeled set with noisy samples, and trains the model on both the labeled and unlabeled data in a semi-supervised manner.To avoid confirmation bias, we simultaneously train two diverged networks where each network uses the dataset division from the other network.During the semi-supervised training phase, we improve the MixMatch strategy by performing label co-refinement and label co-guessing on labeled and unlabeled samples, respectively.Experiments on multiple benchmark datasets demonstrate substantial improvements over state-of-the-art methods.Code is available at https://github.com/LiJunnan1992/DivideMix .", "target": ["Wir schlagen einen neuartigen halb-überwachten Lernansatz mit SOTA-Leistung zur Bekämpfung des Lernens mit verrauschten Etiketten vor."]} +{"source": "We present a new algorithm to train a robust neural network against adversarial attacks. Our algorithm is motivated by the following two ideas.First, although recent work has demonstrated that fusing randomness can improve the robustness of neural networks (Liu 2017), we noticed that adding noise blindly to all the layers is not the optimal way to incorporate randomness. Instead, we model randomness under the framework of Bayesian Neural Network (BNN) to formally learn the posterior distribution of models in a scalable way.Second, we formulate the mini-max problem in BNN to learn the best model distribution under adversarial attacks, leading to an adversarial-trained Bayesian neural net.Experiment results demonstrate that the proposed algorithm achieves state-of-the-art performance under strong attacks.On CIFAR-10 with VGG network, our model leads to 14% accuracy improvement compared with adversarial training (Madry 2017) and random self-ensemble (Liu, 2017) under PGD attack with 0.035 distortion, and the gap becomes even larger on a subset of ImageNet.", "target": ["Wir entwickeln eine Trainingsmethode für Bayes'sche neuronale Netze, die eine viel stärkere Verteidigung gegen White-Box-Angriffe bietet."]} +{"source": "Federated learning distributes model training among a multitude of agents, who, guided by privacy concerns, perform training using their local data but share only model parameter updates, for iterative aggregation at the server.In this work, we explore the threat of model poisoning attacks on federated learning initiated by a single, non-colluding malicious agent where the adversarial objective is to cause the model to misclassify a set of chosen inputs with high confidence.We explore a number of strategies to carry out this attack, starting with simple boosting of the malicious agent's update to overcome the effects of other agents' updates.To increase attack stealth, we propose an alternating minimization strategy, which alternately optimizes for the training loss and the adversarial objective.We follow up by using parameter estimation for the benign agents' updates to improve on attack success.Finally, we use a suite of interpretability techniques to generate visual explanations of model decisions for both benign and malicious models and show that the explanations are nearly visually indistinguishable.Our results indicate that even a highly constrained adversary can carry out model poisoning attacks while simultaneously maintaining stealth, thus highlighting the vulnerability of the federated learning setting and the need to develop effective defense strategies.", "target": ["Wirksame Modellvergiftungsangriffe auf das föderierte Lernen, die eine gezielte Fehlklassifizierung der gewünschten Eingaben mit hoher Zuverlässigkeit bewirken können"]} +{"source": "Despite rapid advances in speech recognition, current models remain brittle to superficial perturbations to their inputs.Small amounts of noise can destroy the performance of an otherwise state-of-the-art model.To harden models against background noise, practitioners often perform data augmentation, adding artificially-noised examples to the training set, carrying over the original label.In this paper, we hypothesize that a clean example and its superficially perturbed counterparts shouldn't merely map to the same class--- they should map to the same representation.We propose invariant-representation-learning (IRL): At each training iteration, for each training example, we sample a noisy counterpart.We then apply a penalty term to coerce matched representations at each layer (above some chosen layer).Our key results, demonstrated on the LibriSpeech dataset are the following:(i) IRL significantly reduces character error rates (CER)on both `clean' (3.3% vs 6.5%) and `other' (11.0% vs 18.1%) test sets;(ii) on several out-of-domain noise settings (different from those seen during training), IRL's benefits are even more pronounced.Careful ablations confirm that our results are not simply due to shrinking activations at the chosen layers.", "target": ["In dieser Arbeit stellen wir die Hypothese auf, dass oberflächlich gestörte Datenpunkte nicht nur der gleichen Klasse zugeordnet werden sollten, sondern auch der gleichen Repräsentation zugeordnet werden sollten."]} +{"source": "In this paper we design a harmonic acoustic model for pitch detection.This model arranges conventional convolution and sparse convolution in a way such that the global harmonic patterns captured by sparse convolution are composed of the enough number of local patterns captured by layers of conventional convolution.When trained on the MAPS dataset, the harmonic model outperforms all existing pitch detection systems trained on the same dataset.Most impressively, when trained on MAPS with simple data augmentation, the harmonic model with an LSTM layer on top surpasses an up-to-date, more complex pitch detection system trained on the MAESTRO dataset to which complicated data augmentation is applied and whose training split is an order-of-magnitude larger than the training split of MAPS.The harmonic model has demonstrated potential to be used for advanced automatic music transcription (AMT) systems.", "target": ["Harmonisch akustisches Modell"]} +{"source": "Learning domain-invariant representation is a dominant approach for domain generalization.However, previous methods based on domain invariance overlooked the underlying dependency of classes on domains, which is responsible for the trade-off between classification accuracy and the invariance.This study proposes a novel method {\\em adversarial feature learning under accuracy constraint (AFLAC)}, which maximizes domain invariance within a range that does not interfere with accuracy.Empirical validations show that the performance of AFLAC is superior to that of baseline methods, supporting the importance of considering the dependency and the efficacy of the proposed method to overcome the problem.", "target": ["Den durch die Abhängigkeit der Klassen von den Domänen verursachten Kompromiss durch die Verbesserung der kontradiktorischen Netze der Domänen anzugehen"]} +{"source": "Recent advances in deep generative models have lead to remarkable progress in synthesizing high quality images.Following their successful application in image processing and representation learning, an important next step is to consider videos.Learning generative models of video is a much harder task, requiring a model to capture the temporal dynamics of a scene, in addition to the visual presentation of objects.While recent generative models of video have had some success, current progress is hampered by the lack of qualitative metrics that consider visual quality, temporal coherence, and diversity of samples.To this extent we propose Fréchet Video Distance (FVD), a new metric for generative models of video based on FID.We contribute a large-scale human study, which confirms that FVD correlates well with qualitative human judgment of generated videos.", "target": ["Wir schlagen FVD vor: eine neue Metrik für generative Videomodelle auf der Grundlage von FID. Eine groß angelegte Humanstudie bestätigt, dass FVD gut mit der qualitativen menschlichen Beurteilung von generierten Videos korreliert."]} +{"source": "Despite advances in deep learning, artificial neural networks do not learn the same way as humans do.Today, neural networks can learn multiple tasks when trained on them jointly, but cannot maintain performance on learnt tasks when tasks are presented one at a time -- this phenomenon called catastrophic forgetting is a fundamental challenge to overcome before neural networks can learn continually from incoming data.In this work, we derive inspiration from human memory to develop an architecture capable of learning continuously from sequentially incoming tasks, while averting catastrophic forgetting.Specifically, our model consists of a dual memory architecture to emulate the complementary learning systems (hippocampus and the neocortex) in the human brain and maintains a consolidated long-term memory via generative replay of past experiences.We(i) substantiate our claim that replay should be generative,(ii) show the benefits of generative replay and dual memory via experiments, and(iii) demonstrate improved performance retention even for small models with low capacity.Our architecture displays many important characteristics of the human memory and provides insights on the connection between sleep and learning in humans.", "target": ["Eine dem menschlichen Gehirn nachempfundene duale Gedächtnisarchitektur zum Erlernen von sequentiell eingehenden Aufgaben bei gleichzeitiger Vermeidung von katastrophalem Vergessen."]} +{"source": "Most research on lifelong learning applies to images or games, but not language.We present LAMOL, a simple yet effective method for lifelong language learning (LLL) based on language modeling.LAMOL replays pseudo-samples of previous tasks while requiring no extra memory or model capacity.Specifically, LAMOL is a language model that simultaneously learns to solve the tasks and generate training samples.When the model is trained for a new task, it generates pseudo-samples of previous tasks for training alongside data for the new task.The results show that LAMOL prevents catastrophic forgetting without any sign of intransigence and can perform five very different language tasks sequentially with only one model. Overall, LAMOL outperforms previous methods by a considerable margin and is only 2-3% worse than multitasking, which is usually considered the LLL upper bound.The source code is available at https://github.com/jojotenya/LAMOL.", "target": ["Sprachmodellierung für lebenslanges Sprachenlernen."]} +{"source": "Deep learning natural language processing models often use vector word embeddings, such as word2vec or GloVe, to represent words.A discrete sequence of words can be much more easily integrated with downstream neural layers if it is represented as a sequence of continuous vectors.Also, semantic relationships between words, learned from a text corpus, can be encoded in the relative configurations of the embedding vectors.However, storing and accessing embedding vectors for all words in a dictionary requires large amount of space, and may stain systems with limited GPU memory.Here, we used approaches inspired by quantum computing to propose two related methods, word2ket and word2ketXS, for storing word embedding matrix during training and inference in a highly efficient way.Our approach achieves a hundred-fold or more reduction in the space required to store the embeddings with almost no relative drop in accuracy in practical natural language processing tasks.", "target": ["Wir nutzen Ideen aus der Quanteninformatik, um Worteinbettungen vorzuschlagen, die mit viel weniger trainierbaren Parametern auskommen."]} +{"source": "One of the distinguishing aspects of human language is its compositionality, which allows us to describe complex environments with limited vocabulary.Previously, it has been shown that neural network agents can learn to communicate in a highly structured, possibly compositional language based on disentangled input (e.g. hand- engineered features).Humans, however, do not learn to communicate based on well-summarized features.In this work, we train neural agents to simultaneously develop visual perception from raw image pixels, and learn to communicate with a sequence of discrete symbols.The agents play an image description game where the image contains factors such as colors and shapes.We train the agents using the obverter technique where an agent introspects to generate messages that maximize its own understanding.Through qualitative analysis, visualization and a zero-shot test, we show that the agents can develop, out of raw image pixels, a language with compositional properties, given a proper pressure from the environment.", "target": ["Wir trainieren neuronale Netzwerk-Agenten, um eine Sprache mit kompositorischen Eigenschaften aus rohen Pixel-Eingaben zu entwickeln."]} +{"source": "The ability to forecast a set of likely yet diverse possible future behaviors of an agent (e.g., future trajectories of a pedestrian) is essential for safety-critical perception systems (e.g., autonomous vehicles).In particular, a set of possible future behaviors generated by the system must be diverse to account for all possible outcomes in order to take necessary safety precautions.It is not sufficient to maintain a set of the most likely future outcomes because the set may only contain perturbations of a dominating single outcome (major mode).While generative models such as variational autoencoders (VAEs) have been shown to be a powerful tool for learning a distribution over future trajectories, randomly drawn samples from the learned implicit likelihood model may not be diverse -- the likelihood model is derived from the training data distribution and the samples will concentrate around the major mode of the data.In this work, we propose to learn a diversity sampling function (DSF) that generates a diverse yet likely set of future trajectories.The DSF maps forecasting context features to a set of latent codes which can be decoded by a generative model (e.g., VAE) into a set of diverse trajectory samples.Concretely, the process of identifying the diverse set of samples is posed as DSF parameter estimation.To learn the parameters of the DSF, the diversity of the trajectory samples is evaluated by a diversity loss based on a determinantal point process (DPP).Gradient descent is performed over the DSF parameters, which in turn moves the latent codes of the sample set to find an optimal set of diverse yet likely trajectories.Our method is a novel application of DPPs to optimize a set of items (forecasted trajectories) in continuous space.We demonstrate the diversity of the trajectories produced by our approach on both low-dimensional 2D trajectory data and high-dimensional human motion data.", "target": ["Wir lernen eine Diversity-Sampling-Funktion mit DPPs, um einen vielfältigen Satz von Proben aus einem generativen Modell zu erhalten."]} +{"source": "There is mounting evidence that pretraining can be valuable for neural network language understanding models, but we do not yet have a clear understanding of how the choice of pretraining objective affects the type of linguistic information that models learn.With this in mind, we compare four objectives---language modeling, translation, skip-thought, and autoencoding---on their ability to induce syntactic and part-of-speech information, holding constant the genre and quantity of training data.We find that representations from language models consistently perform best on our syntactic auxiliary prediction tasks, even when trained on relatively small amounts of data, which suggests that language modeling may be the best data-rich pretraining task for transfer learning applications requiring syntactic information.We also find that a randomly-initialized, frozen model can perform strikingly well on our auxiliary tasks, but that this effect disappears when the amount of training data for the auxiliary tasks is reduced.", "target": ["Repräsentationen von Sprachmodellen schneiden bei der Vorhersage von syntaktischen Hilfsmitteln durchweg besser ab als Übersetzungsencoder."]} +{"source": "We consider the problem of generating configurations that satisfy physical constraints for optimal material nano-pattern design, where multiple (and often conflicting) properties need to be simultaneously satisfied. Consider, for example, the trade-off between thermal resistance, electrical conductivity, and mechanical stability needed to design a nano-porous template with optimal thermoelectric efficiency. To that end, we leverage the posterior regularization framework andshow that this constraint satisfaction problem can be formulated as sampling froma Gibbs distribution. The main challenges come from the black-box nature ofthose physical constraints, since they are obtained via solving highly non-linearPDEs.To overcome those difficulties, we introduce Surrogate-based Constrained Langevin dynamics for black-box sampling.We explore two surrogate approaches.The first approach exploits zero-order approximation of gradients in the Langevin Sampling and we refer to it as Zero-Order Langevin.In practice, this approach can be prohibitive since we still need to often query the expensive PDE solvers.The second approach approximates the gradients in the Langevin dynamics with deep neural networks, allowing us an efficient sampling strategy using the surrogate model.We prove the convergence of those two approaches when the target distribution is log-concave and smooth.We show the effectiveness of both approaches in designing optimal nano-porous material configurations, where the goal is to produce nano-pattern templates with low thermal conductivity and reasonable mechanical stability.", "target": ["Wir schlagen ein auf Surrogate basierendes beschränktes Langevin-Sampling vor, das im Design von nano-porigen Materialkonfigurationen Anwendung findet."]} +{"source": "There is growing interest in geometrically-inspired embeddings for learning hierarchies, partial orders, and lattice structures, with natural applications to transitive relational data such as entailment graphs.Recent work has extended these ideas beyond deterministic hierarchies to probabilistically calibrated models, which enable learning from uncertain supervision and inferring soft-inclusions among concepts, while maintaining the geometric inductive bias of hierarchical embedding models.We build on the Box Lattice model of Vilnis et al. (2018), which showed promising results in modeling soft-inclusions through an overlapping hierarchy of sets, parameterized as high-dimensional hyperrectangles (boxes).However, the hard edges of the boxes present difficulties for standard gradient based optimization; that work employed a special surrogate function for the disjoint case, but we find this method to be fragile. In this work, we present a novel hierarchical embedding model, inspired by a relaxation of box embeddings into parameterized density functions using Gaussian convolutions over the boxes.Our approach provides an alternative surrogate to the original lattice measure that improves the robustness of optimization in the disjoint case, while also preserving the desirable properties with respect to the original lattice.We demonstrate increased or matching performance on WordNet hypernymy prediction, Flickr caption entailment, and a MovieLens-based market basket dataset.We show especially marked improvements in the case of sparse data, where many conditional probabilities should be low, and thus boxes should be nearly disjoint.", "target": ["Verbesserung hierarchischer Einbettungsmodelle durch Kernelglättung"]} +{"source": "We present a weakly-supervised data augmentation approach to improve Named Entity Recognition (NER) in a challenging domain: extracting biomedical entities (e.g., proteins) from the scientific literature.First, we train a neural NER (NNER) model over a small seed of fully-labeled examples.Second, we use a reference set of entity names (e.g., proteins in UniProt) to identify entity mentions with high precision, but low recall, on an unlabeled corpus.Third, we use the NNER model to assign weak labels to the corpus.Finally, we retrain our NNER model iteratively over the augmented training set, including the seed, the reference-set examples, and the weakly-labeled examples, which results in refined labels.We show empirically that this augmented bootstrapping process significantly improves NER performance, and discuss the factors impacting the efficacy of the approach.", "target": ["Augmented-Bootstrapping-Ansatz, der Informationen aus einer Referenzmenge mit iterativen Verfeinerungen von Soft-Labels kombiniert, um die Erkennung von Namensbestandteilen aus biomedizinischer Literatur zu verbessern."]} +{"source": "Quantum machine learning methods have the potential to facilitate learning using extremely large datasets.While the availability of data for training machine learning models is steadily increasing, oftentimes it is much easier to collect feature vectors that to obtain the corresponding labels.One of the approaches for addressing this issue is to use semi-supervised learning, which leverages not only the labeled samples, but also unlabeled feature vectors.Here, we present a quantum machine learning algorithm for training Semi-Supervised Kernel Support Vector Machines.The algorithm uses recent advances in quantum sample-based Hamiltonian simulation to extend the existing Quantum LS-SVM algorithm to handle the semi-supervised term in the loss, while maintaining the same quantum speedup as the Quantum LS-SVM.", "target": ["Wir erweitern die Quanten-SVMs auf eine halbüberwachte Umgebung, um mit dem wahrscheinlichen Problem vieler fehlender Klassenbezeichnungen in großen Datensätzen umzugehen."]} +{"source": "Deep neural networks have become the state-of-the-art models in numerous machine learning tasks.However, general guidance to network architecture design is still missing.In our work, we bridge deep neural network design with numerical differential equations.We show that many effective networks, such as ResNet, PolyNet, FractalNet and RevNet, can be interpreted as different numerical discretizations of differential equations.This finding brings us a brand new perspective on the design of effective deep architectures.We can take advantage of the rich knowledge in numerical analysis to guide us in designing new and potentially more effective deep networks.As an example, we propose a linear multi-step architecture (LM-architecture) which is inspired by the linear multi-step method solving ordinary differential equations.The LM-architecture is an effective structure that can be used on any ResNet-like networks.In particular, we demonstrate that LM-ResNet and LM-ResNeXt (i.e. the networks obtained by applying the LM-architecture on ResNet and ResNeXt respectively) can achieve noticeably higher accuracy than ResNet and ResNeXt on both CIFAR and ImageNet with comparable numbers of trainable parameters.In particular, on both CIFAR and ImageNet, LM-ResNet/LM-ResNeXt can significantly compress (>50%) the original networks while maintaining a similar performance.This can be explained mathematically using the concept of modified equation from numerical analysis.Last but not least, we also establish a connection between stochastic control and noise injection in the training process which helps to improve generalization of the networks.Furthermore, by relating stochastic training strategy with stochastic dynamic system, we can easily apply stochastic training to the networks with the LM-architecture.As an example, we introduced stochastic depth to LM-ResNet and achieve significant improvement over the original LM-ResNet on CIFAR10.", "target": ["Dieses Werk schlägt eine Brücke zwischen tiefen Netzwerkarchitekturen und numerischen (stochastischen) Differentialgleichungen. Diese neue Perspektive ermöglicht neue Designs von effektiveren tiefen neuronalen Netzen."]} +{"source": "Transforming a graphical user interface screenshot created by a designer into computer code is a typical task conducted by a developer in order to build customized software, websites, and mobile applications.In this paper, we show that deep learning methods can be leveraged to train a model end-to-end to automatically generate code from a single input image with over 77% of accuracy for three different platforms (i.e. iOS, Android and web-based technologies).", "target": ["CNN und LSTM, um markup-ähnlichen Code zu erzeugen, der Bilder von grafischen Benutzeroberflächen beschreibt."]} +{"source": "Computer vision tasks such as image classification, image retrieval and few-shot learning are currently dominated by Euclidean and spherical embeddings, so that the final decisions about class belongings or the degree of similarity are made using linear hyperplanes, Euclidean distances, or spherical geodesic distances (cosine similarity).In this work, we demonstrate that in many practical scenarios hyperbolic embeddings provide a better alternative.", "target": ["Wir zeigen, dass hyperbolische Einbettungen für anspruchsvolle Computer-Vision-Aufgaben nützlich sind, insbesondere für die Klassifizierung von wenigen Bildern."]} +{"source": "High-dimensional time series are common in many domains.Since human cognition is not optimized to work well in high-dimensional spaces, these areas could benefit from interpretable low-dimensional representations.However, most representation learning algorithms for time series data are difficult to interpret.This is due to non-intuitive mappings from data features to salient properties of the representation and non-smoothness over time.To address this problem, we propose a new representation learning framework building on ideas from interpretable discrete dimensionality reduction and deep generative modeling.This framework allows us to learn discrete representations of time series, which give rise to smooth and interpretable embeddings with superior clustering performance.We introduce a new way to overcome the non-differentiability in discrete representation learning and present a gradient-based version of the traditional self-organizing map algorithm that is more performant than the original.Furthermore, to allow for a probabilistic interpretation of our method, we integrate a Markov model in the representation space.This model uncovers the temporal transition structure, improves clustering performance even further and provides additional explanatory insights as well as a natural representation of uncertainty.We evaluate our model in terms of clustering performance and interpretability on static (Fashion-)MNIST data, a time series of linearly interpolated (Fashion-)MNIST images, a chaotic Lorenz attractor system with two macro states, as well as on a challenging real world medical time series application on the eICU data set.Our learned representations compare favorably with competitor methods and facilitate downstream tasks on the real world data.", "target": ["Wir stellen eine Methode zum Erlernen interpretierbarer Repräsentationen von Zeitreihen vor, die auf Ideen von Variations-Autoencodern, selbstorganisierenden Karten und probabilistischen Modellen beruht."]} +{"source": "We propose Significance-Offset Convolutional Neural Network, a deep convolutional network architecture for regression of multivariate asynchronous time series. The model is inspired by standard autoregressive (AR) models and gating mechanisms used in recurrent neural networks. It involves an AR-like weighting system, where the final predictor is obtained as a weighted sum of adjusted regressors, while the weights are data-dependent functions learnt through a convolutional network.The architecture was designed for applications on asynchronous time series and is evaluated on such datasets: a hedge fund proprietary dataset of over 2 million quotes for a credit derivative index, an artificially generated noisy autoregressive series and household electricity consumption dataset. The pro-posed architecture achieves promising results as compared to convolutional and recurrent neural networks.The code for the numerical experiments and the architecture implementation will be shared online to make the research reproducible.", "target": ["Convolutional Architecture zum Lernen datenabhängiger Gewichte für die autoregressive Vorhersage von Zeitreihen."]} +{"source": "MixUp is a data augmentation scheme in which pairs of training samples and their corresponding labels are mixed using linear coefficients.Without label mixing, MixUp becomes a more conventional scheme: input samples are moved but their original labels are retained.Because samples are preferentially moved in the direction of other classes \\iffalse -- which are typically clustered in input space -- \\fi we refer to this method as directional adversarial training, or DAT.We show that under two mild conditions, MixUp asymptotically convergences to a subset of DAT.We define untied MixUp (UMixUp), a superset of MixUp wherein training labels are mixed with different linear coefficients to those of their corresponding samples.We show that under the same mild conditions, untied MixUp converges to the entire class of DAT schemes.Motivated by the understanding that UMixUp is both a generalization of MixUp and a form of adversarial training, we experiment with different datasets and loss functions to show that UMixUp provides improved performance over MixUp.In short, we present a novel interpretation of MixUp as belonging to a class highly analogous to adversarial training, and on this basis we introduce a simple generalization which outperforms MixUp.", "target": ["Wir stellen eine neuartige Interpretation von MixUp vor, die einer Klasse angehört, die in hohem Maße mit dem gegnerischen Training vergleichbar ist, und auf dieser Grundlage führen wir eine einfache Verallgemeinerung ein, die MixUp übertrifft"]} +{"source": "Plan recognition aims to look for target plans to best explain the observed actions based on plan libraries and/or domain models.Despite the success of previous approaches on plan recognition, they mostly rely on correct action observations. Recent advances in visual activity recognition have the potential of enabling applications such as automated video surveillance.Effective approaches for such problems would require the ability to recognize the plans of agents from video information.Traditional plan recognition algorithms rely on access to detailed planning domain models.One recent promising direction involves learning approximate (or shallow) domain models directly from the observed activity sequences.Such plan recognition approaches expect observed action sequences as inputs.However, visual inference results are often noisy and uncertain, typically represented as a distribution over possible actions.In this work, we develop a visual plan recognition framework that recognizes plans with an approximate domain model learned from uncertain visual data.", "target": ["Behandlung von Unsicherheiten in der visuellen Wahrnehmung für die Planerkennung"]} +{"source": "We consider the task of answering complex multi-hop questions using a corpus as a virtual knowledge base (KB).In particular, we describe a neural module, DrKIT, that traverses textual data like a virtual KB, softly following paths of relations between mentions of entities in the corpus.At each step the operation uses a combination of sparse-matrix TFIDF indices and maximum inner product search (MIPS) on a special index of contextual representations.This module is differentiable, so the full system can be trained completely end-to-end using gradient based methods, starting from natural language inputs.We also describe a pretraining scheme for the index mention encoder by generating hard negative examples using existing knowledge bases.We show that DrKIT improves accuracy by 9 points on 3-hop questions in the MetaQA dataset, cutting the gap between text-based and KB-based state-of-the-art by 70%.DrKIT is also very efficient, processing upto 10x more queries per second than existing state-of-the-art QA systems.", "target": ["Differenzierbarer Multi-Hop-Zugriff auf eine textuelle Wissensbasis mit indizierten kontextuellen Darstellungen"]} +{"source": "In spite of their great success, traditional factorization algorithms typically do not support features (e.g., Matrix Factorization), or their complexity scales quadratically with the number of features (e.g, Factorization Machine).On the other hand, neural methods allow large feature sets, but are often designed for a specific application.We propose novel deep factorization methods that allow efficient and flexible feature representation.For example, we enable describing items with natural language with complexity linear to the vocabulary size—this enables prediction for unseen items and avoids the cold start problem.We show that our architecture can generalize some previously published single-purpose neural architectures.Our experiments suggest improved training times and accuracy compared to shallow methods.", "target": ["Skalierbarer Allzweck-Faktorisierungsalgorithmus - hilft auch, das Kaltstartproblem zu umgehen."]} +{"source": "Augmented Reality (AR) can assist with physical tasks such as object assembly through the use of “situated instructions”.These instructions can be in the form of videos, pictures, text or guiding animations, where the most helpful media among these is highly dependent on both the user and the nature of the task.Our work supports the authoring of AR tutorials for assembly tasks with little overhead beyond simply performing the task itself.The presented system, AuthAR reduces the time and effort required to build interactive AR tutorials by automatically generating key components of the AR tutorial while the author is assembling the physical pieces.Further, the system guides authors through the process of adding videos, pictures, text and animations to the tutorial.This concurrent assembly and tutorial generation approach allows for authoring of portable tutorials that fit the preferences of different end users.", "target": ["Wir stellen ein System zur Erstellung von Tutorials mit gemischten Medien vor, das die Erstellung von Videos, Bildern, Text und dynamischen Anweisungen an Ort und Stelle rationalisiert."]} +{"source": "Monitoring patients in ICU is a challenging and high-cost task.Hence, predicting the condition of patients during their ICU stay can help provide better acute care and plan the hospital's resources.There has been continuous progress in machine learning research for ICU management, and most of this work has focused on using time series signals recorded by ICU instruments.In our work, we show that adding clinical notes as another modality improves the performance of the model for three benchmark tasks: in-hospital mortality prediction, modeling decompensation, and length of stay forecasting that play an important role in ICU management.While the time-series data is measured at regular intervals, doctor notes are charted at irregular times, making it challenging to model them together.We propose a method to model them jointly, achieving considerable improvement across benchmark tasks over baseline time-series model.", "target": ["Wir konnten nachweisen, dass die Verwendung von klinischen Notizen in Verbindung mit ICU-Instrumenten die Leistung bei ICU-Management-Benchmark-Aufgaben verbessert."]} +{"source": "Existing works in deep Multi-Agent Reinforcement Learning (MARL) mainly focus on coordinating cooperative agents to complete certain tasks jointly.However, in many cases of the real world, agents are self-interested such as employees in a company and clubs in a league.Therefore, the leader, i.e., the manager of the company or the league, needs to provide bonuses to followers for efficient coordination, which we call expensive coordination.The main difficulties of expensive coordination are thati) the leader has to consider the long-term effect and predict the followers' behaviors when assigning bonuses andii) the complex interactions between followers make the training process hard to converge, especially when the leader's policy changes with time.In this work, we address this problem through an event-based deep RL approach.Our main contributions are threefold.(1) We model the leader's decision-making process as a semi-Markov Decision Process and propose a novel multi-agent event-based policy gradient to learn the leader's long-term policy.(2) We exploit the leader-follower consistency scheme to design a follower-aware module and a follower-specific attention module to predict the followers' behaviors and make accurate response to their behaviors.(3) We propose an action abstraction-based policy gradient algorithm to reduce the followers' decision space and thus accelerate the training process of followers.Experiments in resource collections, navigation, and the predator-prey game reveal that our approach outperforms the state-of-the-art methods dramatically.", "target": ["Wir schlagen einen ereignisbasierten Policy-Gradienten vor, um den Anführer zu trainieren, und einen Action-Abstraction-Policy-Gradienten, um die Follower im Anführer-Follower-Markov-Spiel zu trainieren."]} +{"source": "Recent work has studied the emergence of language among deep reinforcement learning agents that must collaborate to solve a task.Of particular interest are the factors that cause language to be compositional---i.e., express meaning by combining words which themselves have meaning.Evolutionary linguists have found that in addition to structural priors like those already studied in deep learning, the dynamics of transmitting language from generation to generation contribute significantly to the emergence of compositionality.In this paper, we introduce these cultural evolutionary dynamics into language emergence by periodically replacing agents in a population to create a knowledge gap, implicitly inducing cultural transmission of language.We show that this implicit cultural transmission encourages the resulting languages to exhibit better compositional generalization.", "target": ["Wir nutzen kulturelle Übertragung, um Kompositionalität in Sprachen zu fördern, die aus der Interaktion zwischen neuronalen Agenten entstehen."]} +{"source": "Based on our observation that there exists a dramatic drop for the singular values of the fully connected layers or a single feature map of the convolutional layer, and that the dimension of the concatenated feature vector almost equals the summation of the dimension on each feature map, we propose a singular value decomposition (SVD) based approach to estimate the dimension of the deep manifolds for a typical convolutional neural network VGG19.We choose three categories from the ImageNet, namely Persian Cat, Container Ship and Volcano, and determine the local dimension of the deep manifolds of the deep layers through the tangent space of a target image.Through several augmentation methods, we found that the Gaussian noise method is closer to the intrinsic dimension, as by adding random noise to an image we are moving in an arbitrary dimension, and when the rank of the feature matrix of the augmented images does not increase we are very closeto the local dimension of the manifold.We also estimate the dimension of the deep manifold based on the tangent space for each of the maxpooling layers.Our results show that the dimensions of different categories are close to each other and decline quickly along the convolutional layers and fully connected layers.Furthermore, we show that the dimensions decline quickly inside the Conv5 layer.Our work provides new insights for the intrinsic structure of deep neural networks and helps unveiling the inner organization of the black box of deep neural networks.", "target": ["Wir schlagen eine SVD-basierte Methode vor, um die lokale Dimension der Aktivierungsverteiler in tiefen neuronalen Netzen zu untersuchen."]} +{"source": "Large pre-trained Transformers such as BERT have been tremendously effective for many NLP tasks. However, inference in these large-capacity models is prohibitively slow and expensive. Transformers are essentially a stack of self-attention layers which encode each input position using the entire input sequence as its context. However, we find that it may not be necessary to apply this expensive sequence-wide self-attention over at all layers. Based on this observation, we propose a decomposition to a pre-trained Transformer that allows the lower layers to process segments of the input independently enabling parallelism and caching. We show that the information loss due to this decomposition can be recovered in the upper layers with auxiliary supervision during fine-tuning. We evaluate de-composition with pre-trained BERT models on five different paired-input tasks in question answering, sentence similarity, and natural language inference. Results show that decomposition enables faster inference (up to 4x), significant memory reduction (up to 70%) while retaining most (up to 99%) of the original performance. We will release the code at.", "target": ["Die Inferenz in großen Transformatoren ist aufgrund der Self-Attention in mehreren Schichten teuer. Wir zeigen, dass eine einfache Zerlegungstechnik zu einem schnelleren Modell mit geringem Speicherbedarf führen kann, das genauso genau ist wie die ursprünglichen Modelle."]} +{"source": "Exploration while learning representations is one of the main challenges DeepReinforcement Learning (DRL) faces today.As the learned representation is dependant in the observed data, the exploration strategy has a crucial role.The popular DQN algorithm has improved significantly the capabilities of ReinforcementLearning (RL) algorithms to learn state representations from raw data, yet, it usesa naive exploration strategy which is statistically inefficient.The RandomizedLeast Squares Value Iteration (RLSVI) algorithm (Osband et al., 2016), on theother hand, explores and generalizes efficiently via linearly parameterized valuefunctions.However, it is based on hand-designed state representation that requiresprior engineering work for every environment.In this paper, we propose a DeepLearning adaptation for RLSVI.Rather than using hand-design state representation, we use a state representation that is being learned directly from the data by aDQN agent.As the representation is being optimized during the learning process,a key component for the suggested method is a likelihood matching mechanism,which adapts to the changing representations.We demonstrate the importance ofthe various properties of our algorithm on a toy problem and show that our methodoutperforms DQN in five Atari benchmarks, reaching competitive results with theRainbow algorithm.", "target": ["Eine Deep Learning-Anpassung der Randomized Least Squares Value Iteration"]} +{"source": "The complexity of large-scale neural networks can lead to poor understanding of their internal details.We show that this opaqueness provides an opportunity for adversaries to embed unintended functionalities into the network in the form of Trojan horse attacks.Our novel framework hides the existence of a malicious network within a benign transport network.Our attack is flexible, easy to execute, and difficult to detect.We prove theoretically that the malicious network's detection is computationally infeasible and demonstrate empirically that the transport network does not compromise its disguise.Our attack exposes an important, previously unknown loophole that unveils a new direction in machine learning security.", "target": ["Die Parameter eines trainierten neuronalen Netzwerks können vertauscht werden, um ein völlig separates Modell für eine andere Aufgabe zu erstellen, was die Einbettung von trojanischen Netwerken in ein anderes Netz ermöglicht."]} +{"source": "In this paper, we introduce Random Path Generative Adversarial Network (RPGAN) --- an alternative scheme of GANs that can serve as a tool for generative model analysis.While the latent space of a typical GAN consists of input vectors, randomly sampled from the standard Gaussian distribution, the latent space of RPGAN consists of random paths in a generator network.As we show, this design allows to associate different layers of the generator with different regions of the latent space, providing their natural interpretability.With experiments on standard benchmarks, we demonstrate that RPGAN reveals several interesting insights about roles that different layers play in the image generation process.Aside from interpretability, the RPGAN model also provides competitive generation quality and allows efficient incremental learning on new data.", "target": ["Wir stellen ein alternatives GAN-Design vor, das auf zufälligen Routen im Generator basiert und als Werkzeug für die Interpretierbarkeit generativer Modelle dienen kann."]} +{"source": "Deep artificial neural networks can achieve an extremely small difference between training and test accuracies on identically distributed training and test sets, which is a standard measure of generalization.However, the training and test sets may not be sufficiently representative of the empirical sample set, which consists of real-world input samples.When samples are drawn from an underrepresented or unrepresented subset during inference, the gap between the training and inference accuracies can be significant.To address this problem, we first reformulate a classification algorithm as a procedure for searching for a source code that maps input features to classes.We then derive a necessary and sufficient condition for generalization using a universal cognitive similarity metric, namely information distance, based on Kolmogorov complexity.Using this condition, we formulate an optimization problem to learn a more general classification function.To achieve this end, we extend the input features by concatenating encodings of them, and then train the classifier on the extended features.As an illustration of this idea, we focus on image classification, where we use channel codes on the input features as a systematic way to improve the degree to which the training and test sets are representative of the empirical sample set.To showcase our theoretical findings, considering that corrupted or perturbed input features belong to the empirical sample set, but typically not to the training and test sets, we demonstrate through extensive systematic experiments that, as a result of learning a more general classification function, a model trained on encoded input features is significantly more robust to common corruptions, e.g., Gaussian and shot noise, as well as adversarial perturbations, e.g., those found via projected gradient descent, than the model trained on uncoded input features.", "target": ["Wir präsentieren einen theoretischen und experimentellen Rahmen für die Definition, das Verständnis und das Erreichen von Generalisierung und damit von Robustheit beim Deep Learning, indem wir auf die algorithmische Informationstheorie und die Code-Theorie zurückgreifen."]} +{"source": "Many approaches to causal discovery are limited by their inability to discriminate between Markov equivalent graphs given only observational data.We formulate causal discovery as a marginal likelihood based Bayesian model selection problem.We adopt a parameterization based on the notion of the independence of causal mechanisms which renders Markov equivalent graphs distinguishable.We complement this with an empirical Bayesian approach to setting priors so that the actual underlying causal graph is assigned a higher marginal likelihood than its alternatives.Adopting a Bayesian approach also allows for straightforward modeling of unobserved confounding variables, for which we provide a variational algorithm to approximate the marginal likelihood, since this desirable feat renders the computation of the marginal likelihood intractable.We believe that the Bayesian approach to causal discovery both allows the rich methodology of Bayesian inference to be used in various difficult aspects of this problem and provides a unifying framework to causal discovery research.We demonstrate promising results in experiments conducted on real data, supporting our modeling approach and our inference methodology.", "target": ["Wir stellen die Entdeckung kausaler Strukturen als eine Bayes'sche Modellauswahl dar, die es uns ermöglicht, zwischen Markov-äquivalenten Graphen zu unterscheiden, um den einzigartigen kausalen Graphen zu identifizieren."]} +{"source": "The goal of compressed sensing is to learn a structured signal $x$ from a limited number of noisy linear measurements $y\\approx Ax$. In traditional compressed sensing, ``structure'' is represented by sparsity in some known basis. Inspired by the success of deep learning in modeling images, recent work starting with~\\cite{BDJP17} has instead considered structure to come from a generative model $G: \\R^k \\to \\R^n$. We present two results establishing the difficulty of this latter task, showing that existing bounds are tight. First, we provide a lower bound matching the~\\cite{BDJP17} upper bound for compressed sensing from $L$-Lipschitz generative models $G$. In particular, there exists such a function that requires roughly $\\Omega(k \\log L)$ linear measurements for sparse recovery to be possible. This holds even for the more relaxed goal of \\emph{nonuniform} recovery. Second, we show that generative models generalize sparsity as a representation of structure. In particular, we construct a ReLU-based neural network $G: \\R^{2k} \\to \\R^n$ with $O(1)$ layers and $O(kn)$ activations per layer, such that the range of $G$ contains all $k$-sparse vectors.", "target": ["Untere Schranke für Compressed Sensing mit generativen Modellen, die den bekannten oberen Schranken entspricht"]} +{"source": "We hypothesize that end-to-end neural image captioning systems work seemingly well because they exploit and learn ‘distributional similarity’ in a multimodal feature space, by mapping a test image to similar training images in this space and generating a caption from the same space.To validate our hypothesis, we focus on the ‘image’ side of image captioning, and vary the input image representation but keep the RNN text generation model of a CNN-RNN constant.We propose a sparse bag-of-objects vector as an interpretable representation to investigate our distributional similarity hypothesis.We found that image captioning models(i) are capable of separating structure from noisy input representations;(ii) experience virtually no significant performance loss when a high dimensional representation is compressed to a lower dimensional space;(iii) cluster images with similar visual and linguistic information together;(iv) are heavily reliant on test sets with a similar distribution as the training set;(v) repeatedly generate the same captions by matching images and ‘retrieving’ a caption in the joint visual-textual space.Our experiments all point to one fact: that our distributional similarity hypothesis holds.We conclude that, regardless of the image representation, image captioning systems seem to match images and generate captions in a learned joint image-text semantic subspace.", "target": ["In diesem Beitrag wird eine empirische Analyse der Rolle verschiedener Arten von Bildrepräsentationen vorgestellt und die Eigenschaften dieser Repräsentationen für die Aufgabe der Bildbeschriftung untersucht."]} +{"source": "We present a sequence-to-action parsing approach for the natural language to SQL task that incrementally fills the slots of a SQL query with feasible actions from a pre-defined inventory.To account for the fact that typically there are multiple correct SQL queries with the same or very similar semantics, we draw inspiration from syntactic parsing techniques and propose to train our sequence-to-action models with non-deterministic oracles.We evaluate our models on the WikiSQL dataset and achieve an execution accuracy of 83.7% on the test set, a 2.1% absolute improvement over the models trained with traditional static oracles assuming a single correct target SQL query.When further combined with the execution-guided decoding strategy, our model sets a new state-of-the-art performance at an execution accuracy of 87.1%.", "target": ["Wir entwerfen inkrementelle Sequenz-zu-Aktion-Parser für Text-zu-SQL-Aufgaben und erzielen SOTA-Ergebnisse. Wir verbessern sie weiter, indem wir nicht-deterministische Orakel verwenden, um mehrere korrekte Aktionssequenzen zu ermöglichen."]} +{"source": "We propose Efficient Neural Architecture Search (ENAS), a faster and less expensive approach to automated model design than previous methods.In ENAS, a controller learns to discover neural network architectures by searching for an optimal path within a larger model.The controller is trained with policy gradient to select a path that maximizes the expected reward on the validation set.Meanwhile the model corresponding to the selected path is trained to minimize the cross entropy loss.On the Penn Treebank dataset, ENAS can discover a novel architecture thats achieves a test perplexity of 57.8, which is state-of-the-art among automatic model design methods on Penn Treebank.On the CIFAR-10 dataset, ENAS can design novel architectures that achieve a test error of 2.89%, close to the 2.65% achieved by standard NAS (Zoph et al., 2017).Most importantly, our experiments show that ENAS is more than 10x faster and 100x less resource-demanding than NAS.", "target": ["Ein Ansatz, der die Suche nach neuronalen Architekturen um das 10-fache beschleunigt und dabei 100-mal weniger Rechenressourcen benötigt."]} +{"source": "Nowadays, deep neural networks (DNNs) have become the main instrument for machine learning tasks within a wide range of domains, including vision, NLP, and speech.Meanwhile, in an important case of heterogenous tabular data, the advantage of DNNs over shallow counterparts remains questionable.In particular, there is no sufficient evidence that deep learning machinery allows constructing methods that outperform gradient boosting decision trees (GBDT), which are often the top choice for tabular problems.In this paper, we introduce Neural Oblivious Decision Ensembles (NODE), a new deep learning architecture, designed to work with any tabular data.In a nutshell, the proposed NODE architecture generalizes ensembles of oblivious decision trees, but benefits from both end-to-end gradient-based optimization and the power of multi-layer hierarchical representation learning.With an extensive experimental comparison to the leading GBDT packages on a large number of tabular datasets, we demonstrate the advantage of the proposed NODE architecture, which outperforms the competitors on most of the tasks.We open-source the PyTorch implementation of NODE and believe that it will become a universal framework for machine learning on tabular data.", "target": ["Wir schlagen eine neue DNN-Architektur für Deep Learning auf tabellarischen Daten vor"]} +{"source": "Person re-identification (re-ID) aims at identifying the same persons' images across different cameras.However, domain diversities between different datasets pose an evident challenge for adapting the re-ID model trained on one dataset to another one.State-of-the-art unsupervised domain adaptation methods for person re-ID transferred the learned knowledge from the source domain by optimizing with pseudo labels created by clustering algorithms on the target domain.Although they achieved state-of-the-art performances, the inevitable label noise caused by the clustering procedure was ignored.Such noisy pseudo labels substantially hinders the model's capability on further improving feature representations on the target domain.In order to mitigate the effects of noisy pseudo labels, we propose to softly refine the pseudo labels in the target domain by proposing an unsupervised framework, Mutual Mean-Teaching (MMT), to learn better features from the target domain via off-line refined hard pseudo labels and on-line refined soft pseudo labels in an alternative training manner. In addition, the common practice is to adopt both the classification loss and the triplet loss jointly for achieving optimal performances in person re-ID models.However, conventional triplet loss cannot work with softly refined labels.To solve this problem, a novel soft softmax-triplet loss is proposed to support learning with soft pseudo triplet labels for achieving the optimal domain adaptation performance.The proposed MMT framework achieves considerable improvements of 14.4%, 18.2%, 13.1% and 16.4% mAP on Market-to-Duke, Duke-to-Market, Market-to-MSMT and Duke-to-MSMT unsupervised domain adaptation tasks.", "target": ["Ein Rahmenwerk, das die Online-Verfeinerung von Pseudo-Etiketten mit einem neuartigen Softmax-Triplet-Verlust für die unüberwachte Domänenanpassung bei der Re-Identifizierung von Personen durchführt."]} +{"source": "We present the first end-to-end verifier of audio classifiers.Compared to existing methods, our approach enables analysis of both, the entire audio processing stage as well as recurrent neural network architectures (e.g., LSTM).The audio processing is verified using novel convex relaxations tailored to feature extraction operations used in audio (e.g., Fast Fourier Transform) while recurrent architectures are certified via a novel binary relaxation for the recurrent unit update.We show the verifier scales to large networks while computing significantly tighter bounds than existing methods for common audio classification benchmarks: on the challenging Google Speech Commands dataset we certify 95% more inputs than the interval approximation (only prior scalable method), for a perturbation of -90dB.", "target": ["Wir stellen den ersten Ansatz vor, um die Robustheit neuronaler Netze gegen geräuschbasierte Störungen im Audiobereich zu zertifizieren."]} +{"source": "Since deep neural networks are over-parameterized, they can memorize noisy examples.We address such memorizing issue in the presence of annotation noise.From the fact that deep neural networks cannot generalize neighborhoods of the features acquired via memorization, we hypothesize that noisy examples do not consistently incur small losses on the network under a certain perturbation.Based on this, we propose a novel training method called Learning with Ensemble Consensus (LEC) that prevents overfitting noisy examples by eliminating them using the consensus of an ensemble of perturbed networks.One of the proposed LECs, LTEC outperforms the current state-of-the-art methods on noisy MNIST, CIFAR-10, and CIFAR-100 in an efficient manner.", "target": ["In dieser Arbeit wird eine Methode zur Erzeugung und Verwendung von Ensembles vorgestellt, um verrauschte Beispiele in Gegenwart von Annotationsrauschen effektiv zu identifizieren."]} +{"source": "Recovering sparse conditional independence graphs from data is a fundamental problem in machine learning with wide applications.A popular formulation of the problem is an $\\ell_1$ regularized maximum likelihood estimation.Many convex optimization algorithms have been designed to solve this formulation to recover the graph structure.Recently, there is a surge of interest to learn algorithms directly based on data, and in this case, learn to map empirical covariance to the sparse precision matrix.However, it is a challenging task in this case, since the symmetric positive definiteness (SPD) and sparsity of the matrix are not easy to enforce in learned algorithms, and a direct mapping from data to precision matrix may contain many parameters.We propose a deep learning architecture, GLAD, which uses an Alternating Minimization (AM) algorithm as our model inductive bias, and learns the model parameters via supervised learning.We show that GLAD learns a very compact and effective model for recovering sparse graphs from data.", "target": ["Ein datengesteuerter Lernalgorithmus, der auf der Abrollung der Alternierenden Minimierungsoptimierung für die Wiederherstellung spärlicher Graphen basiert."]} +{"source": "Counterfactual regret minimization (CFR) is a fundamental and effective technique for solving Imperfect Information Games (IIG).However, the original CFR algorithm only works for discrete states and action spaces, and the resulting strategy is maintained as a tabular representation.Such tabular representation limits the method from being directly applied to large games.In this paper, we propose a double neural representation for the IIGs, where one neural network represents the cumulative regret, and the other represents the average strategy. Such neural representations allow us to avoid manual game abstraction and carry out end-to-end optimization.To make the learning efficient, we also developed several novel techniques including a robust sampling method and a mini-batch Monte Carlo Counterfactual Regret Minimization (MCCFR) method, which may be of independent interests. Empirically, on games tractable to tabular approaches, neural strategies trained with our algorithm converge comparably to their tabular counterparts, and significantly outperform those based on deep reinforcement learning. On extremely large games with billions of decision nodes, our approach achieved strong performance while using hundreds of times less memory than the tabular CFR.On head-to-head matches of hands-up no-limit texas hold'em, our neural agent beat the strong agent ABS-CFR by $9.8\\pm4.1$ chips per game.It's a successful application of neural CFR in large games.", "target": ["Wir haben einen doppelten neuronalen Rahmen vorgeschlagen, um ein groß angelegtes unvollkommene Informationen Spiel zu lösen."]} +{"source": "We present the first verification that a neural network for perception tasks producesa correct output within a specified tolerance for every input of interest.We definecorrectness relative to a specification which identifies 1) a state space consisting ofall relevant states of the world and 2) an observation process that produces neuralnetwork inputs from the states of the world.Tiling the state and input spaces witha finite number of tiles, obtaining ground truth bounds from the state tiles andnetwork output bounds from the input tiles, then comparing the ground truth andnetwork output bounds delivers an upper bound on the network output error forany input of interest.Results from two case studies highlight the ability of ourtechnique to deliver tight error bounds for all inputs of interest and show how theerror bounds vary over the state and input spaces.", "target": ["Wir präsentieren den ersten Nachweis, dass ein neuronales Netz für Wahrnehmungsaufgaben für jede Eingabe von Interesse eine korrekte Ausgabe innerhalb einer bestimmten Toleranz erzeugt."]} +{"source": "Deep generative models have achieved remarkable progress in recent years.Despite this progress, quantitative evaluation and comparison of generative models remains as one of the important challenges.One of the most popular metrics for evaluating generative models is the log-likelihood.While the direct computation of log-likelihood can be intractable, it has been recently shown that the log-likelihood of some of the most interesting generative models such as variational autoencoders (VAE) or generative adversarial networks (GAN) can be efficiently estimated using annealed importance sampling (AIS).In this work, we argue that the log-likelihood metric by itself cannot represent all the different performance characteristics of generative models, and propose to use rate distortion curves to evaluate and compare deep generative models.We show that we can approximate the entire rate distortion curve using one single run of AIS for roughly the same computational cost as a single log-likelihood estimate.We evaluate lossy compression rates of different deep generative models such as VAEs, GANs (and its variants) and adversarial autoencoders (AAE) on MNIST and CIFAR10, and arrive at a number of insights not obtainable from log-likelihoods alone.", "target": ["Wir untersuchen Annäherungen an die Ratenverzerrung, um tiefe generative Modelle zu evaluieren, und zeigen, dass Ratenverzerrungskurven mehr Erkenntnisse über das Modell liefern als die Log-Likelihood allein, während sie ungefähr die gleichen Berechnungskosten erfordern."]} +{"source": "Although reinforcement learning methods can achieve impressive results in simulation, the real world presents two major challenges: generating samples is exceedingly expensive, and unexpected perturbations or unseen situations cause proficient but specialized policies to fail at test time.Given that it is impractical to train separate policies to accommodate all situations the agent may see in the real world, this work proposes to learn how to quickly and effectively adapt online to new tasks.To enable sample-efficient learning, we consider learning online adaptation in the context of model-based reinforcement learning.Our approach uses meta-learning to train a dynamics model prior such that, when combined with recent data, this prior can be rapidly adapted to the local context.Our experiments demonstrate online adaptation for continuous control tasks on both simulated and real-world agents.We first show simulated agents adapting their behavior online to novel terrains, crippled body parts, and highly-dynamic environments.We also illustrate the importance of incorporating online adaptation into autonomous agents that operate in the real world by applying our method to a real dynamic legged millirobot: We demonstrate the agent's learned ability to quickly adapt online to a missing leg, adjust to novel terrains and slopes, account for miscalibration or errors in pose estimation, and compensate for pulling payloads.", "target": ["Ein modellbasierter Meta-RL-Algorithmus, der es einem realen Roboter ermöglicht, sich in dynamischen Umgebungen online anzupassen"]} +{"source": "Model-free deep reinforcement learning approaches have shown superhuman performance in simulated environments (e.g., Atari games, Go, etc).During training, these approaches often implicitly construct a latent space that contains key information for decision making.In this paper, we learn a forward model on this latent space and apply it to model-based planning in miniature Real-time Strategy game with incomplete information (MiniRTS).We first show that the latent space constructed from existing actor-critic models contains relevant information of the game, and design training procedure to learn forward models.We also show that our learned forward model can predict meaningful future state and is usable for latent space Monte-Carlo Tree Search (MCTS), in terms of win rates against rule-based agents.", "target": ["Das Papier analysiert den latenten Raum, der durch modellfreie Ansätze in einem Miniaturspiel mit unvollständiger Information erlernt wird, trainiert ein Vorwärtsmodell im latenten Raum und wendet es auf die Monte-Carlo-Baumsuche an, was zu einer positiven Leistung führt."]} +{"source": "Several state of the art convolutional networks rely on inter-connecting different layers to ease the flow of information and gradient between their input and output layers.These techniques have enabled practitioners to successfully train deep convolutional networks with hundreds of layers.Particularly, a novel way of interconnecting layers was introduced as the Dense Convolutional Network (DenseNet) and has achieved state of the art performance on relevant image recognition tasks.Despite their notable empirical success, their theoretical understanding is still limited.In this work, we address this problem by analyzing the effect of layer interconnection on the overall expressive power of a convolutional network.In particular, the connections used in DenseNet are compared with other types of inter-layer connectivity.We carry out a tensor analysis on the expressive power inter-connections on convolutional arithmetic circuits (ConvACs) and relate our results to standard convolutional networks.The analysis leads to performance bounds and practical guidelines for design of ConvACs.The generalization of these results are discussed for other kinds of convolutional networks via generalized tensor decompositions.", "target": ["Wir analysieren die Ausdruckskraft der in DenseNets verwendeten Verbindungen über Tensorzerlegungen."]} +{"source": "We consider the following central question in the field of Deep Reinforcement Learning (DRL):How can we use implicit human feedback to accelerate and optimize the training of a DRL algorithm?State-of-the-art methods rely on any human feedback to be provided explicitly, requiring the active participation of humans (e.g., expert labeling, demonstrations, etc.).In this work, we investigate an alternative paradigm, where non-expert humans are silently observing (and assessing) the agent interacting with the environment.The human's intrinsic reactions to the agent's behavior is sensed as implicit feedback by placing electrodes on the human scalp and monitoring what are known as event-related electric potentials.The implicit feedback is then used to augment the agent's learning in the RL tasks.We develop a system to obtain and accurately decode the implicit human feedback (specifically error-related event potentials) for state-action pairs in an Atari-type environment.As a baseline contribution, we demonstrate the feasibility of capturing error-potentials of a human observer watching an agent learning to play several different Atari-games using an electroencephalogram (EEG) cap, and then decoding the signals appropriately and using them as an auxiliary reward function to a DRL algorithm with the intent of accelerating its learning of the game.Building atop the baseline, we then make the following novel contributions in our work:(i) We argue that the definition of error-potentials is generalizable across different environments; specifically we show that error-potentials of an observer can be learned for a specific game, and the definition used as-is for another game without requiring re-learning of the error-potentials. (ii) We propose two different frameworks to combine recent advances in DRL into the error-potential based feedback system in a sample-efficient manner, allowing humans to provide implicit feedback while training in the loop, or prior to the training of the RL agent.(iii) Finally, we scale the implicit human feedback (via ErrP) based RL to reasonably complex environments (games) and demonstrate the significance of our approach through synthetic and real user experiments.", "target": ["Wir nutzen implizites menschliches Feedback (über Fehlerpotentiale, EEG), um das Training eines DRL-Algorithmus auf praktische Weise zu beschleunigen und zu optimieren."]} +{"source": "Deep learning has demonstrated abilities to learn complex structures, but they can be restricted by available data.Recently, Consensus Networks (CNs) were proposed to alleviate data sparsity by utilizing features from multiple modalities, but they too have been limited by the size of labeled data.In this paper, we extend CN to Transductive Consensus Networks (TCNs), suitable for semi-supervised learning.In TCNs, different modalities of input are compressed into latent representations, which we encourage to become indistinguishable during iterative adversarial training.To understand TCNs two mechanisms, consensus and classification, we put forward its three variants in ablation studies on these mechanisms.To further investigate TCN models, we treat the latent representations as probability distributions and measure their similarities as the negative relative Jensen-Shannon divergences.We show that a consensus state beneficial for classification desires a stable but imperfect similarity between the representations.Overall, TCNs outperform or align with the best benchmark algorithms given 20 to 200 labeled samples on the Bank Marketing and the DementiaBank datasets.", "target": ["TCN für multimodales halb-überwachtes Lernen + Ablationsstudie seiner Mechanismen + Interpretationen latenter Repräsentationen"]} +{"source": "Several first order stochastic optimization methods commonly used in the Euclidean domain such as stochastic gradient descent (SGD), accelerated gradient descent or variance reduced methods have already been adapted to certain Riemannian settings.However, some of the most popular of these optimization tools - namely Adam, Adagrad and the more recent Amsgrad - remain to be generalized to Riemannian manifolds.We discuss the difficulty of generalizing such adaptive schemes to the most agnostic Riemannian setting, and then provide algorithms and convergence proofs for geodesically convex objectives in the particular case of a product of Riemannian manifolds, in which adaptivity is implemented across manifolds in the cartesian product.Our generalization is tight in the sense that choosing the Euclidean space as Riemannian manifold yields the same algorithms and regret bounds as those that were already known for the standard algorithms.Experimentally, we show faster convergence and to a lower train loss value for Riemannian adaptive methods over their corresponding baselines on the realistic task of embedding the WordNet taxonomy in the Poincare ball.", "target": ["Anpassung von Adam, Amsgrad, Adagrad an Riemannsche Mannigfaltigkeiten."]} +{"source": "We study the problem of defending deep neural network approaches for image classification from physically realizable attacks.First, we demonstrate that the two most scalable and effective methods for learning robust models, adversarial training with PGD attacks and randomized smoothing, exhibit very limited effectiveness against three of the highest profile physical attacks.Next, we propose a new abstract adversarial model, rectangular occlusion attacks, in which an adversary places a small adversarially crafted rectangle in an image, and develop two approaches for efficiently computing the resulting adversarial examples.Finally, we demonstrate that adversarial training using our new attack yields image classification models that exhibit high robustness against the physically realizable attacks we study, offering the first effective generic defense against such attacks.", "target": ["Verteidigung gegen physikalisch realisierbare Angriffe auf die Bildklassifizierung"]} +{"source": "Continual learning is the problem of sequentially learning new tasks or knowledge while protecting previously acquired knowledge.However, catastrophic forgetting poses a grand challenge for neural networks performing such learning process.Thus, neural networks that are deployed in the real world often struggle in scenarios where the data distribution is non-stationary (concept drift), imbalanced, or not always fully available, i.e., rare edge cases.We propose a Differentiable Hebbian Consolidation model which is composed of a Differentiable Hebbian Plasticity (DHP) Softmax layer that adds a rapid learning plastic component (compressed episodic memory) to the fixed (slow changing) parameters of the softmax output layer; enabling learned representations to be retained for a longer timescale.We demonstrate the flexibility of our method by integrating well-known task-specific synaptic consolidation methods to penalize changes in the slow weights that are important for each target task.We evaluate our approach on the Permuted MNIST, Split MNIST and Vision Datasets Mixture benchmarks, and introduce an imbalanced variant of Permuted MNIST --- a dataset that combines the challenges of class imbalance and concept drift.Our proposed model requires no additional hyperparameters and outperforms comparable baselines by reducing forgetting.", "target": ["Hebbsche plastische Gewichte können sich wie ein komprimierter episodischer Gedächtnisspeicher in neuronalen Netzwerken verhalten und in Kombination mit aufgabenspezifischer synaptischer Konsolidierung die Fähigkeit verbessern, katastrophales Vergessen beim kontinuierlichen Lernen zu lindern."]} +{"source": "In order to choose a neural network architecture that will be effective for a particular modeling problem, one must understand the limitations imposed by each of the potential options.These limitations are typically described in terms of information theoretic bounds, or by comparing the relative complexity needed to approximate example functions between different architectures.In this paper, we examine the topological constraints that the architecture of a neural network imposes on the level sets of all the functions that it is able to approximate.This approach is novel for both the nature of the limitations and the fact that they are independent of network depth for a broad family of activation functions.", "target": ["Diese Arbeit beweist, dass dünne neuronale Netze bestimmte Funktionen nicht annähern können, unabhängig davon, wie tief sie sind."]} +{"source": "Program verification offers a framework for ensuring program correctness and therefore systematically eliminating different classes of bugs.Inferring loop invariants is one of the main challenges behind automated verification of real-world programs which often contain many loops.In this paper, we present Continuous Logic Network (CLN), a novel neural architecture for automatically learning loop invariants directly from program execution traces.Unlike existing neural networks, CLNs can learn precise and explicit representations of formulas in Satisfiability Modulo Theories (SMT) for loop invariants from program execution traces.We develop a new sound and complete semantic mapping for assigning SMT formulas to continuous truth values that allows CLNs to be trained efficiently.We use CLNs to implement a new inference system for loop invariants, CLN2INV, that significantly outperforms existing approaches on the popular Code2Inv dataset.CLN2INV is the first tool to solve all 124 theoretically solvable problems in the Code2Inv dataset.Moreover, CLN2INV takes only 1.1 second on average for each problem, which is 40 times faster than existing approaches.We further demonstrate that CLN2INV can even learn 12 significantly more complex loop invariants than the ones required for the Code2Inv dataset.", "target": ["Wir stellen das Continuous Logic Network (CLN) vor, eine neuartige neuronale Architektur zum automatischen Lernen von Schleifeninvarianten und allgemeinen SMT-Formeln."]} +{"source": "Single cell RNA sequencing (scRNAseq) technology enables quantifying gene expression profiles by individual cells within cancer.Dimension reduction methods have been commonly used for cell clustering analysis and visualization of the data.Current dimension reduction methods tend overly eliminate the expression variations correspond to less dominating characteristics, such we fail to find the homogenious properties of cancer development.In this paper, we proposed a new and clustering analysis method for scRNAseq data, namely BBSC, via implementing a binarization of the gene expression profile into on/off frequency changes with a Boolean matrix factorization.The low rank representation of expression matrix recovered by BBSC increase the resolution in identifying distinct cell types or functions.Application of BBSC on two cancer scRNAseq data successfully discovered both homogeneous and heterogeneous cancer cell clusters.Further finding showed potential in preventing cancer progression.", "target": ["Unsere Ergebnisse werfen ein Licht auf die Verhinderung des Fortschreitens von Krebs"]} +{"source": "While normalizing flows have led to significant advances in modeling high-dimensional continuous distributions, their applicability to discrete distributions remains unknown.In this paper, we show that flows can in fact be extended to discrete events---and under a simple change-of-variables formula not requiring log-determinant-Jacobian computations.Discrete flows have numerous applications.We display proofs of concept under 2 flow architectures: discrete autoregressive flows enable bidirectionality, allowing for example tokens in text to depend on both left-to-right and right-to-left contexts in an exact language model; and discrete bipartite flows (i.e., with layer structure from RealNVP) enable parallel generation such as exact nonautoregressive text modeling.", "target": ["Wir erweitern autoregressive Flüsse und RealNVP auf diskrete Daten."]} +{"source": "We present a Deep Neural Network with Spike Assisted Feature Extraction (SAFE-DNN) to improve robustness of classification under stochastic perturbation of inputs.The proposed network augments a DNN with unsupervised learning of low-level features using spiking neuron network (SNN) with Spike-Time-Dependent-Plasticity (STDP).The complete network learns to ignore local perturbation while performing global feature detection and classification.The experimental results on CIFAR-10 and ImageNet subset demonstrate improved noise robustness for multiple DNN architectures without sacrificing accuracy on clean images.", "target": ["Eine geräuschresistente Deep-Learning-Architektur."]} +{"source": "Neural embeddings have been used with great success in Natural Language Processing (NLP) where they provide compact representations that encapsulate word similarity and attain state-of-the-art performance in a range of linguistic tasks.The success of neural embeddings has prompted significant amounts of research into applications in domains other than language.One such domain is graph-structured data, where embeddings of vertices can be learned that encapsulate vertex similarity and improve performance on tasks including edge prediction and vertex labelling.For both NLP and graph-based tasks, embeddings in high-dimensional Euclidean spaces have been learned.However, recent work has shown that the appropriate isometric space for embedding complex networks is not the flat Euclidean space, but a negatively curved hyperbolic space.We present a new concept that exploits these recent insights and propose learning neural embeddings of graphs in hyperbolic space.We provide experimental evidence that hyperbolic embeddings significantly outperform Euclidean embeddings on vertex classification tasks for several real-world public datasets.", "target": ["Wir lernen neuronale Einbettungen von Graphen im hyperbolischen statt im euklidischen Raum"]} +{"source": "The International Competition on Knowledge Engineering for Planning and Scheduling (ICKEPS) plays a pivotal role in fostering the development of new Knowledge Engineering (KE) tools, and in emphasising the importance of principled approaches for all the different KE aspects that are needed for the successful long-term use of planning in real-world applications. In this paper, as an exercise in synthesis and for the sake of stimulating thoughts and discussion, we review the format of previous ICKEPS, to suggest alternative formats for future competitions, ideally to motivate someone to step up and organise the next ones.", "target": ["Ideen für zukünftige ICKEPS"]} +{"source": "We show that generating English Wikipedia articles can be approached as a multi-document summarization of source documents.We use extractive summarizationto coarsely identify salient information and a neural abstractive model to generatethe article.For the abstractive model, we introduce a decoder-only architecturethat can scalably attend to very long sequences, much longer than typical encoder-decoder architectures used in sequence transduction.We show that this model cangenerate fluent, coherent multi-sentence paragraphs and even whole Wikipediaarticles.When given reference documents, we show it can extract relevant factualinformation as reflected in perplexity, ROUGE scores and human evaluations.", "target": ["Wir generieren Wikipedia-Artikel abstrakt auf der Grundlage von Texten aus Quellendokumenten."]} +{"source": "Abstract Stochastic gradient descent (SGD) and Adam are commonly used to optimize deep neural networks, but choosing one usually means making tradeoffs between speed, accuracy and stability.Here we present an intuition for why the tradeoffs exist as well as a method for unifying the two in a continuous way.This makes it possible to control the way models are trained in much greater detail.We show that for default parameters, the new algorithm equals or outperforms SGD and Adam across a range of models for image classification tasks and outperforms SGD for language modeling tasks.", "target": ["Ein Algorithmus zur Vereinheitlichung von SGD und Adam und empirische Untersuchung seiner Leistung"]} +{"source": "The use of imitation learning to learn a single policy for a complex task that has multiple modes or hierarchical structure can be challenging.In fact, previous work has shown that when the modes are known, learning separate policies for each mode or sub-task can greatly improve the performance of imitation learning.In this work, we discover the interaction between sub-tasks from their resulting state-action trajectory sequences using a directed graphical model.We propose a new algorithm based on the generative adversarial imitation learning framework which automatically learns sub-task policies from unsegmented demonstrations.Our approach maximizes the directed information flow in the graphical model between sub-task latent variables and their generated trajectories.We also show how our approach connects with the existing Options framework, which is commonly used to learn hierarchical policies.", "target": ["Lernen hierarchischer Strategien aus unsegmentierten Demonstrationen unter Verwendung gerichteter Informationen"]} +{"source": "The Convolutional Neural Network (CNN) has been successfully applied in many fields during recent decades; however it lacks the ability to utilize prior domain knowledge when dealing with many realistic problems.We present a framework called Geometric Operator Convolutional Neural Network (GO-CNN) that uses domain knowledge, wherein the kernel of the first convolutional layer is replaced with a kernel generated by a geometric operator function.This framework integrates many conventional geometric operators, which allows it to adapt to a diverse range of problems.Under certain conditions, we theoretically analyze the convergence and the bound of the generalization errors between GO-CNNs and common CNNs.Although the geometric operator convolution kernels have fewer trainable parameters than common convolution kernels, the experimental results indicate that GO-CNN performs more accurately than common CNN on CIFAR-10/100.Furthermore, GO-CNN reduces dependence on the amount of training examples and enhances adversarial stability.", "target": ["Traditionelle Bildverarbeitungsalgorithmen werden mit Convolutional Neural Networks, einem neuen neuronalen Netz, kombiniert."]} +{"source": "Determinantal point processes (DPPs) is an effective tool to deliver diversity on multiple machine learning and computer vision tasks.Under deep learning framework, DPP is typically optimized via approximation, which is not straightforward and has some conflict with diversity requirement.We note, however, there has been no deep learning paradigms to optimize DPP directly since it involves matrix inversion which may result in highly computational instability.This fact greatly hinders the wide use of DPP on some specific objectives where DPP serves as a term to measure the feature diversity.In this paper, we devise a simple but effective algorithm to address this issue to optimize DPP term directly expressed with L-ensemble in spectral domain over gram matrix, which is more flexible than learning on parametric kernels.By further taking into account some geometric constraints, our algorithm seeks to generate valid sub-gradients of DPP term in case when the DPP gram matrix is not invertible (no gradients exist in this case).In this sense, our algorithm can be easily incorporated with multiple deep learning tasks.Experiments show the effectiveness of our algorithm, indicating promising performance for practical learning problems.", "target": ["Wir haben eine spezielle Backpropagation-Methode über einen geeigneten spektralen Subgradienten vorgeschlagen, um den deterministischen Punktprozess in den Deep Learning-Rahmen zu integrieren."]} +{"source": "The quality of a machine translation system depends largely on the availability of sizable parallel corpora.For the recently popular Neural Machine Translation (NMT) framework, data sparsity problem can become even more severe.With large amount of tunable parameters, the NMT model may overfit to the existing language pairs while failing to understand the general diversity in language.In this paper, we advocate to broadcast every sentence pair as two groups of similar sentences to incorporate more diversity in language expressions, which we name as parallel cluster.Then we define a more general cluster-to-cluster correspondence score and train our model to maximize this score.Since direct maximization is difficult, we derive its lower-bound as our surrogate objective, which is found to generalize point-point Maximum Likelihood Estimation (MLE) and point-to-cluster Reward Augmented Maximum Likelihood (RAML) algorithms as special cases.Based on this novel objective function, we delineate four potential systems to realize our cluster-to-cluster framework and test their performances in three recognized translation tasks, each task with forward and reverse translation directions.In each of the six experiments, our proposed four parallel systems have consistently proved to outperform the MLE baseline, RL (Reinforcement Learning) and RAML systems significantly.Finally, we have performed case study to empirically analyze the strength of the cluster-to-cluster NMT framework.", "target": ["Wir entwickeln einen neuartigen Cluster-to-Cluster-Rahmen für das NMT-Training, der die Vielfalt der Ausgangs- und Zielsprache besser verstehen kann."]} +{"source": "The capability of making interpretable and self-explanatory decisions is essential for developing responsible machine learning systems.In this work, we study the learning to explain the problem in the scope of inductive logic programming (ILP).We propose Neural Logic Inductive Learning (NLIL), an efficient differentiable ILP framework that learns first-order logic rules that can explain the patterns in the data.In experiments, compared with the state-of-the-art models, we find NLIL is able to search for rules that are x10 times longer while remaining x3 times faster.We also show that NLIL can scale to large image datasets, i.e. Visual Genome, with 1M entities.", "target": ["Ein effizientes differenzierbares ILP-Modell, das logische Regeln erster Ordnung erlernt, die die Daten erklären können."]} +{"source": "Neural network-based classifiers parallel or exceed human-level accuracy on many common tasks and are used in practical systems.Yet, neural networks are susceptible to adversarial examples, carefully perturbed inputs that cause networks to misbehave in arbitrarily chosen ways.When generated with standard methods, these examples do not consistently fool a classifier in the physical world due to a combination of viewpoint shifts, camera noise, and other natural transformations.Adversarial examples generated using standard techniques require complete control over direct input to the classifier, which is impossible in many real-world systems.We introduce the first method for constructing real-world 3D objects that consistently fool a neural network across a wide distribution of angles and viewpoints.We present a general-purpose algorithm for generating adversarial examples that are robust across any chosen distribution of transformations.We demonstrate its application in two dimensions, producing adversarial images that are robust to noise, distortion, and affine transformation.Finally, we apply the algorithm to produce arbitrary physical 3D-printed adversarial objects, demonstrating that our approach works end-to-end in the real world.Our results show that adversarial examples are a practical concern for real-world systems.", "target": ["Wir stellen eine neue Methode zur Synthese von Gegenbeispielen vor, die in der physischen Welt robust sind, und verwenden sie zur Herstellung der ersten 3D-Gegenobjekte."]} +{"source": "Representations of sets are challenging to learn because operations on sets should be permutation-invariant.To this end, we propose a Permutation-Optimisation module that learns how to permute a set end-to-end.The permuted set can be further processed to learn a permutation-invariant representation of that set, avoiding a bottleneck in traditional set models.We demonstrate our model's ability to learn permutations and set representations with either explicit or implicit supervision on four datasets, on which we achieve state-of-the-art results: number sorting, image mosaics, classification from image mosaics, and visual question answering.", "target": ["Lernen, wie man eine Menge permutiert und dann die permutierte Menge mit RNN kodiert, um eine Mengendarstellung zu erhalten."]} +{"source": "The physical design of a robot and the policy that controls its motion are inherently coupled.However, existing approaches largely ignore this coupling, instead choosing to alternate between separate design and control phases, which requires expert intuition throughout and risks convergence to suboptimal designs.In this work, we propose a method that jointly optimizes over the physical design of a robot and the corresponding control policy in a model-free fashion, without any need for expert supervision.Given an arbitrary robot morphology, our method maintains a distribution over the design parameters and uses reinforcement learning to train a neural network controller.Throughout training, we refine the robot distribution to maximize the expected reward.This results in an assignment to the robot parameters and neural network policy that are jointly optimal.We evaluate our approach in the context of legged locomotion, and demonstrate that it discovers novel robot designs and walking gaits for several different morphologies, achieving performance comparable to or better than that of hand-crafted designs.", "target": ["Verwenden Sie Deep Reinforcement Learning, um die physischen Attribute eines Roboters zusammen mit einer Steuerungsstrategie zu entwerfen."]} +{"source": "Deep learning enables training of large and flexible function approximators from scratch at the cost of large amounts of data.Applications of neural networks often consider learning in the context of a single task.However, in many scenarios what we hope to learn is not just a single task, but a model that can be used to solve multiple different tasks.Such multi-task learning settings have the potential to improve data efficiency and generalization by sharing data and representations across tasks.However, in some challenging multi-task learning settings, particularly in reinforcement learning, it is very difficult to learn a single model that can solve all the tasks while realizing data efficiency and performance benefits.Learning each of the tasks independently from scratch can actually perform better in such settings, but it does not benefit from the representation sharing that multi-task learning can potentially provide.In this work, we develop an approach that endows a single model with the ability to represent both extremes: joint training and independent training.To this end, we introduce matrix-interleaving (Mint), a modification to standard neural network models that projects the activations for each task into a different learned subspace, represented by a per-task and per-layer matrix.By learning these matrices jointly with the other model parameters, the optimizer itself can decide how much to share representations between tasks.On three challenging multi-task supervised learning and reinforcement learning problems with varying degrees of shared task structure, we find that this model consistently matches or outperforms joint training and independent training, combining the best elements of both.", "target": ["Wir schlagen einen Ansatz vor, der ein einziges Modell mit der Fähigkeit ausstattet, beide Extreme zu repräsentieren: gemeinsames Training und unabhängiges Training, was zu effektivem Multi-Task-Lernen führt."]} +{"source": "Training agents to operate in one environment often yields overfitted models that are unable to generalize to the changes in that environment.However, due to the numerous variations that can occur in the real-world, the agent is often required to be robust in order to be useful.This has not been the case for agents trained with reinforcement learning (RL) algorithms.In this paper, we investigate the overfitting of RL agents to the training environments in visual navigation tasks.Our experiments show that deep RL agents can overfit even when trained on multiple environments simultaneously. We propose a regularization method which combines RL with supervised learning methods by adding a term to the RL objective that would encourage the invariance of a policy to variations in the observations that ought not to affect the action taken.The results of this method, called invariance regularization, show an improvement in the generalization of policies to environments not seen during training.", "target": ["Wir schlagen einen Regularisierungsterm vor, der, wenn er dem Ziel des Reinforcement Learnings hinzugefügt wird, es der Strategie ermöglicht, die Belohnung zu maximieren und gleichzeitig zu lernen, sich gegenüber irrelevanten Änderungen innerhalb der Eingabe nicht zu verändern."]} +{"source": "Though visual information has been introduced for enhancing neural machine translation (NMT), its effectiveness strongly relies on the availability of large amounts of bilingual parallel sentence pairs with manual image annotations.In this paper, we present a universal visual representation learned over the monolingual corpora with image annotations, which overcomes the lack of large-scale bilingual sentence-image pairs, thereby extending image applicability in NMT.In detail, a group of images with similar topics to the source sentence will be retrieved from a light topic-image lookup table learned over the existing sentence-image pairs, and then is encoded as image representations by a pre-trained ResNet.An attention layer with a gated weighting is to fuse the visual information and text information as input to the decoder for predicting target translations.In particular, the proposed method enables the visual information to be integrated into large-scale text-only NMT in addition to the multimodel NMT.Experiments on four widely used translation datasets, including the WMT'16 English-to-Romanian, WMT'14 English-to-German, WMT'14 English-to-French, and Multi30K, show that the proposed approach achieves significant improvements over strong baselines.", "target": ["In dieser Arbeit wurde eine universelle visuelle Darstellung für die neuronale maschinelle Übersetzung (NMT) vorgeschlagen, die Bilder mit ähnlichen Themen wie der Ausgangssatz verwendet und die Anwendbarkeit von Bildern in der NMT erweitert."]} +{"source": "This paper introduces a novel framework for learning algorithms to solve online combinatorial optimization problems.Towards this goal, we introduce a number of key ideas from traditional algorithms and complexity theory.First, we draw a new connection between primal-dual methods and reinforcement learning.Next, we introduce the concept of adversarial distributions (universal and high-entropy training sets), which are distributions that encourage the learner to find algorithms that work well in the worst case.We test our new ideas on a number of optimization problem such as the AdWords problem, the online knapsack problem, and the secretary problem.Our results indicate that the models have learned behaviours that are consistent with the traditional optimal algorithms for these problems.", "target": ["Durch die Kombination von Ideen aus dem traditionellen Algorithmenentwurf und dem Reinforcement Learning stellen wir einen neuartigen Rahmen für Lernalgorithmen vor, die online kombinatorische Optimierungsprobleme lösen."]} +{"source": "Despite their popularity and successes, deep neural networks are poorly understood theoretically and treated as 'black box' systems.Using a functional view of these networks gives us a useful new lens with which to understand them.This allows us us to theoretically or experimentally probe properties of these networks, including the effect of standard initializations, the value of depth, the underlying loss surface, and the origins of generalization.One key result is that generalization results from smoothness of the functional approximation, combined with a flat initial approximation.This smoothness increases with number of units, explaining why massively overparamaterized networks continue to generalize well.", "target": ["Ein funktionaler Ansatz zeigt, dass eine flache Initialisierung, die durch den Gradientenabstieg erhalten bleibt, zu einer Generalisierungsfähigkeit führt."]} +{"source": "It is well-known that deeper neural networks are harder to train than shallower ones.In this short paper, we use the (full) eigenvalue spectrum of the Hessian to explore how the loss landscape changes as the network gets deeper, and as residual connections are added to the architecture.Computing a series of quantitative measures on the Hessian spectrum, we show that the Hessian eigenvalue distribution in deeper networks has substantially heavier tails (equivalently, more outlier eigenvalues), which makes the network harder to optimize with first-order methods.We show that adding residual connections mitigates this effect substantially, suggesting a mechanism by which residual connections improve training.", "target": ["Die Netzwerktiefe erhöht die Ausreißereigenwerte in der Hessian. Residuale Verbindungen mildern dies ab."]} +{"source": "In the context of optimization, a gradient of a neural network indicates the amount a specific weight should change with respect to the loss.Therefore, small gradients indicate a good value of the weight that requires no change and can be kept frozen during training.This paper provides an experimental study on the importance of a neural network weights, and to which extent do they need to be updated.We wish to show that starting from the third epoch, freezing weights which have no informative gradient and are less likely to be changed during training, results in a very slight drop in the overall accuracy (and in sometimes better).We experiment on the MNIST, CIFAR10 and Flickr8k datasets using several architectures (VGG19,ResNet-110 and DenseNet-121).On CIFAR10, we show that freezing 80% of the VGG19 network parameters from the third epoch onwards results in 0.24% drop in accuracy, while freezing 50% of Resnet-110 parameters results in 0.9% drop in accuracy and finally freezing 70% of Densnet-121 parameters results in 0.57% drop in accuracy.Furthermore, to experiemnt with real-life applications, we train an image captioning model with attention mechanism on the Flickr8k dataset using LSTM networks, freezing 60% of the parameters from the third epoch onwards, resulting in a better BLEU-4 score than the fully trained model.Our source code can be found in the appendix.", "target": ["Eine experimentelle Arbeit, die beweist, wie viele redundante Gewichte erst ab der dritten Epoche eingefroren werden können, wobei die Genauigkeit nur sehr geringfügig abnimmt."]} +{"source": "Like humans, deep networks learn better when samples are organized and introduced in a meaningful order or curriculum.While conventional approaches to curriculum learning emphasize the difficulty of samples as the core incremental strategy, it forces networks to learn from small subsets of data while introducing pre-computation overheads.In this work, we propose Learning with Incremental Labels and Adaptive Compensation (LILAC), which introduces a novel approach to curriculum learning.LILAC emphasizes incrementally learning labels instead of incrementally learning difficult samples.It works in two distinct phases: first, in the incremental label introduction phase, we unmask ground-truth labels in fixed increments during training, to improve the starting point from which networks learn.In the adaptive compensation phase, we compensate for failed predictions by adaptively altering the target vector to a smoother distribution.We evaluate LILAC against the closest comparable methods in batch and curriculum learning and label smoothing, across three standard image benchmarks, CIFAR-10, CIFAR-100, and STL-10.We show that our method outperforms batch learning with higher mean recognition accuracy as well as lower standard deviation in performance consistently across all benchmarks.We further extend LILAC to state-of-the-art performance across CIFAR-10 using simple data augmentation while exhibiting label order invariance among other important properties.", "target": ["Ein neuartiger Ansatz für das Lernen von Lehrplänen durch inkrementelles Lernen von Beschriftungen und adaptive Glättung von Beschriftungen für falsch klassifizierte Proben, was die durchschnittliche Leistung erhöht und die Standardabweichung verringert."]} +{"source": "Words in natural language follow a Zipfian distribution whereby some words are frequent but most are rare.Learning representations for words in the ``long tail'' of this distribution requires enormous amounts of data. Representations of rare words trained directly on end tasks are usually poor, requiring us to pre-train embeddings on external data, or treat all rare words as out-of-vocabulary words with a unique representation.We provide a method for predicting embeddings of rare words on the fly from small amounts of auxiliary data with a network trained end-to-end for the downstream task.We show that this improves results against baselines where embeddings are trained on the end task for reading comprehension, recognizing textual entailment and language modeling.", "target": ["Wir schlagen eine Methode vor, um mit seltenen Wörtern umzugehen, indem wir ihre Einbettung aus Definitionen berechnen."]} +{"source": "The capability of reliably detecting out-of-distribution samples is one of the key factors in deploying a good classifier, as the test distribution always does not match with the training distribution in most real-world applications.In this work, we propose a deep generative classifier which is effective to detect out-of-distribution samples as well as classify in-distribution samples, by integrating the concept of Gaussian discriminant analysis into deep neural networks.Unlike the discriminative (or softmax) classifier that only focuses on the decision boundary partitioning its latent space into multiple regions, our generative classifier aims to explicitly model class-conditional distributions as separable Gaussian distributions.Thereby, we can define the confidence score by the distance between a test sample and the center of each distribution.Our empirical evaluation on multi-class images and tabular data demonstrate that the generative classifier achieves the best performances in distinguishing out-of-distribution samples, and also it can be generalized well for various types of deep neural networks.", "target": ["Diese Arbeit schlägt einen tiefen generativen Klassifikator vor, der sowohl Beispiele außerhalb der Verteilung erkennen als auch Beispiele innerhalb der Verteilung klassifizieren kann, indem das Konzept der Gaußschen Diskriminanzanalyse in tiefe neuronale Netze integriert wird."]} +{"source": "One of the most prevalent symptoms among the elderly population, dementia, can be detected by classifiers trained on linguistic features extracted from narrative transcripts.However, these linguistic features are impacted in a similar but different fashion by the normal aging process.Aging is therefore a confounding factor, whose effects have been hard for machine learning classifiers to isolate. In this paper, we show that deep neural network (DNN) classifiers can infer ages from linguistic features, which is an entanglement that could lead to unfairness across age groups.We show this problem is caused by undesired activations of v-structures in causality diagrams, and it could be addressed with fair representation learning.We build neural network classifiers that learn low-dimensional representations reflecting the impacts of dementia yet discarding the effects of age.To evaluate these classifiers, we specify a model-agnostic score $\\Delta_{eo}^{(N)}$ measuring how classifier results are disentangled from age.Our best models outperform baseline neural network classifiers in disentanglement, while compromising accuracy by as little as 2.56\\% and 2.25\\% on DementiaBank and the Famous People dataset respectively.", "target": ["Aufzeigen, dass das Alter die Erkennung von kognitiven Beeinträchtigungen behindert + Lösen von Problemen mit dem Lernen von fairen Repräsentationen + Vorschlagen von Metriken und Modellen."]} +{"source": "Much of the focus in the design of deep neural networks had been on improving accuracy, leading to more powerful yet highly complex network architectures that are difficult to deploy in practical scenarios. As a result, there has been a recent interest in the design of quantitative metrics for evaluating deep neural networks that accounts for more than just model accuracy as the sole indicator of network performance. In this study, we continue the conversation towards universal metrics for evaluating the performance of deep neural networks for practical on-device edge usage by introducing NetScore, a new metric designed specifically to provide a quantitative assessment of the balance between accuracy, computational complexity, and network architecture complexity of a deep neural network. In what is one of the largest comparative analysis between deep neural networks in literature, the NetScore metric, the top-1 accuracy metric, and the popular information density metric were compared across a diverse set of 60 different deep convolutional neural networks for image classification on the ImageNet Large Scale Visual Recognition Challenge (ILSVRC 2012) dataset. The evaluation results across these three metrics for this diverse set of networks are presented in this study to act as a reference guide for practitioners in the field.", "target": ["Wir stellen NetScore vor, eine neue Metrik, die eine quantitative Bewertung des Gleichgewichts zwischen Genauigkeit, Rechenaufwand und Komplexität der Netzwerkarchitektur eines tiefen neuronalen Netzwerks ermöglicht."]} +{"source": "Deep Neural Networks (DNNs) are vulnerable to adversarial attacks, especially white-box targeted attacks.This paper studies the problem of how aggressive white-box targeted attacks can be to go beyond widely used Top-1 attacks.We propose to learn ordered Top-k attacks (k>=1), which enforce the Top-k predicted labels of an adversarial example to be the k (randomly) selected and ordered labels (the ground-truth label is exclusive).Two methods are presented.First, we extend the vanilla Carlini-Wagner (C&W) method and use it as a strong baseline.Second, we present an adversarial distillation framework consisting of two components:(i) Computing an adversarial probability distribution for any given ordered Top-$k$ targeted labels.(ii) Learning adversarial examples by minimizing the Kullback-Leibler (KL) divergence between the adversarial distribution and the predicted distribution, together with the perturbation energy penalty.In computing adversarial distributions, we explore how to leverage label semantic similarities, leading to knowledge-oriented attacks.In experiments, we test Top-k (k=1,2,5,10) attacks in the ImageNet-1000 val dataset using two popular DNNs trained with the clean ImageNet-1000 train dataset, ResNet-50 and DenseNet-121.Overall, the adversarial distillation approach obtains the best results, especially by large margin when computation budget is limited..It reduces the perturbation energy consistently with the same attack success rate on all the four k's, and improve the attack success rate by large margin against the modified C&W method for k=10.", "target": ["geordnete Top-k-Angriffe"]} +{"source": "Neural message passing algorithms for semi-supervised classification on graphs have recently achieved great success.However, for classifying a node these methods only consider nodes that are a few propagation steps away and the size of this utilized neighborhood is hard to extend.In this paper, we use the relationship between graph convolutional networks (GCN) and PageRank to derive an improved propagation scheme based on personalized PageRank.We utilize this propagation procedure to construct a simple model, personalized propagation of neural predictions (PPNP), and its fast approximation, APPNP.Our model's training time is on par or faster and its number of parameters on par or lower than previous models.It leverages a large, adjustable neighborhood for classification and can be easily combined with any neural network.We show that this model outperforms several recently proposed methods for semi-supervised classification in the most thorough study done so far for GCN-like models.Our implementation is available online.", "target": ["Personalisierte Propagierung neuronaler Vorhersagen (PPNP) verbessert graphische neuronale Netze, indem es sie in Vorhersage und Propagierung mittels personalisiertem PageRank trennt."]} +{"source": "Most of recent work in cross-lingual word embeddings is severely Anglocentric.The vast majority of lexicon induction evaluation dictionaries are between English and another language, and the English embedding space is selected by default as the hub when learning in a multilingual setting.With this work, however, we challenge these practices.First, we show that the choice of hub language can significantly impact downstream lexicon induction performance.Second, we both expand the current evaluation dictionary collection to include all language pairs using triangulation, and also create new dictionaries for under-represented languages.Evaluating established methods over all these language pairs sheds light into their suitability and presents new challenges for the field.Finally, in our analysis we identify general guidelines for strong cross-lingual embeddings baselines, based on more than just Anglocentric experiments.", "target": ["Die Wahl der Zielsprache wirkt sich auf die Qualität der sprachübergreifenden Einbettungen aus, die nicht nur anhand englischsprachiger Wörterbücher bewertet werden sollten."]} +{"source": "Interpreting generative adversarial network (GAN) training as approximate divergence minimization has been theoretically insightful, has spurred discussion, and has lead to theoretically and practically interesting extensions such as f-GANs and Wasserstein GANs.For both classic GANs and f-GANs, there is an original variant of training and a \"non-saturating\" variant which uses an alternative form of generator update.The original variant is theoretically easier to study, but the alternative variant frequently performs better and is recommended for use in practice.The alternative generator update is often regarded as a simple modification to deal with optimization issues, and it appears to be a common misconception that the two variants minimize the same divergence.In this short note we derive the divergences approximately minimized by the original and alternative variants of GAN and f-GAN training.This highlights important differences between the two variants.For example, we show that the alternative variant of KL-GAN training actually minimizes the reverse KL divergence, and that the alternative variant of conventional GAN training minimizes a \"softened\" version of the reverse KL.We hope these results may help to clarify some of the theoretical discussion surrounding the divergence minimization view of GAN training.", "target": ["Das typische GAN-Training optimiert nicht Jensen-Shannon, sondern so etwas wie eine umgekehrte KL-Divergenz."]} +{"source": "REINFORCE can be used to train models in structured prediction settings to directly optimize the test-time objective.However, the common case of sampling one prediction per datapoint (input) is data-inefficient.We show that by drawing multiple samples (predictions) per datapoint, we can learn with significantly less data, as we freely obtain a REINFORCE baseline to reduce variance.Additionally we derive a REINFORCE estimator with baseline, based on sampling without replacement.Combined with a recent technique to sample sequences without replacement using Stochastic Beam Search, this improves the training procedure for a sequence model that predicts the solution to the Travelling Salesman Problem.", "target": ["Wir zeigen, dass wir durch das Ziehen mehrerer Stichproben (Vorhersagen) pro Eingabe (Datenpunkt) mit weniger Daten lernen können, da wir eine REINFORCE-Basislinie erhalten."]} +{"source": "Reinforcement learning (RL) is a powerful technique to train an agent to perform a task. However, an agent that is trained using RL is only capable of achieving the single task that is specified via its reward function. Such an approach does not scale well to settings in which an agent needs to perform a diverse set of tasks, such as navigating to varying positions in a room or moving objects to varying locations. Instead, we propose a method that allows an agent to automatically discover the range of tasks that it is capable of performing in its environment. We use a generator network to propose tasks for the agent to try to achieve, each task being specified as reaching a certain parametrized subset of the state-space. The generator network is optimized using adversarial training to produce tasks that are always at the appropriate level of difficulty for the agent. Our method thus automatically produces a curriculum of tasks for the agent to learn. We show that, by using this framework, an agent can efficiently and automatically learn to perform a wide set of tasks without requiring any prior knowledge of its environment (Videos and code available at: https://sites.google.com/view/goalgeneration4rl).Our method can also learn to achieve tasks with sparse rewards, which pose significant challenges for traditional RL methods.", "target": ["Wir lösen Multitasking-Probleme effizient mit einem automatischen Algorithmus zur Erstellung von Lehrplänen auf der Grundlage eines generativen Modells, das die Leistung des Lernagenten verfolgt."]} +{"source": "A wide range of defenses have been proposed to harden neural networks against adversarial attacks.However, a pattern has emerged in which the majority of adversarial defenses are quickly broken by new attacks. Given the lack of success at generating robust defenses, we are led to ask a fundamental question: Are adversarial attacks inevitable?This paper analyzes adversarial examples from a theoretical perspective, and identifies fundamental bounds on the susceptibility of a classifier to adversarial attacks. We show that, for certain classes of problems, adversarial examples are inescapable. Using experiments, we explore the implications of theoretical guarantees for real-world problems and discuss how factors such as dimensionality and image complexity limit a classifier's robustness against adversarial examples.", "target": ["In dieser Arbeit werden Klassen von Problemen identifiziert, für die negative Beispiele unausweichlich sind, und es werden grundlegende Grenzen für die Anfälligkeit jedes Klassifizierers für negative Beispiele abgeleitet."]} +{"source": "For computer vision applications, prior works have shown the efficacy of reducing numeric precision of model parameters (network weights) in deep neural networks.Activation maps, however, occupy a large memory footprint during both the training and inference step when using mini-batches of inputs.One way to reduce this large memory footprint is to reduce the precision of activations.However, past works have shown that reducing the precision of activations hurts model accuracy.We study schemes to train networks from scratch using reduced-precision activations without hurting accuracy.We reduce the precision of activation maps (along with model parameters) and increase the number of filter maps in a layer, and find that this scheme matches or surpasses the accuracy of the baseline full-precision network.As a result, one can significantly improve the execution efficiency (e.g. reduce dynamic memory footprint, memory band- width and computational energy) and speed up the training and inference process with appropriate hardware support.We call our scheme WRPN -- wide reduced-precision networks.We report results and show that WRPN scheme is better than previously reported accuracies on ILSVRC-12 dataset while being computationally less expensive compared to previously reported reduced-precision networks.", "target": ["Die Verringerung der Genauigkeit (auf 4 Bits, 2 Bits und sogar binär) und die Erweiterung der Filterbänke führt zu ebenso genauen Netzen wie die, die mit FP32-Gewichten und -Aktivierungen erhalten werden."]} +{"source": "We investigate methods for semi-supervised learning (SSL) of a neural linear-chain conditional random field (CRF) for Named Entity Recognition (NER) by treating the tagger as the amortized variational posterior in a generative model of text given tags.We first illustrate how to incorporate a CRF in a VAE, enabling end-to-end training on semi-supervised data.We then investigate a series of increasingly complex deep generative models of tokens given tags enabled by end-to-end optimization, comparing the proposed models against supervised and strong CRF SSL baselines on the Ontonotes5 NER dataset.We find that our best proposed model consistently improves performance by $\\approx 1\\%$ F1 in low- and moderate-resource regimes and easily addresses degenerate model behavior in a more difficult, partially supervised setting.", "target": ["Wir betten eine CRF in eine VAE von Token und NER-Tags für halbüberwachtes Lernen ein und zeigen Verbesserungen in ressourcenarmen Umgebungen."]} +{"source": "To make deep neural networks feasible in resource-constrained environments (such as mobile devices), it is beneficial to quantize models by using low-precision weights.One common technique for quantizing neural networks is the straight-through gradient method, which enables back-propagation through the quantization mapping.Despite its empirical success, little is understood about why the straight-through gradient method works.Building upon a novel observation that the straight-through gradient method is in fact identical to the well-known Nesterov’s dual-averaging algorithm on a quantization constrained optimization problem, we propose a more principled alternative approach, called ProxQuant , that formulates quantized network training as a regularized learning problem instead and optimizes it via the prox-gradient method.ProxQuant does back-propagation on the underlying full-precision vector and applies an efficient prox-operator in between stochastic gradient steps to encourage quantizedness.For quantizing ResNets and LSTMs, ProxQuant outperforms state-of-the-art results on binary quantization and is on par with state-of-the-art on multi-bit quantization.For binary quantization, our analysis shows both theoretically and experimentally that ProxQuant is more stable than the straight-through gradient method (i.e. BinaryConnect), challenging the indispensability of the straight-through gradient method and providing a powerful alternative.", "target": ["Ein prinzipieller Rahmen für die Modellquantisierung unter Verwendung der proximalen Gradientenmethode."]} +{"source": "The vulnerabilities of deep neural networks against adversarial examples have become a significant concern for deploying these models in sensitive domains.Devising a definitive defense against such attacks is proven to be challenging, and the methods relying on detecting adversarial samples are only valid when the attacker is oblivious to the detection mechanism.In this paper, we consider the adversarial detection problem under the robust optimization framework.We partition the input space into subspaces and train adversarial robust subspace detectors using asymmetrical adversarial training (AAT).The integration of the classifier and detectors presents a detection mechanism that provides a performance guarantee to the adversary it considered.We demonstrate that AAT promotes the learning of class-conditional distributions, which further gives rise to generative detection/classification approaches that are both robust and more interpretable.We provide comprehensive evaluations of the above methods, and demonstrate their competitive performances and compelling properties on adversarial detection and robust classification problems.", "target": ["Eine neue generative Modellierungstechnik, die auf asymmetrischem adversarialem Training basiert, und ihre Anwendungen zur Erkennung von adversarialem Beispiel und robuster Klassifizierung"]} +{"source": "Exploration is a key component of successful reinforcement learning, but optimal approaches are computationally intractable, so researchers have focused on hand-designing mechanisms based on exploration bonuses and intrinsic reward, some inspired by curious behavior in natural systems. In this work, we propose a strategy for encoding curiosity algorithms as programs in a domain-specific language and searching, during a meta-learning phase, for algorithms that enable RL agents to perform well in new domains. Our rich language of programs, which can combine neural networks with other building blocks including nearest-neighbor modules and can choose its own loss functions, enables the expression of highly generalizable programs that perform well in domains as disparate as grid navigation with image input, acrobot, lunar lander, ant and hopper. To make this approach feasible, we develop several pruning techniques, including learning to predict a program's success based on its syntactic properties. We demonstrate the effectiveness of the approach empirically, finding curiosity strategies that are similar to those in published literature, as well as novel strategies that are competitive with them and generalize well.", "target": ["Das Meta-Lernen von neugierigen Algorithmen durch die Suche in einem reichhaltigen Raum von Programmen führt zu neuartigen Mechanismen, die sich über sehr unterschiedliche Bereiche des Reinforcement Learnings verallgemeinern lassen."]} +{"source": "Many machine learning algorithms represent input data with vector embeddings or discrete codes.When inputs exhibit compositional structure (e.g. objects built from parts or procedures from subroutines), it is natural to ask whether this compositional structure is reflected in the the inputs’ learned representations.While the assessment of compositionality in languages has received significant attention in linguistics and adjacent fields, the machine learning literature lacks general-purpose tools for producing graded measurements of compositional structure in more general (e.g. vector-valued) representation spaces.We describe a procedure for evaluating compositionality by measuring how well the true representation-producing model can be approximated by a model that explicitly composes a collection of inferred representational primitives.We use the procedure to provide formal and empirical characterizations of compositional structure in a variety of settings, exploring the relationship between compositionality and learning dynamics, human judgments, representational similarity, and generalization.", "target": ["In diesem Beitrag wird ein einfaches Verfahren zur Bewertung der Kompositionsstruktur in gelernten Repräsentationen vorgeschlagen und das Verfahren verwendet, um die Rolle der Kompositionalität bei vier Lernproblemen zu untersuchen."]} +{"source": "In this paper, we propose an end-to-end deep learning model, called E2Efold, for RNA secondary structure prediction which can effectively take into account the inherent constraints in the problem.The key idea of E2Efold is to directly predict the RNA base-pairing matrix, and use an unrolled constrained programming algorithm as a building block in the architecture to enforce constraints.With comprehensive experiments on benchmark datasets, we demonstrate the superior performance of E2Efold: it predicts significantly better structures compared to previous SOTA (29.7% improvement in some cases in F1 scores and even larger improvement for pseudoknotted structures) and runs as efficient as the fastest algorithms in terms of inference time.", "target": ["Ein DL-Modell für die Vorhersage von RNA-Sekundärstrukturen, das in der Architektur einen ausgerollten Algorithmus zur Durchsetzung von Beschränkungen verwendet."]} +{"source": "Learning in recurrent neural networks (RNNs) is most often implemented by gradient descent using backpropagation through time (BPTT), but BPTT does not model accurately how the brain learns.Instead, many experimental results on synaptic plasticity can be summarized as three-factor learning rules involving eligibility traces of the local neural activity and a third factor.We present here eligibility propagation (e-prop), a new factorization of the loss gradients in RNNs that fits the framework of three factor learning rules when derived for biophysical spiking neuron models.When tested on the TIMIT speech recognition benchmark, it is competitive with BPTT both for training artificial LSTM networks and spiking RNNs.Further analysis suggests that the diversity of learning signals and the consideration of slow internal neural dynamics are decisive to the learning efficiency of e-prop.", "target": ["Wir stellen eine Alternative zu BPTT vor, die mit experimentellen Daten zur synaptischen Plastizität kompatibel ist und mit BPTT in Benchmarks für maschinelles Lernen konkurriert."]} +{"source": "Recurrent neural networks (RNNs) are an effective representation of control policies for a wide range of reinforcement and imitation learning problems.RNN policies, however, are particularly difficult to explain, understand, and analyze due to their use of continuous-valued memory vectors and observation features.In this paper, we introduce a new technique, Quantized Bottleneck Insertion, to learn finite representations of these vectors and features.The result is a quantized representation of the RNN that can be analyzed to improve our understanding of memory use and general behavior.We present results of this approach on synthetic environments and six Atari games.The resulting finite representations are surprisingly small in some cases, using as few as 3 discrete memory states and 10 observations for a perfect Pong policy.We also show that these finite policy representations lead to improved interpretability.", "target": ["Extrahieren einer endlichen Zustandsmaschine aus einem rekurrenten neuronalen Netz mittels Quantisierung zum Zwecke der Interpretierbarkeit mit Experimenten auf dem Atari."]} +{"source": "Building upon the recent success of deep reinforcement learning methods, we investigate the possibility of on-policy reinforcement learning improvement by reusing the data from several consecutive policies.On-policy methods bring many benefits, such as ability to evaluate each resulting policy.However, they usually discard all the information about the policies which existed before.In this work, we propose adaptation of the replay buffer concept, borrowed from the off-policy learning setting, to the on-policy algorithms.To achieve this, the proposed algorithm generalises the Q-, value and advantage functions for data from multiple policies.The method uses trust region optimisation, while avoiding some of the common problems of the algorithms such as TRPO or ACKTR: it uses hyperparameters to replace the trust region selection heuristics, as well as the trainable covariance matrix instead of the fixed one.In many cases, the method not only improves the results comparing to the state-of-the-art trust region on-policy learning algorithms such as ACKTR and TRPO, but also with respect to their off-policy counterpart DDPG.", "target": ["Wir untersuchen die theoretischen und praktischen Beweise für die Verbesserung des On-Policy Reinforcement Learning durch die Wiederverwendung der Daten von mehreren aufeinanderfolgenden Policies."]} +{"source": "Convolutional Neural Networks (CNN) have been successful in processing data signals that are uniformly sampled in the spatial domain (e.g., images).However, most data signals do not natively exist on a grid, and in the process of being sampled onto a uniform physical grid suffer significant aliasing error and information loss.Moreover, signals can exist in different topological structures as, for example, points, lines, surfaces and volumes.It has been challenging to analyze signals with mixed topologies (for example, point cloud with surface mesh).To this end, we develop mathematical formulations for Non-Uniform Fourier Transforms (NUFT) to directly, and optimally, sample nonuniform data signals of different topologies defined on a simplex mesh into the spectral domain with no spatial sampling error.The spectral transform is performed in the Euclidean space, which removes the translation ambiguity from works on the graph spectrum.Our representation has four distinct advantages: (1) the process causes no spatial sampling error during initial sampling, (2) the generality of this approach provides a unified framework for using CNNs to analyze signals of mixed topologies, (3) it allows us to leverage state-of-the-art backbone CNN architectures for effective learning without having to design a particular architecture for a particular data structure in an ad-hoc fashion, and (4) the representation allows weighted meshes where each element has a different weight (i.e., texture) indicating local properties.We achieve good results on-par with state-of-the-art for 3D shape retrieval task, and new state-of-the-art for point cloud to surface reconstruction task.", "target": ["Wir nutzen die nicht-euklidische Fourier-Transformation von Formen, die durch einen simpliziellen Komplex definiert sind, für Deep Learning und erzielen damit deutlich bessere Ergebnisse als mit punktbasierten Sampling-Techniken, die in der aktuellen Literatur zum 3D-Lernen verwendet werden."]} +{"source": "Discovering causal structure among a set of variables is a fundamental problem in many empirical sciences.Traditional score-based casual discovery methods rely on various local heuristics to search for a Directed Acyclic Graph (DAG) according to a predefined score function.While these methods, e.g., greedy equivalence search, may have attractive results with infinite samples and certain model assumptions, they are less satisfactory in practice due to finite data and possible violation of assumptions.Motivated by recent advances in neural combinatorial optimization, we propose to use Reinforcement Learning (RL) to search for the DAG with the best scoring.Our encoder-decoder model takes observable data as input and generates graph adjacency matrices that are used to compute rewards.The reward incorporates both the predefined score function and two penalty terms for enforcing acyclicity.In contrast with typical RL applications where the goal is to learn a policy, we use RL as a search strategy and our final output would be the graph, among all graphs generated during training, that achieves the best reward.We conduct experiments on both synthetic and real datasets, and show that the proposed approach not only has an improved search ability but also allows for a flexible score function under the acyclicity constraint.", "target": ["Wir wenden Reinforcement Learning auf die Score-basierte Kausalerkennung an und erzielen vielversprechende Ergebnisse sowohl auf synthetischen als auch auf realen Datensätzen"]} +{"source": "The topic modeling discovers the latent topic probability of given the text documents.To generate the more meaningful topic that better represents the given document, we proposed a universal method which can be used in the data preprocessing stage.The method consists of three steps.First, it generates the word/word-pair from every single document.Second, it applies a two way parallel TF-IDF algorithm to word/word-pair for semantic filtering.Third, it uses the k-means algorithm to merge the word pairs that have the similar semantic meaning.Experiments are carried out on the Open Movie Database (OMDb), Reuters Dataset and 20NewsGroup Dataset and use the mean Average Precision score as the evaluation metric.Comparing our results with other state-of-the-art topic models, such as Latent Dirichlet allocation and traditional Restricted Boltzmann Machines.Our proposed data preprocessing can improve the generated topic accuracy by up to 12.99\\%.How the number of clusters and the number of word pairs should be adjusted for different type of text document is also discussed.", "target": ["Wir haben eine universelle Methode vorgeschlagen, die in der Vorverarbeitungsphase der Daten verwendet werden kann, um das aussagekräftigere Thema zu generieren, das das gegebene Dokument besser repräsentiert"]} +{"source": "Multi-task learning has been successful in modeling multiple related tasks with large, carefully curated labeled datasets.By leveraging the relationships among different tasks, multi-task learning framework can improve the performance significantly.However, most of the existing works are under the assumption that the predefined tasks are related to each other.Thus, their applications on real-world are limited, because rare real-world problems are closely related.Besides, the understanding of relationships among tasks has been ignored by most of the current methods.Along this line, we propose a novel multi-task learning framework - Learning To Transfer Via Modelling Multi-level Task Dependency, which constructed attention based dependency relationships among different tasks.At the same time, the dependency relationship can be used to guide what knowledge should be transferred, thus the performance of our model also be improved.To show the effectiveness of our model and the importance of considering multi-level dependency relationship, we conduct experiments on several public datasets, on which we obtain significant improvements over current methods.", "target": ["Wir schlagen ein neuartiges Multitasking-Lernsystem vor, das automatisch Multi-View-Abhängigkeitsbeziehungen extrahiert und diese zur Steuerung des Wissenstransfers zwischen verschiedenen Aufgaben verwendet."]} +{"source": "We design simple and quantifiable testing of global translation-invariance in deep learning models trained on the MNIST dataset.Experiments on convolutional and capsules neural networks show that both models have poor performance in dealing with global translation-invariance; however, the performance improved by using data augmentation.Although the capsule network is better on the MNIST testing dataset, the convolutional neural network generally has better performance on the translation-invariance.", "target": ["Prüfung der globalen Translationsinvarianz in Convolutional und Kapselnetzen"]} +{"source": "Gaussian processes are ubiquitous in nature and engineering.A case in point is a class of neural networks in the infinite-width limit, whose priors correspond to Gaussian processes.Here we perturbatively extend this correspondence to finite-width neural networks, yielding non-Gaussian processes as priors.The methodology developed herein allows us to track the flow of preactivation distributions by progressively integrating out random variables from lower to higher layers, reminiscent of renormalization-group flow.We further develop a perturbative prescription to perform Bayesian inference with weakly non-Gaussian priors.", "target": ["Wir entwickeln eine analytische Methode zur Untersuchung der Bayes'schen Inferenz von neuronalen Netzen mit endlicher Breite und stellen fest, dass sich das Bild des Renormierungsgruppenflusses auf natürliche Weise ergibt."]} +{"source": "Distillation is a method to transfer knowledge from one model to another and often achieves higher accuracy with the same capacity.In this paper, we aim to provide a theoretical understanding on what mainly helps with the distillation.Our answer is \"early stopping\".Assuming that the teacher network is overparameterized, we argue that the teacher network is essentially harvesting dark knowledge from the data via early stopping.This can be justified by a new concept, Anisotropic In- formation Retrieval (AIR), which means that the neural network tends to fit the informative information first and the non-informative information (including noise) later.Motivated by the recent development on theoretically analyzing overparame- terized neural networks, we can characterize AIR by the eigenspace of the Neural Tangent Kernel(NTK).AIR facilities a new understanding of distillation.With that, we further utilize distillation to refine noisy labels.We propose a self-distillation al- gorithm to sequentially distill knowledge from the network in the previous training epoch to avoid memorizing the wrong labels.We also demonstrate, both theoret- ically and empirically, that self-distillation can benefit from more than just early stopping.Theoretically, we prove convergence of the proposed algorithm to the ground truth labels for randomly initialized overparameterized neural networks in terms of l2 distance, while the previous result was on convergence in 0-1 loss.The theoretical result ensures the learned neural network enjoy a margin on the training data which leads to better generalization.Empirically, we achieve better testing accuracy and entirely avoid early stopping which makes the algorithm more user-friendly.", "target": ["Theoretisches verstehen des Regularisierungseffekt der Destillation. Wir zeigen, dass ein frühes Anhalten in diesem Prozess essentiell ist. Aus dieser Perspektive haben wir eine Destillationsmethode für das Lernen mit korrupiertem Label und mit theoretischen Garantien entwickelt."]} +{"source": "Lifelong learning poses considerable challenges in terms of effectiveness (minimizing prediction errors for all tasks) and overall computational tractability for real-time performance. This paper addresses continuous lifelong multitask learning by jointly re-estimating the inter-task relations (\\textit{output} kernel) and the per-task model parameters at each round, assuming data arrives in a streaming fashion.We propose a novel algorithm called \\textit{Online Output Kernel Learning Algorithm} (OOKLA) for lifelong learning setting.To avoid the memory explosion, we propose a robust budget-limited versions of the proposed algorithm that efficiently utilize the relationship between the tasks to bound the total number of representative examples in the support set. In addition, we propose a two-stage budgeted scheme for efficiently tackling the task-specific budget constraints in lifelong learning.Our empirical results over three datasets indicate superior AUC performance for OOKLA and its budget-limited cousins over strong baselines.", "target": ["ein neuartiger Ansatz für lebenslanges Online-Lernen unter Verwendung von Output-Kerneln."]} +{"source": "Minecraft is a videogame that offers many interesting challenges for AI systems.In this paper, we focus in construction scenarios where an agent must build a complex structure made of individual blocks.As higher-level objects are formed of lower-level objects, the construction can naturally be modelled as a hierarchical task network.We model a house-construction scenario in classical and HTN planning and compare the advantages and disadvantages of both kinds of models.", "target": ["Wir modellieren ein Hausbau-Szenario in Minecraft in klassischer und HTN-Planung und vergleichen die Vor- und Nachteile beider Modellarten."]} +{"source": "Attacks on natural language models are difficult to compare due to their different definitions of what constitutes a successful attack.We present a taxonomy of constraints to categorize these attacks.For each constraint, we present a real-world use case and a way to measure how well generated samples enforce the constraint.We then employ our framework to evaluate two state-of-the art attacks which fool models with synonym substitution.These attacks claim their adversarial perturbations preserve the semantics and syntactical correctness of the inputs, but our analysis shows these constraints are not strongly enforced.For a significant portion of these adversarial examples, a grammar checker detects an increase in errors.Additionally, human studies indicate that many of these adversarial examples diverge in semantic meaning from the input or do not appear to be human-written.Finally, we highlight the need for standardized evaluation of attacks that share constraints.Without shared evaluation metrics, it is up to researchers to set thresholds that determine the trade-off between attack quality and attack success.We recommend well-designed human studies to determine the best threshold to approximate human judgement.", "target": ["Wir stellen einen Rahmen für die Bewertung von Gegenbeispielen in der natürlichen Sprachverarbeitung vor und zeigen, dass generierte Gegenbeispiele oft nicht semantikerhaltend, syntaktisch korrekt oder unverdächtig sind."]} +{"source": "In order for machine learning to be deployed and trusted in many applications, it is crucial to be able to reliably explain why the machine learning algorithm makes certain predictions.For example, if an algorithm classifies a given pathology image to be a malignant tumor, then the doctor may need to know which parts of the image led the algorithm to this classification.How to interpret black-box predictors is thus an important and active area of research. A fundamental question is: how much can we trust the interpretation itself?In this paper, we show that interpretation of deep learning predictions is extremely fragile in the following sense: two perceptively indistinguishable inputs with the same predicted label can be assigned very different}interpretations.We systematically characterize the fragility of the interpretations generated by several widely-used feature-importance interpretation methods (saliency maps, integrated gradient, and DeepLIFT) on ImageNet and CIFAR-10.Our experiments show that even small random perturbation can change the feature importance and new systematic perturbations can lead to dramatically different interpretations without changing the label.We extend these results to show that interpretations based on exemplars (e.g. influence functions) are similarly fragile.Our analysis of the geometry of the Hessian matrix gives insight on why fragility could be a fundamental challenge to the current interpretation approaches.", "target": ["Können wir der Erklärung eines neuronalen Netzes für seine Vorhersage vertrauen? Wir untersuchen die Robustheit verschiedener gängiger Vorstellungen von der Interpretierbarkeit neuronaler Netze, einschließlich herausragender Zuordnungen und Einflussfunktionen, und entwerfen Gegenbeispiele zu ihnen."]} +{"source": "Stochastic AUC maximization has garnered an increasing interest due to better fit to imbalanced data classification.However, existing works are limited to stochastic AUC maximization with a linear predictive model, which restricts its predictive power when dealing with extremely complex data.In this paper, we consider stochastic AUC maximization problem with a deep neural network as the predictive model.Building on the saddle point reformulation of a surrogated loss of AUC, the problem can be cast into a {\\it non-convex concave} min-max problem.The main contribution made in this paper is to make stochastic AUC maximization more practical for deep neural networks and big data with theoretical insights as well.In particular, we propose to explore Polyak-\\L{}ojasiewicz (PL) condition that has been proved and observed in deep learning, which enables us to develop new stochastic algorithms with even faster convergence rate and more practical step size scheme.An AdaGrad-style algorithm is also analyzed under the PL condition with adaptive convergence rate.Our experimental results demonstrate the effectiveness of the proposed algorithms.", "target": ["In dem Papier werden zwei Algorithmen für das stochastische AUC-Maximierungsproblem mit modernster Komplexität bei Verwendung eines tiefen neuronalen Netzes als Vorhersagemodell entwickelt, die auch durch empirische Studien verifiziert werden."]} +{"source": "Designing rewards for Reinforcement Learning (RL) is challenging because it needs to convey the desired task, be efficient to optimize, and be easy to compute.The latter is particularly problematic when applying RL to robotics, where detecting whether the desired configuration is reached might require considerable supervision and instrumentation.Furthermore, we are often interested in being able to reach a wide range of configurations, hence setting up a different reward every time might be unpractical.Methods like Hindsight Experience Replay (HER) have recently shown promise to learn policies able to reach many goals, without the need of a reward.Unfortunately, without tricks like resetting to points along the trajectory, HER might take a very long time to discover how to reach certain areas of the state-space.In this work we investigate different approaches to incorporate demonstrations to drastically speed up the convergence to a policy able to reach any goal, also surpassing the performance of an agent trained with other Imitation Learning algorithms.Furthermore, our method can be used when only trajectories without expert actions are available, which can leverage kinestetic or third person demonstration.", "target": ["Wir lösen zielgerichtete Aufgaben, indem wir die Algorithmen des Einsichts-Erfahrungs-Wiederholens und des Nachahmungs-Lernens kombinieren, wobei wir eine schnellere Konvergenz als der erste und eine höhere Endleistung als der zweite Algorithmus zeigen."]} +{"source": "Bayesian neural networks, which both use the negative log-likelihood loss function and average their predictions using a learned posterior over the parameters, have been used successfully across many scientific fields, partly due to their ability to `effortlessly' extract desired representations from many large-scale datasets.However, generalization bounds for this setting is still missing.In this paper, we present a new PAC-Bayesian generalization bound for the negative log-likelihood loss which utilizes the \\emph{Herbst Argument} for the log-Sobolev inequality to bound the moment generating function of the learners risk.", "target": ["Wir leiten einen neuen PAC-Bayesian Bound für unbeschränkte Verlustfunktionen (z.B. Negative Log-Likelihood) ab."]} +{"source": "Data augmentation techniques, e.g., flipping or cropping, which systematically enlarge the training dataset by explicitly generating more training samples, are effective in improving the generalization performance of deep neural networks.In the supervised setting, a common practice for data augmentation is to assign the same label to all augmented samples of the same source.However, if the augmentation results in large distributional discrepancy among them (e.g., rotations), forcing their label invariance may be too difficult to solve and often hurts the performance.To tackle this challenge, we suggest a simple yet effective idea of learning the joint distribution of the original and self-supervised labels of augmented samples.The joint learning framework is easier to train, and enables an aggregated inference combining the predictions from different augmented samples for improving the performance.Further, to speed up the aggregation process, we also propose a knowledge transfer technique, self-distillation, which transfers the knowledge of augmentation into the model itself.We demonstrate the effectiveness of our data augmentation framework on various fully-supervised settings including the few-shot and imbalanced classification scenarios.", "target": ["Wir schlagen eine einfache Technik zur selbstüberwachten Datenerweiterung vor, die die Leistung vollständig überwachter Szenarien verbessert, einschließlich des Few-Shot Learnings und der unausgewogenen Klassifizierung."]} +{"source": "Long short-term memory (LSTM) networks allow to exhibit temporal dynamic behavior with feedback connections and seem a natural choice for learning sequences of 3D meshes.We introduce an approach for dynamic mesh representations as used for numerical simulations of car crashes.To bypass the complication of using 3D meshes, we transform the surface mesh sequences into spectral descriptors that efficiently encode the shape.A two branch LSTM based network architecture is chosen to learn the representations and dynamics of the crash during the simulation.The architecture is based on unsupervised video prediction by an LSTM without any convolutional layer.It uses an encoder LSTM to map an input sequence into a fixed length vector representation.On this representation one decoder LSTM performs the reconstruction of the input sequence, while the other decoder LSTM predicts the future behavior by receiving initial steps of the sequence as seed.The spatio-temporal error behavior of the model is analysed to study how well the model can extrapolate the learned spectral descriptors into the future, that is, how well it has learned to represent the underlying dynamical structural mechanics.Considering that only a few training examples are available, which is the typical case for numerical simulations, the network performs very well.", "target": ["Eine zweigliedrige LSTM-basierte Netzwerkarchitektur lernt die Repräsentation und Dynamik von 3D-Netzen aus numerischen Crash-Simulationen."]} +{"source": "The purpose of an encoding model is to predict brain activity given a stimulus.In this contribution, we attempt at estimating a whole brain encoding model of auditory perception in a naturalistic stimulation setting.We analyze data from an open dataset, in which 16 subjects watched a short movie while their brain activity was being measured using functional MRI.We extracted feature vectors aligned with the timing of the audio from the movie, at different layers of a Deep Neural Network pretrained on the classification of auditory scenes.fMRI data was parcellated using hierarchical clustering on 500 parcels, and encoding models were estimated using a fully connected neural network with one hidden layer, trained to predict the signals for each parcel from the DNN features.Individual encoding models were successfully trained and predicted brain activity on unseen data, in parcels located in the superior temporal lobe, as well as dorsolateral prefrontal regions, which are usually considered as areas involved in auditory and language processing.Taken together, this contribution extends previous attempts on estimating encoding models, by showing the ability to model brain activity using a generic DNN (ie not specifically trained for this purpose) to extract auditory features, suggesting a degree of similarity between internal DNN representations and brain activity in naturalistic settings.", "target": ["Merkmalsvektoren aus SoundNet können die Gehirnaktivität von Personen, die einen Film ansehen, in auditiven und sprachbezogenen Gehirnregionen vorhersagen."]} +{"source": "In this paper, we describe the \"implicit autoencoder\" (IAE), a generative autoencoder in which both the generative path and the recognition path are parametrized by implicit distributions.We use two generative adversarial networks to define the reconstruction and the regularization cost functions of the implicit autoencoder, and derive the learning rules based on maximum-likelihood learning.Using implicit distributions allows us to learn more expressive posterior and conditional likelihood distributions for the autoencoder.Learning an expressive conditional likelihood distribution enables the latent code to only capture the abstract and high-level information of the data, while the remaining information is captured by the implicit conditional likelihood distribution.For example, we show that implicit autoencoders can disentangle the global and local information, and perform deterministic or stochastic reconstructions of the images.We further show that implicit autoencoders can disentangle discrete underlying factors of variation from the continuous factors in an unsupervised fashion, and perform clustering and semi-supervised learning.", "target": ["Wir schlagen einen generativen Autoencoder vor, der aussagekräftige posteriore und bedingte Wahrscheinlichkeitsverteilungen unter Verwendung impliziter Verteilungen erlernen kann, und trainieren das Modell mit einer neuen Formulierung der ELBO."]} +{"source": "Strong inductive biases allow children to learn in fast and adaptable ways.Children use the mutual exclusivity (ME) bias to help disambiguate how words map to referents, assuming that if an object has one label then it does not need another.In this paper, we investigate whether or not standard neural architectures have a ME bias, demonstrating that they lack this learning assumption.Moreover, we show that their inductive biases are poorly matched to lifelong learning formulations of classification and translation.We demonstrate that there is a compelling case for designing neural networks that reason by mutual exclusivity, which remains an open challenge.", "target": ["Kinder nutzen die gegenseitige Ausschließlichkeit (Mutual Exclusivity, ME), um neue Wörter zu lernen, während herkömmliche neuronale Netze die gegenteilige Tendenz aufweisen, was das Lernen in naturalistischen Szenarien wie dem lebenslangen Lernen behindert."]} +{"source": "Cortical neurons process and integrate information on multiple timescales.In addition, these timescales or temporal receptive fields display functional and hierarchical organization.For instance, areas important for working memory (WM), such as prefrontal cortex, utilize neurons with stable temporal receptive fields and long timescales to support reliable representations of stimuli.Despite of the recent advances in experimental techniques, the underlying mechanisms for the emergence of neuronal timescales long enough to support WM are unclear and challenging to investigate experimentally.Here, we demonstrate that spiking recurrent neural networks (RNNs) designed to perform a WM task reproduce previously observed experimental findings and that these models could be utilized in the future to study how neuronal timescales specific to WM emerge.", "target": ["Rekurrente neuronale Netzwerke mit Spikes, die eine Arbeitsgedächtnisaufgabe ausführen, verwenden lange heterogene Zeitskalen, die denen im präfrontalen Kortex verblüffend ähnlich sind."]} +{"source": "Conventional deep learning classifiers are static in the sense that they are trained ona predefined set of classes and learning to classify a novel class typically requiresre-training.In this work, we address the problem of Low-shot network-expansionlearning.We introduce a learning framework which enables expanding a pre-trained(base) deep network to classify novel classes when the number of examples for thenovel classes is particularly small.We present a simple yet powerful distillationmethod where the base network is augmented with additional weights to classifythe novel classes, while keeping the weights of the base network unchanged.Weterm this learning hard distillation, since we preserve the response of the networkon the old classes to be equal in both the base and the expanded network.Weshow that since only a small number of weights needs to be trained, the harddistillation excels for low-shot training scenarios.Furthermore, hard distillationavoids detriment to classification performance on the base classes.Finally, weshow that low-shot network expansion can be done with a very small memoryfootprint by using a compact generative model of the base classes training datawith only a negligible degradation relative to learning with the full training set.", "target": ["In dieser Arbeit befassen wir uns mit dem Problem des Lernens von Low-shot-Netzerweiterungen"]} +{"source": "People ask questions that are far richer, more informative, and more creative than current AI systems.We propose a neural program generation framework for modeling human question asking, which represents questions as formal programs and generates programs with an encoder-decoder based deep neural network.From extensive experiments using an information-search game, we show that our method can ask optimal questions in synthetic settings, and predict which questions humans are likely to ask in unconstrained settings.We also propose a novel grammar-based question generation framework trained with reinforcement learning, which is able to generate creative questions without supervised data.", "target": ["Wir stellen ein Modell des menschlichen Fragenstellens vor, das neuronale Netze und symbolische Programme kombiniert, die lernen können, gute Fragen mit oder ohne überwachte Beispiele zu generieren."]} +{"source": "The classification of images taken in special imaging environments except air is the first challenge in extending the applications of deep learning.We report on an UW-Net (Underwater Network), a new convolutional neural network (CNN) based network for underwater image classification.In this model, we simulate the visual correlation of background attention with image understanding for special environments, such as fog and underwater by constructing an inception-attention (I-A) module.The experimental results demonstrate that the proposed UW-Net achieves an accuracy of 99.3% on underwater image classification, which is significantly better than other image classification networks, such as AlexNet, InceptionV3, ResNet and Se-ResNet.Moreover, we demonstrate the proposed IA module can be used to boost the performance of the existing object recognition networks.By substituting the inception module with the I-A module, the Inception-ResnetV2 network achieves a 10.7% top1 error rate and a 0% top5 error rate on the subset of ILSVRC-2012, which further illustrates the function of the background attention in the image classifications.", "target": ["Ein visueller Verständigungsmechanismus für eine besondere Umgebung"]} +{"source": "Deep networks have achieved impressive results across a variety of important tasks.However, a known weakness is a failure to perform well when evaluated on data which differ from the training distribution, even if these differences are very small, as is the case with adversarial examples. We propose \\emph{Fortified Networks}, a simple transformation of existing networks, which “fortifies” the hidden layers in a deep network by identifying when the hidden states are off of the data manifold, and maps these hidden states back to parts of the data manifold where the network performs well.Our principal contribution is to show that fortifying these hidden states improves the robustness of deep networks and our experiments (i) demonstrate improved robustness to standard adversarial attacks in both black-box and white-box threat models; (ii) suggest that our improvements are not primarily due to the problem of deceptively good results due to degraded quality in the gradient signal (the gradient masking problem) and (iii) show the advantage of doing this fortification in the hidden layers instead of the input space. We demonstrate improvements in adversarial robustness on three datasets (MNIST, Fashion MNIST, CIFAR10), across several attack parameters, both white-box and black-box settings, and the most widely studied attacks (FGSM, PGD, Carlini-Wagner). We show that these improvements are achieved across a wide variety of hyperparameters.", "target": ["Besseres adversariales Training durch Lernen der Abbildung auf die Datenvielfalt mit Autoencodern in den verborgenen Zuständen."]} +{"source": "Neural networks could misclassify inputs that are slightly different from their training data, which indicates a small margin between their decision boundaries and the training dataset.In this work, we study the binary classification of linearly separable datasets and show that linear classifiers could also have decision boundaries that lie close to their training dataset if cross-entropy loss is used for training.In particular, we show that if the features of the training dataset lie in a low-dimensional affine subspace and the cross-entropy loss is minimized by using a gradient method, the margin between the training points and the decision boundary could be much smaller than the optimal value.This result is contrary to the conclusions of recent related works such as (Soudry et al., 2018), and we identify the reason for this contradiction.In order to improve the margin, we introduce differential training, which is a training paradigm that uses a loss function defined on pairs of points from each class.We show that the decision boundary of a linear classifier trained with differential training indeed achieves the maximum margin.The results reveal the use of cross-entropy loss as one of the hidden culprits of adversarial examples and introduces a new direction to make neural networks robust against them.", "target": ["Wir zeigen, dass die Minimierung des Cross-Entropy-Loss mit Hilfe einer Gradientenmethode zu einer sehr schlechten Marge führen kann, wenn die Merkmale des Datensatzes in einem niedrigdimensionalen Unterraum liegen."]} +{"source": "The concepts of unitary evolution matrices and associative memory have boosted the field of Recurrent Neural Networks (RNN) to state-of-the-art performance in a variety of sequential tasks. However, RNN still has a limited capacity to manipulate long-term memory. To bypass this weakness the most successful applications of RNN use external techniques such as attention mechanisms.In this paper we propose a novel RNN model that unifies the state-of-the-art approaches: Rotational Unit of Memory (RUM).The core of RUM is its rotational operation, which is, naturally, a unitary matrix, providing architectures with the power to learn long-term dependencies by overcoming the vanishing and exploding gradients problem. Moreover, the rotational unit also serves as associative memory.We evaluate our model on synthetic memorization, question answering and language modeling tasks. RUM learns the Copying Memory task completely and improves the state-of-the-art result in the Recall task. RUM’s performance in the bAbI Question Answering task is comparable to that of models with attention mechanism.We also improve the state-of-the-art result to 1.189 bits-per-character (BPC) loss in the Character Level Penn Treebank (PTB) task, which is to signify the applications of RUM to real-world sequential data.The universality of our construction, at the core of RNN, establishes RUM as a promising approach to language modeling, speech recognition and machine translation.", "target": ["Ein neuartiges RNN-Modell, das die derzeitigen Modelle in einer Reihe von sequentiellen Aufgaben deutlich übertrifft."]} +{"source": "While many recent advances in deep reinforcement learning rely on model-free methods, model-based approaches remain an alluring prospect for their potential to exploit unsupervised data to learn environment dynamics.One prospect is to pursue hybrid approaches, as in AlphaGo, which combines Monte-Carlo Tree Search (MCTS)—a model-based method—with deep-Q networks (DQNs)—a model-free method.MCTS requires generating rollouts, which is computationally expensive.In this paper, we propose to simulate roll-outs, exploiting the latest breakthroughs in image-to-image transduction, namely Pix2Pix GANs, to predict the dynamics of the environment.Our proposed algorithm, generative adversarial tree search (GATS), simulates rollouts up to a specified depth using both a GAN- based dynamics model and a reward predictor.GATS employs MCTS for planning over the simulated samples and uses DQN to estimate the Q-function at the leaf states.Our theoretical analysis establishes some favorable properties of GATS vis-a-vis the bias-variance trade-off and empirical results show that on 5 popular Atari games, the dynamics and reward predictors converge quickly to accurate solutions.However, GATS fails to outperform DQNs in 4 out of 5 games.Notably, in these experiments, MCTS has only short rollouts (up to tree depth 4), while previous successes of MCTS have involved tree depth in the hundreds.We present a hypothesis for why tree search with short rollouts can fail even given perfect modeling.", "target": ["Überraschend negative Ergebnisse bei modellbasierter + modelltiefer RL"]} +{"source": "We present a novel architecture of GAN for a disentangled representation learning.The new model architecture is inspired by Information Bottleneck (IB) theory thereby named IB-GAN.IB-GAN objective is similar to that of InfoGAN but has a crucial difference; a capacity regularization for mutual information is adopted, thanks to which the generator of IB-GAN can harness a latent representation in disentangled and interpretable manner.To facilitate the optimization of IB-GAN in practice, a new variational upper-bound is derived.With experiments on CelebA, 3DChairs, and dSprites datasets, we demonstrate that the visual quality of samples generated by IB-GAN is often better than those by β-VAEs.Moreover, IB-GAN achieves much higher disentanglement metrics score than β-VAEs or InfoGAN on the dSprites dataset.", "target": ["Inspiriert von der Theorie des Informationsengpasses schlagen wir eine neue GAN-Architektur für das Erlernen einer entflochtenen Darstellung vor."]} +{"source": "We investigate the training and performance of generative adversarial networks using the Maximum Mean Discrepancy (MMD) as critic, termed MMD GANs.As our main theoretical contribution, we clarify the situation with bias in GAN loss functions raised by recent work: we show that gradient estimators used in the optimization process for both MMD GANs and Wasserstein GANs are unbiased, but learning a discriminator based on samples leads to biased gradients for the generator parameters.We also discuss the issue of kernel choice for the MMD critic, and characterize the kernel corresponding to the energy distance used for the Cramér GAN critic.Being an integral probability metric, the MMD benefits from training strategies recently developed for Wasserstein GANs.In experiments, the MMD GAN is able to employ a smaller critic network than the Wasserstein GAN, resulting in a simpler and faster-training algorithm with matching performance.We also propose an improved measure of GAN convergence, the Kernel Inception Distance, and show how to use it to dynamically adapt learning rates during GAN training.", "target": ["Erklärung der Verzerrungen bei MMD-GANs; MMD-GANs arbeiten mit kleineren kritischen Netzen als WGAN-GPs; neue GAN-Bewertungsmetrik."]} +{"source": "We extend the Consensus Network framework to Transductive Consensus Network (TCN), a semi-supervised multi-modal classification framework, and identify its two mechanisms: consensus and classification.By putting forward three variants as ablation studies, we show both mechanisms should be functioning together.Overall, TCNs outperform or align with the best benchmark algorithms when only 20 to 200 labeled data points are available.", "target": ["Ein halbüberwachtes multimodales Klassifizierungssystem, TCN, das verschiedene Benchmarks übertrifft."]} +{"source": "Separating mixed distributions is a long standing challenge for machine learning and signal processing.Applications include: single-channel multi-speaker separation (cocktail party problem), singing voice separation and separating reflections from images.Most current methods either rely on making strong assumptions on the source distributions (e.g. sparsity, low rank, repetitiveness) or rely on having training samples of each source in the mixture.In this work, we tackle the scenario of extracting an unobserved distribution additively mixed with a signal from an observed (arbitrary) distribution.We introduce a new method: Neural Egg Separation - an iterative method that learns to separate the known distribution from progressively finer estimates of the unknown distribution.In some settings, Neural Egg Separation is initialization sensitive, we therefore introduce GLO Masking which ensures a good initialization.Extensive experiments show that our method outperforms current methods that use the same level of supervision and often achieves similar performance to full supervision.", "target": ["Eine iterative neuronale Methode zur Extraktion von Signalen, die nur gemischt mit anderen Signalen beobachtet werden"]} +{"source": "In health, machine learning is increasingly common, yet neural network embedding (representation) learning is arguably under-utilized for physiological signals. This inadequacy stands out in stark contrast to more traditional computer science domains, such as computer vision (CV), and natural language processing (NLP). For physiological signals, learning feature embeddings is a natural solution to data insufficiency caused by patient privacy concerns -- rather than share data, researchers may share informative embedding models (i.e., representation models), which map patient data to an output embedding. Here, we present the PHASE (PHysiologicAl Signal Embeddings) framework, which consists of three components: i) learning neural network embeddings of physiological signals, ii) predicting outcomes based on the learned embedding, and iii) interpreting the prediction results by estimating feature attributions in the \"stacked\" models (i.e., feature embedding model followed by prediction model). PHASE is novel in three ways: 1) To our knowledge, PHASE is the first instance of transferal of neural networks to create physiological signal embeddings.2) We present a tractable method to obtain feature attributions through stacked models. We prove that our stacked model attributions can approximate Shapley values -- attributions known to have desirable properties -- for arbitrary sets of models.3) PHASE was extensively tested in a cross-hospital setting including publicly available data. In our experiments, we show that PHASE significantly outperforms alternative embeddings -- such as raw, exponential moving average/variance, and autoencoder -- currently in use.Furthermore, we provide evidence that transferring neural network embedding/representation learners between distinct hospitals still yields performant embeddings and offer recommendations when transference is ineffective.", "target": ["Einbettung physiologischer Signale für die Vorhersageleistung und die Krankenhausübertragung mit einer allgemeinen Shapley-Wert-Interpretationsmethode für gestapelte Modelle."]} +{"source": "We consider the dictionary learning problem, where the aim is to model the given data as a linear combination of a few columns of a matrix known as a dictionary, where the sparse weights forming the linear combination are known as coefficients.Since the dictionary and coefficients, parameterizing the linear model are unknown, the corresponding optimization is inherently non-convex.This was a major challenge until recently, when provable algorithms for dictionary learning were proposed.Yet, these provide guarantees only on the recovery of the dictionary, without explicit recovery guarantees on the coefficients.Moreover, any estimation error in the dictionary adversely impacts the ability to successfully localize and estimate the coefficients.This potentially limits the utility of existing provable dictionary learning methods in applications where coefficient recovery is of interest.To this end, we develop NOODL: a simple Neurally plausible alternating Optimization-based Online Dictionary Learning algorithm, which recovers both the dictionary and coefficients exactly at a geometric rate, when initialized appropriately.Our algorithm, NOODL, is also scalable and amenable for large scale distributed implementations in neural architectures, by which we mean that it only involves simple linear and non-linear operations.Finally, we corroborate these theoretical results via experimental evaluation of the proposed algorithm with the current state-of-the-art techniques.", "target": ["Wir stellen einen beweisbaren Algorithmus vor, der beide Faktoren des Wörterbuch-Lernmodells exakt wiederherstellt."]} +{"source": "Real-world Relation Extraction (RE) tasks are challenging to deal with, either due to limited training data or class imbalance issues.In this work, we present Data Augmented Relation Extraction (DARE), a simple method to augment training data by properly finetuning GPT2 to generate examples for specific relation types.The generated training data is then used in combination with the gold dataset to train a BERT-based RE classifier.In a series of experiments we show the advantages of our method, which leads in improvements of up to 11 F1 score points compared to a strong baseline.Also, DARE achieves new state-of-the-art in three widely used biomedical RE datasets surpassing the previous best results by 4.7 F1 points on average.", "target": ["Datenerweiterte Relationsextraktion mit GPT-2"]} +{"source": "This paper addresses the problem of representing a system's belief using multi-variate normal distributions (MND) where the underlying model is based on a deep neural network (DNN).The major challenge with DNNs is the computational complexity that is needed to obtain model uncertainty using MNDs.To achieve a scalable method, we propose a novel approach that expresses the parameter posterior in sparse information form.Our inference algorithm is based on a novel Laplace Approximation scheme, which involves a diagonal correction of the Kronecker-factored eigenbasis.As this makes the inversion of the information matrix intractable - an operation that is required for full Bayesian analysis, we devise a low-rank approximation of this eigenbasis and a memory-efficient sampling scheme.We provide both a theoretical analysis and an empirical evaluation on various benchmark data sets, showing the superiority of our approach over existing methods.", "target": ["Ein approximativer Inferenzalgorithmus für Deep Learning"]} +{"source": "The ever-increasing size of modern datasets combined with the difficulty of obtaining label information has made semi-supervised learning of significant practical importance in modern machine learning applications.In comparison to supervised learning, the key difficulty in semi-supervised learning is how to make full use of the unlabeled data.In order to utilize manifold information provided by unlabeled data, we propose a novel regularization called the tangent-normal adversarial regularization, which is composed by two parts.The two parts complement with each other and jointly enforce the smoothness along two different directions that are crucial for semi-supervised learning.One is applied along the tangent space of the data manifold, aiming to enforce local invariance of the classifier on the manifold, while the other is performed on the normal space orthogonal to the tangent space, intending to impose robustness on the classifier against the noise causing the observed data deviating from the underlying data manifold. Both of the two regularizers are achieved by the strategy of virtual adversarial training.Our method has achieved state-of-the-art performance on semi-supervised learning tasks on both artificial dataset and practical datasets.", "target": ["Wir schlagen eine neuartige Regularisierungsstrategie vor, die auf adversarialem Training basiert und die Leistung von semi-supervised learning deutlich verbessern kann."]} +{"source": "Universal approximation property of neural networks is one of the motivations to use these models in various real-world problems.However, this property is not the only characteristic that makes neural networks unique as there is a wide range of other approaches with similar property.Another characteristic which makes these models interesting is that they can be trained with the backpropagation algorithm which allows an efficient gradient computation and gives these universal approximators the ability to efficiently learn complex manifolds from a large amount of data in different domains.Despite their abundant use in practice, neural networks are still not well understood and a broad range of ongoing research is to study the interpretability of neural networks.On the other hand, topological data analysis (TDA) relies on strong theoretical framework of (algebraic) topology along with other mathematical tools for analyzing possibly complex datasets.In this work, we leverage a universal approximation theorem originating from algebraic topology to build a connection between TDA and common neural network training framework.We introduce the notion of automatic subdivisioning and devise a particular type of neural networks for regression tasks: Simplicial Complex Networks (SCNs).SCN's architecture is defined with a set of bias functions along with a particular policy during the forward pass which alternates the common architecture search framework in neural networks.We believe the view of SCNs can be used as a step towards building interpretable deep learning models.Finally, we verify its performance on a set of regression problems.", "target": ["Eine neuartige Methode für überwachtes Lernen durch Unterteilung des Eingaberaums in Verbindung mit Funktionsannäherung."]} +{"source": "The goal of network representation learning is to learn low-dimensional node embeddings that capture the graph structure and are useful for solving downstream tasks.However, despite the proliferation of such methods there is currently no study of their robustness to adversarial attacks.We provide the first adversarial vulnerability analysis on the widely used family of methods based on random walks.We derive efficient adversarial perturbations that poison the network structure and have a negative effect on both the quality of the embeddings and the downstream tasks.We further show that our attacks are transferable since they generalize to many models, and are successful even when the attacker is restricted.", "target": ["Angriffe auf unüberwachte Knoteneinbettungen auf der Grundlage der Eigenwert-Störungstheorie."]} +{"source": "Hierarchical reinforcement learning methods offer a powerful means of planning flexible behavior in complicated domains.However, learning an appropriate hierarchical decomposition of a domain into subtasks remains a substantial challenge.We present a novel algorithm for subtask discovery, based on the recently introduced multitask linearly-solvable Markov decision process (MLMDP) framework.The MLMDP can perform never-before-seen tasks by representing them as a linear combination of a previously learned basis set of tasks.In this setting, the subtask discovery problem can naturally be posed as finding an optimal low-rank approximation of the set of tasks the agent will face in a domain.We use non-negative matrix factorization to discover this minimal basis set of tasks, and show that the technique learns intuitive decompositions in a variety of domains.Our method has several qualitatively desirable features: it is not limited to learning subtasks with single goal states, instead learning distributed patterns of preferred states; it learns qualitatively different hierarchical decompositions in the same domain depending on the ensemble of tasks the agent will face; and it may be straightforwardly iterated to obtain deeper hierarchical decompositions.", "target": ["Wir stellen einen neuartigen Algorithmus für die hierarchische Erkennung von Teilaufgaben vor, der den Rahmen des linearen Markov-Entscheidungsprozesses für mehrere Aufgaben nutzt."]} +{"source": "This paper introduces a framework for solving combinatorial optimization problems by learning from input-output examples of optimization problems.We introduce a new memory augmented neural model in which the memory is not resettable (i.e the information stored in the memory after processing an input example is kept for the next seen examples).We used deep reinforcement learning to train a memory controller agent to store useful memories.Our model was able to outperform hand-crafted solver on Binary Linear Programming (Binary LP).The proposed model is tested on different Binary LP instances with large number of variables (up to 1000 variables) and constrains (up to 700 constrains).", "target": ["Wir schlagen ein Gedächtnisnetzwerkmodell vor, um binäre LP-Instanzen zu lösen, bei denen die Gedächtnisinformationen für eine langfristige Verwendung aufbewahrt werden."]} +{"source": "Sequence-to-sequence (Seq2Seq) models with attention have excelled at tasks which involve generating natural language sentences such as machine translation, image captioning and speech recognition.Performance has further been improved by leveraging unlabeled data, often in the form of a language model.In this work, we present the Cold Fusion method, which leverages a pre-trained language model during training, and show its effectiveness on the speech recognition task.We show that Seq2Seq models with Cold Fusion are able to better utilize language information enjoyingi) faster convergence and better generalization, andii) almost complete transfer to a new domain while using less than 10% of the labeled training data.", "target": ["Wir stellen eine neuartige Methode vor, um Seq2Seq-Modelle mit Sprachmodellen zu trainieren, die schneller konvergieren, besser verallgemeinern und mit weniger als 10 % der gelabelten Daten fast vollständig auf eine neue Domäne übertragen werden können."]} +{"source": "A central capability of intelligent systems is the ability to continuously build upon previous experiences to speed up and enhance learning of new tasks.Two distinct research paradigms have studied this question.Meta-learning views this problem as learning a prior over model parameters that is amenable for fast adaptation on a new task, but typically assumes the set of tasks are available together as a batch.In contrast, online (regret based) learning considers a sequential setting in which problems are revealed one after the other, but conventionally train only a single model without any task-specific adaptation.This work introduces an online meta-learning setting, which merges ideas from both the aforementioned paradigms to better capture the spirit and practice of continual lifelong learning.We propose the follow the meta leader (FTML) algorithm which extends the MAML algorithm to this setting.Theoretically, this work provides an O(logT) regret guarantee for the FTML algorithm.Our experimental evaluation on three different large-scale tasks suggest that the proposed algorithm significantly outperforms alternatives based on traditional online learning approaches.", "target": ["Wir führen die Online-Meta-Learning-Problemstellung ein, um den Geist und die Praxis des kontinuierlichen lebenslangen Lernens besser zu erfassen."]} +{"source": "We investigate the extent to which individual attention heads in pretrained transformer language models, such as BERT and RoBERTa, implicitly capture syntactic dependency relations.We employ two methods—taking the maximum attention weight and computing the maximum spanning tree—to extract implicit dependency relations from the attention weights of each layer/head, and compare them to the ground-truth Universal Dependency (UD) trees.We show that, for some UD relation types, there exist heads that can recover the dependency type significantly better than baselines on parsed English text, suggesting that some self-attention heads act as a proxy for syntactic structure.We also analyze BERT fine-tuned on two datasets—the syntax-oriented CoLA and the semantics-oriented MNLI—to investigate whether fine-tuning affects the patterns of their self-attention, but we do not observe substantial differences in the overall dependency relations extracted using our methods.Our results suggest that these models have some specialist attention heads that track individual dependency types, but no generalist head that performs holistic parsing significantly better than a trivial baseline, and that analyzing attention weights directly may not reveal much of the syntactic knowledge that BERT-style models are known to learn.", "target": ["Die Aufmerksamkeitsgewichte legen nicht vollständig offen, was BERT über die Syntax weiß."]} +{"source": "State-of-the-art face super-resolution methods employ deep convolutional neural networks to learn a mapping between low- and high-resolution facial patterns by exploring local appearance knowledge.However, most of these methods do not well exploit facial structures and identity information, and struggle to deal with facial images that exhibit large pose variation and misalignment.In this paper, we propose a novel face super-resolution method that explicitly incorporates 3D facial priors which grasp the sharp facial structures.Firstly, the 3D face rendering branch is set up to obtain 3D priors of salient facial structures and identity knowledge.Secondly, the Spatial Attention Mechanism is used to better exploit this hierarchical information (i.e. intensity similarity, 3D facial structure, identity content) for the super-resolution problem.Extensive experiments demonstrate that the proposed algorithm achieves superior face super-resolution results and outperforms the state-of-the-art.", "target": ["Wir schlagen eine neuartige Methode zur Gesichtsauflösung vor, die explizit 3D-Gesichtsprioritäten einbezieht, die die scharfen Gesichtsstrukturen erfassen."]} +{"source": "We present Cross-View Training (CVT), a simple but effective method for deep semi-supervised learning.On labeled examples, the model is trained with standard cross-entropy loss.On an unlabeled example, the model first performs inference (acting as a \"teacher\") to produce soft targets.The model then learns from these soft targets (acting as a ``\"student\").We deviate from prior work by adding multiple auxiliary student prediction layers to the model.The input to each student layer is a sub-network of the full model that has a restricted view of the input (e.g., only seeing one region of an image).The students can learn from the teacher (the full model) because the teacher sees more of each example.Concurrently, the students improve the quality of the representations used by the teacher as they learn to make predictions with limited data.When combined with Virtual Adversarial Training, CVT improves upon the current state-of-the-art on semi-supervised CIFAR-10 and semi-supervised SVHN.We also apply CVT to train models on five natural language processing tasks using hundreds of millions of sentences of unlabeled data.On all tasks CVT substantially outperforms supervised learning alone, resulting in models that improve upon or are competitive with the current state-of-the-art.", "target": ["Selbsttraining mit verschiedenen Ansichten der Eingabe liefert hervorragende Ergebnisse für die halbüberwachte Bilderkennung, das Sequenz-Tagging und das Parsing von Abhängigkeiten."]} +{"source": "I show how it can be beneficial to express Metropolis accept/reject decisions in terms of comparison with a uniform [0,1] value, and to then update this uniform value non-reversibly, as part of the Markov chain state, rather than sampling it independently each iteration.This provides a small improvement for random walk Metropolis and Langevin updates in high dimensions. It produces a larger improvement when using Langevin updates with persistent momentum, giving performance comparable to that of Hamiltonian Monte Carlo (HMC) with long trajectories. This is of significance when some variables are updated by other methods, since if HMC is used, these updates can be done only between trajectories, whereas they can be done more often with Langevin updates. This is seen for a Bayesian neural network model, in which connection weights are updated by persistent Langevin or HMC, while hyperparameters are updated by Gibbs sampling.", "target": ["Eine nicht umkehrbare Entscheidung über die Annahme oder Ablehnung kann von Vorteil sein."]} +{"source": "We introduce ES-MAML, a new framework for solving the model agnostic meta learning (MAML) problem based on Evolution Strategies (ES).Existing algorithms for MAML are based on policy gradients, and incur significant difficulties when attempting to estimate second derivatives using backpropagation on stochastic policies.We show how ES can be applied to MAML to obtain an algorithm which avoids the problem of estimating second derivatives, and is also conceptually simple and easy to implement.Moreover, ES-MAML can handle new types of nonsmooth adaptation operators, and other techniques for improving performance and estimation of ES methods become applicable.We show empirically that ES-MAML is competitive with existing methods and often yields better adaptation with fewer queries.", "target": ["Wir bieten einen neuen Rahmen für MAML in der ES/Blackbox-Umgebung und zeigen, dass er deterministische und lineare Strategien, bessere Exploration und nicht-differenzierbare Anpassungsoperatoren ermöglicht."]} +{"source": "Transforming one probability distribution to another is a powerful tool in Bayesian inference and machine learning.Some prominent examples are constrained-to-unconstrained transformations of distributions for use in Hamiltonian Monte-Carlo and constructing flexible and learnable densities such as normalizing flows.We present Bijectors.jl, a software package for transforming distributions implemented in Julia, available at github.com/TuringLang/Bijectors.jl.The package provides a flexible and composable way of implementing transformations of distributions without being tied to a computational framework. We demonstrate the use of Bijectors.jl on improving variational inference by encoding known statistical dependencies into the variational posterior using normalizing flows, providing a general approach to relaxing the mean-field assumption usually made in variational inference.", "target": ["Wir stellen einen Software-Rahmen für die Transformation von Verteilungen vor und demonstrieren dessen Flexibilität bei der Lockerung von Mean-Field-Annahmen in der Variationsinferenz durch die Verwendung von Kopplungsflüssen zur Replikation von Strukturen aus dem generativen Zielmodell."]} +{"source": "Multi-domain learning (MDL) aims at obtaining a model with minimal average risk across multiple domains.Our empirical motivation is automated microscopy data, where cultured cells are imaged after being exposed to known and unknown chemical perturbations, and each dataset displays significant experimental bias.This paper presents a multi-domain adversarial learning approach, MuLANN, to leverage multiple datasets with overlapping but distinct class sets, in a semi-supervised setting.Our contributions include:i) a bound on the average- and worst-domain risk in MDL, obtained using the H-divergence;ii) a new loss to accommodate semi-supervised multi-domain learning and domain adaptation;iii) the experimental validation of the approach, improving on the state of the art on two standard image benchmarks, and a novel bioimage dataset, Cell.", "target": ["Adversarial Domain adaptation und Multi-domain learning: Ein neuer Loss um multi- und single-domain Klassen in halbüberwachten Umgebungen zu steuern."]} +{"source": "We introduce our Distribution Regression Network (DRN) which performs regression from input probability distributions to output probability distributions.Compared to existing methods, DRN learns with fewer model parameters and easily extends to multiple input and multiple output distributions.On synthetic and real-world datasets, DRN performs similarly or better than the state-of-the-art.Furthermore, DRN generalizes the conventional multilayer perceptron (MLP).In the framework of MLP, each node encodes a real number, whereas in DRN, each node encodes a probability distribution.", "target": ["Ein lernendes Netz, das den MLP-Rahmen verallgemeinert, um eine Regression von Verteilung zu Verteilung durchzuführen"]} +{"source": "Existing sequence prediction methods are mostly concerned with time-independent sequences, in which the actual time span between events is irrelevant and the distance between events is simply the difference between their order positions in the sequence.While this time-independent view of sequences is applicable for data such as natural languages, e.g., dealing with words in a sentence, it is inappropriate and inefficient for many real world events that are observed and collected at unequally spaced points of time as they naturally arise, e.g., when a person goes to a grocery store or makes a phone call.The time span between events can carry important information about the sequence dependence of human behaviors.In this work, we propose a set of methods for using time in sequence prediction.Because neural sequence models such as RNN are more amenable for handling token-like input, we propose two methods for time-dependent event representation, based on the intuition on how time is tokenized in everyday life and previous work on embedding contextualization.We also introduce two methods for using next event duration as regularization for training a sequence prediction model.We discuss these methods based on recurrent neural nets.We evaluate these methods as well as baseline models on five datasets that resemble a variety of sequence prediction tasks.The experiments revealed that the proposed methods offer accuracy gain over baseline models in a range of settings.", "target": ["Vorgeschlagene Methoden für zeitabhängige Ereignisdarstellung und Regularisierung für Sequenzvorhersage; Evaluierung dieser Methoden an fünf Datensätzen, die eine Reihe von Sequenzvorhersageaufgaben beinhalten."]} +{"source": "Off-policy reinforcement learning algorithms promise to be applicable in settings where only a fixed data-set (batch) of environment interactions is available and no new experience can be acquired.This property makes these algorithms appealing for real world problems such as robot control.In practice, however, standard off-policy algorithms fail in the batch setting for continuous control.In this paper, we propose a simple solution to this problem.It admits the use of data generated by arbitrary behavior policies and uses a learned prior -- the advantage-weighted behavior model (ABM) -- to bias the RL policy towards actions that have previously been executed and are likely to be successful on the new task.Our method can be seen as an extension of recent work on batch-RL that enables stable learning from conflicting data-sources.We find improvements on competitive baselines in a variety of RL tasks -- including standard continuous control benchmarks and multi-task learning for simulated and real-world robots.", "target": ["Wir entwickeln eine Methode für stabiles Offline Reinforcement Learning aus aufgezeichneten Daten. Der Schlüssel dazu ist die Regularisierung der RL-Politik in Richtung eines gelernten \"vorteilhaft gewichteten\" Modells der Daten."]} +{"source": "One of the main challenges in applying graph convolutional neural networks on gene-interaction data is the lack of understanding of the vector space to which they belong and also the inherent difficulties involved in representing those interactions on a significantly lower dimension, viz Euclidean spaces.The challenge becomes more prevalent when dealing with various types of heterogeneous data.We introduce a systematic, generalized method, called iSOM-GSN, used to transform ``multi-omic'' data with higher dimensions onto a two-dimensional grid.Afterwards, we apply a convolutional neural network to predict disease states of various types.Based on the idea of Kohonen's self-organizing map, we generate a two-dimensional grid for each sample for a given set of genes that represent a gene similarity network. We have tested the model to predict breast and prostate cancer using gene expression, DNA methylation and copy number alteration, yielding prediction accuracies in the 94-98% range for tumor stages of breast cancer and calculated Gleason scores of prostate cancer with just 11 input genes for both cases.The scheme not only outputs nearly perfect classification accuracy, but also provides an enhanced scheme for representation learning, visualization, dimensionality reduction, and interpretation of the results.", "target": ["In diesem Beitrag wird ein Deep-Learning-Modell vorgestellt, das selbstorganisierende Zuordnungen und neuronale Convolutional Networks für das Repräsentationslernen von Multi-omics-Daten kombiniert."]} +{"source": "State-of-the-art performances on language comprehension tasks are achieved by huge language models pre-trained on massive unlabeled text corpora, with very light subsequent fine-tuning in a task-specific supervised manner.It seems the pre-training procedure learns a very good common initialization for further training on various natural language understanding tasks, such that only few steps need to be taken in the parameter space to learn each task.In this work, using Bidirectional Encoder Representations from Transformers (BERT) as an example, we verify this hypothesis by showing that task-specific fine-tuned language models are highly close in parameter space to the pre-trained one.Taking advantage of such observations, we further show that the fine-tuned versions of these huge models, having on the order of $10^8$ floating-point parameters, can be made very computationally efficient.First, fine-tuning only a fraction of critical layers suffices.Second, fine-tuning can be adequately performed by learning a binary multiplicative mask on pre-trained weights, \\textit{i.e.} by parameter-sparsification.As a result, with a single effort, we achieve three desired outcomes: (1) learning to perform specific tasks, (2) saving memory by storing only binary masks of certain layers for each task, and (3) saving compute on appropriate hardware by performing sparse operations with model parameters.", "target": ["Sparsifizierung als Feinabstimmung von Sprachmodellen"]} +{"source": "We present a simple and effective algorithm designed to address the covariate shift problem in imitation learning.It operates by training an ensemble of policies on the expert demonstration data, and using the variance of their predictions as a cost which is minimized with RL together with a supervised behavioral cloning cost.Unlike adversarial imitation methods, it uses a fixed reward function which is easy to optimize.We prove a regret bound for the algorithm in the tabular setting which is linear in the time horizon multiplied by a coefficient which we show to be low for certain problems in which behavioral cloning fails.We evaluate our algorithm empirically across multiple pixel-based Atari environments and continuous control tasks, and show that it matches or significantly outperforms behavioral cloning and generative adversarial imitation learning.", "target": ["Verfahren zur Behandlung von Kovariatenverschiebungen beim Imitation Learning unter Verwendung von Ensemble-Unsicherheit"]} +{"source": "We present and discuss a simple image preprocessing method for learning disentangled latent factors. In particular, we utilize the implicit inductive bias contained in features from networks pretrained on the ImageNet database. We enhance this bias by explicitly fine-tuning such pretrained networks on tasks useful for the NeurIPS2019 disentanglement challenge, such as angle and position estimation or color classification.Furthermore, we train a VAE on regionally aggregate feature maps, and discuss its disentanglement performance using metrics proposed in recent literature.", "target": ["Wir verwenden eine überwachte Feinabstimmung der Merkmalsvektoren, um die Übertragung von der Simulation auf die reale Welt zu verbessern."]} +{"source": "A reinforcement learning agent that needs to pursue different goals across episodes requires a goal-conditional policy.In addition to their potential to generalize desirable behavior to unseen goals, such policies may also enable higher-level planning based on subgoals.In sparse-reward environments, the capacity to exploit information about the degree to which an arbitrary goal has been achieved while another goal was intended appears crucial to enable sample efficient learning.However, reinforcement learning agents have only recently been endowed with such capacity for hindsight.In this paper, we demonstrate how hindsight can be introduced to policy gradient methods, generalizing this idea to a broad class of successful algorithms.Our experiments on a diverse selection of sparse-reward environments show that hindsight leads to a remarkable increase in sample efficiency.", "target": ["Wir führen die Fähigkeit ein, Informationen darüber zu nutzen, inwieweit ein beliebiges Ziel erreicht wurde, während ein anderes Ziel für die Regeln der Gradientenmethoden vorgesehen war."]} +{"source": "Computer vision has undergone a dramatic revolution in performance, driven in large part through deep features trained on large-scale supervised datasets.However, much of these improvements have focused on static image analysis; video understanding has seen rather modest improvements.Even though new datasets and spatiotemporal models have been proposed, simple frame-by-frame classification methods often still remain competitive.We posit that current video datasets are plagued with implicit biases over scene and object structure that can dwarf variations in temporal structure.In this work, we build a video dataset with fully observable and controllable object and scene bias, and which truly requires spatiotemporal understanding in order to be solved.Our dataset, named CATER, is rendered synthetically using a library of standard 3D objects, and tests the ability to recognize compositions of object movements that require long-term reasoning.In addition to being a challenging dataset, CATER also provides a plethora of diagnostic tools to analyze modern spatiotemporal video architectures by being completely observable and controllable.Using CATER, we provide insights into some of the most recent state of the art deep video architectures.", "target": ["Wir schlagen einen neuen Benchmark für das Verstehen von Videos vor, dessen Aufgaben im Gegensatz zu den meisten bestehenden Videodatensätzen von vornherein zeitliche Schlussfolgerungen erfordern."]} +{"source": "We address the efficiency issues caused by the straggler effect in the recently emerged federated learning, which collaboratively trains a model on decentralized non-i.i.d. (non-independent and identically distributed) data across massive worker devices without exchanging training data in the unreliable and heterogeneous networks.We propose a novel two-stage analysis on the error bounds of general federated learning, which provides practical insights into optimization.As a result, we propose a novel easy-to-implement federated learning algorithm that uses asynchronous settings and strategies to control discrepancies between the global model and delayed models and adjust the number of local epochs with the estimation of staleness to accelerate convergence and resist performance deterioration caused by stragglers.Experiment results show that our algorithm converges fast and robust on the existence of massive stragglers.", "target": ["Wir schlagen einen effizienten und robusten asynchronen Algorithmus für föderiertes Lernen vor, der die Existenz von Nachzüglern berücksichtigt."]} +{"source": "Long Short-Term Memory (LSTM) units have the ability to memorise and use long-term dependencies between inputs to generate predictions on time series data.We introduce the concept of modifying the cell state (memory) of LSTMs using rotation matrices parametrised by a new set of trainable weights.This addition shows significant increases of performance on some of the tasks from the bAbI dataset.", "target": ["Das Hinzufügen eines neuen Satzes von Gewichten zum LSTM, die den Zellspeicher drehen, verbessert die Leistung bei einigen bAbI-Aufgaben."]} +{"source": "We address the problem of marginal inference for an exponential family defined over the set of permutation matrices.This problem is known to quickly become intractable as the size of the permutation increases, since its involves the computation of the permanent of a matrix, a #P-hard problem.We introduce Sinkhorn variational marginal inference as a scalable alternative, a method whose validity is ultimately justified by the so-called Sinkhorn approximation of the permanent.We demonstrate the efectiveness of our method in the problem of probabilistic identification of neurons in the worm C.elegans", "target": ["Neue Methodik zur variablen marginalen Inferenz von Permutationen auf der Grundlage des Sinkhorn-Algorithmus, angewandt auf die probabilistische Identifizierung von Neuronen"]} +{"source": "The robustness of neural networks to adversarial examples has received great attention due to security implications.Despite various attack approaches to crafting visually imperceptible adversarial examples, little has been developed towards a comprehensive measure of robustness.In this paper, we provide theoretical justification for converting robustness analysis into a local Lipschitz constant estimation problem, and propose to use the Extreme Value Theory for efficient evaluation.Our analysis yields a novel robustness metric called CLEVER, which is short for Cross Lipschitz Extreme Value for nEtwork Robustness.The proposed CLEVER score is attack-agnostic and is computationally feasible for large neural networks.Experimental results on various networks, including ResNet, Inception-v3 and MobileNet, show that(i) CLEVER is aligned with the robustness indication measured by the $\\ell_2$ and $\\ell_\\infty$ norms of adversarial examples from powerful attacks, and(ii) defended networks using defensive distillation or bounded ReLU indeed give better CLEVER scores.To the best of our knowledge, CLEVER is the first attack-independent robustness metric that can be applied to any neural network classifiers.", "target": ["Wir schlagen die erste angriffsunabhängige Robustheitsmetrik, a.k.a. CLEVER, vor, die auf jeden Klassifikator eines neuronalen Netzes angewendet werden kann."]} +{"source": "Multi-agent collaboration is required by numerous real-world problems.Although distributed setting is usually adopted by practical systems, local range communication and information aggregation still matter in fulfilling complex tasks.For multi-agent reinforcement learning, many previous studies have been dedicated to design an effective communication architecture.However, existing models usually suffer from an ossified communication structure, e.g., most of them predefine a particular communication mode by specifying a fixed time frequency and spatial scope for agents to communicate regardless of necessity.Such design is incapable of dealing with multi-agent scenarios that are capricious and complicated, especially when only partial information is available.Motivated by this, we argue that the solution is to build a spontaneous and self-organizing communication (SSoC) learning scheme.By treating the communication behaviour as an explicit action, SSoC learns to organize communication in an effective and efficient way.Particularly, it enables each agent to spontaneously decide when and who to send messages based on its observed states.In this way, a dynamic inter-agent communication channel is established in an online and self-organizing manner.The agents also learn how to adaptively aggregate the received messages and its own hidden states to execute actions.Various experiments have been conducted to demonstrate that SSoC really learns intelligent message passing among agents located far apart.With such agile communications, we observe that effective collaboration tactics emerge which have not been mastered by the compared baselines.", "target": ["In diesem Beitrag wird ein Lernverfahren für spontane und selbstorganisierende Kommunikation (SSoC) für Multi-Agenten-RL-Aufgaben vorgeschlagen."]} +{"source": "We study the BERT language representation model and the sequence generation model with BERT encoder for multi-label text classification task.We experiment with both models and explore their special qualities for this setting.We also introduce and examine experimentally a mixed model, which is an ensemble of multi-label BERT and sequence generating BERT models.Our experiments demonstrated that BERT-based models and the mixed model, in particular, outperform current baselines in several metrics achieving state-of-the-art results on three well-studied multi-label classification datasets with English texts and two private Yandex Taxi datasets with Russian texts.", "target": ["Über die Verwendung von BERT als Encoder für die sequentielle Vorhersage von Etiketten bei der Klassifizierung von Texten mit mehreren Etiketten"]} +{"source": "Click Through Rate (CTR) prediction is a critical task in industrial applications, especially for online social and commerce applications.It is challenging to find a proper way to automatically discover the effective cross features in CTR tasks.We propose a novel model for CTR tasks, called Deep neural networks with Encoder enhanced Factorization Machine (DeepEnFM).Instead of learning the cross features directly, DeepEnFM adopts the Transformer encoder as a backbone to align the feature embeddings with the clues of other fields.The embeddings generated from encoder are beneficial for the further feature interactions.Particularly, DeepEnFM utilizes a bilinear approach to generate different similarity functions with respect to different field pairs.Furthermore, the max-pooling method makes DeepEnFM feasible to capture both the supplementary and suppressing information among different attention heads.Our model is validated on the Criteo and Avazu datasets, and achieves state-of-art performance.", "target": ["DNN und Encoder verbesserte FM mit bilinearer Aufmerksamkeit und Max-Pooling für CTR"]} +{"source": "For autonomous agents to successfully operate in the real world, the ability to anticipate future scene states is a key competence.In real-world scenarios, future states become increasingly uncertain and multi-modal, particularly on long time horizons.Dropout based Bayesian inference provides a computationally tractable, theoretically well grounded approach to learn different hypotheses/models to deal with uncertain futures and make predictions that correspond well to observations -- are well calibrated.However, it turns out that such approaches fall short to capture complex real-world scenes, even falling behind in accuracy when compared to the plain deterministic approaches.This is because the used log-likelihood estimate discourages diversity.In this work, we propose a novel Bayesian formulation for anticipating future scene states which leverages synthetic likelihoods that encourage the learning of diverse models to accurately capture the multi-modal nature of future scene states.We show that our approach achieves accurate state-of-the-art predictions and calibrated probabilities through extensive experiments for scene anticipation on Cityscapes dataset.Moreover, we show that our approach generalizes across diverse tasks such as digit generation and precipitation forecasting.", "target": ["Dropout-basierte Bayes'sche Inferenz wird erweitert, um mit Multimodalität umgehen zu können, und wird anhand von Aufgaben zur Szenenvorhersage evaluiert."]} +{"source": "Conditional generative adversarial networks (cGAN) have led to large improvements in the task of conditional image generation, which lies at the heart of computer vision.The major focus so far has been on performance improvement, while there has been little effort in making cGAN more robust to noise.The regression (of the generator) might lead to arbitrarily large errors in the output, which makes cGAN unreliable for real-world applications.In this work, we introduce a novel conditional GAN model, called RoCGAN, which leverages structure in the target space of the model to address the issue.Our model augments the generator with an unsupervised pathway, which promotes the outputs of the generator to span the target manifold even in the presence of intense noise.We prove that RoCGAN share similar theoretical properties as GAN and experimentally verify that our model outperforms existing state-of-the-art cGAN architectures by a large margin in a variety of domains including images from natural scenes and faces.", "target": ["Wir stellen eine neue Art von bedingtem GAN vor, das darauf abzielt, die Struktur im Zielraum des Generators zu nutzen. Wir erweitern den Generator mit einem neuen, unbeaufsichtigten Pfad, um die Zielstruktur zu lernen."]} +{"source": "Though deep neural networks have achieved the state of the art performance in visual classification, recent studies have shown that they are all vulnerable to the attack of adversarial examples.To solve the problem, some regularization adversarial training methods, constraining the output label or logit, have been studied.In this paper, we propose a novel regularized adversarial training framework ATLPA,namely Adversarial Tolerant Logit Pairing with Attention.Instead of constraining a hard distribution (e.g., one-hot vectors or logit) in adversarial training, ATLPA uses Tolerant Logit which consists of confidence distribution on top-k classes and captures inter-class similarities at the image level.Specifically, in addition to minimizing the empirical loss, ATLPA encourages attention map for pairs of examples to be similar.When applied to clean examples and their adversarial counterparts, ATLPA improves accuracy on adversarial examples over adversarial training.We evaluate ATLPA with the state of the art algorithms, the experiment results show that our method outperforms these baselines with higher accuracy.Compared with previous work, our work is evaluated under highly challenging PGD attack: the maximum perturbation $\\epsilon$ is 64 and 128 with 10 to 200 attack iterations.", "target": ["In diesem Papier schlagen wir ein neuartiges regularisiertes adversariales Trainingssystem ATLPA vor, nämlich Adversarial Tolerant Logit Pairing with Attention."]} +{"source": "A fundamental trait of intelligence is the ability to achieve goals in the face of novel circumstances.In this work, we address one such setting which requires solving a task with a novel set of actions.Empowering machines with this ability requires generalization in the way an agent perceives its available actions along with the way it uses these actions to solve tasks.Hence, we propose a framework to enable generalization over both these aspects: understanding an action’s functionality, and using actions to solve tasks through reinforcement learning.Specifically, an agent interprets an action’s behavior using unsupervised representation learning over a collection of data samples reflecting the diverse properties of that action.We employ a reinforcement learning architecture which works over these action representations, and propose regularization metrics essential for enabling generalization in a policy.We illustrate the generalizability of the representation learning method and policy, to enable zero-shot generalization to previously unseen actions on challenging sequential decision-making environments.Our results and videos can be found at sites.google.com/view/action-generalization/", "target": ["Wir befassen uns mit dem Problem der Generalisierung des Reinforcement Learnings auf ungesehenen Actionräumen."]} +{"source": "Temporal point processes are the dominant paradigm for modeling sequences of events happening at irregular intervals.The standard way of learning in such models is by estimating the conditional intensity function. However, parameterizing the intensity function usually incurs several trade-offs.We show how to overcome the limitations of intensity-based approaches by directly modeling the conditional distribution of inter-event times. We draw on the literature on normalizing flows to design models that are flexible and efficient.We additionally propose a simple mixture model that matches the flexibility of flow-based models, but also permits sampling and computing moments in closed form. The proposed models achieve state-of-the-art performance in standard prediction tasks and are suitable for novel applications, such as learning sequence embeddings and imputing missing data.", "target": ["Lernen in zeitlichen Punktprozessen durch Modellierung der bedingten Dichte, nicht der bedingten Intensität."]} +{"source": "We propose a novel yet simple neural network architecture for topic modelling.The method is based on training an autoencoder structure where the bottleneck represents the space of the topics distribution and the decoder outputs represent the space of the words distributions over the topics.We exploit an auxiliary decoder to prevent mode collapsing in our model. A key feature for an effective topic modelling method is having sparse topics and words distributions, where there is a trade-off between the sparsity level of topics and words.This feature is implemented in our model by L-2 regularization and the model hyperparameters take care of the trade-off. We show in our experiments that our model achieves competitive results compared to the state-of-the-art deep models for topic modelling, despite its simple architecture and training procedure.The “New York Times” and “20 Newsgroups” datasets are used in the experiments.", "target": ["Ein tiefes Modell für die Themenmodellierung"]} +{"source": "Voice Conversion (VC) is a task of converting perceived speaker identity from a source speaker to a particular target speaker.Earlier approaches in the literature primarily find a mapping between the given source-target speaker-pairs.Developing mapping techniques for many-to-many VC using non-parallel data, including zero-shot learning remains less explored areas in VC.Most of the many-to-many VC architectures require training data from all the target speakers for whom we want to convert the voices.In this paper, we propose a novel style transfer architecture, which can also be extended to generate voices even for target speakers whose data were not used in the training (i.e., case of zero-shot learning).In particular, propose Adaptive Generative Adversarial Network (AdaGAN), new architectural training procedure help in learning normalized speaker-independent latent representation, which will be used to generate speech with different speaking styles in the context of VC.We compare our results with the state-of-the-art StarGAN-VC architecture.In particular, the AdaGAN achieves 31.73%, and 10.37% relative improvement compared to the StarGAN in MOS tests for speech quality and speaker similarity, respectively.The key strength of the proposed architectures is that it yields these results with less computational complexity.AdaGAN is 88.6% less complex than StarGAN-VC in terms of FLoating Operation Per Second (FLOPS), and 85.46% less complex in terms of trainable parameters.", "target": ["Neuartiges adaptives, auf Instanznormalisierung basierendes GAN-Framework für nicht parallele Many-to-many- und Zero-Shot-VC."]} +{"source": "Self-attention-based Transformer has demonstrated the state-of-the-art performances in a number of natural language processing tasks.Self attention is able to model long-term dependencies, but it may suffer from the extraction of irrelevant information in the context.To tackle the problem, we propose a novel model called Sparse Transformer.Sparse Transformer is able to improve the concentration of attention on the global context through an explicit selection of the most relevant segments.Extensive experimental results on a series of natural language processing tasks, including neural machine translation, image captioning, and language modeling, all demonstrate the advantages of Sparse Transformer in model performance. Sparse Transformer reaches the state-of-the-art performances in the IWSLT 2015 English-to-Vietnamese translation and IWSLT 2014 German-to-English translation.In addition, we conduct qualitative analysis to account for Sparse Transformer's superior performance.", "target": ["Diese Arbeit schlägt Sparse Transformer vor, um die Konzentration der Aufmerksamkeit auf den globalen Kontext durch eine explizite Auswahl der relevantesten Segmente für das Lernen von Sequenz zu Sequenz zu verbessern."]} +{"source": "Human observers can learn to recognize new categories of objects from a handful of examples, yet doing so with machine perception remains an open challenge.We hypothesize that data-efficient recognition is enabled by representations which make the variability in natural signals more predictable, as suggested by recent perceptual evidence.We therefore revisit and improve Contrastive Predictive Coding, a recently-proposed unsupervised learning framework, and arrive at a representation which enables generalization from small amounts of labeled data.When provided with only 1% of ImageNet labels (i.e. 13 per class), this model retains a strong classification performance, 73% Top-5 accuracy, outperforming supervised networks by 28% (a 65% relative improvement) and state-of-the-art semi-supervised methods by 14%.We also find this representation to serve as a useful substrate for object detection on the PASCAL-VOC 2007 dataset, approaching the performance of representations trained with a fully annotated ImageNet dataset.", "target": ["Nicht überwachte Repräsentationen, die mit Contrastive Predictive Coding gelernt wurden, ermöglichen eine dateneffiziente Bildklassifizierung."]} +{"source": "We demonstrate the possibility of what we call sparse learning: accelerated training of deep neural networks that maintain sparse weights throughout training while achieving dense performance levels.We accomplish this by developing sparse momentum, an algorithm which uses exponentially smoothed gradients (momentum) to identify layers and weights which reduce the error efficiently.Sparse momentum redistributes pruned weights across layers according to the mean momentum magnitude of each layer.Within a layer, sparse momentum grows weights according to the momentum magnitude of zero-valued weights.We demonstrate state-of-the-art sparse performance on MNIST, CIFAR-10, and ImageNet, decreasing the mean error by a relative 8%, 15%, and 6% compared to other sparse algorithms.Furthermore, we show that sparse momentum reliably reproduces dense performance levels while providing up to 5.61x faster training.In our analysis, ablations show that the benefits of momentum redistribution and growth increase with the depth and size of the network.", "target": ["Die Neuverteilung und Vergrößerung der Gewichte entsprechend der Impulsgröße ermöglicht das Training von spärlichen Netzen aus zufälligen Initialisierungen, die dichte Leistungsniveaus mit 5 % bis 50 % Gewichten erreichen können, während das Training um das bis zu 5,6-fache beschleunigt wird."]} +{"source": "To provide principled ways of designing proper Deep Neural Network (DNN) models, it is essential to understand the loss surface of DNNs under realistic assumptions.We introduce interesting aspects for understanding the local minima and overall structure of the loss surface.The parameter domain of the loss surface can be decomposed into regions in which activation values (zero or one for rectified linear units) are consistent.We found that, in each region, the loss surface have properties similar to that of linear neural networks where every local minimum is a global minimum.This means that every differentiable local minimum is the global minimum of the corresponding region.We prove that for a neural network with one hidden layer using rectified linear units under realistic assumptions.There are poor regions that lead to poor local minima, and we explain why such regions exist even in the overparameterized DNNs.", "target": ["Die Verlustfläche von neuronalen Netzen ist eine disjunkte Vereinigung von Regionen, bei der jedes lokale Minimum ein globales Minimum der entsprechenden Region ist."]} +{"source": "Contextualized word representations, such as ELMo and BERT, were shown to perform well on a various of semantic and structural (syntactic) task.In this work, we tackle the task of unsupervised disentanglement between semantics and structure in neural language representations: we aim to learn a transformation of the contextualized vectors, that discards the lexical semantics, but keeps the structural information.To this end, we automatically generate groups of sentences which are structurally similar but semantically different, and use metric-learning approach to learn a transformation that emphasizes the structural component that is encoded in the vectors.We demonstrate that our transformation clusters vectors in space by structural properties, rather than by lexical semantics.Finally, we demonstrate the utility of our distilled representations by showing that they outperform the original contextualized representations in few-shot parsing setting.", "target": ["Wir destillieren aus Sprachmodellen Repräsentationen für die Syntax durch unüberwachtes metrisches Lernen"]} +{"source": "We introduce a new memory architecture for navigation in previously unseen environments, inspired by landmark-based navigation in animals.The proposed semi-parametric topological memory (SPTM) consists of a (non-parametric) graph with nodes corresponding to locations in the environment and a (parametric) deep network capable of retrieving nodes from the graph based on observations.The graph stores no metric information, only connectivity of locations corresponding to the nodes.We use SPTM as a planning module in a navigation system.Given only 5 minutes of footage of a previously unseen maze, an SPTM-based navigation agent can build a topological map of the environment and use it to confidently navigate towards goals.The average success rate of the SPTM agent in goal-directed navigation across test environments is higher than the best-performing baseline by a factor of three.", "target": ["Wir stellen eine neue Speicherarchitektur für die Navigation in bisher unbekannten Umgebungen vor, die von der Landmarken-basierten Navigation bei Tieren inspiriert ist."]} +{"source": "The available resolution in our visual world is extremely high, if not infinite.Existing CNNs can be applied in a fully convolutional way to images of arbitrary resolution, but as the size of the input increases, they can not capture contextual information.In addition, computational requirements scale linearly to the number of input pixels, and resources are allocated uniformly across the input, no matter how informative different image regions are.We attempt to address these problems by proposing a novel architecture that traverses an image pyramid in a top-down fashion, while it uses a hard attention mechanism to selectively process only the most informative image parts.We conduct experiments on MNIST and ImageNet datasets, and we show that our models can significantly outperform fully convolutional counterparts, when the resolution of the input is that big that the receptive field of the baselines can not adequately cover the objects of interest.Gains in performance come for less FLOPs, because of the selective processing that we follow.Furthermore, our attention mechanism makes our predictions more interpretable, and creates a trade-off between accuracy and complexity that can be tuned both during training and testing time.", "target": ["Wir schlagen eine neuartige Architektur vor, die eine Bildpyramide von oben nach unten durchläuft und dabei nur die informativsten Regionen auf dem Weg besucht."]} +{"source": "Hyperparameter optimization can be formulated as a bilevel optimization problem, where the optimal parameters on the training set depend on the hyperparameters.We aim to adapt regularization hyperparameters for neural networks by fitting compact approximations to the best-response function, which maps hyperparameters to optimal weights and biases.We show how to construct scalable best-response approximations for neural networks by modeling the best-response as a single network whose hidden units are gated conditionally on the regularizer.We justify this approximation by showing the exact best-response for a shallow linear network with L2-regularized Jacobian can be represented by a similar gating mechanism.We fit this model using a gradient-based hyperparameter optimization algorithm which alternates between approximating the best-response around the current hyperparameters and optimizing the hyperparameters using the approximate best-response function.Unlike other gradient-based approaches, we do not require differentiating the training loss with respect to the hyperparameters, allowing us to tune discrete hyperparameters, data augmentation hyperparameters, and dropout probabilities.Because the hyperparameters are adapted online, our approach discovers hyperparameter schedules that can outperform fixed hyperparameter values.Empirically, our approach outperforms competing hyperparameter optimization methods on large-scale deep learning problems.We call our networks, which update their own hyperparameters online during training, Self-Tuning Networks (STNs).", "target": ["Wir verwenden ein Hypernetzwerk zur Vorhersage optimaler Gewichte bei gegebenen Hyperparametern und trainieren alles gemeinsam."]} +{"source": "Conditional Generative Adversarial Networks (cGANs) are finding increasingly widespread use in many application domains.Despite outstanding progress, quantitative evaluation of such models often involves multiple distinct metrics to assess different desirable properties, such as image quality, conditional consistency, and intra-conditioning diversity.In this setting, model benchmarking becomes a challenge, as each metric may indicate a different \"best\" model.In this paper, we propose the Frechet Joint Distance (FJD), which is defined as the Frechet distance between joint distributions of images and conditioning, allowing it to implicitly capture the aforementioned properties in a single metric.We conduct proof-of-concept experiments on a controllable synthetic dataset, which consistently highlight the benefits of FJD when compared to currently established metrics.Moreover, we use the newly introduced metric to compare existing cGAN-based models for a variety of conditioning modalities (e.g. class labels, object masks, bounding boxes, images, and text captions).We show that FJD can be used as a promising single metric for model benchmarking.", "target": ["Wir schlagen eine neue Metrik zur Bewertung von bedingten GANs vor, die Bildqualität, bedingte Konsistenz und Vielfalt innerhalb der Konditionierung in einem einzigen Maß erfasst."]} +{"source": "Combining deep model-free reinforcement learning with on-line planning is a promising approach to building on the successes of deep RL.On-line planning with look-ahead trees has proven successful in environments where transition models are known a priori.However, in complex environments where transition models need to be learned from data, the deficiencies of learned models have limited their utility for planning.To address these challenges, we propose TreeQN, a differentiable, recursive, tree-structured model that serves as a drop-in replacement for any value function network in deep RL with discrete actions.TreeQN dynamically constructs a tree by recursively applying a transition model in a learned abstract state space and then aggregating predicted rewards and state-values using a tree backup to estimate Q-values.We also propose ATreeC, an actor-critic variant that augments TreeQN with a softmax layer to form a stochastic policy network.Both approaches are trained end-to-end, such that the learned model is optimised for its actual use in the tree.We show that TreeQN and ATreeC outperform n-step DQN and A2C on a box-pushing task, as well as n-step DQN and value prediction networks (Oh et al., 2017) on multiple Atari games.Furthermore, we present ablation studies that demonstrate the effect of different auxiliary losses on learning transition models.", "target": ["Wir stellen TreeQN und ATreeC vor, neue Architekturen für tiefes Reinforcement Learning in diskreten Handlungsdomänen, die differenzierbare Online-Baumplanung in die Handlungswertfunktion oder -politik integrieren."]} +{"source": "Multi-label classification (MLC) is the task of assigning a set of target labels for a given sample.Modeling the combinatorial label interactions in MLC has been a long-haul challenge.Recurrent neural network (RNN) based encoder-decoder models have shown state-of-the-art performance for solving MLC.However, the sequential nature of modeling label dependencies through an RNN limits its ability in parallel computation, predicting dense labels, and providing interpretable results.In this paper, we propose Message Passing Encoder-Decoder (MPED) Networks, aiming to provide fast, accurate, and interpretable MLC.MPED networks model the joint prediction of labels by replacing all RNNs in the encoder-decoder architecture with message passing mechanisms and dispense with autoregressive inference entirely. The proposed models are simple, fast, accurate, interpretable, and structure-agnostic (can be used on known or unknown structured data).Experiments on seven real-world MLC datasets show the proposed models outperform autoregressive RNN models across five different metrics with a significant speedup during training and testing time.", "target": ["Wir schlagen Message-Passing-Encoder-Decoder-Netzwerke für eine schnelle und genaue Modellierung von Label-Abhängigkeiten für Multi-Label-Klassifikation vor."]} +{"source": "Recent few-shot learning algorithms have enabled models to quickly adapt to new tasks based on only a few training samples.Previous few-shot learning works have mainly focused on classification and reinforcement learning.In this paper, we propose a few-shot meta-learning system that focuses exclusively on regression tasks.Our model is based on the idea that the degree of freedom of the unknown function can be significantly reduced if it is represented as a linear combination of a set of sparsifying basis functions.This enables a few labeled samples to approximate the function.We design a Basis Function Learner network to encode basis functions for a task distribution, and a Weights Generator network to generate the weight vector for a novel task.We show that our model outperforms the current state of the art meta-learning methods in various regression tasks.", "target": ["Wir schlagen eine Methode zur Durchführung von Few-Shot Regression vor, indem wir eine Reihe von Basisfunktionen zur Darstellung der Funktionsverteilung lernen."]} +{"source": "Large-scale pre-trained language model, such as BERT, has recently achieved great success in a wide range of language understanding tasks.However, it remains an open question how to utilize BERT for text generation tasks.In this paper, we present a novel approach to addressing this challenge in a generic sequence-to-sequence (Seq2Seq) setting.We first propose a new task, Conditional Masked Language Modeling (C-MLM), to enable fine-tuning of BERT on target text-generation dataset.The fine-tuned BERT (i.e., teacher) is then exploited as extra supervision to improve conventional Seq2Seq models (i.e., student) for text generation.By leveraging BERT's idiosyncratic bidirectional nature, distilling the knowledge learned from BERT can encourage auto-regressive Seq2Seq models to plan ahead, imposing global sequence-level supervision for coherent text generation.Experiments show that the proposed approach significantly outperforms strong baselines of Transformer on multiple text generation tasks, including machine translation (MT) and text summarization.Our proposed model also achieves new state-of-the-art results on the IWSLT German-English and English-Vietnamese MT datasets.", "target": ["Wir schlagen einen modellunabhängigen Weg vor, BERT für die Texterzeugung zu nutzen und erreichen Verbesserungen gegenüber Transformer bei 2 Aufgaben in 4 Datensätzen."]} +{"source": "Humans have the remarkable ability to correctly classify images despite possible degradation.Many studies have suggested that this hallmark of human vision results from the interaction between feedforward signals from bottom-up pathways of the visual cortex and feedback signals provided by top-down pathways.Motivated by such interaction, we propose a new neuro-inspired model, namely Convolutional Neural Networks with Feedback (CNN-F).CNN-F extends CNN with a feedback generative network, combining bottom-up and top-down inference to perform approximate loopy belief propagation. We show that CNN-F's iterative inference allows for disentanglement of latent variables across layers.We validate the advantages of CNN-F over the baseline CNN.Our experimental results suggest that the CNN-F is more robust to image degradation such as pixel noise, occlusion, and blur. Furthermore, we show that the CNN-F is capable of restoring original images from the degraded ones with high reconstruction accuracy while introducing negligible artifacts.", "target": ["CNN-F erweitert CNN um ein generatives Netz mit Rückkopplung für robustes Sehen."]} +{"source": "We develop new approximation and statistical learning theories of convolutional neural networks (CNNs) via the ResNet-type structure where the channel size, filter size, and width are fixed.It is shown that a ResNet-type CNN is a universal approximator and its expression ability is no worse than fully-connected neural networks (FNNs) with a \\textit{block-sparse} structure even if the size of each layer in the CNN is fixed.Our result is general in the sense that we can automatically translate any approximation rate achieved by block-sparse FNNs into that by CNNs.Thanks to the general theory, it is shown that learning on CNNs satisfies optimality in approximation and estimation of several important function classes.As applications, we consider two types of function classes to be estimated: the Barron class and H\\\"older class.We prove the clipped empirical risk minimization (ERM) estimator can achieve the same rate as FNNs even the channel size, filter size, and width of CNNs are constant with respect to the sample size.This is minimax optimal (up to logarithmic factors) for the H\\\"older class.Our proof is based on sophisticated evaluations of the covering number of CNNs and the non-trivial parameter rescaling technique to control the Lipschitz constant of CNNs to be constructed.", "target": ["Es wird gezeigt, dass CNNs vom Typ ResNet ein universeller Approximator sind und dass ihre Ausdrucksfähigkeit nicht schlechter ist als die von vollständig verbundenen neuronalen Netzen (FNNs) mit einer \\textit{block-sparse} Struktur, selbst wenn die Größe jeder Schicht im CNN festgelegt ist."]} +{"source": "Few shot image classification aims at learning a classifier from limited labeled data.Generating the classification weights has been applied in many meta-learning approaches for few shot image classification due to its simplicity and effectiveness.However, we argue that it is difficult to generate the exact and universal classification weights for all the diverse query samples from very few training samples.In this work, we introduce Attentive Weights Generation for few shot learning via Information Maximization (AWGIM), which addresses current issues by two novel contributions.i) AWGIM generates different classification weights for different query samples by letting each of query samples attends to the whole support set.ii) To guarantee the generated weights adaptive to different query sample, we re-formulate the problem to maximize the lower bound of mutual information between generated weights and query as well as support data.As far as we can see, this is the first attempt to unify information maximization into few shot learning.Both two contributions are proved to be effective in the extensive experiments and we show that AWGIM is able to achieve state-of-the-art performance on benchmark datasets.", "target": ["Ein neuartiges \"few shot\"-Lernverfahren zur Erzeugung abfragespezifischer Klassifikationsgewichte durch Informationsmaximierung."]} +{"source": "Conversational question answering (CQA) is a novel QA task that requires the understanding of dialogue context.Different from traditional single-turn machine reading comprehension (MRC), CQA is a comprehensive task comprised of passage reading, coreference resolution, and contextual understanding.In this paper, we propose an innovative contextualized attention-based deep neural network, SDNet, to fuse context into traditional MRC models.Our model leverages both inter-attention and self-attention to comprehend the conversation and passage.Furthermore, we demonstrate a novel method to integrate the BERT contextual model as a sub-module in our network.Empirical results show the effectiveness of SDNet.On the CoQA leaderboard, it outperforms the previous best model's F1 score by 1.6%.Our ensemble model further improves the F1 score by 2.7%.", "target": ["Eine neuronale Methode zur Beantwortung von Gesprächsfragen mit einem Aufmerksamkeitsmechanismus und einer neuartigen Verwendung von BERT als kontextuellem Einbettungsmittel"]} +{"source": "Generative adversarial networks (GANs) are one of the most popular approaches when it comes to training generative models, among which variants of Wasserstein GANs are considered superior to the standard GAN formulation in terms of learning stability and sample quality.However, Wasserstein GANs require the critic to be 1-Lipschitz, which is often enforced implicitly by penalizing the norm of its gradient, or by globally restricting its Lipschitz constant via weight normalization techniques.Training with a regularization term penalizing the violation of the Lipschitz constraint explicitly, instead of through the norm of the gradient, was found to be practically infeasible in most situations.Inspired by Virtual Adversarial Training, we propose a method called Adversarial Lipschitz Regularization, and show that using an explicit Lipschitz penalty is indeed viable and leads to competitive performance when applied to Wasserstein GANs, highlighting an important connection between Lipschitz regularization and adversarial training.", "target": ["Alternative zur Gradientenstrafe"]} +{"source": "Multi-task learning promises to use less data, parameters, and time than training separate single-task models.But realizing these benefits in practice is challenging.In particular, it is difficult to define a suitable architecture that has enough capacity to support many tasks while not requiring excessive compute for each individual task.There are difficult trade-offs when deciding how to allocate parameters and layers across a large set of tasks.To address this, we propose a method for automatically searching over multi-task architectures that accounts for resource constraints.We define a parameterization of feature sharing strategies for effective coverage and sampling of architectures.We also present a method for quick evaluation of such architectures with feature distillation.Together these contributions allow us to quickly optimize for parameter-efficient multi-task models.We benchmark on Visual Decathlon, demonstrating that we can automatically search for and identify architectures that effectively make trade-offs between task resource requirements while maintaining a high level of final performance.", "target": ["automatische Suche nach Multi-Task-Architekturen, die die Nutzung von Funktionen pro Aufgabe reduzieren"]} +{"source": "As distributed approaches to natural language semantics have developed and diversified, embedders for linguistic units larger than words (e.g., sentences) have come to play an increasingly important role. To date, such embedders have been evaluated using benchmark tasks (e.g., GLUE) and linguistic probes. We propose a comparative approach, nearest neighbor overlap (N2O), that quantifies similarity between embedders in a task-agnostic manner. N2O requires only a collection of examples and is simple to understand: two embedders are more similar if, for the same set of inputs, there is greater overlap between the inputs' nearest neighbors. We use N2O to compare 21 sentence embedders and show the effects of different design choices and architectures.", "target": ["Wir schlagen Nearest Neighbor Overlap vor, ein Verfahren, das die Ähnlichkeit zwischen Einbettungsprogrammen aufgabenunabhängig quantifiziert, und verwenden es, um 21 Satzeinbettungsprogramme zu vergleichen."]} +{"source": "Generative Adversarial Networks (GANs) can achieve state-of-the-art sample quality in generative modelling tasks but suffer from the mode collapse problem.Variational Autoencoders (VAE) on the other hand explicitly maximize a reconstruction-based data log-likelihood forcing it to cover all modes, but suffer from poorer sample quality.Recent works have proposed hybrid VAE-GAN frameworks which integrate a GAN-based synthetic likelihood to the VAE objective to address both the mode collapse and sample quality issues, with limited success.This is because the VAE objective forces a trade-off between the data log-likelihood and divergence to the latent prior.The synthetic likelihood ratio term also shows instability during training.We propose a novel objective with a ``\"Best-of-Many-Samples\" reconstruction cost and a stable direct estimate of the synthetic likelihood.This enables our hybrid VAE-GAN framework to achieve high data log-likelihood and low divergence to the latent prior at the same time and shows significant improvement over both hybrid VAE-GANS and plain GANs in mode coverage and quality.", "target": ["Wir schlagen ein neues Ziel für das Training hybrider VAE-GANs vor, das zu einer deutlichen Verbesserung der Modusabdeckung und der Qualität führt."]} +{"source": "In order to mimic the human ability of continual acquisition and transfer of knowledge across various tasks, a learning system needs the capability for life-long learning, effectively utilizing the previously acquired skills.As such, the key challenge is to transfer and generalize the knowledge learned from one task to other tasks, avoiding interference from previous knowledge and improving the overall performance.In this paper, within the continual learning paradigm, we introduce a method that effectively forgets the less useful data samples continuously across different tasks.The method uses statistical leverage score information to measure the importance of the data samples in every task and adopts frequent directions approach to enable a life-long learning property.This effectively maintains a constant training size across all tasks.We first provide some mathematical intuition for the method and then demonstrate its effectiveness with experiments on variants of MNIST and CIFAR100 datasets.", "target": ["Eine neue Methode verwendet statistische Leverage-Score-Informationen, um die Wichtigkeit der Datenproben in jeder Aufgabe zu messen, und nimmt häufige Richtungen an, um eine lebenslange Lerneigenschaft zu ermöglichen."]} +{"source": "Convolutional neural networks (CNNs) are inherently equivariant to translation.Efforts to embed other forms of equivariance have concentrated solely on rotation.We expand the notion of equivariance in CNNs through the Polar Transformer Network (PTN).PTN combines ideas from the Spatial Transformer Network (STN) and canonical coordinate representations.The result is a network invariant to translation and equivariant to both rotation and scale.PTN is trained end-to-end and composed of three distinct stages: a polar origin predictor, the newly introduced polar transformer module and a classifier.PTN achieves state-of-the-art on rotated MNIST and the newly introduced SIM2MNIST dataset, an MNIST variation obtained by adding clutter and perturbing digits with translation, rotation and scaling.The ideas of PTN are extensible to 3D which we demonstrate through the Cylindrical Transformer Network.", "target": ["Wir lernen Merkmalskarten, die invariant gegenüber Translation und äquivariant gegenüber Rotation und Skalierung sind."]} +{"source": "Meta-learning allows an intelligent agent to leverage prior learning episodes as a basis for quickly improving performance on a novel task.Bayesian hierarchical modeling provides a theoretical framework for formalizing meta-learning as inference for a set of parameters that are shared across tasks.Here, we reformulate the model-agnostic meta-learning algorithm (MAML) of Finn et al. (2017) as a method for probabilistic inference in a hierarchical Bayesian model.In contrast to prior methods for meta-learning via hierarchical Bayes, MAML is naturally applicable to complex function approximators through its use of a scalable gradient descent procedure for posterior inference.Furthermore, the identification of MAML as hierarchical Bayes provides a way to understand the algorithm’s operation as a meta-learning procedure, as well as an opportunity to make use of computational strategies for efficient inference.We use this opportunity to propose an improvement to the MAML algorithm that makes use of techniques from approximate inference and curvature estimation.", "target": ["Ein spezieller gradientenbasierter Meta-Lernalgorithmus, MAML, ist äquivalent zu einer Inferenzprozedur in einem hierarchischen Bayes'schen Modell. Wir nutzen diese Verbindung, um MAML durch Methoden der approximativen Inferenz und Krümmungsschätzung zu verbessern."]} +{"source": "This work provides an automatic machine learning (AutoML) modelling architecture called Autostacker.Autostacker improves the prediction accuracy of machine learning baselines by utilizing an innovative hierarchical stacking architecture and an efficient parameter search algorithm.Neither prior domain knowledge about the data nor feature preprocessing is needed.We significantly reduce the time of AutoML with a naturally inspired algorithm - Parallel Hill Climbing (PHC).By parallelizing PHC, Autostacker can provide candidate pipelines with sufficient prediction accuracy within a short amount of time.These pipelines can be used as is or as a starting point for human experts to build on.By focusing on the modelling process, Autostacker breaks the tradition of following fixed order pipelines by exploring not only single model pipeline but also innovative combinations and structures.As we will show in the experiment section, Autostacker achieves significantly better performance both in terms of test accuracy and time cost comparing with human initial trials and recent popular AutoML system.", "target": ["Automatisiertes maschinelles Lernsystem mit effizientem Suchalgorithmus und innovativer Struktur zur Bereitstellung besserer Modellgrundlagen."]} +{"source": "Surrogate models can be used to accelerate approximate Bayesian computation (ABC).In one such framework the discrepancy between simulated and observed data is modelled with a Gaussian process.So far principled strategies have been proposed only for sequential selection of the simulation locations.To address this limitation, we develop Bayesian optimal design strategies to parallellise the expensive simulations.We also touch the problem of quantifying the uncertainty of the ABC posterior due to the limited budget of simulations.", "target": ["Wir schlagen prinzipielle Batch-Bayesian-Versuchsplanungsstrategien und eine Methode zur Quantifizierung der Unsicherheit der Posterior-Zusammenfassungen in einem auf Gauß-Process-Surrogate basierenden Rahmen für approximative Bayesianische Berechnungen vor."]} +{"source": "Randomly initialized first-order optimization algorithms are the method of choice for solving many high-dimensional nonconvex problems in machine learning, yet general theoretical guarantees cannot rule out convergence to critical points of poor objective value.For some highly structured nonconvex problems however, the success of gradient descent can be understood by studying the geometry of the objective.We study one such problem -- complete orthogonal dictionary learning, and provide converge guarantees for randomly initialized gradient descent to the neighborhood of a global optimum.The resulting rates scale as low order polynomials in the dimension even though the objective possesses an exponential number of saddle points.This efficient convergence can be viewed as a consequence of negative curvature normal to the stable manifolds associated with saddle points, and we provide evidence that this feature is shared by other nonconvex problems of importance as well.", "target": ["Wir liefern eine effiziente Konvergenzrate für den Gradientenabstieg beim Lernziel des vollständigen orthogonalen Wörterbuchs auf der Grundlage einer geometrischen Analyse."]} +{"source": "Coding theory is a central discipline underpinning wireline and wireless modems that are the workhorses of the information age.Progress in coding theory is largely driven by individual human ingenuity with sporadic breakthroughs over the past century.In this paper we study whether it is possible to automate the discovery of decoding algorithms via deep learning.We study a family of sequential codes parametrized by recurrent neural network (RNN) architectures.We show that cre- atively designed and trained RNN architectures can decode well known sequential codes such as the convolutional and turbo codes with close to optimal performance on the additive white Gaussian noise (AWGN) channel, which itself is achieved by breakthrough algorithms of our times (Viterbi and BCJR decoders, representing dynamic programing and forward-backward algorithms).We show strong gen- eralizations, i.e., we train at a specific signal to noise ratio and block length but test at a wide range of these quantities, as well as robustness and adaptivity to deviations from the AWGN setting.", "target": ["Wir zeigen, dass kreativ gestaltete und trainierte RNN-Architekturen bekannte sequentielle Codes dekodieren können und nahezu optimale Leistungen erzielen."]} +{"source": "Adam is shown not being able to converge to the optimal solution in certain cases.Researchers recently propose several algorithms to avoid the issue of non-convergence of Adam, but their efficiency turns out to be unsatisfactory in practice.In this paper, we provide a new insight into the non-convergence issue of Adam as well as other adaptive learning rate methods.We argue that there exists an inappropriate correlation between gradient $g_t$ and the second moment term $v_t$ in Adam ($t$ is the timestep), which results in that a large gradient is likely to have small step size while a small gradient may have a large step size.We demonstrate that such unbalanced step sizes are the fundamental cause of non-convergence of Adam, and we further prove that decorrelating $v_t$ and $g_t$ will lead to unbiased step size for each gradient, thus solving the non-convergence problem of Adam.Finally, we propose AdaShift, a novel adaptive learning rate method that decorrelates $v_t$ and $g_t$ by temporal shifting, i.e., using temporally shifted gradient $g_{t-n}$ to calculate $v_t$.The experiment results demonstrate that AdaShift is able to address the non-convergence issue of Adam, while still maintaining a competitive performance with Adam in terms of both training speed and generalization.", "target": ["Wir analysieren und lösen das Problem der Nicht-Konvergenz von Adam."]} +{"source": "Most domain adaptation methods consider the problem of transferring knowledge to the target domain from a single source dataset.However, in practical applications, we typically have access to multiple sources.In this paper we propose the first approach for Multi-Source Domain Adaptation (MSDA) based on Generative Adversarial Networks.Our method is inspired by the observation that the appearance of a given image depends on three factors: the domain, the style (characterized in terms of low-level features variations) and the content.For this reason we propose to project the image features onto a space where only the dependence from the content is kept, and then re-project this invariant representation onto the pixel space using the target domain and style.In this way, new labeled images can be generated which are used to train a final target classifier.We test our approach using common MSDA benchmarks, showing that it outperforms state-of-the-art methods.", "target": ["In diesem Papier schlagen wir eine generative Methode zur Anpassung von Multisource-Domänen vor, die auf der Zerlegung von Inhalts-, Stil- und Domänenfaktoren basiert."]} +{"source": "Inferring the most likely configuration for a subset of variables of a joint distribution given the remaining ones – which we refer to as co-generation – is an important challenge that is computationally demanding for all but the simplest settings.This task has received a considerable amount of attention, particularly for classical ways of modeling distributions like structured prediction.In contrast, almost nothing is known about this task when considering recently proposed techniques for modeling high-dimensional distributions, particularly generative adversarial nets (GANs).Therefore, in this paper, we study the occurring challenges for co-generation with GANs.To address those challenges we develop an annealed importance sampling (AIS) based Hamiltonian Monte Carlo (HMC) co-generation algorithm.The presented approach significantly outperforms classical gradient-based methods on synthetic data and on CelebA.", "target": ["Verwendung von geglättetem Wichtigkeits-Sampling für das Co-Generation Problem."]} +{"source": "Implicit probabilistic models are models defined naturally in terms of a sampling procedure and often induces a likelihood function that cannot be expressed explicitly.We develop a simple method for estimating parameters in implicit models that does not require knowledge of the form of the likelihood function or any derived quantities, but can be shown to be equivalent to maximizing likelihood under some conditions.Our result holds in the non-asymptotic parametric setting, where both the capacity of the model and the number of data examples are finite.We also demonstrate encouraging experimental results.", "target": ["Wir entwickeln eine neue Methode zur Schätzung von Parametern ohne Likelihood, die unter bestimmten Bedingungen der maximalen Likelihood gleichwertig ist"]} +{"source": "While most approaches to the problem of Inverse Reinforcement Learning (IRL) focus on estimating a reward function that best explains an expert agent’s policy or demonstrated behavior on a control task, it is often the case that such behavior is more succinctly represented by a simple reward combined with a set of hard constraints.In this setting, the agent is attempting to maximize cumulative rewards subject to these given constraints on their behavior.We reformulate the problem of IRL on Markov Decision Processes (MDPs) such that, given a nominal model of the environment and a nominal reward function, we seek to estimate state, action, and feature constraints in the environment that motivate an agent’s behavior.Our approach is based on the Maximum Entropy IRL framework, which allows us to reason about the likelihood of an expert agent’s demonstrations given our knowledge of an MDP.Using our method, we can infer which constraints can be added to the MDP to most increase the likelihood of observing these demonstrations.We present an algorithm which iteratively infers the Maximum Likelihood Constraint to best explain observed behavior, and we evaluate its efficacy using both simulated behavior and recorded data of humans navigating around an obstacle.", "target": ["Unsere Methode ermittelt Einschränkungen bei der Aufgabenausführung, indem sie das Prinzip der maximalen Entropie nutzt, um zu quantifizieren, inwieweit sich Demonstrationen von dem erwarteten, nicht eingeschränkten Verhalten unterscheiden."]} +{"source": "The success of reinforcement learning in the real world has been limited to instrumented laboratory scenarios, often requiring arduous human supervision to enable continuous learning.In this work, we discuss the required elements of a robotic system that can continually and autonomously improve with data collected in the real world, and propose a particular instantiation of such a system.Subsequently, we investigate a number of challenges of learning without instrumentation -- including the lack of episodic resets, state estimation, and hand-engineered rewards -- and propose simple, scalable solutions to these challenges.We demonstrate the efficacy of our proposed system on dexterous robotic manipulation tasks in simulation and the real world, and also provide an insightful analysis and ablation study of the challenges associated with this learning paradigm.", "target": ["System zum Erlernen von Roboteraufgaben in der realen Welt mit Reinforcement Learning ohne Instrumentierung"]} +{"source": "We study the problem of cross-lingual voice conversion in non-parallel speech corpora and one-shot learning setting.Most prior work require either parallel speech corpora or enough amount of training data from a target speaker.However, we convert an arbitrary sentences of an arbitrary source speaker to target speaker's given only one target speaker training utterance.To achieve this, we formulate the problem as learning disentangled speaker-specific and context-specific representations and follow the idea of [1] which uses Factorized Hierarchical Variational Autoencoder (FHVAE).After training FHVAE on multi-speaker training data, given arbitrary source and target speakers' utterance, we estimate those latent representations and then reconstruct the desired utterance of converted voice to that of target speaker.We use multi-language speech corpus to learn a universal model that works for all of the languages.We investigate the use of a one-hot language embedding to condition the model on the language of the utterance being queried and show the effectiveness of the approach.We conduct voice conversion experiments with varying size of training utterances and it was able to achieve reasonable performance with even just one training utterance.We also investigate the effect of using or not using the language conditioning.Furthermore, we visualize the embeddings of the different languages and sexes.Finally, in the subjective tests, for one language and cross-lingual voice conversion, our approach achieved moderately better or comparable results compared to the baseline in speech quality and similarity.", "target": ["Wir verwenden einen Variational Autoencoder, um Stil und Inhalt zu trennen, und erreichen die Sprachumwandlung durch Modifizierung der Stileinbettung und -dekodierung. Wir verwenden ein mehrsprachiges Sprachkorpus und untersuchen dessen Auswirkungen."]} +{"source": "We propose an end-to-end-trainable attention module for convolutional neural network (CNN) architectures built for image classification.The module takes as input the 2D feature vector maps which form the intermediate representations of the input image at different stages in the CNN pipeline, and outputs a 2D matrix of scores for each map.Standard CNN architectures are modified through the incorporation of this module, and trained under the constraint that a convex combination of the intermediate 2D feature vectors, as parametrised by the score matrices, must alone be used for classification.Incentivised to amplify the relevant and suppress the irrelevant or misleading, the scores thus assume the role of attention values.Our experimental observations provide clear evidence to this effect: the learned attention maps neatly highlight the regions of interest while suppressing background clutter.Consequently, the proposed function is able to bootstrap standard CNN architectures for the task of image classification, demonstrating superior generalisation over 6 unseen benchmark datasets.When binarised, our attention maps outperform other CNN-based attention maps, traditional saliency maps, and top object proposals for weakly supervised segmentation as demonstrated on the Object Discovery dataset.We also demonstrate improved robustness against the fast gradient sign method of adversarial attack.", "target": ["Diese Arbeit schlägt eine Methode vor, die CNNs dazu zwingt, räumliche Aufmerksamkeit zu nutzen, um objektzentriertere Darstellungen zu lernen, die in verschiedener Hinsicht besser sind."]} +{"source": "Recurrent neural network(RNN) is an effective neural network in solving very complex supervised and unsupervised tasks.There has been a significant improvement in RNN field such as natural language processing, speech processing, computer vision and other multiple domains.This paper deals with RNN application on different use cases like Incident Detection , Fraud Detection , and Android Malware Classification.The best performing neural network architecture is chosenby conducting different chain of experiments for different network parameters and structures.The network is run up to 1000 epochs with learning rate set in the range of 0.01 to 0.5.Obviously, RNN performed very well when compared to classical machine learning algorithms.This is mainly possible because RNNs implicitly extracts the underlying features and also identifies the characteristics of the data.This lead to better accuracy.", "target": ["Rekurrente neuronale Netze für Anwendungsfälle im Bereich der Cybersicherheit"]} +{"source": "Anatomical studies demonstrate that brain reformats input information to generate reliable responses for performing computations.However, it remains unclear how neural circuits encode complex spatio-temporal patterns.We show that neural dynamics are strongly influenced by the phase alignment between the input and the spontaneous chaotic activity.Input alignment along the dominant chaotic projections causes the chaotic trajectories to become stable channels (or attractors), hence, improving the computational capability of a recurrent network.Using mean field analysis, we derive the impact of input alignment on the overall stability of attractors formed.Our results indicate that input alignment determines the extent of intrinsic noise suppression and hence, alters the attractor state stability, thereby controlling the network's inference ability.", "target": ["Input-Strukturierung entlang des Chaos für Stabilität"]} +{"source": "Generative adversarial networks are a learning framework that rely on training a discriminator to estimate a measure of difference between a target and generated distributions.GANs, as normally formulated, rely on the generated samples being completely differentiable w.r.t.the generative parameters, and thus do not work for discrete data.We introduce a method for training GANs with discrete data that uses the estimated difference measure from the discriminator to compute importance weights for generated samples, thus providing a policy gradient for training the generator.The importance weights have a strong connection to the decision boundary of the discriminator, and we call our method boundary-seeking GANs (BGANs).We demonstrate the effectiveness of the proposed algorithm with discrete image and character-based natural language generation. In addition, the boundary-seeking objective extends to continuous data, which can be used to improve stability of training, and we demonstrate this on Celeba, Large-scale Scene Understanding (LSUN) bedrooms, and Imagenet without conditioning.", "target": ["Wir befassen uns mit dem Training von GANs mit diskreten Daten, indem wir einen Policy-Gradienten formulieren, der sich über f-Divergenzen hinweg verallgemeinern lässt"]} +{"source": "Policy gradient methods have enjoyed great success in deep reinforcement learning but suffer from high variance of gradient estimates.The high variance problem is particularly exasperated in problems with long horizons or high-dimensional action spaces.To mitigate this issue, we derive a bias-free action-dependent baseline for variance reduction which fully exploits the structural form of the stochastic policy itself and does not make any additional assumptions about the MDP.We demonstrate and quantify the benefit of the action-dependent baseline through both theoretical analysis as well as numerical results, including an analysis of the suboptimality of the optimal state-dependent baseline.The result is a computationally efficient policy gradient algorithm, which scales to high-dimensional control problems, as demonstrated by a synthetic 2000-dimensional target matching task.Our experimental results indicate that action-dependent baselines allow for faster learning on standard reinforcement learning benchmarks and high-dimensional hand manipulation and synthetic tasks.Finally, we show that the general idea of including additional information in baselines for improved variance reduction can be extended to partially observed and multi-agent tasks.", "target": ["Handlungsabhängige Grundlinien können verzerrungsfrei sein und eine größere Varianzreduzierung bewirken als nur zustandsabhängige Grundlinien für politische Gradientenmethoden."]} +{"source": "The cost of annotating training data has traditionally been a bottleneck for supervised learning approaches.The problem is further exacerbated when supervised learning is applied to a number of correlated tasks simultaneously since the amount of labels required scales with the number of tasks.To mitigate this concern, we propose an active multitask learning algorithm that achieves knowledge transfer between tasks.The approach forms a so-called committee for each task that jointly makes decisions and directly shares data across similar tasks.Our approach reduces the number of queries needed during training while maintaining high accuracy on test data.Empirical results on benchmark datasets show significant improvements on both accuracy and number of query requests.", "target": ["Wir schlagen einen aktiven Multitask-Lernalgorithmus vor, der einen Wissenstransfer zwischen den Aufgaben ermöglicht."]} +{"source": "Detection of photo manipulation relies on subtle statistical traces, notoriously removed by aggressive lossy compression employed online.We demonstrate that end-to-end modeling of complex photo dissemination channels allows for codec optimization with explicit provenance objectives.We design a lightweight trainable lossy image codec, that delivers competitive rate-distortion performance, on par with best hand-engineered alternatives, but has lower computational footprint on modern GPU-enabled platforms.Our results show that significant improvements in manipulation detection accuracy are possible at fractional costs in bandwidth/storage.Our codec improved the accuracy from 37% to 86% even at very low bit-rates, well below the practicality of JPEG (QF 20).", "target": ["Wir lernen einen effizienten verlustbehafteten Bildcodec kennen, der so optimiert werden kann, dass er eine zuverlässige Erkennung von Fotomanipulationen zu einem Bruchteil der Kosten in Bezug auf Nutzlast/Qualität und sogar bei niedrigen Bitraten ermöglicht."]} +{"source": "Recurrent Neural Networks have long been the dominating choice for sequence modeling.However, it severely suffers from two issues: impotent in capturing very long-term dependencies and unable to parallelize the sequential computation procedure.Therefore, many non-recurrent sequence models that are built on convolution and attention operations have been proposed recently.Notably, models with multi-head attention such as Transformer have demonstrated extreme effectiveness in capturing long-term dependencies in a variety of sequence modeling tasks.Despite their success, however, these models lack necessary components to model local structures in sequences and heavily rely on position embeddings that have limited effects and require a considerable amount of design efforts.In this paper, we propose the R-Transformer which enjoys the advantages of both RNNs and the multi-head attention mechanism while avoids their respective drawbacks.The proposed model can effectively capture both local structures and global long-term dependencies in sequences without any use of position embeddings.We evaluate R-Transformer through extensive experiments with data from a wide range of domains and the empirical results show that R-Transformer outperforms the state-of-the-art methods by a large margin in most of the tasks.", "target": ["In diesem Papier wird ein effektives generisches Sequenzmodell vorgeschlagen, das die Stärken von RNNs und Multi-Head Attention nutzt."]} +{"source": "Many tasks in natural language processing and related domains require high precision output that obeys dataset-specific constraints.This level of fine-grained control can be difficult to obtain in large-scale neural network models.In this work, we propose a structured latent-variable approach that adds discrete control states within a standard autoregressive neural paradigm.Under this formulation, we can include a range of rich, posterior constraints to enforce task-specific knowledge that is effectively trained into the neural model.This approach allows us to provide arbitrary grounding of internal model decisions, without sacrificing any representational power of neural models.Experiments consider applications of this approach for text generation and part-of-speech induction.For natural language generation, we find that this method improves over standard benchmarks, while also providing fine-grained control.", "target": ["Ein strukturierter latent-variabler Ansatz, der diskrete Kontrollzustände innerhalb eines standardmäßigen autoregressiven neuronalen Paradigmas hinzufügt, um eine willkürliche Grundlage für interne Modellentscheidungen zu schaffen, ohne die Darstellungsfähigkeit neuronaler Modelle zu beeinträchtigen."]} +{"source": "Suppose a deep classification model is trained with samples that need to be kept private for privacy or confidentiality reasons.In this setting, can an adversary obtain the private samples if the classification model is given to the adversary?We call this reverse engineering against the classification model the Classifier-to-Generator (C2G) Attack.This situation arises when the classification model is embedded into mobile devices for offline prediction (e.g., object recognition for the automatic driving car and face recognition for mobile phone authentication).For C2G attack, we introduce a novel GAN, PreImageGAN.In PreImageGAN, the generator is designed to estimate the the sample distribution conditioned by the preimage of classification model $f$, $P(X|f(X)=y)$, where $X$ is the random variable on the sample space and $y$ is the probability vector representing the target label arbitrary specified by the adversary.In experiments, we demonstrate PreImageGAN works successfully with hand-written character recognition and face recognition.In character recognition, we show that, given a recognition model of hand-written digits, PreImageGAN allows the adversary to extract alphabet letter images without knowing that the model is built for alphabet letter images.In face recognition, we show that, when an adversary obtains a face recognition model for a set of individuals, PreImageGAN allows the adversary to extract face images of specific individuals contained in the set, even when the adversary has no knowledge of the face of the individuals.", "target": ["Schätzung der Verteilung der Trainingsdaten aus dem trainierten Klassifikator unter Verwendung von GAN."]} +{"source": "The goal of standard compressive sensing is to estimate an unknown vector from linear measurements under the assumption of sparsity in some basis.Recently, it has been shown that significantly fewer measurements may be required if the sparsity assumption is replaced by the assumption that the unknown vector lies near the range of a suitably-chosen generative model. In particular, in (Bora {\\em et al.}, 2017) it was shown that roughly $O(k\\log L)$ random Gaussian measurements suffice for accurate recovery when the $k$-input generative model is bounded and $L$-Lipschitz, and that $O(kd \\log w)$ measurements suffice for $k$-input ReLU networks with depth $d$ and width $w$. In this paper, we establish corresponding algorithm-independent lower bounds on the sample complexity using tools from minimax statistical analysis. In accordance with the above upper bounds, our results are summarized as follows: (i) We construct an $L$-Lipschitz generative model capable of generating group-sparse signals, and show that the resulting necessary number of measurements is $\\Omega(k \\log L)$;(ii) Using similar ideas, we construct two-layer ReLU networks of high width requiring $\\Omega(k \\log w)$ measurements, as well as lower-width deep ReLU networks requiring $\\Omega(k d)$ measurements. As a result, we establish that the scaling laws derived in (Bora {\\em et al.}, 2017) are optimal or near-optimal in the absence of further assumptions.", "target": ["Wir stellen fest, dass die in (Bora et al., 2017) abgeleiteten Skalierungsgesetze optimal oder nahezu optimal sind, wenn keine weiteren Annahmen getroffen werden."]} +{"source": "Discovering and exploiting the causal structure in the environment is a crucial challenge for intelligent agents.Here we explore whether modern deep reinforcement learning can be used to train agents to perform causal reasoning.We adopt a meta-learning approach, where the agent learns a policy for conducting experiments via causal interventions, in order to support a subsequent task which rewards making accurate causal inferences.We also found the agent could make sophisticated counterfactual predictions, as well as learn to draw causal inferences from purely observational data.Though powerful formalisms for causal reasoning have been developed, applying them in real-world domains can be difficult because fitting to large amounts of high dimensional data often requires making idealized assumptions.Our results suggest that causal reasoning in complex settings may benefit from powerful learning-based approaches.More generally, this work may offer new strategies for structured exploration in reinforcement learning, by providing agents with the ability to perform—and interpret—experiments.", "target": ["Meta-Lernen eines Lernalgorithmus, der in der Lage ist, kausale Schlussfolgerungen zu ziehen"]} +{"source": "Sentiment classification is an active research area with several applications including analysis of political opinions, classifying comments, movie reviews, news reviews and product reviews.To employ rule based sentiment classification, we require sentiment lexicons.However, manual construction of sentiment lexicon is time consuming and costly for resource-limited languages.To bypass manual development time and costs, we tried to build Amharic Sentiment Lexicons relying on corpus based approach.The intention of this approach is to handle sentiment terms specific to Amharic language from Amharic Corpus.Small set of seed terms are manually prepared from three parts of speech such as noun, adjective and verb.We developed algorithms for constructing Amharic sentiment lexicons automatically from Amharic news corpus.Corpus based approach is proposed relying on the word co-occurrence distributional embedding including frequency based embedding (i.e. Positive Point-wise Mutual Information PPMI).First we build word-context unigram frequency count matrix and transform it to point-wise mutual Information matrix.Using this matrix, we computed the cosine distance of mean vector of seed lists and each word in the corpus vocabulary.Based on the threshold value, the top closest words to the mean vector of seed list are added to the lexicon.Then the mean vector of the new sentiment seed list is updated and process is repeated until we get sufficient terms in the lexicon.Using PPMI with threshold value of 100 and 200, we got corpus based Amharic Sentiment lexicons of size 1811 and 3794 respectively by expanding 519 seeds.Finally, the lexicon generated in corpus based approach is evaluated.", "target": ["Entwicklung eines korpusbasierten Algorithmus zur Erstellung eines amharischen Sentiment-Lexikons auf der Grundlage eines Korpus"]} +{"source": "Optimistic initialisation is an effective strategy for efficient exploration in reinforcement learning (RL).In the tabular case, all provably efficient model-free algorithms rely on it.However, model-free deep RL algorithms do not use optimistic initialisation despite taking inspiration from these provably efficient tabular algorithms.In particular, in scenarios with only positive rewards, Q-values are initialised at their lowest possible values due to commonly used network initialisation schemes, a pessimistic initialisation.Merely initialising the network to output optimistic Q-values is not enough, since we cannot ensure that they remain optimistic for novel state-action pairs, which is crucial for exploration.We propose a simple count-based augmentation to pessimistically initialised Q-values that separates the source of optimism from the neural network.We show that this scheme is provably efficient in the tabular setting and extend it to the deep RL setting.Our algorithm, Optimistic Pessimistically Initialised Q-Learning (OPIQ), augments the Q-value estimates of a DQN-based agent with count-derived bonuses to ensure optimism during both action selection and bootstrapping.We show that OPIQ outperforms non-optimistic DQN variants that utilise a pseudocount-based intrinsic motivation in hard exploration tasks, and that it predicts optimistic estimates for novel state-action pairs.", "target": ["Wir ergänzen die Q-Wert-Schätzungen mit einem zählbasierten Bonus, der für Optimismus bei der Aktionsauswahl und beim Bootstrapping sorgt, selbst wenn die Q-Wert-Schätzungen pessimistisch sind."]} +{"source": "Model-free deep reinforcement learning (RL) algorithms have been demonstrated on a range of challenging decision making and control tasks.However, these methods typically suffer from two major challenges: very high sample complexity and brittle convergence properties, which necessitate meticulous hyperparameter tuning.Both of these challenges severely limit the applicability of such methods to complex, real-world domains.In this paper, we propose soft actor-critic, an off-policy actor-critic deep RL algorithm based on the maximum entropy reinforcement learning framework.In this framework, the actor aims to maximize expected reward while also maximizing entropy - that is, succeed at the task while acting as randomly as possible.Prior deep RL methods based on this framework have been formulated as either off-policy Q-learning, or on-policy policy gradient methods.By combining off-policy updates with a stable stochastic actor-critic formulation, our method achieves state-of-the-art performance on a range of continuous control benchmark tasks, outperforming prior on-policy and off-policy methods.Furthermore, we demonstrate that, in contrast to other off-policy algorithms, our approach is very stable, achieving very similar performance across different random seeds.", "target": ["Wir schlagen Soft Actor-Critic vor, einen Deep-RL-Algorithmus, der auf dem Maximum-Entropy Reinforcement Learning Framework basiert."]} +{"source": "In many settings, it is desirable to learn decision-making and control policies through learning or from expert demonstrations.The most common approaches under this framework are Behaviour Cloning (BC), and Inverse Reinforcement Learning (IRL).Recent methods for IRL have demonstrated the capacity to learn effective policies with access to a very limited set of demonstrations, a scenario in which BC methods often fail.Unfortunately, directly comparing the algorithms for these methods does not provide adequate intuition for understanding this difference in performance.This is the motivating factor for our work.We begin by presenting $f$-MAX, a generalization of AIRL (Fu et al., 2018), a state-of-the-art IRL method.$f$-MAX provides grounds for more directly comparing the objectives for LfD.We demonstrate that $f$-MAX, and by inheritance AIRL, is a subset of the cost-regularized IRL framework laid out by Ho & Ermon (2016).We conclude by empirically evaluating the factors of difference between various LfD objectives in the continuous control domain.", "target": ["Die Verteilungsanpassung durch Divergenzminimierung bietet eine gemeinsame Grundlage für den Vergleich von gegnerischen Maximum-Entropy Inverse Reinforcement Learning-Methoden mit Behaviour Cloning."]} +{"source": "Value-based methods constitute a fundamental methodology in planning and deep reinforcement learning (RL).In this paper, we propose to exploit the underlying structures of the state-action value function, i.e., Q function, for both planning and deep RL.In particular, if the underlying system dynamics lead to some global structures of the Q function, one should be capable of inferring the function better by leveraging such structures.Specifically, we investigate the low-rank structure, which widely exists for big data matrices.We verify empirically the existence of low-rank Q functions in the context of control and deep RL tasks (Atari games).As our key contribution, by leveraging Matrix Estimation (ME) techniques, we propose a general framework to exploit the underlying low-rank structure in Q functions, leading to a more efficient planning procedure for classical control, and additionally, a simple scheme that can be applied to any value-based RL techniques to consistently achieve better performance on ''low-rank'' tasks.Extensive experiments on control tasks and Atari games confirm the efficacy of our approach.", "target": ["Wir schlagen einen generischen Rahmen vor, der es ermöglicht, die Low-Rank-Struktur sowohl beim Planungs- als auch beim Deep Reinforcement Learning zu nutzen."]} +{"source": "Learned representations of source code enable various software developer tools, e.g., to detect bugs or to predict program properties.At the core of code representations often are word embeddings of identifier names in source code, because identifiers account for the majority of source code vocabulary and convey important semantic information.Unfortunately, there currently is no generally accepted way of evaluating the quality of word embeddings of identifiers, and current evaluations are biased toward specific downstream tasks.This paper presents IdBench, the first benchmark for evaluating to what extent word embeddings of identifiers represent semantic relatedness and similarity.The benchmark is based on thousands of ratings gathered by surveying 500 software developers.We use IdBench to evaluate state-of-the-art embedding techniques proposed for natural language, an embedding technique specifically designed for source code, and lexical string distance functions, as these are often used in current developer tools.Our results show that the effectiveness of embeddings varies significantly across different embedding techniques and that the best available embeddings successfully represent semantic relatedness.On the downside, no existing embedding provides a satisfactory representation of semantic similarities, e.g., because embeddings consider identifiers with opposing meanings as similar, which may lead to fatal mistakes in downstream developer tools.IdBench provides a gold standard to guide the development of novel embeddings that address the current limitations.", "target": ["Ein Benchmark zur Bewertung der neuronalen Einbettung von Bezeichnern im Quellcode."]} +{"source": "Generative adversarial nets (GANs) are widely used to learn the data sampling process and their performance may heavily depend on the loss functions, given a limited computational budget.This study revisits MMD-GAN that uses the maximum mean discrepancy (MMD) as the loss function for GAN and makes two contributions.First, we argue that the existing MMD loss function may discourage the learning of fine details in data as it attempts to contract the discriminator outputs of real data.To address this issue, we propose a repulsive loss function to actively learn the difference among the real data by simply rearranging the terms in MMD.Second, inspired by the hinge loss, we propose a bounded Gaussian kernel to stabilize the training of MMD-GAN with the repulsive loss function.The proposed methods are applied to the unsupervised image generation tasks on CIFAR-10, STL-10, CelebA, and LSUN bedroom datasets.Results show that the repulsive loss function significantly improves over the MMD loss at no additional computational cost and outperforms other representative loss functions.The proposed methods achieve an FID score of 16.21 on the CIFAR-10 dataset using a single DCGAN network and spectral normalization.", "target": ["Durch Umstellung der Terme in der maximalen mittleren Diskrepanz ergibt sich eine viel bessere Verlustfunktion für den Diskriminator generativer adversarischer Netze"]} +{"source": "Deep neural networks have shown incredible performance for inference tasks in a variety of domains.Unfortunately, most current deep networks are enormous cloud-based structures that require significant storage space, which limits scaling of deep learning as a service (DLaaS) and use for on-device augmented intelligence. This paper finds algorithms that directly use lossless compressed representations of deep feedforward networks (with synaptic weights drawn from discrete sets), to perform inference without full decompression.The basic insight that allows less rate than naive approaches is the recognition that the bipartite graph layers of feedforward networks have a kind of permutation invariance to the labeling of nodes, in terms of inferential operation and that the inference operation depends locally on the edges directly connected to it.We also provide experimental results of our approach on the MNIST dataset.", "target": ["In diesem Werk werden Algorithmen entwickelt, die direkt verlustfrei komprimierte Darstellungen von tiefen Feedforward-Netzen verwenden, um Inferenzen ohne vollständige Dekompression durchzuführen."]} +{"source": "Generative adversarial networks (GANs) form a generative modeling approach known for producing appealing samples, but they are notably difficult to train.One common way to tackle this issue has been to propose new formulations of the GAN objective.Yet, surprisingly few studies have looked at optimization methods designed for this adversarial training.In this work, we cast GAN optimization problems in the general variational inequality framework.Tapping into the mathematical programming literature, we counter some common misconceptions about the difficulties of saddle point optimization and propose to extend methods designed for variational inequalities to the training of GANs.We apply averaging, extrapolation and a computationally cheaper variant that we call extrapolation from the past to the stochastic gradient method (SGD) and Adam.", "target": ["Wir gliedern GANs in den Rahmen der Variationsungleichung ein und importieren Techniken aus dieser Literatur, um GANs besser zu optimieren; wir geben algorithmische Erweiterungen an und testen empirisch ihre Leistung beim Training von GANs."]} +{"source": "In order to efficiently learn with small amount of data on new tasks, meta-learning transfers knowledge learned from previous tasks to the new ones.However, a critical challenge in meta-learning is the task heterogeneity which cannot be well handled by traditional globally shared meta-learning methods.In addition, current task-specific meta-learning methods may either suffer from hand-crafted structure design or lack the capability to capture complex relations between tasks.In this paper, motivated by the way of knowledge organization in knowledge bases, we propose an automated relational meta-learning (ARML) framework that automatically extracts the cross-task relations and constructs the meta-knowledge graph.When a new task arrives, it can quickly find the most relevant structure and tailor the learned structure knowledge to the meta-learner.As a result, the proposed framework not only addresses the challenge of task heterogeneity by a learned meta-knowledge graph, but also increases the model interpretability.We conduct extensive experiments on 2D toy regression and few-shot image classification and the results demonstrate the superiority of ARML over state-of-the-art baselines.", "target": ["Lösung des Problems der Aufgabenheterogenität beim Meta-Lernen durch Einführung eines Meta-Wissensgraphen"]} +{"source": "In this paper, a deep boosting algorithm is developed tolearn more discriminative ensemble classifier by seamlessly combining a set of base deep CNNs (base experts)with diverse capabilities, e.g., these base deep CNNs aresequentially trained to recognize a set of object classes in an easy-to-hard way according to theirlearning complexities.Our experimental results have demonstratedthat our deep boosting algorithm can significantly improve theaccuracy rates on large-scale visual recognition.", "target": ["Es wird ein Deep-Boosting-Algorithmus entwickelt, um durch die nahtlose Kombination einer Reihe von Deep-CNNs einen diskriminanteren Ensemble-Klassifikator zu lernen."]} +{"source": "We present a method for translating music across musical instruments and styles.This method is based on unsupervised training of a multi-domain wavenet autoencoder, with a shared encoder and a domain-independent latent space that is trained end-to-end on waveforms.Employing a diverse training dataset and large net capacity, the single encoder allows us to translate also from musical domains that were not seen during training.We evaluate our method on a dataset collected from professional musicians, and achieve convincing translations.We also study the properties of the obtained translation and demonstrate translating even from a whistle, potentially enabling the creation of instrumental music by untrained humans.", "target": ["Eine automatische Methode zur Umwandlung von Musik zwischen Instrumenten und Stilen"]} +{"source": "Most existing defenses against adversarial attacks only consider robustness to L_p-bounded distortions.In reality, the specific attack is rarely known in advance and adversaries are free to modify images in ways which lie outside any fixed distortion model; for example, adversarial rotations lie outside the set of L_p-bounded distortions.In this work, we advocate measuring robustness against a much broader range of unforeseen attacks, attacks whose precise form is unknown during defense design.We propose several new attacks and a methodology for evaluating a defense against a diverse range of unforeseen distortions.First, we construct novel adversarial JPEG, Fog, Gabor, and Snow distortions to simulate more diverse adversaries.We then introduce UAR, a summary metric that measures the robustness of a defense against a given distortion. Using UAR to assess robustness against existing and novel attacks, we perform an extensive study of adversarial robustness.We find that evaluation against existing L_p attacks yields redundant information which does not generalize to other attacks; we instead recommend evaluating against our significantly more diverse set of attacks.We further find that adversarial training against either one or multiple distortions fails to confer robustness to attacks with other distortion types. These results underscore the need to evaluate and study robustness against unforeseen distortions.", "target": ["Wir schlagen mehrere neue Angriffe und eine Methodik zur Messung der Robustheit gegenüber unvorhergesehenen Angriffen vor."]} +{"source": "Deep neural networks (DNNs) have witnessed as a powerful approach in this year by solving long-standing Artificialintelligence (AI) supervised and unsupervised tasks exists in natural language processing, speech processing, computer vision and others.In this paper, we attempt to apply DNNs on three different cyber security use cases: Android malware classification, incident detection and fraud detection.The data set of each use case contains real known benign and malicious activities samples.These use cases are part of Cybersecurity Data Mining Competition (CDMC) 2017.The efficient network architecture for DNNs is chosen by conducting various trails of experiments for network parameters and network structures.The experiments of such chosen efficient configurations of DNNs are run up to 1000 epochs with learning rate set in the range [0.01-0.5].Experiments of DNNs performed well in comparison to the classical machine learning algorithm in all cases of experiments of cyber security use cases.This is due to the fact that DNNs implicitly extract and build better features, identifies the characteristics of the data that lead to better accuracy.The best accuracy obtained by DNNs and XGBoost on Android malware classification 0.940 and 0.741, incident detection 1.00 and 0.997, and fraud detection 0.972 and 0.916 respectively.The accuracy obtained by DNNs varies -0.05%, +0.02%, -0.01% from the top scored system in CDMC 2017 tasks.", "target": ["Deep-Net: Deep Neural Network für Anwendungsfälle der Cybersicherheit"]} +{"source": "In this paper, we present an approach to learn recomposable motor primitives across large-scale and diverse manipulation demonstrations.Current approaches to decomposing demonstrations into primitives often assume manually defined primitives and bypass the difficulty of discovering these primitives.On the other hand, approaches in primitive discovery put restrictive assumptions on the complexity of a primitive, which limit applicability to narrow tasks.Our approach attempts to circumvent these challenges by jointly learning both the underlying motor primitives and recomposing these primitives to form the original demonstration.Through constraints on both the parsimony of primitive decomposition and the simplicity of a given primitive, we are able to learn a diverse set of motor primitives, as well as a coherent latent representation for these primitives.We demonstrate, both qualitatively and quantitatively, that our learned primitives capture semantically meaningful aspects of a demonstration.This allows us to compose these primitives in a hierarchical reinforcement learning setup to efficiently solve robotic manipulation tasks like reaching and pushing.", "target": ["Wir lernen einen Raum motorischer Primitive aus unkommentierten Roboterdemonstrationen und zeigen, dass diese Primitive semantisch sinnvoll sind und für neue Roboteraufgaben zusammengestellt werden können."]} +{"source": "Using modern deep learning models to make predictions on time series data from wearable sensors generally requires large amounts of labeled data.However, labeling these large datasets can be both cumbersome and costly.In this paper, we apply weak supervision to time series data, and programmatically label a dataset from sensors worn by patients with Parkinson's.We then built a LSTM model that predicts when these patients exhibit clinically relevant freezing behavior (inability to make effective forward stepping).We show that (1) when our model is trained using patient-specific data (prior sensor sessions), we come within 9% AUROC of a model trained using hand-labeled data and (2) when we assume no prior observations of subjects, our weakly supervised model matched performance with hand-labeled data.These results demonstrate that weak supervision may help reduce the need to painstakingly hand label time series training data.", "target": ["Wir demonstrieren die Machbarkeit eines schwach überwachten Zeitreihenklassifizierungsansatzes für tragbare Sensordaten."]} +{"source": "Learning semantic correspondence between the structured data (e.g., slot-value pairs) and associated texts is a core problem for many downstream NLP applications, e.g., data-to-text generation.Recent neural generation methods require to use large scale training data.However, the collected data-text pairs for training are usually loosely corresponded, where texts contain additional or contradicted information compare to its paired input.In this paper, we propose a local-to-global alignment (L2GA) framework to learn semantic correspondences from loosely related data-text pairs.First, a local alignment model based on multi-instance learning is applied to build the semantic correspondences within a data-text pair.Then, a global alignment model built on top of a memory guided conditional random field (CRF) layer is designed to exploit dependencies among alignments in the entire training corpus, where the memory is used to integrate the alignment clues provided by the local alignment model.Therefore, it is capable of inducing missing alignments for text spans that are not supported by its imperfect paired input.Experiments on recent restaurant dataset show that our proposed method can improve the alignment accuracy and as a by product, our method is also applicable to induce semantically equivalent training data-text pairs for neural generation models.", "target": ["Wir schlagen ein lokal-globales Alignment-Framework vor, um semantische Korrespondenzen aus verrauschten Daten-Text-Paaren mit schwacher Überwachung zu lernen."]} +{"source": "Imitation learning algorithms provide a simple and straightforward approach for training control policies via standard supervised learning methods.By maximizing the likelihood of good actions provided by an expert demonstrator, supervised imitation learning can produce effective policies without the algorithmic complexities and optimization challenges of reinforcement learning, at the cost of requiring an expert demonstrator -- typically a person -- to provide the demonstrations.In this paper, we ask: can we use imitation learning to train effective policies without any expert demonstrations?The key observation that makes this possible is that, in the multi-task setting, trajectories that are generated by a suboptimal policy can still serve as optimal examples for other tasks.In particular, in the setting where the tasks correspond to different goals, every trajectory is a successful demonstration for the state that it actually reaches.Informed by this observation, we propose a very simple algorithm for learning behaviors without any demonstrations, user-provided reward functions, or complex reinforcement learning methods.Our method simply maximizes the likelihood of actions the agent actually took in its own previous rollouts, conditioned on the goal being the state that it actually reached.Although related variants of this approach have been proposed previously in imitation learning settings with example demonstrations, we present the first instance of this approach as a method for learning goal-reaching policies entirely from scratch.We present a theoretical result linking self-supervised imitation learning and reinforcement learning, and empirical results showing that it performs competitively with more complex reinforcement learning methods on a range of challenging goal reaching problems.", "target": ["Lernen, wie man Ziele von Grund auf erreicht, indem man Imitation Learning mit Datenumetikettierung einsetzt"]} +{"source": "Neural networks have recently shown excellent performance on numerous classi- fication tasks.These networks often have a large number of parameters and thus require much data to train.When the number of training data points is small, however, a network with high flexibility will quickly overfit the training data, resulting in a large model variance and a poor generalization performance.To address this problem, we propose a new ensemble learning method called InterBoost for small-sample image classification.In the training phase, InterBoost first randomly generates two complementary datasets to train two base networks of the same structure, separately, and then next two complementary datasets for further training the networks are generated through interaction (or information sharing) between the two base networks trained previously.This interactive training process continues iteratively until a stop criterion is met.In the testing phase, the outputs of the two networks are combined to obtain one final score for classification.Detailed analysis of the method is provided for an in-depth understanding of its mechanism.", "target": ["In dieser Arbeit haben wir eine Ensemble-Methode namens InterBoost für das Training neuronaler Netze zur Klassifizierung kleiner Stichproben vorgeschlagen. Die Methode hat eine bessere Generalisierungsleistung als andere Ensemble-Methoden und reduziert die Varianz erheblich."]} +{"source": "Interpreting neural networks is a crucial and challenging task in machine learning.In this paper, we develop a novel framework for detecting statistical interactions captured by a feedforward multilayer neural network by directly interpreting its learned weights.Depending on the desired interactions, our method can achieve significantly better or similar interaction detection performance compared to the state-of-the-art without searching an exponential solution space of possible interactions.We obtain this accuracy and efficiency by observing that interactions between input features are created by the non-additive effect of nonlinear activation functions, and that interacting paths are encoded in weight matrices.We demonstrate the performance of our method and the importance of discovered interactions via experimental results on both synthetic datasets and real-world application datasets.", "target": ["Wir erkennen statistische Interaktionen, die von einem mehrschichtigen neuronalen Netz mit Vorwärtskopplung erfasst werden, indem wir seine gelernten Gewichte direkt interpretieren."]} +{"source": "The neural linear model is a simple adaptive Bayesian linear regression method that has recently been used in a number of problems ranging from Bayesian optimization to reinforcement learning.Despite its apparent successes in these settings, to the best of our knowledge there has been no systematic exploration of its capabilities on simple regression tasks.In this work we characterize these on the UCI datasets, a popular benchmark for Bayesian regression models, as well as on the recently introduced ''gap'' datasets, which are better tests of out-of-distribution uncertainty.We demonstrate that the neural linear model is a simple method that shows competitive performance on these tasks.", "target": ["Wir vergleichen das neuronale lineare Modell mit den UCI- und UCI-\"Gap\"-Datensätzen."]} +{"source": "The reproducibility of reinforcement-learning research has been highlighted as a key challenge area in the field.In this paper, we present a case study in reproducing the results of one groundbreaking algorithm, AlphaZero, a reinforcement learning system that learns how to play Go at a superhuman level given only the rules of the game.We describe Minigo, a reproduction of the AlphaZero system using publicly available Google Cloud Platform infrastructure and Google Cloud TPUs.The Minigo system includes both the central reinforcement learning loop as well as auxiliary monitoring and evaluation infrastructure.With ten days of training from scratch on 800 Cloud TPUs, Minigo can play evenly against LeelaZero and ELF OpenGo, two of the strongest publicly available Go AIs.We discuss the difficulties of scaling a reinforcement learning system and the monitoring systems required to understand the complex interplay of hyperparameter configurations.", "target": ["Wir haben AlphaZero auf der Google Cloud Platform reproduziert"]} +{"source": "Generative adversarial networks (GANs) train implicit generative models through solving minimax problems.Such minimax problems are known as nonconvex- nonconcave, for which the dynamics of first-order methods are not well understood.In this paper, we consider GANs in the type of the integral probability metrics (IPMs) with the generator represented by an overparametrized neural network.When the discriminator is solved to approximate optimality in each iteration, we prove that stochastic gradient descent on a regularized IPM objective converges globally to a stationary point with a sublinear rate.Moreover, we prove that when the width of the generator network is sufficiently large and the discriminator function class has enough discriminative ability, the obtained stationary point corresponds to a generator that yields a distribution that is close to the distribution of the observed data in terms of the total variation.To the best of our knowledge, we seem to first establish both the global convergence and global optimality of training GANs when the generator is parametrized by a neural network.", "target": ["Wir weisen globale Konvergenz zur Optimalität für IPM-basierte GANs nach, bei denen der Generator ein überparametrisiertes neuronales Netz ist."]} +{"source": "We present network embedding algorithms that capture information about a node from the local distribution over node attributes around it, as observed over random walks following an approach similar to Skip-gram.Observations from neighborhoods of different sizes are either pooled (AE) or encoded distinctly in a multi-scale approach (MUSAE). Capturing attribute-neighborhood relationships over multiple scales is useful for a diverse range of applications, including latent feature identification across disconnected networks with similar attributes.We prove theoretically that matrices of node-feature pointwise mutual information are implicitly factorized by the embeddings.Experiments show that our algorithms are robust, computationally efficient and outperform comparable models on social, web and citation network datasets.", "target": ["Wir entwickeln effiziente Verfahren zur approximativen Einbettung von Netzwerken mit mehreren Skalen und nachweisbaren Eigenschaften."]} +{"source": "Few-shot classification aims to learn a classifier to recognize unseen classes during training with limited labeled examples.While significant progress has been made, the growing complexity of network designs, meta-learning algorithms, and differences in implementation details make a fair comparison difficult.In this paper, we present1) a consistent comparative analysis of several representative few-shot classification algorithms, with results showing that deeper backbones significantly reduce the gap across methods including the baseline,2) a slightly modified baseline method that surprisingly achieves competitive performance when compared with the state-of-the-art on both the mini-ImageNet and the CUB datasets, and3) a new experimental setting for evaluating the cross-domain generalization ability for few-shot classification algorithms.Our results reveal that reducing intra-class variation is an important factor when the feature backbone is shallow, but not as critical when using deeper backbones.In a realistic, cross-domain evaluation setting, we show that a baseline method with a standard fine-tuning practice compares favorably against other state-of-the-art few-shot learning algorithms.", "target": ["Eine detaillierte empirische Studie über die Few-Shot Klassifizierung, die die Herausforderungen bei der Standardbewertung aufzeigt und eine neue Richtung angibt."]} +{"source": "Temporal logics are useful for describing dynamic system behavior, and have been successfully used as a language for goal definitions during task planning.Prior works on inferring temporal logic specifications have focused on \"summarizing\" the input dataset -- i.e., finding specifications that are satisfied by all plan traces belonging to the given set.In this paper, we examine the problem of inferring specifications that describe temporal differences between two sets of plan traces.We formalize the concept of providing such contrastive explanations, then present a Bayesian probabilistic model for inferring contrastive explanations as linear temporal logic specifications.We demonstrate the efficacy, scalability, and robustness of our model for inferring correct specifications across various benchmark planning domains and for a simulated air combat mission.", "target": ["Wir stellen ein Bayes'sches Inferenzmodell vor, um kontrastive Erklärungen (als LTL-Spezifikationen) abzuleiten, die beschreiben, wie sich zwei Sätze von Planspuren unterscheiden."]} +{"source": "This work tackles the problem of characterizing and understanding the decision boundaries of neural networks with piece-wise linear non-linearity activations.We use tropical geometry, a new development in the area of algebraic geometry, to provide a characterization of the decision boundaries of a simple neural network of the form (Affine, ReLU, Affine).Specifically, we show that the decision boundaries are a subset of a tropical hypersurface, which is intimately related to a polytope formed by the convex hull of two zonotopes.The generators of the zonotopes are precise functions of the neural network parameters.We utilize this geometric characterization to shed light and new perspective on three tasks.In doing so, we propose a new tropical perspective for the lottery ticket hypothesis, where we see the effect of different initializations on the tropical geometric representation of the decision boundaries.Also, we leverage this characterization as a new set of tropical regularizers, which deal directly with the decision boundaries of a network.We investigate the use of these regularizers in neural network pruning (removing network parameters that do not contribute to the tropical geometric representation of the decision boundaries) and in generating adversarial input attacks (with input perturbations explicitly perturbing the decision boundaries geometry to change the network prediction of the input).", "target": ["Die tropische Geometrie kann genutzt werden, um die Entscheidungsgrenzen neuronaler Netze darzustellen und interessante Erkenntnisse zu gewinnen."]} +{"source": "First-order methods such as stochastic gradient descent (SGD) are currently the standard algorithm for training deep neural networks.Second-order methods, despite their better convergence rate, are rarely used in practice due to the pro- hibitive computational cost in calculating the second-order information.In this paper, we propose a novel Gram-Gauss-Newton (GGN) algorithm to train deep neural networks for regression problems with square loss.Our method draws inspiration from the connection between neural network optimization and kernel regression of neural tangent kernel (NTK).Different from typical second-order methods that have heavy computational cost in each iteration, GGN only has minor overhead compared to first-order methods such as SGD.We also give theoretical results to show that for sufficiently wide neural networks, the convergence rate of GGN is quadratic.Furthermore, we provide convergence guarantee for mini-batch GGN algorithm, which is, to our knowledge, the first convergence result for the mini-batch version of a second-order method on overparameterized neural net- works.Preliminary experiments on regression tasks demonstrate that for training standard networks, our GGN algorithm converges much faster and achieves better performance than SGD.", "target": ["Eine neuartige Gram-Gauß-Newton-Methode zum Trainieren neuronaler Netze, inspiriert durch den neuronalen Tangens-Kernel und die Gauß-Newton-Methode, mit schneller Konvergenzgeschwindigkeit, sowohl theoretisch als auch experimentell."]} +{"source": "Recent pretrained sentence encoders achieve state of the art results on language understanding tasks, but does this mean they have implicit knowledge of syntactic structures?We introduce a grammatically annotated development set for the Corpus of Linguistic Acceptability (CoLA; Warstadt et al., 2018), which we use to investigate the grammatical knowledge of three pretrained encoders, including the popular OpenAI Transformer (Radford et al., 2018) and BERT (Devlin et al., 2018).We fine-tune these encoders to do acceptability classification over CoLA and compare the models’ performance on the annotated analysis set.Some phenomena, e.g. modification by adjuncts, are easy to learn for all models, while others, e.g. long-distance movement, are learned effectively only by models with strong overall performance, and others still, e.g. morphological agreement, are hardly learned by any model.", "target": ["Wir untersuchen das implizite syntaktische Wissen von Satzeinbettungen anhand eines neuen Analysesets von grammatikalisch annotierten Sätzen mit Akzeptanzurteilen."]} +{"source": "When considering simultaneously a finite number of tasks, multi-output learning enables one to account for the similarities of the tasks via appropriate regularizers.We propose a generalization of the classical setting to a continuum of tasks by using vector-valued RKHSs.", "target": ["Wir schlagen eine Erweiterung des Multi-Output-Lernens auf ein Kontinuum von Aufgaben unter Verwendung von Operator-bewerteten Kerneln vor."]} +{"source": "We analyze the joint probability distribution on the lengths of thevectors of hidden variables in different layers of a fully connecteddeep network, when the weights and biases are chosen randomly according toGaussian distributions, and the input is binary-valued. We showthat, if the activation function satisfies a minimal set ofassumptions, satisfied by all activation functions that we know thatare used in practice, then, as the width of the network gets large,the ``length process'' converges in probability to a length mapthat is determined as a simple function of the variances of therandom weights and biases, and the activation function.We also show that this convergence may fail for activation functions that violate our assumptions.", "target": ["Wir beweisen, dass für Aktivierungsfunktionen, die einige Bedingungen erfüllen, die Längen der Vektoren der versteckten Variablen zu einer Längenzuordnung konvergieren, wenn ein tiefes Netzwerk breit wird."]} +{"source": "Data augmentation is one of the most effective approaches for improving the accuracy of modern machine learning models, and it is also indispensable to train a deep model for meta-learning.However, most current data augmentation implementations applied in meta-learning are the same as those used in the conventional image classification.In this paper, we introduce a new data augmentation method for meta-learning, which is named as ``Task Level Data Augmentation'' (referred to Task Aug).The basic idea of Task Aug is to increase the number of image classes rather than the number of images in each class.In contrast, with a larger amount of classes, we can sample more diverse task instances during training.This allows us to train a deep network by meta-learning methods with little over-fitting.Experimental results show that our approach achieves state-of-the-art performance on miniImageNet, CIFAR-FS, and FC100 few-shot learning benchmarks.Once paper is accepted, we will provide the link to code.", "target": ["Wir schlagen einen Ansatz zur Datenerweiterung für das Meta-Lernen vor und beweisen, dass er zulässig ist."]} +{"source": "In this paper, we present a general framework for distilling expectations with respect to the Bayesian posterior distribution of a deep neural network, significantly extending prior work on a method known as ``Bayesian Dark Knowledge.\" Our generalized framework applies to the case of classification models and takes as input the architecture of a ``teacher\" network, a general posterior expectation of interest, and the architecture of a ``student\" network.The distillation method performs an online compression of the selected posterior expectation using iteratively generated Monte Carlo samples from the parameter posterior of the teacher model.We further consider the problem of optimizing the student model architecture with respect to an accuracy-speed-storage trade-off.We present experimental results investigating multiple data sets, distillation targets, teacher model architectures, and approaches to searching for student model architectures.We establish the key result that distilling into a student model with an architecture that matches the teacher, as is done in Bayesian Dark Knowledge, can lead to sub-optimal performance.Lastly, we show that student architecture search methods can identify student models with significantly improved performance.", "target": ["Ein allgemeiner Rahmen für die Destillation von Bayes'schen posterioren Erwartungen für tiefe neuronale Netze."]} +{"source": "Variational Autoencoders (VAEs) have proven to be powerful latent variable models.How- ever, the form of the approximate posterior can limit the expressiveness of the model.Categorical distributions are flexible and useful building blocks for example in neural memory layers.We introduce the Hierarchical Discrete Variational Autoencoder (HD-VAE): a hi- erarchy of variational memory layers.The Concrete/Gumbel-Softmax relaxation allows maximizing a surrogate of the Evidence Lower Bound by stochastic gradient ascent.We show that, when using a limited number of latent variables, HD-VAE outperforms the Gaussian baseline on modelling multiple binary image datasets.Training very deep HD-VAE remains a challenge due to the relaxation bias that is induced by the use of a surrogate objective.We introduce a formal definition and conduct a preliminary theoretical and empirical study of the bias.", "target": ["In diesem Papier führen wir eine diskrete Hierarchie kategorialer latenter Variablen ein, die wir mit Hilfe der Concrete/Gumbel-Softmax-Relaxation trainieren, und wir leiten eine obere Schranke für die absolute Differenz zwischen dem unbiased und dem biased Ziel ab."]} +{"source": "In this paper, we propose a novel technique for improving the stochastic gradient descent (SGD) method to train deep networks, which we term \\emph{PowerSGD}.The proposed PowerSGD method simply raises the stochastic gradient to a certain power $\\gamma\\in[0,1]$ during iterations and introduces only one additional parameter, namely, the power exponent $\\gamma$ (when $\\gamma=1$, PowerSGD reduces to SGD).We further propose PowerSGD with momentum, which we term \\emph{PowerSGDM}, and provide convergence rate analysis on both PowerSGD and PowerSGDM methods.Experiments are conducted on popular deep learning models and benchmark datasets.Empirical results show that the proposed PowerSGD and PowerSGDM obtain faster initial training speed than adaptive gradient methods, comparable generalization ability with SGD, and improved robustness to hyper-parameter selection and vanishing gradients.PowerSGD is essentially a gradient modifier via a nonlinear transformation.As such, it is orthogonal and complementary to other techniques for accelerating gradient-based optimization.", "target": ["Wir schlagen eine neue Klasse von Optimierern für die beschleunigte nicht-konvexe Optimierung mittels einer nichtlinearen Gradiententransformation vor."]} +{"source": "We aim to build complex humanoid agents that integrate perception, motor control, and memory.In this work, we partly factor this problem into low-level motor control from proprioception and high-level coordination of the low-level skills informed by vision.We develop an architecture capable of surprisingly flexible, task-directed motor control of a relatively high-DoF humanoid body by combining pre-training of low-level motor controllers with a high-level, task-focused controller that switches among low-level sub-policies.The resulting system is able to control a physically-simulated humanoid body to solve tasks that require coupling visual perception from an unstabilized egocentric RGB camera during locomotion in the environment.Supplementary video link: https://youtu.be/fBoir7PNxPk", "target": ["Lösen Sie Aufgaben, die die visuell gesteuerte Fortbewegung von Humanoiden beinhalten, indem Sie das Fortbewegungsverhalten aus Bewegungserfassungsdaten wiederverwenden."]} +{"source": "The gap between the empirical success of deep learning and the lack of strong theoretical guarantees calls for studying simpler models.By observing that a ReLU neuron is a product of a linear function with a gate (the latter determines whether the neuron is active or not), where both share a jointly trained weight vector, we propose to decouple the two.We introduce GaLU networks — networks in which each neuron is a product of a Linear Unit, defined by a weight vector which is being trained, with a Gate, defined by a different weight vector which is not being trained.Generally speaking, given a base model and a simpler version of it, the two parameters that determine the quality of the simpler version are whether its practical performance is close enough to the base model and whether it is easier to analyze it theoretically.We show that GaLU networks perform similarly to ReLU networks on standard datasets and we initiate a study of their theoretical properties, demonstrating that they are indeed easier to analyze.We believe that further research of GaLU networks may be fruitful for the development of a theory of deep learning.", "target": ["Wir schlagen Gated Linear Unit-Netzwerke vor - ein Modell, das bei realen Daten ähnliche Leistungen wie ReLU-Netzwerke erbringt, aber theoretisch viel einfacher zu analysieren ist."]} +{"source": "Machine learning systems often encounter Out-of-Distribution (OoD) errors when dealing with testing data coming from a different distribution from the one used for training.With their growing use in critical applications, it becomes important to develop systems that are able to accurately quantify its predictive uncertainty and screen out these anomalous inputs.However, unlike standard learning tasks, there is currently no well established guiding principle for designing architectures that can accurately quantify uncertainty.Moreover, commonly used OoD detection approaches are prone to errors and even sometimes assign higher likelihoods to OoD samples.To address these problems, we first seek to identify guiding principles for designing uncertainty-aware architectures, by proposing Neural Architecture Distribution Search (NADS).Unlike standard neural architecture search methods which seek for a single best performing architecture, NADS searches for a distribution of architectures that perform well on a given task, allowing us to identify building blocks common among all uncertainty aware architectures.With this formulation, we are able to optimize a stochastic outlier detection objective and construct an ensemble of models to perform OoD detection.We perform multiple OoD detection experiments and observe that our NADS performs favorably compared to state-of-the-art OoD detection methods.", "target": ["Wir schlagen eine Architektur-Suchmethode vor, um eine Verteilung von Architekturen zu identifizieren und daraus ein Bayes'sches Ensemble zur Ausreißererkennung zu konstruieren."]} +{"source": "Modern applications from Autonomous Vehicles to Video Surveillance generate massive amounts of image data.In this work we propose a novel image outlier detection approach (IOD for short) that leverages the cutting-edge image classifier to discover outliers without using any labeled outlier.We observe that although intuitively the confidence that a convolutional neural network (CNN) has that an image belongs to a particular class could serve as outlierness measure to each image, directly applying this confidence to detect outlier does not work well.This is because CNN often has high confidence on an outlier image that does not belong to any target class due to its generalization ability that ensures the high accuracy in classification.To solve this issue, we propose a Deep Neural Forest-based approach that harmonizes the contradictory requirements of accurately classifying images and correctly detecting the outlier images.Our experiments using several benchmark image datasets including MNIST, CIFAR-10, CIFAR-100, and SVHN demonstrate the effectiveness of our IOD approach for outlier detection, capturing more than 90% of outliers generated by injecting one image dataset into another, while still preserving the classification accuracy of the multi-class classification problem.", "target": ["Ein neuartiger Ansatz zur Erkennung von Ausreißern in Bilddaten unter Beibehaltung der Klassifizierungsgenauigkeit der Bildklassifizierung"]} +{"source": "This paper introduces CloudLSTM, a new branch of recurrent neural models tailored to forecasting over data streams generated by geospatial point-cloud sources.We design a Dynamic Point-cloud Convolution (D-Conv) operator as the core component of CloudLSTMs, which performs convolution directly over point-clouds and extracts local spatial features from sets of neighboring points that surround different elements of the input.This operator maintains the permutation invariance of sequence-to-sequence learning frameworks, while representing neighboring correlations at each time step -- an important aspect in spatiotemporal predictive learning.The D-Conv operator resolves the grid-structural data requirements of existing spatiotemporal forecasting models and can be easily plugged into traditional LSTM architectures with sequence-to-sequence learning and attention mechanisms. We apply our proposed architecture to two representative, practical use cases that involve point-cloud streams, i.e. mobile service traffic forecasting and air quality indicator forecasting.Our results, obtained with real-world datasets collected in diverse scenarios for each use case, show that CloudLSTM delivers accurate long-term predictions, outperforming a variety of neural network models.", "target": ["In diesem Beitrag wird CloudLSTM vorgestellt, ein neuer Zweig rekurrenter neuronaler Modelle, der auf die Vorhersage von Datenströmen zugeschnitten ist, die von raumbezogenen Punkthäufungen erzeugt werden."]} +{"source": "Knowledge Graphs (KG), composed of entities and relations, provide a structured representation of knowledge.For easy access to statistical approaches on relational data, multiple methods to embed a KG as components of R^d have been introduced.We propose TransINT, a novel and interpretable KG embedding method that isomorphically preserves the implication ordering among relations in the embedding space.TransINT maps set of entities (tied by a relation) to continuous sets of vectors that are inclusion-ordered isomorphically to relation implications.With a novel parameter sharing scheme, TransINT enables automatic training on missing but implied facts without rule grounding.We achieve new state-of-the-art performances with signficant margins in Link Prediction and Triple Classification on FB122 dataset, with boosted performance even on test instances that cannot be inferred by logical rules.The angles between the continuous sets embedded by TransINT provide an interpretable way to mine semantic relatedness and implication rules among relations.", "target": ["Wir schlagen TransINT vor, ein neuartiges und interpretierbares KG-Einbettungsverfahren, das die Implikationsordnung zwischen den Relationen im Einbettungsraum auf erklärbare, robuste und geometrisch kohärente Weise isomorph bewahrt."]} +{"source": "Unsupervised domain adaptive object detection aims to learn a robust detector on the domain shift circumstance, where the training (source) domain is label-rich with bounding box annotations, while the testing (target) domain is label-agnostic and the feature distributions between training and testing domains are dissimilar or even totally different.In this paper, we propose a gradient detach based Stacked Complementary Losses (SCL) method that uses detection objective (cross entropy and smooth l1 regression) as the primary objective, and cuts in several auxiliary losses in different network stages to utilize information from the complement data (target images) that can be effective in adapting model parameters to both source and target domains.A gradient detach operation is applied between detection and context sub-networks during training to force networks to learn discriminative representations.We argue that the conventional training with primary objective mainly leverages the information from the source-domain for maximizing likelihood and ignores the complement data in shallow layers of networks, which leads to an insufficient integration within different domains.Thus, our proposed method is a more syncretic adaptation learning process.We conduct comprehensive experiments on seven datasets, the results demonstrate that our method performs favorably better than the state-of-the-art methods by a large margin.For instance, from Cityscapes to FoggyCityscapes, we achieve 37.9% mAP, outperforming the previous art Strong-Weak by 3.6%.", "target": ["Wir stellen eine neue, auf Gradientenablösung basierende, komplementär-objektive Trainingsstrategie zur domänenadaptiven Objekterkennung vor."]} +{"source": "Convolutional neural networks (CNN) have become the most successful and popular approach in many vision-related domains.While CNNs are particularly well-suited for capturing a proper hierarchy of concepts from real-world images, they are limited to domains where data is abundant.Recent attempts have looked into mitigating this data scarcity problem by casting their original single-task problem into a new multi-task learning (MTL) problem.The main goal of this inductive transfer mechanism is to leverage domain-specific information from related tasks, in order to improve generalization on the main task.While recent results in the deep learning (DL) community have shown the promising potential of training task-specific CNNs in a soft parameter sharing framework, integrating the recent DL advances for improving knowledge sharing is still an open problem.In this paper, we propose the Deep Collaboration Network (DCNet), a novel approach for connecting task-specific CNNs in a MTL framework.We define connectivity in terms of two distinct non-linear transformation blocks.One aggregates task-specific features into global features, while the other merges back the global features with each task-specific network.Based on the observation that task relevance depends on depth, our transformation blocks use skip connections as suggested by residual network approaches, to more easily deactivate unrelated task-dependent features.To validate our approach, we employed facial landmark detection (FLD) datasets as they are readily amenable to MTL, given the number of tasks they include.Experimental results show that we can achieve up to 24.31% relative improvement in landmark failure rate over other state-of-the-art MTL approaches.We finally perform an ablation study showing that our approach effectively allows knowledge sharing, by leveraging domain-specific features at particular depths from tasks that we know are related.", "target": ["Wir schlagen einen neuen Ansatz für die Verbindung von aufgabenspezifischen Netzwerken in einer Multi-Task-Lernumgebung vor, der auf den jüngsten Fortschritten im Bereich der Residual Networks basiert."]} +{"source": "Zero-Shot Learning (ZSL) is a classification task where some classes referred as unseen classes have no labeled training images.Instead, we only have side information (or description) about seen and unseen classes, often in the form of semantic or descriptive attributes.Lack of training images from a set of classes restricts the use of standard classification techniques and losses, including the popular cross-entropy loss.The key step in tackling ZSL problem is bridging visual to semantic space via learning a nonlinear embedding.A well established approach is to obtain the semantic representation of the visual information and perform classification in the semantic space.In this paper, we propose a novel architecture of casting ZSL as a fully connected neural-network with cross-entropy loss to embed visual space to semantic space.During training in order to introduce unseen visual information to the network, we utilize soft-labeling based on semantic similarities between seen and unseen classes.To the best of our knowledge, such similarity based soft-labeling is not explored for cross-modal transfer and ZSL.We evaluate the proposed model on five benchmark datasets for zero-shot learning, AwA1, AwA2, aPY, SUN and CUB datasets, and show that, despite the simplicity, our approach achieves the state-of-the-art performance in Generalized-ZSL setting on all of these datasets and outperforms the state-of-the-art for some datasets.", "target": ["Verwendung des Cross-Entropie-Verlustes für Zero Shot Learning mit Soft Labeling auf ungesehenen Klassen: eine einfache und effektive Lösung, die auf fünf ZSL-Benchmark-Datensätzen die beste Leistung erzielt."]} +{"source": "In complex tasks, such as those with large combinatorial action spaces, random exploration may be too inefficient to achieve meaningful learning progress.In this work, we use a curriculum of progressively growing action spaces to accelerate learning.We assume the environment is out of our control, but that the agent may set an internal curriculum by initially restricting its action space.Our approach uses off-policy reinforcement learning to estimate optimal value functions for multiple action spaces simultaneously and efficiently transfers data, value estimates, and state representations from restricted action spaces to the full task.We show the efficacy of our approach in proof-of-concept control tasks and on challenging large-scale StarCraft micromanagement tasks with large, multi-agent action spaces.", "target": ["Die schrittweise Vergrößerung des verfügbaren Aktionsraums ist ein großartiger Lehrplan für Lernagenten"]} +{"source": "Recently, researchers have discovered that the state-of-the-art object classifiers can be fooled easily by small perturbations in the input unnoticeable to human eyes. It is known that an attacker can generate strong adversarial examples if she knows the classifier parameters. Conversely, a defender can robustify the classifier by retraining if she has the adversarial examples. The cat-and-mouse game nature of attacks and defenses raises the question of the presence of equilibria in the dynamics.In this paper, we present a neural-network based attack class to approximate a larger but intractable class of attacks, and formulate the attacker-defender interaction as a zero-sum leader-follower game.We present sensitivity-penalized optimization algorithms to find minimax solutions, which are the best worst-case defenses against whitebox attacks.Advantages of the learning-based attacks and defenses compared to gradient-based attacks and defenses are demonstrated with MNIST and CIFAR-10.", "target": ["Eine spieltheoretische Lösung für gegnerische Angriffe und Verteidigungsmaßnahmen."]} +{"source": "Supervised learning with irregularly sampled time series have been a challenge to Machine Learning methods due to the obstacle of dealing with irregular time intervals.Some papers introduced recently recurrent neural network models that deals with irregularity, but most of them rely on complex mechanisms to achieve a better performance.This work propose a novel method to represent timestamps (hours or dates) as dense vectors using sinusoidal functions, called Time Embeddings.As a data input method it and can be applied to most machine learning models.The method was evaluated with two predictive tasks from MIMIC III, a dataset of irregularly sampled time series of electronic health records.Our tests showed an improvement to LSTM-based and classical machine learning models, specially with very irregular data.", "target": ["Eine neuartige Methode zur Erstellung dichter Zeitdeskriptoren (Time Embeddings), um einfache Modelle zum Verständnis zeitlicher Strukturen zu erstellen"]} +{"source": "Community detection in graphs can be solved via spectral methods or posterior inference under certain probabilistic graphical models.Focusing on random graph families such as the stochastic block model, recent research has unified both approaches and identified both statistical and computational detection thresholds in terms of the signal-to-noise ratio.By recasting community detection as a node-wise classification problem on graphs, we can also study it from a learning perspective.We present a novel family of Graph Neural Networks (GNNs) for solving community detection problems in a supervised learning setting.We show that, in a data-driven manner and without access to the underlying generative models, they can match or even surpass the performance of the belief propagation algorithm on binary and multiclass stochastic block models, which is believed to reach the computational threshold in these cases.In particular, we propose to augment GNNs with the non-backtracking operator defined on the line graph of edge adjacencies.The GNNs are achieved good performance on real-world datasets. In addition, we perform the first analysis of the optimization landscape of using (linear) GNNs to solve community detection problems, demonstrating that under certain simplifications and assumptions, the loss value at any local minimum is close to the loss value at the global minimum/minima.", "target": ["Wir schlagen eine neuartige Architektur für neuronale Graphen-Netzwerke vor, die auf der Non-Backtracking-Matrix basiert, die über die Kanten-Adjazenzen definiert ist, und demonstrieren ihre Effektivität bei der Erkennung von Gemeinschaften in Graphen."]} +{"source": "Residual networks (Resnets) have become a prominent architecture in deep learning.However, a comprehensive understanding of Resnets is still a topic of ongoing research.A recent view argues that Resnets perform iterative refinement of features.We attempt to further expose properties of this aspect.To this end, we study Resnets both analytically and empirically.We formalize the notion of iterative refinement in Resnets by showing that residual architectures naturally encourage features to move along the negative gradient of loss during the feedforward phase.In addition, our empirical analysis suggests that Resnets are able to perform both representation learning and iterative refinement.In general, a Resnet block tends to concentrate representation learning behavior in the first few layers while higher layers perform iterative refinement of features.Finally we observe that sharing residual layers naively leads to representation explosion and hurts generalization performance, and show that simple existing strategies can help alleviating this problem.", "target": ["Residuale Verbindungen führen wirklich iterative Inferenz durch"]} +{"source": "We develop end-to-end learned reconstructions for lensless mask-based cameras, including an experimental system for capturing aligned lensless and lensed images for training. Various reconstruction methods are explored, on a scale from classic iterative approaches (based on the physical imaging model) to deep learned methods with many learned parameters. In the middle ground, we present several variations of unrolled alternating direction method of multipliers (ADMM) with varying numbers of learned parameters.The network structure combines knowledge of the physical imaging model with learned parameters updated from the data, which compensate for artifacts caused by physical approximations.Our unrolled approach is 20X faster than classic methods and produces better reconstruction quality than both the classic and deep methods on our experimental system.", "target": ["Wir verbessern die Rekonstruktionszeit und -qualität bei einem experimentellen maskenbasierten linsenlosen Imager mit einem End-to-End-Lernansatz, der das Wissen über das Bildgebungsmodell einbezieht."]} +{"source": "Deep learning, a rebranding of deep neural network research works, has achieved a remarkable success in recent years.With multiple hidden layers, deep learning models aim at computing the hierarchical feature representations of the observational data.Meanwhile, due to its severe disadvantages in data consumption, computational resources, parameter tuning costs and the lack of result explainability, deep learning has also suffered from lots of criticism.In this paper, we will introduce a new representation learning model, namely “Sample-Ensemble Genetic Evolutionary Network” (SEGEN), which can serve as an alternative approach to deep learning models.Instead of building one single deep model, based on a set of sampled sub-instances, SEGEN adopts a genetic-evolutionary learning strategy to build a group of unit models generations by generations.The unit models incorporated in SEGEN can be either traditional machine learning models or the recent deep learning models with a much “narrower” and “shallower” architecture.The learning results of each instance at the final generation will be effectively combined from each unit model via diffusive propagation and ensemble learning strategies.From the computational perspective, SEGEN requires far less data, fewer computational resources and parameter tuning efforts, but has sound theoretic interpretability of the learning process and results.Extensive experiments have been done on several different real-world benchmark datasets, and the experimental results obtained by SEGEN have demonstrated its advantages over the state-of-the-art representation learning models.", "target": ["Wir stellen ein neues Repräsentations-Lernmodell vor, nämlich das \"Sample-Ensemble Genetic Evolutionary Network\" (SEGEN), das als alternativer Ansatz für Deep-Learning-Modelle dienen kann."]} +{"source": "How can we teach artificial agents to use human language flexibly to solve problems in a real-world environment?We have one example in nature of agents being able to solve this problem: human babies eventually learn to use human language to solve problems, and they are taught with an adult human-in-the-loop.Unfortunately, current machine learning methods (e.g. from deep reinforcement learning) are too data inefficient to learn a language in this way (3).An outstanding goal is finding an algorithm with a suitable ‘language learning prior’ that allows it to learn human language, while minimizing the number of required human interactions.In this paper, we propose to learn such a prior in simulation, leveraging the increasing amount of available compute for machine learning experiments (1).We call our approach Learning to Learn to Communicate (L2C).Specifically, in L2C we train a meta-learning agent in simulation to interact with populations of pre-trained agents, each with their own distinct communication protocol.Once the meta-learning agent is able to quickly adapt to each population of agents, it can be deployed in new populations unseen during training, including populations of humans.To show the promise of the L2C framework, we conduct some preliminary experiments in a Lewis signaling game (4), where we show that agentstrained with L2C are able to learn a simple form of human language (represented by a hand-coded compositional language) in fewer iterations than randomly initialized agents.", "target": ["Wir schlagen vor, Meta-Lernen für effizienteres Sprachenlernen zu nutzen, und zwar durch eine Art \"Domänen-Randomisierung\"."]} +{"source": "We study the problem of fitting task-specific learning rate schedules from the perspective of hyperparameter optimization. This allows us to explicitly search for schedules that achieve good generalization.We describe the structure of the gradient of a validation error w.r.t.the learning rates, the hypergradient, and based on this we introduce a novel online algorithm.Our method adaptively interpolates between two recently proposed techniques (Franceschi et al., 2017; Baydin et al.,2018), featuring increased stability and faster convergence.We show empirically that the proposed technique compares favorably with baselines and related methodsin terms of final test accuracy.", "target": ["MARTHE: eine neue Methode zur Anpassung aufgabenspezifischer Lernratenpläne aus der Perspektive der Hyperparameteroptimierung"]} +{"source": "Recent years have witnessed some exciting developments in the domain of generating images from scene-based text descriptions.These approaches have primarily focused on generating images from a static text description and are limited to generating images in a single pass.They are unable to generate an image interactively based on an incrementally additive text description (something that is more intuitive and similar to the way we describe an image). We propose a method to generate an image incrementally based on a sequence of graphs of scene descriptions (scene-graphs).We propose a recurrent network architecture that preserves the image content generated in previous steps and modifies the cumulative image as per the newly provided scene information.Our model utilizes Graph Convolutional Networks (GCN) to cater to variable-sized scene graphs along with Generative Adversarial image translation networks to generate realistic multi-object images without needing any intermediate supervision during training.We experiment with Coco-Stuff dataset which has multi-object images along with annotations describing the visual scene and show that our model significantly outperforms other approaches on the same dataset in generating visually consistent images for incrementally growing scene graphs.", "target": ["Interaktive Bilderzeugung aus inkrementell wachsenden Szenegraphen in mehreren Schritten unter Verwendung von GANs unter Beibehaltung des Inhalts der in den vorherigen Schritten erzeugten Bilder"]} +{"source": "In some important computer vision domains, such as medical or hyperspectral imaging, we care about the classification of tiny objects in large images.However, most Convolutional Neural Networks (CNNs) for image classification were developed using biased datasets that contain large objects, in mostly central image positions.To assess whether classical CNN architectures work well for tiny object classification we build a comprehensive testbed containing two datasets: one derived from MNIST digits and one from histopathology images.This testbed allows controlled experiments to stress-test CNN architectures with a broad spectrum of signal-to-noise ratios.Our observations indicate that: (1) There exists a limit to signal-to-noise below which CNNs fail to generalize and that this limit is affected by dataset size - more data leading to better performances; however, the amount of training data required for the model to generalize scales rapidly with the inverse of the object-to-image ratio (2) in general, higher capacity models exhibit better generalization; (3) when knowing the approximate object sizes, adapting receptive field is beneficial; and (4) for very small signal-to-noise ratio the choice of global pooling operation affects optimization, whereas for relatively large signal-to-noise values, all tested global pooling operations exhibit similar performance.", "target": ["Wir untersuchen Klassifizierungsszenarien mit geringem und sehr geringem Signal-Rausch-Verhältnis, bei denen Objekte, die mit dem Klassenlabel korrelieren, einen winzigen Teil des gesamten Bildes einnehmen (z. B. medizinische oder hyperspektrale Bildgebung)."]} +{"source": "Recent trends of incorporating attention mechanisms in vision have led researchers to reconsider the supremacy of convolutional layers as a primary building block.Beyond helping CNNs to handle long-range dependencies, Ramachandran et al. (2019) showed that attention can completely replace convolution and achieve state-of-the-art performance on vision tasks.This raises the question: do learned attention layers operate similarly to convolutional layers?This work provides evidence that attention layers can perform convolution and, indeed, they often learn to do so in practice.Specifically, we prove that a multi-head self-attention layer with sufficient number of heads is at least as expressive as any convolutional layer.Our numerical experiments then show that self-attention layers attend to pixel-grid patterns similarly to CNN layers, corroborating our analysis.Our code is publicly available.", "target": ["Eine Selbstaufmerksamkeitsschicht kann Convolution durchführen und lernt dies oft in der Praxis."]} +{"source": "We introduce a “learning-based” algorithm for the low-rank decomposition problem: given an $n \\times d$ matrix $A$, and a parameter $k$, compute a rank-$k$ matrix $A'$ that minimizes the approximation loss $||A- A'||_F$.The algorithm uses a training set of input matrices in order to optimize its performance.Specifically, some of the most efficient approximate algorithms for computing low-rank approximations proceed by computing a projection $SA$, where $S$ is a sparse random $m \\times n$ “sketching matrix”, and then performing the singular value decomposition of $SA$.We show how to replace the random matrix $S$ with a “learned” matrix of the same sparsity to reduce the error.Our experiments show that, for multiple types of data sets, a learned sketch matrix can substantially reduce the approximation loss compared to a random matrix $S$, sometimes by one order of magnitude.We also study mixed matrices where only some of the rows are trained and the remaining ones are random, and show that matrices still offer improved performance while retaining worst-case guarantees.", "target": ["Lernbasierte Algorithmen können die Leistung klassischer Algorithmen für das Low-Rank-Approximationsproblem verbessern und gleichzeitig die Worst-Case-Garantie beibehalten."]} +{"source": "Neural conversational models are widely used in applications like personal assistants and chat bots.These models seem to give better performance when operating on word level.However, for fusion languages like French, Russian and Polish vocabulary size sometimes become infeasible since most of the words have lots of word forms.We propose a neural network architecture for transforming normalized text into a grammatically correct one.Our model efficiently employs correspondence between normalized and target words and significantly outperforms character-level models while being 2x faster in training and 20\\% faster at evaluation.We also propose a new pipeline for building conversational models: first generate a normalized answer and then transform it into a grammatically correct one using our network.The proposed pipeline gives better performance than character-level conversational models according to assessor testing.", "target": ["Vorgeschlagene Architektur zur Lösung der Aufgabe der morphologischen Übereinstimmung"]} +{"source": "This paper proposes the use of spectral element methods \\citep{canuto_spectral_1988} for fast and accurate training of Neural Ordinary Differential Equations (ODE-Nets; \\citealp{Chen2018NeuralOD}) for system identification.This is achieved by expressing their dynamics as a truncated series of Legendre polynomials.The series coefficients, as well as the network weights, are computed by minimizing the weighted sum of the loss function and the violation of the ODE-Net dynamics.The problem is solved by coordinate descent that alternately minimizes, with respect to the coefficients and the weights, two unconstrained sub-problems using standard backpropagation and gradient methods.The resulting optimization scheme is fully time-parallel and results in a low memory footprint.Experimental comparison to standard methods, such as backpropagation through explicit solvers and the adjoint technique \\citep{Chen2018NeuralOD}, on training surrogate models of small and medium-scale dynamical systems shows that it is at least one order of magnitude faster at reaching a comparable value of the loss function.The corresponding testing MSE is one order of magnitude smaller as well, suggesting generalization capabilities increase.", "target": ["In diesem Beitrag wird die Verwendung von Spektralelementmethoden für ein schnelles und genaues Training von neuronalen gewöhnlichen Differentialgleichungen zur Systemidentifikation vorgeschlagen."]} +{"source": "Exploration in sparse reward reinforcement learning remains an open challenge.Many state-of-the-art methods use intrinsic motivation to complement the sparse extrinsic reward signal, giving the agent more opportunities to receive feedback during exploration.Commonly these signals are added as bonus rewards, which results in a mixture policy that neither conducts exploration nor task fulfillment resolutely.In this paper, we instead learn separate intrinsic and extrinsic task policies and schedule between these different drives to accelerate exploration and stabilize learning.Moreover, we introduce a new type of intrinsic reward denoted as successor feature control (SFC), which is general and not task-specific.It takes into account statistics over complete trajectories and thus differs from previous methods that only use local information to evaluate intrinsic motivation.We evaluate our proposed scheduled intrinsic drive (SID) agent using three different environments with pure visual inputs: VizDoom, DeepMind Lab and DeepMind Control Suite.The results show a substantially improved exploration efficiency with SFC and the hierarchical usage of the intrinsic drives.A video of our experimental results can be found at https://gofile.io/?c=HpEwTd.", "target": ["Ein neues intrinsisches Belohnungssignal auf der Grundlage von Nachfolgemerkmalen und eine neuartige Methode zur Kombination von extrinsischer und intrinsischer Belohnung."]} +{"source": "Meta-Reinforcement learning approaches aim to develop learning procedures that can adapt quickly to a distribution of tasks with the help of a few examples.Developing efficient exploration strategies capable of finding the most useful samples becomes critical in such settings.Existing approaches to finding efficient exploration strategies add auxiliary objectives to promote exploration by the pre-update policy, however, this makes the adaptation using a few gradient steps difficult as the pre-update (exploration) and post-update (exploitation) policies are quite different.Instead, we propose to explicitly model a separate exploration policy for the task distribution.Having two different policies gives more flexibility in training the exploration policy and also makes adaptation to any specific task easier.We show that using self-supervised or supervised learning objectives for adaptation stabilizes the training process and also demonstrate the superior performance of our model compared to prior works in this domain.", "target": ["Wir schlagen vor, eine separate Explorationsstrategie zu verwenden, um die Trajektorien vor der Adaption in MAML zu sammeln. Wir zeigen auch, dass die Verwendung eines selbstüberwachten Ziels in der inneren Schleife zu einem stabileren Training und einer viel besseren Leistung führt."]} +{"source": "The “Supersymmetric Artificial Neural Network” in deep learning (denoted (x; θ, bar{θ})Tw), espouses the importance of considering biological constraints in the aim of further generalizing backward propagation. Looking at the progression of ‘solution geometries’; going from SO(n) representation (such as Perceptron like models) to SU(n) representation (such as UnitaryRNNs) has guaranteed richer and richer representations in weight space of the artificial neural network, and hence better and better hypotheses were generatable.The Supersymmetric Artificial Neural Network explores a natural step forward, namely SU(m|n) representation.These supersymmetric biological brain representations (Perez et al.) can be represented by supercharge compatible special unitary notation SU(m|n), or (x; θ, bar{θ})Tw parameterized by θ, bar{θ}, which are supersymmetric directions, unlike θ seen in the typical non-supersymmetric deep learning model.Notably, Supersymmetric values can encode or represent more information than the typical deep learning model, in terms of “partner potential” signals for example.", "target": ["Verallgemeinerung der Rückwärtsausbreitung unter Verwendung formaler Methoden der Supersymmetrie."]} +{"source": "Regularization-based continual learning approaches generally prevent catastrophic forgetting by augmenting the training loss with an auxiliary objective.However in most practical optimization scenarios with noisy data and/or gradients, it is possible that stochastic gradient descent can inadvertently change critical parameters.In this paper, we argue for the importance of regularizing optimization trajectories directly.We derive a new co-natural gradient update rule for continual learning whereby the new task gradients are preconditioned with the empirical Fisher information of previously learnt tasks.We show that using the co-natural gradient systematically reduces forgetting in continual learning.Moreover, it helps combat overfitting when learning a new task in a low resource scenario.", "target": ["Die Regulierung der Optimierungskurve mit den Fisher-Informationen der alten Aufgaben reduziert das katastrophale Vergessen erheblich"]} +{"source": "We study the problem of generating source code in a strongly typed,Java-like programming language, given a label (for example a set ofAPI calls or types) carrying a small amount of information about thecode that is desired.The generated programs are expected to respect a`\"realistic\" relationship between programs and labels, as exemplifiedby a corpus of labeled programs available during training.Two challenges in such *conditional program generation* are thatthe generated programs must satisfy a rich set of syntactic andsemantic constraints, and that source code contains many low-levelfeatures that impede learning. We address these problems by traininga neural generator not on code but on *program sketches*, ormodels of program syntax that abstract out names and operations thatdo not generalize across programs.During generation, we infer aposterior distribution over sketches, then concretize samples fromthis distribution into type-safe programs using combinatorialtechniques. We implement our ideas in a system for generatingAPI-heavy Java code, and show that it can often predict the entirebody of a method given just a few API calls or data types that appearin the method.", "target": ["Wir geben eine Methode zur Generierung von typsicheren Programmen in einer Java-ähnlichen Sprache an, die eine kleine Menge an syntaktischen Informationen über den gewünschten Code enthält."]} +{"source": "We propose an approach for sequence modeling based on autoregressive normalizing flows.Each autoregressive transform, acting across time, serves as a moving reference frame for modeling higher-level dynamics.This technique provides a simple, general-purpose method for improving sequence modeling, with connections to existing and classical techniques.We demonstrate the proposed approach both with standalone models, as well as a part of larger sequential latent variable models.Results are presented on three benchmark video datasets, where flow-based dynamics improve log-likelihood performance over baseline models.", "target": ["Wir zeigen, wie autoregressive Flüsse verwendet werden können, um sequentielle latente Variablenmodelle zu verbessern."]} +{"source": "It is well-known that many machine learning models are susceptible to adversarial attacks, in which an attacker evades a classifier by making small perturbations to inputs.This paper discusses how industrial copyright detection tools, which serve a central role on the web, are susceptible to adversarial attacks.We discuss a range of copyright detection systems, and why they are particularly vulnerable to attacks. These vulnerabilities are especially apparent for neural network based systems. As proof of concept, we describe a well-known music identification method and implement this system in the form of a neural net.We then attack this system using simple gradient methods.Adversarial music created this way successfully fools industrial systems, including the AudioTag copyright detector and YouTube's Content ID system.Our goal is to raise awareness of the threats posed by adversarial examples in this space and to highlight the importance of hardening copyright detection systems to attacks.", "target": ["Widersprüchliche Beispiele können das Urheberrechtserkennungssystem von YouTube täuschen"]} +{"source": "Equilibrium Propagation (EP) is a learning algorithm that bridges Machine Learning and Neuroscience, by computing gradients closely matching those of Backpropagation Through Time (BPTT), but with a learning rule local in space.Given an input x and associated target y, EP proceeds in two phases: in the first phase neurons evolve freely towards a first steady state; in the second phase output neurons are nudged towards y until they reach a second steady state.However, in existing implementations of EP, the learning rule is not local in time:the weight update is performed after the dynamics of the second phase have converged and requires information of the first phase that is no longer available physically.This is a major impediment to the biological plausibility of EP and its efficient hardware implementation.In this work, we propose a version of EP named Continual Equilibrium Propagation (C-EP) where neuron and synapse dynamics occur simultaneously throughout the second phase, so that the weight update becomes local in time.We prove theoretically that, provided the learning rates are sufficiently small, at each time step of the second phase the dynamics of neurons and synapses follow the gradients of the loss given by BPTT (Theorem 1).We demonstrate training with C-EP on MNIST and generalize C-EP to neural networks where neurons are connected by asymmetric connections.We show through experiments that the more the network updates follows the gradients of BPTT, the best it performs in terms of training.These results bring EP a step closer to biology while maintaining its intimate link with backpropagation.", "target": ["Wir schlagen eine kontinuierliche Version der Gleichgewichtsausbreitung vor, bei der Neuronen- und Synapsendynamik während der gesamten zweiten Phase gleichzeitig auftreten, mit theoretischen Garantien und numerischen Simulationen."]} +{"source": "There are two main lines of research on visual reasoning: neural module network (NMN) with explicit multi-hop reasoning through handcrafted neural modules, and monolithic network with implicit reasoning in the latent feature space.The former excels in interpretability and compositionality, while the latter usually achieves better performance due to model flexibility and parameter efficiency. In order to bridge the gap of the two, we present Meta Module Network (MMN), a novel hybrid approach that can efficiently utilize a Meta Module to perform versatile functionalities, while preserving compositionality and interpretability through modularized design.The proposed model first parses an input question into a functional program through a Program Generator.Instead of handcrafting a task-specific network to represent each function like traditional NMN, we use Recipe Encoder to translate the functions into their corresponding recipes (specifications), which are used to dynamically instantiate the Meta Module into Instance Modules.To endow different instance modules with designated functionality, a Teacher-Student framework is proposed, where a symbolic teacher pre-executes against the scene graphs to provide guidelines for the instantiated modules (student) to follow.In a nutshell, MMN adopts the meta module to increase its parameterization efficiency, and uses recipe encoding to improve its generalization ability over NMN.Experiments conducted on the GQA benchmark demonstrates that: (1) MMN achieves significant improvement over both NMN and monolithic network baselines; (2) MMN is able to generalize to unseen but related functions.", "target": ["Wir schlagen ein neues Meta-Modul-Netzwerk vor, um einige der Einschränkungen früherer neuronaler Modulnetzwerke zu beseitigen und eine starke Leistung auf realistischen visuellen Reasoning-Datensätzen zu erzielen."]} +{"source": "We propose a new perspective on adversarial attacks against deep reinforcement learning agents.Our main contribution is CopyCAT, a targeted attack able to consistently lure an agent into following an outsider's policy.It is pre-computed, therefore fast inferred, and could thus be usable in a real-time scenario.We show its effectiveness on Atari 2600 games in the novel read-only setting.In the latter, the adversary cannot directly modify the agent's state -its representation of the environment- but can only attack the agent's observation -its perception of the environment.Directly modifying the agent's state would require a write-access to the agent's inner workings and we argue that this assumption is too strong in realistic settings.", "target": ["Wir schlagen einen neuen Angriff vor, um die volle Kontrolle über neuronale Richtlinien in realistischen Umgebungen zu übernehmen."]} +{"source": "Cold-start and efficiency issues of the Top-k recommendation are critical to large-scale recommender systems.Previous hybrid recommendation methods are effective to deal with the cold-start issues by extracting real latent factors of cold-start items(users) from side information, but they still suffer low efficiency in online recommendation caused by the expensive similarity search in real latent space.This paper presents a collaborative generated hashing (CGH) to improve the efficiency by denoting users and items as binary codes, which applies to various settings: cold-start users, cold-start items and warm-start ones.Specifically, CGH is designed to learn hash functions of users and items through the Minimum Description Length (MDL) principle; thus, it can deal with various recommendation settings.In addition, CGH initiates a new marketing strategy through mining potential users by a generative step.To reconstruct effective users, the MDL principle is used to learn compact and informative binary codes from the content data.Extensive experiments on two public datasets show the advantages for recommendations in various settings over competing baselines and analyze the feasibility of the application in marketing.", "target": ["Es kann effektive Hash-Codes für eine effiziente Kaltstart-Empfehlung generieren und gleichzeitig eine praktikable Marketingstrategie bieten."]} +{"source": "Recent efforts to combine Representation Learning with Formal Methods, commonly known as the Neuro-Symbolic Methods, have given rise to a new trend of applying rich neural architectures to solve classical combinatorial optimization problems.In this paper, we propose a neural framework that can learn to solve the Circuit Satisfiability problem.Our framework is built upon two fundamental contributions: a rich embedding architecture that encodes the problem structure and an end-to-end differentiable training procedure that mimics Reinforcement Learning and trains the model directly toward solving the SAT problem.The experimental results show the superior out-of-sample generalization performance of our framework compared to the recently developed NeuroSAT method.", "target": ["Wir schlagen einen neuronalen Rahmen vor, der lernen kann, das Problem der Erfüllbarkeit von Schaltkreisen anhand von (unbeschrifteten) Schaltkreisinstanzen zu lösen."]} +{"source": "Sequence generation models such as recurrent networks can be trained with a diverse set of learning algorithms.For example, maximum likelihood learning is simple and efficient, yet suffers from the exposure bias problem.Reinforcement learning like policy gradient addresses the problem but can have prohibitively poor exploration efficiency.A variety of other algorithms such as RAML, SPG, and data noising, have also been developed in different perspectives.This paper establishes a formal connection between these algorithms.We present a generalized entropy regularized policy optimization formulation, and show that the apparently divergent algorithms can all be reformulated as special instances of the framework, with the only difference being the configurations of reward function and a couple of hyperparameters.The unified interpretation offers a systematic view of the varying properties of exploration and learning efficiency.Besides, based on the framework, we present a new algorithm that dynamically interpolates among the existing algorithms for improved learning.Experiments on machine translation and text summarization demonstrate the superiority of the proposed algorithm.", "target": ["Eine einheitliche Sichtweise verschiedener Lernalgorithmen für die Sequenzerzeugung, wie MLE, RL, RAML, Datenrauschen usw."]} +{"source": "We are reporting the SHINRA project, a project for structuring Wikipedia with collaborative construction scheme.The goal of the project is to create a huge and well-structured knowledge base to be used in NLP applications, such as QA, Dialogue systems and explainable NLP systems.It is created based on a scheme of ”Resource by Collaborative Contribution (RbCC)”.We conducted a shared task of structuring Wikipedia, and at the same, submitted results are used to construct a knowledge base.There are machine readable knowledge bases such as CYC, DBpedia, YAGO, Freebase Wikidata and so on, but each of them has problems to be solved.CYC has a coverage problem, and others have a coherence problem due to the fact that these are based on Wikipedia and/or created by many but inherently incoherent crowd workers.In order to solve the later problem, we started a project for structuring Wikipedia using automatic knowledge base construction shared-task.The automatic knowledge base construction shared-tasks have been popular and well studied for decades.However, these tasks are designed only to compare the performances of different systems, and to find which system ranks the best on limited test data.The results of the participated systems are not shared and the systems may be abandoned once the task is over.We believe this situation can be improved by the following changes:1. designing the shared-task to construct knowledge base rather than evaluating only limited test data2. making the outputs of all the systems open to public so that we can run ensemble learning to create the better results than the best systems3. repeating the task so that we can run the task with the larger and better training data from the output of the previous task (bootstrapping and active learning)We conducted “SHINRA2018” with the above mentioned scheme and in this paperwe report the results and the future directions of the project.The task is to extract the values of the pre-defined attributes from Wikipedia pages.We have categorized most of the entities in Japanese Wikipedia (namely 730 thousand entities) into the 200 ENE categories.Based on this data, the shared-task is to extract the values of the attributes from Wikipedia pages.We gave out the 600 training data and the participants are required to submit the attribute-values for all remaining entities of the same category type.Then 100 data out of them for each category are used to evaluate the system output in the shared-task.We conducted a preliminary ensemble learning on the outputs and found 15 F1 score improvement on a category and the average of 8 F1 score improvements on all 5 categories we tested over a strong baseline.Based on this promising results, we decided to conduct three tasks in 2019; multi-lingual categorization task (ML), extraction for the same 5 categories in Japanese with a larger training data (JP-5) and extraction for 34 new categories in Japanese (JP-34).", "target": ["Wir führen ein \"Resource by Collaborative Construction\"-Schema ein, um KB, strukturierte Wikipedia zu erstellen"]} +{"source": "Recent image super-resolution(SR) studies leverage very deep convolutional neural networks and the rich hierarchical features they offered, which leads to better reconstruction performance than conventional methods.However, the small receptive fields in the up-sampling and reconstruction process of those models stop them to take full advantage of global contextual information.This causes problems for further performance improvement.In this paper, inspired by image reconstruction principles of human visual system, we propose an image super-resolution global reasoning network (SRGRN) to effectively learn the correlations between different regions of an image, through global reasoning.Specifically, we propose global reasoning up-sampling module (GRUM) and global reasoning reconstruction block (GRRB).They construct a graph model to perform relation reasoning on regions of low resolution (LR) images.They aim to reason the interactions between different regions in the up-sampling and reconstruction process and thus leverage more contextual information to generate accurate details.Our proposed SRGRN are more robust and can handle low resolution images that are corrupted by multiple types of degradation.Extensive experiments on different benchmark data-sets show that our model outperforms other state-of-the-art methods.Also our model is lightweight and consumes less computing power, which makes it very suitable for real life deployment.", "target": ["Ein hochmodernes Modell auf der Grundlage globaler Schlussfolgerungen für die Superauflösung von Bildern"]} +{"source": "Graph Neural Networks (GNNs) have received tremendous attention recently due to their power in handling graph data for different downstream tasks across different application domains.The key of GNN is its graph convolutional filters, and recently various kinds of filters are designed.However, there still lacks in-depth analysis on (1) Whether there exists a best filter that can perform best on all graph data; (2) Which graph properties will influence the optimal choice of graph filter; (3) How to design appropriate filter adaptive to the graph data.In this paper, we focus on addressing the above three questions.We first propose a novel assessment tool to evaluate the effectiveness of graph convolutional filters for a given graph.Using the assessment tool, we find out that there is no single filter as a `silver bullet' that perform the best on all possible graphs.In addition, different graph structure properties will influence the optimal graph convolutional filter's design choice.Based on these findings, we develop Adaptive Filter Graph Neural Network (AFGNN), a simple but powerful model that can adaptively learn task-specific filter.For a given graph, it leverages graph filter assessment as regularization and learns to combine from a set of base filters.Experiments on both synthetic and real-world benchmark datasets demonstrate that our proposed model can indeed learn an appropriate filter and perform well on graph tasks.", "target": ["Vorschlag für einen Bewertungsrahmen zur Analyse und zum Erlernen von Graph Convolutional Filtern"]} +{"source": "The advance of node pooling operations in Graph Neural Networks (GNNs) has lagged behind the feverish design of new message-passing techniques, and pooling remains an important and challenging endeavor for the design of deep architectures.In this paper, we propose a pooling operation for GNNs that leverages a differentiable unsupervised loss based on the minCut optimization objective.For each node, our method learns a soft cluster assignment vector that depends on the node features, the target inference task (e.g., a graph classification loss), and, thanks to the minCut objective, also on the connectivity structure of the graph.Graph pooling is obtained by applying the matrix of assignment vectors to the adjacency matrix and the node features.We validate the effectiveness of the proposed pooling method on a variety of supervised and unsupervised tasks.", "target": ["Eine neue Pooling-Schicht für GNNs, die lernt, wie Knoten entsprechend ihrer Eigenschaften, der Graphenkonnektivität und dem Ziel der nachgelagerten Aufgabe zu poolen sind."]} +{"source": "Knowledge graphs are structured representations of real world facts.However, they typically contain only a small subset of all possible facts.Link prediction is the task of inferring missing facts based on existing ones.We propose TuckER, a relatively simple yet powerful linear model based on Tucker decomposition of the binary tensor representation of knowledge graph triples.By using this particular decomposition, parameters are shared between relations, enabling multi-task learning.TuckER outperforms previous state-of-the-art models across several standard link prediction datasets.", "target": ["Wir schlagen TuckER vor, ein relativ einfaches, aber leistungsfähiges lineares Modell zur Linkvorhersage in Wissensgraphen, das auf der Tucker-Dekomposition der binären Tensordarstellung von Wissensgraphen-Tripeln basiert."]} +{"source": "With innovations in architecture design, deeper and wider neural network models deliver improved performance on a diverse variety of tasks.But the increased memory footprint of these models presents a challenge during training, when all intermediate layer activations need to be stored for back-propagation.Limited GPU memory forces practitioners to make sub-optimal choices: either train inefficiently with smaller batches of examples; or limit the architecture to have lower depth and width, and fewer layers at higher spatial resolutions.This work introduces an approximation strategy that significantly reduces a network's memory footprint during training, but has negligible effect on training performance and computational expense.During the forward pass, we replace activations with lower-precision approximations immediately after they have been used by subsequent layers, thus freeing up memory.The approximate activations are then used during the backward pass.This approach limits the accumulation of errors across the forward and backward pass---because the forward computation across the network still happens at full precision, and the approximation has a limited effect when computing gradients to a layer's input.Experiments, on CIFAR and ImageNet, show that using our approach with 8- and even 4-bit fixed-point approximations of 32-bit floating-point activations has only a minor effect on training and validation performance, while affording significant savings in memory usage.", "target": ["Ein Algorithmus zur Verringerung des Speicherbedarfs für das Training tiefer Netzwerke, der auf einer Annäherungsstrategie basiert."]} +{"source": "Estimating the frequencies of elements in a data stream is a fundamental task in data analysis and machine learning.The problem is typically addressed using streaming algorithms which can process very large data using limited storage.Today's streaming algorithms, however, cannot exploit patterns in their input to improve performance.We propose a new class of algorithms that automatically learn relevant patterns in the input data and use them to improve its frequency estimates. The proposed algorithms combine the benefits of machine learning with the formal guarantees available through algorithm theory. We prove that our learning-based algorithms have lower estimation errors than their non-learning counterparts. We also evaluate our algorithms on two real-world datasets and demonstrate empirically their performance gains.", "target": ["Algorithmen für Datenströme können durch Deep Learning verbessert werden, ohne dass die Leistung beeinträchtigt wird."]} +{"source": "Link prediction in simple graphs is a fundamental problem in which new links between nodes are predicted based on the observed structure of the graph.However, in many real-world applications, there is a need to model relationships among nodes which go beyond pairwise associations.For example, in a chemical reaction, relationship among the reactants and products is inherently higher-order.Additionally, there is need to represent the direction from reactants to products.Hypergraphs provide a natural way to represent such complex higher-order relationships.Even though Graph Convolutional Networks (GCN) have recently emerged as a powerful deep learning-based approach for link prediction over simple graphs, their suitability for link prediction in hypergraphs is unexplored -- we fill this gap in this paper and propose Neural Hyperlink Predictor (NHP). NHP adapts GCNs for link prediction in hypergraphs. We propose two variants of NHP --NHP-U and NHP-D -- for link prediction over undirected and directed hypergraphs, respectively.To the best of our knowledge, NHP-D is the first method for link prediction over directed hypergraphs.Through extensive experiments on multiple real-world datasets, we show NHP's effectiveness.", "target": ["Wir schlagen den Neural Hyperlink Predictor (NHP) vor. NHP adaptiert graphische Convolutional Networks für die Link-Vorhersage in Hypergraphen"]} +{"source": "In this paper, we present a method for adversarial decomposition of text representation.This method can be used to decompose a representation of an input sentence into several independent vectors, where each vector is responsible for a specific aspect of the input sentence.We evaluate the proposed method on two case studies: the conversion between different social registers and diachronic language change.We show that the proposed method is capable of fine-grained con- trolled change of these aspects of the input sentence.For example, our model is capable of learning a continuous (rather than categorical) representation of the style of the sentence, in line with the reality of language use.The model uses adversarial-motivational training and includes a special motivational loss, which acts opposite to the discriminator and encourages a better decomposition.Finally, we evaluate the obtained meaning embeddings on a downstream task of para- phrase detection and show that they are significantly better than embeddings of a regular autoencoder.", "target": ["Eine Methode, die separate Repräsentationen für die Bedeutung und die Form eines Satzes lernt"]} +{"source": "We were approached by a group of healthcare providers who are involved in the care of chronic patients looking for potential technologies to facilitate the process of reviewing patient-generated data during clinical visits. Aiming at understanding the healthcare providers' attitudes towards reviewing patient-generated data, we (1) conducted a focus group with a mixed group of healthcare providers.Next, to gain the patients' perspectives, we (2) interviewed eight chronic patients, collected a sample of their data and designed a series of visualizations representing patient data we collected.Last, we (3) sought feedback on the visualization designs from healthcare providers who requested this exploration.We found four factors shaping patient-generated data: data & context, patient's motivation, patient's time commitment, and patient's support circle.Informed by the results of our studies, we discussed the importance of designing patient-generated visualizations for individuals by considering both patient and healthcare provider rather than designing with the purpose of generalization and provided guidelines for designing future patient-generated data visualizations.", "target": ["Wir haben untersucht, welche Visualisierungen chronische Patienten dabei unterstützen können, ihre Gesundheitsdaten bei Arztbesuchen zu präsentieren und zu überprüfen."]} +{"source": "Recently, neural-network based forward dynamics models have been proposed that attempt to learn the dynamics of physical systems in a deterministic way.While near-term motion can be predicted accurately, long-term predictions suffer from accumulating input and prediction errors which can lead to plausible but different trajectories that diverge from the ground truth. A system that predicts distributions of the future physical states for long time horizons based on its uncertainty is thus a promising solution. In this work, we introduce a novel robust Monte Carlo sampling based graph-convolutional dropout method that allows us to sample multiple plausible trajectories for an initial state given a neural-network based forward dynamics predictor. By introducing a new shape preservation loss and training our dynamics model recurrently, we stabilize long-term predictions.We show that our model’s long-term forward dynamics prediction errors on complicated physical interactions of rigid and deformable objects of various shapes are significantly lower than existing strong baselines.Lastly, we demonstrate how generating multiple trajectories with our Monte Carlo dropout method can be used to train model-free reinforcement learning agents faster and to better solutions on simple manipulation tasks.", "target": ["Wir schlagen einen stochastisch differenzierbaren Vorwärtsdynamik-Prädiktor vor, der in der Lage ist, mehrere physikalisch plausible Trajektorien unter demselben anfänglichen Eingangszustand abzutasten, und zeigen, dass er verwendet werden kann, um modellfreie Strategien effizienter zu trainieren."]} +{"source": "There has been a large amount of interest, both in the past and particularly recently, into the relative advantage of different families of universal function approximators, for instance neural networks, polynomials, rational functions, etc.However, current research has focused almost exclusively on understanding this problem in a worst case setting: e.g. characterizing the best L1 or L_{infty} approximation in a box (or sometimes, even under an adversarially constructed data distribution.)In this setting many classical tools from approximation theory can be effectively used.However, in typical applications we expect data to be high dimensional, but structured -- so, it would only be important to approximate the desired function well on the relevant part of its domain, e.g. a small manifold on which real input data actually lies.Moreover, even within this domain the desired quality of approximation may not be uniform; for instance in classification problems, the approximation needs to be more accurate near the decision boundary.These issues, to the best of our knowledge, have remain unexplored until now.\tWith this in mind, we analyze the performance of neural networks and polynomial kernels in a natural regression setting where the data enjoys sparse latent structure, and the labels depend in a simple way on the latent variables.We give an almost-tight theoretical analysis of the performance of both neural networks and polynomials for this problem, as well as verify our theory with simulations.Our results both involve new (complex-analytic) techniques, which may be of independent interest, and show substantial qualitative differences with what is known in the worst-case setting.", "target": ["Beyond-Worst-Case-Analyse der Repräsentationskraft von ReLU-Netzen und polynomialen Kerneln - insbesondere bei Vorhandensein einer spärlichen latenten Struktur."]} +{"source": "Recent deep generative models can provide photo-realistic images as well as visual or textual content embeddings useful to address various tasks of computer vision and natural language processing.Their usefulness is nevertheless often limited by the lack of control over the generative process or the poor understanding of the learned representation.To overcome these major issues, very recent works have shown the interest of studying the semantics of the latent space of generative models.In this paper, we propose to advance on the interpretability of the latent space of generative models by introducing a new method to find meaningful directions in the latent space of any generative model along which we can move to control precisely specific properties of the generated image like position or scale of the object in the image.Our method is weakly supervised and particularly well suited for the search of directions encoding simple transformations of the generated image, such as translation, zoom or color variations.We demonstrate the effectiveness of our method qualitatively and quantitatively, both for GANs and variational auto-encoders.", "target": ["Ein Modell zur Steuerung der Bilderzeugung mit GAN und beta-VAE in Bezug auf Maßstab und Position der Objekte"]} +{"source": "Modern neural network architectures use structured linear transformations, such as low-rank matrices, sparse matrices, permutations, and the Fourier transform, to improve inference speed and reduce memory usage compared to general linear maps.However, choosing which of the myriad structured transformations to use (and its associated parameterization) is a laborious task that requires trading off speed, space, and accuracy.We consider a different approach: we introduce a family of matrices called kaleidoscope matrices (K-matrices) that provably capture any structured matrix with near-optimal space (parameter) and time (arithmetic operation) complexity.We empirically validate that K-matrices can be automatically learned within end-to-end pipelines to replace hand-crafted procedures, in order to improve model quality.For example, replacing channel shuffles in ShuffleNet improves classification accuracy on ImageNet by up to 5%.Learnable K-matrices can also simplify hand-engineered pipelines---we replace filter bank feature computation in speech data preprocessing with a kaleidoscope layer, resulting in only 0.4% loss in accuracy on the TIMIT speech recognition task.K-matrices can also capture latent structure in models: for a challenging permuted image classification task, adding a K-matrix to a standard convolutional architecture can enable learning the latent permutation and improve accuracy by over 8 points.We provide a practically efficient implementation of our approach, and use K-matrices in a Transformer network to attain 36% faster end-to-end inference speed on a language translation task.", "target": ["Wir schlagen eine differenzierbare Familie von \"Kaleidoskop-Matrizen\" vor, beweisen, dass alle strukturierten Matrizen in dieser Form dargestellt werden können, und verwenden sie, um handgefertigte lineare Zuordnungen in Deep-Learning-Modellen zu ersetzen."]} +{"source": "We propose a method, called Label Embedding Network, which can learn label representation (label embedding) during the training process of deep networks.With the proposed method, the label embedding is adaptively and automatically learned through back propagation.The original one-hot represented loss function is converted into a new loss function with soft distributions, such that the originally unrelated labels have continuous interactions with each other during the training process.As a result, the trained model can achieve substantially higher accuracy and with faster convergence speed.Experimental results based on competitive tasks demonstrate the effectiveness of the proposed method, and the learned label embedding is reasonable and interpretable.The proposed method achieves comparable or even better results than the state-of-the-art systems.", "target": ["Lernen der Label-Repräsentation für tiefe Netzwerke"]} +{"source": "Decentralized training of deep learning models is a key element for enabling data privacy and on-device learning over networks, as well as for efficient scaling to large compute clusters.As current approaches are limited by network bandwidth, we propose the use of communication compression in the decentralized training context.We show that Choco-SGD achieves linear speedup in the number of workers for arbitrary high compression ratios on general non-convex functions, and non-IID training data. We demonstrate the practical performance of the algorithm in two key scenarios: the training of deep learning models(i) over decentralized user devices, connected by a peer-to-peer network and(ii) in a datacenter.", "target": ["Wir schlagen Choco-SGD--dezentralisiertes SGD mit komprimierter Kommunikation--für nicht-konvexe Ziele vor und zeigen seine starke Leistung in verschiedenen Deep-Learning-Anwendungen (On-Device-Lernen, Rechenzentrumsfall)."]} +{"source": "We show that Entropy-SGD (Chaudhari et al., 2017), when viewed as a learning algorithm, optimizes a PAC-Bayes bound on the risk of a Gibbs (posterior) classifier, i.e., a randomized classifier obtained by a risk-sensitive perturbation of the weights of a learned classifier.Entropy-SGD works by optimizing the bound’s prior, violating the hypothesis of the PAC-Bayes theorem that the prior is chosen independently of the data.Indeed, available implementations of Entropy-SGD rapidly obtain zero training error on random labels and the same holds of the Gibbs posterior.In order to obtain a valid generalization bound, we show that an ε-differentially private prior yields a valid PAC-Bayes bound, a straightforward consequence of results connecting generalization with differential privacy.Using stochastic gradient Langevin dynamics (SGLD) to approximate the well-known exponential release mechanism, we observe that generalization error on MNIST (measured on held out data) falls within the (empirically nonvacuous) bounds computed under the assumption that SGLD produces perfect samples.In particular, Entropy-SGLD can be configured to yield relatively tight generalization bounds and still fit real labels, although these same settings do not obtain state-of-the-art performance.", "target": ["Wir zeigen, dass Entropie-SGD den Vorgänger einer PAC-Bayes-Schranke optimiert und damit gegen die Anforderung verstößt, dass der Vorgänger unabhängig von den Daten sein muss; wir verwenden Differential Privacy, um dieses Problem zu lösen und die Verallgemeinerung zu verbessern."]} +{"source": "In this paper, we investigate learning the deep neural networks for automated optical inspection in industrial manufacturing.Our preliminary result has shown the stunning performance improvement by transfer learning from the completely dissimilar source domain: ImageNet.Further study for demystifying this improvement shows that the transfer learning produces a highly compressible network, which was not the case for the network learned from scratch.The experimental result shows that there is a negligible accuracy drop in the network learned by transfer learning until it is compressed to 1/128 reduction of the number of convolution filters.This result is contrary to the compression without transfer learning which loses more than 5% accuracy at the same compression rate.", "target": ["Wir zeigen experimentell, dass Transfer-Lernen spärliche Merkmale im Netzwerk erzeugt und dadurch ein komprimierbareres Netzwerk produziert."]} +{"source": "The key challenge in semi-supervised learning is how to effectively leverage unlabeled data to improve learning performance.The classical label propagation method, despite its popularity, has limited modeling capability in that it only exploits graph information for making predictions.In this paper, we consider label propagation from a graph signal processing perspective and decompose it into three components: signal, filter, and classifier.By extending the three components, we propose a simple generalized label propagation (GLP) framework for semi-supervised learning.GLP naturally integrates graph and data feature information, and offers the flexibility of selecting appropriate filters and domain-specific classifiers for different applications.Interestingly, GLP also provides new insight into the popular graph convolutional network and elucidates its working mechanisms.Extensive experiments on three citation networks, one knowledge graph, and one image dataset demonstrate the efficiency and effectiveness of GLP.", "target": ["Wir erweitern die klassischen Label-Propation-Methoden zur gemeinsamen Modellierung von Graphen- und Merkmalsinformationen aus der Perspektive der Graphenfilterung und zeigen Verbindungen zu den Graph Convolutional Networks auf."]} +{"source": "Because the choice and tuning of the optimizer affects the speed, and ultimately the performance of deep learning, there is significant past and recent research in this area.Yet, perhaps surprisingly, there is no generally agreed-upon protocol for the quantitative and reproducible evaluation of optimization strategies for deep learning.We suggest routines and benchmarks for stochastic optimization, with special focus on the unique aspects of deep learning, such as stochasticity, tunability and generalization.As the primary contribution, we present DeepOBS, a Python package of deep learning optimization benchmarks.The package addresses key challenges in the quantitative assessment of stochastic optimizers, and automates most steps of benchmarking.The library includes a wide and extensible set of ready-to-use realistic optimization problems, such as training Residual Networks for image classification on ImageNet or character-level language prediction models, as well as popular classics like MNIST and CIFAR-10.The package also provides realistic baseline results for the most popular optimizers on these test problems, ensuring a fair comparison to the competition when benchmarking new optimizers, and without having to run costly experiments.It comes with output back-ends that directly produce LaTeX code for inclusion in academic publications.It supports TensorFlow and is available open source.", "target": ["Wir bieten ein Softwarepaket an, das die Bewertung von Deep-Learning-Optimierern drastisch vereinfacht, automatisiert und verbessert."]} +{"source": "Pre-trained word embeddings are the primarymethod for transfer learning in several Natural Language Processing (NLP) tasks.Recentworks have focused on using unsupervisedtechniques such as language modeling to obtain these embeddings.In contrast, this workfocuses on extracting representations frommultiple pre-trained supervised models, whichenriches word embeddings with task and domain specific knowledge.Experiments performed in cross-task, cross-domain and crosslingual settings indicate that such supervisedembeddings are helpful, especially in the lowresource setting, but the extent of gains is dependent on the nature of the task and domain.", "target": ["Extrahieren von kontextuellen Einbettungen aus einem standardmäßigen überwachten Modell. Unterstützt nachgelagerte NLP-Modelle in ressourcenarmen Umgebungen"]} +{"source": "We build a theoretical framework for understanding practical meta-learning methods that enables the integration of sophisticated formalizations of task-similarity with the extensive literature on online convex optimization and sequential prediction algorithms in order to provide within-task performance guarantees.Our approach improves upon recent analyses of parameter-transfer by enabling the task-similarity to be learned adaptively and by improving transfer-risk bounds in the setting of statistical learning-to-learn.It also leads to straightforward derivations of average-case regret bounds for efficient algorithms in settings where the task-environment changes dynamically or the tasks share a certain geometric structure.", "target": ["Praktische adaptive Algorithmen für gradientenbasiertes Meta-Lernen mit nachweisbaren Garantien."]} +{"source": "In this work, we propose a self-supervised method to learn sentence representations with an injection of linguistic knowledge.Multiple linguistic frameworks propose diverse sentence structures from which semantic meaning might be expressed out of compositional words operations.We aim to take advantage of this linguist diversity and learn to represent sentences by contrasting these diverse views.Formally, multiple views of the same sentence are mapped to close representations.On the contrary, views from other sentences are mapped further.By contrasting different linguistic views, we aim at building embeddings which better capture semantic and which are less sensitive to the sentence outward form.", "target": ["Unser Ziel ist es, die Vielfalt der sprachlichen Strukturen für die Erstellung von Satzrepräsentationen zu nutzen."]} +{"source": "The peripheral nervous system represents the input/output system for the brain.Cuff electrodes implanted on the peripheral nervous system allow observation and control over this system, however, the data produced by these electrodes have a low signal-to-noise ratio and a complex signal content.In this paper, we consider the analysis of neural data recorded from the vagus nerve in animal models, and develop an unsupervised learner based on convolutional neural networks that is able to simultaneously de-noise and cluster regions of the data by signal content.", "target": ["Die unüberwachte Analyse von Daten, die vom peripheren Nervensystem aufgezeichnet wurden, entrauscht und kategorisiert die Signale."]} +{"source": "Adversarial attacks on convolutional neural networks (CNN) have gained significant attention and there have been active research efforts on defense mechanisms.Stochastic input transformation methods have been proposed, where the idea is to recover the image from adversarial attack by random transformation, and to take the majority vote as consensus among the random samples.However, the transformation improves the accuracy on adversarial images at the expense of the accuracy on clean images.While it is intuitive that the accuracy on clean images would deteriorate, the exact mechanism in which how this occurs is unclear.In this paper, we study the distribution of softmax induced by stochastic transformations.We observe that with random transformations on the clean images, although the mass of the softmax distribution could shift to the wrong class, the resulting distribution of softmax could be used to correct the prediction.Furthermore, on the adversarial counterparts, with the image transformation, the resulting shapes of the distribution of softmax are similar to the distributions from the clean images.With these observations, we propose a method to improve existing transformation-based defenses.We train a separate lightweight distribution classifier to recognize distinct features in the distributions of softmax outputs of transformed images.Our empirical studies show that our distribution classifier, by training on distributions obtained from clean images only, outperforms majority voting for both clean and adversarial images.Our method is generic and can be integrated with existing transformation-based defenses.", "target": ["Wir verbessern die bestehenden transformationsbasierten Abwehrmechanismen, indem wir einen Verteilungsklassifikator für die Verteilung von Softmax aus transformierten Bildern verwenden."]} +{"source": "Reinforcement learning in multi-agent scenarios is important for real-world applications but presents challenges beyond those seen in single-agent settings.We present an actor-critic algorithm that trains decentralized policies in multi-agent settings, using centrally computed critics that share an attention mechanism which selects relevant information for each agent at every timestep.This attention mechanism enables more effective and scalable learning in complex multi-agent environments, when compared to recent approaches.Our approach is applicable not only to cooperative settings with shared rewards, but also individualized reward settings, including adversarial settings, and it makes no assumptions about the action spaces of the agents.As such, it is flexible enough to be applied to most multi-agent learning problems", "target": ["Wir schlagen einen Ansatz zum Erlernen dezentraler Strategien in Multi-Agenten-Umgebungen unter Verwendung aufmerksamkeitsbasierter Kritiker vor und zeigen vielversprechende Ergebnisse in Umgebungen mit komplexen Interaktionen."]} +{"source": "The recent expansion of machine learning applications to molecular biology proved to have a significant contribution to our understanding of biological systems, and genome functioning in particular.Technological advances enabled the collection of large epigenetic datasets, including information about various DNA binding factors (ChIP-Seq) and DNA spatial structure (Hi-C).Several studies have confirmed the correlation between DNA binding factors and Topologically Associating Domains (TADs) in DNA structure.However, the information about physical proximity represented by genomic coordinate was not yet used for the improvement of the prediction models.In this research, we focus on Machine Learning methods for prediction of folding patterns of DNA in a classical model organism Drosophila melanogaster.The paper considers linear models with four types of regularization, Gradient Boosting and Recurrent Neural Networks for the prediction of chromatin folding patterns from epigenetic marks.The bidirectional LSTM RNN model outperformed all the models and gained the best prediction scores.This demonstrates the utilization of complex models and the importance of memory of sequential DNA states for the chromatin folding.We identify informative epigenetic features that lead to the further conclusion of their biological significance.", "target": ["Wir wenden RNN an, um das biologische Problem der Vorhersage von Chromatinfaltungsmustern aus epigenetischen Markierungen zu lösen, und zeigen zum ersten Mal, dass die Nutzung des Speichers für sequenzielle Zustände auf dem DNA-Molekül für die beste Leistung von Bedeutung ist."]} +{"source": "Previous work showed empirically that large neural networks can be significantly reduced in size while preserving their accuracy.Model compression became a central research topic, as it is crucial for deployment of neural networks on devices with limited computational and memory resources.The majority of the compression methods are based on heuristics and offer no worst-case guarantees on the trade-off between the compression rate and the approximation error for an arbitrarily new sample.We propose the first efficient, data-independent neural pruning algorithm with a provable trade-off between its compression rate and the approximation error for any future test sample.Our method is based on the coreset framework, which finds a small weighted subset of points that provably approximates the original inputs.Specifically, we approximate the output of a layer of neurons by a coreset of neurons in the previous layer and discard the rest.We apply this framework in a layer-by-layer fashion from the top to the bottom.Unlike previous works, our coreset is data independent, meaning that it provably guarantees the accuracy of the function for any input $x\\in \\mathbb{R}^d$, including an adversarial one.We demonstrate the effectiveness of our method on popular network architectures.In particular, our coresets yield 90% compression of the LeNet-300-100 architecture on MNIST while improving the accuracy.", "target": ["Wir schlagen eine effiziente, nachweisbare und datenunabhängige Methode zur Netzwerkkompression durch neuronales Pruning unter Verwendung von Neuronen-Coresets vor - eine neuartige Konstruktion, die in dieser Arbeit vorgeschlagen wird."]} +{"source": "Planning problems in partially observable environments cannot be solved directly with convolutional networks and require some form of memory.But, even memory networks with sophisticated addressing schemes are unable to learn intelligent reasoning satisfactorily due to the complexity of simultaneously learning to access memory and plan.To mitigate these challenges we propose the Memory Augmented Control Network (MACN).The network splits planning into a hierarchical process.At a lower level, it learns to plan in a locally observed space.At a higher level, it uses a collection of policies computed on locally observed spaces to learn an optimal plan in the global environment it is operating in.The performance of the network is evaluated on path planning tasks in environments in the presence of simple and complex obstacles and in addition, is tested for its ability to generalize to new environments not seen in the training set.", "target": ["Memory Augmented Network zur Planung in teilweise beobachtbaren Umgebungen."]} +{"source": "Contextualized word representations such as ELMo and BERT have become the de facto starting point for incorporating pretrained representations for downstream NLP tasks.In these settings, contextual representations have largely made obsolete their static embedding predecessors such as Word2Vec and GloVe.However, static embeddings do have their advantages in that they are straightforward to understand and faster to use.Additionally, embedding analysis methods for static embeddings are far more diverse and mature than those available for their dynamic counterparts.In this work, we introduce simple methods for generating static lookup table embeddings from existing pretrained contextual representations and demonstrate they outperform Word2Vec and GloVe embeddings on a variety of word similarity and word relatedness tasks.In doing so, our results also reveal insights that may be useful for subsequent downstream tasks using our embeddings or the original contextual models.Further, we demonstrate the increased potential for analysis by applying existing approaches for estimating social bias in word embeddings.Our analysis constitutes the most comprehensive study of social bias in contextual word representations (via the proxy of our distilled embeddings) and reveals a number of inconsistencies in current techniques for quantifying social bias in word embeddings.We publicly release our code and distilled word embeddings to support reproducible research and the broader NLP community.", "target": ["Ein Verfahren zur Destillation von kontextuellen Modellen in statische Einbettungen; wir wenden unsere Methode auf 9 populäre Modelle an und zeigen deutliche Verbesserungen der Darstellungsqualität im Vergleich zu Word2Vec/GloVe sowie ein verbessertes Analysepotenzial durch eine gründliche Untersuchung sozialer Verzerrungen."]} +{"source": "The brain performs unsupervised learning and (perhaps) simultaneous supervised learning.This raises the question as to whether a hybrid of supervised and unsupervised methods will produce better learning.Inspired by the rich space of Hebbian learning rules, we set out to directly learn the unsupervised learning rule on local information that best augments a supervised signal.We present the Hebbian-augmented training algorithm (HAT) for combining gradient-based learning with an unsupervised rule on pre-synpatic activity, post-synaptic activities, and current weights.We test HAT's effect on a simple problem (Fashion-MNIST) and find consistently higher performance than supervised learning alone.This finding provides empirical evidence that unsupervised learning on synaptic activities provides a strong signal that can be used to augment gradient-based methods. We further find that the meta-learned update rule is a time-varying function; thus, it is difficult to pinpoint an interpretable Hebbian update rule that aids in training. We do find that the meta-learner eventually degenerates into a non-Hebbian rule that preserves important weights so as not to disturb the learner's convergence.", "target": ["Das Metalernen von unbeaufsichtigten Aktualisierungsregeln für neuronale Netze verbessert die Leistung und zeigt möglicherweise, wie Neuronen im Gehirn ohne Zugang zu globalen Bezeichnungen lernen."]} +{"source": "Deep convolutional networks often append additive constant (\"bias\") terms to their convolution operations, enabling a richer repertoire of functional mappings.Biases are also used to facilitate training, by subtracting mean response over batches of training images (a component of \"batch normalization\").Recent state-of-the-art blind denoising methods seem to require these terms for their success.Here, however, we show that bias terms used in most CNNs (additive constants, including those used for batch normalization) interfere with the interpretability of these networks, do not help performance, and in fact prevent generalization of performance to noise levels not including in the training data.In particular, bias-free CNNs (BF-CNNs) are locally linear, and hence amenable to direct analysis with linear-algebraic tools.These analyses provide interpretations of network functionality in terms of projection onto a union of low-dimensional subspaces, connecting the learning-based method to more traditional denoising methodology.Additionally, BF-CNNs generalize robustly, achieving near-state-of-the-art performance at noise levels well beyond the range over which they have been trained.", "target": ["Wir zeigen, dass das Entfernen von konstanten Termen aus CNN-Architekturen die Interpretierbarkeit der Entrauschungsmethode mittels linearer Algebra-Techniken ermöglicht und außerdem die Generalisierungsleistung über verschiedene Lärmpegel hinweg erhöht."]} +{"source": "The selection of initial parameter values for gradient-based optimization of deep neural networks is one of the most impactful hyperparameter choices in deep learning systems, affecting both convergence times and model performance.Yet despite significant empirical and theoretical analysis, relatively little has been proved about the concrete effects of different initialization schemes.In this work, we analyze the effect of initialization in deep linear networks, and provide for the first time a rigorous proof that drawing the initial weights from the orthogonal group speeds up convergence relative to the standard Gaussian initialization with iid weights.We show that for deep networks, the width needed for efficient convergence to a global minimum with orthogonal initializations is independent of the depth, whereas the width needed for efficient convergence with Gaussian initializations scales linearly in the depth.Our results demonstrate how the benefits of a good initialization can persist throughout learning, suggesting an explanation for the recent empirical successes found by initializing very deep non-linear networks according to the principle of dynamical isometry.", "target": ["Wir liefern zum ersten Mal einen rigorosen Beweis dafür, dass orthogonale Initialisierung die Konvergenz für tiefe lineare Netze im Vergleich zur Gaußschen Initialisierung beschleunigt."]} +{"source": "Survival function estimation is used in many disciplines, but it is most common in medical analytics in the form of the Kaplan-Meier estimator.Sensitive data (patient records) is used in the estimation without any explicit control on the information leakage, which is a significant privacy concern.We propose a first differentially private estimator of the survival function and show that it can be easily extended to provide differentially private confidence intervals and test statistics without spending any extra privacy budget.We further provide extensions for differentially private estimation of the competing risk cumulative incidence function.Using nine real-life clinical datasets, we provide empirical evidence that our proposed method provides good utility while simultaneously providing strong privacy guarantees.", "target": ["Eine erste differentially private Schätzung der Überlebensfunktion"]} +{"source": "Neural networks can learn to extract statistical properties from data, but they seldom make use of structured information from the label space to help representation learning.Although some label structure can implicitly be obtained when training on huge amounts of data, in a few-shot learning context where little data is available, making explicit use of the label structure can inform the model to reshape the representation space to reflect a global sense of class dependencies. We propose a meta-learning framework, Conditional class-Aware Meta-Learning (CAML), that conditionally transforms feature representations based on a metric space that is trained to capture inter-class dependencies.This enables a conditional modulation of the feature representations of the base-learner to impose regularities informed by the label space.Experiments show that the conditional transformation in CAML leads to more disentangled representations and achieves competitive results on the miniImageNet benchmark.", "target": ["CAML ist eine Instanz von MAML mit bedingten Klassenabhängigkeiten."]} +{"source": "We study the problem of multiset prediction.The goal of multiset prediction is to train a predictor that maps an input to a multiset consisting of multiple items.Unlike existing problems in supervised learning, such as classification, ranking and sequence generation, there is no known order among items in a target multiset, and each item in the multiset may appear more than once, making this problem extremely challenging.In this paper, we propose a novel multiset loss function by viewing this problem from the perspective of sequential decision making.The proposed multiset loss function is empirically evaluated on two families of datasets, one synthetic and the other real, with varying levels of difficulty, against various baseline loss functions including reinforcement learning, sequence, and aggregated distribution matching loss functions.The experiments reveal the effectiveness of the proposed loss function over the others.", "target": ["Wir untersuchen das Problem der Multiset-Vorhersage und schlagen eine neuartige Multiset-Verlustfunktion vor, die durch Analysen und empirische Belege ihre Wirksamkeit unter Beweis stellt."]} +{"source": "Understanding theoretical properties of deep and locally connected nonlinear network, such as deep convolutional neural network (DCNN), is still a hard problem despite its empirical success.In this paper, we propose a novel theoretical framework for such networks with ReLU nonlinearity.The framework bridges data distribution with gradient descent rules, favors disentangled representations and is compatible with common regularization techniques such as Batch Norm, after a novel discovery of its projection nature.The framework is built upon teacher-student setting, by projecting the student's forward/backward pass onto the teacher's computational graph.We do not impose unrealistic assumptions (e.g., Gaussian inputs, independence of activation, etc).Our framework could help facilitate theoretical analysis of many practical issues, e.g. disentangled representations in deep networks.", "target": ["In diesem Beitrag wird ein theoretischer Rahmen vorgestellt, der die Datenverteilung für tiefe und lokal verbundene ReLU-Netze explizit modelliert."]} +{"source": "Multi-agent cooperation is an important feature of the natural world.Many tasks involve individual incentives that are misaligned with the common good, yet a wide range of organisms from bacteria to insects and humans are able to overcome their differences and collaborate.Therefore, the emergence of cooperative behavior amongst self-interested individuals is an important question for the fields of multi-agent reinforcement learning (MARL) and evolutionary theory.Here, we study a particular class of multi-agent problems called intertemporal social dilemmas (ISDs), where the conflict between the individual and the group is particularly sharp.By combining MARL with appropriately structured natural selection, we demonstrate that individual inductive biases for cooperation can be learned in a model-free way.To achieve this, we introduce an innovative modular architecture for deep reinforcement learning agents which supports multi-level selection.We present results in two challenging environments, and interpret these in the context of cultural and ecological evolution.", "target": ["Wir stellen einen biologisch inspirierten modularen Evolutionsalgorithmus vor, in dem Deep-RL-Agenten lernen, in einem schwierigen sozialen Multi-Agenten-Spiel zu kooperieren, was helfen könnte, die Evolution des Altruismus zu erklären."]} +{"source": "In adversarial attacks to machine-learning classifiers, small perturbations are added to input that is correctly classified.The perturbations yield adversarial examples, which are virtually indistinguishable from the unperturbed input, and yet are misclassified.In standard neural networks used for deep learning, attackers can craft adversarial examples from most input to cause a misclassification of their choice. We introduce a new type of network units, called RBFI units, whose non-linear structure makes them inherently resistant to adversarial attacks.On permutation-invariant MNIST, in absence of adversarial attacks, networks using RBFI units match the performance of networks using sigmoid units, and are slightly below the accuracy of networks with ReLU units.When subjected to adversarial attacks based on projected gradient descent or fast gradient-sign methods, networks with RBFI units retain accuracies above 75%, while ReLU or Sigmoid see their accuracies reduced to below 1%.Further, RBFI networks trained on regular input either exceed or closely match the accuracy of sigmoid and ReLU network trained with the help of adversarial examples.The non-linear structure of RBFI units makes them difficult to train using standard gradient descent.We show that RBFI networks of RBFI units can be efficiently trained to high accuracies using pseudogradients, computed using functions especially crafted to facilitate learning instead of their true derivatives.", "target": ["Wir stellen eine Art von neuronalem Netz vor, das strukturell resistent gegen feindliche Angriffe ist, selbst wenn es auf unaugmentierten Trainingssätzen trainiert wird. Diese Widerstandsfähigkeit ist auf die Stabilität der Netzeinheiten gegenüber Eingabestörungen zurückzuführen."]} +{"source": "We show that there exists an inherent tension between the goal of adversarial robustness and that of standard generalization. Specifically, training robust models may not only be more resource-consuming, but also lead to a reduction of standard accuracy.We demonstrate that this trade-off between the standard accuracy of a model and its robustness to adversarial perturbations provably exists even in a fairly simple and natural setting.These findings also corroborate a similar phenomenon observed in practice.Further, we argue that this phenomenon is a consequence of robust classifiers learning fundamentally different feature representations than standard classifiers.These differences, in particular, seem to result in unexpected benefits: the features learned by robust models tend to align better with salient data characteristics and human perception.", "target": ["Wir zeigen, dass die adversariale Robustheit zwar auf Kosten der normalen Klassifizierungsleistung geht, aber auch unerwartete Vorteile mit sich bringt."]} +{"source": "Large deep neural networks are powerful, but exhibit undesirable behaviors such as memorization and sensitivity to adversarial examples.In this work, we propose mixup, a simple learning principle to alleviate these issues.In essence, mixup trains a neural network on convex combinations of pairs of examples and their labels. By doing so, mixup regularizes the neural network to favor simple linear behavior in-between training examples. Our experiments on the ImageNet-2012, CIFAR-10, CIFAR-100, Google commands and UCI datasets show that mixup improves the generalization of state-of-the-art neural network architectures. We also find that mixup reduces the memorization of corrupt labels, increases the robustness to adversarial examples, and stabilizes the training of generative adversarial networks.", "target": ["Training auf konvexen Kombinationen zwischen zufälligen Trainingsbeispielen und ihren Bezeichnungen verbessert die Generalisierung in tiefen neuronalen Netzen"]} +{"source": "We present a novel approach to spike sorting for high-density multielectrode probes using the Neural Clustering Process (NCP), a recently introduced neural architecture that performs scalable amortized approximate Bayesian inference for efficient probabilistic clustering.To optimally encode spike waveforms for clustering, we extended NCP by adding a convolutional spike encoder, which is learned end-to-end with the NCP network.Trained purely on labeled synthetic spikes from a simple generative model, the NCP spike sorting model shows promising performance for clustering multi-channel spike waveforms.The model provides higher clustering quality than an alternative Bayesian algorithm, finds more spike templates with clear receptive fields on real data and recovers more ground truth neurons on hybrid test data compared to a recent spike sorting algorithm.Furthermore, NCP is able to handle the clustering uncertainty of ambiguous small spikes by GPU-parallelized posterior sampling.The source code is publicly available.", "target": ["Wir stellen einen neuartigen Ansatz zur Spike-Sortierung vor, der den Neural Clustering Process (NCP) verwendet, eine kürzlich eingeführte neuronale Architektur, die skalierbare amortisierte approximative Bayes'sche Inferenz für effizientes probabilistisches Clustering durchführt."]} +{"source": "The goal of the paper is to propose an algorithm for learning the most generalizable solution from given training data.It is shown that Bayesian approach leads to a solution that dependent on statistics of training data and not on particularsamples.The solution is stable under perturbations of training data because it is defined by an integral contribution of multiple maxima of the likelihood and not by a single global maximum.Specifically, the Bayesian probability distributionof parameters (weights) of a probabilistic model given by a neural network is estimated via recurrent variational approximations.Derived recurrent update rules correspond to SGD-type rules for finding a minimum of an effective loss that is an average of an original negative log-likelihood over the Gaussian distributions of weights, which makes it a function of means and variances.The effective loss is convex for large variances and non-convex in the limit of small variances.Among stationary solutions of the update rules there are trivial solutions with zero variances at local minima of the original loss and a single non-trivial solution with finite variances that is a critical point at the end of convexity of the effective lossin the mean-variance space.At the critical point both first- and second-order gradients of the effective loss w.r.t. means are zero.The empirical study confirms that the critical point represents the most generalizable solution.While the location ofthe critical point in the weight space depends on specifics of the used probabilistic model some properties at the critical point are universal and model independent.", "target": ["Vorgeschlagene Methode zum Finden der am besten verallgemeinerbaren Lösung, die gegenüber Störungen der Trainingsdaten stabil ist."]} +{"source": "Humans rely on episodic memory constantly, in remembering the name of someone they met 10 minutes ago, the plot of a movie as it unfolds, or where they parked the car.Endowing reinforcement learning agents with episodic memory is a key step on the path toward replicating human-like general intelligence.We analyze why standard RL agents lack episodic memory today, and why existing RL tasks don't require it.We design a new form of external memory called Masked Experience Memory, or MEM, modeled after key features of human episodic memory.To evaluate episodic memory we define an RL task based on the common children's game of Concentration.We find that a MEM RL agent leverages episodic memory effectively to master Concentration, unlike the baseline agents we tested.", "target": ["Implementierung und Evaluierung des episodischen Speichers für RL."]} +{"source": "Parameters are one of the most critical components of machine learning models.As datasets and learning domains change, it is often necessary and time-consuming to re-learn entire models.Rather than re-learning the parameters from scratch, replacing learning with optimization, we propose a framework building upon the theory of \\emph{optimal transport} to adapt model parameters by discovering correspondences between models and data, significantly amortizing the training cost.We demonstrate our idea on the challenging problem of creating probabilistic spatial representations for autonomous robots.Although recent mapping techniques have facilitated robust occupancy mapping, learning all spatially-diverse parameters in such approximate Bayesian models demand considerable computational time, discouraging them to be used in real-world robotic mapping.Considering the fact that the geometric features a robot would observe with its sensors are similar across various environments, in this paper, we demonstrate how to re-use parameters and hyperparameters learned in different domains.This adaptation is computationally more efficient than variational inference and Monte Carlo techniques.A series of experiments conducted on realistic settings verified the possibility of transferring thousands of such parameters with a negligible time and memory cost, enabling large-scale mapping in urban environments.", "target": ["Wir stellen eine Methode zur Anpassung von Hyperparametern probabilistischer Modelle durch optimalen Transport mit Anwendungen in der Robotik vor."]} +{"source": "An algorithm is introduced for learning a predictive state representation with off-policy temporal difference (TD) learning that is then used to learn to steer a vehicle with reinforcement learning. There are three components being learned simultaneously: (1) the off-policy predictions as a compact representation of state, (2) the behavior policy distribution for estimating the off-policy predictions, and (3) the deterministic policy gradient for learning to act. A behavior policy discriminator is learned and used for estimating the important sampling ratios needed to learn the predictive representation off-policy with general value functions (GVFs). A linear deterministic policy gradient method is used to train the agent with only the predictive representations while the predictions are being learned. All three components are combined, demonstrated and evaluated on the problem of steering the vehicle from images in the TORCS racing simulator environment.Steering from only images is a challenging problem where evaluation is completed on a held-out set of tracks that were never seen during training in order to measure the generalization of the predictions and controller. Experiments show the proposed method is able to steer smoothly and navigate many but not all of the tracks available in TORCS with performance that exceeds DDPG using only images as input and approaches the performance of an ideal non-vision based kinematics model.", "target": ["Ein Algorithmus zum Erlernen einer prädiktiven Zustandsrepräsentation mit allgemeinen Wertfunktionen und Off-Policy-Lernen wird auf das Problem der sichtbasierten Lenkung beim autonomen Fahren angewendet."]} +{"source": "For numerous domains, including for instance earth observation, medical imaging, astrophysics,..., available image and signal datasets often irregular space-time sampling patterns and large missing data rates.These sampling properties is a critical issue to apply state-of-the-art learning-based (e.g., auto-encoders, CNNs,...) to fully benefit from the available large-scale observations and reach breakthroughs in the reconstruction and identification of processes of interest.In this paper, we address the end-to-end learning of representations of signals, images and image sequences from irregularly-sampled data, {\\em i.e.} when the training data involved missing data.From an analogy to Bayesian formulation, we consider energy-based representations.Two energy forms are investigated: one derived from auto-encoders and one relating to Gibbs energies.The learning stage of these energy-based representations (or priors) involve a joint interpolation issue, which resorts to solving an energy minimization problem under observation constraints.Using a neural-network-based implementation of the considered energy forms, we can state an end-to-end learning scheme from irregularly-sampled data.We demonstrate the relevance of the proposed representations for different case-studies: namely, multivariate time series, 2{\\sc } images and image sequences.", "target": ["Wir befassen uns mit dem durchgängigen Lernen von energiebasierten Repräsentationen für Signal- und Bildbeobachtungsdaten mit unregelmäßigen Probemustern."]} +{"source": "Credit assignment in Meta-reinforcement learning (Meta-RL) is still poorly understood.Existing methods either neglect credit assignment to pre-adaptation behavior or implement it naively.This leads to poor sample-efficiency during meta-training as well as ineffective task identification strategies.This paper provides a theoretical analysis of credit assignment in gradient-based Meta-RL.Building on the gained insights we develop a novel meta-learning algorithm that overcomes both the issue of poor credit assignment and previous difficulties in estimating meta-policy gradients.By controlling the statistical distance of both pre-adaptation and adapted policies during meta-policy search, the proposed algorithm endows efficient and stable meta-learning.Our approach leads to superior pre-adaptation policy behavior and consistently outperforms previous Meta-RL algorithms in sample-efficiency, wall-clock time, and asymptotic performance.", "target": ["Ein neuartiger und theoretisch fundierter Algorithmus für Meta-Reinforcement Learning"]} +{"source": "When training a neural network for a desired task, one may prefer to adapt a pretrained network rather than start with a randomly initialized one -- due to lacking enough training data, performing lifelong learning where the system has to learn a new task while being previously trained for other tasks, or wishing to encode priors in the network via preset weights.The most commonly employed approaches for network adaptation are fine-tuning and using the pre-trained network as a fixed feature extractor, among others. In this paper we propose a straightforward alternative: Side-Tuning.Side-tuning adapts a pretrained network by training a lightweight \"side\" network that is fused with the (unchanged) pre-rained network using a simple additive process.This simple method works as well as or better than existing solutions while it resolves some of the basic issues with fine-tuning, fixed features, and several other common baselines.In particular, side-tuning is less prone to overfitting when little training data is available, yields better results than using a fixed feature extractor, and doesn't suffer from catastrophic forgetting in lifelong learning. We demonstrate the performance of side-tuning under a diverse set of scenarios, including lifelong learning (iCIFAR, Taskonomy), reinforcement learning, imitation learning (visual navigation in Habitat), NLP question-answering (SQuAD v2), and single-task transfer learning (Taskonomy), with consistently promising results.", "target": ["Beim Side-Tuning wird ein bereits trainiertes Netz angepasst, indem ein leichtgewichtiges \"Side\"-Netz trainiert wird, das mit dem (unveränderten) bereits trainierten Netz durch einen einfachen additiven Prozess verschmolzen wird."]} +{"source": "In this paper, we present a technique for generating artificial datasets that retain statistical properties of the real data while providing differential privacy guarantees with respect to this data.We include a Gaussian noise layer in the discriminator of a generative adversarial network to make the output and the gradients differentially private with respect to the training data, and then use the generator component to synthesise privacy-preserving artificial dataset.Our experiments show that under a reasonably small privacy budget we are able to generate data of high quality and successfully train machine learning models on this artificial data.", "target": ["Trainieren von GANs mit differential privacy, um künstliche Datensätze zu erzeugen, die den Datenschutz wahren."]} +{"source": "This paper presents two methods to disentangle and interpret contextual effects that are encoded in a pre-trained deep neural network.Unlike convolutional studies that visualize image appearances corresponding to the network output or a neural activation from a global perspective, our research aims to clarify how a certain input unit (dimension) collaborates with other units (dimensions) to constitute inference patterns of the neural network and thus contribute to the network output.The analysis of local contextual effects w.r.t. certain input units is of special values in real applications.In particular, we used our methods to explain the gaming strategy of the alphaGo Zero model in experiments, and our method successfully disentangled the rationale of each move during the game.", "target": ["In diesem Beitrag werden Methoden zur Entflechtung und Interpretation von Kontexteffekten vorgestellt, die in einem tiefen neuronalen Netz kodiert sind."]} +{"source": "The main goal of network pruning is imposing sparsity on the neural network by increasing the number of parameters with zero value in order to reduce the architecture size and the computational speedup.", "target": ["Vorschlag einer neuen Methode, die auf der geführten Aufmerksamkeit basiert, um die Seltenheit in tiefen neuronalen Netzen zu verstärken."]} +{"source": "Giving provable guarantees for learning neural networks is a core challenge of machine learning theory.Most prior work gives parameter recovery guarantees for one hidden layer networks, however, the networks used in practice have multiple non-linear layers.In this work, we show how we can strengthen such results to deeper networks -- we address the problem of uncovering the lowest layer in a deep neural network under the assumption that the lowest layer uses a high threshold before applying the activation, the upper network can be modeled as a well-behaved polynomial and the input distribution is gaussian.", "target": ["Wir können die unterste Schicht eines tiefen neuronalen Netzes nachweislich wiederherstellen, wenn die unterste Schicht eine Aktivierung mit \"hohem Schwellenwert\" verwendet und das obige Netz ein \"wohlerzogenes\" Polynom ist."]} +{"source": "Federated learning involves training and effectively combining machine learning models from distributed partitions of data (i.e., tasks) on edge devices, and be naturally viewed as a multi- task learning problem.While Federated Averaging (FedAvg) is the leading optimization method for training non-convex models in this setting, its behavior is not well understood in realistic federated settings when the devices/tasks are statistically heterogeneous, i.e., where each device collects data in a non-identical fashion.In this work, we introduce a framework, called FedProx, to tackle statistical heterogeneity.FedProx encompasses FedAvg as a special case.We provide convergence guarantees for FedProx through a device dissimilarity assumption.Our empirical evaluation validates our theoretical analysis and demonstrates the improved robustness and stability of FedProx for learning in heterogeneous networks.", "target": ["Wir stellen FedProx vor, einen Rahmen zur Bewältigung statistischer Heterogenität in föderierten Umgebungen mit Konvergenzgarantien und verbesserter Robustheit und Stabilität."]} +{"source": "Referential games offer a grounded learning environment for neural agents which accounts for the fact that language is functionally used to communicate.However, they do not take into account a second constraint considered to be fundamental for the shape of human language: that it must be learnable by new language learners and thus has to overcome a transmission bottleneck.In this work, we insert such a bottleneck in a referential game, by introducing a changing population of agents in which new agents learn by playing with more experienced agents.We show that mere cultural transmission results in a substantial improvement in language efficiency and communicative success, measured in convergence speed, degree of structure in the emerged languages and within-population consistency of the language.However, as our core contribution, we show that the optimal situation is to co-evolve language and agents.When we allow the agent population to evolve through genotypical evolution, we achieve across the board improvements on all considered metrics.These results stress that for language emergence studies cultural evolution is important, but also the suitability of the architecture itself should be considered.", "target": ["Wir ermöglichen sowohl die kulturelle Evolution von Sprache als auch die genetische Evolution von Agenten in einem referentiellen Spiel, indem wir eine neue Sprachübertragungs Engine verwenden."]} +{"source": "The need for large amounts of training image data with clearly defined features is a major obstacle to applying generative adversarial networks(GAN) on image generation where training data is limited but diverse, since insufficient latent feature representation in the already scarce data often leads to instability and mode collapse during GAN training.To overcome the hurdle of limited data when applying GAN to limited datasets, we propose in this paper the strategy of \\textit{parallel recurrent data augmentation}, where the GAN model progressively enriches its training set with sample images constructed from GANs trained in parallel at consecutive training epochs.Experiments on a variety of small yet diverse datasets demonstrate that our method, with little model-specific considerations, produces images of better quality as compared to the images generated without such strategy.The source code and generated images of this paper will be made public after review.", "target": ["Wir haben eine neuartige, einfache und effiziente Methode zur Datenerweiterung eingeführt, die die Leistung bestehender GANs steigert, wenn die Trainingsdaten begrenzt und vielfältig sind."]} +{"source": "We develop a stochastic whole-brain and body simulator of the nematode roundworm Caenorhabditis elegans (C. elegans) and show that it is sufficiently regularizing to allow imputation of latent membrane potentials from partial calcium fluorescence imaging observations.This is the first attempt we know of to ``complete the circle,'' where an anatomically grounded whole-connectome simulator is used to impute a time-varying ``brain'' state at single-cell fidelity from covariates that are measurable in practice. Using state of the art Bayesian machine learning methods to condition on readily obtainable data, our method paves the way for neuroscientists to recover interpretable connectome-wide state representations, automatically estimate physiologically relevant parameter values from data, and perform simulations investigating intelligent lifeforms in silico.", "target": ["Wir entwickeln einen Simulator für das gesamte Konnektom und den Körper von C. elegans und demonstrieren eine gemeinsame Zustandsraum- und Parameterinferenz in diesem Simulator."]} +{"source": "The high-quality node embeddings learned from the Graph Neural Networks (GNNs) have been applied to a wide range of node-based applications and some of them have achieved state-of-the-art (SOTA) performance.However, when applying node embeddings learned from GNNs to generate graph embeddings, the scalar node representation may not suffice to preserve the node/graph properties efficiently, resulting in sub-optimal graph embeddings.Inspired by the Capsule Neural Network (CapsNet), we propose the Capsule Graph Neural Network (CapsGNN), which adopts the concept of capsules to address the weakness in existing GNN-based graph embeddings algorithms.By extracting node features in the form of capsules, routing mechanism can be utilized to capture important information at the graph level.As a result, our model generates multiple embeddings for each graph to capture graph properties from different aspects.The attention module incorporated in CapsGNN is used to tackle graphs with various sizes which also enables the model to focus on critical parts of the graphs.Our extensive evaluations with 10 graph-structured datasets demonstrate that CapsGNN has a powerful mechanism that operates to capture macroscopic properties of the whole graph by data-driven.It outperforms other SOTA techniques on several graph classification tasks, by virtue of the new instrument.", "target": ["Inspiriert von CapsNet schlagen wir eine neue Architektur für Grapheneinbettungen auf der Basis von Knotenmerkmalen vor, die aus GNN extrahiert werden."]} +{"source": "We introduce a novel framework for generative models based on Restricted Kernel Machines (RKMs) with multi-view generation and uncorrelated feature learning capabilities, called Gen-RKM.To incorporate multi-view generation, this mechanism uses a shared representation of data from various views.The mechanism is flexible to incorporate both kernel-based, (deep) neural network and convolutional based models within the same setting.To update the parameters of the network, we propose a novel training procedure which jointly learns the features and shared representation.Experiments demonstrate the potential of the framework through qualitative evaluation of generated samples.", "target": ["Gen-RKM: ein neuer Rahmen für generative Modelle unter Verwendung von Restricted Kernel Machines mit Multiview-Generierung und unkorreliertem Merkmalslernen."]} +{"source": "Work on the problem of contextualized word representation—the development of reusable neural network components for sentence understanding—has recently seen a surge of progress centered on the unsupervised pretraining task of language modeling with methods like ELMo (Peters et al., 2018).This paper contributes the first large-scale systematic study comparing different pretraining tasks in this context, both as complements to language modeling and as potential alternatives.The primary results of the study support the use of language modeling as a pretraining task and set a new state of the art among comparable models using multitask learning with language models.However, a closer look at these results reveals worryingly strong baselines and strikingly varied results across target tasks, suggesting that the widely-used paradigm of pretraining and freezing sentence encoders may not be an ideal platform for further work.", "target": ["Wir vergleichen viele Aufgaben und Aufgabenkombinationen zum Vortraining von BiLSTMs auf Satzebene für NLP-Aufgaben. Die Sprachmodellierung ist die beste einzelne Pretraining-Aufgabe, aber auch einfache Basisaufgaben schneiden gut ab."]} +{"source": "In this paper, we study the adversarial attack and defence problem in deep learning from the perspective of Fourier analysis.We first explicitly compute the Fourier transform of deep ReLU neural networks and show that there exist decaying but non-zero high frequency components in the Fourier spectrum of neural networks.We then demonstrate that the vulnerability of neural networks towards adversarial samples can be attributed to these insignificant but non-zero high frequency components.Based on this analysis, we propose to use a simple post-averaging technique to smooth out these high frequency components to improve the robustness of neural networks against adversarial attacks.Experimental results on the ImageNet and the CIFAR-10 datasets have shown that our proposed method is universally effective to defend many existing adversarial attacking methods proposed in the literature, including FGSM, PGD, DeepFool and C&W attacks.Our post-averaging method is simple since it does not require any re-training, and meanwhile it can successfully defend over 80-96% of the adversarial samples generated by these methods without introducing significant performance degradation (less than 2%) on the original clean images.", "target": ["Ein Einblick in den Grund der gegnerischen Verwundbarkeit, eine effektive Verteidigungsmethode gegen Adversarial Attacks."]} +{"source": "Reinforcement learning methods that continuously learn neural networks by episode generation with game tree search have been successful in two-person complete information deterministic games such as chess, shogi, and Go.However, there are only reports of practical cases and there are little evidence to guarantee the stability and the final performance of learning process.In this research, the coordination of episode generation was focused on.By means of regarding the entire system as game tree search, the new method can handle the trade-off between exploitation and exploration during episode generation.The experiments with a small problem showed that it had robust performance compared to the existing method, Alpha Zero.", "target": ["Anwendung der Monte-Carlo-Baumsuche auf die Episodenerzeugung in Alpha Zero"]} +{"source": "Message-passing neural networks (MPNNs) have been successfully applied in a wide variety of applications in the real world.However, two fundamental weaknesses of MPNNs' aggregators limit their ability to represent graph-structured data: losing the structural information of nodes in neighborhoods and lacking the ability to capture long-range dependencies in disassortative graphs.Few studies have noticed the weaknesses from different perspectives.From the observations on classical neural network and network geometry, we propose a novel geometric aggregation scheme for graph neural networks to overcome the two weaknesses. The behind basic idea is the aggregation on a graph can benefit from a continuous space underlying the graph.The proposed aggregation scheme is permutation-invariant and consists of three modules, node embedding, structural neighborhood, and bi-level aggregation.We also present an implementation of the scheme in graph convolutional networks, termed Geom-GCN, to perform transductive learning on graphs.Experimental results show the proposed Geom-GCN achieved state-of-the-art performance on a wide range of open datasets of graphs.", "target": ["Bei neuronalen Netzen mit Graphen kann die Aggregation in einem Graphen von einem kontinuierlichen Raum profitieren, der dem Graphen zugrunde liegt."]} +{"source": "We consider the task of program synthesis in the presence of a reward function over the output of programs, where the goal is to find programs with maximal rewards.We introduce a novel iterative optimization scheme, where we train an RNN on a dataset of K best programs from a priority queue of the generated programs so far.Then, we synthesize new programs and add them to the priority queue by sampling from the RNN.We benchmark our algorithm called priority queue training (PQT) against genetic algorithm and reinforcement learning baselines on a simple but expressive Turing complete programming language called BF.Our experimental results show that our deceptively simple PQT algorithm significantly outperforms the baselines.By adding a program length penalty to the reward function, we are able to synthesize short, human readable programs.", "target": ["Wir verwenden einen einfachen Suchalgorithmus mit einem RNN und einer Prioritätswarteschlange, um Lösungen für Codierungsaufgaben zu finden."]} +{"source": "We present graph wavelet neural network (GWNN), a novel graph convolutional neural network (CNN), leveraging graph wavelet transform to address the shortcomings of previous spectral graph CNN methods that depend on graph Fourier transform.Different from graph Fourier transform, graph wavelet transform can be obtained via a fast algorithm without requiring matrix eigendecomposition with high computational cost.Moreover, graph wavelets are sparse and localized in vertex domain, offering high efficiency and good interpretability for graph convolution.The proposed GWNN significantly outperforms previous spectral graph CNNs in the task of graph-based semi-supervised classification on three benchmark datasets: Cora, Citeseer and Pubmed.", "target": ["Wir präsentieren das Graph Wavelet Neural Network (GWNN), ein neuartiges Graph-Convolutional Neural Network (CNN), das die Graph Wavelet Transformation nutzt, um die Unzulänglichkeit früherer spektraler Graph CNN Methoden zu beheben, die auf der Graph Fourier Transformation beruhen."]} +{"source": "Learning rate decay (lrDecay) is a \\emph{de facto} technique for training modern neural networks.It starts with a large learning rate and then decays it multiple times.It is empirically observed to help both optimization and generalization.Common beliefs in how lrDecay works come from the optimization analysis of (Stochastic) Gradient Descent:1) an initially large learning rate accelerates training or helps the network escape spurious local minima;2) decaying the learning rate helps the network converge to a local minimum and avoid oscillation.Despite the popularity of these common beliefs, experiments suggest that they are insufficient in explaining the general effectiveness of lrDecay in training modern neural networks that are deep, wide, and nonconvex.We provide another novel explanation: an initially large learning rate suppresses the network from memorizing noisy data while decaying the learning rate improves the learning of complex patterns.The proposed explanation is validated on a carefully-constructed dataset with tractable pattern complexity.And its implication, that additional patterns learned in later stages of lrDecay are more complex and thus less transferable, is justified in real-world datasets.We believe that this alternative explanation will shed light into the design of better training strategies for modern neural networks.", "target": ["Wir liefern eine weitere neue Erklärung für den Verfall der Lernrate: Eine anfänglich hohe Lernrate verhindert, dass sich das Netzwerk verrauschte Daten merkt, während ein Verfall der Lernrate das Erlernen komplexer Muster verbessert."]} +{"source": "We show how an ensemble of $Q^*$-functions can be leveraged for more effective exploration in deep reinforcement learning.We build on well established algorithms from the bandit setting, and adapt them to the $Q$-learning setting.We propose an exploration strategy based on upper-confidence bounds (UCB).Our experiments show significant gains on the Atari benchmark.", "target": ["Die Anpassung der UCB-Exploration an das Ensemble Q-learning verbessert frühere Methoden wie Double DQN, A3C+ im Atari-Benchmark"]} +{"source": "We focus on the problem of learning a single motor module that can flexibly express a range of behaviors for the control of high-dimensional physically simulated humanoids.To do this, we propose a motor architecture that has the general structure of an inverse model with a latent-variable bottleneck.We show that it is possible to train this model entirely offline to compress thousands of expert policies and learn a motor primitive embedding space.The trained neural probabilistic motor primitive system can perform one-shot imitation of whole-body humanoid behaviors, robustly mimicking unseen trajectories.Additionally, we demonstrate that it is also straightforward to train controllers to reuse the learned motor primitive space to solve tasks, and the resulting movements are relatively naturalistic.To support the training of our model, we compare two approaches for offline policy cloning, including an experience efficient method which we call linear feedback policy cloning.We encourage readers to view a supplementary video (https://youtu.be/CaDEf-QcKwA ) summarizing our results.", "target": ["Neural Probabilistic Motor Primitives komprimieren die Richtlinien zur Bewegungserfassung in ein flexibles Modell, das one-shot Imitation und Wiederverwendung als Low-Level-Controller kann."]} +{"source": "Data augmentation is a useful technique to enlarge the size of the training set and prevent overfitting for different machine learning tasks when training data is scarce.However, current data augmentation techniques rely heavily on human design and domain knowledge, and existing automated approaches are yet to fully exploit the latent features in the training dataset.In this paper we propose \\textit{Parallel Adaptive GAN Data Augmentation}(PAGANDA), where the training set adaptively enriches itself with sample images automatically constructed from Generative Adversarial Networks (GANs) trained in parallel.We demonstrate by experiments that our data augmentation strategy, with little model-specific considerations, can be easily adapted to cross-domain deep learning/machine learning tasks such as image classification and image inpainting, while significantly improving model performance in both tasks.Our source code and experimental details are available at \\url{https://github.com/miaojiang1987/k-folder-data-augmentation-gan/}.", "target": ["Wir stellen eine automatisierte adaptive Datenerweiterung vor, die für mehrere unterschiedliche Aufgaben funktioniert."]} +{"source": "Deep neural networks with millions of parameters may suffer from poor generalizations due to overfitting.To mitigate the issue, we propose a new regularization method that penalizes the predictive distribution between similar samples.In particular, we distill the predictive distribution between different samples of the same label and augmented samples of the same source during training.In other words, we regularize the dark knowledge (i.e., the knowledge on wrong predictions) of a single network, i.e., a self-knowledge distillation technique, to force it output more meaningful predictions. We demonstrate the effectiveness of the proposed method via experiments on various image classification tasks: it improves not only the generalization ability, but also the calibration accuracy of modern neural networks.", "target": ["Wir schlagen eine neue Regularisierungstechnik vor, die auf der Wissensdestillation basiert."]} +{"source": "In this paper, we consider the specific problem of word-level language modeling and investigate strategies for regularizing and optimizing LSTM-based models.We propose the weight-dropped LSTM, which uses DropConnect on hidden-to-hidden weights, as a form of recurrent regularization.Further, we introduce NT-ASGD, a non-monotonically triggered (NT) variant of the averaged stochastic gradient method (ASGD), wherein the averaging trigger is determined using a NT condition as opposed to being tuned by the user.Using these and other regularization strategies, our ASGD Weight-Dropped LSTM (AWD-LSTM) achieves state-of-the-art word level perplexities on two data sets: 57.3 on Penn Treebank and 65.8 on WikiText-2.In exploring the effectiveness of a neural cache in conjunction with our proposed model, we achieve an even lower state-of-the-art perplexity of 52.8 on Penn Treebank and 52.0 on WikiText-2.We also explore the viability of the proposed regularization and optimization strategies in the context of the quasi-recurrent neural network (QRNN) and demonstrate comparable performance to the AWD-LSTM counterpart.The code for reproducing the results is open sourced and is available at https://github.com/salesforce/awd-lstm-lm.", "target": ["Effektive Regularisierungs- und Optimierungsstrategien für LSTM-basierte Sprachmodelle erreichen SOTA auf PTB und WT2."]} +{"source": "Various methods of measuring unit selectivity have been developed with the aim of better understanding how neural networks work. But the different measures provide divergent estimates of selectivity, and this has led to different conclusions regarding the conditions in which selective object representations are learned and the functional relevance of these representations.In an attempt to better characterize object selectivity, we undertake a comparison of various selectivity measures on a large set of units in AlexNet, including localist selectivity, precision, class-conditional mean activity selectivity (CCMAS), network dissection, the human interpretation of activation maximization (AM) images, and standard signal-detection measures. We find that the different measures provide different estimates of object selectivity, with precision and CCMAS measures providing misleadingly high estimates.Indeed, the most selective units had a poor hit-rate or a high false-alarm rate (or both) in object classification, making them poor object detectors. We fail to find any units that are even remotely as selective as the 'grandmother cell' units reported in recurrent neural networks.In order to generalize these results, we compared selectivity measures on a few units in VGG-16 and GoogLeNet trained on the ImageNet or Places-365 datasets that have been described as 'object detectors'.Again, we find poor hit-rates and high false-alarm rates for object classification.", "target": ["Auf der Suche nach Objektdetektoren unter Verwendung vieler verschiedener Selektivitätsmaße; CNNs sind leicht selektiv, aber nicht genug, um als Objektdetektoren bezeichnet zu werden."]} +{"source": "The folding structure of the DNA molecule combined with helper molecules, also referred to as the chromatin, is highly relevant for the functional properties of DNA.The chromatin structure is largely determined by the underlying primary DNA sequence, though the interaction is not yet fully understood.In this paper we develop a convolutional neural network that takes an image-representation of primary DNA sequence as its input, and predicts key determinants of chromatin structure.The method is developed such that it is capable of detecting interactions between distal elements in the DNA sequence, which are known to be highly relevant.Our experiments show that the method outperforms several existing methods both in terms of prediction accuracy and training time.", "target": ["Eine Methode zur Umwandlung von DNA-Sequenzen in 2D-Bilder unter Verwendung von raumfüllenden Hilbert-Kurven zur Verbesserung der Stärken von CNNs"]} +{"source": "This paper introduces a novel method to perform transfer learning across domains and tasks, formulating it as a problem of learning to cluster.The key insight is that, in addition to features, we can transfer similarity information and this is sufficient to learn a similarity function and clustering network to perform both domain adaptation and cross-task transfer learning.We begin by reducing categorical information to pairwise constraints, which only considers whether two instances belong to the same class or not (pairwise semantic similarity).This similarity is category-agnostic and can be learned from data in the source domain using a similarity network.We then present two novel approaches for performing transfer learning using this similarity function.First, for unsupervised domain adaptation, we design a new loss function to regularize classification with a constrained clustering loss, hence learning a clustering network with the transferred similarity metric generating the training inputs.Second, for cross-task learning (i.e., unsupervised clustering with unseen categories), we propose a framework to reconstruct and estimate the number of semantic clusters, again using the clustering network.Since the similarity network is noisy, the key is to use a robust clustering algorithm, and we show that our formulation is more robust than the alternative constrained and unconstrained clustering approaches.Using this method, we first show state of the art results for the challenging cross-task problem, applied on Omniglot and ImageNet.Our results show that we can reconstruct semantic clusters with high accuracy.We then evaluate the performance of cross-domain transfer using images from the Office-31 and SVHN-MNIST tasks and present top accuracy on both datasets. Our approach doesn't explicitly deal with domain discrepancy.If we combine with a domain adaptation loss, it shows further improvement.", "target": ["Ein lernfähiges Clustering-Ziel zur Erleichterung des bereichs- und aufgabenübergreifenden Transfer Lernens"]} +{"source": "Recent advances in cross-lingual word embeddings have primarily relied on mapping-based methods, which project pretrained word embeddings from different languages into a shared space through a linear transformation.However, these approaches assume word embedding spaces are isomorphic between different languages, which has been shown not to hold in practice (Søgaard et al., 2018), and fundamentally limits their performance.This motivates investigating joint learning methods which can overcome this impediment, by simultaneously learning embeddings across languages via a cross-lingual term in the training objective.Given the abundance of parallel data available (Tiedemann, 2012), we propose a bilingual extension of the CBOW method which leverages sentence-aligned corpora to obtain robust cross-lingual word and sentence representations.Our approach significantly improves cross-lingual sentence retrieval performance over all other approaches, as well as convincingly outscores mapping methods while maintaining parity with jointly trained methods on word-translation.It also achieves parity with a deep RNN method on a zero-shot cross-lingual document classification task, requiring far fewer computational resources for training and inference.As an additional advantage, our bilingual method also improves the quality of monolingual word vectors despite training on much smaller datasets. We make our code and models publicly available.", "target": ["Gemeinsame Methode zum Erlernen von sprachenübergreifenden Einbettungen mit modernster Leistung für sprachenübergreifende Aufgaben und einsprachige Qualität"]} +{"source": "To act and plan in complex environments, we posit that agents should have a mental simulator of the world with three characteristics:(a) it should build an abstract state representing the condition of the world;(b) it should form a belief which represents uncertainty on the world;(c) it should go beyond simple step-by-step simulation, and exhibit temporal abstraction.Motivated by the absence of a model satisfying all these requirements, we propose TD-VAE, a generative sequence model that learns representations containing explicit beliefs about states several steps into the future, and that can be rolled out directly without single-step transitions.TD-VAE is trained on pairs of temporally separated time points, using an analogue of temporal difference learning used in reinforcement learning.", "target": ["Generatives Modell zeitlicher Daten, das einen online belief state aufbaut, im latenten Raum operiert und sprunghafte Vorhersagen und Zustandsänderungen vornimmt."]} +{"source": "This paper introduces an information theoretic co-training objective for unsupervised learning. We consider the problem of predicting the future. Rather than predict future sensations (image pixels or sound waves) we predict ``hypotheses'' to be confirmed by future sensations. More formally, we assume a population distribution on pairs $(x,y)$ where we can think of $x$ as a past sensation and $y$ as a future sensation. We train both a predictor model $P_\\Phi(z|x)$ and a confirmation model $P_\\Psi(z|y)$ where we view $z$ as hypotheses (when predicted) or facts (when confirmed). For a population distribution on pairs $(x,y)$ we focus on the problem of measuring the mutual information between $x$ and $y$. By the data processing inequality this mutual information is at least as large as the mutual information between $x$ and $z$ under the distribution on triples $(x,z,y)$ defined by the confirmation model $P_\\Psi(z|y)$.The information theoretic training objective for $P_\\Phi(z|x)$ and $P_\\Psi(z|y)$ can be viewed as a form of co-training where we want the prediction from $x$ to match the confirmation from $y$.We give experiments on applications to learning phonetics on the TIMIT dataset.", "target": ["Stellt ein informationstheoretisches Trainingsziel für Co-Training vor und demonstriert dessen Leistungsfähigkeit beim unüberwachten Lernen von Phonetik."]} +{"source": "Generative models for singing voice have been mostly concerned with the task of \"singing voice synthesis,\" i.e., to produce singing voice waveforms given musical scores and text lyrics.In this work, we explore a novel yet challenging alternative: singing voice generation without pre-assigned scores and lyrics, in both training and inference time.In particular, we experiment with three different schemes:1) free singer, where the model generates singing voices without taking any conditions;2) accompanied singer, where the model generates singing voices over a waveform of instrumental music; and3) solo singer, where the model improvises a chord sequence first and then uses that to generate voices.We outline the associated challenges and propose a pipeline to tackle these new tasks.This involves the development of source separation and transcription models for data preparation, adversarial networks for audio generation, and customized metrics for evaluation.", "target": ["Unsere Modelle erzeugen Gesangsstimmen ohne Text und Partitur. Sie nehmen eine Begleitung als Eingabe und geben Gesangsstimmen aus."]} +{"source": "The carbon footprint of natural language processing (NLP) research has been increasing in recent years due to its reliance on large and inefficient neural network implementations.Distillation is a network compression technique which attempts to impart knowledge from a large model to a smaller one.We use teacher-student distillation to improve the efficiency of the Biaffine dependency parser which obtains state-of-the-art performance with respect to accuracy and parsing speed (Dozat & Manning, 2016). When distilling to 20% of the original model’s trainable parameters, we only observe an average decrease of ∼1 point for both UAS and LAS across a number of diverse Universal Dependency treebanks while being 2.26x (1.21x) faster than the baseline model on CPU (GPU) at inference time.We also observe a small increase in performance when compressing to 80% for some treebanks. Finally, through distillation we attain a parser which is not only faster but also more accurate than the fastest modern parser on the Penn Treebank.", "target": ["Wir steigern die Effizienz von Dependenzparsern mit neuronalen Netzen durch Lehrer-Schüler-Destillation."]} +{"source": "While autoencoders are a key technique in representation learning for continuous structures, such as images or wave forms, developing general-purpose autoencoders for discrete structures, such as text sequence or discretized images, has proven to be more challenging.In particular, discrete inputs make it more difficult to learn a smooth encoder that preserves the complex local relationships in the input space.In this work, we propose an adversarially regularized autoencoder (ARAE) with the goal of learning more robust discrete-space representations.ARAE jointly trains both a rich discrete-space encoder, such as an RNN, and a simpler continuous space generator function, while using generative adversarial network (GAN) training to constrain the distributions to be similar.This method yields a smoother contracted code space that maps similar inputs to nearby codes, and also an implicit latent variable GAN model for generation.Experiments on text and discretized images demonstrate that the GAN model produces clean interpolations and captures the multimodality of the original space, and that the autoencoder produces improvements in semi-supervised learning as well as state-of-the-art results in unaligned text style transfer task using only a shared continuous-space representation.", "target": ["Adversarially Regularized Autoencoders lernen glatte Repräsentationen diskreter Strukturen und ermöglichen so interessante Ergebnisse bei der Texterzeugung, wie z.B. unaligned style transfer, semi-supervised learning und latent space interpolation and arithmetic."]} +{"source": "When data arise from multiple latent subpopulations, machine learning frameworks typically estimate parameter values independently for each sub-population.In this paper, we proposeto overcome these limits by considering samples as tasks in a multitask learning framework.", "target": ["Wir stellen eine Methode zur Schätzung von Sammlungen von Regressionsmodellen vor, bei der jedes Modell auf eine einzelne Stichprobe zugeschnitten ist."]} +{"source": "In this work, we first conduct mathematical analysis on the memory, which isdefined as a function that maps an element in a sequence to the current output,of three RNN cells; namely, the simple recurrent neural network (SRN), the longshort-term memory (LSTM) and the gated recurrent unit (GRU).Based on theanalysis, we propose a new design, called the extended-long short-term memory(ELSTM), to extend the memory length of a cell.Next, we present a multi-taskRNN model that is robust to previous erroneous predictions, called the dependentbidirectional recurrent neural network (DBRNN), for the sequence-in-sequenceout(SISO) problem.Finally, the performance of the DBRNN model with theELSTM cell is demonstrated by experimental results.", "target": ["Eine rekurrente neuronale Netzwerkzelle mit erweitertem Long Short Term Memory und ein Multi-Task-RNN-Modell für Sequence-in-Sequence-out Probleme"]} +{"source": "Many recently trained neural networks employ large numbers of parameters to achieve good performance.One may intuitively use the number of parameters required as a rough gauge of the difficulty of a problem.But how accurate are such notions?How many parameters are really needed?In this paper we attempt to answer this question by training networks not in their native parameter space, but instead in a smaller, randomly oriented subspace.We slowly increase the dimension of this subspace, note at which dimension solutions first appear, and define this to be the intrinsic dimension of the objective landscape.The approach is simple to implement, computationally tractable, and produces several suggestive conclusions.Many problems have smaller intrinsic dimensions than one might suspect, and the intrinsic dimension for a given dataset varies little across a family of models with vastly different sizes.This latter result has the profound implication that once a parameter space is large enough to solve a problem, extra parameters serve directly to increase the dimensionality of the solution manifold.Intrinsic dimension allows some quantitative comparison of problem difficulty across supervised, reinforcement, and other types of learning where we conclude, for example, that solving the inverted pendulum problem is 100 times easier than classifying digits from MNIST, and playing Atari Pong from pixels is about as hard as classifying CIFAR-10.In addition to providing new cartography of the objective landscapes wandered by parameterized models, the method is a simple technique for constructively obtaining an upper bound on the minimum description length of a solution.A byproduct of this construction is a simple approach for compressing networks, in some cases by more than 100 times.", "target": ["Wir trainieren in zufälligen Unterräumen des Parameterraums, um zu messen, wie viele Dimensionen tatsächlich erforderlich sind, um eine Lösung zu finden."]} +{"source": "Graph Neural Networks as a combination of Graph Signal Processing and Deep Convolutional Networks shows great power in pattern recognition in non-Euclidean domains.In this paper, we propose a new method to deploy two pipelines based on the duality of a graph to improve accuracy.By exploring the primal graph and its dual graph where nodes and edges can be treated as one another, we have exploited the benefits of both vertex features and edge features.As a result, we have arrived at a framework that has great potential in both semisupervised and unsupervised learning.", "target": ["Ein primäres duales neuronales Netzmodell für halb-überwachtes Lernen"]} +{"source": "We describe two end-to-end autoencoding models for semi-supervised graph-based dependency parsing.The first model is a Local Autoencoding Parser (LAP) encoding the input using continuous latent variables in a sequential manner; The second model is a Global Autoencoding Parser (GAP) encoding the input into dependency trees as latent variables, with exact inference.Both models consist of two parts: an encoder enhanced by deep neural networks (DNN) that can utilize the contextual information to encode the input into latent variables, and a decoder which is a generative model able to reconstruct the input.Both LAP and GAP admit a unified structure with different loss functions for labeled and unlabeled data with shared parameters.We conducted experiments on WSJ and UD dependency parsing data sets, showing that our models can exploit the unlabeled data to boost the performance given a limited amount of labeled data.", "target": ["Wir beschreiben zwei End-to-End-Auto-Coding-Parser für semi-supervised graph-based dependency parsing."]} +{"source": "We improve previous end-to-end differentiable neural networks (NNs) with fastweight memories.A gate mechanism updates fast weights at every time step ofa sequence through two separate outer-product-based matrices generated by slowparts of the net.The system is trained on a complex sequence to sequence variationof the Associative Retrieval Problem with roughly 70 times more temporalmemory (i.e. time-varying variables) than similar-sized standard recurrent NNs(RNNs).In terms of accuracy and number of parameters, our architecture outperformsa variety of RNNs, including Long Short-Term Memory, Hypernetworks,and related fast weight architectures.", "target": ["Ein verbessertes Fast-Weight-Netzwerk, das bessere Ergebnisse bei einer allgemeinen Spielzeugaufgabe zeigt."]} +{"source": "The field of deep learning has been craving for an optimization method that shows outstanding property for both optimization and generalization. We propose a method for mathematical optimization based on flows along geodesics, that is, the shortest paths between two points, with respect to the Riemannian metric induced by a non-linear function.In our method, the flows refer to Exponentially Decaying Flows (EDF), as they can be designed to converge on the local solutions exponentially.In this paper, we conduct experiments to show its high performance on optimization benchmarks (i.e., convergence properties), as well as its potential for producing good machine learning benchmarks (i.e., generalization properties).", "target": ["Einführung einer neuen Optimierungsmethode und ihre Anwendung auf Deep Learning."]} +{"source": "We introduce Quantum Graph Neural Networks (QGNN), a new class of quantum neural network ansatze which are tailored to represent quantum processes which have a graph structure, and are particularly suitable to be executed on distributed quantum systems over a quantum network.Along with this general class of ansatze, we introduce further specialized architectures, namely, Quantum Graph Recurrent Neural Networks (QGRNN) and Quantum Graph Convolutional Neural Networks (QGCNN).We provide four example applications of QGNN's: learning Hamiltonian dynamics of quantum systems, learning how to create multipartite entanglement in a quantum network, unsupervised learning for spectral clustering, and supervised learning for graph isomorphism classification.", "target": ["Einführung einer neuen Klasse von neuronalen Quantennetzen zum Erlernen graphbasierter Darstellungen auf Quantencomputern."]} +{"source": "Data breaches involve information being accessed by unauthorized parties.Our research concerns user perception of data breaches, especially issues relating to accountability.A preliminary study indicated many people had weak understanding of the issues, and felt they themselves were somehow responsible.We speculated that this impression might stem from organizational communication strategies.We therefore compared texts from organizations with external sources, such as the news media.This suggested that organizations use well-known crisis communication methods to reduce their reputational damage, and that these strategies align with repositioning of the narrative elements involved in the story.We then conducted a quantitative study, asking participants to rate either organizational texts or news texts about breaches.The findings of this study were in line with our document analysis, and suggest that organizational communication affects the users' perception of victimization, attitudes in data protection, and accountability.Our study suggests some software design and legal implications supporting users to protect themselves and develop better mental models of security breaches.", "target": ["In dieser Arbeit haben wir den Einfluss von Kommunikationsstrategien auf die mentalen Modelle von Nutzern einer Datenschutzverletzung untersucht."]} +{"source": "Goal recognition is the problem of inferring the correct goal towards which an agent executes a plan, given a set of goal hypotheses, a domain model, and a (possibly noisy) sample of the plan being executed. This is a key problem in both cooperative and competitive agent interactions and recent approaches have produced fast and accurate goal recognition algorithms. In this paper, we leverage advances in operator-counting heuristics computed using linear programs over constraints derived from classical planning problems to solve goal recognition problems. Our approach uses additional operator-counting constraints derived from the observations to efficiently infer the correct goal, and serves as basis for a number of further methods with additional constraints.", "target": ["Ein Zielerkennungsansatz, der auf einer Heuristik der Operatorenzählung basiert, um das Störungen im Datensatz zu berücksichtigen."]} +{"source": "It can be challenging to train multi-task neural networks that outperform or even match their single-task counterparts.To help address this, we propose using knowledge distillation where single-task models teach a multi-task model.We enhance this training with teacher annealing, a novel method that gradually transitions the model from distillation to supervised learning, helping the multi-task model surpass its single-task teachers.We evaluate our approach by multi-task fine-tuning BERT on the GLUE benchmark.Our method consistently improves over standard single-task and multi-task training.", "target": ["die Destillation von Einzelaufgabenmodellen in ein Multiaufgabenmodell verbessert die Leistung beim Verstehen natürlicher Sprache"]} +{"source": "The detection of out of distribution samples for image classification has been widely researched.Safety critical applications, such as autonomous driving, would benefit from the ability to localise the unusual objects causing the image to be out of distribution.This paper adapts state-of-the-art methods for detecting out of distribution images for image classification to the new task of detecting out of distribution pixels, which can localise the unusual objects.It further experimentally compares the adapted methods on two new datasets derived from existing semantic segmentation datasets using PSPNet and DeeplabV3+ architectures, as well as proposing a new metric for the task.The evaluation shows that the performance ranking of the compared methods does not transfer to the new task and every method performs significantly worse than their image-level counterparts.", "target": ["Evaluierung von Methoden zur Erkennung von Verteilungsfehlern auf Pixelebene an zwei neuen realen Datensätzen mit PSPNet und DeeplabV3+."]} +{"source": "We present a domain adaptation method for transferring neural representations from label-rich source domains to unlabeled target domains.Recent adversarial methods proposed for this task learn to align features across domains by ``fooling'' a special domain classifier network.However, a drawback of this approach is that the domain classifier simply labels the generated features as in-domain or not, without considering the boundaries between classes.This means that ambiguous target features can be generated near class boundaries, reducing target classification accuracy.We propose a novel approach, Adversarial Dropout Regularization (ADR), which encourages the generator to output more discriminative features for the target domain.Our key idea is to replace the traditional domain critic with a critic that detects non-discriminative features by using dropout on the classifier network.The generator then learns to avoid these areas of the feature space and thus creates better features.We apply our ADR approach to the problem of unsupervised domain adaptation for image classification and semantic segmentation tasks, and demonstrate significant improvements over the state of the art.", "target": ["Wir stellen eine neue adversarische Methode zur Anpassung neuronaler Repräsentationen vor, die auf einer Kritik basiert, die nicht-diskriminierende Merkmale erkennt."]} +{"source": "The use of deep learning for a wide range of data problems has increased the need for understanding and diagnosing these models, and deep learning interpretation techniques have become an essential tool for data analysts.Although numerous model interpretation methods have been proposed in recent years, most of these procedures are based on heuristics with little or no theoretical guarantees.In this work, we propose a statistical framework for saliency estimation for black box computer vision models.We build a model-agnostic estimation procedure that is statistically consistent and passes the saliency checks of Adebayo et al. (2018).Our method requires solving a linear program, whose solution can be efficiently computed in polynomial time.Through our theoretical analysis, we establish an upper bound on the number of model evaluations needed to recover the region of importance with high probability, and build a new perturbation scheme for estimation of local gradients that is shown to be more efficient than the commonly used random perturbation schemes.Validity of the new method is demonstrated through sensitivity analysis.", "target": ["Wir schlagen einen statistischen Rahmen und ein theoretisch konsistentes Verfahren zur Schätzung der Auffälligkeit vor."]} +{"source": "We explore the idea of compositional set embeddings that can be used to infer notjust a single class, but the set of classes associated with the input data (e.g., image,video, audio signal).This can be useful, for example, in multi-object detection inimages, or multi-speaker diarization (one-shot learning) in audio.In particular, wedevise and implement two novel models consisting of (1) an embedding functionf trained jointly with a “composite” function g that computes set union opera-tions between the classes encoded in two embedding vectors; and (2) embeddingf trained jointly with a “query” function h that computes whether the classes en-coded in one embedding subsume the classes encoded in another embedding.Incontrast to prior work, these models must both perceive the classes associatedwith the input examples, and also encode the relationships between different classlabel sets.In experiments conducted on simulated data, OmniGlot, and COCOdatasets, the proposed composite embedding models outperform baselines basedon traditional embedding approaches.", "target": ["Wir haben untersucht, wie eine neuartige Methode der kompositorischen Mengeneinbettung nicht nur eine einzelne Klasse, sondern eine ganze Menge von Klassen, die mit den Eingabedaten verbunden sind, wahrnehmen und darstellen kann."]} +{"source": "In this work, we propose a goal-driven collaborative task that contains language, vision, and action in a virtual environment as its core components.Specifically, we develop a Collaborative image-Drawing game between two agents, called CoDraw.Our game is grounded in a virtual world that contains movable clip art objects.The game involves two players: a Teller and a Drawer.The Teller sees an abstract scene containing multiple clip art pieces in a semantically meaningful configuration, while the Drawer tries to reconstruct the scene on an empty canvas using available clip art pieces.The two players communicate via two-way communication using natural language.We collect the CoDraw dataset of ~10K dialogs consisting of ~138K messages exchanged between human agents.We define protocols and metrics to evaluate the effectiveness of learned agents on this testbed, highlighting the need for a novel \"crosstalk\" condition which pairs agents trained independently on disjoint subsets of the training data for evaluation.We present models for our task, including simple but effective baselines and neural network approaches trained using a combination of imitation learning and goal-driven training.All models are benchmarked using both fully automated evaluation and by playing the game with live human agents.", "target": ["Wir stellen einen Datensatz, Modelle und Trainings- und Evaluationsprotokolle für eine kollaborative Zeichenaufgabe vor, die es ermöglicht, zielgerichtete und wahrnehmungs- und handlungsbasierte Spracherzeugung und Sprachverständnis zu untersuchen."]} +{"source": "Presence of bias and confounding effects is inarguably one of the most critical challenges in machine learning applications that has alluded to pivotal debates in the recent years.Such challenges range from spurious associations of confounding variables in medical studies to the bias of race in gender or face recognition systems.One solution is to enhance datasets and organize them such that they do not reflect biases, which is a cumbersome and intensive task.The alternative is to make use of available data and build models considering these biases.Traditional statistical methods apply straightforward techniques such as residualization or stratification to precomputed features to account for confounding variables.However, these techniques are not in general applicable to end-to-end deep learning methods.In this paper, we propose a method based on the adversarial training strategy to learn discriminative features unbiased and invariant to the confounder(s).This is enabled by incorporating a new adversarial loss function that encourages a vanished correlation between the bias and learned features.We apply our method to a synthetic, a medical diagnosis, and a gender classification (Gender Shades) dataset.Our results show that the learned features by our method not only result in superior prediction performance but also are uncorrelated with the bias or confounder variables.The code is available at http://blinded_for_review/.", "target": ["Wir schlagen eine Methode vor, die auf der adversen Trainingsstrategie basiert, um diskriminierende Merkmale zu erlernen, die unvoreingenommen und invariant gegenüber den Störfaktoren sind, indem wir eine Verlustfunktion einbauen, die eine verschwindende Korrelation zwischen der Verzerrung und den erlernten Merkmalen fördert."]} +{"source": "Existing neural question answering (QA) models are required to reason over and draw complicated inferences from a long context for most large-scale QA datasets.However, if we view QA as a combined retrieval and reasoning task, we can assume the existence of a minimal context which is necessary and sufficient to answer a given question.Recent work has shown that a sentence selector module that selects a shorter context and feeds it to the downstream QA model achieves performance comparable to a QA model trained on full context, while also being more interpretable.Recent work has also shown that most state-of-the-art QA models break when adversarially generated sentences are appended to the context.While humans are immune to such distractor sentences, QA models get easily misled into selecting answers from these sentences.We hypothesize that the sentence selector module can filter out extraneous context, thereby allowing the downstream QA model to focus and reason over the parts of the context that are relevant to the question.In this paper, we show that the sentence selector itself is susceptible to adversarial inputs.However, we demonstrate that a pipeline consisting of a sentence selector module followed by the QA model can be made more robust to adversarial attacks in comparison to a QA model trained on full context.Thus, we provide evidence towards a modular approach for question answering that is more robust and interpretable.", "target": ["Ein modularer Ansatz, der aus einem Satzselektionsmodul und einem anschließenden QS-Modell besteht, kann im Vergleich zu einem QS-Modell, das auf der Grundlage des gesamten Kontexts trainiert wurde, robuster gegenüber Angriffen durch Angreifer gemacht werden."]} +{"source": "Multi-relational graph embedding which aims at achieving effective representations with reduced low-dimensional parameters, has been widely used in knowledge base completion.Although knowledge base data usually contains tree-like or cyclic structure, none of existing approaches can embed these data into a compatible space that in line with the structure.To overcome this problem, a novel framework, called Riemannian TransE, is proposed in this paper to embed the entities in a Riemannian manifold.Riemannian TransE models each relation as a move to a point and defines specific novel distance dissimilarity for each relation, so that all the relations are naturally embedded in correspondence to the structure of data.Experiments on several knowledge base completion tasks have shown that, based on an appropriate choice of manifold, Riemannian TransE achieves good performance even with a significantly reduced parameters.", "target": ["Multirelationale Grapheneinbettung mit Riemannschen Mannigfaltigkeiten und TransE-ähnlicher Verlustfunktion."]} +{"source": "Catastrophic forgetting in neural networks is one of the most well-known problems in continual learning.Previous attempts on addressing the problem focus on preventing important weights from changing.Such methods often require task boundaries to learn effectively and do not support backward transfer learning.In this paper, we propose a meta-learning algorithm which learns to reconstruct the gradients of old tasks w.r.t. the current parameters and combines these reconstructed gradients with the current gradient to enable continual learning and backward transfer learning from the current task to previous tasks.Experiments on standard continual learning benchmarks show that our algorithm can effectively prevent catastrophic forgetting and supports backward transfer learning.", "target": ["Wir schlagen einen Meta-Lernalgorithmus für kontinuierliches Lernen vor, der das Problem des katastrophalen Vergessens wirksam verhindern und das Rückwärts-Transferlernen unterstützen kann."]} +{"source": "We give a formal procedure for computing preimages of convolutional network outputs using the dual basis defined from the set of hyperplanes associated with the layers of the network.We point out the special symmetry associated with arrangements of hyperplanes of convolutional networks that take the form of regular multidimensional polyhedral cones.We discuss the efficiency of of large number of layers of nested cones that result from incremental small size convolutions in order to give a good compromise between efficient contraction of data to low dimensions and shaping of preimage manifolds.We demonstrate how a specific network flattens a non linear input manifold to an affine output manifold and discuss it's relevance to understanding classification properties of deep networks.", "target": ["Analyse von tiefen Convolutional Networks in Bezug auf die zugehörige Anordnung von Hyperebenen"]} +{"source": "Meta learning has been making impressive progress for fast model adaptation.However, limited work has been done on learning fast uncertainty adaption for Bayesian modeling.In this paper, we propose to achieve the goal by placing meta learning on the space of probability measures, inducing the concept of meta sampling for fast uncertainty adaption.Specifically, we propose a Bayesian meta sampling framework consisting of two main components: a meta sampler and a sample adapter.The meta sampler is constructed by adopting a neural-inverse-autoregressive-flow (NIAF) structure, a variant of the recently proposed neural autoregressive flows, to efficiently generate meta samples to be adapted.The sample adapter moves meta samples to task-specific samples, based on a newly proposed and general Bayesian sampling technique, called optimal-transport Bayesian sampling.The combination of the two components allows a simple learning procedure for themeta sampler to be developed, which can be efficiently optimized via standard back-propagation.Extensive experimental results demonstrate the efficiency and effectiveness of the proposed framework, obtaining better sample quality and fasteruncertainty adaption compared to related methods.", "target": ["Wir haben eine Bayes'sche Meta-Sampling-Methode zur Anpassung der Modellunsicherheit beim Meta-Lernen vorgeschlagen"]} +{"source": "We propose a context-adaptive entropy model for use in end-to-end optimized image compression.Our model exploits two types of contexts, bit-consuming contexts and bit-free contexts, distinguished based upon whether additional bitallocation is required.Based on these contexts, we allow the model to more accurately estimate the distribution of each latent representation with a more generalized form of the approximation models, which accordingly leads to anenhanced compression performance.Based on the experimental results, the proposed method outperforms the traditional image codecs, such as BPG and JPEG2000, as well as other previous artificial-neural-network (ANN) based approaches, in terms of the peak signal-to-noise ratio (PSNR) and multi-scale structural similarity (MS-SSIM) index.The test code is publicly available at https://github.com/JooyoungLeeETRI/CA_Entropy_Model.", "target": ["Kontextadaptives Entropiemodell zur Verwendung in der durchgängig optimierten Bildkompression, das die Kompressionsleistung erheblich verbessert"]} +{"source": "Deep networks often perform well on the data distribution on which they are trained, yet give incorrect (and often very confident) answers when evaluated on points from off of the training distribution.This is exemplified by the adversarial examples phenomenon but can also be seen in terms of model generalization and domain shift. Ideally, a model would assign lower confidence to points unlike those from the training distribution. We propose a regularizer which addresses this issue by training with interpolated hidden states and encouraging the classifier to be less confident at these points. Because the hidden states are learned, this has an important effect of encouraging the hidden states for a class to be concentrated in such a way so that interpolations within the same class or between two different classes do not intersect with the real data points from other classes. This has a major advantage in that it avoids the underfitting which can result from interpolating in the input space. We prove that the exact condition for this problem of underfitting to be avoided by Manifold Mixup is that the dimensionality of the hidden states exceeds the number of classes, which is often the case in practice. Additionally, this concentration can be seen as making the features in earlier layers more discriminative. We show that despite requiring no significant additional computation, Manifold Mixup achieves large improvements over strong baselines in supervised learning, robustness to single-step adversarial attacks, semi-supervised learning, and Negative Log-Likelihood on held out samples.", "target": ["Eine Methode zum Erlernen besserer Repräsentationen, die als Regularisierer fungiert und trotz ihres geringen zusätzlichen Rechenaufwands Verbesserungen im Vergleich zu starken Baselines bei überwachten und halb-überwachten Lernaufgaben erzielt."]} +{"source": "Sometimes SRS (Stereotactic Radio Surgery) requires using sphere packing on a Region of Interest (ROI) such as cancer to determine a treatment plan. We have developed a sphere packing algorithm which packs non-intersecting spheres inside the ROI. The region of interest in our case are those voxels which are identified as cancer tissues. In this paper, we analyze the rotational invariant properties of our sphere-packing algorithm which is based on distance transformations.Epsilon-Rotation invariant means the ability to arbitrary rotate the 3D ROI while keeping the volume properties remaining (almost) same within some limit of epsilon.The applied rotations produce spherical packing which remains highly correlated as we analyze the geometrically properties of sphere packing before and after the rotation of the volume data for the ROI.Our novel sphere packing algorithm has high degree of rotation invariance within the range of +/- epsilon.Our method used a shape descriptor derived from the values of the disjoint set of spheres form the distance-based sphere packing algorithm to extract the invariant descriptor from the ROI.We demonstrated by implementing these ideas using Slicer3D platform available for our research. The data is based on sing MRI Stereotactic images.We presented several performance results on different benchmarks data of over 30 patients in Slicer3D platform.", "target": ["Packen von Interessensgebieten (ROI) wie z. B. Krebsregionen, die in 3D-Volumendaten identifiziert wurden, Packen von Bereichen innerhalb des ROI, Drehen des ROI, Messen des Unterschieds in der gepackten Bereichen vor und nach der Drehung."]} +{"source": "We show implicit filter level sparsity manifests in convolutional neural networks (CNNs) which employ Batch Normalization and ReLU activation, and are trained using adaptive gradient descent techniques with L2 regularization or weight decay.Through an extensive empirical study (Anonymous, 2019) we hypothesize the mechanism be hind the sparsification process.We find that the interplay of various phenomena influences the strength of L2 and weight decay regularizers, leading the supposedly non sparsity inducing regularizers to induce filter sparsity. In this workshop article we summarize some of our key findings and experiments, and present additional results on modern network architectures such as ResNet-50.", "target": ["Seltenheit auf Filterebene entsteht implizit in CNNs, die mit adaptiven Gradientenabstiegsansätzen trainiert werden, aufgrund verschiedener Phänomene, und das Ausmaß der Seltenheit kann versehentlich durch verschiedene, scheinbar nicht zusammenhängende Hyperparameter beeinflusst werden."]} +{"source": "Humans can learn a variety of concepts and skills incrementally over the course of their lives while exhibiting an array of desirable properties, such as non-forgetting, concept rehearsal, forward transfer and backward transfer of knowledge, few-shot learning, and selective forgetting.Previous approaches to lifelong machine learning can only demonstrate subsets of these properties, often by combining multiple complex mechanisms. In this Perspective, we propose a powerful unified framework that can demonstrate all of the properties by utilizing a small number of weight consolidation parameters in deep neural networks.In addition, we are able to draw many parallels between the behaviours and mechanisms of our proposed framework and those surrounding human learning, such as memory loss or sleep deprivation.This Perspective serves as a conduit for two-way inspiration to further understand lifelong learning in machines and humans.", "target": ["Indem wir Parallelen zum menschlichen Lernen ziehen, schlagen wir einen einheitlichen Rahmen vor, um viele lebenslange Lernfähigkeiten in neuronalen Netzen zu zeigen, indem wir eine kleine Anzahl von Gewichtskonsolidierungsparametern verwenden."]} +{"source": "Limited angle CT reconstruction is an under-determined linear inverse problem that requires appropriate regularization techniques to be solved.In this work we study how pre-trained generative adversarial networks (GANs) can be used to clean noisy, highly artifact laden reconstructions from conventional techniques, by effectively projecting onto the inferred image manifold.In particular, we use a robust version of the popularly used GAN prior for inverse problems, based on a recent technique called corruption mimicking, that significantly improves the reconstruction quality.The proposed approach operates in the image space directly, as a result of which it does not need to be trained or require access to the measurement model, is scanner agnostic, and can work over a wide range of sensing scenarios.", "target": ["Wir zeigen, dass robuste GAN priors besser funktionieren als GAN priors für die CT-Rekonstruktion mit begrenztem Winkel, die ein stark unterbestimmtes inverses Problem darstellt."]} +{"source": "We propose an effective multitask learning setup for reducing distant supervision noise by leveraging sentence-level supervision.We show how sentence-level supervision can be used to improve the encoding of individual sentences, and to learn which input sentences are more likely to express the relationship between a pair of entities.We also introduce a novel neural architecture for collecting signals from multiple input sentences, which combines the benefits of attention and maxpooling.The proposed method increases AUC by 10% (from 0.261 to 0.284), and outperforms recently published results on the FB-NYT dataset.", "target": ["Eine neue Form der Aufmerksamkeit, die sich gut für die Überwachung aus der Ferne eignet, und ein Multitasking-Ansatz, um Anmerkungen auf Satzebene hinzuzufügen."]} +{"source": "The attention layer in a neural network model provides insights into the model’s reasoning behind its prediction, which are usually criticized for being opaque.Recently, seemingly contradictory viewpoints have emerged about the interpretability of attention weights (Jain & Wallace, 2019; Vig & Belinkov, 2019).Amid such confusion arises the need to understand attention mechanism more systematically.In this work, we attempt to fill this gap by giving a comprehensive explanation which justifies both kinds of observations (i.e., when is attention interpretable and when it is not).Through a series of experiments on diverse NLP tasks, we validate our observations and reinforce our claim of interpretability of attention through manual evaluation.", "target": ["Analyse des Aufmerksamkeitsmechanismus bei verschiedenen NLP-Aufgaben."]} +{"source": "Generative models forsource code are an interesting structured prediction problem, requiring to reason about both hard syntactic and semantic constraints as well as about natural, likely programs.We present a novel model for this problem that uses a graph to represent the intermediate state of the generated output.Our model generates code by interleaving grammar-driven expansion steps with graph augmentation and neural message passing steps.An experimental evaluation shows that our new model can generate semantically meaningful expressions, outperforming a range of strong baselines.", "target": ["Die Darstellung von Programmen als Graphen inklusive Semantik hilft bei der Programmerstellung"]} +{"source": "The inference of models, prediction of future symbols, and entropy rate estimation of discrete-time, discrete-event processes is well-worn ground.However, many time series are better conceptualized as continuous-time, discrete-event processes.Here, we provide new methods for inferring models, predicting future symbols, and estimating the entropy rate of continuous-time, discrete-event processes.The methods rely on an extension of Bayesian structural inference that takes advantage of neural network’s universal approximation power.Based on experiments with simple synthetic data, these new methods seem to be competitive with state-of- the-art methods for prediction and entropy rate estimation as long as the correct model is inferred.", "target": ["Eine neue Methode zur Ableitung eines Modells, zur Schätzung der Entropierate und zur Vorhersage zeitkontinuierlicher, ereignisdiskreter Prozesse."]} +{"source": "Recent studies show that convolutional neural networks (CNNs) are vulnerable under various settings, including adversarial examples, backdoor attacks, and distribution shifting. Motivated by the findings that human visual system pays more attention to global structure (e.g., shape) for recognition while CNNs are biased towards local texture features in images, we propose a unified framework EdgeGANRob based on robust edge features to improve the robustness of CNNs in general, which first explicitly extracts shape/structure features from a given image and then reconstructs a new image by refilling the texture information with a trained generative adversarial network (GAN).In addition, to reduce the sensitivity of edge detection algorithm to adversarial perturbation, we propose a robust edge detection approach Robust Canny based on the vanilla Canny algorithm.To gain more insights, we also compare EdgeGANRob with its simplified backbone procedure EdgeNetRob, which performs learning tasks directly on the extracted robust edge features.We find that EdgeNetRob can help boost model robustness significantly but at the cost of the clean model accuracy.EdgeGANRob, on the other hand, is able to improve clean model accuracy compared with EdgeNetRob and without losing the robustness benefits introduced by EdgeNetRob.Extensive experiments show that EdgeGANRob is resilient in different learning tasks under diverse settings.", "target": ["Ein vereinheitlichtes Modell zur Verbesserung der Robustheit des Modells bei mehreren Aufgaben"]} +{"source": "Learning rich representation from data is an important task for deep generative models such as variational auto-encoder (VAE).However, by extracting high-level abstractions in the bottom-up inference process, the goal of preserving all factors of variations for top-down generation is compromised.Motivated by the concept of “starting small”, we present a strategy to progressively learn independent hierarchical representations from high- to low-levels of abstractions.The model starts with learning the most abstract representation, and then progressively grow the network architecture to introduce new representations at different levels of abstraction.We quantitatively demonstrate the ability of the presented model to improve disentanglement in comparison to existing works on two benchmark datasets using three disentanglement metrics, including a new metric we proposed to complement the previously-presented metric of mutual information gap.We further present both qualitative and quantitative evidence on how the progression of learning improves disentangling of hierarchical representations.By drawing on the respective advantage of hierarchical representation learning and progressive learning, this is to our knowledge the first attempt to improve disentanglement by progressively growing the capacity of VAE to learn hierarchical representations.", "target": ["Wir haben eine progressive Lernmethode vorgeschlagen, um das Lernen und die Entflechtung latenter Repräsentationen auf verschiedenen Abstraktionsebenen zu verbessern."]} +{"source": "Deep latent variable models are powerful tools for representation learning.In this paper, we adopt the deep information bottleneck model, identify its shortcomings and propose a model that circumvents them.To this end, we apply a copula transformation which, by restoring the invariance properties of the information bottleneck method, leads to disentanglement of the features in the latent space.Building on that, we show how this transformation translates to sparsity of the latent space in the new model. We evaluate our method on artificial and real data.", "target": ["Wir wenden die Copula-Transformation auf den Deep Information Bottleneck an, was zu wiederhergestellten Invarianzeigenschaften und einem entwirrten latenten Raum mit überlegenen Vorhersagefähigkeiten führt."]} +{"source": "State-of-the-art machine learning methods exhibit limited compositional generalization.At the same time, there is a lack of realistic benchmarks that comprehensively measure this ability, which makes it challenging to find and evaluate improvements.We introduce a novel method to systematically construct such benchmarks by maximizing compound divergence while guaranteeing a small atom divergence between train and test sets, and we quantitatively compare this method to other approaches for creating compositional generalization benchmarks.We present a large and realistic natural language question answering dataset that is constructed according to this method, and we use it to analyze the compositional generalization ability of three machine learning architectures.We find that they fail to generalize compositionally and that there is a surprisingly strong negative correlation between compound divergence and accuracy.We also demonstrate how our method can be used to create new compositionality benchmarks on top of the existing SCAN dataset, which confirms these findings.", "target": ["Benchmark und Methode zur Messung der kompositorischen Generalisierung durch Maximierung der Divergenz der Verbindungshäufigkeit bei geringer Divergenz der Atomhäufigkeit."]} +{"source": "To understand how object vision develops in infancy and childhood, it will be necessary to develop testable computational models.Deep neural networks (DNNs) have proven valuable as models of adult vision, but it is not yet clear if they have any value as models of development.As a first model, we measured learning in a DNN designed to mimic the architecture and representational geometry of the visual system (CORnet).We quantified the development of explicit object representations at each level of this network through training by freezing the convolutional layers and training an additional linear decoding layer.We evaluate decoding accuracy on the whole ImageNet validation set, and also for individual visual classes.CORnet, however, uses supervised training and because infants have only extremely impoverished access to labels they must instead learn in an unsupervised manner.We therefore also measured learning in a state-of-the-art unsupervised network (DeepCluster).CORnet and DeepCluster differ in both supervision and in the convolutional networks at their heart, thus to isolate the effect of supervision, we ran a control experiment in which we trained the convolutional network from DeepCluster (an AlexNet variant) in a supervised manner.We make predictions on how learning should develop across brain regions in infants.In all three networks, we also tested for a relationship in the order in which infants and machines acquire visual classes, and found only evidence for a counter-intuitive relationship.We discuss the potential reasons for this.", "target": ["Unüberwachte Netze lernen von unten nach oben; Maschinen und Kleinkinder lernen visuelle Klassen in unterschiedlicher Reihenfolge"]} +{"source": "We show that in a variety of large-scale deep learning scenarios the gradient dynamically converges to a very small subspace after a short period of training.The subspace is spanned by a few top eigenvectors of the Hessian (equal to the number of classes in the dataset), and is mostly preserved over long periods of training.A simple argument then suggests that gradient descent may happen mostly in this subspace.We give an example of this effect in a solvable model of classification, and we comment on possible implications for optimization and learning.", "target": ["Für Klassifizierungsprobleme mit k Klassen zeigen wir, dass der Gradient dazu neigt, in einem winzigen, sich langsam entwickelnden Unterraum zu leben, der von den Eigenvektoren aufgespannt wird, die den k-größten Eigenwerten der Hessian entsprechen."]} +{"source": "Answering questions that require multi-hop reasoning at web-scale necessitates retrieving multiple evidence documents, one of which often has little lexical or semantic relationship to the question.This paper introduces a new graph-based recurrent retrieval approach that learns to retrieve reasoning paths over the Wikipedia graph to answer multi-hop open-domain questions.Our retriever model trains a recurrent neural network that learns to sequentially retrieve evidence paragraphs in the reasoning path by conditioning on the previously retrieved documents. Our reader model ranks the reasoning paths and extracts the answer span included in the best reasoning path.Experimental results show state-of-the-art results in three open-domain QA datasets, showcasing the effectiveness and robustness of our method.Notably, our method achieves significant improvement in HotpotQA, outperforming the previous best model by more than 14 points.", "target": ["Ein graphenbasierter rekurrenter Retriever, der lernt, Argumentationspfade über den Wikipedia-Graphen abzurufen, übertrifft den aktuellen Stand der Technik bei HotpotQA um mehr als 14 Punkte."]} +{"source": "Stereo matching is one of the important basic tasks in the computer vision field.In recent years, stereo matching algorithms based on deep learning have achieved excellent performance and become the mainstream research direction.Existing algorithms generally use deep convolutional neural networks (DCNNs) to extract more abstract semantic information, but we believe that the detailed information of the spatial structure is more important for stereo matching tasks.Based on this point of view, this paper proposes a shallow feature extraction network with a large receptive field.The network consists of three parts: a primary feature extraction module, an atrous spatial pyramid pooling (ASPP) module and a feature fusion module.The primary feature extraction network contains only three convolution layers.This network utilizes the basic feature extraction ability of the shallow network to extract and retain the detailed information of the spatial structure.In this paper, the dilated convolution and atrous spatial pyramid pooling (ASPP) module is introduced to increase the size of receptive field.In addition, a feature fusion module is designed, which integrates the feature maps with multiscale receptive fields and mutually complements the feature information of different scales.We replaced the feature extraction part of the existing stereo matching algorithms with our shallow feature extraction network, and achieved state-of-the-art performance on the KITTI 2015 dataset.Compared with the reference network, the number of parameters is reduced by 42%, and the matching accuracy is improved by 1.9%.", "target": ["Wir haben ein flaches Feature-Extraktionsnetzwerk mit einem großen rezeptiven Feld für Stereo-Matching-Aufgaben eingeführt, das eine einfache Struktur verwendet, um eine bessere Leistung zu erzielen."]} +{"source": "We propose a new output layer for deep neural networks that permits the use of logged contextual bandit feedback for training.Such contextual bandit feedback can be available in huge quantities (e.g., logs of search engines, recommender systems) at little cost, opening up a path for training deep networks on orders of magnitude more data.To this effect, we propose a Counterfactual Risk Minimization (CRM) approach for training deep networks using an equivariant empirical risk estimator with variance regularization, BanditNet, and show how the resulting objective can be decomposed in a way that allows Stochastic Gradient Descent (SGD) training.We empirically demonstrate the effectiveness of the method by showing how deep networks -- ResNets in particular -- can be trained for object recognition without conventionally labeled images.", "target": ["Diese Arbeit schlägt eine neue Ausgabeschicht für tiefe Netzwerke vor, die die Verwendung von protokolliertem kontextuellem Bandit-Feedback für das Training ermöglicht."]} +{"source": "Protein classification is responsible for the biological sequence, we came up with an idea whichdeals with the classification of proteomics using deep learning algorithm.This algorithm focusesmainly to classify sequences of protein-vector which is used for the representation of proteomics.Selection of the type protein representation is challenging based on which output in terms ofaccuracy is depended on, The protein representation used here is n-gram i.e. 3-gram and Kerasembedding used for biological sequences like protein.In this paper we are working on the Proteinclassification to show the strength and representation of biological sequence of the proteins", "target": ["Klassifizierung von Proteinfamilien mit Deep Learning"]} +{"source": "In this work, we attempt to answer a critical question: whether there exists some input sequence that will cause a well-trained discrete-space neural network sequence-to-sequence (seq2seq) model to generate egregious outputs (aggressive, malicious, attacking, etc.).And if such inputs exist, how to find them efficiently.We adopt an empirical methodology, in which we first create lists of egregious output sequences, and then design a discrete optimization algorithm to find input sequences that will cause the model to generate them.Moreover, the optimization algorithm is enhanced for large vocabulary search and constrained to search for input sequences that are likely to be input by real-world users.In our experiments, we apply this approach to dialogue response generation models trained on three real-world dialogue data-sets: Ubuntu, Switchboard and OpenSubtitles, testing whether the model can generate malicious responses.We demonstrate that given the trigger inputs our algorithm finds, a significant number of malicious sentences are assigned large probability by the model, which reveals an undesirable consequence of standard seq2seq training.", "target": ["Ziel dieses Beitrags ist es, eine empirische Antwort auf die Frage zu geben, ob ein gut trainiertes Dialogantwortmodell bösartige Antworten ausgeben kann."]} +{"source": "In the problem of unsupervised learning of disentangled representations, one of the promising methods is to penalize the total correlation of sampled latent vari-ables. Unfortunately, this well-motivated strategy often fail to achieve disentanglement due to a problematic difference between the sampled latent representation and its corresponding mean representation. We provide a theoretical explanation that low total correlation of sample distribution cannot guarantee low total correlation of the mean representation.We prove that for the mean representation of arbitrarily high total correlation, there exist distributions of latent variables of abounded total correlation. However, we still believe that total correlation could be a key to the disentanglement of unsupervised representative learning, and we propose a remedy, RTC-VAE, which rectifies the total correlation penalty. Experiments show that our model has a more reasonable distribution of the mean representation compared with baseline models, e.g.,β-TCVAE and FactorVAE.", "target": ["das gesamte Problem der STOA VAEs theoretisch und qualitativ diagnostiziert"]} +{"source": "Visual attention mechanisms have been widely used in image captioning models.In this paper, to better link the image structure with the generated text, we replace the traditional softmax attention mechanism by two alternative sparsity-promoting transformations: sparsemax and Total-Variation Sparse Attention (TVmax).With sparsemax, we obtain sparse attention weights, selecting relevant features. In order to promote sparsity and encourage fusing of the related adjacent spatial locations, we propose TVmax. By selecting relevant groups of features, the TVmax transformation improves interpretability.We present results in the Microsoft COCO and Flickr30k datasets, obtaining gains in comparison to softmax. TVmax outperforms the other compared attention mechanisms in terms of human-rated caption quality and attention relevance.", "target": ["Wir schlagen einen neuen, spärlichen und strukturierten Aufmerksamkeitsmechanismus, TVmax, vor, der die Spärlichkeit fördert und dazu anregt, dass die Gewichtung verwandter benachbarter Orte gleich ist."]} +{"source": "Although there are more than 65,000 languages in the world, the pronunciations of many phonemes sound similar across the languages.When people learn a foreign language, their pronunciation often reflect their native language's characteristics.That motivates us to investigate how the speech synthesis network learns the pronunciation when multi-lingual dataset is given.In this study, we train the speech synthesis network bilingually in English and Korean, and analyze how the network learns the relations of phoneme pronunciation between the languages.Our experimental result shows that the learned phoneme embedding vectors are located closer if their pronunciations are similar across the languages.Based on the result, we also show that it is possible to train networks that synthesize English speaker's Korean speech and vice versa.In another experiment, we train the network with limited amount of English dataset and large Korean dataset, and analyze the required amount of dataset to train a resource-poor language with the help of resource-rich languages.", "target": ["Gelernte Phonemeeinbettungen mehrsprachiger neuronaler Sprachsynthesenetzwerke könnten die Beziehungen der Phonemeaussprache zwischen den Sprachen darstellen."]} +{"source": "Generative Adversarial Networks (GANs) have recently achieved impressive results for many real-world applications, and many GAN variants have emerged with improvements in sample quality and training stability.However, visualization and understanding of GANs is largely missing.How does a GAN represent our visual world internally?What causes the artifacts in GAN results?How do architectural choices affect GAN learning?Answering such questions could enable us to develop new insights and better models.In this work, we present an analytic framework to visualize and understand GANs at the unit-, object-, and scene-level.We first identify a group of interpretable units that are closely related to object concepts with a segmentation-based network dissection method.Then, we quantify the causal effect of interpretable units by measuring the ability of interventions to control objects in the output.Finally, we examine the contextual relationship between these units and their surrounding by inserting the discovered object concepts into new images.We show several practical applications enabled by our framework, from comparing internal representations across different layers, models, and datasets, to improving GANs by locating and removing artifact-causing units, to interactively manipulating objects in the scene.We provide open source interpretation tools to help peer researchers and practitioners better understand their GAN models.", "target": ["GAN-Darstellungen werden im Detail untersucht, und es werden Sätze von Darstellungseinheiten gefunden, die die Erzeugung semantischer Konzepte im Output steuern."]} +{"source": "Historically, the pursuit of efficient inference has been one of the driving forces be-hind the research into new deep learning architectures and building blocks.Some of the recent examples include: the squeeze-and-excitation module of (Hu et al.,2018), depthwise separable convolutions in Xception (Chollet, 2017), and the inverted bottleneck in MobileNet v2 (Sandler et al., 2018). Notably, in all of these cases, the resulting building blocks enabled not only higher efficiency, but also higher accuracy, and found wide adoption in the field.In this work, we further expand the arsenal of efficient building blocks for neural network architectures; but instead of combining standard primitives (such as convolution), we advocate for the replacement of these dense primitives with their sparse counterparts. While the idea of using sparsity to decrease the parameter count is not new (Mozer & Smolensky, 1989), the conventional wisdom is that this reduction in theoretical FLOPs does not translate into real-world efficiency gains. We aim to correct this misconception by introducing a family of efficient sparse kernels for several hardware platforms, which we plan to open-source for the benefit of the community.Equipped with our efficient implementation of sparse primitives, we show that sparse versions of MobileNet v1 and MobileNet v2 architectures substantially outperform strong dense baselines on the efficiency-accuracy curve. On Snapdragon 835 our sparse networks outperform their dense equivalents by 1.1−2.2x – equivalent to approximately one entire generation of improvement. We hope that our findings will facilitate wider adoption of sparsity as a tool for creating efficient and accurate deep learning architectures.", "target": ["Sparse MobileNets sind schneller als Dense Nets mit den entsprechenden Kerneln."]} +{"source": "In this work, we address the semi-supervised classification of graph data, where the categories of those unlabeled nodes are inferred from labeled nodes as well as graph structures.Recent works often solve this problem with the advanced graph convolution in a conventional supervised manner, but the performance could be heavily affected when labeled data is scarce.Here we propose a Graph Inference Learning (GIL) framework to boost the performance of node classification by learning the inference of node labels on graph topology.To bridge the connection of two nodes, we formally define a structure relation by encapsulating node attributes, between-node paths and local topological structures together, which can make inference conveniently deduced from one node to another node.For learning the inference process, we further introduce meta-optimization on structure relations from training nodes to validation nodes, such that the learnt graph inference capability can be better self-adapted into test nodes.Comprehensive evaluations on four benchmark datasets (including Cora, Citeseer, Pubmed and NELL) demonstrate the superiority of our GIL when compared with other state-of-the-art methods in the semi-supervised node classification task.", "target": ["Wir schlagen einen neuartigen Rahmen für das Erlernen von Grapheninferenzen vor, indem wir Strukturbeziehungen aufbauen, um unbekannte Knotenbeschriftungen aus den beschrifteten Knoten in einer durchgängigen Weise abzuleiten."]} +{"source": "This paper proposes a method for efficient training of Q-function for continuous-state Markov Decision Processes (MDP), such that the traces of the resulting policies satisfy a Linear Temporal Logic (LTL) property.LTL, a modal logic, can express a wide range of time-dependent logical properties including safety and liveness.We convert the LTL property into a limit deterministic Buchi automaton with which a synchronized product MDP is constructed.The control policy is then synthesised by a reinforcement learning algorithm assuming that no prior knowledge is available from the MDP.The proposed method is evaluated in a numerical study to test the quality of the generated control policy and is compared against conventional methods for policy synthesis such as MDP abstraction (Voronoi quantizer) and approximate dynamic programming (fitted value iteration).", "target": ["Da Sicherheit zu einem kritischen Begriff beim maschinellen Lernen wird, glauben wir, dass diese Arbeit als Grundlage für eine Reihe von Forschungsrichtungen dienen kann, wie z. B. für sicherheitsbewusste Lernalgorithmen."]} +{"source": "We introduce two approaches for conducting efficient Bayesian inference in stochastic simulators containing nested stochastic sub-procedures, i.e., internal procedures for which the density cannot be calculated directly such as rejection sampling loops.The resulting class of simulators are used extensively throughout the sciences and can be interpreted as probabilistic generative models.However, drawing inferences from them poses a substantial challenge due to the inability to evaluate even their unnormalised density, preventing the use of many standard inference procedures like Markov Chain Monte Carlo (MCMC).To address this, we introduce inference algorithms based on a two-step approach that first approximates the conditional densities of the individual sub-procedures, before using these approximations to run MCMC methods on the full program.Because the sub-procedures can be dealt with separately and are lower-dimensional than that of the overall problem, this two-step process allows them to be isolated and thus be tractably dealt with, without placing restrictions on the overall dimensionality of the problem.We demonstrate the utility of our approach on a simple, artificially constructed simulator.", "target": ["Wir stellen zwei Ansätze für eine effiziente und skalierbare Inferenz in stochastischen Simulatoren vor, für die die Dichte nicht direkt ausgewertet werden kann, z.B. aufgrund von Rückweisungsschleifen."]} +{"source": "While adversarial training can improve robust accuracy (against an adversary), it sometimes hurts standard accuracy (when there is no adversary).Previous work has studied this tradeoff between standard and robust accuracy, but only in the setting where no predictor performs well on both objectives in the infinite data limit.In this paper, we show that even when the optimal predictor with infinite data performs well on both objectives, a tradeoff can still manifest itself with finite data.Furthermore, since our construction is based on a convex learning problem, we rule out optimization concerns, thus laying bare a fundamental tension between robustness and generalization.Finally, we show that robust self-training mostly eliminates this tradeoff by leveraging unlabeled data.", "target": ["Selbst wenn es bei unendlich vielen Daten keinen Kompromiss gibt, kann das adversarische Training selbst bei einem konvexen Problem eine schlechtere Standardgenauigkeit aufweisen."]} +{"source": "Skip connections are increasingly utilized by deep neural networks to improve accuracy and cost-efficiency.In particular, the recent DenseNet is efficient in computation and parameters, and achieves state-of-the-art predictions by directly connecting each feature layer to all previous ones.However, DenseNet's extreme connectivity pattern may hinder its scalability to high depths, and in applications like fully convolutional networks, full DenseNet connections are prohibitively expensive. This work first experimentally shows that one key advantage of skip connections is to have short distances among feature layers during backpropagation.Specifically, using a fixed number of skip connections, the connection patterns with shorter backpropagation distance among layers have more accurate predictions.Following this insight, we propose a connection template, Log-DenseNet, which, in comparison to DenseNet, only slightly increases the backpropagation distances among layers from 1 to ($1 + \\log_2 L$), but uses only $L\\log_2 L$ total connections instead of $O(L^2)$.Hence, \\logdenses are easier to scale than DenseNets, and no longer require careful GPU memory management.We demonstrate the effectiveness of our design principle by showing better performance than DenseNets on tabula rasa semantic segmentation, and competitive results on visual recognition.", "target": ["Wir zeigen, dass Shortcut-Verbindungen in Mustern platziert werden sollten, die die Abstände zwischen den Schichten während der Backpropagation minimieren, und entwerfen Netzwerke, die log L Abstände mit L log(L) Verbindungen erreichen."]} +{"source": "Building agents to interact with the web would allow for significant improvements in knowledge understanding and representation learning.However, web navigation tasks are difficult for current deep reinforcement learning (RL) models due to the large discrete action space and the varying number of actions between the states.In this work, we introduce DOM-Q-NET, a novel architecture for RL-based web navigation to address both of these problems.It parametrizes Q functions with separate networks for different action categories: clicking a DOM element and typing a string input. Our model utilizes a graph neural network to represent the tree-structured HTML of a standard web page. We demonstrate the capabilities of our model on the MiniWoB environment where we can match or outperform existing work without the use of expert demonstrations.Furthermore, we show 2x improvements in sample efficiency when training in the multi-task setting, allowing our model to transfer learned behaviours across tasks.", "target": ["Graphenbasiertes Deep Q Network für die Webnavigation"]} +{"source": "Learning disentangled representations that correspond to factors of variation in real-world data is critical to interpretable and human-controllable machine learning.Recently, concerns about the viability of learning disentangled representations in a purely unsupervised manner has spurred a shift toward the incorporation of weak supervision.However, there is currently no formalism that identifies when and how weak supervision will guarantee disentanglement.To address this issue, we provide a theoretical framework—including a calculus of disentanglement— to assist in analyzing the disentanglement guarantees (or lack thereof) conferred by weak supervision when coupled with learning algorithms based on distribution matching.We empirically verify the guarantees and limitations of several weak supervision methods (restricted labeling, match-pairing, and rank-pairing), demonstrating the predictive power and usefulness of our theoretical framework.", "target": ["Wir haben einen theoretischen Rahmen für die schwach überwachte Entflechtung konstruiert und viele Experimente durchgeführt, um die Theorie zu untermauern."]} +{"source": "Despite the growing interest in continual learning, most of its contemporary works have been studied in a rather restricted setting where tasks are clearly distinguishable, and task boundaries are known during training.However, if our goal is to develop an algorithm that learns as humans do, this setting is far from realistic, and it is essential to develop a methodology that works in a task-free manner.Meanwhile, among several branches of continual learning, expansion-based methods have the advantage of eliminating catastrophic forgetting by allocating new resources to learn new data.In this work, we propose an expansion-based approach for task-free continual learning.Our model, named Continual Neural Dirichlet Process Mixture (CN-DPM), consists of a set of neural network experts that are in charge of a subset of the data.CN-DPM expands the number of experts in a principled way under the Bayesian nonparametric framework.With extensive experiments, we show that our model successfully performs task-free continual learning for both discriminative and generative tasks such as image classification and image generation.", "target": ["Wir schlagen zum ersten Mal einen expansionsbasierten Ansatz für aufgabenfreies kontinuierliches Lernen vor. Unser Modell besteht aus einer Reihe von neuronalen Netz Experten und erweitert die Anzahl der Experten nach dem Bayes'schen nichtparametrischen Prinzip."]} +{"source": "Stochastic Gradient Descent (SGD) with Nesterov's momentum is a widely used optimizer in deep learning, which is observed to have excellent generalization performance.However, due to the large stochasticity, SGD with Nesterov's momentum is not robust, i.e., its performance may deviate significantly from the expectation.In this work, we propose Amortized Nesterov's Momentum, a special variant of Nesterov's momentum which has more robust iterates, faster convergence in the early stage and higher efficiency.Our experimental results show that this new momentum achieves similar (sometimes better) generalization performance with little-to-no tuning.In the convex case, we provide optimal convergence rates for our new methods and discuss how the theorems explain the empirical results.", "target": ["Amortisierung der Impulse von Nesterov für ein robusteres, leichteres und schnelleres Deep Learning Training."]} +{"source": "A state-of-the-art generative model, a ”factorized action variational autoencoder (FAVAE),” is presented for learning disentangled and interpretable representations from sequential data via the information bottleneck without supervision.The purpose of disentangled representation learning is to obtain interpretable and transferable representations from data.We focused on the disentangled representation of sequential data because there is a wide range of potential applications if disentanglement representation is extended to sequential data such as video, speech, and stock price data.Sequential data is characterized by dynamic factors and static factors: dynamic factors are time-dependent, and static factors are independent of time.Previous works succeed in disentangling static factors and dynamic factors by explicitly modeling the priors of latent variables to distinguish between static and dynamic factors.However, this model can not disentangle representations between dynamic factors, such as disentangling ”picking” and ”throwing” in robotic tasks.In this paper, we propose new model that can disentangle multiple dynamic factors.Since our method does not require modeling priors, it is capable of disentangling ”between” dynamic factors.In experiments, we show that FAVAE can extract the disentangled dynamic factors.", "target": ["Wir schlagen ein neues Modell vor, das mehrere dynamische Faktoren in sequentiellen Daten entflechten kann"]} +{"source": "Generative networks are promising models for specifying visual transformations.Unfortunately, certification of generative models is challenging as one needs to capture sufficient non-convexity so to produce precise bounds on the output.Existing verification methods either fail to scale to generative networks or do not capture enough non-convexity.In this work, we present a new verifier, called ApproxLine, that can certify non-trivial properties of generative networks.ApproxLine performs both deterministic and probabilistic abstract interpretation and captures infinite sets of outputs of generative networks.We show that ApproxLine can verify interesting interpolations in the network's latent space.", "target": ["Wir verifizieren deterministische und probabilistische Eigenschaften neuronaler Netze mit Hilfe nicht-konvexer Relaxationen über sichtbare Transformationen, die durch generative Modelle spezifiziert werden"]} +{"source": "Multi-view video summarization (MVS) lacks researchers’ attention due to their major challenges of inter-view correlations and overlapping of cameras.Most of the prior MVS works are offline, relying on only summary, needing extra communication bandwidth and transmission time with no focus on uncertain environments.Different from the existing methods, we propose edge intelligence based MVS and spatio-temporal features based activity recognition for IoT environments.We segment the multi-view videos on each slave device over edge into shots using light-weight CNN object detection model and compute mutual information among them to generate summary.Our system does not rely on summary only but encode and transmit it to a master device with neural computing stick (NCS) for intelligently computing inter-view correlations and efficiently recognizing activities, thereby saving computation resources, communication bandwidth, and transmission time.Experiments report an increase of 0.4 in F-measure score on MVS Office dataset as well as 0.2% and 2% increase in activity recognition accuracy over UCF-50 and YouTube 11 datasets, respectively, with lower storage and transmission time compared to state-of-the-art.The time complexity is decreased from 1.23 to 0.45 secs for a single frame processing, thereby generating 0.75 secs faster MVS.Furthermore, we made a new dataset by synthetically adding fog to an MVS dataset to show the adaptability of our system for both certain and uncertain surveillance environments.", "target": ["Ein effizientes Multi-View-Video-Zusammenfassungsschema für die Aktivitätserkennung in IoT-Umgebungen."]} +{"source": "A central goal in the study of the primate visual cortex and hierarchical models for object recognition is understanding how and why single units trade off invariance versus sensitivity to image transformations.For example, in both deep networks and visual cortex there is substantial variation from layer-to-layer and unit-to-unit in the degree of translation invariance.Here, we provide theoretical insight into this variation and its consequences for encoding in a deep network.Our critical insight comes from the fact that rectification simultaneously decreases response variance and correlation across responses to transformed stimuli, naturally inducing a positive relationship between invariance and dynamic range.Invariant input units then tend to drive the network more than those sensitive to small image transformations.We discuss consequences of this relationship for AI: deep nets naturally weight invariant units over sensitive units, and this can be strengthened with training, perhaps contributing to generalization performance.Our results predict a signature relationship between invariance and dynamic range that can now be tested in future neurophysiological studies.", "target": ["Die Entzerrung in tiefen neuronalen Netzen führt natürlich dazu, dass sie eine invariante Darstellung bevorzugen."]} +{"source": "We study the use of the Wave-U-Net architecture for speech enhancement, a model introduced by Stoller et al for the separation of music vocals and accompaniment. This end-to-end learning method for audio source separation operates directly in the time domain, permitting the integrated modelling of phase information and being able to take large temporal contexts into account. Our experiments show that the proposed method improves several metrics, namely PESQ, CSIG, CBAK, COVL and SSNR, over the state-of-the-art with respect to the speech enhancement task on the Voice Bank corpus (VCTK) dataset.We find that a reduced number of hidden layers is sufficient for speech enhancement in comparison to the original system designed for singing voice separation in music.We see this initial result as an encouraging signal to further explore speech enhancement in the time-domain, both as an end in itself and as a pre-processing step to speech recognition systems.", "target": ["Die Wave-U-Net-Architektur, die vor kurzem von Stoller et al. für die Trennung von Musikquellen eingeführt wurde, ist äußerst effektiv für die Sprachanhebung und übertrifft den Stand der Technik."]} +{"source": "Reinforcement learning (RL) is a powerful framework for solving problems by exploring and learning from mistakes.However, in the context of autonomous vehicle (AV) control, requiring an agent to make mistakes, or even allowing mistakes, can be quite dangerous and costly in the real world.For this reason, AV RL is generally only viable in simulation.Because these simulations have imperfect representations, particularly with respect to graphics, physics, and human interaction, we find motivation for a framework similar to RL, suitable to the real world.To this end, we formulate a learning framework that learns from restricted exploration by having a human demonstrator do the exploration.Existing work on learning from demonstration typically either assumes the collected data is performed by an optimal expert, or requires potentially dangerous exploration to find the optimal policy.We propose an alternative framework that learns continuous control from only safe behavior.One of our key insights is that the problem becomes tractable if the feedback score that rates the demonstration applies to the atomic action, as opposed to the entire sequence of actions.We use human experts to collect driving data as well as to label the driving data through a framework we call ``Backseat Driver'', giving us state-action pairs matched with scalar values representing the score for the action.We call the more general learning framework ReNeg, since it learns a regression from states to actions given negative as well as positive examples.We empirically validate several models in the ReNeg framework, testing on lane-following with limited data.We find that the best solution in this context outperforms behavioral cloning has strong connections to stochastic policy gradient approaches.", "target": ["Wir stellen einen neuartigen Rahmen für das Lernen aus Demonstrationen vor, der kontinuierliches menschliches Feedback nutzt; wir evaluieren diesen Rahmen für die kontinuierliche Steuerung von autonomen Fahrzeugen."]} +{"source": "We explore the use of Vector Quantized Variational AutoEncoder (VQ-VAE) models for large scale image generation.To this end, we scale and enhance the autoregressive priors used in VQ-VAE to generate synthetic samples of much higher coherence and fidelity than possible before. We use simple feed-forward encoder and decoder networks, thus our model is an attractive candidate for applications where the encoding and decoding speed is critical.Additionally, this allows us to only sample autoregressively in the compressed latent space, which is an order of magnitude faster than sampling in the pixel space, especially for large images.We demonstrate that a multi-scale hierarchical organization of VQ-VAE, augmented with powerful priors over the latent codes, is able to generate samples with quality that rivals that of state of the art Generative Adversarial Networks on multifaceted datasets such as ImageNet, while not suffering from GAN's known shortcomings such as mode collapse and lack of diversity.", "target": ["skalieren und verbessern VQ-VAE mit leistungsstarken Prioren, um nahezu realistische Bilder zu erzeugen."]} +{"source": "We present a graph neural network assisted Monte Carlo Tree Search approach for the classical traveling salesman problem (TSP).We adopt a greedy algorithm framework to construct the optimal solution to TSP by adding the nodes successively.A graph neural network (GNN) is trained to capture the local and global graph structure and give the prior probability of selecting each vertex every step.The prior probability provides a heuristics for MCTS, and the MCTS output is an improved probability for selecting the successive vertex, as it is the feedback information by fusing the prior with the scouting procedure.Experimental results on TSP up to 100 nodes demonstrate that the proposed method obtains shorter tours than other learning-based methods.", "target": ["Ein Graph Neural Network Assisted Monte Carlo Tree Search Ansatz für das Problem des Handlungsreisenden"]} +{"source": "Graph neural networks have recently achieved great successes in predicting quantum mechanical properties of molecules.These models represent a molecule as a graph using only the distance between atoms (nodes) and not the spatial direction from one atom to another.However, directional information plays a central role in empirical potentials for molecules, e.g. in angular potentials.To alleviate this limitation we propose directional message passing, in which we embed the messages passed between atoms instead of the atoms themselves.Each message is associated with a direction in coordinate space.These directional message embeddings are rotationally equivariant since the associated directions rotate with the molecule.We propose a message passing scheme analogous to belief propagation, which uses the directional information by transforming messages based on the angle between them.Additionally, we use spherical Bessel functions to construct a theoretically well-founded, orthogonal radial basis that achieves better performance than the currently prevalent Gaussian radial basis functions while using more than 4x fewer parameters.We leverage these innovations to construct the directional message passing neural network (DimeNet).DimeNet outperforms previous GNNs on average by 77% on MD17 and by 41% on QM9.", "target": ["Direktionale Nachrichtenübermittlung beinhaltet räumliche Richtungsinformationen zur Verbesserung von neuronalen Graphennetzen."]} +{"source": "Training an agent to solve control tasks directly from high-dimensional images with model-free reinforcement learning (RL) has proven difficult.The agent needs to learn a latent representation together with a control policy to perform the task.Fitting a high-capacity encoder using a scarce reward signal is not only extremely sample inefficient, but also prone to suboptimal convergence.Two ways to improve sample efficiency are to learn a good feature representation and use off-policy algorithms.We dissect various approaches of learning good latent features, and conclude that the image reconstruction loss is the essential ingredient that enables efficient and stable representation learning in image-based RL.Following these findings, we devise an off-policy actor-critic algorithm with an auxiliary decoder that trains end-to-end and matches state-of-the-art performance across both model-free and model-based algorithms on many challenging control tasks.We release our code to encourage future research on image-based RL.", "target": ["Wir entwerfen eine einfache und effiziente modellfreie Off-Policy-Methode für bildbasiertes Reinforcement Learning, die den modernsten modellbasierten Methoden in Bezug auf die Effizienz der Beispiele entspricht."]} +{"source": "Large deep neural networks require huge memory to run and their running speed is sometimes too slow for real applications.Therefore network size reduction with keeping accuracy is crucial for practical applications.We present a novel neural network operator, chopout, with which neural networks are trained, even in a single training process, so as to truncated sub-networks perform as well as possible.Chopout is easy to implement and integrate into most type of existing neural networks.Furthermore it enables to reduce size of networks and latent representations even after training just by truncating layers.We show its effectiveness through several experiments.", "target": ["Wir stellen einen neuartigen, einfachen Operator, Chopout, vor, mit dem neuronale Netze sogar in einem einzigen Trainingsprozess so trainiert werden, dass die abgeschnittenen Teilnetze die bestmögliche Leistung erbringen."]} +{"source": "Generative Adversarial Networks (GANs) have been shown to produce realistically looking synthetic images with remarkable success, yet their performance seems less impressive when the training set is highly diverse.In order to provide a better fit to the target data distribution when the dataset includes many different classes, we propose a variant of the basic GAN model, a Multi-Modal Gaussian-Mixture GAN (GM-GAN), where the probability distribution over the latent space is a mixture of Gaussians.We also propose a supervised variant which is capable of conditional sample synthesis.In order to evaluate the model's performance, we propose a new scoring method which separately takes into account two (typically conflicting) measures - diversity vs. quality of the generated data. Through a series of experiments, using both synthetic and real-world datasets, we quantitatively show that GM-GANs outperform baselines, both when evaluated using the commonly used Inception Score, and when evaluated using our own alternative scoring method.In addition, we qualitatively demonstrate how the unsupervised variant of GM-GAN tends to map latent vectors sampled from different Gaussians in the latent space to samples of different classes in the data space.We show how this phenomenon can be exploited for the task of unsupervised clustering, and provide quantitative evaluation showing the superiority of our method for the unsupervised clustering of image datasets.Finally, we demonstrate a feature which further sets our model apart from other GAN models: the option to control the quality-diversity trade-off by altering, post-training, the probability distribution of the latent space.This allows one to sample higher quality and lower diversity samples, or vice versa, according to one's needs.", "target": ["Multimodale Guassianische Verteilung des latenten Raums in GAN-Modellen verbessert die Leistung und ermöglicht einen Kompromiss zwischen Qualität und Vielfalt"]} +{"source": "Distributed optimization is essential for training large models on large datasets.Multiple approaches have been proposed to reduce the communication overhead in distributed training, such as synchronizing only after performing multiple local SGD steps, and decentralized methods (e.g., using gossip algorithms) to decouple communications among workers.Although these methods run faster than AllReduce-based methods, which use blocking communication before every update, the resulting models may be less accurate after the same number of updates.Inspired by the BMUF method of Chen & Huo (2016), we propose a slow momentum (SloMo) framework, where workers periodically synchronize and perform a momentum update, after multiple iterations of a base optimization algorithm.Experiments on image classification and machine translation tasks demonstrate that SloMo consistently yields improvements in optimization and generalization performance relative to the base optimizer, even when the additional overhead is amortized over many updates so that the SloMo runtime is on par with that of the base optimizer.We provide theoretical convergence guarantees showing that SloMo converges to a stationary point of smooth non-convex losses.Since BMUF is a particular instance of the SloMo framework, our results also correspond to the first theoretical convergence guarantees for BMUF.", "target": ["SlowMo verbessert die Optimierungs- und Generalisierungsleistung kommunikationseffizienter dezentraler Algorithmen ohne Geschwindigkeitseinbußen."]} +{"source": "Structural planning is important for producing long sentences, which is a missing part in current language generation models.In this work, we add a planning phase in neural machine translation to control the coarse structure of output sentences.The model first generates some planner codes, then predicts real output words conditioned on them.The codes are learned to capture the coarse structure of the target sentence.In order to learn the codes, we design an end-to-end neural network with a discretization bottleneck, which predicts the simplified part-of-speech tags of target sentences.Experiments show that the translation performance are generally improved by planning ahead.We also find that translations with different structures can be obtained by manipulating the planner codes.", "target": ["Die syntaktische Struktur der Übersetzung mit Hilfe von Codes planen"]} +{"source": "Interpolation of data in deep neural networks has become a subject of significant research interest. We prove that over-parameterized single layer fully connected autoencoders do not merely interpolate, but rather, memorize training data: they produce outputs in (a non-linear version of) the span of the training examples.In contrast to fully connected autoencoders, we prove that depth is necessary for memorization in convolutional autoencoders. Moreover, we observe that adding nonlinearity to deep convolutional autoencoders results in a stronger form of memorization: instead of outputting points in the span of the training images, deep convolutional autoencoders tend to output individual training images. Since convolutional autoencoder components are building blocks of deep convolutional networks, we envision that our findings will shed light on the important question of the inductive bias in over-parameterized deep networks.", "target": ["Wir identifizieren Memorisierung als induktive Verzerrung der Interpolation in überparametrisierten, vollständig vernetzten und faltbaren Autokodierern."]} +{"source": "The tremendous success of deep neural networks has motivated the need to better understand the fundamental properties of these networks, but many of the theoretical results proposed have only been for shallow networks.In this paper, we study an important primitive for understanding the meaningful input space of a deep network: span recovery.For $k