text
stringlengths 29
3.31k
| label
sequencelengths 1
11
|
---|---|
Satellite image time series, bolstered by their growing availability, are at
the forefront of an extensive effort towards automated Earth monitoring by
international institutions. In particular, large-scale control of agricultural
parcels is an issue of major political and economic importance. In this regard,
hybrid convolutional-recurrent neural architectures have shown promising
results for the automated classification of satellite image time series.We
propose an alternative approach in which the convolutional layers are
advantageously replaced with encoders operating on unordered sets of pixels to
exploit the typically coarse resolution of publicly available satellite images.
We also propose to extract temporal features using a bespoke neural
architecture based on self-attention instead of recurrent networks. We
demonstrate experimentally that our method not only outperforms previous
state-of-the-art approaches in terms of precision, but also significantly
decreases processing time and memory requirements. Lastly, we release a large
open-access annotated dataset as a benchmark for future work on satellite image
time series. | [
"cs.CV"
] |
Previous methods for skeleton-based gesture recognition mostly arrange the
skeleton sequence into a pseudo picture or spatial-temporal graph and apply
deep Convolutional Neural Network (CNN) or Graph Convolutional Network (GCN)
for feature extraction. Although achieving superior results, these methods have
inherent limitations in dynamically capturing local features of interactive
hand parts, and the computing efficiency still remains a serious issue. In this
work, the self-attention mechanism is introduced to alleviate this problem.
Considering the hierarchical structure of hand joints, we propose an efficient
hierarchical self-attention network (HAN) for skeleton-based gesture
recognition, which is based on pure self-attention without any CNN, RNN or GCN
operators. Specifically, the joint self-attention module is used to capture
spatial features of fingers, the finger self-attention module is designed to
aggregate features of the whole hand. In terms of temporal features, the
temporal self-attention module is utilized to capture the temporal dynamics of
the fingers and the entire hand. Finally, these features are fused by the
fusion self-attention module for gesture classification. Experiments show that
our method achieves competitive results on three gesture recognition datasets
with much lower computational complexity. | [
"cs.CV"
] |
Many time series are effectively generated by a combination of deterministic
continuous flows along with discrete jumps sparked by stochastic events.
However, we usually do not have the equation of motion describing the flows, or
how they are affected by jumps. To this end, we introduce Neural Jump
Stochastic Differential Equations that provide a data-driven approach to learn
continuous and discrete dynamic behavior, i.e., hybrid systems that both flow
and jump. Our approach extends the framework of Neural Ordinary Differential
Equations with a stochastic process term that models discrete events. We then
model temporal point processes with a piecewise-continuous latent trajectory,
where the discontinuities are caused by stochastic events whose conditional
intensity depends on the latent state. We demonstrate the predictive
capabilities of our model on a range of synthetic and real-world marked point
process datasets, including classical point processes (such as Hawkes
processes), awards on Stack Overflow, medical records, and earthquake
monitoring. | [
"cs.LG",
"stat.ML"
] |
We present a novel approach to image manipulation and understanding by
simultaneously learning to segment object masks, paste objects to another
background image, and remove them from original images. For this purpose, we
develop a novel generative model for compositional image generation, SEIGAN
(Segment-Enhance-Inpaint Generative Adversarial Network), which learns these
three operations together in an adversarial architecture with additional cycle
consistency losses. To train, SEIGAN needs only bounding box supervision and
does not require pairing or ground truth masks. SEIGAN produces better
generated images (evaluated by human assessors) than other approaches and
produces high-quality segmentation masks, improving over other adversarially
trained approaches and getting closer to the results of fully supervised
training. | [
"cs.CV",
"cs.LG",
"cs.NE"
] |
Object detection and motion parameters estimation are crucial tasks for
self-driving vehicle safe navigation in a complex urban environment. In this
work we propose a novel real-time approach of temporal context aggregation for
motion detection and motion parameters estimation based on 3D point cloud
sequence. We introduce an ego-motion compensation layer to achieve real-time
inference with performance comparable to a naive odometric transform of the
original point cloud sequence. Not only is the proposed architecture capable of
estimating the motion of common road participants like vehicles or pedestrians
but also generalizes to other object categories which are not present in
training data. We also conduct an in-deep analysis of different temporal
context aggregation strategies such as recurrent cells and 3D convolutions.
Finally, we provide comparison results of our state-of-the-art model with
existing solutions on KITTI Scene Flow dataset. | [
"cs.CV",
"cs.RO"
] |
We explore the problem of classification within a medical image data-set
based on a feature vector extracted from the deepest layer of pre-trained
Convolution Neural Networks. We have used feature vectors from several
pre-trained structures, including networks with/without transfer learning to
evaluate the performance of pre-trained deep features versus CNNs which have
been trained by that specific dataset as well as the impact of transfer
learning with a small number of samples. All experiments are done on Kimia
Path24 dataset which consists of 27,055 histopathology training patches in 24
tissue texture classes along with 1,325 test patches for evaluation. The result
shows that pre-trained networks are quite competitive against training from
scratch. As well, fine-tuning does not seem to add any tangible improvement for
VGG16 to justify additional training while we observed considerable improvement
in retrieval and classification accuracy when we fine-tuned the Inception
structure. | [
"cs.CV"
] |
Visual-semantic embedding is an interesting research topic because it is
useful for various tasks, such as visual question answering (VQA), image-text
retrieval, image captioning, and scene graph generation. In this paper, we
focus on zero-shot image retrieval using sentences as queries and present a
survey of the technological trends in this area. First, we provide a
comprehensive overview of the history of the technology, starting with a
discussion of the early studies of image-to-text matching and how the
technology has evolved over time. In addition, a description of the datasets
commonly used in experiments and a comparison of the evaluation results of each
method are presented. We also introduce the implementation available on github
for use in confirming the accuracy of experiments and for further improvement.
We hope that this survey paper will encourage researchers to further develop
their research on bridging images and languages. | [
"cs.CV"
] |
In this paper, we propose a novel approach to convert given speech audio to a
photo-realistic speaking video of a specific person, where the output video has
synchronized, realistic, and expressive rich body dynamics. We achieve this by
first generating 3D skeleton movements from the audio sequence using a
recurrent neural network (RNN), and then synthesizing the output video via a
conditional generative adversarial network (GAN). To make the skeleton movement
realistic and expressive, we embed the knowledge of an articulated 3D human
skeleton and a learned dictionary of personal speech iconic gestures into the
generation process in both learning and testing pipelines. The former prevents
the generation of unreasonable body distortion, while the later helps our model
quickly learn meaningful body movement through a few recorded videos. To
produce photo-realistic and high-resolution video with motion details, we
propose to insert part attention mechanisms in the conditional GAN, where each
detailed part, e.g. head and hand, is automatically zoomed in to have their own
discriminators. To validate our approach, we collect a dataset with 20
high-quality videos from 1 male and 1 female model reading various documents
under different topics. Compared with previous SoTA pipelines handling similar
tasks, our approach achieves better results by a user study. | [
"cs.CV",
"cs.LG",
"eess.AS"
] |
Kernel mean embedding is a useful tool to compare probability measures.
Despite its usefulness, kernel mean embedding considers infinite-dimensional
features, which are challenging to handle in the context of differentially
private data generation. A recent work proposes to approximate the kernel mean
embedding of data distribution using finite-dimensional random features, where
the sensitivity of the features becomes analytically tractable. More
importantly, this approach significantly reduces the privacy cost, compared to
other known privatization methods (e.g., DP-SGD), as the approximate kernel
mean embedding of the data distribution is privatized only once and can then be
repeatedly used during training of a generator without incurring any further
privacy cost. However, the required number of random features is excessively
high, often ten thousand to a hundred thousand, which worsens the sensitivity
of the approximate kernel mean embedding. To improve the sensitivity, we
propose to replace random features with Hermite polynomial features. Unlike the
random features, the Hermite polynomial features are ordered, where the
features at the low orders contain more information on the distribution than
those at the high orders. Hence, a relatively low order of Hermite polynomial
features can more accurately approximate the mean embedding of the data
distribution compared to a significantly higher number of random features. As a
result, using the Hermite polynomial features, we significantly improve the
privacy-accuracy trade-off, reflected in the high quality and diversity of the
generated data, when tested on several heterogeneous tabular datasets, as well
as several image benchmark datasets. | [
"cs.LG",
"cs.CR",
"stat.ML"
] |
Single-stage object detectors have been widely applied in computer vision
applications due to their high efficiency. However, we find that the loss
functions adopted by single-stage object detectors hurt the localization
accuracy seriously. Firstly, the standard cross-entropy loss for classification
is independent of the localization task and drives all the positive examples to
learn as high classification scores as possible regardless of localization
accuracy during training. As a result, there will be many detections that have
high classification scores but low IoU or detections that have low
classification scores but high IoU. Secondly, for the standard smooth L1 loss,
the gradient is dominated by the outliers that have poor localization accuracy
during training. The above two problems will decrease the localization accuracy
of single-stage detectors. In this work, IoU-balanced loss functions that
consist of IoU-balanced classification loss and IoU-balanced localization loss
are proposed to solve the above problems. The IoU-balanced classification loss
pays more attention to positive examples with high IoU and can enhance the
correlation between classification and localization tasks. The IoU-balanced
localization loss decreases the gradient of examples with low IoU and increases
the gradient of examples with high IoU, which can improve the localization
accuracy of models. Extensive experiments on challenging public datasets such
as MS COCO, PASCAL VOC and Cityscapes demonstrate that both IoU-balanced losses
can bring substantial improvement for the popular single-stage detectors,
especially for the localization accuracy. On COCO test-dev, the proposed
methods can substantially improve AP by $1.0\%\sim1.7\%$ and AP75 by
$1.0\%\sim2.4\%$. On PASCAL VOC, it can also substantially improve AP by
$1.3\%\sim1.5\%$ and AP80, AP90 by $1.6\%\sim3.9\%$. | [
"cs.CV"
] |
This paper aims at developing an integrated system of clothing co-parsing, in
order to jointly parse a set of clothing images (unsegmented but annotated with
tags) into semantic configurations. We propose a data-driven framework
consisting of two phases of inference. The first phase, referred as "image
co-segmentation", iterates to extract consistent regions on images and jointly
refines the regions over all images by employing the exemplar-SVM (E-SVM)
technique [23]. In the second phase (i.e. "region co-labeling"), we construct a
multi-image graphical model by taking the segmented regions as vertices, and
incorporate several contexts of clothing configuration (e.g., item location and
mutual interactions). The joint label assignment can be solved using the
efficient Graph Cuts algorithm. In addition to evaluate our framework on the
Fashionista dataset [30], we construct a dataset called CCP consisting of 2098
high-resolution street fashion photos to demonstrate the performance of our
system. We achieve 90.29% / 88.23% segmentation accuracy and 65.52% / 63.89%
recognition rate on the Fashionista and the CCP datasets, respectively, which
are superior compared with state-of-the-art methods. | [
"cs.CV",
"68U01"
] |
The capability of reinforcement learning (RL) agent directly depends on the
diversity of learning scenarios the environment generates and how closely it
captures real-world situations. However, existing environments/simulators lack
the support to systematically model distributions over initial states and
transition dynamics. Furthermore, in complex domains such as soccer, the space
of possible scenarios is infinite, which makes it impossible for one research
group to provide a comprehensive set of scenarios to train, test, and benchmark
RL algorithms. To address this issue, for the first time, we adopt an existing
formal scenario specification language, SCENIC, to intuitively model and
generate interactive scenarios. We interfaced SCENIC to Google Research Soccer
environment to create a platform called SCENIC4RL. Using this platform, we
provide a dataset consisting of 36 scenario programs encoded in SCENIC and
demonstration data generated from a subset of them. We share our experimental
results to show the effectiveness of our dataset and the platform to train,
test, and benchmark RL algorithms. More importantly, we open-source our
platform to enable RL community to collectively contribute to constructing a
comprehensive set of scenarios. | [
"cs.LG",
"cs.AI"
] |
This paper addresses representational block named Hierarchical-Split Block,
which can be taken as a plug-and-play block to upgrade existing convolutional
neural networks, improves model performance significantly in a network.
Hierarchical-Split Block contains many hierarchical split and concatenate
connections within one single residual block. We find multi-scale features is
of great importance for numerous vision tasks. Moreover, Hierarchical-Split
block is very flexible and efficient, which provides a large space of potential
network architectures for different applications. In this work, we present a
common backbone based on Hierarchical-Split block for tasks: image
classification, object detection, instance segmentation and semantic image
segmentation/parsing. Our approach shows significant improvements over all
these core tasks in comparison with the baseline. As shown in Figure1, for
image classification, our 50-layers network(HS-ResNet50) achieves 81.28% top-1
accuracy with competitive latency on ImageNet-1k dataset. It also outperforms
most state-of-the-art models. The source code and models will be available on:
https://github.com/PaddlePaddle/PaddleClas | [
"cs.CV"
] |
Molecular property prediction (e.g., energy) is an essential problem in
chemistry and biology. Unfortunately, many supervised learning methods usually
suffer from the problem of scarce labeled molecules in the chemical space,
where such property labels are generally obtained by Density Functional Theory
(DFT) calculation which is extremely computational costly. An effective
solution is to incorporate the unlabeled molecules in a semi-supervised
fashion. However, learning semi-supervised representation for large amounts of
molecules is challenging, including the joint representation issue of both
molecular essence and structure, the conflict between representation and
property leaning. Here we propose a novel framework called Active
Semi-supervised Graph Neural Network (ASGN) by incorporating both labeled and
unlabeled molecules. Specifically, ASGN adopts a teacher-student framework. In
the teacher model, we propose a novel semi-supervised learning method to learn
general representation that jointly exploits information from molecular
structure and molecular distribution. Then in the student model, we target at
property prediction task to deal with the learning loss conflict. At last, we
proposed a novel active learning strategy in terms of molecular diversities to
select informative data during the whole framework learning. We conduct
extensive experiments on several public datasets. Experimental results show the
remarkable performance of our ASGN framework. | [
"cs.LG",
"stat.ML"
] |
Current deep reinforcement learning (DRL) algorithms utilize randomness in
simulation environments to assume complete coverage in the state space.
However, particularly in high dimensions, relying on randomness may lead to
gaps in coverage of the trained DRL neural network model, which in turn may
lead to drastic and often fatal real-world situations. To the best of the
author's knowledge, the assessment of coverage for DRL is lacking in current
research literature. Therefore, in this paper, a novel measure, Approximate
Pseudo-Coverage (APC), is proposed for assessing the coverage in DRL
applications. We propose to calculate APC by projecting the high dimensional
state space on to a lower dimensional manifold and quantifying the occupied
space. Furthermore, we utilize an exploration-exploitation strategy for
coverage maximization using Rapidly-Exploring Random Tree (RRT). The efficacy
of the assessment and the acceleration of coverage is demonstrated on standard
tasks such as Cartpole, highway-env. | [
"cs.LG",
"cs.AI"
] |
Scene graph generation has received growing attention with the advancements
in image understanding tasks such as object detection, attributes and
relationship prediction,~\etc. However, existing datasets are biased in terms
of object and relationship labels, or often come with noisy and missing
annotations, which makes the development of a reliable scene graph prediction
model very challenging. In this paper, we propose a novel scene graph
generation algorithm with external knowledge and image reconstruction loss to
overcome these dataset issues. In particular, we extract commonsense knowledge
from the external knowledge base to refine object and phrase features for
improving generalizability in scene graph generation. To address the bias of
noisy object annotations, we introduce an auxiliary image reconstruction path
to regularize the scene graph generation network. Extensive experiments show
that our framework can generate better scene graphs, achieving the
state-of-the-art performance on two benchmark datasets: Visual Relationship
Detection and Visual Genome datasets. | [
"cs.CV"
] |
Unsupervised learning on imbalanced data is challenging because, when given
imbalanced data, current model is often dominated by the major category and
ignores the categories with small amount of data. We develop a latent variable
model that can cope with imbalanced data by dividing the latent space into a
shared space and a private space. Based on Gaussian Process Latent Variable
Models, we propose a new kernel formulation that enables the separation of
latent space and derives an efficient variational inference method. The
performance of our model is demonstrated with an imbalanced medical image
dataset. | [
"cs.LG",
"stat.ML"
] |
Labelled image datasets have played a critical role in high-level image
understanding. However, the process of manual labelling is both time-consuming
and labor intensive. To reduce the cost of manual labelling, there has been
increased research interest in automatically constructing image datasets by
exploiting web images. Datasets constructed by existing methods tend to have a
weak domain adaptation ability, which is known as the "dataset bias problem".
To address this issue, we present a novel image dataset construction framework
that can be generalized well to unseen target domains. Specifically, the given
queries are first expanded by searching the Google Books Ngrams Corpus to
obtain a rich semantic description, from which the visually non-salient and
less relevant expansions are filtered out. By treating each selected expansion
as a "bag" and the retrieved images as "instances", image selection can be
formulated as a multi-instance learning problem with constrained positive bags.
We propose to solve the employed problems by the cutting-plane and
concave-convex procedure (CCCP) algorithm. By using this approach, images from
different distributions can be kept while noisy images are filtered out. To
verify the effectiveness of our proposed approach, we build an image dataset
with 20 categories. Extensive experiments on image classification,
cross-dataset generalization, diversity comparison and object detection
demonstrate the domain robustness of our dataset. | [
"cs.CV",
"cs.MM"
] |
Adversarial examples are inputs intentionally perturbed with the aim of
forcing a machine learning model to produce a wrong prediction, while the
changes are not easily detectable by a human. Although this topic has been
intensively studied in the image domain, classification tasks in the audio
domain have received less attention. In this paper we address the existence of
universal perturbations for speech command classification. We provide evidence
that universal attacks can be generated for speech command classification
tasks, which are able to generalize across different models to a significant
extent. Additionally, a novel analytical framework is proposed for the
evaluation of universal perturbations under different levels of universality,
demonstrating that the feasibility of generating effective perturbations
decreases as the universality level increases. Finally, we propose a more
detailed and rigorous framework to measure the amount of distortion introduced
by the perturbations, demonstrating that the methods employed by convention are
not realistic in audio-based problems. | [
"cs.LG",
"eess.AS",
"stat.ML"
] |
Understanding the implication of point cloud is still challenging to achieve
the goal of classification or segmentation due to the irregular and sparse
structure of point cloud. As we have known, PointNet architecture as a
ground-breaking work for point cloud which can learn efficiently shape features
directly on unordered 3D point cloud and have achieved favorable performance.
However, this model fail to consider the fine-grained semantic information of
local structure for point cloud. Afterwards, many valuable works are proposed
to enhance the performance of PointNet by means of semantic features of local
patch for point cloud. In this paper, a multi-scale receptive fields graph
attention network (named after MRFGAT) for point cloud classification is
proposed. By focusing on the local fine features of point cloud and applying
multi attention modules based on channel affinity, the learned feature map for
our network can well capture the abundant features information of point cloud.
The proposed MRFGAT architecture is tested on ModelNet10 and ModelNet40
datasets, and results show it achieves state-of-the-art performance in shape
classification tasks. | [
"cs.CV"
] |
Fashion trend forecasting is of great research significance in providing
useful suggestions for both fashion companies and fashion lovers. Although
various studies have been devoted to tackling this challenging task, they only
studied limited fashion elements with highly seasonal or simple patterns, which
could hardly reveal the real complex fashion trends. Moreover, the mainstream
solutions for this task are still statistical-based and solely focus on
time-series data modeling, which limit the forecast accuracy. Towards
insightful fashion trend forecasting, previous work [1] proposed to analyze
more fine-grained fashion elements which can informatively reveal fashion
trends. Specifically, it focused on detailed fashion element trend forecasting
for specific user groups based on social media data. In addition, it proposed a
neural network-based method, namely KERN, to address the problem of fashion
trend modeling and forecasting. In this work, to extend the previous work, we
propose an improved model named Relation Enhanced Attention Recurrent (REAR)
network. Compared to KERN, the REAR model leverages not only the relations
among fashion elements but also those among user groups, thus capturing more
types of correlations among various fashion trends. To further improve the
performance of long-range trend forecasting, the REAR method devises a sliding
temporal attention mechanism, which is able to capture temporal patterns on
future horizons better. Extensive experiments and more analysis have been
conducted on the FIT and GeoStyle datasets to evaluate the performance of REAR.
Experimental and analytical results demonstrate the effectiveness of the
proposed REAR model in fashion trend forecasting, which also show the
improvement of REAR compared to the KERN. | [
"cs.LG",
"cs.IR",
"cs.MM"
] |
Reasoning over multiple modalities, e.g. in Visual Question Answering (VQA),
requires an alignment of semantic concepts across domains. Despite the
widespread success of end-to-end learning, today's multimodal pipelines by and
large leverage pre-extracted, fixed features from object detectors, typically
Faster R-CNN, as representations of the visual world. The obvious downside is
that the visual representation is not specifically tuned to the multimodal task
at hand. At the same time, while transformer-based object detectors have gained
popularity, they have not been employed in today's multimodal pipelines. We
address both shortcomings with TxT, a transformer-based crossmodal pipeline
that enables fine-tuning both language and visual components on the downstream
task in a fully end-to-end manner. We overcome existing limitations of
transformer-based detectors for multimodal reasoning regarding the integration
of global context and their scalability. Our transformer-based multimodal model
achieves considerable gains from end-to-end learning for multimodal question
answering. | [
"cs.CV",
"cs.CL"
] |
Nowadays, deep learning techniques are widely used for lane detection, but
application in low-light conditions remains a challenge until this day.
Although multi-task learning and contextual-information-based methods have been
proposed to solve the problem, they either require additional manual
annotations or introduce extra inference overhead respectively. In this paper,
we propose a style-transfer-based data enhancement method, which uses
Generative Adversarial Networks (GANs) to generate images in low-light
conditions, that increases the environmental adaptability of the lane detector.
Our solution consists of three parts: the proposed SIM-CycleGAN, light
conditions style transfer and lane detection network. It does not require
additional manual annotations nor extra inference overhead. We validated our
methods on the lane detection benchmark CULane using ERFNet. Empirically, lane
detection model trained using our method demonstrated adaptability in low-light
conditions and robustness in complex scenarios. Our code for this paper will be
publicly available. | [
"cs.CV"
] |
Graph neural networks have become an important tool for modeling structured
data. In many real-world systems, intricate hidden information may exist, e.g.,
heterogeneity in nodes/edges, static node/edge attributes, and spatiotemporal
node/edge features. However, most existing methods only take part of the
information into consideration. In this paper, we present the Co-evolved Meta
Graph Neural Network (CoMGNN), which applies meta graph attention to
heterogeneous graphs with co-evolution of node and edge states. We further
propose a spatiotemporal adaption of CoMGNN (ST-CoMGNN) for modeling
spatiotemporal patterns on nodes and edges. We conduct experiments on two
large-scale real-world datasets. Experimental results show that our models
significantly outperform the state-of-the-art methods, demonstrating the
effectiveness of encoding diverse information from different aspects. | [
"cs.LG",
"cs.SI",
"stat.ML"
] |
Though used extensively, the concept and process of machine learning (ML)
personalization have generally received little attention from academics,
practitioners, and the general public. We describe the ML approach as relying
on the metaphor of the person as a feature vector and contrast this with
humanistic views of the person. In light of the recent calls by the IEEE to
consider the effects of ML on human well-being, we ask whether ML
personalization can be reconciled with these humanistic views of the person,
which highlight the importance of moral and social identity. As human behavior
increasingly becomes digitized, analyzed, and predicted, to what extent do our
subsequent decisions about what to choose, buy, or do, made both by us and
others, reflect who we are as persons? This paper first explicates the term
personalization by considering ML personalization and highlights its relation
to humanistic conceptions of the person, then proposes several dimensions for
evaluating the degree of personalization of ML personalized scores. By doing
so, we hope to contribute to current debate on the issues of algorithmic bias,
transparency, and fairness in machine learning. | [
"stat.ML",
"cs.HC",
"cs.LG",
"cs.SI"
] |
Generative Adversarial Networks (GANs) can achieve state-of-the-art sample
quality in generative modelling tasks but suffer from the mode collapse
problem. Variational Autoencoders (VAE) on the other hand explicitly maximize a
reconstruction-based data log-likelihood forcing it to cover all modes, but
suffer from poorer sample quality. Recent works have proposed hybrid VAE-GAN
frameworks which integrate a GAN-based synthetic likelihood to the VAE
objective to address both the mode collapse and sample quality issues, with
limited success. This is because the VAE objective forces a trade-off between
the data log-likelihood and divergence to the latent prior. The synthetic
likelihood ratio term also shows instability during training. We propose a
novel objective with a "Best-of-Many-Samples" reconstruction cost and a stable
direct estimate of the synthetic likelihood. This enables our hybrid VAE-GAN
framework to achieve high data log-likelihood and low divergence to the latent
prior at the same time and shows significant improvement over both hybrid
VAE-GANS and plain GANs in mode coverage and quality. | [
"cs.LG",
"stat.ML"
] |
Advances in neural network based classifiers have transformed automatic
feature learning from a pipe dream of stronger AI to a routine and expected
property of practical systems. Since the emergence of AlexNet every winning
submission of the ImageNet challenge has employed end-to-end representation
learning, and due to the utility of good representations for transfer learning,
representation learning has become as an important and distinct task from
supervised learning. At present, this distinction is inconsequential, as
supervised methods are state-of-the-art in learning transferable
representations. But recent work has shown that generative models can also be
powerful agents of representation learning. Will the representations learned
from these generative methods ever rival the quality of those from their
supervised competitors? In this work, we argue in the affirmative, that from an
information theoretic perspective, generative models have greater potential for
representation learning. Based on several experimentally validated assumptions,
we show that supervised learning is upper bounded in its capacity for
representation learning in ways that certain generative models, such as
Generative Adversarial Networks (GANs) are not. We hope that our analysis will
provide a rigorous motivation for further exploration of generative
representation learning. | [
"cs.LG",
"cs.AI",
"stat.ML"
] |
Extracting the instantaneous heart rate (iHR) from face videos has been well
studied in recent years. It is well known that changes in skin color due to
blood flow can be captured using conventional cameras. One of the main
limitations of methods that rely on this principle is the need of an
illumination source. Moreover, they have to be able to operate under different
light conditions. One way to avoid these constraints is using infrared cameras,
allowing the monitoring of iHR under low light conditions. In this work, we
present a simple, principled signal extraction method that recovers the iHR
from infrared face videos. We tested the procedure on 7 participants, for whom
we recorded an electrocardiogram simultaneously with their infrared face video.
We checked that the recovered signal matched the ground truth iHR, showing that
infrared is a promising alternative to conventional video imaging for heart
rate monitoring, especially in low light conditions. Code is available at
https://github.com/natalialmg/IR_iHR | [
"cs.CV"
] |
In this paper, we explore the task of generating photo-realistic face images
from lines. Previous methods based on conditional generative adversarial
networks (cGANs) have shown their power to generate visually plausible images
when a conditional image and an output image share well-aligned structures.
However, these models fail to synthesize face images with a whole set of
well-defined structures, e.g. eyes, noses, mouths, etc., especially when the
conditional line map lacks one or several parts. To address this problem, we
propose a conditional self-attention generative adversarial network (CSAGAN).
We introduce a conditional self-attention mechanism to cGANs to capture
long-range dependencies between different regions in faces. We also build a
multi-scale discriminator. The large-scale discriminator enforces the
completeness of global structures and the small-scale discriminator encourages
fine details, thereby enhancing the realism of generated face images. We
evaluate the proposed model on the CelebA-HD dataset by two perceptual user
studies and three quantitative metrics. The experiment results demonstrate that
our method generates high-quality facial images while preserving facial
structures. Our results outperform state-of-the-art methods both quantitatively
and qualitatively. | [
"cs.CV",
"eess.IV"
] |
It has been arduous to assess the progress of a policy learning algorithm in
the domain of hierarchical task with high dimensional action space due to the
lack of a commonly accepted benchmark. In this work, we propose a new
light-weight benchmark task called Diner Dash for evaluating the performance in
a complicated task with high dimensional action space. In contrast to the
traditional Atari games that only have a flat structure of goals and very few
actions, the proposed benchmark task has a hierarchical task structure and size
of 57 for the action space and hence can facilitate the development of policy
learning in complicated tasks. On top of that, we introduce Decomposed Policy
Graph Modelling (DPGM), an algorithm that combines both graph modelling and
deep learning to allow explicit domain knowledge embedding and achieves
significant improvement comparing to the baseline. In the experiments, we have
shown the effectiveness of the domain knowledge injection via a specially
designed imitation algorithm as well as results of other popular algorithms. | [
"cs.LG",
"cs.AI",
"stat.ML"
] |
We study the effect of the stochastic gradient noise on the training of
generative adversarial networks (GANs) and show that it can prevent the
convergence of standard game optimization methods, while the batch version
converges. We address this issue with a novel stochastic variance-reduced
extragradient (SVRE) optimization algorithm, which for a large class of games
improves upon the previous convergence rates proposed in the literature. We
observe empirically that SVRE performs similarly to a batch method on MNIST
while being computationally cheaper, and that SVRE yields more stable GAN
training on standard datasets. | [
"stat.ML",
"cs.LG",
"math.OC"
] |
Kronecker Products (KP) have been used to compress IoT RNN Applications by
15-38x compression factors, achieving better results than traditional
compression methods. However when KP is applied to large Natural Language
Processing tasks, it leads to significant accuracy loss (approx 26%). This
paper proposes a way to recover accuracy otherwise lost when applying KP to
large NLP tasks, by allowing additional degrees of freedom in the KP matrix.
More formally, we propose doping, a process of adding an extremely sparse
overlay matrix on top of the pre-defined KP structure. We call this compression
method doped kronecker product compression. To train these models, we present a
new solution to the phenomenon of co-matrix adaption (CMA), which uses a new
regularization scheme called co matrix dropout regularization (CMR). We present
experimental results that demonstrate compression of a large language model
with LSTM layers of size 25 MB by 25x with 1.4% loss in perplexity score. At
25x compression, an equivalent pruned network leads to 7.9% loss in perplexity
score, while HMD and LMF lead to 15% and 27% loss in perplexity score
respectively. | [
"cs.LG",
"cs.CL",
"stat.ML"
] |
Knowledge graph reasoning is a critical task in natural language processing.
The task becomes more challenging on temporal knowledge graphs, where each fact
is associated with a timestamp. Most existing methods focus on reasoning at
past timestamps and they are not able to predict facts happening in the future.
This paper proposes Recurrent Event Network (RE-NET), a novel autoregressive
architecture for predicting future interactions. The occurrence of a fact
(event) is modeled as a probability distribution conditioned on temporal
sequences of past knowledge graphs. Specifically, our RE-NET employs a
recurrent event encoder to encode past facts and uses a neighborhood aggregator
to model the connection of facts at the same timestamp. Future facts can then
be inferred in a sequential manner based on the two modules. We evaluate our
proposed method via link prediction at future times on five public datasets.
Through extensive experiments, we demonstrate the strength of RENET, especially
on multi-step inference over future timestamps, and achieve state-of-the-art
performance on all five datasets. Code and data can be found at
https://github.com/INK-USC/RE-Net. | [
"cs.LG",
"cs.AI",
"cs.CL",
"stat.ML"
] |
In unsupervised learning, dimensionality reduction is an important tool for
data exploration and visualization. Because these aims are typically
open-ended, it can be useful to frame the problem as looking for patterns that
are enriched in one dataset relative to another. These pairs of datasets occur
commonly, for instance a population of interest vs. control or signal vs.
signal free recordings.However, there are few methods that work on sets of data
as opposed to data points or sequences. Here, we present a probabilistic model
for dimensionality reduction to discover signal that is enriched in the target
dataset relative to the background dataset. The data in these sets do not need
to be paired or grouped beyond set membership. By using a probabilistic model
where some structure is shared amongst the two datasets and some is unique to
the target dataset, we are able to recover interesting structure in the latent
space of the target dataset. The method also has the advantages of a
probabilistic model, namely that it allows for the incorporation of prior
information, handles missing data, and can be generalized to different
distributional assumptions. We describe several possible variations of the
model and demonstrate the application of the technique to de-noising, feature
selection, and subgroup discovery settings. | [
"stat.ML",
"cs.LG"
] |
We propose a hybrid recurrent Video Colorization with Hybrid Generative
Adversarial Network (VCGAN), an improved approach to video colorization using
end-to-end learning. The VCGAN addresses two prevalent issues in the video
colorization domain: Temporal consistency and unification of colorization
network and refinement network into a single architecture. To enhance
colorization quality and spatiotemporal consistency, the mainstream of
generator in VCGAN is assisted by two additional networks, i.e., global feature
extractor and placeholder feature extractor, respectively. The global feature
extractor encodes the global semantics of grayscale input to enhance
colorization quality, whereas the placeholder feature extractor acts as a
feedback connection to encode the semantics of the previous colorized frame in
order to maintain spatiotemporal consistency. If changing the input for
placeholder feature extractor as grayscale input, the hybrid VCGAN also has the
potential to perform image colorization. To improve the consistency of far
frames, we propose a dense long-term loss that smooths the temporal disparity
of every two remote frames. Trained with colorization and temporal losses
jointly, VCGAN strikes a good balance between color vividness and video
continuity. Experimental results demonstrate that VCGAN produces higher-quality
and temporally more consistent colorful videos than existing approaches. | [
"cs.CV",
"cs.MM"
] |
This work presents a practical solution to the problem of call center agent
malpractice. A semi-supervised framework comprising of non-linear power
transformation, neural feature learning and k-means clustering is outlined. We
put these building blocks together and tune the parameters so that the best
performance was obtained. The data used in the experiments is obtained from our
in-house call center. It is made up of recorded agent-customer conversations
which have been annotated using a convolutional neural network based segmenter.
The methods provided a means of tuning the parameters of the neural network to
achieve a desirable result. We show that, using our proposed framework, it is
possible to significantly reduce the malpractice classification error of a
k-means-only clustering model which would serve the same purpose. Additionally,
by presenting the amount of silence per call as a key performance indicator, we
show that the proposed system has enhanced agents performance at our call
center since deployment. | [
"cs.LG",
"cs.SD",
"eess.AS"
] |
We present a novel method for synthesizing both temporally and geometrically
consistent street-view panoramic video from a single satellite image and camera
trajectory. Existing cross-view synthesis approaches focus on images, while
video synthesis in such a case has not yet received enough attention. For
geometrical and temporal consistency, our approach explicitly creates a 3D
point cloud representation of the scene and maintains dense 3D-2D
correspondences across frames that reflect the geometric scene configuration
inferred from the satellite view. As for synthesis in the 3D space, we
implement a cascaded network architecture with two hourglass modules to
generate point-wise coarse and fine features from semantics and per-class
latent vectors, followed by projection to frames and an upsampling module to
obtain the final realistic video. By leveraging computed correspondences, the
produced street-view video frames adhere to the 3D geometric scene structure
and maintain temporal consistency. Qualitative and quantitative experiments
demonstrate superior results compared to other state-of-the-art synthesis
approaches that either lack temporal consistency or realistic appearance. To
the best of our knowledge, our work is the first one to synthesize cross-view
images to video. | [
"cs.CV"
] |
Optical flow estimation is one of the most studied problems in computer
vision, yet recent benchmark datasets continue to reveal problem areas of
today's approaches. Occlusions have remained one of the key challenges. In this
paper, we propose a symmetric optical flow method to address the well-known
chicken-and-egg relation between optical flow and occlusions. In contrast to
many state-of-the-art methods that consider occlusions as outliers, possibly
filtered out during post-processing, we highlight the importance of joint
occlusion reasoning in the optimization and show how to utilize occlusion as an
important cue for estimating optical flow. The key feature of our model is to
fully exploit the symmetry properties that characterize optical flow and
occlusions in the two consecutive images. Specifically through utilizing
forward-backward consistency and occlusion-disocclusion symmetry in the energy,
our model jointly estimates optical flow in both forward and backward
direction, as well as consistent occlusion maps in both views. We demonstrate
significant performance benefits on standard benchmarks, especially from the
occlusion-disocclusion symmetry. On the challenging KITTI dataset we report the
most accurate two-frame results to date. | [
"cs.CV"
] |
Multishot Magnetic Resonance Imaging (MRI) is a promising imaging modality
that can produce a high-resolution image with relatively less data acquisition
time. The downside of multishot MRI is that it is very sensitive to subject
motion and even small amounts of motion during the scan can produce artifacts
in the final MR image that may cause misdiagnosis. Numerous efforts have been
made to address this issue; however, all of these proposals are limited in
terms of how much motion they can correct and the required computational time.
In this paper, we propose a novel generative networks based conjugate gradient
SENSE (CG-SENSE) reconstruction framework for motion correction in multishot
MRI. The proposed framework first employs CG-SENSE reconstruction to produce
the motion-corrupted image and then a generative adversarial network (GAN) is
used to correct the motion artifacts. The proposed method has been rigorously
evaluated on synthetically corrupted data on varying degrees of motion, numbers
of shots, and encoding trajectories. Our analyses (both quantitative as well as
qualitative/visual analysis) establishes that the proposed method significantly
robust and outperforms state-of-the-art motion correction techniques and also
reduces severalfold of computational times. | [
"cs.CV"
] |
The rapid advancement in the field of deep learning and high performance
computing has highly augmented the scope of video based vehicle counting
system. In this paper, the authors deploy several state of the art object
detection and tracking algorithms to detect and track different classes of
vehicles in their regions of interest (ROI). The goal of correctly detecting
and tracking vehicles' in their ROI is to obtain an accurate vehicle count.
Multiple combinations of object detection models coupled with different
tracking systems are applied to access the best vehicle counting framework. The
models' addresses challenges associated to different weather conditions,
occlusion and low-light settings and efficiently extracts vehicle information
and trajectories through its computationally rich training and feedback cycles.
The automatic vehicle counts resulting from all the model combinations are
validated and compared against the manually counted ground truths of over 9
hours' traffic video data obtained from the Louisiana Department of
Transportation and Development. Experimental results demonstrate that the
combination of CenterNet and Deep SORT, Detectron2 and Deep SORT, and YOLOv4
and Deep SORT produced the best overall counting percentage for all vehicles. | [
"cs.CV"
] |
In many multi-agent spatiotemporal systems, the agents are under the
influence of shared, unobserved variables (e.g., the play a team is executing
in a game of basketball). As a result, the trajectories of the agents are often
statistically dependent at any given time step; however, almost universally,
multi-agent models implicitly assume the agents' trajectories are statistically
independent at each time step. In this paper, we introduce baller2vec++, a
multi-entity Transformer that can effectively model coordinated agents.
Specifically, baller2vec++ applies a specially designed self-attention mask to
a mixture of location and "look-ahead" trajectory sequences to learn the
distributions of statistically dependent agent trajectories. We show that,
unlike baller2vec (baller2vec++'s predecessor), baller2vec++ can learn to
emulate the behavior of perfectly coordinated agents in a simulated toy
dataset. Additionally, when modeling the trajectories of professional
basketball players, baller2vec++ outperforms baller2vec by a wide margin. | [
"cs.LG",
"cs.MA"
] |
Recent advancements in Convolutional Neural Networks have yielded super-human
levels of performance in image recognition tasks [13, 25]; however, with
increasing volumes of parcels crossing UK borders each year, classification of
threats becomes integral to the smooth operation of UK borders. In this work we
propose the first pipeline to effectively process Dual-Energy X-Ray scanner
output, and perform classification capable of distinguishing between firearm
families (Assault Rifle, Revolver, Self-Loading Pistol,Shotgun, and Sub-Machine
Gun) from this output. With this pipeline we compare re-cent Convolutional
Neural Network architectures against the X-Ray baggage domain via Transfer
Learning and show ResNet50 to be most suitable to classification - outlining a
number of considerations for operational success within the domain. | [
"cs.CV"
] |
End-to-end learning of communications systems is a fascinating novel concept
that has so far only been validated by simulations for block-based
transmissions. It allows learning of transmitter and receiver implementations
as deep neural networks (NNs) that are optimized for an arbitrary
differentiable end-to-end performance metric, e.g., block error rate (BLER). In
this paper, we demonstrate that over-the-air transmissions are possible: We
build, train, and run a complete communications system solely composed of NNs
using unsynchronized off-the-shelf software-defined radios (SDRs) and
open-source deep learning (DL) software libraries. We extend the existing ideas
towards continuous data transmission which eases their current restriction to
short block lengths but also entails the issue of receiver synchronization. We
overcome this problem by introducing a frame synchronization module based on
another NN. A comparison of the BLER performance of the "learned" system with
that of a practical baseline shows competitive performance close to 1 dB, even
without extensive hyperparameter tuning. We identify several practical
challenges of training such a system over actual channels, in particular the
missing channel gradient, and propose a two-step learning procedure based on
the idea of transfer learning that circumvents this issue. | [
"stat.ML",
"cs.IT",
"math.IT"
] |
Many classic methods have shown non-local self-similarity in natural images
to be an effective prior for image restoration. However, it remains unclear and
challenging to make use of this intrinsic property via deep networks. In this
paper, we propose a non-local recurrent network (NLRN) as the first attempt to
incorporate non-local operations into a recurrent neural network (RNN) for
image restoration. The main contributions of this work are: (1) Unlike existing
methods that measure self-similarity in an isolated manner, the proposed
non-local module can be flexibly integrated into existing deep networks for
end-to-end training to capture deep feature correlation between each location
and its neighborhood. (2) We fully employ the RNN structure for its parameter
efficiency and allow deep feature correlation to be propagated along adjacent
recurrent states. This new design boosts robustness against inaccurate
correlation estimation due to severely degraded images. (3) We show that it is
essential to maintain a confined neighborhood for computing deep feature
correlation given degraded images. This is in contrast to existing practice
that deploys the whole image. Extensive experiments on both image denoising and
super-resolution tasks are conducted. Thanks to the recurrent non-local
operations and correlation propagation, the proposed NLRN achieves superior
results to state-of-the-art methods with much fewer parameters. | [
"cs.CV"
] |
Neural architecture search has attracted wide attentions in both academia and
industry. To accelerate it, researchers proposed weight-sharing methods which
first train a super-network to reuse computation among different operators,
from which exponentially many sub-networks can be sampled and efficiently
evaluated. These methods enjoy great advantages in terms of computational
costs, but the sampled sub-networks are not guaranteed to be estimated
precisely unless an individual training process is taken. This paper owes such
inaccuracy to the inevitable mismatch between assembled network layers, so that
there is a random error term added to each estimation. We alleviate this issue
by training a graph convolutional network to fit the performance of sampled
sub-networks so that the impact of random errors becomes minimal. With this
strategy, we achieve a higher rank correlation coefficient in the selected set
of candidates, which consequently leads to better performance of the final
architecture. In addition, our approach also enjoys the flexibility of being
used under different hardware constraints, since the graph convolutional
network has provided an efficient lookup table of the performance of
architectures in the entire search space. | [
"cs.LG",
"cs.CV",
"stat.ML"
] |
Representation learning over graph structured data has been mostly studied in
static graph settings while efforts for modeling dynamic graphs are still
scant. In this paper, we develop a novel hierarchical variational model that
introduces additional latent random variables to jointly model the hidden
states of a graph recurrent neural network (GRNN) to capture both topology and
node attribute changes in dynamic graphs. We argue that the use of high-level
latent random variables in this variational GRNN (VGRNN) can better capture
potential variability observed in dynamic graphs as well as the uncertainty of
node latent representation. With semi-implicit variational inference developed
for this new VGRNN architecture (SI-VGRNN), we show that flexible non-Gaussian
latent representations can further help dynamic graph analytic tasks. Our
experiments with multiple real-world dynamic graph datasets demonstrate that
SI-VGRNN and VGRNN consistently outperform the existing baseline and
state-of-the-art methods by a significant margin in dynamic link prediction. | [
"cs.LG",
"stat.ML"
] |
This work proposes a novel attentive graph neural network (AGNN) for
zero-shot video object segmentation (ZVOS). The suggested AGNN recasts this
task as a process of iterative information fusion over video graphs.
Specifically, AGNN builds a fully connected graph to efficiently represent
frames as nodes, and relations between arbitrary frame pairs as edges. The
underlying pair-wise relations are described by a differentiable attention
mechanism. Through parametric message passing, AGNN is able to efficiently
capture and mine much richer and higher-order relations between video frames,
thus enabling a more complete understanding of video content and more accurate
foreground estimation. Experimental results on three video segmentation
datasets show that AGNN sets a new state-of-the-art in each case. To further
demonstrate the generalizability of our framework, we extend AGNN to an
additional task: image object co-segmentation (IOCS). We perform experiments on
two famous IOCS datasets and observe again the superiority of our AGNN model.
The extensive experiments verify that AGNN is able to learn the underlying
semantic/appearance relationships among video frames or related images, and
discover the common objects. | [
"cs.CV"
] |
Predicting molecular conformations (or 3D structures) from molecular graphs
is a fundamental problem in many applications. Most existing approaches are
usually divided into two steps by first predicting the distances between atoms
and then generating a 3D structure through optimizing a distance geometry
problem. However, the distances predicted with such two-stage approaches may
not be able to consistently preserve the geometry of local atomic
neighborhoods, making the generated structures unsatisfying. In this paper, we
propose an end-to-end solution for molecular conformation prediction called
ConfVAE based on the conditional variational autoencoder framework.
Specifically, the molecular graph is first encoded in a latent space, and then
the 3D structures are generated by solving a principled bilevel optimization
program. Extensive experiments on several benchmark data sets prove the
effectiveness of our proposed approach over existing state-of-the-art
approaches. Code is available at
\url{https://github.com/MinkaiXu/ConfVAE-ICML21}. | [
"cs.LG",
"q-bio.BM"
] |
Camera pose regression methods apply a single forward pass to the query image
to estimate the camera pose. As such, they offer a fast and light-weight
alternative to traditional localization schemes based on image retrieval. Pose
regression approaches simultaneously learn two regression tasks, aiming to
jointly estimate the camera position and orientation using a single embedding
vector computed by a convolutional backbone. We propose an attention-based
approach for pose regression, where the convolutional activation maps are used
as sequential inputs. Transformers are applied to encode the sequential
activation maps as latent vectors, used for camera pose regression. This allows
us to pay attention to spatially-varying deep features. Using two Transformer
heads, we separately focus on the features for camera position and orientation,
based on how informative they are per task. Our proposed approach is shown to
compare favorably to contemporary pose regressors schemes and achieves
state-of-the-art accuracy across multiple outdoor and indoor benchmarks. In
particular, to the best of our knowledge, our approach is the only method to
attain sub-meter average accuracy across outdoor scenes. We make our code
publicly available from here. | [
"cs.CV",
"cs.AI"
] |
In this paper, we propose a flexible notion of characteristic functions
defined on graph vertices to describe the distribution of vertex features at
multiple scales. We introduce FEATHER, a computationally efficient algorithm to
calculate a specific variant of these characteristic functions where the
probability weights of the characteristic function are defined as the
transition probabilities of random walks. We argue that features extracted by
this procedure are useful for node level machine learning tasks. We discuss the
pooling of these node representations, resulting in compact descriptors of
graphs that can serve as features for graph classification algorithms. We
analytically prove that FEATHER describes isomorphic graphs with the same
representation and exhibits robustness to data corruption. Using the node
feature characteristic functions we define parametric models where evaluation
points of the functions are learned parameters of supervised classifiers.
Experiments on real world large datasets show that our proposed algorithm
creates high quality representations, performs transfer learning efficiently,
exhibits robustness to hyperparameter changes, and scales linearly with the
input size. | [
"cs.LG",
"cs.DM",
"cs.SI",
"stat.ML"
] |
This paper explores the similarities of output layers in Neural Networks
(NNs) with logistic regression to explain importance of inputs by Z-scores. The
network analyzed, a network for fusion of Synthetic Aperture Radar (SAR) and
Microwave Radiometry (MWR) data, is applied to prediction of arctic sea ice.
With the analysis the importance of MWR relative to SAR is found to favor MWR
components. Further, as the model represents image features at different
scales, the relative importance of these are as well analyzed. The suggested
methodology offers a simple and easy framework for analyzing output layer
components and can reduce the number of components for further analysis with
e.g. common NN visualization methods. | [
"cs.CV"
] |
Interpretability in Graph Convolutional Networks (GCNs) has been explored to
some extent in computer vision in general, yet, in the medical domain, it
requires further examination. Moreover, most of the interpretability approaches
for GCNs, especially in the medical domain, focus on interpreting the model in
a post hoc fashion. In this paper, we propose an interpretable graph
learning-based model which 1) interprets the clinical relevance of the input
features towards the task, 2) uses the explanation to improve the model
performance and, 3) learns a population level latent graph that may be used to
interpret the cohort's behavior. In a clinical scenario, such a model can
assist the clinical experts in better decision-making for diagnosis and
treatment planning. The main novelty lies in the interpretable attention module
(IAM), which directly operates on multi-modal features. Our IAM learns the
attention for each feature based on the unique interpretability-specific
losses. We show the application on two publicly available datasets, Tadpole and
UKBB, for three tasks of disease, age, and gender prediction. Our proposed
model shows superior performance with respect to compared methods with an
increase in an average accuracy of 3.2% for Tadpole, 1.6% for UKBB Gender, and
2% for the UKBB Age prediction task. Further, we show exhaustive validation and
clinical interpretation of our results. | [
"cs.CV",
"cs.LG"
] |
Understanding the relationships between biomedical terms like viruses, drugs,
and symptoms is essential in the fight against diseases. Many attempts have
been made to introduce the use of machine learning to the scientific process of
hypothesis generation(HG), which refers to the discovery of meaningful implicit
connections between biomedical terms. However, most existing methods fail to
truly capture the temporal dynamics of scientific term relations and also
assume unobserved connections to be irrelevant (i.e., in a positive-negative
(PN) learning setting). To break these limits, we formulate this HG problem as
future connectivity prediction task on a dynamic attributed graph via
positive-unlabeled (PU) learning. Then, the key is to capture the temporal
evolution of node pair (term pair) relations from just the positive and
unlabeled data. We propose a variational inference model to estimate the
positive prior, and incorporate it in the learning of node pair embeddings,
which are then used for link prediction. Experiment results on real-world
biomedical term relationship datasets and case study analyses on a COVID-19
dataset validate the effectiveness of the proposed model. | [
"cs.LG",
"stat.ML"
] |
Intrinsic isometric shape matching has become the standard approach for pose
invariant correspondence estimation among deformable shapes. Most existing
approaches assume global consistency, i.e., the metric structure of the whole
manifold must not change significantly. While global isometric matching is well
understood, only a few heuristic solutions are known for partial matching.
Partial matching is particularly important for robustness to topological noise
(incomplete data and contacts), which is a common problem in real-world 3D
scanner data. In this paper, we introduce a new approach to partial, intrinsic
isometric matching. Our method is based on the observation that isometries are
fully determined by purely local information: a map of a single point and its
tangent space fixes an isometry for both global and the partial maps. From this
idea, we develop a new representation for partial isometric maps based on
equivalence classes of correspondences between pairs of points and their
tangent spaces. From this, we derive a local propagation algorithm that find
such mappings efficiently. In contrast to previous heuristics based on RANSAC
or expectation maximization, our method is based on a simple and sound
theoretical model and fully deterministic. We apply our approach to register
partial point clouds and compare it to the state-of-the-art methods, where we
obtain significant improvements over global methods for real-world data and
stronger guarantees than previous heuristic partial matching algorithms. | [
"cs.CV",
"cs.GR"
] |
Graph Convolution Networks (GCN) are widely used in learning graph
representations due to their effectiveness and efficiency. However, they suffer
from the notorious over-smoothing problem, in which the learned representations
of densely connected nodes converge to alike vectors when many (>3) graph
convolutional layers are stacked. In this paper, we argue that
there-normalization trick used in GCN leads to overly homogeneous information
propagation, which is the source of over-smoothing. To address this problem, we
propose Graph Highway Networks(GHNet) which utilize gating units to
automatically balance the trade-off between homogeneity and heterogeneity in
the GCN learning process. The gating units serve as direct highways to maintain
heterogeneous information from the node itself after feature propagation. This
design enables GHNet to achieve much larger receptive fields per node without
over-smoothing and thus access to more of the graph connectivity information.
Experimental results on benchmark datasets demonstrate the superior performance
of GHNet over GCN and related models. | [
"cs.LG",
"stat.ML"
] |
Robust training methods against perturbations to the input data have received
great attention in the machine learning literature. A standard approach in this
direction is adversarial training which learns a model using
adversarially-perturbed training samples. However, adversarial training
performs suboptimally against perturbations structured across samples such as
universal and group-sparse shifts that are commonly present in biological data
such as gene expression levels of different tissues. In this work, we seek to
close this optimality gap and introduce Group-Structured Adversarial Training
(GSAT) which learns a model robust to perturbations structured across samples.
We formulate GSAT as a non-convex concave minimax optimization problem which
minimizes a group-structured optimal transport cost. Specifically, we focus on
the applications of GSAT for group-sparse and rank-constrained perturbations
modeled using group and nuclear norm penalties. In order to solve GSAT's
non-smooth optimization problem in those cases, we propose a new minimax
optimization algorithm called GDADMM by combining Gradient Descent Ascent (GDA)
and Alternating Direction Method of Multipliers (ADMM). We present several
applications of the GSAT framework to gain robustness against structured
perturbations for image recognition and computational biology datasets. | [
"cs.LG",
"stat.ML"
] |
Predicting human perceptual similarity is a challenging subject of ongoing
research. The visual process underlying this aspect of human vision is thought
to employ multiple different levels of visual analysis (shapes, objects,
texture, layout, color, etc). In this paper, we postulate that the perception
of image similarity is not an explicitly learned capability, but rather one
that is a byproduct of learning others. This claim is supported by leveraging
representations learned from a diverse set of visual tasks and using them
jointly to predict perceptual similarity. This is done via simple feature
concatenation, without any further learning. Nevertheless, experiments
performed on the challenging Totally-Looks-Like (TLL) benchmark significantly
surpass recent baselines, closing much of the reported gap towards prediction
of human perceptual similarity. We provide an analysis of these results and
discuss them in a broader context of emergent visual capabilities and their
implications on the course of machine-vision research. | [
"cs.CV",
"cs.AI",
"cs.LG"
] |
Graph-level representations are critical in various real-world applications,
such as predicting the properties of molecules. But in practice, precise graph
annotations are generally very expensive and time-consuming. To address this
issue, graph contrastive learning constructs instance discrimination task which
pulls together positive pairs (augmentation pairs of the same graph) and pushes
away negative pairs (augmentation pairs of different graphs) for unsupervised
representation learning. However, since for a query, its negatives are
uniformly sampled from all graphs, existing methods suffer from the critical
sampling bias issue, i.e., the negatives likely having the same semantic
structure with the query, leading to performance degradation. To mitigate this
sampling bias issue, in this paper, we propose a Prototypical Graph Contrastive
Learning (PGCL) approach. Specifically, PGCL models the underlying semantic
structure of the graph data via clustering semantically similar graphs into the
same group, and simultaneously encourages the clustering consistency for
different augmentations of the same graph. Then given a query, it performs
negative sampling via drawing the graphs from those clusters that differ from
the cluster of query, which ensures the semantic difference between query and
its negative samples. Moreover, for a query, PGCL further reweights its
negative samples based on the distance between their prototypes (cluster
centroids) and the query prototype such that those negatives having moderate
prototype distance enjoy relatively large weights. This reweighting strategy is
proved to be more effective than uniform sampling. Experimental results on
various graph benchmarks testify the advantages of our PGCL over
state-of-the-art methods. | [
"cs.LG",
"cs.AI"
] |
We build on the recently proposed EigenGame that views eigendecomposition as
a competitive game. EigenGame's updates are biased if computed using
minibatches of data, which hinders convergence and more sophisticated
parallelism in the stochastic setting. In this work, we propose an unbiased
stochastic update that is asymptotically equivalent to EigenGame, enjoys
greater parallelism allowing computation on datasets of larger sample sizes,
and outperforms EigenGame in experiments. We present applications to finding
the principal components of massive datasets and performing spectral clustering
of graphs. We analyze and discuss our proposed update in the context of
EigenGame and the shift in perspective from optimization to games. | [
"stat.ML",
"cs.AI",
"cs.LG"
] |
In order to keep track of the operational state of power grid, the world's
largest sensor systems, smart grid, was built by deploying hundreds of millions
of smart meters. Such system makes it possible to discover and make quick
response to any hidden threat to the entire power grid. Non-technical losses
(NTLs) have always been a major concern for its consequent security risks as
well as immeasurable revenue loss. However, various causes of NTL may have
different characteristics reflected in the data. Accurately capturing these
anomalies faced with such large scale of collected data records is rather
tricky as a result. In this paper, we proposed a new methodology of detecting
abnormal electricity consumptions. We did a transformation of the collected
time-series data which turns it into an image representation that could well
reflect users' relatively long term consumption behaviors. Inspired by the
excellent neural network architecture used for objective detection in computer
vision domain, we designed our deep learning model that takes the transformed
images as input and yields joint featured inferred from the multiple aspects
the input provides. Considering the limited labeled samples, especially the
abnormal ones, we used our model in a semi-supervised fashion that is brought
out in recent years. The model is tested on samples which are verified by
on-field inspections and our method showed significant improvement. | [
"cs.LG",
"stat.ML"
] |
Convolutional Neural Networks have been highly successful in performing a
host of computer vision tasks such as object recognition, object detection,
image segmentation and texture synthesis. In 2015, Gatys et. al [7] show how
the style of a painter can be extracted from an image of the painting and
applied to another normal photograph, thus recreating the photo in the style of
the painter. The method has been successfully applied to a wide range of images
and has since spawned multiple applications and mobile apps. In this paper, the
neural style transfer algorithm is applied to fashion so as to synthesize new
custom clothes. We construct an approach to personalize and generate new custom
clothes based on a users preference and by learning the users fashion choices
from a limited set of clothes from their closet. The approach is evaluated by
analyzing the generated images of clothes and how well they align with the
users fashion style. | [
"cs.CV",
"cs.AI",
"cs.IR",
"cs.NE"
] |
When learning policies for real-world domains, two important questions arise:
(i) how to efficiently use pre-collected off-policy, non-optimal behavior data;
and (ii) how to mediate among different competing objectives and constraints.
We thus study the problem of batch policy learning under multiple constraints,
and offer a systematic solution. We first propose a flexible meta-algorithm
that admits any batch reinforcement learning and online learning procedure as
subroutines. We then present a specific algorithmic instantiation and provide
performance guarantees for the main objective and all constraints. To certify
constraint satisfaction, we propose a new and simple method for off-policy
policy evaluation (OPE) and derive PAC-style bounds. Our algorithm achieves
strong empirical results in different domains, including in a challenging
problem of simulated car driving subject to multiple constraints such as lane
keeping and smooth driving. We also show experimentally that our OPE method
outperforms other popular OPE techniques on a standalone basis, especially in a
high-dimensional setting. | [
"cs.LG",
"cs.AI",
"math.OC",
"stat.ML"
] |
In this paper we consider the problem of multi-view face detection. While
there has been significant research on this problem, current state-of-the-art
approaches for this task require annotation of facial landmarks, e.g. TSM [25],
or annotation of face poses [28, 22]. They also require training dozens of
models to fully capture faces in all orientations, e.g. 22 models in HeadHunter
method [22]. In this paper we propose Deep Dense Face Detector (DDFD), a method
that does not require pose/landmark annotation and is able to detect faces in a
wide range of orientations using a single model based on deep convolutional
neural networks. The proposed method has minimal complexity; unlike other
recent deep learning object detection methods [9], it does not require
additional components such as segmentation, bounding-box regression, or SVM
classifiers. Furthermore, we analyzed scores of the proposed face detector for
faces in different orientations and found that 1) the proposed method is able
to detect faces from different angles and can handle occlusion to some extent,
2) there seems to be a correlation between dis- tribution of positive examples
in the training set and scores of the proposed face detector. The latter
suggests that the proposed methods performance can be further improved by using
better sampling strategies and more sophisticated data augmentation techniques.
Evaluations on popular face detection benchmark datasets show that our
single-model face detector algorithm has similar or better performance compared
to the previous methods, which are more complex and require annotations of
either different poses or facial landmarks. | [
"cs.CV",
"I.4"
] |
Recently, adversarial erasing for weakly-supervised object attention has been
deeply studied due to its capability in localizing integral object regions.
However, such a strategy raises one key problem that attention regions will
gradually expand to non-object regions as training iterations continue, which
significantly decreases the quality of the produced attention maps. To tackle
such an issue as well as promote the quality of object attention, we introduce
a simple yet effective Self-Erasing Network (SeeNet) to prohibit attentions
from spreading to unexpected background regions. In particular, SeeNet
leverages two self-erasing strategies to encourage networks to use reliable
object and background cues for learning to attention. In this way, integral
object regions can be effectively highlighted without including much more
background regions. To test the quality of the generated attention maps, we
employ the mined object regions as heuristic cues for learning semantic
segmentation models. Experiments on Pascal VOC well demonstrate the superiority
of our SeeNet over other state-of-the-art methods. | [
"cs.CV"
] |
Video question answering has recently received a lot of attention from
multimodal video researchers. Most video question answering datasets are
usually in the form of multiple-choice. But, the model for the multiple-choice
task does not infer the answer. Rather it compares the answer candidates for
picking the correct answer. Furthermore, it makes it difficult to extend to
other tasks. In this paper, we challenge the existing multiple-choice video
question answering by changing it to open-ended video question answering. To
tackle open-ended question answering, we use the pretrained GPT2 model. The
model is fine-tuned with video inputs and subtitles. An ablation study is
performed by changing the existing DramaQA dataset to an open-ended question
answering, and it shows that performance can be improved using video metadata. | [
"cs.CV",
"cs.AI"
] |
Generative adversarial networks (GAN) have recently been shown to be
efficient for speech enhancement. However, most, if not all, existing speech
enhancement GANs (SEGAN) make use of a single generator to perform one-stage
enhancement mapping. In this work, we propose to use multiple generators that
are chained to perform multi-stage enhancement mapping, which gradually refines
the noisy input signals in a stage-wise fashion. Furthermore, we study two
scenarios: (1) the generators share their parameters and (2) the generators'
parameters are independent. The former constrains the generators to learn a
common mapping that is iteratively applied at all enhancement stages and
results in a small model footprint. On the contrary, the latter allows the
generators to flexibly learn different enhancement mappings at different stages
of the network at the cost of an increased model size. We demonstrate that the
proposed multi-stage enhancement approach outperforms the one-stage SEGAN
baseline, where the independent generators lead to more favorable results than
the tied generators. The source code is available at
http://github.com/pquochuy/idsegan. | [
"cs.LG",
"cs.SD",
"eess.AS",
"stat.ML"
] |
The use of Reinforcement Learning in real-world scenarios is strongly limited
by issues of scale. Most RL learning algorithms are unable to deal with
problems composed of hundreds or sometimes even dozens of possible actions, and
therefore cannot be applied to many real-world problems. We consider the RL
problem in the supervised classification framework where the optimal policy is
obtained through a multiclass classifier, the set of classes being the set of
actions of the problem. We introduce error-correcting output codes (ECOCs) in
this setting and propose two new methods for reducing complexity when using
rollouts-based approaches. The first method consists in using an ECOC-based
classifier as the multiclass classifier, reducing the learning complexity from
O(A2) to O(Alog(A)). We then propose a novel method that profits from the
ECOC's coding dictionary to split the initial MDP into O(log(A)) seperate
two-action MDPs. This second method reduces learning complexity even further,
from O(A2) to O(log(A)), thus rendering problems with large action sets
tractable. We finish by experimentally demonstrating the advantages of our
approach on a set of benchmark problems, both in speed and performance. | [
"cs.LG",
"stat.ML",
"68T05"
] |
Prior studies show that the key to face anti-spoofing lies in the subtle
image pattern, termed "spoof trace", e.g., color distortion, 3D mask edge,
Moire pattern, and many others. Designing a generic anti-spoofing model to
estimate those spoof traces can improve not only the generalization of the
spoof detection, but also the interpretability of the model's decision. Yet,
this is a challenging task due to the diversity of spoof types and the lack of
ground truth in spoof traces. This work designs a novel adversarial learning
framework to disentangle the spoof traces from input faces as a hierarchical
combination of patterns at multiple scales. With the disentangled spoof traces,
we unveil the live counterpart of the original spoof face, and further
synthesize realistic new spoof faces after a proper geometric correction. Our
method demonstrates superior spoof detection performance on both seen and
unseen spoof scenarios while providing visually convincing estimation of spoof
traces. Code is available at https://github.com/yaojieliu/ECCV20-STDN. | [
"cs.CV"
] |
We propose a new statistical model for computational linguistics. Rather than
trying to estimate directly the probability distribution of a random sentence
of the language, we define a Markov chain on finite sets of sentences with many
finite recurrent communicating classes and define our language model as the
invariant probability measures of the chain on each recurrent communicating
class. This Markov chain, that we call a communication model, recombines at
each step randomly the set of sentences forming its current state, using some
grammar rules. When the grammar rules are fixed and known in advance instead of
being estimated on the fly, we can prove supplementary mathematical properties.
In particular, we can prove in this case that all states are recurrent states,
so that the chain defines a partition of its state space into finite recurrent
communicating classes. We show that our approach is a decisive departure from
Markov models at the sentence level and discuss its relationships with Context
Free Grammars. Although the toric grammars we use are closely related to
Context Free Grammars, the way we generate the language from the grammar is
qualitatively different. Our communication model has two purposes. On the one
hand, it is used to define indirectly the probability distribution of a random
sentence of the language. On the other hand it can serve as a (crude) model of
language transmission from one speaker to another speaker through the
communication of a (large) set of sentences. | [
"stat.ML",
"cs.CL",
"math.PR",
"62M09, 62P99, 68T50, 91F20, 03B65, 91E40, 60J20"
] |
In this paper, we propose an extraction method of HOG
(histograms-of-oriented-gradients) features from encryption-then-compression
(EtC) images for privacy-preserving machine learning, where EtC images are
images encrypted by a block-based encryption method proposed for EtC systems
with JPEG compression, and HOG is a feature descriptor used in computer vision
for the purpose of object detection and image classification. Recently, cloud
computing and machine learning have been spreading in many fields. However, the
cloud computing has serious privacy issues for end users, due to unreliability
of providers and some accidents. Accordingly, we propose a novel block-based
extraction method of HOG features, and the proposed method enables us to carry
out any machine learning algorithms without any influence, under some
conditions. In an experiment, the proposed method is applied to a face image
recognition problem under the use of two kinds of classifiers: linear support
vector machine (SVM), gaussian SVM, to demonstrate the effectiveness. | [
"cs.CV",
"eess.IV"
] |
Adversarial learning has been successfully embedded into deep networks to
learn transferable features, which reduce distribution discrepancy between the
source and target domains. Existing domain adversarial networks assume fully
shared label space across domains. In the presence of big data, there is strong
motivation of transferring both classification and representation models from
existing big domains to unknown small domains. This paper introduces partial
transfer learning, which relaxes the shared label space assumption to that the
target label space is only a subspace of the source label space. Previous
methods typically match the whole source domain to the target domain, which are
prone to negative transfer for the partial transfer problem. We present
Selective Adversarial Network (SAN), which simultaneously circumvents negative
transfer by selecting out the outlier source classes and promotes positive
transfer by maximally matching the data distributions in the shared label
space. Experiments demonstrate that our models exceed state-of-the-art results
for partial transfer learning tasks on several benchmark datasets. | [
"cs.LG"
] |
In this paper, we propose an approach for filter-level pruning with
hierarchical knowledge distillation based on the teacher, teaching-assistant,
and student framework. Our method makes use of teaching assistants at
intermediate pruning levels that share the same architecture and weights as the
target student. We propose to prune each model independently using the gradient
information from its corresponding teacher. By considering the relative sizes
of each student-teacher pair, this formulation provides a natural trade-off
between the capacity gap for knowledge distillation and the bias of the filter
saliency updates. Our results show improvements in the attainable accuracy and
model compression across the CIFAR10 and ImageNet classification tasks using
the VGG16and ResNet50 architectures. We provide an extensive evaluation that
demonstrates the benefits of using a varying number of teaching assistant
models at different sizes. | [
"cs.CV"
] |
We present a new algorithm for general reinforcement learning where the true
environment is known to belong to a finite class of N arbitrary models. The
algorithm is shown to be near-optimal for all but O(N log^2 N) time-steps with
high probability. Infinite classes are also considered where we show that
compactness is a key criterion for determining the existence of uniform
sample-complexity bounds. A matching lower bound is given for the finite case. | [
"cs.LG"
] |
In this paper, we propose a novel white balance adjustment, called "spatially
varying white balancing," for single, mixed, and non-uniform illuminants. By
using n diagonal matrices along with a weight, the proposed method can reduce
lighting effects on all spatially varying colors in an image under such
illumination conditions. In contrast, conventional white balance adjustments do
not consider the correcting of all colors except under a single illuminant.
Also, multi-color balance adjustments can map multiple colors into
corresponding ground truth colors, although they may cause the rank deficiency
problem to occur as a non-diagonal matrix is used, unlike white balancing. In
an experiment, the effectiveness of the proposed method is shown under mixed
and non-uniform illuminants, compared with conventional white and multi-color
balancing. Moreover, under a single illuminant, the proposed method has almost
the same performance as the conventional white balancing. | [
"cs.CV"
] |
Recent advances in semantic image segmentation have mostly been achieved by
training deep convolutional neural networks (CNNs). We show how to improve
semantic segmentation through the use of contextual information; specifically,
we explore `patch-patch' context between image regions, and `patch-background'
context. For learning from the patch-patch context, we formulate Conditional
Random Fields (CRFs) with CNN-based pairwise potential functions to capture
semantic correlations between neighboring patches. Efficient piecewise training
of the proposed deep structured model is then applied to avoid repeated
expensive CRF inference for back propagation. For capturing the
patch-background context, we show that a network design with traditional
multi-scale image input and sliding pyramid pooling is effective for improving
performance. Our experimental results set new state-of-the-art performance on a
number of popular semantic segmentation datasets, including NYUDv2, PASCAL VOC
2012, PASCAL-Context, and SIFT-flow. In particular, we achieve an
intersection-over-union score of 78.0 on the challenging PASCAL VOC 2012
dataset. | [
"cs.CV"
] |
Point clouds are often sparse and incomplete, which imposes difficulties for
real-world applications. Existing shape completion methods tend to generate
rough shapes without fine-grained details. Considering this, we introduce a
two-branch network for shape completion. The first branch is a cascaded shape
completion sub-network to synthesize complete objects, where we propose to use
the partial input together with the coarse output to preserve the object
details during the dense point reconstruction. The second branch is an
auto-encoder to reconstruct the original partial input. The two branches share
a same feature extractor to learn an accurate global feature for shape
completion. Furthermore, we propose two strategies to enable the training of
our network when ground truth data are not available. This is to mitigate the
dependence of existing approaches on large amounts of ground truth training
data that are often difficult to obtain in real-world applications.
Additionally, our proposed strategies are also able to improve the
reconstruction quality for fully supervised learning. We verify our approach in
self-supervised, semi-supervised and fully supervised settings with superior
performances. Quantitative and qualitative results on different datasets
demonstrate that our method achieves more realistic outputs than
state-of-the-art approaches on the point cloud completion task. | [
"cs.CV",
"cs.AI"
] |
There are time series that are amenable to recurrent neural network (RNN)
solutions when treated as sequences, but some series, e.g. asynchronous time
series, provide a richer variation of feature types than current RNN cells take
into account. In order to address such situations, we introduce a unified RNN
that handles five different feature types, each in a different manner. Our RNN
framework separates sequential features into two groups dependent on their
frequency, which we call sparse and dense features, and which affect cell
updates differently. Further, we also incorporate time features at the
sequential level that relate to the time between specified events in the
sequence and are used to modify the cell's memory state. We also include two
types of static (whole sequence level) features, one related to time and one
not, which are combined with the encoder output. The experiments show that the
modeling framework proposed does increase performance compared to standard
cells. | [
"stat.ML",
"cs.LG"
] |
Single molecule fluorescence microscopy is a powerful technique for
uncovering detailed information about biological systems, both in vitro and in
vivo. In such experiments, the inherently low signal to noise ratios mean that
accurate algorithms to separate true signal and background noise are essential
to generate meaningful results. To this end, we have developed a new and robust
method to reduce noise in single molecule fluorescence images by using a
Gaussian Markov Random Field (GMRF) prior in a Bayesian framework. Two
different strategies are proposed to build the prior - an intrinsic GMRF, with
a stationary relationship between pixels and a heterogeneous intrinsic GMRF,
with a differently weighted relationship between pixels classified as molecules
and background. Testing with synthetic and real experimental fluorescence
images demonstrates that the heterogeneous intrinsic GMRF is superior to other
conventional de-noising approaches. | [
"cs.CV"
] |
We present Deep Graph Infomax (DGI), a general approach for learning node
representations within graph-structured data in an unsupervised manner. DGI
relies on maximizing mutual information between patch representations and
corresponding high-level summaries of graphs---both derived using established
graph convolutional network architectures. The learnt patch representations
summarize subgraphs centered around nodes of interest, and can thus be reused
for downstream node-wise learning tasks. In contrast to most prior approaches
to unsupervised learning with GCNs, DGI does not rely on random walk
objectives, and is readily applicable to both transductive and inductive
learning setups. We demonstrate competitive performance on a variety of node
classification benchmarks, which at times even exceeds the performance of
supervised learning. | [
"stat.ML",
"cs.IT",
"cs.LG",
"cs.SI",
"math.IT"
] |
We present Wiki-CS, a novel dataset derived from Wikipedia for benchmarking
Graph Neural Networks. The dataset consists of nodes corresponding to Computer
Science articles, with edges based on hyperlinks and 10 classes representing
different branches of the field. We use the dataset to evaluate semi-supervised
node classification and single-relation link prediction models. Our experiments
show that these methods perform well on a new domain, with structural
properties different from earlier benchmarks. The dataset is publicly
available, along with the implementation of the data pipeline and the benchmark
experiments, at https://github.com/pmernyei/wiki-cs-dataset . | [
"cs.LG",
"cs.SI",
"stat.ML"
] |
Few-shot learning that trains image classifiers over few labeled examples per
category is a challenging task. In this paper, we propose to exploit an
additional big dataset with different categories to improve the accuracy of
few-shot learning over our target dataset. Our approach is based on the
observation that images can be decomposed into objects, which may appear in
images from both the additional dataset and our target dataset. We use the
object-level relation learned from the additional dataset to infer the
similarity of images in our target dataset with unseen categories. Nearest
neighbor search is applied to do image classification, which is a
non-parametric model and thus does not need fine-tuning. We evaluate our
algorithm on two popular datasets, namely Omniglot and MiniImagenet. We obtain
8.5\% and 2.7\% absolute improvements for 5-way 1-shot and 5-way 5-shot
experiments on MiniImagenet, respectively. Source code will be published upon
acceptance. | [
"cs.CV",
"cs.AI",
"cs.LG"
] |
Human Activity Recognition (HAR) from devices like smartphone accelerometers
is a fundamental problem in ubiquitous computing. Machine learning based
recognition models often perform poorly when applied to new users that were not
part of the training data. Previous work has addressed this challenge by
personalizing general recognition models to the unique motion pattern of a new
user in a static batch setting. They require target user data to be available
upfront. The more challenging online setting has received less attention. No
samples from the target user are available in advance, but they arrive
sequentially. Additionally, the motion pattern of users may change over time.
Thus, adapting to new and forgetting old information must be traded off.
Finally, the target user should not have to do any work to use the recognition
system by, say, labeling any activities. Our work addresses all of these
challenges by proposing an unsupervised online domain adaptation algorithm.
Both classification and personalization happen continuously and incrementally
in real time. Our solution works by aligning the feature distributions of all
subjects, be they sources or the target, in hidden neural network layers. To
this end, we normalize the input of a layer with user-specific mean and
variance statistics. During training, these statistics are computed over
user-specific batches. In the online phase, they are estimated incrementally
for any new target user. | [
"cs.LG",
"eess.SP",
"stat.ML"
] |
We are interested in understanding how well Transformer language models
(TLMs) can perform reasoning tasks when trained on knowledge encoded in the
form of natural language. We investigate their systematic generalization
abilities on a logical reasoning task in natural language, which involves
reasoning over relationships between entities grounded in first-order logical
proofs. Specifically, we perform soft theorem-proving by leveraging TLMs to
generate natural language proofs. We test the generated proofs for logical
consistency, along with the accuracy of the final inference. We observe
length-generalization issues when evaluated on longer-than-trained sequences.
However, we observe TLMs improve their generalization performance after being
exposed to longer, exhaustive proofs. In addition, we discover that TLMs are
able to generalize better using backward-chaining proofs compared to their
forward-chaining counterparts, while they find it easier to generate forward
chaining proofs. We observe that models that are not trained to generate proofs
are better at generalizing to problems based on longer proofs. This suggests
that Transformers have efficient internal reasoning strategies that are harder
to interpret. These results highlight the systematic generalization behavior of
TLMs in the context of logical reasoning, and we believe this work motivates
deeper inspection of their underlying reasoning strategies. | [
"cs.LG",
"cs.AI",
"cs.CL",
"stat.ML"
] |
Recognizing multiple labels of images is a fundamental but challenging task
in computer vision, and remarkable progress has been attained by localizing
semantic-aware image regions and predicting their labels with deep
convolutional neural networks. The step of hypothesis regions (region
proposals) localization in these existing multi-label image recognition
pipelines, however, usually takes redundant computation cost, e.g., generating
hundreds of meaningless proposals with non-discriminative information and
extracting their features, and the spatial contextual dependency modeling among
the localized regions are often ignored or over-simplified. To resolve these
issues, this paper proposes a recurrent attention reinforcement learning
framework to iteratively discover a sequence of attentional and informative
regions that are related to different semantic objects and further predict
label scores conditioned on these regions. Besides, our method explicitly
models long-term dependencies among these attentional regions that help to
capture semantic label co-occurrence and thus facilitate multi-label
recognition. Extensive experiments and comparisons on two large-scale
benchmarks (i.e., PASCAL VOC and MS-COCO) show that our model achieves superior
performance over existing state-of-the-art methods in both performance and
efficiency as well as explicitly identifying image-level semantic labels to
specific object regions. | [
"cs.CV"
] |
With various face presentation attacks arising under unseen scenarios, face
anti-spoofing (FAS) based on domain generalization (DG) has drawn growing
attention due to its robustness. Most existing methods utilize DG frameworks to
align the features to seek a compact and generalized feature space. However,
little attention has been paid to the feature extraction process for the FAS
task, especially the influence of normalization, which also has a great impact
on the generalization of the learned representation. To address this issue, we
propose a novel perspective of face anti-spoofing that focuses on the
normalization selection in the feature extraction process. Concretely, an
Adaptive Normalized Representation Learning (ANRL) framework is devised, which
adaptively selects feature normalization methods according to the inputs,
aiming to learn domain-agnostic and discriminative representation. Moreover, to
facilitate the representation learning, Dual Calibration Constraints are
designed, including Inter-Domain Compatible loss and Inter-Class Separable
loss, which provide a better optimization direction for generalizable
representation. Extensive experiments and visualizations are presented to
demonstrate the effectiveness of our method against the SOTA competitors. | [
"cs.CV"
] |
Learning robust value functions given raw observations and rewards is now
possible with model-free and model-based deep reinforcement learning
algorithms. There is a third alternative, called Successor Representations
(SR), which decomposes the value function into two components -- a reward
predictor and a successor map. The successor map represents the expected future
state occupancy from any given state and the reward predictor maps states to
scalar rewards. The value function of a state can be computed as the inner
product between the successor map and the reward weights. In this paper, we
present DSR, which generalizes SR within an end-to-end deep reinforcement
learning framework. DSR has several appealing properties including: increased
sensitivity to distal reward changes due to factorization of reward and world
dynamics, and the ability to extract bottleneck states (subgoals) given
successor maps trained under a random policy. We show the efficacy of our
approach on two diverse environments given raw pixel observations -- simple
grid-world domains (MazeBase) and the Doom game engine. | [
"stat.ML",
"cs.AI",
"cs.LG",
"cs.NE"
] |
Recent advances in Generative Adversarial Networks (GANs) have shown
increasing success in generating photorealistic images. But they also raise
challenges to visual forensics and model attribution. We present the first
study of learning GAN fingerprints towards image attribution and using them to
classify an image as real or GAN-generated. For GAN-generated images, we
further identify their sources. Our experiments show that (1) GANs carry
distinct model fingerprints and leave stable fingerprints in their generated
images, which support image attribution; (2) even minor differences in GAN
training can result in different fingerprints, which enables fine-grained model
authentication; (3) fingerprints persist across different image frequencies and
patches and are not biased by GAN artifacts; (4) fingerprint finetuning is
effective in immunizing against five types of adversarial image perturbations;
and (5) comparisons also show our learned fingerprints consistently outperform
several baselines in a variety of setups. | [
"cs.CV",
"cs.CR",
"cs.CY",
"cs.GR",
"cs.LG"
] |
Graph embedding techniques have been increasingly deployed in a multitude of
different applications that involve learning on non-Euclidean data. However,
existing graph embedding models either fail to incorporate node attribute
information during training or suffer from node attribute noise, which
compromises the accuracy. Moreover, very few of them scale to large graphs due
to their high computational complexity and memory usage. In this paper we
propose GraphZoom, a multi-level framework for improving both accuracy and
scalability of unsupervised graph embedding algorithms. GraphZoom first
performs graph fusion to generate a new graph that effectively encodes the
topology of the original graph and the node attribute information. This fused
graph is then repeatedly coarsened into much smaller graphs by merging nodes
with high spectral similarities. GraphZoom allows any existing embedding
methods to be applied to the coarsened graph, before it progressively refine
the embeddings obtained at the coarsest level to increasingly finer graphs. We
have evaluated our approach on a number of popular graph datasets for both
transductive and inductive tasks. Our experiments show that GraphZoom can
substantially increase the classification accuracy and significantly accelerate
the entire graph embedding process by up to 40.8x, when compared to the
state-of-the-art unsupervised embedding methods. | [
"cs.LG",
"stat.ML"
] |
We propose a real-time DNN-based technique to segment hand and object of
interacting motions from depth inputs. Our model is called DenseAttentionSeg,
which contains a dense attention mechanism to fuse information in different
scales and improves the results quality with skip-connections. Besides, we
introduce a contour loss in model training, which helps to generate accurate
hand and object boundaries. Finally, we propose and release our InterSegHands
dataset, a fine-scale hand segmentation dataset containing about 52k depth maps
of hand-object interactions. Our experiments evaluate the effectiveness of our
techniques and datasets, and indicate that our method outperforms the current
state-of-the-art deep segmentation methods on interaction segmentation. | [
"cs.CV"
] |
The main challenge of single image super resolution (SISR) is the recovery of
high frequency details such as tiny textures. However, most of the
state-of-the-art methods lack specific modules to identify high frequency
areas, causing the output image to be blurred. We propose an attention-based
approach to give a discrimination between texture areas and smooth areas. After
the positions of high frequency details are located, high frequency
compensation is carried out. This approach can incorporate with previously
proposed SISR networks. By providing high frequency enhancement, better
performance and visual effect are achieved. We also propose our own SISR
network composed of DenseRes blocks. The block provides an effective way to
combine the low level features and high level features. Extensive benchmark
evaluation shows that our proposed method achieves significant improvement over
the state-of-the-art works in SISR. | [
"cs.CV"
] |
The capability of the self-attention mechanism to model the long-range
dependencies has catapulted its deployment in vision models. Unlike convolution
operators, self-attention offers infinite receptive field and enables
compute-efficient modeling of global dependencies. However, the existing
state-of-the-art attention mechanisms incur high compute and/or parameter
overheads, and hence unfit for compact convolutional neural networks (CNNs). In
this work, we propose a simple yet effective "Ultra-Lightweight Subspace
Attention Mechanism" (ULSAM), which infers different attention maps for each
feature map subspace. We argue that leaning separate attention maps for each
feature subspace enables multi-scale and multi-frequency feature
representation, which is more desirable for fine-grained image classification.
Our method of subspace attention is orthogonal and complementary to the
existing state-of-the-arts attention mechanisms used in vision models. ULSAM is
end-to-end trainable and can be deployed as a plug-and-play module in the
pre-existing compact CNNs. Notably, our work is the first attempt that uses a
subspace attention mechanism to increase the efficiency of compact CNNs. To
show the efficacy of ULSAM, we perform experiments with MobileNet-V1 and
MobileNet-V2 as backbone architectures on ImageNet-1K and three fine-grained
image classification datasets. We achieve $\approx$13% and $\approx$25%
reduction in both the FLOPs and parameter counts of MobileNet-V2 with a 0.27%
and more than 1% improvement in top-1 accuracy on the ImageNet-1K and
fine-grained image classification datasets (respectively). Code and trained
models are available at https://github.com/Nandan91/ULSAM. | [
"cs.CV",
"I.5.1; I.5.2; I.5.4"
] |
Existing graph neural networks (GNNs) largely rely on node embeddings, which
represent a node as a vector by its identity, type, or content. However, graphs
with unlabeled nodes widely exist in real-world applications (e.g., anonymized
social networks). Previous GNNs either assign random labels to nodes (which
introduces artefacts to the GNN) or assign one embedding to all nodes (which
fails to distinguish one node from another). In this paper, we analyze the
limitation of existing approaches in two types of classification tasks, graph
classification and node classification. Inspired by our analysis, we propose
two techniques, Dynamic Labeling and Preferential Dynamic Labeling, that
satisfy desired properties statistically or asymptotically for each type of the
task. Experimental results show that we achieve high performance in various
graph-related tasks. | [
"cs.LG",
"cs.AI"
] |
An important component of autoencoders is the method by which the information
capacity of the latent representation is minimized or limited. In this work,
the rank of the covariance matrix of the codes is implicitly minimized by
relying on the fact that gradient descent learning in multi-layer linear
networks leads to minimum-rank solutions. By inserting a number of extra linear
layers between the encoder and the decoder, the system spontaneously learns
representations with a low effective dimension. The model, dubbed Implicit
Rank-Minimizing Autoencoder (IRMAE), is simple, deterministic, and learns
compact latent spaces. We demonstrate the validity of the method on several
image generation and representation learning tasks. | [
"cs.LG",
"cs.CV",
"stat.ML"
] |
Datasets in sleep science present challenges for machine learning algorithms
due to differences in recording setups across clinics. We investigate two deep
transfer learning strategies for overcoming the channel mismatch problem for
cases where two datasets do not contain exactly the same setup leading to
degraded performance in single-EEG models. Specifically, we train a baseline
model on multivariate polysomnography data and subsequently replace the first
two layers to prepare the architecture for single-channel
electroencephalography data. Using a fine-tuning strategy, our model yields
similar performance to the baseline model (F1=0.682 and F1=0.694,
respectively), and was significantly better than a comparable single-channel
model. Our results are promising for researchers working with small databases
who wish to use deep learning models pre-trained on larger databases. | [
"cs.CV",
"eess.SP",
"stat.AP",
"stat.ML"
] |
Aerial image categorization plays an indispensable role in remote sensing and
artificial intelligence. In this paper, we propose a new aerial image
categorization framework, focusing on organizing the local patches of each
aerial image into multiple discriminative subgraphs. The subgraphs reflect both
the geometric property and the color distribution of an aerial image. First,
each aerial image is decomposed into a collection of regions in terms of their
color intensities. Thereby region connected graph (RCG), which models the
connection between the spatial neighboring regions, is constructed to encode
the spatial context of an aerial image. Second, a subgraph mining technique is
adopted to discover the frequent structures in the RCGs constructed from the
training aerial images. Thereafter, a set of refined structures are selected
among the frequent ones toward being highly discriminative and low redundant.
Lastly, given a new aerial image, its sub-RCGs corresponding to the refined
structures are extracted. They are further quantized into a discriminative
vector for SVM classification. Thorough experimental results validate the
effectiveness of the proposed method. In addition, the visualized mined
subgraphs show that the discriminative topologies of each aerial image are
discovered. | [
"cs.CV"
] |
Explainable Artificial Intelligence (XAI) has in recent years become a
well-suited framework to generate human understandable explanations of black
box models. In this paper, we present a novel XAI visual explanation algorithm
denoted SIDU that can effectively localize entire object regions responsible
for prediction in a full extend. We analyze its robustness and effectiveness
through various computational and human subject experiments. In particular, we
assess the SIDU algorithm using three different types of evaluations
(Application, Human and Functionally-Grounded) to demonstrate its superior
performance. The robustness of SIDU is further studied in presence of
adversarial attack on black box models to better understand its performance. | [
"cs.CV",
"cs.AI",
"cs.HC",
"cs.LG"
] |
In the computational prediction of chemical compound properties, molecular
descriptors and fingerprints encoded to low dimensional vectors are used. The
selection of proper molecular descriptors and fingerprints is both important
and challenging as the performance of such models is highly dependent on
descriptors. To overcome this challenge, natural language processing models
that utilize simplified molecular input line-entry system as input were
studied, and several transformer-variant models achieved superior results when
compared with conventional methods. In this study, we explored the structural
differences of the transformer-variant model and proposed a new self-attention
based model. The representation learning performance of the self-attention
module was evaluated in a multi-task learning environment using imbalanced
chemical datasets. The experiment results showed that our model achieved
competitive outcomes on several benchmark datasets. The source code of our
experiment is available at https://github.com/arwhirang/sa-mtl and the dataset
is available from the same URL. | [
"cs.LG"
] |
Off-the-shelf convolutional neural network features achieve outstanding
results in many image retrieval tasks. However, their invariance to target data
is pre-defined by the network architecture and training data. Existing image
retrieval approaches require fine-tuning or modification of pre-trained
networks to adapt to variations unique to the target data. In contrast, our
method enhances the invariance of off-the-shelf features by aggregating
features extracted from images augmented at test-time, with augmentations
guided by a policy learned through reinforcement learning. The learned policy
assigns different magnitudes and weights to the selected transformations, which
are selected from a list of image transformations. Policies are evaluated using
a metric learning protocol to learn the optimal policy. The model converges
quickly and the cost of each policy iteration is minimal as we propose an
off-line caching technique to greatly reduce the computational cost of
extracting features from augmented images. Experimental results on large
trademark retrieval (METU trademark dataset) and landmark retrieval (ROxford5k
and RParis6k scene datasets) tasks show that the learned ensemble of
transformations is highly effective for improving performance, and is
practical, and transferable. | [
"cs.CV"
] |
Visual Recognition is one of the fundamental challenges in AI, where the goal
is to understand the semantics of visual data. Employing mid-level
representation, in particular, shifted the paradigm in visual recognition. The
mid-level image/video representation involves discovering and training a set of
mid-level visual patterns (e.g., parts and attributes) and represent a given
image/video utilizing them. The mid-level patterns can be extracted from images
and videos using the motion and appearance information of visual phenomenas.
This thesis targets employing mid-level representations for different
high-level visual recognition tasks, namely (i)image understanding and
(ii)video understanding.
In the case of image understanding, we focus on object detection/recognition
task. We investigate on discovering and learning a set of mid-level patches to
be used for representing the images of an object category. We specifically
employ the discriminative patches in a subcategory-aware webly-supervised
fashion. We, additionally, study the outcomes provided by employing the
subcategory-based models for undoing dataset bias. | [
"cs.CV"
] |
Graph Convolutional Networks (GCNs) have shown to be effective in handling
unordered data like point clouds and meshes. In this work we propose novel
approaches for graph convolution, pooling and unpooling, inspired from finite
differences and algebraic multigrid frameworks. We form a parameterized
convolution kernel based on discretized differential operators, leveraging the
graph mass, gradient and Laplacian. This way, the parameterization does not
depend on the graph structure, only on the meaning of the network convolutions
as differential operators. To allow hierarchical representations of the input,
we propose pooling and unpooling operations that are based on algebraic
multigrid methods, which are mainly used to solve partial differential
equations on unstructured grids. To motivate and explain our method, we compare
it to standard convolutional neural networks, and show their similarities and
relations in the case of a regular grid. Our proposed method is demonstrated in
various experiments like classification and part-segmentation, achieving on par
or better than state of the art results. We also analyze the computational cost
of our method compared to other GCNs. | [
"cs.CV",
"cs.GR",
"cs.LG"
] |