Integration of variational autoencoder and spatial clustering for adaptive multi-channel neural speech separation

In this paper, we propose a method combining variational autoencoder model of speech with a spatial clustering approach for multi-channel speech separation. The advantage of integrating spatial clustering with a spectral model was shown in several works… As the spectral model, previous works used either factorial generative models of the mixed speech or discriminative neural networks. In our work, we combine the strengths of both approaches, by building a factorial model based on a generative neural network, a variational autoencoder. […]

Read more

Picking BERT’s Brain: Probing for Linguistic Dependencies in Contextualized Embeddings Using Representational Similarity Analysis

As the name implies, contextualized representations of language are typically motivated by their ability to encode context. Which aspects of context are captured by such representations?.. We introduce an approach to address this question using Representational Similarity Analysis (RSA). As case studies, we investigate the degree to which a verb embedding encodes the verb’s subject, a pronoun embedding encodes the pronoun’s antecedent, and a full-sentence representation encodes the sentence’s head word (as determined by a dependency parse). In all cases, […]

Read more

Benchmarking Image Retrieval for Visual Localization

Visual localization, i.e., camera pose estimation in a known scene, is a core component of technologies such as autonomous driving and augmented reality. State-of-the-art localization approaches often rely on image retrieval techniques for one of two tasks: (1) provide an approximate pose estimate or (2) determine which parts of the scene are potentially visible in a given query image… It is common practice to use state-of-the-art image retrieval algorithms for these tasks. These algorithms are often trained for the goal […]

Read more

Cross-Document Event Coreference Resolution Beyond Corpus-Tailored Systems

Cross-document event coreference resolution (CDCR) is an NLP task in which mentions of events need to be identified and clustered throughout a collection of documents. CDCR aims to benefit downstream multi-document applications, but despite recent progress on corpora and model development, downstream improvements from applying CDCR have not been shown yet… The reason lies in the fact that every CDCR system released to date was developed, trained, and tested only on a single respective corpus. This raises strong concerns on […]

Read more

SpinNet: Learning a General Surface Descriptor for 3D Point Cloud Registration

Extracting robust and general 3D local features is key to downstream tasks such as point cloud registration and reconstruction. Existing learning-based local descriptors are either sensitive to rotation transformations, or rely on classical handcrafted features which are neither general nor representative… In this paper, we introduce a new, yet conceptually simple, neural architecture, termed SpinNet, to extract local features which are rotationally invariant whilst sufficiently informative to enable accurate registration. A Spatial Point Transformer is first introduced to map the […]

Read more

Semi-supervised Gated Recurrent Neural Networks for Robotic Terrain Classification

Legged robots are popular candidates for missions in challenging terrains due to the wide variety of locomotion strategies they can employ. Terrain classification is a key enabling technology for autonomous legged robots, as it allows the robot to harness their innate flexibility to adapt their behaviour to the demands of their operating environment… In this paper, we show how highly capable machine learning techniques, namely gated recurrent neural networks, allow our target legged robot to correctly classify the terrain it […]

Read more

Scale setting the M{รถ}bius Domain Wall Fermion on gradient-flowed HISQ action using the omega baryon mass and the gradient-flow scale $w_0$

We report on a sub-percent scale determination using the omega baryon mass and gradient-flow methods. The calculations are performed on 22 ensembles of $N_f=2+1+1$ highly improved, rooted staggered sea-quark configurations generated by the MILC and CalLat Collaborations… The valence quark action used is M”obius Domain-Wall fermions solved on these configurations after a gradient-flow smearing is applied with a flowtime of $t_{rm gf}=1$ in lattice units. The ensembles span four lattice spacings in the range $0.06 lesssim a lesssim 0.15$~fm, six […]

Read more

Lipophilicity Prediction with Multitask Learning and Molecular Substructures Representation

Lipophilicity is one of the factors determining the permeability of the cell membrane to a drug molecule. Hence, accurate lipophilicity prediction is an essential step in the development of new drugs… In this paper, we introduce a novel approach to encoding additional graph information by extracting molecular substructures. By adding a set of generalized atomic features of these substructures to an established Direct Message Passing Neural Network (D-MPNN) we were able to achieve a new state-of-the-art result at the task […]

Read more

Light Higgsinos for Electroweak Naturalness in Mirage-mediated High-scale Supersymmetry

Mirage mediation realized in the KKLT flux compactification can naturally suppress the up-type Higgs soft mass at low energy scales, and consequently it can reduce the degree of electroweak fine-tuning up to a loop factor. Interestingly, this feature holds even in high-scale supersymmetry as long as the gauge coupling unification is achieved for light Higgsinos below TeV… Under the experimental constraints on the observed Higgs boson, it turns out that mirage mediation can exhibit low electroweak fine-tuning better than a […]

Read more

GLGE: A New General Language Generation Evaluation Benchmark

Multi-task benchmarks such as GLUE and SuperGLUE have driven great progress of pretraining and transfer learning in Natural Language Processing (NLP). These benchmarks mostly focus on a range of Natural Language Understanding (NLU) tasks, without considering the Natural Language Generation (NLG) models… In this paper, we present the General Language Generation Evaluation (GLGE), a new multi-task benchmark for evaluating the generalization capabilities of NLG models across eight language generation tasks. For each task, we continue to design three subtasks in […]

Read more
1 3 4 5 6 7 35