**Are you an EPFL student looking for a semester project?**

Work with us on data science and visualisation projects, and deploy your project as an app on top of Graph Search.

Publication# Motivic Donaldson-Thomas invariants for the one-loop quiver with potential

2015

Journal paper

Journal paper

Abstract

We compute the motivic Donaldson-Thomas invariants of the one-loop quiver, with an arbitrary potential. This is the first computation of motivic Donaldson-Thomas invariants to use in an essential way the full machinery of (mu) over cap -equivariant motives, for which we prove a dimensional reduction result similar to that of Behrend, Bryan and Szendroi in their study of degree- zero motivic Donaldson-Thomas invariants. Our result differs from theirs in that it involves nontrivial monodromy.

Official source

This page is automatically generated and may contain information that is not correct, complete, up-to-date, or relevant to your search query. The same applies to every other page on this website. Please make sure to verify the information with EPFL's official sources.

Related concepts (11)

Related publications (32)

Dimensionality reduction

Dimensionality reduction, or dimension reduction, is the transformation of data from a high-dimensional space into a low-dimensional space so that the low-dimensional representation retains some meaningful properties of the original data, ideally close to its intrinsic dimension. Working in high-dimensional spaces can be undesirable for many reasons; raw data are often sparse as a consequence of the curse of dimensionality, and analyzing the data is usually computationally intractable (hard to control or deal with).

Quiver (mathematics)

In graph theory, a quiver is a directed graph where loops and multiple arrows between two vertices are allowed, in other words a multidigraph. They are commonly used in representation theory: a representation V of a quiver assigns a vector space V(x) to each vertex x of the quiver and a linear map V(a) to each arrow a. In , a quiver can be understood to be the underlying structure of a , but without composition or a designation of identity morphisms. That is, there is a forgetful functor from Cat to Quiv.

Nonlinear dimensionality reduction

Nonlinear dimensionality reduction, also known as manifold learning, refers to various related techniques that aim to project high-dimensional data onto lower-dimensional latent manifolds, with the goal of either visualizing the data in the low-dimensional space, or learning the mapping (either from the high-dimensional space to the low-dimensional embedding or vice versa) itself. The techniques described below can be understood as generalizations of linear decomposition methods used for dimensionality reduction, such as singular value decomposition and principal component analysis.

Recent years have witnessed a rise in real-world data captured with rich structural information that can be conveniently depicted by multi-relational graphs. While inference of continuous node features across a simple graph is rather under-studied by the c ...

2021Graphs offer a simple yet meaningful representation of relationships between data. Thisrepresentation is often used in machine learning algorithms in order to incorporate structuralor geometric information about data. However, it can also be used in an inv ...

Pascal Frossard, Effrosyni Simou

In order to perform network analysis tasks, representations that capture the most relevant information in the graph structure are needed. However, existing methods learn representations that cannot be interpreted in a straightforward way and that are relat ...