@kirk86

Robustness Certification of Generative Models

, , and . (2020)cite arxiv:2004.14756Comment: Prior version submitted to ICLR 2020.

Abstract

Generative neural networks can be used to specify continuous transformations between images via latent-space interpolation. However, certifying that all images captured by the resulting path in the image manifold satisfy a given property can be very challenging. This is because this set is highly non-convex, thwarting existing scalable robustness analysis methods, which are often based on convex relaxations. We present ApproxLine, a scalable certification method that successfully verifies non-trivial specifications involving generative models and classifiers. ApproxLine can provide both sound deterministic and probabilistic guarantees, by capturing either infinite non-convex sets of neural network activation vectors or distributions over such sets. We show that ApproxLine is practically useful and can verify interesting interpolations in the networks latent space.

Description

[2004.14756] Robustness Certification of Generative Models

Links and resources

Tags

community

  • @kirk86
  • @dblp
@kirk86's tags highlighted