<?xml version="1.0" encoding="UTF-8"?>
<xml>
<records>
<record>
  <ref-type name="Thesis">32</ref-type>
  <contributors>
    <authors>
      <author>Elsner, Tim</author>
      <author>Kobbelt, Leif</author>
      <author>Shamir, Ariel</author>
    </authors>
    <subsidiary-authors>
      <author>122310</author>
    </subsidiary-authors>
  </contributors>
  <titles>
    <title>Adapting neural representations of visual data for editing and generation</title>
  </titles>
  <periodical/>
  <publisher>RWTH Aachen University</publisher>
  <pub-location>Aachen</pub-location>
  <language>English</language>
  <pages>1 Online-Ressource : Illustrationen</pages>
  <number/>
  <volume/>
  <abstract>Learning and representing the structure of the underlying data is a central aspect of deep learning and ranges from tasks such as image classification to point cloud analysis from laser scanners. This thesis is therefore centred around two questions: First, does imposing the right structure on a visual data representation enable the generation of task-specific edits of an input, and second, can specifically shaping the structure of the latent space aid in learning the underlying data manifold as a whole. For content editing, we show how to learn a latent representation for geometry where the latent information is directly tied to points on the surface of an object to allow intuitive editing. We further demonstrate for images and 3D scenes how to edit the aspect ratio of an input through representing an applied deformation as neural field. As these editing approaches demonstrate, the right representation is crucial for achieving a task. For content generation, we demonstrate how a latent space shaped in spirit of the Gaussian pyramid can be beneficial for generating new images, but further increases the already high computational load. We use this as a motivation for more adaptive representations that require less computational power. We provide a strategy that extends Byte Pair Encoding to compress discretised visual data grids to facilitate generation, adaptively compressing regions of low information density into fewer tokens. We then propose a method that natively learns a representation that is globally adaptive as our final contribution. All of the proposed ideas are centred around controlling and shaping an underlying representation of the information, enabling editing and boosting generation through imposing the right structure on the representation.</abstract>
  <notes>
    <note>Veröffentlicht auf dem Publikationsserver der RWTH Aachen University ; </note>
    <note>Dissertation, RWTH Aachen University, 2025 ; </note>
  </notes>
  <label>2, ; PUB:(DE-HGF)11, ; </label>
  <keywords/>
  <accession-num/>
  <work-type>Dissertation / PhD Thesis</work-type>
  <volume>Dissertation</volume>
  <publisher>RWTH Aachen University</publisher>
  <dates>
    <pub-dates>
      <year>2025</year>
    </pub-dates>
    <year>2025</year>
  </dates>
  <accession-num>RWTH-2025-09736</accession-num>
  <year>2025</year>
  <urls>
    <related-urls>
      <url>https://publications.rwth-aachen.de/record/1021903</url>
    </related-urls>
  </urls>
</record>

</records>
</xml>