Learning with Differentiable Algorithms

dc.contributor.authorPetersen, Felix
dc.date.accessioned2022-09-01T13:18:13Z
dc.date.available2022-09-01T13:18:13Z
dc.date.issued2022eng
dc.description.abstractClassic algorithms and machine learning systems like neural networks are both abundant in everyday life. While classic computer science algorithms are suitable for precise execution of exactly defined tasks such as finding the shortest path in a large graph, neural networks allow learning from data to predict the most likely answer in more complex tasks such as image classification, which cannot be reduced to an exact algorithm. To get the best of both worlds, this thesis explores combining both concepts leading to more robust, better performing, more interpretable, more computationally efficient, and more data efficient architectures. The thesis formalizes the idea of algorithmic supervision, which allows a neural network to learn from or in conjunction with an algorithm. When integrating an algorithm into a neural architecture, it is important that the algorithm is differentiable such that the architecture can be trained end-to-end and gradients can be propagated back through the algorithm in a meaningful way. To make algorithms differentiable, this thesis proposes a general method for continuously relaxing algorithms by perturbing variables and approximating the expectation value in closed form, i.e., without sampling. In addition, this thesis proposes differentiable algorithms, such as differentiable sorting networks, differentiable renderers, and differentiable logic gate networks. Finally, this thesis presents alternative training strategies for learning with algorithms.eng
dc.description.versionpublishedeng
dc.identifier.ppn181565192X
dc.identifier.urihttps://kops.uni-konstanz.de/handle/123456789/58475
dc.language.isoengeng
dc.rightsterms-of-use
dc.rights.urihttps://rightsstatements.org/page/InC/1.0/
dc.subjectmachine learning, optimization, differentiable, continuous relaxation, computer science, algorithms, neural networkseng
dc.subject.ccsComputing methodologies~Machine learning
dc.subject.ddc004eng
dc.titleLearning with Differentiable Algorithmseng
dc.typeDOCTORAL_THESISeng
dspace.entity.typePublication
kops.citation.bibtex
@phdthesis{Petersen2022Learn-58475,
  year={2022},
  title={Learning with Differentiable Algorithms},
  author={Petersen, Felix},
  address={Konstanz},
  school={Universität Konstanz}
}
kops.citation.iso690PETERSEN, Felix, 2022. Learning with Differentiable Algorithms [Dissertation]. Konstanz: University of Konstanzdeu
kops.citation.iso690PETERSEN, Felix, 2022. Learning with Differentiable Algorithms [Dissertation]. Konstanz: University of Konstanzeng
kops.citation.rdf
<rdf:RDF
    xmlns:dcterms="http://purl.org/dc/terms/"
    xmlns:dc="http://purl.org/dc/elements/1.1/"
    xmlns:rdf="http://www.w3.org/1999/02/22-rdf-syntax-ns#"
    xmlns:bibo="http://purl.org/ontology/bibo/"
    xmlns:dspace="http://digital-repositories.org/ontologies/dspace/0.1.0#"
    xmlns:foaf="http://xmlns.com/foaf/0.1/"
    xmlns:void="http://rdfs.org/ns/void#"
    xmlns:xsd="http://www.w3.org/2001/XMLSchema#" > 
  <rdf:Description rdf:about="https://kops.uni-konstanz.de/server/rdf/resource/123456789/58475">
    <dcterms:isPartOf rdf:resource="https://kops.uni-konstanz.de/server/rdf/resource/123456789/36"/>
    <dcterms:rights rdf:resource="https://rightsstatements.org/page/InC/1.0/"/>
    <dcterms:hasPart rdf:resource="https://kops.uni-konstanz.de/bitstream/123456789/58475/3/Petersen_2-1qlgm6rvkz9m08.pdf"/>
    <dspace:hasBitstream rdf:resource="https://kops.uni-konstanz.de/bitstream/123456789/58475/3/Petersen_2-1qlgm6rvkz9m08.pdf"/>
    <dc:date rdf:datatype="http://www.w3.org/2001/XMLSchema#dateTime">2022-09-01T13:18:13Z</dc:date>
    <dcterms:issued>2022</dcterms:issued>
    <dc:creator>Petersen, Felix</dc:creator>
    <dc:language>eng</dc:language>
    <foaf:homepage rdf:resource="http://localhost:8080/"/>
    <dcterms:abstract xml:lang="eng">Classic algorithms and machine learning systems like neural networks are both abundant in everyday life. While classic computer science algorithms are suitable for precise execution of exactly defined tasks such as finding the shortest path in a large graph, neural networks allow learning from data to predict the most likely answer in more complex tasks such as image classification, which cannot be reduced to an exact algorithm. To get the best of both worlds, this thesis explores combining both concepts leading to more robust, better performing, more interpretable, more computationally efficient, and more data efficient architectures. The thesis formalizes the idea of algorithmic supervision, which allows a neural network to learn from or in conjunction with an algorithm. When integrating an algorithm into a neural architecture, it is important that the algorithm is differentiable such that the architecture can be trained end-to-end and gradients can be propagated back through the algorithm in a meaningful way. To make algorithms differentiable, this thesis proposes a general method for continuously relaxing algorithms by perturbing variables and approximating the expectation value in closed form, i.e., without sampling. In addition, this thesis proposes differentiable algorithms, such as differentiable sorting networks, differentiable renderers, and differentiable logic gate networks. Finally, this thesis presents alternative training strategies for learning with algorithms.</dcterms:abstract>
    <dc:contributor>Petersen, Felix</dc:contributor>
    <dc:rights>terms-of-use</dc:rights>
    <dcterms:available rdf:datatype="http://www.w3.org/2001/XMLSchema#dateTime">2022-09-01T13:18:13Z</dcterms:available>
    <void:sparqlEndpoint rdf:resource="http://localhost/fuseki/dspace/sparql"/>
    <dcterms:title>Learning with Differentiable Algorithms</dcterms:title>
    <bibo:uri rdf:resource="https://kops.uni-konstanz.de/handle/123456789/58475"/>
    <dspace:isPartOfCollection rdf:resource="https://kops.uni-konstanz.de/server/rdf/resource/123456789/36"/>
  </rdf:Description>
</rdf:RDF>
kops.date.examination2022-08-22eng
kops.date.yearDegreeGranted2022eng
kops.description.openAccessopenaccessgreen
kops.flag.knbibliographytrue
kops.identifier.nbnurn:nbn:de:bsz:352-2-1qlgm6rvkz9m08
relation.isAuthorOfPublication1cd68b50-d629-4318-a767-bb15e9faff75
relation.isAuthorOfPublication.latestForDiscovery1cd68b50-d629-4318-a767-bb15e9faff75

Dateien

Originalbündel

Gerade angezeigt 1 - 1 von 1
Vorschaubild nicht verfügbar
Name:
Petersen_2-1qlgm6rvkz9m08.pdf
Größe:
7.1 MB
Format:
Adobe Portable Document Format
Beschreibung:
Petersen_2-1qlgm6rvkz9m08.pdf
Petersen_2-1qlgm6rvkz9m08.pdfGröße: 7.1 MBDownloads: 447

Lizenzbündel

Gerade angezeigt 1 - 1 von 1
Vorschaubild nicht verfügbar
Name:
license.txt
Größe:
3.96 KB
Format:
Item-specific license agreed upon to submission
Beschreibung:
license.txt
license.txtGröße: 3.96 KBDownloads: 0