Publikation: Understanding machine-learned density functionals
Dateien
Datum
Autor:innen
Herausgeber:innen
ISSN der Zeitschrift
Electronic ISSN
ISBN
Bibliografische Daten
Verlag
Schriftenreihe
Auflagebezeichnung
DOI (zitierfähiger Link)
Internationale Patentnummer
Angaben zur Forschungsförderung
Projekt
Open Access-Veröffentlichung
Core Facility der Universität Konstanz
Titel in einer weiteren Sprache
Publikationstyp
Publikationsstatus
Erschienen in
Zusammenfassung
Machine learning (ML) is an increasingly popular statistical tool for analyzing either measured or calculated data sets. Here, we explore its application to a well‐defined physics problem, investigating issues of how the underlying physics is handled by ML, and how self‐consistent solutions can be found by limiting the domain in which ML is applied. The particular problem is how to find accurate approximate density functionals for the kinetic energy (KE) of noninteracting electrons. Kernel ridge regression is used to approximate the KE of non‐interacting fermions in a one dimensional box as a functional of their density. The properties of different kernels and methods of cross‐validation are explored, reproducing the physics faithfully in some cases, but not others. We also address how self‐consistency can be achieved with information on only a limited electronic density domain. Accurate constrained optimal densities are found via a modified Euler‐Lagrange constrained minimization of the machine‐learned total energy, despite the poor quality of its functional derivative. A projected gradient descent algorithm is derived using local principal component analysis. Additionally, a sparse grid representation of the density can be used without degrading the performance of the methods. The implications for machine‐learned density functional approximations are discussed.
Zusammenfassung in einer weiteren Sprache
Fachgebiet (DDC)
Schlagwörter
Konferenz
Rezension
Zitieren
ISO 690
LI, Li, John C. SNYDER, Isabelle M. PELASCHIER, Jessica HUANG, Uma-Naresh NIRANJAN, Paul DUNCAN, Matthias RUPP, Klaus-Robert MÜLLER, Kieron BURKE, 2016. Understanding machine-learned density functionals. In: International Journal of Quantum Chemistry. Wiley-Blackwell. 2016, 116(11), pp. 819-833. ISSN 0020-7608. eISSN 1097-461X. Available under: doi: 10.1002/qua.25040BibTex
@article{Li2016Under-52124,
year={2016},
doi={10.1002/qua.25040},
title={Understanding machine-learned density functionals},
number={11},
volume={116},
issn={0020-7608},
journal={International Journal of Quantum Chemistry},
pages={819--833},
author={Li, Li and Snyder, John C. and Pelaschier, Isabelle M. and Huang, Jessica and Niranjan, Uma-Naresh and Duncan, Paul and Rupp, Matthias and Müller, Klaus-Robert and Burke, Kieron}
}RDF
<rdf:RDF
xmlns:dcterms="http://purl.org/dc/terms/"
xmlns:dc="http://purl.org/dc/elements/1.1/"
xmlns:rdf="http://www.w3.org/1999/02/22-rdf-syntax-ns#"
xmlns:bibo="http://purl.org/ontology/bibo/"
xmlns:dspace="http://digital-repositories.org/ontologies/dspace/0.1.0#"
xmlns:foaf="http://xmlns.com/foaf/0.1/"
xmlns:void="http://rdfs.org/ns/void#"
xmlns:xsd="http://www.w3.org/2001/XMLSchema#" >
<rdf:Description rdf:about="https://kops.uni-konstanz.de/server/rdf/resource/123456789/52124">
<dc:contributor>Huang, Jessica</dc:contributor>
<dc:date rdf:datatype="http://www.w3.org/2001/XMLSchema#dateTime">2020-12-15T12:12:18Z</dc:date>
<dc:creator>Huang, Jessica</dc:creator>
<dc:contributor>Li, Li</dc:contributor>
<dc:creator>Niranjan, Uma-Naresh</dc:creator>
<dcterms:isPartOf rdf:resource="https://kops.uni-konstanz.de/server/rdf/resource/123456789/36"/>
<dc:contributor>Niranjan, Uma-Naresh</dc:contributor>
<void:sparqlEndpoint rdf:resource="http://localhost/fuseki/dspace/sparql"/>
<dc:creator>Müller, Klaus-Robert</dc:creator>
<dc:contributor>Pelaschier, Isabelle M.</dc:contributor>
<dc:contributor>Burke, Kieron</dc:contributor>
<dspace:isPartOfCollection rdf:resource="https://kops.uni-konstanz.de/server/rdf/resource/123456789/36"/>
<dc:creator>Pelaschier, Isabelle M.</dc:creator>
<dc:contributor>Duncan, Paul</dc:contributor>
<dc:creator>Burke, Kieron</dc:creator>
<dcterms:available rdf:datatype="http://www.w3.org/2001/XMLSchema#dateTime">2020-12-15T12:12:18Z</dcterms:available>
<dc:creator>Rupp, Matthias</dc:creator>
<foaf:homepage rdf:resource="http://localhost:8080/"/>
<dc:contributor>Snyder, John C.</dc:contributor>
<dc:contributor>Müller, Klaus-Robert</dc:contributor>
<dc:rights>terms-of-use</dc:rights>
<dc:creator>Duncan, Paul</dc:creator>
<dc:language>eng</dc:language>
<dc:creator>Li, Li</dc:creator>
<dcterms:rights rdf:resource="https://rightsstatements.org/page/InC/1.0/"/>
<dcterms:title>Understanding machine-learned density functionals</dcterms:title>
<dc:contributor>Rupp, Matthias</dc:contributor>
<bibo:uri rdf:resource="https://kops.uni-konstanz.de/handle/123456789/52124"/>
<dc:creator>Snyder, John C.</dc:creator>
<dcterms:issued>2016</dcterms:issued>
<dcterms:abstract xml:lang="eng">Machine learning (ML) is an increasingly popular statistical tool for analyzing either measured or calculated data sets. Here, we explore its application to a well‐defined physics problem, investigating issues of how the underlying physics is handled by ML, and how self‐consistent solutions can be found by limiting the domain in which ML is applied. The particular problem is how to find accurate approximate density functionals for the kinetic energy (KE) of noninteracting electrons. Kernel ridge regression is used to approximate the KE of non‐interacting fermions in a one dimensional box as a functional of their density. The properties of different kernels and methods of cross‐validation are explored, reproducing the physics faithfully in some cases, but not others. We also address how self‐consistency can be achieved with information on only a limited electronic density domain. Accurate constrained optimal densities are found via a modified Euler‐Lagrange constrained minimization of the machine‐learned total energy, despite the poor quality of its functional derivative. A projected gradient descent algorithm is derived using local principal component analysis. Additionally, a sparse grid representation of the density can be used without degrading the performance of the methods. The implications for machine‐learned density functional approximations are discussed.</dcterms:abstract>
</rdf:Description>
</rdf:RDF>