Publikation:

Understanding machine-learned density functionals

Lade...
Vorschaubild

Dateien

Zu diesem Dokument gibt es keine Dateien.

Datum

2016

Autor:innen

Li, Li
Snyder, John C.
Pelaschier, Isabelle M.
Huang, Jessica
Niranjan, Uma-Naresh
Duncan, Paul
Müller, Klaus-Robert
Burke, Kieron

Herausgeber:innen

Kontakt

ISSN der Zeitschrift

Electronic ISSN

ISBN

Bibliografische Daten

Verlag

Schriftenreihe

Auflagebezeichnung

URI (zitierfähiger Link)
DOI (zitierfähiger Link)
ArXiv-ID

Internationale Patentnummer

Angaben zur Forschungsförderung

Projekt

Open Access-Veröffentlichung
Core Facility der Universität Konstanz

Gesperrt bis

Titel in einer weiteren Sprache

Publikationstyp
Zeitschriftenartikel
Publikationsstatus
Published

Erschienen in

International Journal of Quantum Chemistry. Wiley-Blackwell. 2016, 116(11), pp. 819-833. ISSN 0020-7608. eISSN 1097-461X. Available under: doi: 10.1002/qua.25040

Zusammenfassung

Machine learning (ML) is an increasingly popular statistical tool for analyzing either measured or calculated data sets. Here, we explore its application to a well‐defined physics problem, investigating issues of how the underlying physics is handled by ML, and how self‐consistent solutions can be found by limiting the domain in which ML is applied. The particular problem is how to find accurate approximate density functionals for the kinetic energy (KE) of noninteracting electrons. Kernel ridge regression is used to approximate the KE of non‐interacting fermions in a one dimensional box as a functional of their density. The properties of different kernels and methods of cross‐validation are explored, reproducing the physics faithfully in some cases, but not others. We also address how self‐consistency can be achieved with information on only a limited electronic density domain. Accurate constrained optimal densities are found via a modified Euler‐Lagrange constrained minimization of the machine‐learned total energy, despite the poor quality of its functional derivative. A projected gradient descent algorithm is derived using local principal component analysis. Additionally, a sparse grid representation of the density can be used without degrading the performance of the methods. The implications for machine‐learned density functional approximations are discussed.

Zusammenfassung in einer weiteren Sprache

Fachgebiet (DDC)
004 Informatik

Schlagwörter

density functional theory, machine learning, orbital free, kinetic energy functional, self‐consistent calculation

Konferenz

Rezension
undefined / . - undefined, undefined

Forschungsvorhaben

Organisationseinheiten

Zeitschriftenheft

Zugehörige Datensätze in KOPS

Zitieren

ISO 690LI, Li, John C. SNYDER, Isabelle M. PELASCHIER, Jessica HUANG, Uma-Naresh NIRANJAN, Paul DUNCAN, Matthias RUPP, Klaus-Robert MÜLLER, Kieron BURKE, 2016. Understanding machine-learned density functionals. In: International Journal of Quantum Chemistry. Wiley-Blackwell. 2016, 116(11), pp. 819-833. ISSN 0020-7608. eISSN 1097-461X. Available under: doi: 10.1002/qua.25040
BibTex
@article{Li2016Under-52124,
  year={2016},
  doi={10.1002/qua.25040},
  title={Understanding machine-learned density functionals},
  number={11},
  volume={116},
  issn={0020-7608},
  journal={International Journal of Quantum Chemistry},
  pages={819--833},
  author={Li, Li and Snyder, John C. and Pelaschier, Isabelle M. and Huang, Jessica and Niranjan, Uma-Naresh and Duncan, Paul and Rupp, Matthias and Müller, Klaus-Robert and Burke, Kieron}
}
RDF
<rdf:RDF
    xmlns:dcterms="http://purl.org/dc/terms/"
    xmlns:dc="http://purl.org/dc/elements/1.1/"
    xmlns:rdf="http://www.w3.org/1999/02/22-rdf-syntax-ns#"
    xmlns:bibo="http://purl.org/ontology/bibo/"
    xmlns:dspace="http://digital-repositories.org/ontologies/dspace/0.1.0#"
    xmlns:foaf="http://xmlns.com/foaf/0.1/"
    xmlns:void="http://rdfs.org/ns/void#"
    xmlns:xsd="http://www.w3.org/2001/XMLSchema#" > 
  <rdf:Description rdf:about="https://kops.uni-konstanz.de/server/rdf/resource/123456789/52124">
    <dc:contributor>Huang, Jessica</dc:contributor>
    <dc:date rdf:datatype="http://www.w3.org/2001/XMLSchema#dateTime">2020-12-15T12:12:18Z</dc:date>
    <dc:creator>Huang, Jessica</dc:creator>
    <dc:contributor>Li, Li</dc:contributor>
    <dc:creator>Niranjan, Uma-Naresh</dc:creator>
    <dcterms:isPartOf rdf:resource="https://kops.uni-konstanz.de/server/rdf/resource/123456789/36"/>
    <dc:contributor>Niranjan, Uma-Naresh</dc:contributor>
    <void:sparqlEndpoint rdf:resource="http://localhost/fuseki/dspace/sparql"/>
    <dc:creator>Müller, Klaus-Robert</dc:creator>
    <dc:contributor>Pelaschier, Isabelle M.</dc:contributor>
    <dc:contributor>Burke, Kieron</dc:contributor>
    <dspace:isPartOfCollection rdf:resource="https://kops.uni-konstanz.de/server/rdf/resource/123456789/36"/>
    <dc:creator>Pelaschier, Isabelle M.</dc:creator>
    <dc:contributor>Duncan, Paul</dc:contributor>
    <dc:creator>Burke, Kieron</dc:creator>
    <dcterms:available rdf:datatype="http://www.w3.org/2001/XMLSchema#dateTime">2020-12-15T12:12:18Z</dcterms:available>
    <dc:creator>Rupp, Matthias</dc:creator>
    <foaf:homepage rdf:resource="http://localhost:8080/"/>
    <dc:contributor>Snyder, John C.</dc:contributor>
    <dc:contributor>Müller, Klaus-Robert</dc:contributor>
    <dc:rights>terms-of-use</dc:rights>
    <dc:creator>Duncan, Paul</dc:creator>
    <dc:language>eng</dc:language>
    <dc:creator>Li, Li</dc:creator>
    <dcterms:rights rdf:resource="https://rightsstatements.org/page/InC/1.0/"/>
    <dcterms:title>Understanding machine-learned density functionals</dcterms:title>
    <dc:contributor>Rupp, Matthias</dc:contributor>
    <bibo:uri rdf:resource="https://kops.uni-konstanz.de/handle/123456789/52124"/>
    <dc:creator>Snyder, John C.</dc:creator>
    <dcterms:issued>2016</dcterms:issued>
    <dcterms:abstract xml:lang="eng">Machine learning (ML) is an increasingly popular statistical tool for analyzing either measured or calculated data sets. Here, we explore its application to a well‐defined physics problem, investigating issues of how the underlying physics is handled by ML, and how self‐consistent solutions can be found by limiting the domain in which ML is applied. The particular problem is how to find accurate approximate density functionals for the kinetic energy (KE) of noninteracting electrons. Kernel ridge regression is used to approximate the KE of non‐interacting fermions in a one dimensional box as a functional of their density. The properties of different kernels and methods of cross‐validation are explored, reproducing the physics faithfully in some cases, but not others. We also address how self‐consistency can be achieved with information on only a limited electronic density domain. Accurate constrained optimal densities are found via a modified Euler‐Lagrange constrained minimization of the machine‐learned total energy, despite the poor quality of its functional derivative. A projected gradient descent algorithm is derived using local principal component analysis. Additionally, a sparse grid representation of the density can be used without degrading the performance of the methods. The implications for machine‐learned density functional approximations are discussed.</dcterms:abstract>
  </rdf:Description>
</rdf:RDF>

Interner Vermerk

xmlui.Submission.submit.DescribeStep.inputForms.label.kops_note_fromSubmitter

Kontakt
URL der Originalveröffentl.

Prüfdatum der URL

Prüfungsdatum der Dissertation

Finanzierungsart

Kommentar zur Publikation

Allianzlizenz
Corresponding Authors der Uni Konstanz vorhanden
Internationale Co-Autor:innen
Universitätsbibliographie
Nein
Begutachtet
Unbekannt
Diese Publikation teilen