Skip to content

GitLab

  • Menu
Projects Groups Snippets
    • Loading...
  • Help
    • Help
    • Support
    • Community forum
    • Submit feedback
    • Contribute to GitLab
  • Sign in
  • P public-wiki
  • Project information
    • Project information
    • Activity
    • Labels
    • Members
  • Repository
    • Repository
    • Files
    • Commits
    • Branches
    • Tags
    • Contributors
    • Graph
    • Compare
  • Issues 1
    • Issues 1
    • List
    • Boards
    • Service Desk
    • Milestones
  • Merge requests 0
    • Merge requests 0
  • Deployments
    • Deployments
    • Releases
  • Monitor
    • Monitor
    • Incidents
  • Analytics
    • Analytics
    • Value stream
    • Repository
  • Wiki
    • Wiki
  • Activity
  • Graph
  • Create a new issue
  • Commits
  • Issue Boards
Collapse sidebar
  • nomad-lab
  • public-wiki
  • Wiki
  • Analytics
  • LASSO_L0

Last edited by Ghiringhelli, Luca Massimiliano (lucamghi) Sep 07, 2016
Page history
This is an old version of this page. You can view the most recent version or browse the history.

LASSO_L0

Lasso+L0 performed on a tailor made feature space was introduced in Phys. Rev. Lett. 114, 105503 (2015).

The application of the method goes through these step:

  • The feature space is generated by creating a list of analytical expressions (the derived features), obtained by combining the selected primary features and operations.
  • The Least Absolute Shrinkage and Selection Operator (LASSO) is applied. In practice the following minimization is performed: LASSOformula1 where P is a vector listing the property of interest (here, the RS - ZB difference in energy) for all data points (here, binary materials), D is a matrix whose columns are the derived features listed for each material, c is the (sparse) vector of coefficients that is found upon minimization, lambda is the regularization parameter that determines the level of sparsity (number of non zero elements) of c, and the subscript 1 stays for L1 (also known as Manhattan) norm, i.e., differently from the usual Euclidean norm (L2 norm), the sum of the absolute values of the elements of the argument, that is a vector. The regularization parameter is decreased in small steps starting from the largest value that gives one non-zero element in c, and 50 distinct features that have non-zero coefficient in c are collected.
  • A L0 optimization is performed, formally written as:

where the subscript 0 stays for the L0 quasinorm, that counts the number of non-zero elements of the argument, and D' is the matrix whose columns are the 50 columns selected from D from the previous step. In practice all singletons, pairs, triplets, ... n-tuples (up to the selected maximum dimension of the descriptor) are listed and for each set a linear least-square regression (LLSR) is performed. The n-tuple that gives the lowest mean square error for the LLSR fit is selected as the resulting n-dimensional descriptor.

Back to analytics wiki home

Clone repository
  • D2.1 Documentation
  • Dev culture
  • Encyclopedia
  • GettingAccess
  • LASSO_L0
  • NOMAD software and source code release
  • ParserAssignment
  • ParsersOverview
  • analytics
    • Home
    • LASSO_L0
    • RSvsZB_LASSO_L0
  • Home
  • nomad meta info
  • plots and pictures for the tutorials
  • remoteviz
    • MPCDF RVS
View All Pages