2016-05-09 10:19:56 +02:00
|
|
|
|
|
|
|
##########################################################################
|
2013-07-23 19:49:42 +02:00
|
|
|
#
|
|
|
|
# TRIQS: a Toolbox for Research in Interacting Quantum Systems
|
|
|
|
#
|
|
|
|
# Copyright (C) 2011 by M. Aichhorn, L. Pourovskii, V. Vildosola
|
|
|
|
#
|
|
|
|
# TRIQS is free software: you can redistribute it and/or modify it under the
|
|
|
|
# terms of the GNU General Public License as published by the Free Software
|
|
|
|
# Foundation, either version 3 of the License, or (at your option) any later
|
|
|
|
# version.
|
|
|
|
#
|
|
|
|
# TRIQS is distributed in the hope that it will be useful, but WITHOUT ANY
|
|
|
|
# WARRANTY; without even the implied warranty of MERCHANTABILITY or FITNESS
|
|
|
|
# FOR A PARTICULAR PURPOSE. See the GNU General Public License for more
|
|
|
|
# details.
|
|
|
|
#
|
|
|
|
# You should have received a copy of the GNU General Public License along with
|
|
|
|
# TRIQS. If not, see <http://www.gnu.org/licenses/>.
|
|
|
|
#
|
2016-05-09 10:19:56 +02:00
|
|
|
##########################################################################
|
2013-07-23 19:49:42 +02:00
|
|
|
|
|
|
|
from types import *
|
|
|
|
import numpy
|
|
|
|
import pytriqs.utility.dichotomy as dichotomy
|
2017-04-05 20:58:01 +02:00
|
|
|
from pytriqs.gf import *
|
2013-07-23 19:49:42 +02:00
|
|
|
import pytriqs.utility.mpi as mpi
|
2014-11-08 00:37:32 +01:00
|
|
|
from pytriqs.archive import *
|
|
|
|
from symmetry import *
|
2016-09-13 11:57:48 +02:00
|
|
|
from block_structure import BlockStructure
|
2015-06-26 18:47:20 +02:00
|
|
|
from sets import Set
|
2015-07-02 15:17:33 +02:00
|
|
|
from itertools import product
|
2016-08-23 15:40:58 +02:00
|
|
|
from warnings import warn
|
2013-07-23 19:49:42 +02:00
|
|
|
|
|
|
|
|
2016-09-13 11:57:48 +02:00
|
|
|
class SumkDFT(object):
|
2013-07-23 19:49:42 +02:00
|
|
|
"""This class provides a general SumK method for combining ab-initio code and pytriqs."""
|
|
|
|
|
2016-05-09 10:19:56 +02:00
|
|
|
def __init__(self, hdf_file, h_field=0.0, use_dft_blocks=False,
|
|
|
|
dft_data='dft_input', symmcorr_data='dft_symmcorr_input', parproj_data='dft_parproj_input',
|
|
|
|
symmpar_data='dft_symmpar_input', bands_data='dft_bands_input', transp_data='dft_transp_input',
|
|
|
|
misc_data='dft_misc_input'):
|
2015-03-12 00:01:12 +01:00
|
|
|
r"""
|
|
|
|
Initialises the class from data previously stored into an hdf5 archive.
|
|
|
|
|
|
|
|
Parameters
|
|
|
|
----------
|
|
|
|
hdf_file : string
|
|
|
|
Name of hdf5 containing the data.
|
|
|
|
h_field : scalar, optional
|
|
|
|
The value of magnetic field to add to the DFT Hamiltonian.
|
|
|
|
The contribution -h_field*sigma is added to diagonal elements of the Hamiltonian.
|
|
|
|
It cannot be used with the spin-orbit coupling on; namely h_field is set to 0 if self.SO=True.
|
|
|
|
use_dft_blocks : boolean, optional
|
|
|
|
If True, the local Green's function matrix for each spin is divided into smaller blocks
|
|
|
|
with the block structure determined from the DFT density matrix of the corresponding correlated shell.
|
2016-09-13 11:57:48 +02:00
|
|
|
|
|
|
|
Alternatively and additionally, the block structure can be analyzed using :meth:`analyse_block_structure <dft.sumk_dft.SumkDFT.analyse_block_structure>`
|
|
|
|
and manipulated using the SumkDFT.block_structre attribute (see :class:`BlockStructure <dft.block_structure.BlockStructure>`).
|
2015-03-12 00:01:12 +01:00
|
|
|
dft_data : string, optional
|
|
|
|
Name of hdf5 subgroup in which DFT data for projector and lattice Green's function construction are stored.
|
|
|
|
symmcorr_data : string, optional
|
|
|
|
Name of hdf5 subgroup in which DFT data on symmetries of correlated shells
|
|
|
|
(symmetry operations, permutaion matrices etc.) are stored.
|
|
|
|
parproj_data : string, optional
|
|
|
|
Name of hdf5 subgroup in which DFT data on non-normalized projectors for non-correlated
|
|
|
|
states (used in the partial density of states calculations) are stored.
|
|
|
|
symmpar_data : string, optional
|
|
|
|
Name of hdf5 subgroup in which DFT data on symmetries of the non-normalized projectors
|
|
|
|
are stored.
|
|
|
|
bands_data : string, optional
|
|
|
|
Name of hdf5 subgroup in which DFT data necessary for band-structure/k-resolved spectral
|
|
|
|
function calculations (projectors, DFT Hamiltonian for a chosen path in the Brillouin zone etc.)
|
|
|
|
are stored.
|
|
|
|
transp_data : string, optional
|
|
|
|
Name of hdf5 subgroup in which DFT data necessary for transport calculations are stored.
|
|
|
|
misc_data : string, optional
|
|
|
|
Name of hdf5 subgroup in which miscellaneous DFT data are stored.
|
2013-07-23 19:49:42 +02:00
|
|
|
"""
|
|
|
|
|
2014-11-15 20:04:54 +01:00
|
|
|
if not type(hdf_file) == StringType:
|
2015-03-12 00:01:12 +01:00
|
|
|
mpi.report("Give a string for the hdf5 filename to read the input!")
|
2013-07-23 19:49:42 +02:00
|
|
|
else:
|
|
|
|
self.hdf_file = hdf_file
|
2014-11-18 11:30:26 +01:00
|
|
|
self.dft_data = dft_data
|
2014-11-03 14:47:55 +01:00
|
|
|
self.symmcorr_data = symmcorr_data
|
|
|
|
self.parproj_data = parproj_data
|
|
|
|
self.symmpar_data = symmpar_data
|
2013-07-23 19:49:42 +02:00
|
|
|
self.bands_data = bands_data
|
2014-11-20 13:17:46 +01:00
|
|
|
self.transp_data = transp_data
|
2015-02-10 11:55:44 +01:00
|
|
|
self.misc_data = misc_data
|
2013-07-23 19:49:42 +02:00
|
|
|
self.h_field = h_field
|
2014-09-22 19:24:33 +02:00
|
|
|
|
2014-12-03 23:12:39 +01:00
|
|
|
# Read input from HDF:
|
2016-05-09 10:19:56 +02:00
|
|
|
things_to_read = ['energy_unit', 'n_k', 'k_dep_projection', 'SP', 'SO', 'charge_below', 'density_required',
|
|
|
|
'symm_op', 'n_shells', 'shells', 'n_corr_shells', 'corr_shells', 'use_rotations', 'rot_mat',
|
|
|
|
'rot_mat_time_inv', 'n_reps', 'dim_reps', 'T', 'n_orbitals', 'proj_mat', 'bz_weights', 'hopping',
|
2014-11-15 18:15:45 +01:00
|
|
|
'n_inequiv_shells', 'corr_to_inequiv', 'inequiv_to_corr']
|
2016-05-09 10:19:56 +02:00
|
|
|
self.subgroup_present, self.value_read = self.read_input_from_hdf(
|
|
|
|
subgrp=self.dft_data, things_to_read=things_to_read)
|
|
|
|
if self.symm_op:
|
|
|
|
self.symmcorr = Symmetry(hdf_file, subgroup=self.symmcorr_data)
|
2013-07-23 19:49:42 +02:00
|
|
|
|
2014-11-15 20:04:54 +01:00
|
|
|
if self.SO and (abs(self.h_field) > 0.000001):
|
|
|
|
self.h_field = 0.0
|
2016-05-09 10:19:56 +02:00
|
|
|
mpi.report(
|
|
|
|
"For SO, the external magnetic field is not implemented, setting it to 0!")
|
2013-07-23 19:49:42 +02:00
|
|
|
|
2016-05-09 10:19:56 +02:00
|
|
|
self.spin_block_names = [['up', 'down'], ['ud']]
|
|
|
|
self.n_spin_blocks = [2, 1]
|
|
|
|
# Convert spin_block_names to indices -- if spin polarized,
|
|
|
|
# differentiate up and down blocks
|
2014-11-14 09:43:28 +01:00
|
|
|
self.spin_names_to_ind = [{}, {}]
|
2016-05-09 10:19:56 +02:00
|
|
|
for iso in range(2): # SO = 0 or 1
|
2014-11-17 10:48:04 +01:00
|
|
|
for isp in range(self.n_spin_blocks[iso]):
|
2016-05-09 10:19:56 +02:00
|
|
|
self.spin_names_to_ind[iso][
|
|
|
|
self.spin_block_names[iso][isp]] = isp * self.SP
|
2013-07-23 19:49:42 +02:00
|
|
|
|
2016-09-13 11:57:48 +02:00
|
|
|
self.block_structure = BlockStructure()
|
2013-07-23 19:49:42 +02:00
|
|
|
|
|
|
|
# GF structure used for the local things in the k sums
|
2016-05-09 10:19:56 +02:00
|
|
|
# Most general form allowing for all hybridisation, i.e. largest
|
|
|
|
# blocks possible
|
|
|
|
self.gf_struct_sumk = [[(sp, range(self.corr_shells[icrsh]['dim'])) for sp in self.spin_block_names[self.corr_shells[icrsh]['SO']]]
|
|
|
|
for icrsh in range(self.n_corr_shells)]
|
2014-12-03 23:12:39 +01:00
|
|
|
# First set a standard gf_struct solver:
|
2016-05-09 10:19:56 +02:00
|
|
|
self.gf_struct_solver = [dict([(sp, range(self.corr_shells[self.inequiv_to_corr[ish]]['dim']))
|
|
|
|
for sp in self.spin_block_names[self.corr_shells[self.inequiv_to_corr[ish]]['SO']]])
|
|
|
|
for ish in range(self.n_inequiv_shells)]
|
|
|
|
# Set standard (identity) maps from gf_struct_sumk <->
|
|
|
|
# gf_struct_solver
|
|
|
|
self.sumk_to_solver = [{} for ish in range(self.n_inequiv_shells)]
|
|
|
|
self.solver_to_sumk = [{} for ish in range(self.n_inequiv_shells)]
|
|
|
|
self.solver_to_sumk_block = [{}
|
|
|
|
for ish in range(self.n_inequiv_shells)]
|
2014-12-03 23:12:39 +01:00
|
|
|
for ish in range(self.n_inequiv_shells):
|
2016-05-09 10:19:56 +02:00
|
|
|
for block, inner_list in self.gf_struct_sumk[self.inequiv_to_corr[ish]]:
|
2014-12-03 23:12:39 +01:00
|
|
|
self.solver_to_sumk_block[ish][block] = block
|
|
|
|
for inner in inner_list:
|
2016-05-09 10:19:56 +02:00
|
|
|
self.sumk_to_solver[ish][
|
|
|
|
(block, inner)] = (block, inner)
|
|
|
|
self.solver_to_sumk[ish][
|
|
|
|
(block, inner)] = (block, inner)
|
|
|
|
# assume no shells are degenerate
|
|
|
|
self.deg_shells = [[] for ish in range(self.n_inequiv_shells)]
|
2014-12-03 23:12:39 +01:00
|
|
|
|
2016-05-09 10:19:56 +02:00
|
|
|
self.chemical_potential = 0.0 # initialise mu
|
|
|
|
self.init_dc() # initialise the double counting
|
2014-12-03 23:12:39 +01:00
|
|
|
|
2016-05-09 10:19:56 +02:00
|
|
|
# Analyse the block structure and determine the smallest gf_struct
|
|
|
|
# blocks and maps, if desired
|
|
|
|
if use_dft_blocks:
|
|
|
|
self.analyse_block_structure()
|
2013-07-23 19:49:42 +02:00
|
|
|
|
2014-11-09 12:10:40 +01:00
|
|
|
################
|
2015-03-12 00:01:12 +01:00
|
|
|
# hdf5 FUNCTIONS
|
2014-11-09 12:10:40 +01:00
|
|
|
################
|
2014-09-22 19:24:33 +02:00
|
|
|
|
2014-12-03 23:12:39 +01:00
|
|
|
def read_input_from_hdf(self, subgrp, things_to_read):
|
2015-03-12 00:01:12 +01:00
|
|
|
r"""
|
|
|
|
Reads data from the HDF file. Prints a warning if a requested dataset is not found.
|
|
|
|
|
|
|
|
Parameters
|
|
|
|
----------
|
|
|
|
subgrp : string
|
|
|
|
Name of hdf5 file subgroup from which the data are to be read.
|
|
|
|
things_to_read : list of strings
|
|
|
|
List of datasets to be read from the hdf5 file.
|
|
|
|
|
|
|
|
Returns
|
|
|
|
-------
|
|
|
|
subgroup_present : boolean
|
|
|
|
Is the subgrp is present in hdf5 file?
|
|
|
|
value_read : boolean
|
|
|
|
Did the reading of requested datasets succeed?
|
|
|
|
|
2013-07-23 19:49:42 +02:00
|
|
|
"""
|
2014-09-22 19:24:33 +02:00
|
|
|
|
2014-11-07 00:55:40 +01:00
|
|
|
value_read = True
|
2016-05-09 10:19:56 +02:00
|
|
|
# initialise variables on all nodes to ensure mpi broadcast works at
|
|
|
|
# the end
|
|
|
|
for it in things_to_read:
|
|
|
|
setattr(self, it, 0)
|
2014-11-25 17:46:04 +01:00
|
|
|
subgroup_present = 0
|
2014-09-22 19:24:33 +02:00
|
|
|
|
2014-11-07 00:55:40 +01:00
|
|
|
if mpi.is_master_node():
|
2016-05-09 10:19:56 +02:00
|
|
|
ar = HDFArchive(self.hdf_file, 'r')
|
2014-11-07 00:55:40 +01:00
|
|
|
if subgrp in ar:
|
|
|
|
subgroup_present = True
|
2013-07-23 19:49:42 +02:00
|
|
|
# first read the necessary things:
|
|
|
|
for it in things_to_read:
|
2014-11-07 00:55:40 +01:00
|
|
|
if it in ar[subgrp]:
|
2016-05-09 10:19:56 +02:00
|
|
|
setattr(self, it, ar[subgrp][it])
|
2013-07-23 19:49:42 +02:00
|
|
|
else:
|
2016-05-09 10:19:56 +02:00
|
|
|
mpi.report("Loading %s failed!" % it)
|
2014-11-07 00:55:40 +01:00
|
|
|
value_read = False
|
2013-07-23 19:49:42 +02:00
|
|
|
else:
|
2016-05-09 10:19:56 +02:00
|
|
|
if (len(things_to_read) != 0):
|
|
|
|
mpi.report(
|
|
|
|
"Loading failed: No %s subgroup in hdf5!" % subgrp)
|
2014-11-07 00:55:40 +01:00
|
|
|
subgroup_present = False
|
|
|
|
value_read = False
|
2013-07-23 19:49:42 +02:00
|
|
|
del ar
|
|
|
|
# now do the broadcasting:
|
2016-05-09 10:19:56 +02:00
|
|
|
for it in things_to_read:
|
|
|
|
setattr(self, it, mpi.bcast(getattr(self, it)))
|
2014-11-07 00:55:40 +01:00
|
|
|
subgroup_present = mpi.bcast(subgroup_present)
|
|
|
|
value_read = mpi.bcast(value_read)
|
2013-07-23 19:49:42 +02:00
|
|
|
|
2014-11-09 12:10:40 +01:00
|
|
|
return subgroup_present, value_read
|
2014-09-22 19:24:33 +02:00
|
|
|
|
2014-12-03 23:12:39 +01:00
|
|
|
def save(self, things_to_save, subgrp='user_data'):
|
2015-03-12 00:01:12 +01:00
|
|
|
r"""
|
|
|
|
Saves data from a list into the HDF file. Prints a warning if a requested data is not found in SumkDFT object.
|
|
|
|
|
|
|
|
Parameters
|
|
|
|
----------
|
|
|
|
things_to_save : list of strings
|
|
|
|
List of datasets to be saved into the hdf5 file.
|
|
|
|
subgrp : string, optional
|
|
|
|
Name of hdf5 file subgroup in which the data are to be stored.
|
|
|
|
"""
|
2013-07-23 19:49:42 +02:00
|
|
|
|
2016-05-09 10:19:56 +02:00
|
|
|
if not (mpi.is_master_node()):
|
|
|
|
return # do nothing on nodes
|
|
|
|
ar = HDFArchive(self.hdf_file, 'a')
|
|
|
|
if not subgrp in ar:
|
|
|
|
ar.create_group(subgrp)
|
|
|
|
for it in things_to_save:
|
2016-09-13 11:57:48 +02:00
|
|
|
if it in [ "gf_struct_sumk", "gf_struct_solver",
|
|
|
|
"solver_to_sumk", "sumk_to_solver", "solver_to_sumk_block"]:
|
|
|
|
warn("It is not recommended to save '{}' individually. Save 'block_structure' instead.".format(it))
|
2014-11-07 00:55:40 +01:00
|
|
|
try:
|
2016-05-09 10:19:56 +02:00
|
|
|
ar[subgrp][it] = getattr(self, it)
|
2014-11-07 00:55:40 +01:00
|
|
|
except:
|
2016-05-09 10:19:56 +02:00
|
|
|
mpi.report("%s not found, and so not saved." % it)
|
2014-09-22 19:24:33 +02:00
|
|
|
del ar
|
|
|
|
|
2014-12-03 23:12:39 +01:00
|
|
|
def load(self, things_to_load, subgrp='user_data'):
|
2015-03-12 00:01:12 +01:00
|
|
|
r"""
|
|
|
|
Loads user data from the HDF file. Raises an exeption if a requested dataset is not found.
|
|
|
|
|
|
|
|
Parameters
|
|
|
|
----------
|
|
|
|
things_to_read : list of strings
|
|
|
|
List of datasets to be read from the hdf5 file.
|
|
|
|
subgrp : string, optional
|
|
|
|
Name of hdf5 file subgroup from which the data are to be read.
|
|
|
|
|
|
|
|
Returns
|
|
|
|
-------
|
|
|
|
list_to_return : list
|
|
|
|
A list containing data read from hdf5.
|
|
|
|
"""
|
2014-12-03 23:12:39 +01:00
|
|
|
|
2016-05-09 10:19:56 +02:00
|
|
|
if not (mpi.is_master_node()):
|
|
|
|
return # do nothing on nodes
|
|
|
|
ar = HDFArchive(self.hdf_file, 'r')
|
|
|
|
if not subgrp in ar:
|
|
|
|
mpi.report("Loading %s failed!" % subgrp)
|
2014-12-03 23:12:39 +01:00
|
|
|
list_to_return = []
|
2016-05-09 10:19:56 +02:00
|
|
|
for it in things_to_load:
|
2014-12-03 23:12:39 +01:00
|
|
|
try:
|
|
|
|
list_to_return.append(ar[subgrp][it])
|
|
|
|
except:
|
2016-05-09 10:19:56 +02:00
|
|
|
raise ValueError, "load: %s not found, and so not loaded." % it
|
2014-12-03 23:12:39 +01:00
|
|
|
del ar
|
|
|
|
return list_to_return
|
|
|
|
|
2014-11-09 12:10:40 +01:00
|
|
|
################
|
|
|
|
# CORE FUNCTIONS
|
|
|
|
################
|
2013-07-23 19:49:42 +02:00
|
|
|
|
2016-05-09 10:19:56 +02:00
|
|
|
def downfold(self, ik, ish, bname, gf_to_downfold, gf_inp, shells='corr', ir=None):
|
2015-03-12 00:01:12 +01:00
|
|
|
r"""
|
|
|
|
Downfolds a block of the Green's function for a given shell and k-point using the corresponding projector matrices.
|
|
|
|
|
|
|
|
Parameters
|
|
|
|
----------
|
|
|
|
ik : integer
|
|
|
|
k-point index for which the downfolding is to be done.
|
|
|
|
ish : integer
|
|
|
|
Shell index of GF to be downfolded.
|
|
|
|
|
|
|
|
- if shells='corr': ish labels all correlated shells (equivalent or not)
|
|
|
|
- if shells='all': ish labels only representative (inequivalent) non-correlated shells
|
|
|
|
|
|
|
|
bname : string
|
|
|
|
Block name of the target block of the lattice Green's function.
|
|
|
|
gf_to_downfold : Gf
|
|
|
|
Block of the Green's function that is to be downfolded.
|
|
|
|
gf_inp : Gf
|
|
|
|
FIXME
|
|
|
|
shells : string, optional
|
|
|
|
|
|
|
|
- if shells='corr': orthonormalized projectors for correlated shells are used for the downfolding.
|
|
|
|
- if shells='all': non-normalized projectors for all included shells are used for the downfolding.
|
|
|
|
|
|
|
|
ir : integer, optional
|
|
|
|
Index of equivalent site in the non-correlated shell 'ish', only used if shells='all'.
|
2016-05-09 10:19:56 +02:00
|
|
|
|
2015-03-12 00:01:12 +01:00
|
|
|
Returns
|
|
|
|
-------
|
|
|
|
gf_downfolded : Gf
|
|
|
|
Downfolded block of the lattice Green's function.
|
|
|
|
"""
|
2016-05-09 10:19:56 +02:00
|
|
|
|
2013-07-23 19:49:42 +02:00
|
|
|
gf_downfolded = gf_inp.copy()
|
2016-05-09 10:19:56 +02:00
|
|
|
# get spin index for proj. matrices
|
|
|
|
isp = self.spin_names_to_ind[self.SO][bname]
|
|
|
|
n_orb = self.n_orbitals[ik, isp]
|
2015-01-08 13:16:44 +01:00
|
|
|
if shells == 'corr':
|
|
|
|
dim = self.corr_shells[ish]['dim']
|
2016-05-09 10:19:56 +02:00
|
|
|
projmat = self.proj_mat[ik, isp, ish, 0:dim, 0:n_orb]
|
2015-01-08 13:16:44 +01:00
|
|
|
elif shells == 'all':
|
2016-05-09 10:19:56 +02:00
|
|
|
if ir is None:
|
|
|
|
raise ValueError, "downfold: provide ir if treating all shells."
|
2015-01-08 13:16:44 +01:00
|
|
|
dim = self.shells[ish]['dim']
|
2016-05-09 10:19:56 +02:00
|
|
|
projmat = self.proj_mat_all[ik, isp, ish, ir, 0:dim, 0:n_orb]
|
2014-09-22 19:24:33 +02:00
|
|
|
|
2016-05-09 10:19:56 +02:00
|
|
|
gf_downfolded.from_L_G_R(
|
|
|
|
projmat, gf_to_downfold, projmat.conjugate().transpose())
|
2013-07-23 19:49:42 +02:00
|
|
|
|
2016-05-09 10:19:56 +02:00
|
|
|
return gf_downfolded
|
2013-07-23 19:49:42 +02:00
|
|
|
|
2016-05-09 10:19:56 +02:00
|
|
|
def upfold(self, ik, ish, bname, gf_to_upfold, gf_inp, shells='corr', ir=None):
|
2015-03-12 00:01:12 +01:00
|
|
|
r"""
|
|
|
|
Upfolds a block of the Green's function for a given shell and k-point using the corresponding projector matrices.
|
|
|
|
|
|
|
|
Parameters
|
|
|
|
----------
|
|
|
|
ik : integer
|
|
|
|
k-point index for which the upfolding is to be done.
|
|
|
|
ish : integer
|
|
|
|
Shell index of GF to be upfolded.
|
|
|
|
|
|
|
|
- if shells='corr': ish labels all correlated shells (equivalent or not)
|
|
|
|
- if shells='all': ish labels only representative (inequivalent) non-correlated shells
|
|
|
|
|
|
|
|
bname : string
|
|
|
|
Block name of the target block of the lattice Green's function.
|
|
|
|
gf_to_upfold : Gf
|
|
|
|
Block of the Green's function that is to be upfolded.
|
|
|
|
gf_inp : Gf
|
|
|
|
FIXME
|
|
|
|
shells : string, optional
|
|
|
|
|
|
|
|
- if shells='corr': orthonormalized projectors for correlated shells are used for the upfolding.
|
|
|
|
- if shells='all': non-normalized projectors for all included shells are used for the upfolding.
|
|
|
|
|
|
|
|
ir : integer, optional
|
|
|
|
Index of equivalent site in the non-correlated shell 'ish', only used if shells='all'.
|
2016-05-09 10:19:56 +02:00
|
|
|
|
2015-03-12 00:01:12 +01:00
|
|
|
Returns
|
|
|
|
-------
|
|
|
|
gf_upfolded : Gf
|
|
|
|
Upfolded block of the lattice Green's function.
|
|
|
|
"""
|
2016-05-09 10:19:56 +02:00
|
|
|
|
2013-07-23 19:49:42 +02:00
|
|
|
gf_upfolded = gf_inp.copy()
|
2016-05-09 10:19:56 +02:00
|
|
|
# get spin index for proj. matrices
|
|
|
|
isp = self.spin_names_to_ind[self.SO][bname]
|
|
|
|
n_orb = self.n_orbitals[ik, isp]
|
2015-01-08 13:16:44 +01:00
|
|
|
if shells == 'corr':
|
|
|
|
dim = self.corr_shells[ish]['dim']
|
2016-05-09 10:19:56 +02:00
|
|
|
projmat = self.proj_mat[ik, isp, ish, 0:dim, 0:n_orb]
|
2015-01-08 13:16:44 +01:00
|
|
|
elif shells == 'all':
|
2016-05-09 10:19:56 +02:00
|
|
|
if ir is None:
|
|
|
|
raise ValueError, "upfold: provide ir if treating all shells."
|
2015-01-08 13:16:44 +01:00
|
|
|
dim = self.shells[ish]['dim']
|
2016-05-09 10:19:56 +02:00
|
|
|
projmat = self.proj_mat_all[ik, isp, ish, ir, 0:dim, 0:n_orb]
|
|
|
|
|
|
|
|
gf_upfolded.from_L_G_R(
|
|
|
|
projmat.conjugate().transpose(), gf_to_upfold, projmat)
|
2013-07-23 19:49:42 +02:00
|
|
|
|
2016-05-09 10:19:56 +02:00
|
|
|
return gf_upfolded
|
2013-07-23 19:49:42 +02:00
|
|
|
|
2016-05-09 10:19:56 +02:00
|
|
|
def rotloc(self, ish, gf_to_rotate, direction, shells='corr'):
|
2015-03-12 00:01:12 +01:00
|
|
|
r"""
|
|
|
|
Rotates a block of the local Green's function from the local frame to the global frame and vice versa.
|
|
|
|
|
|
|
|
Parameters
|
|
|
|
----------
|
|
|
|
ish : integer
|
2015-06-17 15:13:02 +02:00
|
|
|
Shell index of GF to be rotated.
|
2015-03-12 00:01:12 +01:00
|
|
|
|
|
|
|
- if shells='corr': ish labels all correlated shells (equivalent or not)
|
|
|
|
- if shells='all': ish labels only representative (inequivalent) non-correlated shells
|
|
|
|
|
|
|
|
gf_to_rotate : Gf
|
|
|
|
Block of the Green's function that is to be rotated.
|
|
|
|
direction : string
|
|
|
|
The direction of rotation can be either
|
|
|
|
|
|
|
|
- 'toLocal' : global -> local transformation,
|
|
|
|
- 'toGlobal' : local -> global transformation.
|
|
|
|
|
|
|
|
shells : string, optional
|
|
|
|
|
|
|
|
- if shells='corr': the rotation matrix for the correlated shell 'ish' is used,
|
|
|
|
- if shells='all': the rotation matrix for the generic (non-correlated) shell 'ish' is used.
|
|
|
|
|
|
|
|
Returns
|
|
|
|
-------
|
|
|
|
gf_rotated : Gf
|
|
|
|
Rotated block of the local Green's function.
|
|
|
|
"""
|
2013-07-23 19:49:42 +02:00
|
|
|
|
2016-05-09 10:19:56 +02:00
|
|
|
assert ((direction == 'toLocal') or (direction == 'toGlobal')
|
|
|
|
), "rotloc: Give direction 'toLocal' or 'toGlobal'."
|
2013-07-23 19:49:42 +02:00
|
|
|
gf_rotated = gf_to_rotate.copy()
|
2015-01-08 13:26:25 +01:00
|
|
|
if shells == 'corr':
|
|
|
|
rot_mat_time_inv = self.rot_mat_time_inv
|
|
|
|
rot_mat = self.rot_mat
|
|
|
|
elif shells == 'all':
|
|
|
|
rot_mat_time_inv = self.rot_mat_all_time_inv
|
|
|
|
rot_mat = self.rot_mat_all
|
2014-11-09 12:10:40 +01:00
|
|
|
|
2014-11-15 20:04:54 +01:00
|
|
|
if direction == 'toGlobal':
|
2014-09-22 19:24:33 +02:00
|
|
|
|
2015-01-08 13:26:25 +01:00
|
|
|
if (rot_mat_time_inv[ish] == 1) and self.SO:
|
2014-10-28 16:19:28 +01:00
|
|
|
gf_rotated << gf_rotated.transpose()
|
2016-05-09 10:19:56 +02:00
|
|
|
gf_rotated.from_L_G_R(rot_mat[ish].conjugate(
|
|
|
|
), gf_rotated, rot_mat[ish].transpose())
|
2013-07-23 19:49:42 +02:00
|
|
|
else:
|
2016-05-09 10:19:56 +02:00
|
|
|
gf_rotated.from_L_G_R(rot_mat[ish], gf_rotated, rot_mat[
|
|
|
|
ish].conjugate().transpose())
|
2014-09-22 19:24:33 +02:00
|
|
|
|
2014-11-15 20:04:54 +01:00
|
|
|
elif direction == 'toLocal':
|
2014-09-22 19:24:33 +02:00
|
|
|
|
2015-01-08 13:26:25 +01:00
|
|
|
if (rot_mat_time_inv[ish] == 1) and self.SO:
|
2014-10-28 16:19:28 +01:00
|
|
|
gf_rotated << gf_rotated.transpose()
|
2016-05-09 10:19:56 +02:00
|
|
|
gf_rotated.from_L_G_R(rot_mat[ish].transpose(
|
|
|
|
), gf_rotated, rot_mat[ish].conjugate())
|
2013-07-23 19:49:42 +02:00
|
|
|
else:
|
2016-05-09 10:19:56 +02:00
|
|
|
gf_rotated.from_L_G_R(rot_mat[ish].conjugate(
|
|
|
|
).transpose(), gf_rotated, rot_mat[ish])
|
2014-09-22 19:24:33 +02:00
|
|
|
|
2013-07-23 19:49:42 +02:00
|
|
|
return gf_rotated
|
2014-09-22 19:24:33 +02:00
|
|
|
|
2015-03-11 23:53:47 +01:00
|
|
|
def lattice_gf(self, ik, mu=None, iw_or_w="iw", beta=40, broadening=None, mesh=None, with_Sigma=True, with_dc=True):
|
2015-03-12 00:01:12 +01:00
|
|
|
r"""
|
|
|
|
Calculates the lattice Green function for a given k-point from the DFT Hamiltonian and the self energy.
|
|
|
|
|
|
|
|
Parameters
|
|
|
|
----------
|
|
|
|
ik : integer
|
|
|
|
k-point index.
|
|
|
|
mu : real, optional
|
|
|
|
Chemical potential for which the Green's function is to be calculated.
|
|
|
|
If not provided, self.chemical_potential is used for mu.
|
|
|
|
iw_or_w : string, optional
|
|
|
|
|
|
|
|
- `iw_or_w` = 'iw' for a imaginary-frequency self-energy
|
|
|
|
- `iw_or_w` = 'w' for a real-frequency self-energy
|
|
|
|
|
|
|
|
beta : real, optional
|
|
|
|
Inverse temperature.
|
|
|
|
broadening : real, optional
|
|
|
|
Imaginary shift for the axis along which the real-axis GF is calculated.
|
|
|
|
If not provided, broadening will be set to double of the distance between mesh points in 'mesh'.
|
|
|
|
mesh : list, optional
|
|
|
|
Data defining mesh on which the real-axis GF will be calculated, given in the form
|
|
|
|
(om_min,om_max,n_points), where om_min is the minimum omega, om_max is the maximum omega and n_points is the number of points.
|
|
|
|
with_Sigma : boolean, optional
|
|
|
|
If True the GF will be calculated with the self-energy stored in self.Sigmaimp_(w/iw), for real/Matsubara GF, respectively.
|
|
|
|
In this case the mesh is taken from the self.Sigma_imp object.
|
|
|
|
If with_Sigma=True but self.Sigmaimp_(w/iw) is not present, with_Sigma is reset to False.
|
|
|
|
with_dc : boolean, optional
|
|
|
|
if True and with_Sigma=True, the dc correction is substracted from the self-energy before it is included into GF.
|
2016-05-09 10:19:56 +02:00
|
|
|
|
2015-03-12 00:01:12 +01:00
|
|
|
Returns
|
|
|
|
-------
|
|
|
|
G_latt : BlockGf
|
|
|
|
Lattice Green's function.
|
2016-05-09 10:19:56 +02:00
|
|
|
|
2015-03-12 00:01:12 +01:00
|
|
|
"""
|
2016-05-09 10:19:56 +02:00
|
|
|
if mu is None:
|
|
|
|
mu = self.chemical_potential
|
2014-11-14 09:43:28 +01:00
|
|
|
ntoi = self.spin_names_to_ind[self.SO]
|
2014-11-17 10:48:04 +01:00
|
|
|
spn = self.spin_block_names[self.SO]
|
2016-05-09 10:19:56 +02:00
|
|
|
if (iw_or_w != "iw") and (iw_or_w != "w"):
|
|
|
|
raise ValueError, "lattice_gf: Implemented only for Re/Im frequency functions."
|
|
|
|
if not hasattr(self, "Sigma_imp_" + iw_or_w):
|
|
|
|
with_Sigma = False
|
2014-12-07 00:29:39 +01:00
|
|
|
if broadening is None:
|
|
|
|
if mesh is None:
|
|
|
|
broadening = 0.01
|
2016-05-09 10:19:56 +02:00
|
|
|
else: # broadening = 2 * \Delta omega, where \Delta omega is the spacing of omega points
|
|
|
|
broadening = 2.0 * ((mesh[1] - mesh[0]) / (mesh[2] - 1))
|
2014-12-07 00:29:39 +01:00
|
|
|
|
|
|
|
# Are we including Sigma?
|
2014-11-15 20:04:54 +01:00
|
|
|
if with_Sigma:
|
2016-05-09 10:19:56 +02:00
|
|
|
Sigma_imp = getattr(self, "Sigma_imp_" + iw_or_w)
|
2015-04-15 15:35:03 +02:00
|
|
|
sigma_minus_dc = [s.copy() for s in Sigma_imp]
|
2016-05-09 10:19:56 +02:00
|
|
|
if with_dc:
|
|
|
|
sigma_minus_dc = self.add_dc(iw_or_w)
|
2015-04-13 15:11:04 +02:00
|
|
|
if iw_or_w == "iw":
|
2016-05-09 10:19:56 +02:00
|
|
|
# override beta if Sigma_iw is present
|
|
|
|
beta = Sigma_imp[0].mesh.beta
|
2015-12-09 14:55:03 +01:00
|
|
|
mesh = Sigma_imp[0].mesh
|
|
|
|
elif iw_or_w == "w":
|
|
|
|
mesh = Sigma_imp[0].mesh
|
2016-08-23 15:40:58 +02:00
|
|
|
if broadening>0 and mpi.is_master_node():
|
|
|
|
warn('lattice_gf called with Sigma and broadening > 0 (broadening = {}). You might want to explicitly set the broadening to 0.'.format(broadening))
|
2014-12-07 00:29:39 +01:00
|
|
|
else:
|
2015-12-09 14:55:03 +01:00
|
|
|
if iw_or_w == "iw":
|
2016-05-09 10:19:56 +02:00
|
|
|
if beta is None:
|
|
|
|
raise ValueError, "lattice_gf: Give the beta for the lattice GfReFreq."
|
|
|
|
# Default number of Matsubara frequencies
|
|
|
|
mesh = MeshImFreq(beta=beta, S='Fermion', n_max=1025)
|
2015-12-09 14:55:03 +01:00
|
|
|
elif iw_or_w == "w":
|
2016-05-09 10:19:56 +02:00
|
|
|
if mesh is None:
|
|
|
|
raise ValueError, "lattice_gf: Give the mesh=(om_min,om_max,n_points) for the lattice GfReFreq."
|
|
|
|
mesh = MeshReFreq(mesh[0], mesh[1], mesh[2])
|
2014-12-07 00:29:39 +01:00
|
|
|
|
|
|
|
# Check if G_latt is present
|
|
|
|
set_up_G_latt = False # Assume not
|
2016-05-09 10:19:56 +02:00
|
|
|
if not hasattr(self, "G_latt_" + iw_or_w):
|
|
|
|
# Need to create G_latt_(i)w
|
|
|
|
set_up_G_latt = True
|
2014-12-07 00:29:39 +01:00
|
|
|
else: # Check that existing GF is consistent
|
2016-05-09 10:19:56 +02:00
|
|
|
G_latt = getattr(self, "G_latt_" + iw_or_w)
|
2017-10-04 23:50:20 +02:00
|
|
|
GFsize = [gf.target_shape[0] for bname, gf in G_latt]
|
2016-05-09 10:19:56 +02:00
|
|
|
unchangedsize = all([self.n_orbitals[ik, ntoi[spn[isp]]] == GFsize[
|
|
|
|
isp] for isp in range(self.n_spin_blocks[self.SO])])
|
|
|
|
if not unchangedsize:
|
|
|
|
set_up_G_latt = True
|
|
|
|
if (iw_or_w == "iw") and (self.G_latt_iw.mesh.beta != beta):
|
|
|
|
set_up_G_latt = True # additional check for ImFreq
|
2014-12-07 00:29:39 +01:00
|
|
|
|
|
|
|
# Set up G_latt
|
|
|
|
if set_up_G_latt:
|
2016-05-09 10:19:56 +02:00
|
|
|
block_structure = [
|
|
|
|
range(self.n_orbitals[ik, ntoi[sp]]) for sp in spn]
|
|
|
|
gf_struct = [(spn[isp], block_structure[isp])
|
|
|
|
for isp in range(self.n_spin_blocks[self.SO])]
|
|
|
|
block_ind_list = [block for block, inner in gf_struct]
|
2014-12-07 00:29:39 +01:00
|
|
|
if iw_or_w == "iw":
|
2016-05-09 10:19:56 +02:00
|
|
|
glist = lambda: [GfImFreq(indices=inner, mesh=mesh)
|
|
|
|
for block, inner in gf_struct]
|
2014-12-07 00:29:39 +01:00
|
|
|
elif iw_or_w == "w":
|
2016-05-09 10:19:56 +02:00
|
|
|
glist = lambda: [GfReFreq(indices=inner, mesh=mesh)
|
|
|
|
for block, inner in gf_struct]
|
|
|
|
G_latt = BlockGf(name_list=block_ind_list,
|
|
|
|
block_list=glist(), make_copies=False)
|
2014-12-07 00:29:39 +01:00
|
|
|
G_latt.zero()
|
|
|
|
|
|
|
|
if iw_or_w == "iw":
|
|
|
|
G_latt << iOmega_n
|
|
|
|
elif iw_or_w == "w":
|
2016-05-09 10:19:56 +02:00
|
|
|
G_latt << Omega + 1j * broadening
|
2013-07-23 19:49:42 +02:00
|
|
|
|
2016-05-09 10:19:56 +02:00
|
|
|
idmat = [numpy.identity(
|
|
|
|
self.n_orbitals[ik, ntoi[sp]], numpy.complex_) for sp in spn]
|
2013-07-23 19:49:42 +02:00
|
|
|
M = copy.deepcopy(idmat)
|
2014-12-07 00:29:39 +01:00
|
|
|
for ibl in range(self.n_spin_blocks[self.SO]):
|
|
|
|
ind = ntoi[spn[ibl]]
|
2016-05-09 10:19:56 +02:00
|
|
|
n_orb = self.n_orbitals[ik, ind]
|
|
|
|
M[ibl] = self.hopping[ik, ind, 0:n_orb, 0:n_orb] - \
|
|
|
|
(idmat[ibl] * mu) - (idmat[ibl] * self.h_field * (1 - 2 * ibl))
|
2014-12-07 00:29:39 +01:00
|
|
|
G_latt -= M
|
2013-07-23 19:49:42 +02:00
|
|
|
|
2014-11-15 20:04:54 +01:00
|
|
|
if with_Sigma:
|
|
|
|
for icrsh in range(self.n_corr_shells):
|
2016-05-09 10:19:56 +02:00
|
|
|
for bname, gf in G_latt:
|
|
|
|
gf -= self.upfold(ik, icrsh, bname,
|
|
|
|
sigma_minus_dc[icrsh][bname], gf)
|
2013-07-23 19:49:42 +02:00
|
|
|
|
2014-12-07 00:29:39 +01:00
|
|
|
G_latt.invert()
|
2016-05-09 10:19:56 +02:00
|
|
|
setattr(self, "G_latt_" + iw_or_w, G_latt)
|
2013-07-23 19:49:42 +02:00
|
|
|
|
2014-12-07 00:29:39 +01:00
|
|
|
return G_latt
|
2013-07-23 19:49:42 +02:00
|
|
|
|
2016-05-09 10:19:56 +02:00
|
|
|
def set_Sigma(self, Sigma_imp):
|
2015-11-02 11:42:47 +01:00
|
|
|
self.put_Sigma(Sigma_imp)
|
2013-07-23 19:49:42 +02:00
|
|
|
|
2014-11-17 10:48:04 +01:00
|
|
|
def put_Sigma(self, Sigma_imp):
|
2015-03-12 00:01:12 +01:00
|
|
|
r"""
|
|
|
|
Inserts the impurity self-energies into the sumk_dft class.
|
|
|
|
|
|
|
|
Parameters
|
|
|
|
----------
|
|
|
|
Sigma_imp : list of BlockGf (Green's function) objects
|
|
|
|
List containing impurity self-energy for all inequivalent correlated shells.
|
|
|
|
Self-energies for equivalent shells are then automatically set by this function.
|
|
|
|
The self-energies can be of the real or imaginary-frequency type.
|
|
|
|
"""
|
2014-11-17 10:48:04 +01:00
|
|
|
|
2016-05-09 10:19:56 +02:00
|
|
|
assert isinstance(
|
|
|
|
Sigma_imp, list), "put_Sigma: Sigma_imp has to be a list of Sigmas for the correlated shells, even if it is of length 1!"
|
|
|
|
assert len(
|
|
|
|
Sigma_imp) == self.n_inequiv_shells, "put_Sigma: give exactly one Sigma for each inequivalent corr. shell!"
|
2014-11-17 10:48:04 +01:00
|
|
|
|
2014-12-07 00:29:39 +01:00
|
|
|
# init self.Sigma_imp_(i)w:
|
2017-04-05 20:58:01 +02:00
|
|
|
if all( (isinstance(gf, Gf) and isinstance (gf.mesh, MeshImFreq)) for bname, gf in Sigma_imp[0]):
|
2014-11-17 10:48:04 +01:00
|
|
|
# Imaginary frequency Sigma:
|
2016-05-09 10:19:56 +02:00
|
|
|
self.Sigma_imp_iw = [BlockGf(name_block_generator=[(block, GfImFreq(indices=inner, mesh=Sigma_imp[0].mesh))
|
|
|
|
for block, inner in self.gf_struct_sumk[icrsh]], make_copies=False)
|
|
|
|
for icrsh in range(self.n_corr_shells)]
|
2014-12-07 00:29:39 +01:00
|
|
|
SK_Sigma_imp = self.Sigma_imp_iw
|
2017-04-05 20:58:01 +02:00
|
|
|
elif all( isinstance(gf, Gf) and isinstance (gf.mesh, MeshReFreq) for bname, gf in Sigma_imp[0]):
|
2014-11-17 10:48:04 +01:00
|
|
|
# Real frequency Sigma:
|
2016-05-09 10:19:56 +02:00
|
|
|
self.Sigma_imp_w = [BlockGf(name_block_generator=[(block, GfReFreq(indices=inner, mesh=Sigma_imp[0].mesh))
|
|
|
|
for block, inner in self.gf_struct_sumk[icrsh]], make_copies=False)
|
|
|
|
for icrsh in range(self.n_corr_shells)]
|
2014-12-07 00:29:39 +01:00
|
|
|
SK_Sigma_imp = self.Sigma_imp_w
|
2014-11-17 10:48:04 +01:00
|
|
|
else:
|
2014-12-07 00:29:39 +01:00
|
|
|
raise ValueError, "put_Sigma: This type of Sigma is not handled."
|
2014-11-17 10:48:04 +01:00
|
|
|
|
|
|
|
# transform the CTQMC blocks to the full matrix:
|
2014-11-15 20:04:54 +01:00
|
|
|
for icrsh in range(self.n_corr_shells):
|
2016-05-09 10:19:56 +02:00
|
|
|
# ish is the index of the inequivalent shell corresponding to icrsh
|
|
|
|
ish = self.corr_to_inequiv[icrsh]
|
|
|
|
for block, inner in self.gf_struct_solver[ish].iteritems():
|
2014-11-17 10:48:04 +01:00
|
|
|
for ind1 in inner:
|
|
|
|
for ind2 in inner:
|
2016-05-09 10:19:56 +02:00
|
|
|
block_sumk, ind1_sumk = self.solver_to_sumk[
|
|
|
|
ish][(block, ind1)]
|
|
|
|
block_sumk, ind2_sumk = self.solver_to_sumk[
|
|
|
|
ish][(block, ind2)]
|
|
|
|
SK_Sigma_imp[icrsh][block_sumk][
|
|
|
|
ind1_sumk, ind2_sumk] << Sigma_imp[ish][block][ind1, ind2]
|
2013-07-23 19:49:42 +02:00
|
|
|
|
2014-11-17 10:48:04 +01:00
|
|
|
# rotation from local to global coordinate system:
|
|
|
|
if self.use_rotations:
|
|
|
|
for icrsh in range(self.n_corr_shells):
|
2016-05-09 10:19:56 +02:00
|
|
|
for bname, gf in SK_Sigma_imp[icrsh]:
|
|
|
|
gf << self.rotloc(icrsh, gf, direction='toGlobal')
|
2014-09-22 19:24:33 +02:00
|
|
|
|
2016-04-20 19:01:29 +02:00
|
|
|
def extract_G_loc(self, mu=None, iw_or_w='iw', with_Sigma=True, with_dc=True, broadening=None):
|
2015-03-12 00:01:12 +01:00
|
|
|
r"""
|
|
|
|
Extracts the local downfolded Green function by the Brillouin-zone integration of the lattice Green's function.
|
|
|
|
|
|
|
|
Parameters
|
|
|
|
----------
|
|
|
|
mu : real, optional
|
|
|
|
Input chemical potential. If not provided the value of self.chemical_potential is used as mu.
|
|
|
|
with_Sigma : boolean, optional
|
|
|
|
If True then the local GF is calculated with the self-energy self.Sigma_imp.
|
|
|
|
with_dc : boolean, optional
|
|
|
|
If True then the double-counting correction is subtracted from the self-energy in calculating the GF.
|
2016-04-20 19:01:29 +02:00
|
|
|
broadening : float, optional
|
|
|
|
Imaginary shift for the axis along which the real-axis GF is calculated.
|
|
|
|
If not provided, broadening will be set to double of the distance between mesh points in 'mesh'.
|
|
|
|
Only relevant for real-frequency GF.
|
2015-03-12 00:01:12 +01:00
|
|
|
|
|
|
|
Returns
|
|
|
|
-------
|
|
|
|
G_loc_inequiv : list of BlockGf (Green's function) objects
|
|
|
|
List of the local Green's functions for all inequivalent correlated shells,
|
|
|
|
rotated into the corresponding local frames.
|
2016-05-09 10:19:56 +02:00
|
|
|
|
2014-11-17 10:48:04 +01:00
|
|
|
"""
|
2014-11-16 18:02:04 +01:00
|
|
|
|
2016-05-09 10:19:56 +02:00
|
|
|
if mu is None:
|
|
|
|
mu = self.chemical_potential
|
|
|
|
|
|
|
|
if iw_or_w == "iw":
|
|
|
|
G_loc = [self.Sigma_imp_iw[icrsh].copy() for icrsh in range(
|
|
|
|
self.n_corr_shells)] # this list will be returned
|
2016-04-20 19:01:29 +02:00
|
|
|
beta = G_loc[0].mesh.beta
|
2016-05-09 10:19:56 +02:00
|
|
|
G_loc_inequiv = [BlockGf(name_block_generator=[(block, GfImFreq(indices=inner, mesh=G_loc[0].mesh)) for block, inner in self.gf_struct_solver[ish].iteritems()],
|
|
|
|
make_copies=False) for ish in range(self.n_inequiv_shells)]
|
|
|
|
elif iw_or_w == "w":
|
|
|
|
G_loc = [self.Sigma_imp_w[icrsh].copy() for icrsh in range(
|
|
|
|
self.n_corr_shells)] # this list will be returned
|
2016-04-20 19:01:29 +02:00
|
|
|
mesh = G_loc[0].mesh
|
2016-05-09 10:19:56 +02:00
|
|
|
G_loc_inequiv = [BlockGf(name_block_generator=[(block, GfReFreq(indices=inner, mesh=mesh)) for block, inner in self.gf_struct_solver[ish].iteritems()],
|
|
|
|
make_copies=False) for ish in range(self.n_inequiv_shells)]
|
|
|
|
|
|
|
|
for icrsh in range(self.n_corr_shells):
|
|
|
|
G_loc[icrsh].zero() # initialize to zero
|
2014-11-16 18:02:04 +01:00
|
|
|
|
2014-11-17 10:48:04 +01:00
|
|
|
ikarray = numpy.array(range(self.n_k))
|
|
|
|
for ik in mpi.slice_array(ikarray):
|
2016-04-20 19:01:29 +02:00
|
|
|
if iw_or_w == 'iw':
|
2016-05-09 10:19:56 +02:00
|
|
|
G_latt = self.lattice_gf(
|
|
|
|
ik=ik, mu=mu, iw_or_w=iw_or_w, with_Sigma=with_Sigma, with_dc=with_dc, beta=beta)
|
2016-04-20 19:01:29 +02:00
|
|
|
elif iw_or_w == 'w':
|
2016-08-23 15:40:58 +02:00
|
|
|
mesh_parameters = (G_loc[0].mesh.omega_min,G_loc[0].mesh.omega_max,len(G_loc[0].mesh))
|
2016-05-09 10:19:56 +02:00
|
|
|
G_latt = self.lattice_gf(
|
2016-08-23 15:40:58 +02:00
|
|
|
ik=ik, mu=mu, iw_or_w=iw_or_w, with_Sigma=with_Sigma, with_dc=with_dc, broadening=broadening, mesh=mesh_parameters)
|
2016-04-20 19:01:29 +02:00
|
|
|
G_latt *= self.bz_weights[ik]
|
2014-09-22 19:24:33 +02:00
|
|
|
|
2013-07-23 19:49:42 +02:00
|
|
|
for icrsh in range(self.n_corr_shells):
|
2016-05-09 10:19:56 +02:00
|
|
|
# init temporary storage
|
|
|
|
tmp = G_loc[icrsh].copy()
|
|
|
|
for bname, gf in tmp:
|
|
|
|
tmp[bname] << self.downfold(
|
|
|
|
ik, icrsh, bname, G_latt[bname], gf)
|
2015-01-22 10:47:53 +01:00
|
|
|
G_loc[icrsh] += tmp
|
2013-07-23 19:49:42 +02:00
|
|
|
|
2015-03-13 11:10:38 +01:00
|
|
|
# Collect data from mpi
|
2016-05-09 10:19:56 +02:00
|
|
|
for icrsh in range(self.n_corr_shells):
|
|
|
|
G_loc[icrsh] << mpi.all_reduce(
|
|
|
|
mpi.world, G_loc[icrsh], lambda x, y: x + y)
|
2013-07-23 19:49:42 +02:00
|
|
|
mpi.barrier()
|
|
|
|
|
2015-01-22 10:47:53 +01:00
|
|
|
# G_loc[:] is now the sum over k projected to the local orbitals.
|
2014-11-17 10:48:04 +01:00
|
|
|
# here comes the symmetrisation, if needed:
|
2016-05-09 10:19:56 +02:00
|
|
|
if self.symm_op != 0:
|
|
|
|
G_loc = self.symmcorr.symmetrize(G_loc)
|
2013-07-23 19:49:42 +02:00
|
|
|
|
2015-01-22 10:47:53 +01:00
|
|
|
# G_loc is rotated to the local coordinate system:
|
2014-11-15 20:04:54 +01:00
|
|
|
if self.use_rotations:
|
|
|
|
for icrsh in range(self.n_corr_shells):
|
2016-05-09 10:19:56 +02:00
|
|
|
for bname, gf in G_loc[icrsh]:
|
|
|
|
G_loc[icrsh][bname] << self.rotloc(
|
|
|
|
icrsh, gf, direction='toLocal')
|
2013-07-23 19:49:42 +02:00
|
|
|
|
2014-11-17 10:48:04 +01:00
|
|
|
# transform to CTQMC blocks:
|
|
|
|
for ish in range(self.n_inequiv_shells):
|
2016-05-09 10:19:56 +02:00
|
|
|
for block, inner in self.gf_struct_solver[ish].iteritems():
|
2014-11-17 10:48:04 +01:00
|
|
|
for ind1 in inner:
|
|
|
|
for ind2 in inner:
|
2016-05-09 10:19:56 +02:00
|
|
|
block_sumk, ind1_sumk = self.solver_to_sumk[
|
|
|
|
ish][(block, ind1)]
|
|
|
|
block_sumk, ind2_sumk = self.solver_to_sumk[
|
|
|
|
ish][(block, ind2)]
|
|
|
|
G_loc_inequiv[ish][block][ind1, ind2] << G_loc[
|
|
|
|
self.inequiv_to_corr[ish]][block_sumk][ind1_sumk, ind2_sumk]
|
2014-11-17 10:48:04 +01:00
|
|
|
|
|
|
|
# return only the inequivalent shells:
|
2015-01-22 10:47:53 +01:00
|
|
|
return G_loc_inequiv
|
2013-07-23 19:49:42 +02:00
|
|
|
|
2016-09-12 15:29:32 +02:00
|
|
|
def analyse_block_structure(self, threshold=0.00001, include_shells=None, dm=None, hloc=None):
|
2015-03-12 00:01:12 +01:00
|
|
|
r"""
|
|
|
|
Determines the block structure of local Green's functions by analysing the structure of
|
2016-09-12 15:29:32 +02:00
|
|
|
the corresponding density matrices and the local Hamiltonian. The resulting block structures
|
2016-09-13 11:57:48 +02:00
|
|
|
for correlated shells are stored in the :class:`SumkDFT.block_structure <dft.block_structure.BlockStructure>` attribute.
|
2015-03-12 00:01:12 +01:00
|
|
|
|
|
|
|
Parameters
|
|
|
|
----------
|
|
|
|
threshold : real, optional
|
2016-09-12 15:29:32 +02:00
|
|
|
If the difference between density matrix / hloc elements is below threshold,
|
2015-03-12 00:01:12 +01:00
|
|
|
they are considered to be equal.
|
|
|
|
include_shells : list of integers, optional
|
|
|
|
List of correlated shells to be analysed.
|
|
|
|
If include_shells is not provided all correlated shells will be analysed.
|
|
|
|
dm : list of dict, optional
|
|
|
|
List of density matrices from which block stuctures are to be analysed.
|
|
|
|
Each density matrix is a dict {block names: 2d numpy arrays}.
|
|
|
|
If not provided, dm will be calculated from the DFT Hamiltonian by a simple-point BZ integration.
|
2016-09-12 15:29:32 +02:00
|
|
|
hloc : list of dict, optional
|
|
|
|
List of local Hamiltonian matrices from which block stuctures are to be analysed
|
|
|
|
Each Hamiltonian is a dict {block names: 2d numpy arrays}.
|
|
|
|
If not provided, it will be calculated using eff_atomic_levels.
|
2015-03-12 00:01:12 +01:00
|
|
|
"""
|
2013-07-23 19:49:42 +02:00
|
|
|
|
2016-05-09 10:19:56 +02:00
|
|
|
self.gf_struct_solver = [{} for ish in range(self.n_inequiv_shells)]
|
|
|
|
self.sumk_to_solver = [{} for ish in range(self.n_inequiv_shells)]
|
|
|
|
self.solver_to_sumk = [{} for ish in range(self.n_inequiv_shells)]
|
|
|
|
self.solver_to_sumk_block = [{}
|
|
|
|
for ish in range(self.n_inequiv_shells)]
|
2014-11-14 09:19:21 +01:00
|
|
|
|
2016-05-09 10:19:56 +02:00
|
|
|
if dm is None:
|
|
|
|
dm = self.density_matrix(method='using_point_integration')
|
|
|
|
dens_mat = [dm[self.inequiv_to_corr[ish]]
|
|
|
|
for ish in range(self.n_inequiv_shells)]
|
2016-09-12 15:29:32 +02:00
|
|
|
if hloc is None:
|
|
|
|
hloc = self.eff_atomic_levels()
|
2016-10-03 16:56:04 +02:00
|
|
|
H_loc = [hloc[self.corr_to_inequiv[ish]]
|
|
|
|
for ish in range(self.n_corr_shells)]
|
2013-07-23 19:49:42 +02:00
|
|
|
|
2016-05-09 10:19:56 +02:00
|
|
|
if include_shells is None:
|
|
|
|
include_shells = range(self.n_inequiv_shells)
|
2014-11-15 20:04:54 +01:00
|
|
|
for ish in include_shells:
|
2014-09-22 19:24:33 +02:00
|
|
|
|
2014-11-26 16:24:02 +01:00
|
|
|
for sp in self.spin_block_names[self.corr_shells[self.inequiv_to_corr[ish]]['SO']]:
|
2015-06-26 18:47:20 +02:00
|
|
|
n_orb = self.corr_shells[self.inequiv_to_corr[ish]]['dim']
|
2016-05-09 10:19:56 +02:00
|
|
|
# gives an index list of entries larger that threshold
|
|
|
|
dmbool = (abs(dens_mat[ish][sp]) > threshold)
|
2016-09-12 15:29:32 +02:00
|
|
|
hlocbool = (abs(H_loc[ish][sp]) > threshold)
|
2013-07-23 19:49:42 +02:00
|
|
|
|
2016-05-09 10:19:56 +02:00
|
|
|
# Determine off-diagonal entries in upper triangular part of
|
|
|
|
# density matrix
|
2015-06-26 18:47:20 +02:00
|
|
|
offdiag = Set([])
|
|
|
|
for i in range(n_orb):
|
2016-05-09 10:19:56 +02:00
|
|
|
for j in range(i + 1, n_orb):
|
2016-09-12 15:29:32 +02:00
|
|
|
if dmbool[i, j] or hlocbool[i, j]:
|
2016-05-09 10:19:56 +02:00
|
|
|
offdiag.add((i, j))
|
2013-07-23 19:49:42 +02:00
|
|
|
|
2014-11-17 10:48:04 +01:00
|
|
|
# Determine the number of non-hybridising blocks in the gf
|
2016-05-09 10:19:56 +02:00
|
|
|
blocs = [[i] for i in range(n_orb)]
|
2015-06-26 18:47:20 +02:00
|
|
|
while len(offdiag) != 0:
|
|
|
|
pair = offdiag.pop()
|
2016-05-09 10:19:56 +02:00
|
|
|
for b1, b2 in product(blocs, blocs):
|
|
|
|
if (pair[0] in b1) and (pair[1] in b2):
|
|
|
|
if blocs.index(b1) != blocs.index(b2): # In separate blocks?
|
|
|
|
# Merge two blocks
|
|
|
|
b1.extend(blocs.pop(blocs.index(b2)))
|
|
|
|
break # Move on to next pair in offdiag
|
2014-11-15 20:04:54 +01:00
|
|
|
|
2014-11-17 10:48:04 +01:00
|
|
|
# Set the gf_struct for the solver accordingly
|
2015-06-26 18:47:20 +02:00
|
|
|
num_blocs = len(blocs)
|
2014-11-15 20:04:54 +01:00
|
|
|
for i in range(num_blocs):
|
2013-07-23 19:49:42 +02:00
|
|
|
blocs[i].sort()
|
2016-05-09 10:19:56 +02:00
|
|
|
self.gf_struct_solver[ish].update(
|
|
|
|
[('%s_%s' % (sp, i), range(len(blocs[i])))])
|
2014-09-22 19:24:33 +02:00
|
|
|
|
2014-11-14 09:19:21 +01:00
|
|
|
# Construct sumk_to_solver taking (sumk_block, sumk_index) --> (solver_block, solver_inner)
|
2016-05-09 10:19:56 +02:00
|
|
|
# and solver_to_sumk taking (solver_block, solver_inner) -->
|
|
|
|
# (sumk_block, sumk_index)
|
2014-11-15 20:04:54 +01:00
|
|
|
for i in range(num_blocs):
|
2014-11-14 09:19:21 +01:00
|
|
|
for j in range(len(blocs[i])):
|
2014-11-17 10:48:04 +01:00
|
|
|
block_sumk = sp
|
2014-11-14 09:19:21 +01:00
|
|
|
inner_sumk = blocs[i][j]
|
2016-05-09 10:19:56 +02:00
|
|
|
block_solv = '%s_%s' % (sp, i)
|
2014-11-14 09:19:21 +01:00
|
|
|
inner_solv = j
|
2016-05-09 10:19:56 +02:00
|
|
|
self.sumk_to_solver[ish][(block_sumk, inner_sumk)] = (
|
|
|
|
block_solv, inner_solv)
|
|
|
|
self.solver_to_sumk[ish][(block_solv, inner_solv)] = (
|
|
|
|
block_sumk, inner_sumk)
|
2014-11-15 14:16:52 +01:00
|
|
|
self.solver_to_sumk_block[ish][block_solv] = block_sumk
|
2014-09-22 19:24:33 +02:00
|
|
|
|
2014-12-03 23:12:39 +01:00
|
|
|
# Now calculate degeneracies of orbitals
|
2013-07-23 19:49:42 +02:00
|
|
|
dm = {}
|
2016-05-09 10:19:56 +02:00
|
|
|
for block, inner in self.gf_struct_solver[ish].iteritems():
|
2013-07-23 19:49:42 +02:00
|
|
|
# get dm for the blocks:
|
2016-05-09 10:19:56 +02:00
|
|
|
dm[block] = numpy.zeros(
|
|
|
|
[len(inner), len(inner)], numpy.complex_)
|
2014-11-15 20:04:54 +01:00
|
|
|
for ind1 in inner:
|
|
|
|
for ind2 in inner:
|
2016-05-09 10:19:56 +02:00
|
|
|
block_sumk, ind1_sumk = self.solver_to_sumk[
|
|
|
|
ish][(block, ind1)]
|
|
|
|
block_sumk, ind2_sumk = self.solver_to_sumk[
|
|
|
|
ish][(block, ind2)]
|
|
|
|
dm[block][ind1, ind2] = dens_mat[ish][
|
|
|
|
block_sumk][ind1_sumk, ind2_sumk]
|
2014-11-15 20:04:54 +01:00
|
|
|
|
|
|
|
for block1 in self.gf_struct_solver[ish].iterkeys():
|
2016-05-09 10:19:56 +02:00
|
|
|
for block2 in self.gf_struct_solver[ish].iterkeys():
|
2014-11-15 20:04:54 +01:00
|
|
|
if dm[block1].shape == dm[block2].shape:
|
2016-05-09 10:19:56 +02:00
|
|
|
if ((abs(dm[block1] - dm[block2]) < threshold).all()) and (block1 != block2):
|
2014-11-15 20:04:54 +01:00
|
|
|
ind1 = -1
|
|
|
|
ind2 = -2
|
2014-12-03 23:12:39 +01:00
|
|
|
# check if it was already there:
|
2016-05-09 10:19:56 +02:00
|
|
|
for n, ind in enumerate(self.deg_shells[ish]):
|
|
|
|
if block1 in ind:
|
|
|
|
ind1 = n
|
|
|
|
if block2 in ind:
|
|
|
|
ind2 = n
|
2014-11-15 20:04:54 +01:00
|
|
|
if (ind1 < 0) and (ind2 >= 0):
|
|
|
|
self.deg_shells[ish][ind2].append(block1)
|
|
|
|
elif (ind1 >= 0) and (ind2 < 0):
|
|
|
|
self.deg_shells[ish][ind1].append(block2)
|
|
|
|
elif (ind1 < 0) and (ind2 < 0):
|
2016-05-09 10:19:56 +02:00
|
|
|
self.deg_shells[ish].append([block1, block2])
|
2013-07-23 19:49:42 +02:00
|
|
|
|
2016-05-09 10:19:56 +02:00
|
|
|
def density_matrix(self, method='using_gf', beta=40.0):
|
2015-03-12 00:01:12 +01:00
|
|
|
"""Calculate density matrices in one of two ways.
|
|
|
|
|
|
|
|
Parameters
|
|
|
|
----------
|
|
|
|
method : string, optional
|
|
|
|
|
|
|
|
- if 'using_gf': First get lattice gf (g_loc is not set up), then density matrix.
|
|
|
|
It is useful for Hubbard I, and very quick.
|
|
|
|
No assumption on the hopping structure is made (ie diagonal or not).
|
|
|
|
- if 'using_point_integration': Only works for diagonal hopping matrix (true in wien2k).
|
|
|
|
|
|
|
|
beta : float, optional
|
|
|
|
Inverse temperature.
|
|
|
|
|
|
|
|
Returns
|
|
|
|
-------
|
|
|
|
dens_mat : list of dicts
|
|
|
|
Density matrix for each spin in each correlated shell.
|
2014-11-17 10:48:04 +01:00
|
|
|
"""
|
2016-05-09 10:19:56 +02:00
|
|
|
dens_mat = [{} for icrsh in range(self.n_corr_shells)]
|
2014-11-17 10:48:04 +01:00
|
|
|
for icrsh in range(self.n_corr_shells):
|
2014-11-26 16:24:02 +01:00
|
|
|
for sp in self.spin_block_names[self.corr_shells[icrsh]['SO']]:
|
2016-05-09 10:19:56 +02:00
|
|
|
dens_mat[icrsh][sp] = numpy.zeros(
|
|
|
|
[self.corr_shells[icrsh]['dim'], self.corr_shells[icrsh]['dim']], numpy.complex_)
|
2014-11-17 10:48:04 +01:00
|
|
|
|
|
|
|
ikarray = numpy.array(range(self.n_k))
|
|
|
|
for ik in mpi.slice_array(ikarray):
|
|
|
|
|
|
|
|
if method == "using_gf":
|
|
|
|
|
2016-05-09 10:19:56 +02:00
|
|
|
G_latt_iw = self.lattice_gf(
|
|
|
|
ik=ik, mu=self.chemical_potential, iw_or_w="iw", beta=beta)
|
2014-12-07 00:29:39 +01:00
|
|
|
G_latt_iw *= self.bz_weights[ik]
|
|
|
|
dm = G_latt_iw.density()
|
2014-11-17 10:48:04 +01:00
|
|
|
MMat = [dm[sp] for sp in self.spin_block_names[self.SO]]
|
|
|
|
|
|
|
|
elif method == "using_point_integration":
|
|
|
|
|
|
|
|
ntoi = self.spin_names_to_ind[self.SO]
|
|
|
|
spn = self.spin_block_names[self.SO]
|
2016-07-19 16:56:52 +02:00
|
|
|
dims = {sp:self.n_orbitals[ik, ntoi[sp]] for sp in spn}
|
|
|
|
MMat = [numpy.zeros([dims[sp], dims[sp]], numpy.complex_) for sp in spn]
|
2014-11-17 10:48:04 +01:00
|
|
|
|
|
|
|
for isp, sp in enumerate(spn):
|
|
|
|
ind = ntoi[sp]
|
2016-05-09 10:19:56 +02:00
|
|
|
for inu in range(self.n_orbitals[ik, ind]):
|
|
|
|
# only works for diagonal hopping matrix (true in
|
|
|
|
# wien2k)
|
|
|
|
if (self.hopping[ik, ind, inu, inu] - self.h_field * (1 - 2 * isp)) < 0.0:
|
|
|
|
MMat[isp][inu, inu] = 1.0
|
2014-11-17 10:48:04 +01:00
|
|
|
else:
|
2016-05-09 10:19:56 +02:00
|
|
|
MMat[isp][inu, inu] = 0.0
|
2014-11-17 10:48:04 +01:00
|
|
|
|
2016-05-09 10:19:56 +02:00
|
|
|
else:
|
|
|
|
raise ValueError, "density_matrix: the method '%s' is not supported." % method
|
2014-12-03 23:12:39 +01:00
|
|
|
|
2014-11-17 10:48:04 +01:00
|
|
|
for icrsh in range(self.n_corr_shells):
|
2014-11-26 16:24:02 +01:00
|
|
|
for isp, sp in enumerate(self.spin_block_names[self.corr_shells[icrsh]['SO']]):
|
2016-05-09 10:19:56 +02:00
|
|
|
ind = self.spin_names_to_ind[
|
|
|
|
self.corr_shells[icrsh]['SO']][sp]
|
2014-11-26 16:24:02 +01:00
|
|
|
dim = self.corr_shells[icrsh]['dim']
|
2016-05-09 10:19:56 +02:00
|
|
|
n_orb = self.n_orbitals[ik, ind]
|
|
|
|
projmat = self.proj_mat[ik, ind, icrsh, 0:dim, 0:n_orb]
|
2014-11-17 10:48:04 +01:00
|
|
|
if method == "using_gf":
|
2016-05-09 10:19:56 +02:00
|
|
|
dens_mat[icrsh][sp] += numpy.dot(numpy.dot(projmat, MMat[isp]),
|
|
|
|
projmat.transpose().conjugate())
|
2014-11-17 10:48:04 +01:00
|
|
|
elif method == "using_point_integration":
|
2016-05-09 10:19:56 +02:00
|
|
|
dens_mat[icrsh][sp] += self.bz_weights[ik] * numpy.dot(numpy.dot(projmat, MMat[isp]),
|
|
|
|
projmat.transpose().conjugate())
|
2014-11-17 10:48:04 +01:00
|
|
|
|
|
|
|
# get data from nodes:
|
|
|
|
for icrsh in range(self.n_corr_shells):
|
2014-12-07 00:29:39 +01:00
|
|
|
for sp in dens_mat[icrsh]:
|
2016-05-09 10:19:56 +02:00
|
|
|
dens_mat[icrsh][sp] = mpi.all_reduce(
|
|
|
|
mpi.world, dens_mat[icrsh][sp], lambda x, y: x + y)
|
2014-11-17 10:48:04 +01:00
|
|
|
mpi.barrier()
|
|
|
|
|
2016-05-09 10:19:56 +02:00
|
|
|
if self.symm_op != 0:
|
|
|
|
dens_mat = self.symmcorr.symmetrize(dens_mat)
|
2014-11-17 10:48:04 +01:00
|
|
|
|
|
|
|
# Rotate to local coordinate system:
|
|
|
|
if self.use_rotations:
|
|
|
|
for icrsh in range(self.n_corr_shells):
|
2014-12-07 00:29:39 +01:00
|
|
|
for sp in dens_mat[icrsh]:
|
2016-05-09 10:19:56 +02:00
|
|
|
if self.rot_mat_time_inv[icrsh] == 1:
|
|
|
|
dens_mat[icrsh][sp] = dens_mat[icrsh][sp].conjugate()
|
|
|
|
dens_mat[icrsh][sp] = numpy.dot(numpy.dot(self.rot_mat[icrsh].conjugate().transpose(), dens_mat[icrsh][sp]),
|
|
|
|
self.rot_mat[icrsh])
|
2014-11-17 10:48:04 +01:00
|
|
|
|
|
|
|
return dens_mat
|
2014-09-22 19:24:33 +02:00
|
|
|
|
2014-11-16 18:02:04 +01:00
|
|
|
# For simple dft input, get crystal field splittings.
|
2013-07-23 19:49:42 +02:00
|
|
|
def eff_atomic_levels(self):
|
2015-03-12 00:01:12 +01:00
|
|
|
r"""
|
|
|
|
Calculates the effective local Hamiltonian required as an input for
|
|
|
|
the Hubbard I Solver.
|
|
|
|
The local Hamiltonian (effective atomic levels) is calculated by
|
|
|
|
projecting the on-site Bloch Hamiltonian:
|
|
|
|
|
|
|
|
.. math:: H^{loc}_{m m'} = \sum_{k} P_{m \nu}(k) H_{\nu\nu'}(k) P^{*}_{\nu' m'}(k),
|
|
|
|
|
|
|
|
where
|
|
|
|
|
|
|
|
.. math:: H_{\nu\nu'}(k) = [\epsilon_{\nu k} - h_{z} \sigma_{z}] \delta_{\nu\nu'}.
|
|
|
|
|
|
|
|
Parameters
|
|
|
|
----------
|
|
|
|
None
|
|
|
|
|
|
|
|
Returns
|
|
|
|
-------
|
2016-05-11 17:00:52 +02:00
|
|
|
eff_atlevels : gf_struct_sumk like
|
2016-10-03 16:56:04 +02:00
|
|
|
Effective local Hamiltonian :math:`H^{loc}_{m m'}` for each
|
|
|
|
inequivalent correlated shell.
|
2015-03-12 00:01:12 +01:00
|
|
|
|
|
|
|
"""
|
2013-07-23 19:49:42 +02:00
|
|
|
|
|
|
|
# define matrices for inequivalent shells:
|
2016-05-09 10:19:56 +02:00
|
|
|
eff_atlevels = [{} for ish in range(self.n_inequiv_shells)]
|
2014-11-15 18:15:45 +01:00
|
|
|
for ish in range(self.n_inequiv_shells):
|
2014-11-26 16:24:02 +01:00
|
|
|
for sp in self.spin_block_names[self.corr_shells[self.inequiv_to_corr[ish]]['SO']]:
|
2016-05-09 10:19:56 +02:00
|
|
|
eff_atlevels[ish][sp] = numpy.identity(
|
|
|
|
self.corr_shells[self.inequiv_to_corr[ish]]['dim'], numpy.complex_)
|
2014-12-07 00:29:39 +01:00
|
|
|
eff_atlevels[ish][sp] *= -self.chemical_potential
|
2016-05-09 10:19:56 +02:00
|
|
|
eff_atlevels[ish][
|
|
|
|
sp] -= self.dc_imp[self.inequiv_to_corr[ish]][sp]
|
2013-07-23 19:49:42 +02:00
|
|
|
|
|
|
|
# sum over k:
|
2016-05-09 10:19:56 +02:00
|
|
|
if not hasattr(self, "Hsumk"):
|
|
|
|
# calculate the sum over k. Does not depend on mu, so do it only
|
|
|
|
# once:
|
|
|
|
self.Hsumk = [{} for icrsh in range(self.n_corr_shells)]
|
2013-07-23 19:49:42 +02:00
|
|
|
for icrsh in range(self.n_corr_shells):
|
2014-12-07 00:29:39 +01:00
|
|
|
dim = self.corr_shells[icrsh]['dim']
|
2014-11-26 16:24:02 +01:00
|
|
|
for sp in self.spin_block_names[self.corr_shells[icrsh]['SO']]:
|
2016-05-09 10:19:56 +02:00
|
|
|
self.Hsumk[icrsh][sp] = numpy.zeros(
|
|
|
|
[dim, dim], numpy.complex_)
|
2014-11-26 16:24:02 +01:00
|
|
|
for isp, sp in enumerate(self.spin_block_names[self.corr_shells[icrsh]['SO']]):
|
2016-05-09 10:19:56 +02:00
|
|
|
ind = self.spin_names_to_ind[
|
|
|
|
self.corr_shells[icrsh]['SO']][sp]
|
2014-11-15 20:04:54 +01:00
|
|
|
for ik in range(self.n_k):
|
2016-05-09 10:19:56 +02:00
|
|
|
n_orb = self.n_orbitals[ik, ind]
|
2013-07-23 19:49:42 +02:00
|
|
|
MMat = numpy.identity(n_orb, numpy.complex_)
|
2016-05-09 10:19:56 +02:00
|
|
|
MMat = self.hopping[
|
|
|
|
ik, ind, 0:n_orb, 0:n_orb] - (1 - 2 * isp) * self.h_field * MMat
|
|
|
|
projmat = self.proj_mat[ik, ind, icrsh, 0:dim, 0:n_orb]
|
|
|
|
self.Hsumk[icrsh][sp] += self.bz_weights[ik] * numpy.dot(numpy.dot(projmat, MMat),
|
|
|
|
projmat.conjugate().transpose())
|
2013-07-23 19:49:42 +02:00
|
|
|
# symmetrisation:
|
2016-05-09 10:19:56 +02:00
|
|
|
if self.symm_op != 0:
|
|
|
|
self.Hsumk = self.symmcorr.symmetrize(self.Hsumk)
|
2013-07-23 19:49:42 +02:00
|
|
|
|
|
|
|
# Rotate to local coordinate system:
|
2014-11-15 20:04:54 +01:00
|
|
|
if self.use_rotations:
|
|
|
|
for icrsh in range(self.n_corr_shells):
|
2014-12-07 00:29:39 +01:00
|
|
|
for sp in self.Hsumk[icrsh]:
|
2016-05-09 10:19:56 +02:00
|
|
|
if self.rot_mat_time_inv[icrsh] == 1:
|
|
|
|
self.Hsumk[icrsh][sp] = self.Hsumk[
|
|
|
|
icrsh][sp].conjugate()
|
|
|
|
self.Hsumk[icrsh][sp] = numpy.dot(numpy.dot(self.rot_mat[icrsh].conjugate().transpose(), self.Hsumk[icrsh][sp]),
|
|
|
|
self.rot_mat[icrsh])
|
2014-09-22 19:24:33 +02:00
|
|
|
|
2013-07-23 19:49:42 +02:00
|
|
|
# add to matrix:
|
2014-11-15 20:04:54 +01:00
|
|
|
for ish in range(self.n_inequiv_shells):
|
2014-12-07 00:29:39 +01:00
|
|
|
for sp in eff_atlevels[ish]:
|
2016-05-09 10:19:56 +02:00
|
|
|
eff_atlevels[ish][
|
|
|
|
sp] += self.Hsumk[self.inequiv_to_corr[ish]][sp]
|
2013-07-23 19:49:42 +02:00
|
|
|
|
|
|
|
return eff_atlevels
|
|
|
|
|
2014-12-03 23:12:39 +01:00
|
|
|
def init_dc(self):
|
2015-03-12 00:01:12 +01:00
|
|
|
r"""
|
|
|
|
Initializes the double counting terms.
|
2014-12-07 00:29:39 +01:00
|
|
|
|
2015-03-12 00:01:12 +01:00
|
|
|
Parameters
|
|
|
|
----------
|
|
|
|
None
|
|
|
|
|
|
|
|
"""
|
2016-05-09 10:19:56 +02:00
|
|
|
self.dc_imp = [{} for icrsh in range(self.n_corr_shells)]
|
2014-11-15 20:04:54 +01:00
|
|
|
for icrsh in range(self.n_corr_shells):
|
2014-11-26 16:24:02 +01:00
|
|
|
dim = self.corr_shells[icrsh]['dim']
|
|
|
|
spn = self.spin_block_names[self.corr_shells[icrsh]['SO']]
|
2016-05-09 10:19:56 +02:00
|
|
|
for sp in spn:
|
|
|
|
self.dc_imp[icrsh][sp] = numpy.zeros([dim, dim], numpy.float_)
|
2014-11-15 20:04:54 +01:00
|
|
|
self.dc_energ = [0.0 for icrsh in range(self.n_corr_shells)]
|
2013-07-23 19:49:42 +02:00
|
|
|
|
2016-05-09 10:19:56 +02:00
|
|
|
def set_dc(self, dc_imp, dc_energ):
|
2015-03-12 00:01:12 +01:00
|
|
|
r"""
|
|
|
|
Sets double counting corrections to given values.
|
2016-05-09 10:19:56 +02:00
|
|
|
|
2015-03-12 00:01:12 +01:00
|
|
|
Parameters
|
|
|
|
----------
|
|
|
|
dc_imp : gf_struct_sumk like
|
|
|
|
Double-counting self-energy term.
|
|
|
|
dc_energ : list of floats
|
|
|
|
Double-counting energy corrections for each correlated shell.
|
2016-05-09 10:19:56 +02:00
|
|
|
|
2015-03-12 00:01:12 +01:00
|
|
|
"""
|
2014-12-03 23:12:39 +01:00
|
|
|
|
|
|
|
self.dc_imp = dc_imp
|
|
|
|
self.dc_energ = dc_energ
|
|
|
|
|
2016-05-09 10:19:56 +02:00
|
|
|
def calc_dc(self, dens_mat, orb=0, U_interact=None, J_hund=None, use_dc_formula=0, use_dc_value=None):
|
2015-03-12 00:01:12 +01:00
|
|
|
r"""
|
|
|
|
Calculates and sets the double counting corrections.
|
|
|
|
|
|
|
|
If 'use_dc_value' is provided the double-counting term is uniformly initialized
|
|
|
|
with this constant and 'U_interact' and 'J_hund' are ignored.
|
|
|
|
|
|
|
|
If 'use_dc_value' is None the correction is evaluated according to
|
|
|
|
one of the following formulae:
|
|
|
|
|
|
|
|
* use_dc_formula = 0: fully-localised limit (FLL)
|
|
|
|
* use_dc_formula = 1: Held's formula, i.e. mean-field formula for the Kanamori
|
|
|
|
type of the interaction Hamiltonian
|
|
|
|
* use_dc_formula = 2: around mean-field (AMF)
|
|
|
|
|
|
|
|
Note that FLL and AMF formulae were derived assuming a full Slater-type interaction
|
|
|
|
term and should be thus used accordingly. For the Kanamori-type interaction
|
|
|
|
one should use formula 1.
|
|
|
|
|
|
|
|
The double-counting self-energy term is stored in `self.dc_imp` and the energy
|
|
|
|
correction in `self.dc_energ`.
|
|
|
|
|
|
|
|
Parameters
|
|
|
|
----------
|
|
|
|
dens_mat : gf_struct_solver like
|
|
|
|
Density matrix for the specified correlated shell.
|
|
|
|
orb : int, optional
|
|
|
|
Index of an inequivalent shell.
|
|
|
|
U_interact : float, optional
|
|
|
|
Value of interaction parameter `U`.
|
|
|
|
J_hund : float, optional
|
|
|
|
Value of interaction parameter `J`.
|
|
|
|
use_dc_formula : int, optional
|
|
|
|
Type of double-counting correction (see description).
|
|
|
|
use_dc_value : float, optional
|
|
|
|
Value of the double-counting correction. If specified
|
|
|
|
`U_interact`, `J_hund` and `use_dc_formula` are ignored.
|
|
|
|
|
|
|
|
"""
|
2014-01-21 16:25:44 +01:00
|
|
|
|
2014-11-15 20:04:54 +01:00
|
|
|
for icrsh in range(self.n_corr_shells):
|
2013-07-23 19:49:42 +02:00
|
|
|
|
2016-05-09 10:19:56 +02:00
|
|
|
# ish is the index of the inequivalent shell corresponding to icrsh
|
|
|
|
ish = self.corr_to_inequiv[icrsh]
|
|
|
|
if ish != orb:
|
|
|
|
continue # ignore this orbital
|
|
|
|
# *(1+self.corr_shells[icrsh]['SO'])
|
|
|
|
dim = self.corr_shells[icrsh]['dim']
|
2014-11-26 16:24:02 +01:00
|
|
|
spn = self.spin_block_names[self.corr_shells[icrsh]['SO']]
|
2014-01-21 16:25:44 +01:00
|
|
|
|
2016-05-09 10:19:56 +02:00
|
|
|
Ncr = {sp: 0.0 for sp in spn}
|
|
|
|
for block, inner in self.gf_struct_solver[ish].iteritems():
|
2014-12-03 23:12:39 +01:00
|
|
|
bl = self.solver_to_sumk_block[ish][block]
|
2014-11-15 20:04:54 +01:00
|
|
|
Ncr[bl] += dens_mat[block].real.trace()
|
2014-12-03 23:12:39 +01:00
|
|
|
Ncrtot = sum(Ncr.itervalues())
|
2016-05-09 10:19:56 +02:00
|
|
|
for sp in spn:
|
|
|
|
self.dc_imp[icrsh][sp] = numpy.identity(dim, numpy.float_)
|
|
|
|
if self.SP == 0: # average the densities if there is no SP:
|
2014-12-03 23:12:39 +01:00
|
|
|
Ncr[sp] = Ncrtot / len(spn)
|
2016-05-09 10:19:56 +02:00
|
|
|
# correction for SO: we have only one block in this case, but
|
|
|
|
# in DC we need N/2
|
|
|
|
elif self.SP == 1 and self.SO == 1:
|
2014-12-03 23:12:39 +01:00
|
|
|
Ncr[sp] = Ncrtot / 2.0
|
2013-07-23 19:49:42 +02:00
|
|
|
|
2014-12-13 14:14:01 +01:00
|
|
|
if use_dc_value is None:
|
2014-09-22 19:24:33 +02:00
|
|
|
|
2016-05-09 10:19:56 +02:00
|
|
|
if U_interact is None and J_hund is None:
|
|
|
|
raise ValueError, "set_dc: either provide U_interact and J_hund or set use_dc_value to dc value."
|
2014-12-03 23:12:39 +01:00
|
|
|
|
2016-05-09 10:19:56 +02:00
|
|
|
if use_dc_formula == 0: # FLL
|
2014-10-28 16:19:28 +01:00
|
|
|
|
2016-05-09 10:19:56 +02:00
|
|
|
self.dc_energ[icrsh] = U_interact / \
|
|
|
|
2.0 * Ncrtot * (Ncrtot - 1.0)
|
2014-11-17 10:48:04 +01:00
|
|
|
for sp in spn:
|
2016-05-09 10:19:56 +02:00
|
|
|
Uav = U_interact * (Ncrtot - 0.5) - \
|
|
|
|
J_hund * (Ncr[sp] - 0.5)
|
2014-11-17 10:48:04 +01:00
|
|
|
self.dc_imp[icrsh][sp] *= Uav
|
2016-05-09 10:19:56 +02:00
|
|
|
self.dc_energ[icrsh] -= J_hund / \
|
|
|
|
2.0 * (Ncr[sp]) * (Ncr[sp] - 1.0)
|
|
|
|
mpi.report(
|
|
|
|
"DC for shell %(icrsh)i and block %(sp)s = %(Uav)f" % locals())
|
2014-10-28 16:19:28 +01:00
|
|
|
|
2016-05-09 10:19:56 +02:00
|
|
|
elif use_dc_formula == 1: # Held's formula, with U_interact the interorbital onsite interaction
|
2014-10-28 16:19:28 +01:00
|
|
|
|
2016-05-09 10:19:56 +02:00
|
|
|
self.dc_energ[icrsh] = (U_interact + (dim - 1) * (U_interact - 2.0 * J_hund) + (
|
|
|
|
dim - 1) * (U_interact - 3.0 * J_hund)) / (2 * dim - 1) / 2.0 * Ncrtot * (Ncrtot - 1.0)
|
2014-11-17 10:48:04 +01:00
|
|
|
for sp in spn:
|
2016-05-09 10:19:56 +02:00
|
|
|
Uav = (U_interact + (dim - 1) * (U_interact - 2.0 * J_hund) + (dim - 1)
|
|
|
|
* (U_interact - 3.0 * J_hund)) / (2 * dim - 1) * (Ncrtot - 0.5)
|
2014-11-17 10:48:04 +01:00
|
|
|
self.dc_imp[icrsh][sp] *= Uav
|
2016-05-09 10:19:56 +02:00
|
|
|
mpi.report(
|
|
|
|
"DC for shell %(icrsh)i and block %(sp)s = %(Uav)f" % locals())
|
2014-10-28 16:19:28 +01:00
|
|
|
|
2016-05-09 10:19:56 +02:00
|
|
|
elif use_dc_formula == 2: # AMF
|
2014-10-28 16:19:28 +01:00
|
|
|
|
2014-11-15 20:04:54 +01:00
|
|
|
self.dc_energ[icrsh] = 0.5 * U_interact * Ncrtot * Ncrtot
|
2014-11-17 10:48:04 +01:00
|
|
|
for sp in spn:
|
2016-05-09 10:19:56 +02:00
|
|
|
Uav = U_interact * \
|
|
|
|
(Ncrtot - Ncr[sp] / dim) - \
|
|
|
|
J_hund * (Ncr[sp] - Ncr[sp] / dim)
|
2014-11-17 10:48:04 +01:00
|
|
|
self.dc_imp[icrsh][sp] *= Uav
|
2016-05-09 10:19:56 +02:00
|
|
|
self.dc_energ[
|
|
|
|
icrsh] -= (U_interact + (dim - 1) * J_hund) / dim * 0.5 * Ncr[sp] * Ncr[sp]
|
|
|
|
mpi.report(
|
|
|
|
"DC for shell %(icrsh)i and block %(sp)s = %(Uav)f" % locals())
|
2014-09-22 19:24:33 +02:00
|
|
|
|
2016-05-09 10:19:56 +02:00
|
|
|
mpi.report("DC energy for shell %s = %s" %
|
|
|
|
(icrsh, self.dc_energ[icrsh]))
|
2013-07-23 19:49:42 +02:00
|
|
|
|
2016-05-09 10:19:56 +02:00
|
|
|
else: # use value provided for user to determine dc_energ and dc_imp
|
2014-09-22 19:24:33 +02:00
|
|
|
|
2014-12-13 14:14:01 +01:00
|
|
|
self.dc_energ[icrsh] = use_dc_value * Ncrtot
|
2016-05-09 10:19:56 +02:00
|
|
|
for sp in spn:
|
|
|
|
self.dc_imp[icrsh][sp] *= use_dc_value
|
2013-07-23 19:49:42 +02:00
|
|
|
|
2016-05-09 10:19:56 +02:00
|
|
|
mpi.report(
|
|
|
|
"DC for shell %(icrsh)i = %(use_dc_value)f" % locals())
|
|
|
|
mpi.report("DC energy = %s" % self.dc_energ[icrsh])
|
2013-07-23 19:49:42 +02:00
|
|
|
|
2016-05-09 10:19:56 +02:00
|
|
|
def add_dc(self, iw_or_w="iw"):
|
2015-03-12 00:01:12 +01:00
|
|
|
r"""
|
|
|
|
Subtracts the double counting term from the impurity self energy.
|
2016-05-09 10:19:56 +02:00
|
|
|
|
2015-03-12 00:01:12 +01:00
|
|
|
Parameters
|
|
|
|
----------
|
|
|
|
iw_or_w : string, optional
|
|
|
|
|
|
|
|
- `iw_or_w` = 'iw' for a imaginary-frequency self-energy
|
|
|
|
- `iw_or_w` = 'w' for a real-frequency self-energy
|
|
|
|
|
|
|
|
Returns
|
|
|
|
-------
|
|
|
|
sigma_minus_dc : gf_struct_sumk like
|
|
|
|
Self-energy with a subtracted double-counting term.
|
|
|
|
|
|
|
|
"""
|
2014-09-22 19:24:33 +02:00
|
|
|
|
2013-07-23 19:49:42 +02:00
|
|
|
# Be careful: Sigma_imp is already in the global coordinate system!!
|
2016-05-09 10:19:56 +02:00
|
|
|
sigma_minus_dc = [s.copy()
|
|
|
|
for s in getattr(self, "Sigma_imp_" + iw_or_w)]
|
2014-11-15 20:04:54 +01:00
|
|
|
for icrsh in range(self.n_corr_shells):
|
2016-05-09 10:19:56 +02:00
|
|
|
for bname, gf in sigma_minus_dc[icrsh]:
|
2014-10-28 16:19:28 +01:00
|
|
|
# Transform dc_imp to global coordinate system
|
2016-05-09 10:19:56 +02:00
|
|
|
dccont = numpy.dot(self.rot_mat[icrsh], numpy.dot(self.dc_imp[icrsh][
|
|
|
|
bname], self.rot_mat[icrsh].conjugate().transpose()))
|
2014-12-07 00:29:39 +01:00
|
|
|
sigma_minus_dc[icrsh][bname] -= dccont
|
2013-07-23 19:49:42 +02:00
|
|
|
|
2014-12-07 00:29:39 +01:00
|
|
|
return sigma_minus_dc
|
2014-09-22 19:24:33 +02:00
|
|
|
|
2016-05-09 10:19:56 +02:00
|
|
|
def symm_deg_gf(self, gf_to_symm, orb):
|
2015-03-12 00:01:12 +01:00
|
|
|
r"""
|
|
|
|
Averages a GF over degenerate shells.
|
|
|
|
|
|
|
|
Degenerate shells of an inequivalent correlated shell are defined by
|
|
|
|
`self.deg_shells`. This function enforces corresponding degeneracies
|
|
|
|
in the input GF.
|
|
|
|
|
|
|
|
Parameters
|
|
|
|
----------
|
|
|
|
gf_to_symm : gf_struct_solver like
|
|
|
|
Input GF.
|
|
|
|
orb : int
|
|
|
|
Index of an inequivalent shell.
|
|
|
|
|
|
|
|
"""
|
2013-07-23 19:49:42 +02:00
|
|
|
|
2014-11-17 10:48:04 +01:00
|
|
|
for degsh in self.deg_shells[orb]:
|
|
|
|
ss = gf_to_symm[degsh[0]].copy()
|
|
|
|
ss.zero()
|
|
|
|
n_deg = len(degsh)
|
2016-05-09 10:19:56 +02:00
|
|
|
for bl in degsh:
|
|
|
|
ss += gf_to_symm[bl] / (1.0 * n_deg)
|
|
|
|
for bl in degsh:
|
|
|
|
gf_to_symm[bl] << ss
|
2013-07-23 19:49:42 +02:00
|
|
|
|
2016-04-20 19:01:29 +02:00
|
|
|
def total_density(self, mu=None, iw_or_w="iw", with_Sigma=True, with_dc=True, broadening=None):
|
2015-03-12 00:01:12 +01:00
|
|
|
r"""
|
|
|
|
Calculates the total charge within the energy window for a given chemical potential.
|
|
|
|
The chemical potential is either given by parameter `mu` or, if it is not specified,
|
|
|
|
taken from `self.chemical_potential`.
|
|
|
|
|
|
|
|
The total charge is calculated from the trace of the GF in the Bloch basis.
|
2016-04-20 19:01:29 +02:00
|
|
|
By default, a full interacting GF is used. To use the non-interacting GF, set
|
2015-03-12 00:01:12 +01:00
|
|
|
parameter `with_Sigma = False`.
|
|
|
|
|
|
|
|
The number of bands within the energy windows generally depends on `k`. The trace is
|
|
|
|
therefore calculated separately for each `k`-point.
|
|
|
|
|
2014-11-15 20:04:54 +01:00
|
|
|
Since in general n_orbitals depends on k, the calculation is done in the following order:
|
2016-07-08 12:04:31 +02:00
|
|
|
|
|
|
|
.. math:: n_{tot} = \sum_{k} n(k),
|
|
|
|
|
|
|
|
with
|
|
|
|
|
|
|
|
.. math:: n(k) = Tr G_{\nu\nu'}(k, i\omega_{n}).
|
2015-03-12 00:01:12 +01:00
|
|
|
|
|
|
|
The calculation is done in the global coordinate system, if distinction is made between local/global.
|
|
|
|
|
|
|
|
Parameters
|
|
|
|
----------
|
|
|
|
mu : float, optional
|
|
|
|
Input chemical potential. If not specified, `self.chemical_potential` is used instead.
|
2016-04-20 19:01:29 +02:00
|
|
|
iw_or_w : string, optional
|
|
|
|
- `iw_or_w` = 'iw' for a imaginary-frequency self-energy
|
|
|
|
- `iw_or_w` = 'w' for a real-frequency self-energy
|
2015-03-12 00:01:12 +01:00
|
|
|
with_Sigma : boolean, optional
|
|
|
|
If `True` the full interacing GF is evaluated, otherwise the self-energy is not
|
|
|
|
included and the charge would correspond to a non-interacting system.
|
|
|
|
with_dc : boolean, optional
|
|
|
|
Whether or not to subtract the double-counting term from the self-energy.
|
2016-04-20 19:01:29 +02:00
|
|
|
broadening : float, optional
|
|
|
|
Imaginary shift for the axis along which the real-axis GF is calculated.
|
|
|
|
If not provided, broadening will be set to double of the distance between mesh points in 'mesh'.
|
|
|
|
Only relevant for real-frequency GF.
|
2015-03-12 00:01:12 +01:00
|
|
|
|
|
|
|
Returns
|
|
|
|
-------
|
|
|
|
dens : float
|
|
|
|
Total charge :math:`n_{tot}`.
|
2014-09-22 19:24:33 +02:00
|
|
|
|
2013-07-23 19:49:42 +02:00
|
|
|
"""
|
2016-05-09 10:23:57 +02:00
|
|
|
|
2016-05-09 10:19:56 +02:00
|
|
|
if mu is None:
|
|
|
|
mu = self.chemical_potential
|
2013-07-23 19:49:42 +02:00
|
|
|
dens = 0.0
|
2014-11-17 10:48:04 +01:00
|
|
|
ikarray = numpy.array(range(self.n_k))
|
2013-07-23 19:49:42 +02:00
|
|
|
for ik in mpi.slice_array(ikarray):
|
2016-05-09 10:19:56 +02:00
|
|
|
G_latt = self.lattice_gf(
|
|
|
|
ik=ik, mu=mu, iw_or_w=iw_or_w, with_Sigma=with_Sigma, with_dc=with_dc, broadening=broadening)
|
2016-04-20 19:01:29 +02:00
|
|
|
dens += self.bz_weights[ik] * G_latt.total_density()
|
2013-07-23 19:49:42 +02:00
|
|
|
# collect data from mpi:
|
2016-05-09 10:19:56 +02:00
|
|
|
dens = mpi.all_reduce(mpi.world, dens, lambda x, y: x + y)
|
2013-07-23 19:49:42 +02:00
|
|
|
mpi.barrier()
|
2014-09-22 19:24:33 +02:00
|
|
|
|
2013-07-23 19:49:42 +02:00
|
|
|
return dens
|
|
|
|
|
2016-05-09 10:19:56 +02:00
|
|
|
def set_mu(self, mu):
|
2015-03-12 00:01:12 +01:00
|
|
|
r"""
|
|
|
|
Sets a new chemical potential.
|
|
|
|
|
|
|
|
Parameters
|
|
|
|
----------
|
|
|
|
mu : float
|
|
|
|
New value of the chemical potential.
|
2014-11-17 10:48:04 +01:00
|
|
|
|
2015-03-12 00:01:12 +01:00
|
|
|
"""
|
2014-11-17 10:48:04 +01:00
|
|
|
self.chemical_potential = mu
|
|
|
|
|
2016-11-03 11:57:20 +01:00
|
|
|
def calc_mu(self, precision=0.01, iw_or_w='iw', broadening=None, delta=0.5):
|
2015-03-12 00:01:12 +01:00
|
|
|
r"""
|
|
|
|
Searches for the chemical potential that gives the DFT total charge.
|
|
|
|
A simple bisection method is used.
|
|
|
|
|
|
|
|
Parameters
|
|
|
|
----------
|
|
|
|
precision : float, optional
|
|
|
|
A desired precision of the resulting total charge.
|
2016-04-20 19:01:29 +02:00
|
|
|
iw_or_w : string, optional
|
|
|
|
- `iw_or_w` = 'iw' for a imaginary-frequency self-energy
|
|
|
|
- `iw_or_w` = 'w' for a real-frequency self-energy
|
|
|
|
broadening : float, optional
|
|
|
|
Imaginary shift for the axis along which the real-axis GF is calculated.
|
|
|
|
If not provided, broadening will be set to double of the distance between mesh points in 'mesh'.
|
|
|
|
Only relevant for real-frequency GF.
|
2015-03-12 00:01:12 +01:00
|
|
|
|
|
|
|
Returns
|
|
|
|
-------
|
|
|
|
mu : float
|
|
|
|
Value of the chemical potential giving the DFT total charge
|
|
|
|
within specified precision.
|
2013-07-23 19:49:42 +02:00
|
|
|
|
2015-03-12 00:01:12 +01:00
|
|
|
"""
|
2016-05-09 10:19:56 +02:00
|
|
|
F = lambda mu: self.total_density(
|
|
|
|
mu=mu, iw_or_w=iw_or_w, broadening=broadening)
|
2014-11-07 00:55:40 +01:00
|
|
|
density = self.density_required - self.charge_below
|
2013-07-23 19:49:42 +02:00
|
|
|
|
2016-05-09 10:19:56 +02:00
|
|
|
self.chemical_potential = dichotomy.dichotomy(function=F,
|
|
|
|
x_init=self.chemical_potential, y_value=density,
|
2016-11-03 11:57:20 +01:00
|
|
|
precision_on_y=precision, delta_x=delta, max_loops=100,
|
2016-05-09 10:19:56 +02:00
|
|
|
x_name="Chemical Potential", y_name="Total Density",
|
|
|
|
verbosity=3)[0]
|
2013-07-23 19:49:42 +02:00
|
|
|
|
|
|
|
return self.chemical_potential
|
|
|
|
|
2016-03-16 16:18:52 +01:00
|
|
|
def calc_density_correction(self, filename=None, dm_type='wien2k'):
|
2015-03-12 00:01:12 +01:00
|
|
|
r"""
|
|
|
|
Calculates the charge density correction and stores it into a file.
|
2016-05-09 10:19:56 +02:00
|
|
|
|
2015-03-12 00:01:12 +01:00
|
|
|
The charge density correction is needed for charge-self-consistent DFT+DMFT calculations.
|
|
|
|
It represents a density matrix of the interacting system defined in Bloch basis
|
|
|
|
and it is calculated from the sum over Matsubara frequecies of the full GF,
|
|
|
|
|
|
|
|
..math:: N_{\nu\nu'}(k) = \sum_{i\omega_{n}} G_{\nu\nu'}(k, i\omega_{n})
|
2016-05-09 10:19:56 +02:00
|
|
|
|
2015-03-12 00:01:12 +01:00
|
|
|
The density matrix for every `k`-point is stored into a file.
|
|
|
|
|
|
|
|
Parameters
|
|
|
|
----------
|
|
|
|
filename : string
|
|
|
|
Name of the file to store the charge density correction.
|
|
|
|
|
|
|
|
Returns
|
|
|
|
-------
|
|
|
|
(deltaN, dens) : tuple
|
|
|
|
Returns a tuple containing the density matrix `deltaN` and
|
|
|
|
the corresponing total charge `dens`.
|
|
|
|
|
|
|
|
"""
|
2016-03-16 16:18:52 +01:00
|
|
|
assert dm_type in ('vasp', 'wien2k'), "'dm_type' must be either 'vasp' or 'wienk'"
|
2013-07-23 19:49:42 +02:00
|
|
|
|
2016-03-16 16:18:52 +01:00
|
|
|
if filename is None:
|
|
|
|
if dm_type == 'wien2k':
|
|
|
|
filename = 'dens_mat.dat'
|
|
|
|
elif dm_type == 'vasp':
|
|
|
|
filename = 'GAMMA'
|
2017-10-24 09:49:54 +02:00
|
|
|
|
|
|
|
assert type(filename) == StringType, ("calc_density_correction: "
|
|
|
|
"filename has to be a string!")
|
|
|
|
|
2014-11-14 09:43:28 +01:00
|
|
|
ntoi = self.spin_names_to_ind[self.SO]
|
2014-11-17 10:48:04 +01:00
|
|
|
spn = self.spin_block_names[self.SO]
|
|
|
|
dens = {sp: 0.0 for sp in spn}
|
2016-03-16 16:18:52 +01:00
|
|
|
band_en_correction = 0.0
|
|
|
|
|
|
|
|
# Fetch Fermi weights and energy window band indices
|
|
|
|
if dm_type == 'vasp':
|
|
|
|
fermi_weights = 0
|
|
|
|
band_window = 0
|
|
|
|
if mpi.is_master_node():
|
|
|
|
ar = HDFArchive(self.hdf_file,'r')
|
|
|
|
fermi_weights = ar['dft_misc_input']['dft_fermi_weights']
|
|
|
|
band_window = ar['dft_misc_input']['band_window']
|
|
|
|
del ar
|
|
|
|
fermi_weights = mpi.bcast(fermi_weights)
|
|
|
|
band_window = mpi.bcast(band_window)
|
|
|
|
|
|
|
|
# Convert Fermi weights to a density matrix
|
|
|
|
dens_mat_dft = {}
|
|
|
|
for sp in spn:
|
|
|
|
dens_mat_dft[sp] = [fermi_weights[ik, ntoi[sp], :].astype(numpy.complex_) for ik in xrange(self.n_k)]
|
|
|
|
|
2013-07-23 19:49:42 +02:00
|
|
|
|
|
|
|
# Set up deltaN:
|
|
|
|
deltaN = {}
|
2014-11-17 10:48:04 +01:00
|
|
|
for sp in spn:
|
2016-05-09 10:19:56 +02:00
|
|
|
deltaN[sp] = [numpy.zeros([self.n_orbitals[ik, ntoi[sp]], self.n_orbitals[
|
|
|
|
ik, ntoi[sp]]], numpy.complex_) for ik in range(self.n_k)]
|
2014-09-22 19:24:33 +02:00
|
|
|
|
2014-11-17 10:48:04 +01:00
|
|
|
ikarray = numpy.array(range(self.n_k))
|
2013-07-23 19:49:42 +02:00
|
|
|
for ik in mpi.slice_array(ikarray):
|
2016-05-09 10:19:56 +02:00
|
|
|
G_latt_iw = self.lattice_gf(
|
|
|
|
ik=ik, mu=self.chemical_potential, iw_or_w="iw")
|
|
|
|
for bname, gf in G_latt_iw:
|
2014-12-07 00:29:39 +01:00
|
|
|
deltaN[bname][ik] = G_latt_iw[bname].density()
|
2017-01-27 12:19:03 +01:00
|
|
|
|
2014-12-07 00:29:39 +01:00
|
|
|
dens[bname] += self.bz_weights[ik] * G_latt_iw[bname].total_density()
|
2016-03-16 16:18:52 +01:00
|
|
|
if dm_type == 'vasp':
|
|
|
|
# In 'vasp'-mode subtract the DFT density matrix
|
2016-05-10 11:48:28 +02:00
|
|
|
nb = self.n_orbitals[ik, ntoi[bname]]
|
|
|
|
diag_inds = numpy.diag_indices(nb)
|
|
|
|
deltaN[bname][ik][diag_inds] -= dens_mat_dft[bname][ik][:nb]
|
2016-03-16 16:18:52 +01:00
|
|
|
dens[bname] -= self.bz_weights[ik] * dens_mat_dft[bname][ik].sum().real
|
|
|
|
isp = ntoi[bname]
|
2016-12-31 14:46:51 +01:00
|
|
|
b1, b2 = band_window[isp][ik, :2]
|
2016-03-16 16:18:52 +01:00
|
|
|
nb = b2 - b1 + 1
|
|
|
|
assert nb == self.n_orbitals[ik, ntoi[bname]], "Number of bands is inconsistent at ik = %s"%(ik)
|
|
|
|
band_en_correction += numpy.dot(deltaN[bname][ik], self.hopping[ik, isp, :nb, :nb]).trace().real * self.bz_weights[ik]
|
|
|
|
|
2014-12-07 00:29:39 +01:00
|
|
|
# mpi reduce:
|
2014-11-14 09:43:28 +01:00
|
|
|
for bname in deltaN:
|
2013-07-23 19:49:42 +02:00
|
|
|
for ik in range(self.n_k):
|
2016-05-09 10:19:56 +02:00
|
|
|
deltaN[bname][ik] = mpi.all_reduce(
|
|
|
|
mpi.world, deltaN[bname][ik], lambda x, y: x + y)
|
|
|
|
dens[bname] = mpi.all_reduce(
|
|
|
|
mpi.world, dens[bname], lambda x, y: x + y)
|
2013-07-23 19:49:42 +02:00
|
|
|
mpi.barrier()
|
2016-03-16 16:18:52 +01:00
|
|
|
band_en_correction = mpi.all_reduce(mpi.world, band_en_correction, lambda x,y : x+y)
|
2013-07-23 19:49:42 +02:00
|
|
|
|
|
|
|
# now save to file:
|
2016-03-16 16:18:52 +01:00
|
|
|
if dm_type == 'wien2k':
|
|
|
|
if mpi.is_master_node():
|
|
|
|
if self.SP == 0:
|
2017-01-27 12:19:03 +01:00
|
|
|
f = open(filename, 'w')
|
2016-03-16 16:18:52 +01:00
|
|
|
else:
|
2017-01-27 12:19:03 +01:00
|
|
|
f = open(filename + 'up', 'w')
|
|
|
|
f1 = open(filename + 'dn', 'w')
|
2016-03-16 16:18:52 +01:00
|
|
|
# write chemical potential (in Rydberg):
|
2017-01-27 12:19:03 +01:00
|
|
|
f.write("%.14f\n" % (self.chemical_potential / self.energy_unit))
|
|
|
|
if self.SP != 0:
|
|
|
|
f1.write("%.14f\n" %
|
|
|
|
(self.chemical_potential / self.energy_unit))
|
2016-03-16 16:18:52 +01:00
|
|
|
# write beta in rydberg-1
|
2017-01-27 12:19:03 +01:00
|
|
|
f.write("%.14f\n" % (G_latt_iw.mesh.beta * self.energy_unit))
|
|
|
|
if self.SP != 0:
|
|
|
|
f1.write("%.14f\n" % (G_latt_iw.mesh.beta * self.energy_unit))
|
2014-09-22 19:24:33 +02:00
|
|
|
|
2017-01-27 12:19:03 +01:00
|
|
|
if self.SP == 0: # no spin-polarization
|
2014-11-09 12:10:40 +01:00
|
|
|
|
|
|
|
for ik in range(self.n_k):
|
2017-01-27 12:19:03 +01:00
|
|
|
f.write("%s\n" % self.n_orbitals[ik, 0])
|
|
|
|
for inu in range(self.n_orbitals[ik, 0]):
|
|
|
|
for imu in range(self.n_orbitals[ik, 0]):
|
|
|
|
valre = (deltaN['up'][ik][
|
|
|
|
inu, imu].real + deltaN['down'][ik][inu, imu].real) / 2.0
|
|
|
|
valim = (deltaN['up'][ik][
|
|
|
|
inu, imu].imag + deltaN['down'][ik][inu, imu].imag) / 2.0
|
|
|
|
f.write("%.14f %.14f " % (valre, valim))
|
2016-03-16 16:18:52 +01:00
|
|
|
f.write("\n")
|
|
|
|
f.write("\n")
|
|
|
|
f.close()
|
|
|
|
|
2017-01-27 12:19:03 +01:00
|
|
|
elif self.SP == 1: # with spin-polarization
|
2016-03-16 16:18:52 +01:00
|
|
|
|
|
|
|
# dict of filename: (spin index, block_name)
|
2017-01-27 12:19:03 +01:00
|
|
|
if self.SO == 0:
|
|
|
|
to_write = {f: (0, 'up'), f1: (1, 'down')}
|
|
|
|
if self.SO == 1:
|
|
|
|
to_write = {f: (0, 'ud'), f1: (0, 'ud')}
|
2016-03-16 16:18:52 +01:00
|
|
|
for fout in to_write.iterkeys():
|
|
|
|
isp, sp = to_write[fout]
|
|
|
|
for ik in range(self.n_k):
|
2017-01-27 12:19:03 +01:00
|
|
|
fout.write("%s\n" % self.n_orbitals[ik, isp])
|
|
|
|
for inu in range(self.n_orbitals[ik, isp]):
|
|
|
|
for imu in range(self.n_orbitals[ik, isp]):
|
|
|
|
fout.write("%.14f %.14f " % (deltaN[sp][ik][
|
|
|
|
inu, imu].real, deltaN[sp][ik][inu, imu].imag))
|
2017-10-16 10:15:20 +02:00
|
|
|
fout.write("\n")
|
2014-11-09 12:10:40 +01:00
|
|
|
fout.write("\n")
|
2016-03-16 16:18:52 +01:00
|
|
|
fout.close()
|
|
|
|
elif dm_type == 'vasp':
|
|
|
|
assert self.SP == 0, "Spin-polarized density matrix is not implemented"
|
|
|
|
|
|
|
|
if mpi.is_master_node():
|
|
|
|
with open(filename, 'w') as f:
|
|
|
|
f.write(" %i -1 ! Number of k-points, default number of bands\n"%(self.n_k))
|
|
|
|
for ik in xrange(self.n_k):
|
|
|
|
ib1 = band_window[0][ik, 0]
|
|
|
|
ib2 = band_window[0][ik, 1]
|
|
|
|
f.write(" %i %i %i\n"%(ik + 1, ib1, ib2))
|
|
|
|
for inu in xrange(self.n_orbitals[ik, 0]):
|
|
|
|
for imu in xrange(self.n_orbitals[ik, 0]):
|
|
|
|
valre = (deltaN['up'][ik][inu, imu].real + deltaN['down'][ik][inu, imu].real) / 2.0
|
|
|
|
valim = (deltaN['up'][ik][inu, imu].imag + deltaN['down'][ik][inu, imu].imag) / 2.0
|
|
|
|
f.write(" %.14f %.14f"%(valre, valim))
|
|
|
|
f.write("\n")
|
|
|
|
else:
|
|
|
|
raise NotImplementedError("Unknown density matrix type: '%s'"%(dm_type))
|
|
|
|
|
|
|
|
res = deltaN, dens
|
|
|
|
|
|
|
|
if dm_type == 'vasp':
|
|
|
|
res += (band_en_correction,)
|
|
|
|
|
|
|
|
return res
|
2013-07-23 19:49:42 +02:00
|
|
|
|
2014-10-28 16:19:28 +01:00
|
|
|
|
|
|
|
################
|
|
|
|
# FIXME LEAVE UNDOCUMENTED
|
|
|
|
################
|
|
|
|
|
2016-05-09 10:19:56 +02:00
|
|
|
def calc_dc_for_density(self, orb, dc_init, dens_mat, density=None, precision=0.01):
|
2014-10-28 16:19:28 +01:00
|
|
|
"""Searches for DC in order to fulfill charge neutrality.
|
2014-12-07 00:29:39 +01:00
|
|
|
If density is given, then DC is set such that the LOCAL charge of orbital
|
|
|
|
orb coincides with the given density."""
|
2014-10-28 16:19:28 +01:00
|
|
|
|
|
|
|
def F(dc):
|
2016-05-09 10:19:56 +02:00
|
|
|
self.calc_dc(dens_mat=dens_mat, U_interact=0,
|
|
|
|
J_hund=0, orb=orb, use_dc_value=dc)
|
2014-11-15 20:04:54 +01:00
|
|
|
if dens_req is None:
|
2016-05-09 10:19:56 +02:00
|
|
|
return self.total_density(mu=mu)
|
2014-10-28 16:19:28 +01:00
|
|
|
else:
|
|
|
|
return self.extract_G_loc()[orb].total_density()
|
|
|
|
|
2016-05-09 10:19:56 +02:00
|
|
|
if density is None:
|
|
|
|
density = self.density_required - self.charge_below
|
2014-10-28 16:19:28 +01:00
|
|
|
|
2016-05-09 10:19:56 +02:00
|
|
|
dc = dichotomy.dichotomy(function=F,
|
|
|
|
x_init=dc_init, y_value=density,
|
|
|
|
precision_on_y=precision, delta_x=0.5,
|
|
|
|
max_loops=100, x_name="Double Counting", y_name="Total Density",
|
|
|
|
verbosity=3)[0]
|
2014-10-28 16:19:28 +01:00
|
|
|
|
2014-12-07 00:29:39 +01:00
|
|
|
return dc
|
2014-10-28 16:19:28 +01:00
|
|
|
|
|
|
|
def check_projectors(self):
|
2014-12-07 00:29:39 +01:00
|
|
|
"""Calculated the density matrix from projectors (DM = P Pdagger) to check that it is correct and
|
|
|
|
specifically that it matches DFT."""
|
2016-05-09 10:19:56 +02:00
|
|
|
dens_mat = [numpy.zeros([self.corr_shells[icrsh]['dim'], self.corr_shells[icrsh]['dim']], numpy.complex_)
|
|
|
|
for icrsh in range(self.n_corr_shells)]
|
2014-10-28 16:19:28 +01:00
|
|
|
|
|
|
|
for ik in range(self.n_k):
|
2014-11-15 17:03:16 +01:00
|
|
|
for icrsh in range(self.n_corr_shells):
|
2014-11-26 16:24:02 +01:00
|
|
|
dim = self.corr_shells[icrsh]['dim']
|
2016-05-09 10:19:56 +02:00
|
|
|
n_orb = self.n_orbitals[ik, 0]
|
|
|
|
projmat = self.proj_mat[ik, 0, icrsh, 0:dim, 0:n_orb]
|
|
|
|
dens_mat[icrsh][
|
|
|
|
:, :] += numpy.dot(projmat, projmat.transpose().conjugate()) * self.bz_weights[ik]
|
2014-10-28 16:19:28 +01:00
|
|
|
|
2016-05-09 10:19:56 +02:00
|
|
|
if self.symm_op != 0:
|
|
|
|
dens_mat = self.symmcorr.symmetrize(dens_mat)
|
2014-10-28 16:19:28 +01:00
|
|
|
|
|
|
|
# Rotate to local coordinate system:
|
2014-11-15 17:03:16 +01:00
|
|
|
if self.use_rotations:
|
|
|
|
for icrsh in range(self.n_corr_shells):
|
2016-05-09 10:19:56 +02:00
|
|
|
if self.rot_mat_time_inv[icrsh] == 1:
|
|
|
|
dens_mat[icrsh] = dens_mat[icrsh].conjugate()
|
|
|
|
dens_mat[icrsh] = numpy.dot(numpy.dot(self.rot_mat[icrsh].conjugate().transpose(), dens_mat[icrsh]),
|
|
|
|
self.rot_mat[icrsh])
|
2014-10-28 16:19:28 +01:00
|
|
|
|
|
|
|
return dens_mat
|
|
|
|
|
2016-05-09 10:19:56 +02:00
|
|
|
def sorts_of_atoms(self, shells):
|
2014-10-28 16:19:28 +01:00
|
|
|
"""
|
2014-12-07 00:29:39 +01:00
|
|
|
Determine the number of inequivalent sorts.
|
2014-10-28 16:19:28 +01:00
|
|
|
"""
|
2016-05-09 10:19:56 +02:00
|
|
|
sortlst = [shells[i]['sort'] for i in range(len(shells))]
|
2014-12-07 00:29:39 +01:00
|
|
|
n_sorts = len(set(sortlst))
|
|
|
|
return n_sorts
|
2014-10-28 16:19:28 +01:00
|
|
|
|
2016-05-09 10:19:56 +02:00
|
|
|
def number_of_atoms(self, shells):
|
2014-10-28 16:19:28 +01:00
|
|
|
"""
|
2014-12-07 00:29:39 +01:00
|
|
|
Determine the number of inequivalent atoms.
|
2014-10-28 16:19:28 +01:00
|
|
|
"""
|
2016-05-09 10:19:56 +02:00
|
|
|
atomlst = [shells[i]['atom'] for i in range(len(shells))]
|
2014-12-07 00:29:39 +01:00
|
|
|
n_atoms = len(set(atomlst))
|
|
|
|
return n_atoms
|
2016-09-13 11:57:48 +02:00
|
|
|
|
|
|
|
# The following methods are here to ensure backward-compatibility
|
|
|
|
# after introducing the block_structure class
|
|
|
|
def __get_gf_struct_sumk(self):
|
|
|
|
return self.block_structure.gf_struct_sumk
|
|
|
|
def __set_gf_struct_sumk(self,value):
|
|
|
|
self.block_structure.gf_struct_sumk = value
|
|
|
|
gf_struct_sumk = property(__get_gf_struct_sumk,__set_gf_struct_sumk)
|
|
|
|
|
|
|
|
def __get_gf_struct_solver(self):
|
|
|
|
return self.block_structure.gf_struct_solver
|
|
|
|
def __set_gf_struct_solver(self,value):
|
|
|
|
self.block_structure.gf_struct_solver = value
|
|
|
|
gf_struct_solver = property(__get_gf_struct_solver,__set_gf_struct_solver)
|
|
|
|
|
|
|
|
def __get_solver_to_sumk(self):
|
|
|
|
return self.block_structure.solver_to_sumk
|
|
|
|
def __set_solver_to_sumk(self,value):
|
|
|
|
self.block_structure.solver_to_sumk = value
|
|
|
|
solver_to_sumk = property(__get_solver_to_sumk,__set_solver_to_sumk)
|
|
|
|
|
|
|
|
def __get_sumk_to_solver(self):
|
|
|
|
return self.block_structure.sumk_to_solver
|
|
|
|
def __set_sumk_to_solver(self,value):
|
|
|
|
self.block_structure.sumk_to_solver = value
|
|
|
|
sumk_to_solver = property(__get_sumk_to_solver,__set_sumk_to_solver)
|
|
|
|
|
|
|
|
def __get_solver_to_sumk_block(self):
|
|
|
|
return self.block_structure.solver_to_sumk_block
|
|
|
|
def __set_solver_to_sumk_block(self,value):
|
|
|
|
self.block_structure.solver_to_sumk_block = value
|
|
|
|
solver_to_sumk_block = property(__get_solver_to_sumk_block,__set_solver_to_sumk_block)
|