2013-07-23 19:49:42 +02:00
|
|
|
|
2016-05-09 10:19:56 +02:00
|
|
|
##########################################################################
|
2013-07-23 19:49:42 +02:00
|
|
|
#
|
|
|
|
# TRIQS: a Toolbox for Research in Interacting Quantum Systems
|
|
|
|
#
|
|
|
|
# Copyright (C) 2011 by M. Aichhorn, L. Pourovskii, V. Vildosola
|
|
|
|
#
|
|
|
|
# TRIQS is free software: you can redistribute it and/or modify it under the
|
|
|
|
# terms of the GNU General Public License as published by the Free Software
|
|
|
|
# Foundation, either version 3 of the License, or (at your option) any later
|
|
|
|
# version.
|
|
|
|
#
|
|
|
|
# TRIQS is distributed in the hope that it will be useful, but WITHOUT ANY
|
|
|
|
# WARRANTY; without even the implied warranty of MERCHANTABILITY or FITNESS
|
|
|
|
# FOR A PARTICULAR PURPOSE. See the GNU General Public License for more
|
|
|
|
# details.
|
|
|
|
#
|
|
|
|
# You should have received a copy of the GNU General Public License along with
|
|
|
|
# TRIQS. If not, see <http://www.gnu.org/licenses/>.
|
|
|
|
#
|
2016-05-09 10:19:56 +02:00
|
|
|
##########################################################################
|
2013-07-23 19:49:42 +02:00
|
|
|
|
|
|
|
from types import *
|
2014-12-09 16:47:40 +01:00
|
|
|
import numpy
|
2013-07-23 19:49:42 +02:00
|
|
|
from pytriqs.archive import *
|
2014-11-08 23:51:24 +01:00
|
|
|
from converter_tools import *
|
2014-11-20 13:17:46 +01:00
|
|
|
import os.path
|
2013-07-23 19:49:42 +02:00
|
|
|
|
2016-05-09 10:19:56 +02:00
|
|
|
|
2014-11-08 23:51:24 +01:00
|
|
|
class Wien2kConverter(ConverterTools):
|
2013-07-23 19:49:42 +02:00
|
|
|
"""
|
2014-11-18 11:30:26 +01:00
|
|
|
Conversion from Wien2k output to an hdf5 file that can be used as input for the SumkDFT class.
|
2013-07-23 19:49:42 +02:00
|
|
|
"""
|
|
|
|
|
2016-05-09 10:19:56 +02:00
|
|
|
def __init__(self, filename, hdf_filename=None,
|
|
|
|
dft_subgrp='dft_input', symmcorr_subgrp='dft_symmcorr_input',
|
|
|
|
parproj_subgrp='dft_parproj_input', symmpar_subgrp='dft_symmpar_input',
|
|
|
|
bands_subgrp='dft_bands_input', misc_subgrp='dft_misc_input',
|
|
|
|
transp_subgrp='dft_transp_input', repacking=False):
|
2013-07-23 19:49:42 +02:00
|
|
|
"""
|
2015-03-12 00:01:12 +01:00
|
|
|
Initialise the class.
|
|
|
|
|
|
|
|
Parameters
|
|
|
|
----------
|
|
|
|
filename : string
|
|
|
|
Base name of DFT files.
|
|
|
|
hdf_filename : string, optional
|
|
|
|
Name of hdf5 archive to be created.
|
|
|
|
dft_subgrp : string, optional
|
|
|
|
Name of subgroup storing necessary DFT data.
|
|
|
|
symmcorr_subgrp : string, optional
|
|
|
|
Name of subgroup storing correlated-shell symmetry data.
|
|
|
|
parproj_subgrp : string, optional
|
|
|
|
Name of subgroup storing partial projector data.
|
|
|
|
symmpar_subgrp : string, optional
|
|
|
|
Name of subgroup storing partial-projector symmetry data.
|
|
|
|
bands_subgrp : string, optional
|
|
|
|
Name of subgroup storing band data.
|
|
|
|
misc_subgrp : string, optional
|
|
|
|
Name of subgroup storing miscellaneous DFT data.
|
|
|
|
transp_subgrp : string, optional
|
|
|
|
Name of subgroup storing transport data.
|
|
|
|
repacking : boolean, optional
|
|
|
|
Does the hdf5 archive need to be repacked to save space?
|
2016-05-09 10:19:56 +02:00
|
|
|
|
2013-07-23 19:49:42 +02:00
|
|
|
"""
|
|
|
|
|
2016-05-09 10:19:56 +02:00
|
|
|
assert type(
|
|
|
|
filename) == StringType, "Wien2kConverter: Please provide the DFT files' base name as a string."
|
|
|
|
if hdf_filename is None:
|
|
|
|
hdf_filename = filename + '.h5'
|
2015-02-10 16:35:12 +01:00
|
|
|
self.hdf_file = hdf_filename
|
2016-05-09 10:19:56 +02:00
|
|
|
self.dft_file = filename + '.ctqmcout'
|
|
|
|
self.symmcorr_file = filename + '.symqmc'
|
|
|
|
self.parproj_file = filename + '.parproj'
|
|
|
|
self.symmpar_file = filename + '.sympar'
|
|
|
|
self.band_file = filename + '.outband'
|
|
|
|
self.bandwin_file = filename + '.oubwin'
|
|
|
|
self.struct_file = filename + '.struct'
|
|
|
|
self.outputs_file = filename + '.outputs'
|
|
|
|
self.pmat_file = filename + '.pmat'
|
2014-11-18 11:30:26 +01:00
|
|
|
self.dft_subgrp = dft_subgrp
|
2014-11-03 14:47:55 +01:00
|
|
|
self.symmcorr_subgrp = symmcorr_subgrp
|
|
|
|
self.parproj_subgrp = parproj_subgrp
|
|
|
|
self.symmpar_subgrp = symmpar_subgrp
|
|
|
|
self.bands_subgrp = bands_subgrp
|
2015-02-10 11:55:44 +01:00
|
|
|
self.misc_subgrp = misc_subgrp
|
2014-11-03 11:16:32 +01:00
|
|
|
self.transp_subgrp = transp_subgrp
|
2016-05-09 10:19:56 +02:00
|
|
|
self.fortran_to_replace = {'D': 'E'}
|
2013-07-23 19:49:42 +02:00
|
|
|
|
|
|
|
# Checks if h5 file is there and repacks it if wanted:
|
|
|
|
if (os.path.exists(self.hdf_file) and repacking):
|
2014-11-19 18:50:53 +01:00
|
|
|
ConverterTools.repack(self)
|
2014-11-20 13:17:46 +01:00
|
|
|
|
2014-12-09 12:26:00 +01:00
|
|
|
def convert_dft_input(self):
|
2013-07-23 19:49:42 +02:00
|
|
|
"""
|
2015-03-12 00:01:12 +01:00
|
|
|
Reads the appropriate files and stores the data for the
|
|
|
|
|
|
|
|
- dft_subgrp
|
|
|
|
- symmcorr_subgrp
|
|
|
|
- misc_subgrp
|
|
|
|
|
|
|
|
in the hdf5 archive.
|
|
|
|
|
2013-07-23 19:49:42 +02:00
|
|
|
"""
|
|
|
|
|
2016-05-09 10:19:56 +02:00
|
|
|
# Read and write only on the master node
|
|
|
|
if not (mpi.is_master_node()):
|
|
|
|
return
|
|
|
|
mpi.report("Reading input from %s..." % self.dft_file)
|
|
|
|
|
|
|
|
# R is a generator : each R.Next() will return the next number in the
|
|
|
|
# file
|
|
|
|
R = ConverterTools.read_fortran_file(
|
|
|
|
self, self.dft_file, self.fortran_to_replace)
|
2013-07-23 19:49:42 +02:00
|
|
|
try:
|
2014-11-26 16:24:02 +01:00
|
|
|
energy_unit = R.next() # read the energy convertion factor
|
2016-05-09 10:19:56 +02:00
|
|
|
# read the number of k points
|
|
|
|
n_k = int(R.next())
|
|
|
|
k_dep_projection = 1
|
|
|
|
# flag for spin-polarised calculation
|
|
|
|
SP = int(R.next())
|
|
|
|
# flag for spin-orbit calculation
|
|
|
|
SO = int(R.next())
|
2013-07-23 19:49:42 +02:00
|
|
|
charge_below = R.next() # total charge below energy window
|
2016-05-09 10:19:56 +02:00
|
|
|
# total density required, for setting the chemical potential
|
|
|
|
density_required = R.next()
|
2013-07-23 19:49:42 +02:00
|
|
|
symm_op = 1 # Use symmetry groups for the k-sum
|
|
|
|
|
2016-05-09 10:19:56 +02:00
|
|
|
# the information on the non-correlated shells is not important
|
|
|
|
# here, maybe skip:
|
|
|
|
# number of shells (e.g. Fe d, As p, O p) in the unit cell,
|
|
|
|
n_shells = int(R.next())
|
|
|
|
# corresponds to index R in formulas
|
2014-11-26 16:24:02 +01:00
|
|
|
# now read the information about the shells (atom, sort, l, dim):
|
|
|
|
shell_entries = ['atom', 'sort', 'l', 'dim']
|
2016-05-09 10:19:56 +02:00
|
|
|
shells = [{name: int(val) for name, val in zip(
|
|
|
|
shell_entries, R)} for ish in range(n_shells)]
|
|
|
|
|
|
|
|
# number of corr. shells (e.g. Fe d, Ce f) in the unit cell,
|
|
|
|
n_corr_shells = int(R.next())
|
|
|
|
# corresponds to index R in formulas
|
|
|
|
# now read the information about the shells (atom, sort, l, dim, SO
|
|
|
|
# flag, irep):
|
2014-11-26 16:24:02 +01:00
|
|
|
corr_shell_entries = ['atom', 'sort', 'l', 'dim', 'SO', 'irep']
|
2016-05-09 10:19:56 +02:00
|
|
|
corr_shells = [{name: int(val) for name, val in zip(
|
|
|
|
corr_shell_entries, R)} for icrsh in range(n_corr_shells)]
|
2013-07-23 19:49:42 +02:00
|
|
|
|
2016-05-09 10:19:56 +02:00
|
|
|
# determine the number of inequivalent correlated shells and maps,
|
|
|
|
# needed for further reading
|
|
|
|
n_inequiv_shells, corr_to_inequiv, inequiv_to_corr = ConverterTools.det_shell_equivalence(
|
|
|
|
self, corr_shells)
|
2013-07-23 19:49:42 +02:00
|
|
|
|
|
|
|
use_rotations = 1
|
2016-05-09 10:19:56 +02:00
|
|
|
rot_mat = [numpy.identity(
|
|
|
|
corr_shells[icrsh]['dim'], numpy.complex_) for icrsh in range(n_corr_shells)]
|
|
|
|
|
2013-07-23 19:49:42 +02:00
|
|
|
# read the matrices
|
|
|
|
rot_mat_time_inv = [0 for i in range(n_corr_shells)]
|
|
|
|
|
2014-11-26 16:24:02 +01:00
|
|
|
for icrsh in range(n_corr_shells):
|
|
|
|
for i in range(corr_shells[icrsh]['dim']): # read real part:
|
|
|
|
for j in range(corr_shells[icrsh]['dim']):
|
2016-05-09 10:19:56 +02:00
|
|
|
rot_mat[icrsh][i, j] = R.next()
|
|
|
|
# read imaginary part:
|
|
|
|
for i in range(corr_shells[icrsh]['dim']):
|
2014-11-26 16:24:02 +01:00
|
|
|
for j in range(corr_shells[icrsh]['dim']):
|
2016-05-09 10:19:56 +02:00
|
|
|
rot_mat[icrsh][i, j] += 1j * R.next()
|
2013-07-23 19:49:42 +02:00
|
|
|
|
2016-05-09 10:19:56 +02:00
|
|
|
if (SP == 1): # read time inversion flag:
|
2013-07-23 19:49:42 +02:00
|
|
|
rot_mat_time_inv[icrsh] = int(R.next())
|
2016-05-09 10:19:56 +02:00
|
|
|
|
2014-09-22 19:24:33 +02:00
|
|
|
# Read here the info for the transformation of the basis:
|
2014-11-15 18:15:45 +01:00
|
|
|
n_reps = [1 for i in range(n_inequiv_shells)]
|
|
|
|
dim_reps = [0 for i in range(n_inequiv_shells)]
|
2013-07-23 19:49:42 +02:00
|
|
|
T = []
|
2014-11-26 16:24:02 +01:00
|
|
|
for ish in range(n_inequiv_shells):
|
2016-05-09 10:19:56 +02:00
|
|
|
# number of representatives ("subsets"), e.g. t2g and eg
|
|
|
|
n_reps[ish] = int(R.next())
|
|
|
|
dim_reps[ish] = [int(R.next()) for i in range(
|
|
|
|
n_reps[ish])] # dimensions of the subsets
|
|
|
|
|
2014-09-22 19:24:33 +02:00
|
|
|
# The transformation matrix:
|
|
|
|
# is of dimension 2l+1 without SO, and 2*(2l+1) with SO!
|
2016-05-09 10:19:56 +02:00
|
|
|
ll = 2 * corr_shells[inequiv_to_corr[ish]]['l'] + 1
|
2014-11-26 16:24:02 +01:00
|
|
|
lmax = ll * (corr_shells[inequiv_to_corr[ish]]['SO'] + 1)
|
2016-05-09 10:19:56 +02:00
|
|
|
T.append(numpy.zeros([lmax, lmax], numpy.complex_))
|
|
|
|
|
2013-07-23 19:49:42 +02:00
|
|
|
# now read it from file:
|
2014-11-26 16:24:02 +01:00
|
|
|
for i in range(lmax):
|
|
|
|
for j in range(lmax):
|
2016-05-09 10:19:56 +02:00
|
|
|
T[ish][i, j] = R.next()
|
2014-11-26 16:24:02 +01:00
|
|
|
for i in range(lmax):
|
|
|
|
for j in range(lmax):
|
2016-05-09 10:19:56 +02:00
|
|
|
T[ish][i, j] += 1j * R.next()
|
|
|
|
|
2013-07-23 19:49:42 +02:00
|
|
|
# Spin blocks to be read:
|
2016-05-09 10:19:56 +02:00
|
|
|
n_spin_blocs = SP + 1 - SO
|
|
|
|
|
2013-07-23 19:49:42 +02:00
|
|
|
# read the list of n_orbitals for all k points
|
2016-05-09 10:19:56 +02:00
|
|
|
n_orbitals = numpy.zeros([n_k, n_spin_blocs], numpy.int)
|
2013-07-23 19:49:42 +02:00
|
|
|
for isp in range(n_spin_blocs):
|
2014-11-26 16:24:02 +01:00
|
|
|
for ik in range(n_k):
|
2016-05-09 10:19:56 +02:00
|
|
|
n_orbitals[ik, isp] = int(R.next())
|
|
|
|
|
2013-07-23 19:49:42 +02:00
|
|
|
# Initialise the projectors:
|
2016-05-09 10:19:56 +02:00
|
|
|
proj_mat = numpy.zeros([n_k, n_spin_blocs, n_corr_shells, max(
|
|
|
|
[crsh['dim'] for crsh in corr_shells]), numpy.max(n_orbitals)], numpy.complex_)
|
2014-09-22 19:24:33 +02:00
|
|
|
|
2013-07-23 19:49:42 +02:00
|
|
|
# Read the projectors from the file:
|
2014-11-26 16:24:02 +01:00
|
|
|
for ik in range(n_k):
|
2013-07-23 19:49:42 +02:00
|
|
|
for icrsh in range(n_corr_shells):
|
2014-11-26 16:24:02 +01:00
|
|
|
n_orb = corr_shells[icrsh]['dim']
|
2016-05-09 10:19:56 +02:00
|
|
|
# first Real part for BOTH spins, due to conventions in
|
|
|
|
# dmftproj:
|
2013-07-23 19:49:42 +02:00
|
|
|
for isp in range(n_spin_blocs):
|
2014-11-26 16:24:02 +01:00
|
|
|
for i in range(n_orb):
|
|
|
|
for j in range(n_orbitals[ik][isp]):
|
2016-05-09 10:19:56 +02:00
|
|
|
proj_mat[ik, isp, icrsh, i, j] = R.next()
|
2013-07-23 19:49:42 +02:00
|
|
|
# now Imag part:
|
|
|
|
for isp in range(n_spin_blocs):
|
2014-11-26 16:24:02 +01:00
|
|
|
for i in range(n_orb):
|
|
|
|
for j in range(n_orbitals[ik][isp]):
|
2016-05-09 10:19:56 +02:00
|
|
|
proj_mat[ik, isp, icrsh, i, j] += 1j * R.next()
|
|
|
|
|
2013-07-23 19:49:42 +02:00
|
|
|
# now define the arrays for weights and hopping ...
|
2016-05-09 10:19:56 +02:00
|
|
|
# w(k_index), default normalisation
|
|
|
|
bz_weights = numpy.ones([n_k], numpy.float_) / float(n_k)
|
|
|
|
hopping = numpy.zeros([n_k, n_spin_blocs, numpy.max(
|
|
|
|
n_orbitals), numpy.max(n_orbitals)], numpy.complex_)
|
2014-09-22 19:24:33 +02:00
|
|
|
|
2013-07-23 19:49:42 +02:00
|
|
|
# weights in the file
|
2016-05-09 10:19:56 +02:00
|
|
|
for ik in range(n_k):
|
|
|
|
bz_weights[ik] = R.next()
|
|
|
|
|
2013-07-23 19:49:42 +02:00
|
|
|
# if the sum over spins is in the weights, take it out again!!
|
|
|
|
sm = sum(bz_weights)
|
2016-05-09 10:19:56 +02:00
|
|
|
bz_weights[:] /= sm
|
2014-09-22 19:24:33 +02:00
|
|
|
|
2013-07-23 19:49:42 +02:00
|
|
|
# Grab the H
|
2016-05-09 10:19:56 +02:00
|
|
|
# we use now the convention of a DIAGONAL Hamiltonian -- convention
|
|
|
|
# for Wien2K.
|
2013-07-23 19:49:42 +02:00
|
|
|
for isp in range(n_spin_blocs):
|
2016-05-09 10:19:56 +02:00
|
|
|
for ik in range(n_k):
|
|
|
|
n_orb = n_orbitals[ik, isp]
|
2014-11-26 16:24:02 +01:00
|
|
|
for i in range(n_orb):
|
2016-05-09 10:19:56 +02:00
|
|
|
hopping[ik, isp, i, i] = R.next() * energy_unit
|
|
|
|
|
2014-09-22 19:24:33 +02:00
|
|
|
# keep some things that we need for reading parproj:
|
2016-05-09 10:19:56 +02:00
|
|
|
things_to_set = ['n_shells', 'shells', 'n_corr_shells', 'corr_shells',
|
|
|
|
'n_spin_blocs', 'n_orbitals', 'n_k', 'SO', 'SP', 'energy_unit']
|
|
|
|
for it in things_to_set:
|
|
|
|
setattr(self, it, locals()[it])
|
|
|
|
except StopIteration: # a more explicit error if the file is corrupted.
|
|
|
|
raise "Wien2k_converter : reading file %s failed!" % self.dft_file
|
2013-07-23 19:49:42 +02:00
|
|
|
|
|
|
|
R.close()
|
2014-11-03 14:47:55 +01:00
|
|
|
# Reading done!
|
2016-05-09 10:19:56 +02:00
|
|
|
|
2014-11-03 14:47:55 +01:00
|
|
|
# Save it to the HDF:
|
2016-05-09 10:19:56 +02:00
|
|
|
ar = HDFArchive(self.hdf_file, 'a')
|
|
|
|
if not (self.dft_subgrp in ar):
|
|
|
|
ar.create_group(self.dft_subgrp)
|
|
|
|
# The subgroup containing the data. If it does not exist, it is
|
|
|
|
# created. If it exists, the data is overwritten!
|
|
|
|
things_to_save = ['energy_unit', 'n_k', 'k_dep_projection', 'SP', 'SO', 'charge_below', 'density_required',
|
|
|
|
'symm_op', 'n_shells', 'shells', 'n_corr_shells', 'corr_shells', 'use_rotations', 'rot_mat',
|
|
|
|
'rot_mat_time_inv', 'n_reps', 'dim_reps', 'T', 'n_orbitals', 'proj_mat', 'bz_weights', 'hopping',
|
2014-11-15 18:15:45 +01:00
|
|
|
'n_inequiv_shells', 'corr_to_inequiv', 'inequiv_to_corr']
|
2016-05-09 10:19:56 +02:00
|
|
|
for it in things_to_save:
|
|
|
|
ar[self.dft_subgrp][it] = locals()[it]
|
2013-07-23 19:49:42 +02:00
|
|
|
del ar
|
2014-10-31 18:52:32 +01:00
|
|
|
|
2016-05-09 10:19:56 +02:00
|
|
|
# Symmetries are used, so now convert symmetry information for
|
|
|
|
# *correlated* orbitals:
|
|
|
|
self.convert_symmetry_input(orbits=self.corr_shells, symm_file=self.symmcorr_file,
|
|
|
|
symm_subgrp=self.symmcorr_subgrp, SO=self.SO, SP=self.SP)
|
2015-03-12 00:01:12 +01:00
|
|
|
self.convert_misc_input()
|
2014-10-31 18:52:32 +01:00
|
|
|
|
2014-11-03 14:47:55 +01:00
|
|
|
def convert_parproj_input(self):
|
2013-07-23 19:49:42 +02:00
|
|
|
"""
|
2015-03-12 00:01:12 +01:00
|
|
|
Reads the appropriate files and stores the data for the
|
|
|
|
|
|
|
|
- parproj_subgrp
|
|
|
|
- symmpar_subgrp
|
|
|
|
|
|
|
|
in the hdf5 archive.
|
|
|
|
|
2013-07-23 19:49:42 +02:00
|
|
|
"""
|
|
|
|
|
2016-05-09 10:19:56 +02:00
|
|
|
if not (mpi.is_master_node()):
|
|
|
|
return
|
2015-07-07 15:08:53 +02:00
|
|
|
|
|
|
|
# get needed data from hdf file
|
2016-05-09 10:19:56 +02:00
|
|
|
ar = HDFArchive(self.hdf_file, 'a')
|
|
|
|
things_to_read = ['SP', 'SO', 'n_shells',
|
|
|
|
'n_k', 'n_orbitals', 'shells']
|
2015-07-07 15:08:53 +02:00
|
|
|
|
|
|
|
for it in things_to_read:
|
2016-05-09 10:19:56 +02:00
|
|
|
if not hasattr(self, it):
|
|
|
|
setattr(self, it, ar[self.dft_subgrp][it])
|
2015-07-07 15:08:53 +02:00
|
|
|
self.n_spin_blocs = self.SP + 1 - self.SO
|
|
|
|
del ar
|
|
|
|
|
2016-05-09 10:19:56 +02:00
|
|
|
mpi.report("Reading input from %s..." % self.parproj_file)
|
2013-07-23 19:49:42 +02:00
|
|
|
|
2016-05-09 10:19:56 +02:00
|
|
|
dens_mat_below = [[numpy.zeros([self.shells[ish]['dim'], self.shells[ish]['dim']], numpy.complex_) for ish in range(self.n_shells)]
|
|
|
|
for isp in range(self.n_spin_blocs)]
|
2013-07-23 19:49:42 +02:00
|
|
|
|
2016-05-09 10:19:56 +02:00
|
|
|
R = ConverterTools.read_fortran_file(
|
|
|
|
self, self.parproj_file, self.fortran_to_replace)
|
2013-07-23 19:49:42 +02:00
|
|
|
|
|
|
|
n_parproj = [int(R.next()) for i in range(self.n_shells)]
|
|
|
|
n_parproj = numpy.array(n_parproj)
|
2016-05-09 10:19:56 +02:00
|
|
|
|
2013-07-23 19:49:42 +02:00
|
|
|
# Initialise P, here a double list of matrices:
|
2016-05-09 10:19:56 +02:00
|
|
|
proj_mat_all = numpy.zeros([self.n_k, self.n_spin_blocs, self.n_shells, max(
|
|
|
|
n_parproj), max([sh['dim'] for sh in self.shells]), max(self.n_orbitals)], numpy.complex_)
|
|
|
|
|
|
|
|
rot_mat_all = [numpy.identity(
|
|
|
|
self.shells[ish]['dim'], numpy.complex_) for ish in range(self.n_shells)]
|
2013-07-23 19:49:42 +02:00
|
|
|
rot_mat_all_time_inv = [0 for i in range(self.n_shells)]
|
|
|
|
|
|
|
|
for ish in range(self.n_shells):
|
|
|
|
# read first the projectors for this orbital:
|
2014-11-26 16:24:02 +01:00
|
|
|
for ik in range(self.n_k):
|
2013-07-23 19:49:42 +02:00
|
|
|
for ir in range(n_parproj[ish]):
|
2014-11-03 14:47:55 +01:00
|
|
|
|
2013-07-23 19:49:42 +02:00
|
|
|
for isp in range(self.n_spin_blocs):
|
2016-05-09 10:19:56 +02:00
|
|
|
# read real part:
|
|
|
|
for i in range(self.shells[ish]['dim']):
|
2014-11-26 16:24:02 +01:00
|
|
|
for j in range(self.n_orbitals[ik][isp]):
|
2016-05-09 10:19:56 +02:00
|
|
|
proj_mat_all[ik, isp, ish, ir, i, j] = R.next()
|
|
|
|
|
2013-07-23 19:49:42 +02:00
|
|
|
for isp in range(self.n_spin_blocs):
|
2016-05-09 10:19:56 +02:00
|
|
|
# read imaginary part:
|
|
|
|
for i in range(self.shells[ish]['dim']):
|
2014-11-26 16:24:02 +01:00
|
|
|
for j in range(self.n_orbitals[ik][isp]):
|
2016-05-09 10:19:56 +02:00
|
|
|
proj_mat_all[ik, isp, ish,
|
|
|
|
ir, i, j] += 1j * R.next()
|
|
|
|
|
|
|
|
# now read the Density Matrix for this orbital below the energy
|
|
|
|
# window:
|
2013-07-23 19:49:42 +02:00
|
|
|
for isp in range(self.n_spin_blocs):
|
2014-11-26 16:24:02 +01:00
|
|
|
for i in range(self.shells[ish]['dim']): # read real part:
|
|
|
|
for j in range(self.shells[ish]['dim']):
|
2016-05-09 10:19:56 +02:00
|
|
|
dens_mat_below[isp][ish][i, j] = R.next()
|
2013-07-23 19:49:42 +02:00
|
|
|
for isp in range(self.n_spin_blocs):
|
2016-05-09 10:19:56 +02:00
|
|
|
# read imaginary part:
|
|
|
|
for i in range(self.shells[ish]['dim']):
|
2014-11-26 16:24:02 +01:00
|
|
|
for j in range(self.shells[ish]['dim']):
|
2016-05-09 10:19:56 +02:00
|
|
|
dens_mat_below[isp][ish][i, j] += 1j * R.next()
|
|
|
|
if (self.SP == 0):
|
|
|
|
dens_mat_below[isp][ish] /= 2.0
|
2013-07-23 19:49:42 +02:00
|
|
|
|
|
|
|
# Global -> local rotation matrix for this shell:
|
2014-11-26 16:24:02 +01:00
|
|
|
for i in range(self.shells[ish]['dim']): # read real part:
|
|
|
|
for j in range(self.shells[ish]['dim']):
|
2016-05-09 10:19:56 +02:00
|
|
|
rot_mat_all[ish][i, j] = R.next()
|
2014-11-26 16:24:02 +01:00
|
|
|
for i in range(self.shells[ish]['dim']): # read imaginary part:
|
|
|
|
for j in range(self.shells[ish]['dim']):
|
2016-05-09 10:19:56 +02:00
|
|
|
rot_mat_all[ish][i, j] += 1j * R.next()
|
|
|
|
|
2013-07-23 19:49:42 +02:00
|
|
|
if (self.SP):
|
|
|
|
rot_mat_all_time_inv[ish] = int(R.next())
|
|
|
|
|
|
|
|
R.close()
|
2014-11-03 14:47:55 +01:00
|
|
|
# Reading done!
|
2013-07-23 19:49:42 +02:00
|
|
|
|
2014-11-03 14:47:55 +01:00
|
|
|
# Save it to the HDF:
|
2016-05-09 10:19:56 +02:00
|
|
|
ar = HDFArchive(self.hdf_file, 'a')
|
|
|
|
if not (self.parproj_subgrp in ar):
|
|
|
|
ar.create_group(self.parproj_subgrp)
|
|
|
|
# The subgroup containing the data. If it does not exist, it is
|
|
|
|
# created. If it exists, the data is overwritten!
|
|
|
|
things_to_save = ['dens_mat_below', 'n_parproj',
|
|
|
|
'proj_mat_all', 'rot_mat_all', 'rot_mat_all_time_inv']
|
|
|
|
for it in things_to_save:
|
|
|
|
ar[self.parproj_subgrp][it] = locals()[it]
|
2013-07-23 19:49:42 +02:00
|
|
|
del ar
|
|
|
|
|
2016-05-09 10:19:56 +02:00
|
|
|
# Symmetries are used, so now convert symmetry information for *all*
|
|
|
|
# orbitals:
|
|
|
|
self.convert_symmetry_input(orbits=self.shells, symm_file=self.symmpar_file,
|
|
|
|
symm_subgrp=self.symmpar_subgrp, SO=self.SO, SP=self.SP)
|
2013-07-23 19:49:42 +02:00
|
|
|
|
2014-11-03 14:47:55 +01:00
|
|
|
def convert_bands_input(self):
|
2013-07-23 19:49:42 +02:00
|
|
|
"""
|
2015-03-12 00:01:12 +01:00
|
|
|
Reads the appropriate files and stores the data for the bands_subgrp in the hdf5 archive.
|
|
|
|
|
2013-07-23 19:49:42 +02:00
|
|
|
"""
|
|
|
|
|
2016-05-09 10:19:56 +02:00
|
|
|
if not (mpi.is_master_node()):
|
|
|
|
return
|
2013-07-23 19:49:42 +02:00
|
|
|
|
|
|
|
try:
|
2015-07-07 15:08:53 +02:00
|
|
|
# get needed data from hdf file
|
2016-05-09 10:19:56 +02:00
|
|
|
ar = HDFArchive(self.hdf_file, 'a')
|
|
|
|
things_to_read = ['SP', 'SO', 'n_corr_shells',
|
|
|
|
'n_shells', 'corr_shells', 'shells', 'energy_unit']
|
2015-07-07 15:08:53 +02:00
|
|
|
|
|
|
|
for it in things_to_read:
|
2016-05-09 10:19:56 +02:00
|
|
|
if not hasattr(self, it):
|
|
|
|
setattr(self, it, ar[self.dft_subgrp][it])
|
2015-07-07 15:08:53 +02:00
|
|
|
self.n_spin_blocs = self.SP + 1 - self.SO
|
|
|
|
del ar
|
|
|
|
|
2016-05-09 10:19:56 +02:00
|
|
|
mpi.report("Reading input from %s..." % self.band_file)
|
|
|
|
R = ConverterTools.read_fortran_file(
|
|
|
|
self, self.band_file, self.fortran_to_replace)
|
2013-07-23 19:49:42 +02:00
|
|
|
n_k = int(R.next())
|
|
|
|
|
|
|
|
# read the list of n_orbitals for all k points
|
2016-05-09 10:19:56 +02:00
|
|
|
n_orbitals = numpy.zeros([n_k, self.n_spin_blocs], numpy.int)
|
2013-07-23 19:49:42 +02:00
|
|
|
for isp in range(self.n_spin_blocs):
|
2014-11-26 16:24:02 +01:00
|
|
|
for ik in range(n_k):
|
2016-05-09 10:19:56 +02:00
|
|
|
n_orbitals[ik, isp] = int(R.next())
|
2013-07-23 19:49:42 +02:00
|
|
|
|
|
|
|
# Initialise the projectors:
|
2016-05-09 10:19:56 +02:00
|
|
|
proj_mat = numpy.zeros([n_k, self.n_spin_blocs, self.n_corr_shells, max(
|
|
|
|
[crsh['dim'] for crsh in self.corr_shells]), numpy.max(n_orbitals)], numpy.complex_)
|
2013-07-23 19:49:42 +02:00
|
|
|
|
|
|
|
# Read the projectors from the file:
|
2014-11-26 16:24:02 +01:00
|
|
|
for ik in range(n_k):
|
2013-07-23 19:49:42 +02:00
|
|
|
for icrsh in range(self.n_corr_shells):
|
2014-11-26 16:24:02 +01:00
|
|
|
n_orb = self.corr_shells[icrsh]['dim']
|
2016-05-09 10:19:56 +02:00
|
|
|
# first Real part for BOTH spins, due to conventions in
|
|
|
|
# dmftproj:
|
2013-07-23 19:49:42 +02:00
|
|
|
for isp in range(self.n_spin_blocs):
|
2014-11-26 16:24:02 +01:00
|
|
|
for i in range(n_orb):
|
2016-05-09 10:19:56 +02:00
|
|
|
for j in range(n_orbitals[ik, isp]):
|
|
|
|
proj_mat[ik, isp, icrsh, i, j] = R.next()
|
2013-07-23 19:49:42 +02:00
|
|
|
# now Imag part:
|
|
|
|
for isp in range(self.n_spin_blocs):
|
2014-11-26 16:24:02 +01:00
|
|
|
for i in range(n_orb):
|
2016-05-09 10:19:56 +02:00
|
|
|
for j in range(n_orbitals[ik, isp]):
|
|
|
|
proj_mat[ik, isp, icrsh, i, j] += 1j * R.next()
|
|
|
|
|
|
|
|
hopping = numpy.zeros([n_k, self.n_spin_blocs, numpy.max(
|
|
|
|
n_orbitals), numpy.max(n_orbitals)], numpy.complex_)
|
2013-07-23 19:49:42 +02:00
|
|
|
|
|
|
|
# Grab the H
|
|
|
|
# we use now the convention of a DIAGONAL Hamiltonian!!!!
|
|
|
|
for isp in range(self.n_spin_blocs):
|
2016-05-09 10:19:56 +02:00
|
|
|
for ik in range(n_k):
|
|
|
|
n_orb = n_orbitals[ik, isp]
|
2014-11-26 16:24:02 +01:00
|
|
|
for i in range(n_orb):
|
2016-05-09 10:19:56 +02:00
|
|
|
hopping[ik, isp, i, i] = R.next() * self.energy_unit
|
2013-07-23 19:49:42 +02:00
|
|
|
|
|
|
|
# now read the partial projectors:
|
|
|
|
n_parproj = [int(R.next()) for i in range(self.n_shells)]
|
|
|
|
n_parproj = numpy.array(n_parproj)
|
2016-05-09 10:19:56 +02:00
|
|
|
|
2013-07-23 19:49:42 +02:00
|
|
|
# Initialise P, here a double list of matrices:
|
2016-05-09 10:19:56 +02:00
|
|
|
proj_mat_all = numpy.zeros([n_k, self.n_spin_blocs, self.n_shells, max(n_parproj), max(
|
|
|
|
[sh['dim'] for sh in self.shells]), numpy.max(n_orbitals)], numpy.complex_)
|
2013-07-23 19:49:42 +02:00
|
|
|
|
|
|
|
for ish in range(self.n_shells):
|
2014-11-26 16:24:02 +01:00
|
|
|
for ik in range(n_k):
|
2013-07-23 19:49:42 +02:00
|
|
|
for ir in range(n_parproj[ish]):
|
|
|
|
for isp in range(self.n_spin_blocs):
|
2016-05-09 10:19:56 +02:00
|
|
|
|
|
|
|
# read real part:
|
|
|
|
for i in range(self.shells[ish]['dim']):
|
|
|
|
for j in range(n_orbitals[ik, isp]):
|
|
|
|
proj_mat_all[ik, isp, ish,
|
|
|
|
ir, i, j] = R.next()
|
|
|
|
|
|
|
|
# read imaginary part:
|
|
|
|
for i in range(self.shells[ish]['dim']):
|
|
|
|
for j in range(n_orbitals[ik, isp]):
|
|
|
|
proj_mat_all[ik, isp, ish,
|
|
|
|
ir, i, j] += 1j * R.next()
|
2013-07-23 19:49:42 +02:00
|
|
|
|
2015-07-07 15:08:53 +02:00
|
|
|
R.close()
|
|
|
|
|
|
|
|
except KeyError:
|
|
|
|
raise "convert_bands_input : Needed data not found in hdf file. Consider calling convert_dft_input first!"
|
2016-05-09 10:19:56 +02:00
|
|
|
except StopIteration: # a more explicit error if the file is corrupted.
|
2014-09-22 19:24:33 +02:00
|
|
|
raise "Wien2k_converter : reading file band_file failed!"
|
2013-07-23 19:49:42 +02:00
|
|
|
|
2014-11-03 14:47:55 +01:00
|
|
|
# Reading done!
|
2013-07-23 19:49:42 +02:00
|
|
|
|
2014-11-03 14:47:55 +01:00
|
|
|
# Save it to the HDF:
|
2016-05-09 10:19:56 +02:00
|
|
|
ar = HDFArchive(self.hdf_file, 'a')
|
|
|
|
if not (self.bands_subgrp in ar):
|
|
|
|
ar.create_group(self.bands_subgrp)
|
|
|
|
# The subgroup containing the data. If it does not exist, it is
|
|
|
|
# created. If it exists, the data is overwritten!
|
|
|
|
things_to_save = ['n_k', 'n_orbitals', 'proj_mat',
|
|
|
|
'hopping', 'n_parproj', 'proj_mat_all']
|
|
|
|
for it in things_to_save:
|
|
|
|
ar[self.bands_subgrp][it] = locals()[it]
|
2013-07-23 19:49:42 +02:00
|
|
|
del ar
|
|
|
|
|
2015-03-12 00:01:12 +01:00
|
|
|
def convert_misc_input(self):
|
2014-11-03 11:16:32 +01:00
|
|
|
"""
|
2015-03-12 00:01:12 +01:00
|
|
|
Reads additional information on:
|
2016-05-09 10:19:56 +02:00
|
|
|
|
2015-03-12 00:01:12 +01:00
|
|
|
- the band window from :file:`case.oubwin`,
|
|
|
|
- lattice parameters from :file:`case.struct`,
|
|
|
|
- symmetries from :file:`case.outputs`,
|
2016-05-09 10:19:56 +02:00
|
|
|
|
2015-03-12 00:01:12 +01:00
|
|
|
if those Wien2k files are present and stores the data in the hdf5 archive.
|
|
|
|
This function is automatically called by :meth:`convert_dft_input <pytriqs.applications.dft.converters.wien2k_converter.Wien2kConverter.convert_dft_input>`.
|
|
|
|
|
2015-02-10 11:55:44 +01:00
|
|
|
"""
|
|
|
|
|
2016-05-09 10:19:56 +02:00
|
|
|
if not (mpi.is_master_node()):
|
|
|
|
return
|
|
|
|
|
2015-03-12 00:01:12 +01:00
|
|
|
# Check if SP, SO and n_k are already in h5
|
2015-08-10 16:14:51 +02:00
|
|
|
ar = HDFArchive(self.hdf_file, 'r')
|
2016-05-09 10:19:56 +02:00
|
|
|
if not (self.dft_subgrp in ar):
|
|
|
|
raise IOError, "convert_misc_input: No %s subgroup in hdf file found! Call convert_dft_input first." % self.dft_subgrp
|
2015-03-12 00:01:12 +01:00
|
|
|
SP = ar[self.dft_subgrp]['SP']
|
|
|
|
SO = ar[self.dft_subgrp]['SO']
|
|
|
|
n_k = ar[self.dft_subgrp]['n_k']
|
|
|
|
del ar
|
2016-05-09 10:19:56 +02:00
|
|
|
|
2015-02-24 18:14:05 +01:00
|
|
|
things_to_save = []
|
2014-11-03 11:16:32 +01:00
|
|
|
|
2015-02-10 11:55:44 +01:00
|
|
|
# Read relevant data from .oubwin/up/dn files
|
|
|
|
#############################################
|
|
|
|
# band_window: Contains the index of the lowest and highest band within the
|
|
|
|
# projected subspace (used by dmftproj) for each k-point.
|
2014-12-19 11:53:06 +01:00
|
|
|
|
2016-05-09 10:19:56 +02:00
|
|
|
if (SP == 0 or SO == 1):
|
2015-02-10 11:55:44 +01:00
|
|
|
files = [self.bandwin_file]
|
2014-12-11 12:11:48 +01:00
|
|
|
elif SP == 1:
|
2016-05-09 10:19:56 +02:00
|
|
|
files = [self.bandwin_file + 'up', self.bandwin_file + 'dn']
|
|
|
|
else: # SO and SP can't both be 1
|
2015-07-07 15:26:04 +02:00
|
|
|
assert 0, "convert_misc_input: Reading oubwin error! Check SP and SO!"
|
|
|
|
|
|
|
|
band_window = [None for isp in range(SP + 1 - SO)]
|
2014-12-11 12:11:48 +01:00
|
|
|
for isp, f in enumerate(files):
|
2015-02-20 15:15:17 +01:00
|
|
|
if os.path.exists(f):
|
2016-05-09 10:19:56 +02:00
|
|
|
mpi.report("Reading input from %s..." % f)
|
|
|
|
R = ConverterTools.read_fortran_file(
|
|
|
|
self, f, self.fortran_to_replace)
|
2015-07-07 15:26:04 +02:00
|
|
|
n_k_oubwin = int(R.next())
|
|
|
|
if (n_k_oubwin != n_k):
|
2016-05-09 10:19:56 +02:00
|
|
|
mpi.report(
|
|
|
|
"convert_misc_input : WARNING : n_k in case.oubwin is different from n_k in case.klist")
|
|
|
|
assert int(
|
|
|
|
R.next()) == SO, "convert_misc_input: SO is inconsistent in oubwin file!"
|
2015-07-07 15:26:04 +02:00
|
|
|
|
2016-05-09 10:19:56 +02:00
|
|
|
band_window[isp] = numpy.zeros((n_k_oubwin, 2), dtype=int)
|
2015-07-07 15:26:04 +02:00
|
|
|
for ik in xrange(n_k_oubwin):
|
2015-02-20 15:15:17 +01:00
|
|
|
R.next()
|
2016-05-09 10:19:56 +02:00
|
|
|
band_window[isp][ik, 0] = R.next() # lowest band
|
|
|
|
band_window[isp][ik, 1] = R.next() # highest band
|
2015-02-20 15:15:17 +01:00
|
|
|
R.next()
|
2015-02-24 18:14:05 +01:00
|
|
|
things_to_save.append('band_window')
|
2015-02-10 11:55:44 +01:00
|
|
|
|
2016-05-09 10:19:56 +02:00
|
|
|
R.close() # Reading done!
|
2014-11-03 11:16:32 +01:00
|
|
|
|
|
|
|
# Read relevant data from .struct file
|
2014-12-19 11:53:06 +01:00
|
|
|
######################################
|
|
|
|
# lattice_type: bravais lattice type as defined by Wien2k
|
|
|
|
# lattice_constants: unit cell parameters in a. u.
|
|
|
|
# lattice_angles: unit cell angles in rad
|
|
|
|
|
2015-02-20 15:15:17 +01:00
|
|
|
if (os.path.exists(self.struct_file)):
|
2016-05-09 10:19:56 +02:00
|
|
|
mpi.report("Reading input from %s..." % self.struct_file)
|
|
|
|
|
2015-02-20 15:15:17 +01:00
|
|
|
with open(self.struct_file) as R:
|
|
|
|
try:
|
|
|
|
R.readline()
|
|
|
|
lattice_type = R.readline().split()[0]
|
|
|
|
R.readline()
|
2015-03-03 18:43:50 +01:00
|
|
|
temp = R.readline()
|
2016-05-09 10:19:56 +02:00
|
|
|
lattice_constants = numpy.array(
|
|
|
|
[float(temp[0 + 10 * i:10 + 10 * i].strip()) for i in range(3)])
|
|
|
|
lattice_angles = numpy.array(
|
|
|
|
[float(temp[30 + 10 * i:40 + 10 * i].strip()) for i in range(3)]) * numpy.pi / 180.0
|
|
|
|
things_to_save.extend(
|
|
|
|
['lattice_type', 'lattice_constants', 'lattice_angles'])
|
2015-02-20 15:15:17 +01:00
|
|
|
except IOError:
|
2016-05-09 10:19:56 +02:00
|
|
|
raise "convert_misc_input: reading file %s failed" % self.struct_file
|
2014-11-03 11:16:32 +01:00
|
|
|
|
|
|
|
# Read relevant data from .outputs file
|
2014-12-19 11:53:06 +01:00
|
|
|
#######################################
|
2016-05-09 10:19:56 +02:00
|
|
|
# rot_symmetries: matrix representation of all (space group) symmetry
|
|
|
|
# operations
|
|
|
|
|
2015-02-20 15:15:17 +01:00
|
|
|
if (os.path.exists(self.outputs_file)):
|
2016-05-09 10:19:56 +02:00
|
|
|
mpi.report("Reading input from %s..." % self.outputs_file)
|
|
|
|
|
2015-02-20 15:15:17 +01:00
|
|
|
rot_symmetries = []
|
|
|
|
with open(self.outputs_file) as R:
|
|
|
|
try:
|
2014-11-03 11:16:32 +01:00
|
|
|
while 1:
|
2015-02-20 15:15:17 +01:00
|
|
|
temp = R.readline().strip(' ').split()
|
2016-05-09 10:19:56 +02:00
|
|
|
if (temp[0] == 'PGBSYM:'):
|
2015-02-20 15:15:17 +01:00
|
|
|
n_symmetries = int(temp[-1])
|
|
|
|
break
|
|
|
|
for i in range(n_symmetries):
|
|
|
|
while 1:
|
2016-05-09 10:19:56 +02:00
|
|
|
if (R.readline().strip().split()[0] == 'Symmetry'):
|
|
|
|
break
|
|
|
|
sym_i = numpy.zeros((3, 3), dtype=float)
|
2015-02-20 15:15:17 +01:00
|
|
|
for ir in range(3):
|
|
|
|
temp = R.readline().strip().split()
|
|
|
|
for ic in range(3):
|
|
|
|
sym_i[ir, ic] = float(temp[ic])
|
|
|
|
R.readline()
|
|
|
|
rot_symmetries.append(sym_i)
|
2015-02-24 18:14:05 +01:00
|
|
|
things_to_save.extend(['n_symmetries', 'rot_symmetries'])
|
|
|
|
things_to_save.append('rot_symmetries')
|
2015-02-20 15:15:17 +01:00
|
|
|
except IOError:
|
2016-05-09 10:19:56 +02:00
|
|
|
raise "convert_misc_input: reading file %s failed" % self.outputs_file
|
2015-02-10 11:55:44 +01:00
|
|
|
|
|
|
|
# Save it to the HDF:
|
2016-05-09 10:19:56 +02:00
|
|
|
ar = HDFArchive(self.hdf_file, 'a')
|
|
|
|
if not (self.misc_subgrp in ar):
|
|
|
|
ar.create_group(self.misc_subgrp)
|
|
|
|
for it in things_to_save:
|
|
|
|
ar[self.misc_subgrp][it] = locals()[it]
|
2015-02-10 11:55:44 +01:00
|
|
|
del ar
|
|
|
|
|
|
|
|
def convert_transport_input(self):
|
|
|
|
"""
|
2015-03-12 00:01:12 +01:00
|
|
|
Reads the necessary information for transport calculations on:
|
2016-05-09 10:19:56 +02:00
|
|
|
|
2015-03-12 00:01:12 +01:00
|
|
|
- the optical band window and the velocity matrix elements from :file:`case.pmat`
|
|
|
|
|
|
|
|
and stores the data in the hdf5 archive.
|
2016-05-09 10:19:56 +02:00
|
|
|
|
2015-02-10 11:55:44 +01:00
|
|
|
"""
|
2016-05-09 10:19:56 +02:00
|
|
|
|
|
|
|
if not (mpi.is_master_node()):
|
|
|
|
return
|
|
|
|
|
2015-02-10 11:55:44 +01:00
|
|
|
# Check if SP, SO and n_k are already in h5
|
2015-08-10 16:14:51 +02:00
|
|
|
ar = HDFArchive(self.hdf_file, 'r')
|
2016-05-09 10:19:56 +02:00
|
|
|
if not (self.dft_subgrp in ar):
|
|
|
|
raise IOError, "convert_transport_input: No %s subgroup in hdf file found! Call convert_dft_input first." % self.dft_subgrp
|
2015-02-10 11:55:44 +01:00
|
|
|
SP = ar[self.dft_subgrp]['SP']
|
|
|
|
SO = ar[self.dft_subgrp]['SO']
|
|
|
|
n_k = ar[self.dft_subgrp]['n_k']
|
|
|
|
del ar
|
|
|
|
|
|
|
|
# Read relevant data from .pmat/up/dn files
|
|
|
|
###########################################
|
|
|
|
# band_window_optics: Contains the index of the lowest and highest band within the
|
|
|
|
# band window (used by optics) for each k-point.
|
|
|
|
# velocities_k: velocity (momentum) matrix elements between all bands in band_window_optics
|
|
|
|
# and each k-point.
|
|
|
|
|
2016-05-09 10:19:56 +02:00
|
|
|
if (SP == 0 or SO == 1):
|
2015-02-10 11:55:44 +01:00
|
|
|
files = [self.pmat_file]
|
2014-12-09 16:47:40 +01:00
|
|
|
elif SP == 1:
|
2016-05-09 10:19:56 +02:00
|
|
|
files = [self.pmat_file + 'up', self.pmat_file + 'dn']
|
|
|
|
else: # SO and SP can't both be 1
|
2015-02-10 11:55:44 +01:00
|
|
|
assert 0, "convert_transport_input: Reading velocity file error! Check SP and SO!"
|
|
|
|
|
|
|
|
velocities_k = [[] for f in files]
|
|
|
|
band_window_optics = []
|
2014-12-09 16:47:40 +01:00
|
|
|
for isp, f in enumerate(files):
|
2016-05-09 10:19:56 +02:00
|
|
|
if not os.path.exists(f):
|
|
|
|
raise IOError, "convert_transport_input: File %s does not exist" % f
|
|
|
|
mpi.report("Reading input from %s..." % f)
|
2015-02-10 11:55:44 +01:00
|
|
|
|
2016-05-09 10:19:56 +02:00
|
|
|
R = ConverterTools.read_fortran_file(
|
|
|
|
self, f, {'D': 'E', '(': '', ')': '', ',': ' '})
|
2015-02-10 11:55:44 +01:00
|
|
|
band_window_optics_isp = []
|
2014-12-09 16:47:40 +01:00
|
|
|
for ik in xrange(n_k):
|
|
|
|
R.next()
|
2015-02-10 11:55:44 +01:00
|
|
|
nu1 = int(R.next())
|
|
|
|
nu2 = int(R.next())
|
|
|
|
band_window_optics_isp.append((nu1, nu2))
|
|
|
|
n_bands = nu2 - nu1 + 1
|
2016-05-09 10:19:56 +02:00
|
|
|
for _ in range(4):
|
|
|
|
R.next()
|
2015-02-10 11:55:44 +01:00
|
|
|
if n_bands <= 0:
|
2016-05-09 10:19:56 +02:00
|
|
|
velocity_xyz = numpy.zeros((1, 1, 3), dtype=complex)
|
2015-02-10 11:55:44 +01:00
|
|
|
else:
|
2016-05-09 10:19:56 +02:00
|
|
|
velocity_xyz = numpy.zeros(
|
|
|
|
(n_bands, n_bands, 3), dtype=complex)
|
2015-02-10 11:55:44 +01:00
|
|
|
for nu_i in range(n_bands):
|
|
|
|
for nu_j in range(nu_i, n_bands):
|
|
|
|
for i in range(3):
|
2016-05-09 10:19:56 +02:00
|
|
|
velocity_xyz[nu_i][nu_j][
|
|
|
|
i] = R.next() + R.next() * 1j
|
|
|
|
if (nu_i != nu_j):
|
|
|
|
velocity_xyz[nu_j][nu_i][i] = velocity_xyz[
|
|
|
|
nu_i][nu_j][i].conjugate()
|
2015-02-10 11:55:44 +01:00
|
|
|
velocities_k[isp].append(velocity_xyz)
|
|
|
|
band_window_optics.append(numpy.array(band_window_optics_isp))
|
2016-05-09 10:19:56 +02:00
|
|
|
R.close() # Reading done!
|
2014-11-03 11:16:32 +01:00
|
|
|
|
|
|
|
# Put data to HDF5 file
|
|
|
|
ar = HDFArchive(self.hdf_file, 'a')
|
2016-05-09 10:19:56 +02:00
|
|
|
if not (self.transp_subgrp in ar):
|
|
|
|
ar.create_group(self.transp_subgrp)
|
|
|
|
# The subgroup containing the data. If it does not exist, it is
|
|
|
|
# created. If it exists, the data is overwritten!!!
|
2015-02-10 11:55:44 +01:00
|
|
|
things_to_save = ['band_window_optics', 'velocities_k']
|
2016-05-09 10:19:56 +02:00
|
|
|
for it in things_to_save:
|
|
|
|
ar[self.transp_subgrp][it] = locals()[it]
|
2014-11-03 11:16:32 +01:00
|
|
|
del ar
|
|
|
|
|
2014-10-31 18:52:32 +01:00
|
|
|
def convert_symmetry_input(self, orbits, symm_file, symm_subgrp, SO, SP):
|
2013-07-23 19:49:42 +02:00
|
|
|
"""
|
2015-03-12 00:01:12 +01:00
|
|
|
Reads and stores symmetrisation data from symm_file, which can be is case.sympar or case.symqmc.
|
|
|
|
|
|
|
|
Parameters
|
|
|
|
----------
|
|
|
|
orbits : list of dicts
|
|
|
|
This is either shells or corr_shells depending on whether the symmetry
|
|
|
|
information is for correlated shells or partial projectors.
|
|
|
|
symm_file : string
|
|
|
|
Name of the file containing symmetry data.
|
|
|
|
This is case.symqmc for correlated shells and case.sympar for partial projectors.
|
|
|
|
symm_subgrp : string, optional
|
|
|
|
Name of subgroup storing symmetry data.
|
|
|
|
SO : integer
|
|
|
|
Is spin-orbit coupling considered?
|
|
|
|
SP : integer
|
|
|
|
Is the system spin-polarised?
|
|
|
|
|
2013-07-23 19:49:42 +02:00
|
|
|
"""
|
|
|
|
|
2016-05-09 10:19:56 +02:00
|
|
|
if not (mpi.is_master_node()):
|
|
|
|
return
|
|
|
|
mpi.report("Reading input from %s..." % symm_file)
|
2013-07-23 19:49:42 +02:00
|
|
|
|
|
|
|
n_orbits = len(orbits)
|
2014-11-08 23:51:24 +01:00
|
|
|
|
2016-05-09 10:19:56 +02:00
|
|
|
R = ConverterTools.read_fortran_file(
|
|
|
|
self, symm_file, self.fortran_to_replace)
|
2013-07-23 19:49:42 +02:00
|
|
|
|
|
|
|
try:
|
2014-11-14 18:13:43 +01:00
|
|
|
n_symm = int(R.next()) # Number of symmetry operations
|
2013-07-23 19:49:42 +02:00
|
|
|
n_atoms = int(R.next()) # number of atoms involved
|
2016-05-09 10:19:56 +02:00
|
|
|
perm = [[int(R.next()) for i in range(n_atoms)]
|
|
|
|
for j in range(n_symm)] # list of permutations of the atoms
|
|
|
|
if SP:
|
|
|
|
# time inversion for SO coupling
|
|
|
|
time_inv = [int(R.next()) for j in range(n_symm)]
|
2013-07-23 19:49:42 +02:00
|
|
|
else:
|
2016-05-09 10:19:56 +02:00
|
|
|
time_inv = [0 for j in range(n_symm)]
|
2013-07-23 19:49:42 +02:00
|
|
|
|
|
|
|
# Now read matrices:
|
2016-05-09 10:19:56 +02:00
|
|
|
mat = []
|
2014-11-26 16:24:02 +01:00
|
|
|
for i_symm in range(n_symm):
|
2016-05-09 10:19:56 +02:00
|
|
|
|
|
|
|
mat.append([numpy.zeros([orbits[orb]['dim'], orbits[orb][
|
|
|
|
'dim']], numpy.complex_) for orb in range(n_orbits)])
|
2013-07-23 19:49:42 +02:00
|
|
|
for orb in range(n_orbits):
|
2014-11-26 16:24:02 +01:00
|
|
|
for i in range(orbits[orb]['dim']):
|
|
|
|
for j in range(orbits[orb]['dim']):
|
2016-05-09 10:19:56 +02:00
|
|
|
# real part
|
|
|
|
mat[i_symm][orb][i, j] = R.next()
|
2014-11-26 16:24:02 +01:00
|
|
|
for i in range(orbits[orb]['dim']):
|
|
|
|
for j in range(orbits[orb]['dim']):
|
2016-05-09 10:19:56 +02:00
|
|
|
mat[i_symm][orb][i, j] += 1j * \
|
|
|
|
R.next() # imaginary part
|
2013-07-23 19:49:42 +02:00
|
|
|
|
2016-05-09 10:19:56 +02:00
|
|
|
mat_tinv = [numpy.identity(orbits[orb]['dim'], numpy.complex_)
|
2013-07-23 19:49:42 +02:00
|
|
|
for orb in range(n_orbits)]
|
|
|
|
|
2016-05-09 10:19:56 +02:00
|
|
|
if ((SO == 0) and (SP == 0)):
|
|
|
|
# here we need an additional time inversion operation, so read
|
|
|
|
# it:
|
2013-07-23 19:49:42 +02:00
|
|
|
for orb in range(n_orbits):
|
2014-11-26 16:24:02 +01:00
|
|
|
for i in range(orbits[orb]['dim']):
|
|
|
|
for j in range(orbits[orb]['dim']):
|
2016-05-09 10:19:56 +02:00
|
|
|
# real part
|
|
|
|
mat_tinv[orb][i, j] = R.next()
|
2014-11-26 16:24:02 +01:00
|
|
|
for i in range(orbits[orb]['dim']):
|
|
|
|
for j in range(orbits[orb]['dim']):
|
2016-05-09 10:19:56 +02:00
|
|
|
mat_tinv[orb][i, j] += 1j * \
|
|
|
|
R.next() # imaginary part
|
2013-07-23 19:49:42 +02:00
|
|
|
|
2016-05-09 10:19:56 +02:00
|
|
|
except StopIteration: # a more explicit error if the file is corrupted.
|
2014-09-22 19:24:33 +02:00
|
|
|
raise "Wien2k_converter : reading file symm_file failed!"
|
2016-05-09 10:19:56 +02:00
|
|
|
|
2013-07-23 19:49:42 +02:00
|
|
|
R.close()
|
2014-11-03 14:47:55 +01:00
|
|
|
# Reading done!
|
2013-07-23 19:49:42 +02:00
|
|
|
|
|
|
|
# Save it to the HDF:
|
2016-05-09 10:19:56 +02:00
|
|
|
ar = HDFArchive(self.hdf_file, 'a')
|
|
|
|
if not (symm_subgrp in ar):
|
|
|
|
ar.create_group(symm_subgrp)
|
|
|
|
things_to_save = ['n_symm', 'n_atoms', 'perm',
|
|
|
|
'orbits', 'SO', 'SP', 'time_inv', 'mat', 'mat_tinv']
|
|
|
|
for it in things_to_save:
|
|
|
|
ar[symm_subgrp][it] = locals()[it]
|
2013-07-23 19:49:42 +02:00
|
|
|
del ar
|