mirror of
https://github.com/triqs/dft_tools
synced 2024-10-31 19:23:45 +01:00
Minimal changes in the doc
This commit is contained in:
parent
eb8060c290
commit
217f933634
9
doc/_templates/sideb_appli.html
vendored
9
doc/_templates/sideb_appli.html
vendored
@ -1,9 +0,0 @@
|
|||||||
<p>
|
|
||||||
<a href="http://ipht.cea.fr"> <img style="width: 80px; margin: 10px 5px 0 0" src='_static/logo_cea.png' alt="CEA"/> </a>
|
|
||||||
<a href="http://www.cpht.polytechnique.fr"> <img style="width: 80px; margin: 10px 5px 0 5px" src='_static/logo_x.png' alt="Ecole Polytechnique"/> </a>
|
|
||||||
<br>
|
|
||||||
<a href="http://www.cnrs.fr"> <img style="width: 80px; margin: 10px 0 0 5px" src='_static/logo_cnrs.png' alt="CNRS"/> </a>
|
|
||||||
<img style="width: 80px; margin: 10px 0 0 5px" src='_static/logo_erc.jpg' alt="ERC"/>
|
|
||||||
</p>
|
|
||||||
|
|
||||||
|
|
@ -4,11 +4,16 @@
|
|||||||
Changelog
|
Changelog
|
||||||
=========
|
=========
|
||||||
|
|
||||||
This document describes the main changes in TRIQS 1.0 that might
|
This document describes the main changes in TRIQS.
|
||||||
have consequences for your scripts and archives.
|
|
||||||
|
From TRIQS 0.x to TRIQS 1.0
|
||||||
|
---------------------------
|
||||||
|
|
||||||
|
There have been changes from versions 0.x to 1.0 that will most likely have
|
||||||
|
consequences for your scripts and archives.
|
||||||
|
|
||||||
Python classes
|
Python classes
|
||||||
--------------
|
~~~~~~~~~~~~~~
|
||||||
|
|
||||||
The number of python classes in the old 0.x releases was increasing with no
|
The number of python classes in the old 0.x releases was increasing with no
|
||||||
clear naming convention. In TRIQS 1.0 we have unified the naming of the classes
|
clear naming convention. In TRIQS 1.0 we have unified the naming of the classes
|
||||||
@ -17,9 +22,17 @@ following the `PEP naming conventions
|
|||||||
|
|
||||||
* Package and module names: lowercase with underscores
|
* Package and module names: lowercase with underscores
|
||||||
* Class names: CapWords
|
* Class names: CapWords
|
||||||
|
* Function names: lowercase with underscores
|
||||||
* Function and method arguments: lowercase with underscores
|
* Function and method arguments: lowercase with underscores
|
||||||
|
|
||||||
Archives
|
Archives
|
||||||
--------
|
~~~~~~~~
|
||||||
|
|
||||||
|
We provide :download:`an update script <scripts/update_archive.py>` which should
|
||||||
|
help you upgrade your archive. The usage is very simple::
|
||||||
|
|
||||||
|
$ python update_archive.py old_archive new_archive
|
||||||
|
|
||||||
|
where ``old_archive`` is your old archive to be upgraded and ``new_archive`` is
|
||||||
|
the name of the new archive. If you encounter problems please post an
|
||||||
|
issue with a copy of your archive.
|
||||||
|
@ -10,5 +10,6 @@ Table of contents
|
|||||||
documentation
|
documentation
|
||||||
applications
|
applications
|
||||||
issues
|
issues
|
||||||
|
changelog
|
||||||
about
|
about
|
||||||
|
|
||||||
|
@ -80,4 +80,3 @@ Further reading
|
|||||||
installation/install_options
|
installation/install_options
|
||||||
installation/python
|
installation/python
|
||||||
installation/clang
|
installation/clang
|
||||||
installation/changelog
|
|
||||||
|
222
doc/scripts/update_archive.py
Normal file
222
doc/scripts/update_archive.py
Normal file
@ -0,0 +1,222 @@
|
|||||||
|
import h5py
|
||||||
|
import sys
|
||||||
|
import numpy
|
||||||
|
|
||||||
|
if len(sys.argv) < 3:
|
||||||
|
print "Usage: python update_archive.py old_archive new_archive"
|
||||||
|
sys.exit()
|
||||||
|
|
||||||
|
print """
|
||||||
|
This script is an attempt to update your archive to TRIQS 1.0.
|
||||||
|
Please keep a copy of your old archive as this script is
|
||||||
|
** not guaranteed ** to work for your archive.
|
||||||
|
If you encounter any problem please report it on github!
|
||||||
|
"""
|
||||||
|
|
||||||
|
# The dictionary of modifications to make
|
||||||
|
modifications = {
|
||||||
|
'GF': {'SSS': 'BlockGf', '__Name': 'name', '__Note': 'note', '__BlockIndicesList': 'indices'},
|
||||||
|
'GFBloc_ImFreq': {'SSS': 'GfImFreq'},
|
||||||
|
'GFBloc_ReFreq': {'SSS': 'GfReFreq'},
|
||||||
|
'GFBloc_ImTime': {'SSS': 'GfImTime'},
|
||||||
|
'GFBloc_ReTime': {'SSS': 'GfReTime'},
|
||||||
|
'GFBloc_ImLegendre': {'SSS': 'GfLegendre'},
|
||||||
|
'MeshGF': {'SSS': 'MeshGf'},
|
||||||
|
'TailGF': {'SSS': 'TailGf'},
|
||||||
|
'DOS': {'Name': 'name'},
|
||||||
|
'BlockGf': {'__Name': 'name', '__Note': 'note', '__BlockIndicesList': 'indices'}
|
||||||
|
}
|
||||||
|
|
||||||
|
|
||||||
|
def sanitize_imfreq(origin, dest):
|
||||||
|
|
||||||
|
dest.create_dataset(name="data", data=origin["Data"].value.transpose((2,0,1,3)))
|
||||||
|
dest["data"].attrs.create('__complex__', "1")
|
||||||
|
|
||||||
|
dest.create_group(name="indices")
|
||||||
|
exec("indL = %s"%origin["IndicesL"].value)
|
||||||
|
exec("indR = %s"%origin["IndicesR"].value)
|
||||||
|
indL = [ str(i) for i in indL ]
|
||||||
|
indR = [ str(i) for i in indR ]
|
||||||
|
dest["indices"].create_dataset(name="left", data=indL)
|
||||||
|
dest["indices"].create_dataset(name="right", data=indR)
|
||||||
|
|
||||||
|
dest.create_group(name="singularity")
|
||||||
|
dest["singularity"].create_dataset(name="data", data=origin["Tail"]["array"].value.transpose((2,0,1,3)))
|
||||||
|
dest["singularity"]["data"].attrs.create('__complex__', "1")
|
||||||
|
dest["singularity"].create_dataset(name="omin", data=origin["Tail"]["OrderMinMIN"].value)
|
||||||
|
mask = numpy.zeros( dest["singularity"]["data"].shape[0:2], numpy.integer )
|
||||||
|
mask.fill(origin["Tail"]["OrderMax"].value)
|
||||||
|
dest["singularity"].create_dataset(name="mask", data=mask)
|
||||||
|
|
||||||
|
dest.create_group(name="mesh")
|
||||||
|
beta = origin["Mesh"]["Beta"].value
|
||||||
|
pi = numpy.arccos(-1)
|
||||||
|
size = numpy.uint(len(origin["Mesh"]["array"].value))
|
||||||
|
dest["mesh"].create_dataset(name="kind", data=2)
|
||||||
|
dest["mesh"].create_dataset(name="min", data=pi/beta)
|
||||||
|
dest["mesh"].create_dataset(name="max", data=(2*size+1)*pi/beta)
|
||||||
|
dest["mesh"].create_dataset(name="size", data=size)
|
||||||
|
dest["mesh"].create_group(name="domain")
|
||||||
|
dest["mesh"]["domain"].create_dataset(name="beta", data=beta)
|
||||||
|
dest["mesh"]["domain"].create_dataset(name="statistic", data={"Fermion":"F", "Boson":"B"}[origin["Mesh"]["Statistic"].value] )
|
||||||
|
|
||||||
|
return ['Data', 'IndicesL', 'IndicesR', 'Mesh', 'Name', 'Note', 'Tail']
|
||||||
|
|
||||||
|
|
||||||
|
def sanitize_imtime(origin, dest):
|
||||||
|
|
||||||
|
dest.create_dataset(name="data", data=origin["Data"].value.transpose((2,0,1)))
|
||||||
|
|
||||||
|
dest.create_group(name="indices")
|
||||||
|
exec("indL = %s"%origin["IndicesL"].value)
|
||||||
|
exec("indR = %s"%origin["IndicesR"].value)
|
||||||
|
indL = [ str(i) for i in indL ]
|
||||||
|
indR = [ str(i) for i in indR ]
|
||||||
|
dest["indices"].create_dataset(name="left", data=indL)
|
||||||
|
dest["indices"].create_dataset(name="right", data=indR)
|
||||||
|
|
||||||
|
dest.create_group(name="singularity")
|
||||||
|
dest["singularity"].create_dataset(name="data", data=origin["Tail"]["array"].value.transpose((2,0,1,3)))
|
||||||
|
dest["singularity"]["data"].attrs.create('__complex__', "1")
|
||||||
|
dest["singularity"].create_dataset(name="omin", data=origin["Tail"]["OrderMinMIN"].value)
|
||||||
|
mask = numpy.zeros( dest["singularity"]["data"].shape[0:2], numpy.integer )
|
||||||
|
mask.fill(origin["Tail"]["OrderMax"].value)
|
||||||
|
dest["singularity"].create_dataset(name="mask", data=mask)
|
||||||
|
|
||||||
|
dest.create_group(name="mesh")
|
||||||
|
beta = origin["Mesh"]["Beta"].value
|
||||||
|
size = numpy.uint(len(origin["Mesh"]["array"].value))
|
||||||
|
min_t = origin["Mesh"]["array"].value[0]
|
||||||
|
if min_t > 1e-10:
|
||||||
|
kind = 0
|
||||||
|
assert(abs(min_t - 0.5*beta/size) < 1e-10)
|
||||||
|
else:
|
||||||
|
kind = 2
|
||||||
|
dest["mesh"].create_dataset(name="kind", data=kind)
|
||||||
|
dest["mesh"].create_dataset(name="min", data=0.0)
|
||||||
|
dest["mesh"].create_dataset(name="max", data=beta)
|
||||||
|
dest["mesh"].create_dataset(name="size", data=size)
|
||||||
|
dest["mesh"].create_group(name="domain")
|
||||||
|
dest["mesh"]["domain"].create_dataset(name="beta", data=beta)
|
||||||
|
dest["mesh"]["domain"].create_dataset(name="statistic", data={"Fermion":"F", "Boson":"B"}[origin["Mesh"]["Statistic"].value] )
|
||||||
|
|
||||||
|
return ['Data', 'IndicesL', 'IndicesR', 'Mesh', 'Name', 'Note', 'Tail']
|
||||||
|
|
||||||
|
def sanitize_legendre(origin, dest):
|
||||||
|
|
||||||
|
dest.create_dataset(name="data", data=origin["Data"].value.transpose((2,0,1,3)))
|
||||||
|
|
||||||
|
dest.create_group(name="indices")
|
||||||
|
exec("indL = %s"%origin["IndicesL"].value)
|
||||||
|
exec("indR = %s"%origin["IndicesR"].value)
|
||||||
|
indL = [ str(i) for i in indL ]
|
||||||
|
indR = [ str(i) for i in indR ]
|
||||||
|
dest["indices"].create_dataset(name="left", data=indL)
|
||||||
|
dest["indices"].create_dataset(name="right", data=indR)
|
||||||
|
|
||||||
|
dest.create_group(name="mesh")
|
||||||
|
beta = origin["Mesh"]["Beta"].value
|
||||||
|
size = numpy.uint(len(origin["Mesh"]["array"].value))
|
||||||
|
dest["mesh"].create_group(name="domain")
|
||||||
|
dest["mesh"]["domain"].create_dataset(name="beta", data=beta)
|
||||||
|
dest["mesh"]["domain"].create_dataset(name="n_max", data=size)
|
||||||
|
dest["mesh"]["domain"].create_dataset(name="statistic", data={"Fermion":"F", "Boson":"B"}[origin["Mesh"]["Statistic"].value] )
|
||||||
|
|
||||||
|
return ['Data', 'IndicesL', 'IndicesR', 'Mesh', 'Name', 'Note', 'Tail']
|
||||||
|
|
||||||
|
def sanitize_refreq(origin, dest):
|
||||||
|
|
||||||
|
dest.create_dataset(name="data", data=origin["Data"].value.transpose((2,0,1,3)))
|
||||||
|
dest["data"].attrs.create('__complex__', "1")
|
||||||
|
|
||||||
|
dest.create_group(name="indices")
|
||||||
|
exec("indL = %s"%origin["IndicesL"].value)
|
||||||
|
exec("indR = %s"%origin["IndicesR"].value)
|
||||||
|
indL = [ str(i) for i in indL ]
|
||||||
|
indR = [ str(i) for i in indR ]
|
||||||
|
dest["indices"].create_dataset(name="left", data=indL)
|
||||||
|
dest["indices"].create_dataset(name="right", data=indR)
|
||||||
|
|
||||||
|
dest.create_group(name="singularity")
|
||||||
|
dest["singularity"].create_dataset(name="data", data=origin["Tail"]["array"].value.transpose((2,0,1,3)))
|
||||||
|
dest["singularity"]["data"].attrs.create('__complex__', "1")
|
||||||
|
dest["singularity"].create_dataset(name="omin", data=origin["Tail"]["OrderMinMIN"].value)
|
||||||
|
mask = numpy.zeros( dest["singularity"]["data"].shape[0:2], numpy.integer )
|
||||||
|
mask.fill(origin["Tail"]["OrderMax"].value)
|
||||||
|
dest["singularity"].create_dataset(name="mask", data=mask)
|
||||||
|
|
||||||
|
dest.create_group(name="mesh")
|
||||||
|
size = numpy.uint(len(origin["Mesh"]["array"].value))
|
||||||
|
min_w = origin["Mesh"]["array"].value[0]
|
||||||
|
max_w = origin["Mesh"]["array"].value[-1]
|
||||||
|
dest["mesh"].create_dataset(name="kind", data=1)
|
||||||
|
dest["mesh"].create_dataset(name="min", data=min_w)
|
||||||
|
dest["mesh"].create_dataset(name="max", data=max_w)
|
||||||
|
dest["mesh"].create_dataset(name="size", data=size)
|
||||||
|
|
||||||
|
return ['Data', 'IndicesL', 'IndicesR', 'Mesh', 'Name', 'Note', 'Tail']
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
|
def copy_and_correct(origin, dest):
|
||||||
|
|
||||||
|
scheme = None
|
||||||
|
exclude = []
|
||||||
|
|
||||||
|
# copy attributes in destination archive
|
||||||
|
for at in origin.attrs.keys():
|
||||||
|
|
||||||
|
# figure if the group is a TRIQS scheme
|
||||||
|
if at == 'TRIQS_HDF5_data_scheme':
|
||||||
|
|
||||||
|
scheme = origin.attrs['TRIQS_HDF5_data_scheme']
|
||||||
|
try:
|
||||||
|
new_name = modifications[scheme]['SSS']
|
||||||
|
except:
|
||||||
|
new_name = scheme
|
||||||
|
dest.attrs.create('TRIQS_HDF5_data_scheme', new_name)
|
||||||
|
|
||||||
|
else:
|
||||||
|
|
||||||
|
dest.attrs.create(at, origin.attrs[at])
|
||||||
|
|
||||||
|
# some special cases here
|
||||||
|
if scheme == "GFBloc_ImFreq" or scheme == "GfImFreq": exclude = sanitize_imfreq(origin, dest)
|
||||||
|
if scheme == "GFBloc_ImTime" or scheme == "GfImTime": exclude = sanitize_imtime(origin, dest)
|
||||||
|
if scheme == "GFBloc_ReFreq" or scheme == "GfReFreq": exclude = sanitize_refreq(origin, dest)
|
||||||
|
if scheme == "GFBloc_ImLegendre" or scheme == "GfLegendre": exclude = sanitize_legendre(origin, dest)
|
||||||
|
|
||||||
|
# copy the rest
|
||||||
|
for key in [ o for o in origin if o not in exclude ]:
|
||||||
|
|
||||||
|
# key is a dataset
|
||||||
|
if hasattr(origin[key],'value'):
|
||||||
|
|
||||||
|
# check if dataset needs renaming
|
||||||
|
try:
|
||||||
|
new_key = modifications[scheme][key]
|
||||||
|
except:
|
||||||
|
new_key = key
|
||||||
|
|
||||||
|
# create dataset and copy data as well as attributes in destination archive
|
||||||
|
dest.create_dataset(name=new_key, data=origin[key].value)
|
||||||
|
for at in origin[key].attrs.keys():
|
||||||
|
dest[key].attrs.create(at, origin[key].attrs[at])
|
||||||
|
|
||||||
|
# key is a group
|
||||||
|
else:
|
||||||
|
|
||||||
|
dest.create_group(name=key)
|
||||||
|
copy_and_correct(origin[key], dest[key])
|
||||||
|
|
||||||
|
|
||||||
|
# Let's do it
|
||||||
|
A = h5py.File(sys.argv[1],'r')
|
||||||
|
B = h5py.File(sys.argv[2],'w')
|
||||||
|
|
||||||
|
print "Starting update..."
|
||||||
|
copy_and_correct(A,B)
|
||||||
|
print "Update done and written in", sys.argv[2]
|
||||||
|
|
||||||
|
|
@ -1,6 +1,44 @@
|
|||||||
|
|
||||||
|
.. _versions:
|
||||||
|
|
||||||
Version compatibility
|
Version compatibility
|
||||||
=====================
|
=====================
|
||||||
|
|
||||||
As the TRIQS library and applications based on it might live separate lives, it
|
TRIQS version numbering
|
||||||
is important to define a clear version compatibility scheme.
|
-----------------------
|
||||||
|
|
||||||
|
The TRIQS library has three-digit *release* numbers (e.g. 1.0.5) and two-digit
|
||||||
|
*version* numbers (e.g. 1.0). The first two digits of the release number always
|
||||||
|
correspond to the version number.
|
||||||
|
|
||||||
|
New releases within the same version correspond to bug fixes and improvements
|
||||||
|
that do not change the API of the library. In other words, applications based
|
||||||
|
on the TRIQS library do not need to be changed between such releases of the
|
||||||
|
TRIQS library. Instead, a modification of the version number, say from 1.0 to
|
||||||
|
1.1, indicates deeper modifications of the library with possible breaks of the
|
||||||
|
API. In that case, the source codes of the applications might need to be
|
||||||
|
modified.
|
||||||
|
|
||||||
|
TRIQS applications version numbering
|
||||||
|
------------------------------------
|
||||||
|
|
||||||
|
In order to easily identify which versions of an application are compatible
|
||||||
|
with given versions of TRIQS, the applications follow a similar version
|
||||||
|
numbering: three-digit release numbers and two-digit version numbers. The
|
||||||
|
compatibility rule is then simply that *an application is compatible with the
|
||||||
|
TRIQS library if it has the same version number*.
|
||||||
|
|
||||||
|
How to figure the version of the installed TRIQS library
|
||||||
|
--------------------------------------------------------
|
||||||
|
|
||||||
|
In order to figure the version of an installed TRIQS library, you can
|
||||||
|
either:
|
||||||
|
|
||||||
|
* Read the content of the ``version.hpp`` file which is in the
|
||||||
|
``path_to_install_directory/include/triqs/`` diretory.
|
||||||
|
|
||||||
|
* Run ``pytriqs`` and type the following commands::
|
||||||
|
|
||||||
|
from pytriqs.version import *
|
||||||
|
show_version()
|
||||||
|
|
||||||
|
Loading…
Reference in New Issue
Block a user