2014-11-12 14:30:26 +01:00
|
|
|
# -*- coding: utf-8 -*-
|
|
|
|
|
|
|
|
import sqlite3
|
|
|
|
import re
|
2014-11-12 18:12:07 +01:00
|
|
|
import sys
|
|
|
|
import os
|
2014-11-19 11:47:55 +01:00
|
|
|
import time
|
2014-11-12 18:12:07 +01:00
|
|
|
|
2015-01-20 14:20:06 +01:00
|
|
|
debug = True
|
2015-01-09 15:44:03 +01:00
|
|
|
|
|
|
|
|
2015-01-16 09:38:12 +01:00
|
|
|
def checkSQLite3(db_path):
|
|
|
|
|
|
|
|
from os.path import isfile, getsize
|
|
|
|
|
|
|
|
# Check if db file is readable
|
|
|
|
if not os.access(db_path, os.R_OK):
|
|
|
|
print >>sys.stderr, "Db file %s is not readable" % (db_path)
|
|
|
|
raise IOError
|
|
|
|
|
|
|
|
if not isfile(db_path):
|
|
|
|
print >>sys.stderr, "Db file %s is not... a file!" % (db_path)
|
|
|
|
raise IOError
|
|
|
|
|
|
|
|
if getsize(db_path) < 100: # SQLite database file header is 100 bytes
|
|
|
|
print >>sys.stderr, "Db file %s is not a SQLite file!" % (db_path)
|
|
|
|
raise IOError
|
|
|
|
|
|
|
|
with open(db_path, 'rb') as fd:
|
|
|
|
header = fd.read(100)
|
|
|
|
|
|
|
|
if header[:16] != 'SQLite format 3\x00':
|
|
|
|
print >>sys.stderr, "Db file %s is not in SQLiteFormat3!" % (db_path)
|
|
|
|
raise IOError
|
|
|
|
|
|
|
|
# Check if the file system allows I/O on sqlite3 (lustre)
|
|
|
|
# If not, copy on /dev/shm and remove after opening
|
|
|
|
try:
|
|
|
|
EMSL_local(db_path=db_path).get_list_basis_available()
|
|
|
|
except sqlite3.OperationalError:
|
|
|
|
print >>sys.stdrerr, "I/O Error for you file system"
|
|
|
|
print >>sys.stderr, "Try some fixe"
|
|
|
|
new_db_path = "/dev/shm/%d.db" % (os.getpid())
|
|
|
|
os.system("cp %s %s" % (db_path, new_db_path))
|
|
|
|
db_path = new_db_path
|
|
|
|
else:
|
|
|
|
changed = False
|
|
|
|
return db_path, changed
|
|
|
|
|
2015-01-16 16:51:56 +01:00
|
|
|
# Try again to check
|
2015-01-16 09:38:12 +01:00
|
|
|
try:
|
|
|
|
EMSL_local(db_path=db_path).get_list_basis_available()
|
|
|
|
except:
|
|
|
|
print >>sys.stderr, "Sorry..."
|
|
|
|
os.system("rm -f /dev/shm/%d.db" % (os.getpid()))
|
|
|
|
raise
|
|
|
|
else:
|
|
|
|
print >>sys.stderr, "Working !"
|
|
|
|
changed = True
|
|
|
|
return db_path, changed
|
2014-11-12 14:30:26 +01:00
|
|
|
|
|
|
|
|
2015-01-12 14:22:13 +01:00
|
|
|
def install_with_pip(name):
|
|
|
|
|
|
|
|
ins = False
|
|
|
|
d = {'y': True,
|
|
|
|
'n': False}
|
|
|
|
|
|
|
|
while True:
|
|
|
|
choice = raw_input('Do you want to install it ? [Y/N]')
|
|
|
|
try:
|
|
|
|
ins = d[choice.lower()]
|
|
|
|
break
|
|
|
|
except:
|
|
|
|
print "not a valid choice"
|
|
|
|
|
|
|
|
if ins:
|
|
|
|
try:
|
|
|
|
import pip
|
|
|
|
pip.main(['install', name])
|
|
|
|
except:
|
|
|
|
print "You need pip, (http://pip.readthedocs.org/en/latest/installing.html)"
|
|
|
|
sys.exit(1)
|
|
|
|
|
|
|
|
|
2015-01-08 16:12:15 +01:00
|
|
|
def cond_sql_or(table_name, l_value):
|
|
|
|
|
|
|
|
l = []
|
|
|
|
dmy = " OR ".join(['%s = "%s"' % (table_name, i) for i in l_value])
|
|
|
|
if dmy:
|
|
|
|
l.append("(%s)" % dmy)
|
|
|
|
|
|
|
|
return l
|
|
|
|
|
|
|
|
|
2014-11-12 14:30:26 +01:00
|
|
|
class EMSL_dump:
|
|
|
|
|
2015-01-16 09:38:12 +01:00
|
|
|
format_dict = {"g94": "Gaussian94",
|
|
|
|
"gamess-us": "GAMESS-US",
|
|
|
|
"gamess-uk": "GAMESS-UK",
|
|
|
|
"turbomole": "Turbomole",
|
|
|
|
"tx93": "TX93",
|
|
|
|
"molpro": "Molpro",
|
|
|
|
"molproint": "MolproInt",
|
|
|
|
"hondo": "Hondo",
|
|
|
|
"supermolecule": "SuperMolecule",
|
|
|
|
"molcas": "Molcas",
|
|
|
|
"hyperchem": "HyperChem",
|
|
|
|
"dalton": "Dalton",
|
|
|
|
"demon-ks": "deMon-KS",
|
|
|
|
"demon2k": "deMon2k",
|
|
|
|
"aces2": "AcesII"
|
|
|
|
}
|
|
|
|
|
2014-11-12 14:30:26 +01:00
|
|
|
def __init__(self, db_path=None, format="GAMESS-US", contraction="True"):
|
|
|
|
self.db_path = db_path
|
|
|
|
self.format = format
|
2014-11-12 15:35:31 +01:00
|
|
|
self.contraction = str(contraction)
|
2015-01-12 14:22:13 +01:00
|
|
|
try:
|
|
|
|
import requests
|
|
|
|
except:
|
|
|
|
print "You need the requests package"
|
|
|
|
install_with_pip("requests")
|
|
|
|
finally:
|
|
|
|
self.requests = requests
|
2014-11-12 14:30:26 +01:00
|
|
|
|
2015-01-16 09:38:12 +01:00
|
|
|
def get_list_format(self):
|
|
|
|
"""List all the format available in EMSL"""
|
|
|
|
return self.format_dict
|
|
|
|
|
2014-11-12 14:30:26 +01:00
|
|
|
def set_db_path(self, path):
|
|
|
|
"""Define the database path"""
|
|
|
|
self.db_path = path
|
|
|
|
|
2015-01-16 09:38:12 +01:00
|
|
|
def get_dict_ele(self):
|
|
|
|
"""A dict of element"""
|
|
|
|
elt_path = os.path.dirname(sys.argv[0]) + "/src/elts_abrev.dat"
|
|
|
|
|
|
|
|
with open(elt_path, "r") as f:
|
|
|
|
data = f.readlines()
|
|
|
|
|
|
|
|
dict_ele = dict()
|
|
|
|
for i in data:
|
|
|
|
l = i.split("-")
|
|
|
|
dict_ele[l[1].strip().lower()] = l[2].strip().lower()
|
|
|
|
return dict_ele
|
|
|
|
|
2014-11-12 14:30:26 +01:00
|
|
|
def dwl_basis_list_raw(self):
|
2014-11-19 11:47:55 +01:00
|
|
|
print "Download all the name available in EMSL. It can take some time.",
|
2014-11-12 14:30:26 +01:00
|
|
|
sys.stdout.flush()
|
|
|
|
|
|
|
|
"""Download the source code of the iframe who contains the list of the basis set available"""
|
|
|
|
|
|
|
|
url = "https://bse.pnl.gov/bse/portal/user/anon/js_peid/11535052407933/panel/Main/template/content"
|
2014-11-12 16:29:11 +01:00
|
|
|
if debug:
|
2014-11-12 18:12:07 +01:00
|
|
|
import cPickle as pickle
|
|
|
|
dbcache = 'db/cache'
|
|
|
|
if not os.path.isfile(dbcache):
|
|
|
|
page = self.requests.get(url).text
|
|
|
|
file = open(dbcache, 'w')
|
|
|
|
pickle.dump(page, file)
|
|
|
|
else:
|
|
|
|
file = open(dbcache, 'r')
|
|
|
|
page = pickle.load(file)
|
|
|
|
file.close()
|
2014-11-12 16:29:11 +01:00
|
|
|
|
|
|
|
else:
|
2014-11-12 18:12:07 +01:00
|
|
|
page = self.requests.get(url).text
|
2014-11-12 14:30:26 +01:00
|
|
|
|
|
|
|
print "Done"
|
|
|
|
return page
|
|
|
|
|
|
|
|
def bl_raw_to_array(self, data_raw):
|
|
|
|
"""Parse the raw html to create a basis set array whith all the info:
|
|
|
|
url, name,description"""
|
|
|
|
|
2014-11-19 11:47:55 +01:00
|
|
|
d = {}
|
2014-11-12 14:30:26 +01:00
|
|
|
|
|
|
|
for line in data_raw.split('\n'):
|
|
|
|
if "new basisSet(" in line:
|
|
|
|
b = line.find("(")
|
|
|
|
e = line.find(");")
|
|
|
|
|
|
|
|
s = line[b + 1:e]
|
|
|
|
|
|
|
|
tup = eval(s)
|
2015-01-16 16:51:56 +01:00
|
|
|
xml_path = tup[0]
|
2014-11-12 14:30:26 +01:00
|
|
|
name = tup[1]
|
|
|
|
|
2015-01-16 16:51:56 +01:00
|
|
|
elts = re.sub('[["\ \]]', '', tup[3]).split(',')
|
2014-11-12 14:30:26 +01:00
|
|
|
|
2015-01-20 14:20:06 +01:00
|
|
|
des = re.sub('\s+', ' ', tup[-1])
|
2014-11-12 14:30:26 +01:00
|
|
|
|
2015-01-16 16:51:56 +01:00
|
|
|
if "-ecp" in xml_path.lower():
|
2014-11-12 14:30:26 +01:00
|
|
|
continue
|
2015-01-16 16:51:56 +01:00
|
|
|
d[name] = [name, xml_path, des, elts]
|
2014-11-19 11:47:55 +01:00
|
|
|
|
|
|
|
"""Tric for the unicity of the name"""
|
|
|
|
array = [d[key] for key in d]
|
2014-11-12 14:30:26 +01:00
|
|
|
|
2014-11-19 11:47:55 +01:00
|
|
|
array_sort = sorted(array, key=lambda x: x[0])
|
|
|
|
print len(array_sort), "basisset will be download"
|
2014-11-12 14:30:26 +01:00
|
|
|
|
2014-11-19 11:47:55 +01:00
|
|
|
return array_sort
|
2014-11-12 14:30:26 +01:00
|
|
|
|
|
|
|
def basis_data_row_to_array(self, data, name, des, elts):
|
|
|
|
"""Parse the basis data raw html to get a nice tuple"""
|
|
|
|
|
|
|
|
d = []
|
|
|
|
|
|
|
|
b = data.find("$DATA")
|
|
|
|
e = data.find("$END")
|
|
|
|
if (b == -1 or data.find("$DATA$END") != -1):
|
2014-11-19 11:47:55 +01:00
|
|
|
if debug:
|
|
|
|
print data
|
2015-01-08 16:12:15 +01:00
|
|
|
raise Exception("WARNING not DATA")
|
2014-11-12 14:30:26 +01:00
|
|
|
else:
|
2015-01-09 15:44:03 +01:00
|
|
|
data = data.replace("PHOSPHOROUS", "PHOSPHORUS")
|
2015-01-09 16:32:40 +01:00
|
|
|
data = data.replace("D+", "E+")
|
|
|
|
data = data.replace("D-", "E-")
|
|
|
|
|
2015-01-12 14:22:13 +01:00
|
|
|
data = data[b + 5:e - 1].split('\n\n')
|
2015-01-09 15:44:03 +01:00
|
|
|
|
2015-01-16 09:38:12 +01:00
|
|
|
dict_ele = self.get_dict_ele()
|
|
|
|
|
2014-11-12 14:30:26 +01:00
|
|
|
for (elt, data_elt) in zip(elts, data):
|
|
|
|
|
2015-01-09 15:44:03 +01:00
|
|
|
elt_long_th = dict_ele[elt.lower()]
|
|
|
|
elt_long_exp = data_elt.split()[0].lower()
|
|
|
|
|
2015-01-09 16:32:40 +01:00
|
|
|
if "$" in data_elt:
|
2015-01-20 14:20:06 +01:00
|
|
|
if debug:
|
|
|
|
print "Eror",
|
2015-01-26 09:00:53 +01:00
|
|
|
raise Exception("WARNING bad split")
|
2015-01-09 16:32:40 +01:00
|
|
|
|
2015-01-09 15:44:03 +01:00
|
|
|
if elt_long_th == elt_long_exp:
|
2015-01-20 14:20:06 +01:00
|
|
|
d.append([elt, data_elt.strip()])
|
2015-01-09 15:44:03 +01:00
|
|
|
else:
|
2015-01-20 14:20:06 +01:00
|
|
|
if debug:
|
|
|
|
print "th", elt_long_th
|
|
|
|
print "exp", elt_long_exp
|
|
|
|
print "abv", elt
|
2015-01-09 15:44:03 +01:00
|
|
|
raise Exception("WARNING not good ELEMENT")
|
2014-11-12 14:30:26 +01:00
|
|
|
|
2015-01-20 14:20:06 +01:00
|
|
|
return [name, des, d]
|
2014-11-12 14:30:26 +01:00
|
|
|
|
|
|
|
def create_sql(self, list_basis_array):
|
|
|
|
"""Create the sql from the list of basis available data"""
|
|
|
|
|
|
|
|
conn = sqlite3.connect(self.db_path)
|
|
|
|
c = conn.cursor()
|
|
|
|
|
2015-01-20 14:20:06 +01:00
|
|
|
c.execute('''CREATE TABLE basis_tab(
|
|
|
|
basis_id INTEGER PRIMARY KEY AUTOINCREMENT,
|
|
|
|
name text,
|
|
|
|
description text,
|
|
|
|
UNIQUE(name)
|
|
|
|
);''')
|
|
|
|
|
|
|
|
c.execute('''CREATE TABLE data_tab(
|
|
|
|
basis_id INTEGER,
|
|
|
|
elt TEXT,
|
|
|
|
data TEXT,
|
|
|
|
FOREIGN KEY(basis_id)
|
|
|
|
REFERENCES basis_tab(basis_id)
|
|
|
|
);''')
|
|
|
|
|
|
|
|
c.execute(''' CREATE VIEW output_tab AS
|
|
|
|
SELECT basis_id,
|
|
|
|
name,
|
|
|
|
description,
|
|
|
|
elt,
|
|
|
|
data
|
|
|
|
FROM basis_tab
|
|
|
|
NATURAL JOIN data_tab
|
|
|
|
''')
|
|
|
|
|
|
|
|
conn.commit()
|
2014-11-12 14:30:26 +01:00
|
|
|
|
2014-11-12 16:29:11 +01:00
|
|
|
import Queue
|
|
|
|
import threading
|
2014-11-12 14:30:26 +01:00
|
|
|
|
2014-11-12 18:12:07 +01:00
|
|
|
num_worker_threads = 7
|
2015-01-09 15:44:03 +01:00
|
|
|
attemps_max = 20
|
2014-11-19 11:47:55 +01:00
|
|
|
|
2014-11-12 18:12:07 +01:00
|
|
|
q_in = Queue.Queue(num_worker_threads)
|
2014-11-12 16:29:11 +01:00
|
|
|
q_out = Queue.Queue(num_worker_threads)
|
2014-11-12 14:30:26 +01:00
|
|
|
|
2014-11-12 16:29:11 +01:00
|
|
|
def worker():
|
2014-11-19 11:47:55 +01:00
|
|
|
"""get a Job from the q_in, do stuff, when finish put it in the q_out"""
|
2014-11-12 18:12:07 +01:00
|
|
|
while True:
|
2015-01-20 14:20:06 +01:00
|
|
|
name, path_xml, des, elts = q_in.get()
|
2015-01-16 16:51:56 +01:00
|
|
|
|
|
|
|
url = "https://bse.pnl.gov:443/bse/portal/user/anon/js_peid/11535052407933/action/portlets.BasisSetAction/template/courier_content/panel/Main/"
|
|
|
|
url += "/eventSubmit_doDownload/true"
|
|
|
|
|
|
|
|
params = {'bsurl': path_xml, 'bsname': name,
|
|
|
|
'elts': " ".join(elts),
|
|
|
|
'format': self.format,
|
|
|
|
'minimize': self.contraction}
|
2014-11-19 11:47:55 +01:00
|
|
|
|
2015-01-09 15:44:03 +01:00
|
|
|
attemps = 0
|
|
|
|
while attemps < attemps_max:
|
2015-01-16 16:51:56 +01:00
|
|
|
text = self.requests.get(url, params=params).text
|
2014-11-19 11:47:55 +01:00
|
|
|
try:
|
|
|
|
basis_data = self.basis_data_row_to_array(
|
|
|
|
text, name, des, elts)
|
|
|
|
except:
|
|
|
|
time.sleep(0.1)
|
2015-01-09 15:44:03 +01:00
|
|
|
attemps += 1
|
2015-01-26 09:00:53 +01:00
|
|
|
else:
|
|
|
|
break
|
2015-01-09 15:44:03 +01:00
|
|
|
|
|
|
|
try:
|
2015-01-20 14:20:06 +01:00
|
|
|
q_out.put(basis_data)
|
2015-01-09 15:44:03 +01:00
|
|
|
except:
|
2015-01-16 16:51:56 +01:00
|
|
|
if debug:
|
2015-01-20 14:20:06 +01:00
|
|
|
print "Fail on q_out.put", basis_data
|
2015-01-09 15:44:03 +01:00
|
|
|
raise
|
2015-01-26 09:00:53 +01:00
|
|
|
else:
|
|
|
|
q_in.task_done()
|
|
|
|
|
2014-11-12 16:29:11 +01:00
|
|
|
|
|
|
|
def enqueue():
|
2015-01-16 16:51:56 +01:00
|
|
|
for [name, path_xml, des, elts] in list_basis_array:
|
2015-01-20 14:20:06 +01:00
|
|
|
q_in.put([name, path_xml, des, elts])
|
2014-11-19 11:47:55 +01:00
|
|
|
|
2014-11-12 18:12:07 +01:00
|
|
|
return 0
|
2014-11-12 14:30:26 +01:00
|
|
|
|
2014-11-12 16:29:11 +01:00
|
|
|
t = threading.Thread(target=enqueue)
|
|
|
|
t.daemon = True
|
|
|
|
t.start()
|
|
|
|
|
|
|
|
for i in range(num_worker_threads):
|
|
|
|
t = threading.Thread(target=worker)
|
|
|
|
t.daemon = True
|
|
|
|
t.start()
|
|
|
|
|
2014-11-19 11:47:55 +01:00
|
|
|
nb_basis = len(list_basis_array)
|
|
|
|
|
|
|
|
for i in range(nb_basis):
|
2015-01-20 14:20:06 +01:00
|
|
|
name, des, d = q_out.get()
|
|
|
|
q_out.task_done()
|
|
|
|
|
|
|
|
try:
|
|
|
|
c.execute(
|
|
|
|
"INSERT INTO basis_tab(name,description) VALUES (?,?)", [
|
|
|
|
name, des])
|
|
|
|
conn.commit()
|
|
|
|
except sqlite3.IntegrityError:
|
|
|
|
print '{:>3}'.format(i + 1), "/", nb_basis, name, "fail"
|
2014-11-19 11:47:55 +01:00
|
|
|
|
2015-01-20 14:20:06 +01:00
|
|
|
id_ = c.lastrowid
|
2014-11-12 14:30:26 +01:00
|
|
|
try:
|
|
|
|
c.executemany(
|
2015-01-20 14:20:06 +01:00
|
|
|
"INSERT INTO data_tab VALUES (?,?,?)", [
|
|
|
|
[id_] + k for k in d])
|
2014-11-12 14:30:26 +01:00
|
|
|
conn.commit()
|
2014-11-19 11:47:55 +01:00
|
|
|
|
|
|
|
print '{:>3}'.format(i + 1), "/", nb_basis, name
|
2015-01-20 14:20:06 +01:00
|
|
|
|
2014-11-12 14:30:26 +01:00
|
|
|
except:
|
2015-01-16 16:51:56 +01:00
|
|
|
print '{:>3}'.format(i + 1), "/", nb_basis, name, "fail"
|
2014-11-19 11:47:55 +01:00
|
|
|
raise
|
2015-01-20 14:20:06 +01:00
|
|
|
|
2014-11-12 14:30:26 +01:00
|
|
|
conn.close()
|
2014-11-19 11:47:55 +01:00
|
|
|
|
2014-11-12 16:29:11 +01:00
|
|
|
q_in.join()
|
2014-11-12 14:30:26 +01:00
|
|
|
|
|
|
|
def new_db(self):
|
|
|
|
"""Create new_db from scratch"""
|
|
|
|
|
|
|
|
_data = self.dwl_basis_list_raw()
|
|
|
|
array_basis = self.bl_raw_to_array(_data)
|
|
|
|
del _data
|
|
|
|
|
|
|
|
self.create_sql(array_basis)
|
|
|
|
|
|
|
|
|
2015-03-10 19:30:11 +01:00
|
|
|
def string_to_nb_mo(str_l):
|
|
|
|
|
|
|
|
assert len(str_l) == 1
|
|
|
|
|
|
|
|
d = {"S": 3,
|
|
|
|
"P": 5,
|
|
|
|
"D": 7,
|
|
|
|
"F": 9,
|
|
|
|
"L": 8}
|
|
|
|
|
|
|
|
if str_l in d:
|
|
|
|
return d[str_l]
|
|
|
|
# ord("G") = 72 and ord("Z") = 87
|
|
|
|
elif 72 <= ord(str_l) <= 87:
|
|
|
|
# orf("G") = 72 and l = 4 so ofset if 68
|
|
|
|
return 2 * (ord(str_l) - 68) + 1
|
|
|
|
else:
|
|
|
|
raise BaseException
|
2014-11-12 14:30:26 +01:00
|
|
|
|
2015-03-10 19:30:11 +01:00
|
|
|
|
|
|
|
class EMSL_local:
|
|
|
|
"""
|
|
|
|
All the method for using the EMSL db localy
|
|
|
|
"""
|
2014-11-12 14:30:26 +01:00
|
|
|
def __init__(self, db_path=None):
|
|
|
|
self.db_path = db_path
|
|
|
|
|
2015-03-10 19:30:11 +01:00
|
|
|
def get_list_basis_available(self, elts=[], average_mo_number=False):
|
|
|
|
"""
|
|
|
|
return all the basis name who contant all the elts
|
|
|
|
"""
|
2014-11-12 14:30:26 +01:00
|
|
|
conn = sqlite3.connect(self.db_path)
|
|
|
|
c = conn.cursor()
|
2015-03-10 19:30:11 +01:00
|
|
|
# If not elts just get the disctinct name
|
|
|
|
# Else: 1) fetch for geting the run_id available
|
|
|
|
# 2) If average_mo_number:
|
|
|
|
# * Get name,descripption,data
|
|
|
|
# * Then parse it
|
|
|
|
# Else Get name,description
|
|
|
|
# 3) Parse it
|
|
|
|
|
|
|
|
# ~#~#~#~#~#~#~#~#~#~#~#~#~#~#~ #
|
|
|
|
# G e t i n g B a s i s _ i d #
|
|
|
|
# ~#~#~#~#~#~#~#~#~#~#~#~#~#~#~ #
|
2014-11-20 15:33:26 +01:00
|
|
|
if not elts:
|
2015-03-10 19:30:11 +01:00
|
|
|
cmd = """SELECT DISTINCT name, description FROM basis_tab"""
|
|
|
|
c.execute(cmd)
|
|
|
|
info = c.fetchall()
|
|
|
|
else:
|
|
|
|
cmd_ele = ["SELECT DISTINCT basis_id FROM data_tab WHERE elt=?"] * len(elts)
|
|
|
|
cmd = " INTERSECT ".join(cmd_ele) + ";"
|
|
|
|
c.execute(cmd, elts)
|
2015-02-27 14:32:09 +01:00
|
|
|
|
2015-03-10 19:30:11 +01:00
|
|
|
dump = [i[0] for i in c.fetchall()]
|
|
|
|
cmd_basis = " ".join(cond_sql_or("basis_id", dump))
|
|
|
|
cmd_ele = " ".join(cond_sql_or("elt", elts))
|
2015-02-27 14:32:09 +01:00
|
|
|
|
2015-03-10 19:30:11 +01:00
|
|
|
if average_mo_number:
|
|
|
|
cmd = """SELECT DISTINCT name,description,data
|
|
|
|
FROM output_tab"""
|
2014-11-20 15:33:26 +01:00
|
|
|
|
2015-03-10 19:30:11 +01:00
|
|
|
else:
|
|
|
|
cmd = """SELECT DISTINCT name,description
|
|
|
|
FROM output_tab"""
|
2014-11-12 14:30:26 +01:00
|
|
|
|
2015-03-10 19:30:11 +01:00
|
|
|
cmd += " WHERE" + cmd_ele + " AND " + cmd_basis
|
2014-11-12 14:30:26 +01:00
|
|
|
|
2015-03-10 19:30:11 +01:00
|
|
|
c.execute(cmd)
|
|
|
|
info = c.fetchall()
|
2015-01-09 13:17:37 +01:00
|
|
|
|
2015-03-10 19:30:11 +01:00
|
|
|
conn.close()
|
|
|
|
|
|
|
|
# ~#~#~#~#~#~#~ #
|
|
|
|
# P a r s i n g #
|
|
|
|
# ~#~#~#~#~#~#~ #
|
|
|
|
|
|
|
|
dict_info = {}
|
|
|
|
# dict_info[name] = [description, nb_mo, nb_ele]
|
|
|
|
|
|
|
|
if average_mo_number:
|
|
|
|
|
|
|
|
for name, description, data in info:
|
|
|
|
nb_mo = 0
|
|
|
|
nb_ele = 0
|
|
|
|
for line in data.split("\n")[1:]:
|
|
|
|
str_l = line.split()[0]
|
|
|
|
try:
|
|
|
|
nb_mo += string_to_nb_mo(str_l)
|
|
|
|
nb_ele += 1
|
|
|
|
except BaseException:
|
|
|
|
pass
|
|
|
|
|
|
|
|
try:
|
|
|
|
dict_info[name][1] += nb_mo
|
|
|
|
dict_info[name][2] += 1
|
|
|
|
except:
|
|
|
|
dict_info[name] = [description, nb_mo, nb_ele]
|
|
|
|
|
|
|
|
# ~#~#~#~#~#~ #
|
|
|
|
# R e t u r n #
|
|
|
|
# ~#~#~#~#~#~ #
|
|
|
|
|
|
|
|
if average_mo_number:
|
|
|
|
return[[k, v[0], v[1] / v[2]] for k, v in dict_info.iteritems()]
|
|
|
|
else:
|
|
|
|
return [i[:] for i in info]
|
2014-11-12 14:30:26 +01:00
|
|
|
|
|
|
|
def get_list_element_available(self, basis_name):
|
|
|
|
|
|
|
|
conn = sqlite3.connect(self.db_path)
|
|
|
|
c = conn.cursor()
|
|
|
|
|
2015-03-10 19:30:11 +01:00
|
|
|
str_ = "SELECT DISTINCT elt from output_tab WHERE name=:name_us COLLATE NOCASE"
|
|
|
|
c.execute(str_, {"name_us": basis_name})
|
2014-11-12 14:30:26 +01:00
|
|
|
|
|
|
|
data = c.fetchall()
|
|
|
|
|
|
|
|
data = [str(i[0]) for i in data]
|
|
|
|
|
|
|
|
conn.close()
|
|
|
|
return data
|
|
|
|
|
2015-01-09 13:17:37 +01:00
|
|
|
def get_basis(self, basis_name, elts=None, with_l=False):
|
2015-03-10 19:30:11 +01:00
|
|
|
"""
|
|
|
|
Return the data from the basis set
|
|
|
|
"""
|
2015-01-20 14:20:06 +01:00
|
|
|
import re
|
|
|
|
|
2015-01-09 13:17:37 +01:00
|
|
|
def get_list_type(l_line):
|
2015-03-10 19:30:11 +01:00
|
|
|
"""
|
|
|
|
Return the begin and the end of all the type of orbital
|
|
|
|
Usefull for tranforming the L in S, P
|
|
|
|
output : [ [type, begin, end], ...]
|
|
|
|
"""
|
|
|
|
# Example
|
|
|
|
# [[u'S', 1, 5], [u'L', 5, 9], [u'L', 9, 12], [u'D', 16, 18]]"
|
|
|
|
|
2015-01-09 13:17:37 +01:00
|
|
|
l = []
|
|
|
|
for i, line in enumerate(l_line):
|
|
|
|
|
|
|
|
m = re.search(p, line)
|
|
|
|
if m:
|
|
|
|
l.append([m.group(1), i])
|
|
|
|
try:
|
|
|
|
l[-2].append(i)
|
|
|
|
except IndexError:
|
|
|
|
pass
|
|
|
|
|
|
|
|
l[-1].append(i + 1)
|
2015-03-10 19:30:11 +01:00
|
|
|
print l
|
2015-01-09 13:17:37 +01:00
|
|
|
return l
|
|
|
|
|
|
|
|
# __ _
|
|
|
|
# /__ _ _|_ _|_ ._ _ ._ _ _ _. |
|
|
|
|
# \_| (/_ |_ | | (_) | | | _> (_| |
|
|
|
|
# |
|
2014-11-12 14:30:26 +01:00
|
|
|
conn = sqlite3.connect(self.db_path)
|
|
|
|
c = conn.cursor()
|
|
|
|
|
2015-01-08 16:12:15 +01:00
|
|
|
if elts:
|
|
|
|
cmd_ele = "AND " + " ".join(cond_sql_or("elt", elts))
|
|
|
|
else:
|
|
|
|
cmd_ele = ""
|
|
|
|
|
2015-01-20 14:20:06 +01:00
|
|
|
c.execute('''SELECT DISTINCT data from output_tab
|
2015-01-08 16:12:15 +01:00
|
|
|
WHERE name="{basis_name}" COLLATE NOCASE
|
|
|
|
{cmd_ele}'''.format(basis_name=basis_name,
|
|
|
|
cmd_ele=cmd_ele))
|
2014-11-12 14:30:26 +01:00
|
|
|
|
2015-01-09 13:17:37 +01:00
|
|
|
l_data_raw = c.fetchall()
|
2014-11-12 14:30:26 +01:00
|
|
|
conn.close()
|
2015-01-09 13:17:37 +01:00
|
|
|
|
|
|
|
# |_| _. ._ _| | _ || | ||
|
|
|
|
# | | (_| | | (_| | (/_ |_
|
|
|
|
#
|
|
|
|
|
|
|
|
p = re.compile(ur'^(\w)\s+\d+\b')
|
|
|
|
|
|
|
|
l_data = []
|
|
|
|
|
|
|
|
for data_raw in l_data_raw:
|
|
|
|
|
|
|
|
basis = data_raw[0].strip()
|
|
|
|
|
|
|
|
l_line_raw = basis.split("\n")
|
|
|
|
|
2015-03-10 19:30:11 +01:00
|
|
|
# l_line_raw[0] containt the name of the Atom
|
2015-01-09 13:17:37 +01:00
|
|
|
l_line = [l_line_raw[0]]
|
|
|
|
|
2015-01-09 14:26:54 +01:00
|
|
|
for symmetry, begin, end in get_list_type(l_line_raw):
|
2015-01-09 13:17:37 +01:00
|
|
|
|
2015-01-09 14:26:54 +01:00
|
|
|
if not(with_l) and symmetry in "L":
|
2015-01-09 14:01:27 +01:00
|
|
|
|
2015-01-09 14:26:54 +01:00
|
|
|
body_s = []
|
|
|
|
body_p = []
|
2015-01-09 14:01:27 +01:00
|
|
|
|
|
|
|
for i_l in l_line_raw[begin + 1:end]:
|
2015-01-09 14:26:54 +01:00
|
|
|
|
2015-03-10 09:03:22 +01:00
|
|
|
# one L => S & P
|
2015-01-09 14:01:27 +01:00
|
|
|
a = i_l.split()
|
|
|
|
|
|
|
|
common = "{:>3}".format(a[0])
|
|
|
|
common += "{:>15.7f}".format(float(a[1]))
|
|
|
|
|
|
|
|
tail_s = common + "{:>23.7f}".format(float(a[2]))
|
|
|
|
body_s.append(tail_s)
|
|
|
|
|
2015-03-10 09:03:22 +01:00
|
|
|
# Is only a whan only 3 elements, coef for p == coef for s
|
|
|
|
try:
|
|
|
|
tail_p = common + "{:>23.7f}".format(float(a[3]))
|
|
|
|
except IndexError:
|
|
|
|
tail_p = tail_s
|
|
|
|
finally:
|
|
|
|
body_p.append(tail_p)
|
2015-01-09 13:17:37 +01:00
|
|
|
|
|
|
|
l_line += [l_line_raw[begin].replace("L", "S")]
|
2015-01-09 14:01:27 +01:00
|
|
|
l_line += body_s
|
2015-01-09 13:17:37 +01:00
|
|
|
|
|
|
|
l_line += [l_line_raw[begin].replace("L", "P")]
|
2015-01-09 14:01:27 +01:00
|
|
|
l_line += body_p
|
2015-01-09 13:17:37 +01:00
|
|
|
else:
|
|
|
|
l_line += l_line_raw[begin:end]
|
|
|
|
|
|
|
|
l_data.append("\n".join(l_line))
|
|
|
|
|
|
|
|
return l_data
|
|
|
|
|
2014-11-12 14:30:26 +01:00
|
|
|
if __name__ == "__main__":
|
|
|
|
|
|
|
|
e = EMSL_local(db_path="EMSL.db")
|
|
|
|
l = e.get_list_basis_available()
|
|
|
|
for i in l:
|
|
|
|
print i
|
|
|
|
|
|
|
|
l = e.get_list_element_available("pc-0")
|
|
|
|
print l
|
|
|
|
|
|
|
|
l = e.get_basis("cc-pVTZ", ["H", "He"])
|
|
|
|
for i in l:
|
|
|
|
print i
|