Store metadata in each scan
This commit is contained in:
parent
20527e8d2b
commit
b31c0b413c
@ -87,8 +87,8 @@ def create():
|
||||
proposal_textinput.on_change("value", proposal_textinput_callback)
|
||||
|
||||
def _init_datatable():
|
||||
scan_list = [s["idx"] for s in det_data["scan"]]
|
||||
hkl = [f'{s["h"]} {s["k"]} {s["l"]}' for s in det_data["scan"]]
|
||||
scan_list = [s["idx"] for s in det_data]
|
||||
hkl = [f'{s["h"]} {s["k"]} {s["l"]}' for s in det_data]
|
||||
scan_table_source.data.update(
|
||||
scan=scan_list,
|
||||
hkl=hkl,
|
||||
@ -159,8 +159,8 @@ def create():
|
||||
append_upload_button.on_change("value", append_upload_button_callback)
|
||||
|
||||
def _update_table():
|
||||
num_of_peaks = [len(scan.get("peak_indexes", [])) for scan in det_data["scan"]]
|
||||
fit_ok = [(1 if "fit" in scan else 0) for scan in det_data["scan"]]
|
||||
num_of_peaks = [len(scan.get("peak_indexes", [])) for scan in det_data]
|
||||
fit_ok = [(1 if "fit" in scan else 0) for scan in det_data]
|
||||
scan_table_source.data.update(peaks=num_of_peaks, fit=fit_ok)
|
||||
|
||||
def _update_plot(scan):
|
||||
@ -284,7 +284,7 @@ def create():
|
||||
# skip unnecessary update caused by selection drop
|
||||
return
|
||||
|
||||
_update_plot(det_data["scan"][scan_table_source.data["scan"][new[0]]])
|
||||
_update_plot(det_data[scan_table_source.data["scan"][new[0]]])
|
||||
|
||||
scan_table_source = ColumnDataSource(dict(scan=[], hkl=[], peaks=[], fit=[], export=[]))
|
||||
scan_table = DataTable(
|
||||
@ -305,7 +305,7 @@ def create():
|
||||
def _get_selected_scan():
|
||||
selected_index = scan_table_source.selected.indices[0]
|
||||
selected_scan_id = scan_table_source.data["scan"][selected_index]
|
||||
return det_data["scan"][selected_scan_id]
|
||||
return det_data[selected_scan_id]
|
||||
|
||||
def peak_pos_textinput_callback(_attr, _old, new):
|
||||
if new is not None and not peak_pos_textinput_lock:
|
||||
@ -446,7 +446,7 @@ def create():
|
||||
|
||||
def peakfind_all_button_callback():
|
||||
peakfind_params = _get_peakfind_params()
|
||||
for scan in det_data["scan"]:
|
||||
for scan in det_data:
|
||||
pyzebra.ccl_findpeaks(scan, **peakfind_params)
|
||||
|
||||
_update_table()
|
||||
@ -478,7 +478,7 @@ def create():
|
||||
|
||||
def fit_all_button_callback():
|
||||
fit_params = _get_fit_params()
|
||||
for scan in det_data["scan"]:
|
||||
for scan in det_data:
|
||||
# fit_params are updated inplace within `fitccl`
|
||||
pyzebra.fitccl(scan, **deepcopy(fit_params))
|
||||
|
||||
@ -514,8 +514,8 @@ def create():
|
||||
export_data = deepcopy(det_data)
|
||||
for s, export in zip(scan_table_source.data["scan"], scan_table_source.data["export"]):
|
||||
if not export:
|
||||
if "fit" in export_data["scan"][s]:
|
||||
del export_data["scan"][s]["fit"]
|
||||
if "fit" in export_data[s]:
|
||||
del export_data[s]["fit"]
|
||||
|
||||
pyzebra.export_1D(
|
||||
export_data,
|
||||
@ -547,7 +547,7 @@ def create():
|
||||
export_data = deepcopy(det_data)
|
||||
for s, export in zip(scan_table_source.data["scan"], scan_table_source.data["export"]):
|
||||
if not export:
|
||||
del export_data["scan"][s]
|
||||
del export_data[s]
|
||||
|
||||
pyzebra.export_1D(
|
||||
export_data,
|
||||
|
@ -96,16 +96,10 @@ def create():
|
||||
proposal_textinput.on_change("value", proposal_textinput_callback)
|
||||
|
||||
def _init_datatable():
|
||||
scan_list = [s["idx"] for s in det_data["scan"]]
|
||||
scan_list = [s["idx"] for s in det_data]
|
||||
file_list = []
|
||||
extra_meta = det_data.get("extra_meta", {})
|
||||
for scan_id in scan_list:
|
||||
if scan_id in extra_meta:
|
||||
f_path = extra_meta[scan_id]["original_filename"]
|
||||
else:
|
||||
f_path = det_data["meta"]["original_filename"]
|
||||
|
||||
_, f_name = os.path.split(f_path)
|
||||
_, f_name = os.path.split(det_data[scan_id]["meta"]["original_filename"])
|
||||
file_list.append(f_name)
|
||||
|
||||
scan_table_source.data.update(
|
||||
@ -184,8 +178,8 @@ def create():
|
||||
append_upload_button.on_change("value", append_upload_button_callback)
|
||||
|
||||
def _update_table():
|
||||
num_of_peaks = [len(scan.get("peak_indexes", [])) for scan in det_data["scan"]]
|
||||
fit_ok = [(1 if "fit" in scan else 0) for scan in det_data["scan"]]
|
||||
num_of_peaks = [len(scan.get("peak_indexes", [])) for scan in det_data]
|
||||
fit_ok = [(1 if "fit" in scan else 0) for scan in det_data]
|
||||
scan_table_source.data.update(peaks=num_of_peaks, fit=fit_ok)
|
||||
|
||||
def _update_plot():
|
||||
@ -271,12 +265,12 @@ def create():
|
||||
for ind, p in enumerate(scan_table_source.data["param"]):
|
||||
if p:
|
||||
s = scan_table_source.data["scan"][ind]
|
||||
xs.append(np.array(det_data["scan"][s]["om"]))
|
||||
x.extend(det_data["scan"][s]["om"])
|
||||
ys.append(np.array(det_data["scan"][s]["Counts"]))
|
||||
y.extend([float(p)] * len(det_data["scan"][s]["om"]))
|
||||
xs.append(np.array(det_data[s]["om"]))
|
||||
x.extend(det_data[s]["om"])
|
||||
ys.append(np.array(det_data[s]["Counts"]))
|
||||
y.extend([float(p)] * len(det_data[s]["om"]))
|
||||
param.append(float(p))
|
||||
par.extend(det_data["scan"][s]["Counts"])
|
||||
par.extend(det_data[s]["Counts"])
|
||||
|
||||
ov_plot_mline_source.data.update(xs=xs, ys=ys, param=param, color=color_palette(len(xs)))
|
||||
ov_param_plot_scatter_source.data.update(x=x, y=y, param=par)
|
||||
@ -412,7 +406,7 @@ def create():
|
||||
def _get_selected_scan():
|
||||
selected_index = scan_table_source.selected.indices[0]
|
||||
selected_scan_id = scan_table_source.data["scan"][selected_index]
|
||||
return det_data["scan"][selected_scan_id]
|
||||
return det_data[selected_scan_id]
|
||||
|
||||
def peak_pos_textinput_callback(_attr, _old, new):
|
||||
if new is not None and not peak_pos_textinput_lock:
|
||||
@ -553,7 +547,7 @@ def create():
|
||||
|
||||
def peakfind_all_button_callback():
|
||||
peakfind_params = _get_peakfind_params()
|
||||
for scan in det_data["scan"]:
|
||||
for scan in det_data:
|
||||
pyzebra.ccl_findpeaks(scan, **peakfind_params)
|
||||
|
||||
_update_table()
|
||||
@ -585,7 +579,7 @@ def create():
|
||||
|
||||
def fit_all_button_callback():
|
||||
fit_params = _get_fit_params()
|
||||
for scan in det_data["scan"]:
|
||||
for scan in det_data:
|
||||
# fit_params are updated inplace within `fitccl`
|
||||
pyzebra.fitccl(scan, **deepcopy(fit_params))
|
||||
|
||||
@ -621,7 +615,7 @@ def create():
|
||||
export_data = deepcopy(det_data)
|
||||
for s, export in zip(scan_table_source.data["scan"], scan_table_source.data["export"]):
|
||||
if not export:
|
||||
del export_data["scan"][s]
|
||||
del export_data[s]
|
||||
|
||||
pyzebra.export_1D(
|
||||
export_data,
|
||||
@ -648,8 +642,8 @@ def create():
|
||||
export_data = deepcopy(det_data)
|
||||
for s, export in zip(scan_table_source.data["scan"], scan_table_source.data["export"]):
|
||||
if not export:
|
||||
if "fit" in export_data["scan"][s]:
|
||||
del export_data["scan"][s]["fit"]
|
||||
if "fit" in export_data[s]:
|
||||
del export_data[s]["fit"]
|
||||
|
||||
pyzebra.export_1D(
|
||||
export_data,
|
||||
|
@ -98,8 +98,9 @@ def load_1D(filepath):
|
||||
|
||||
|
||||
def parse_1D(fileobj, data_type):
|
||||
metadata = {"data_type": data_type}
|
||||
|
||||
# read metadata
|
||||
metadata = {}
|
||||
for line in fileobj:
|
||||
if "=" in line:
|
||||
variable, value = line.split("=")
|
||||
@ -154,6 +155,9 @@ def parse_1D(fileobj, data_type):
|
||||
counts.extend(map(int, next(fileobj).split()))
|
||||
s["Counts"] = counts
|
||||
|
||||
# add metadata to each scan
|
||||
s["meta"] = metadata
|
||||
|
||||
scan.append(s)
|
||||
|
||||
elif data_type == ".dat":
|
||||
@ -179,21 +183,17 @@ def parse_1D(fileobj, data_type):
|
||||
|
||||
s["om"] = np.array(s["om"])
|
||||
|
||||
s["temp"] = metadata["temp"]
|
||||
try:
|
||||
s["mf"] = metadata["mf"]
|
||||
except KeyError:
|
||||
if "mf" not in metadata:
|
||||
print("Magnetic field is not present in dat file")
|
||||
|
||||
s["omega"] = metadata["omega"]
|
||||
s["n_points"] = len(s["om"])
|
||||
s["monitor"] = s["Monitor1"][0]
|
||||
s["twotheta"] = metadata["twotheta"]
|
||||
s["chi"] = metadata["chi"]
|
||||
s["phi"] = metadata["phi"]
|
||||
s["nu"] = metadata["nu"]
|
||||
|
||||
s["idx"] = 1
|
||||
|
||||
# add metadata to the scan
|
||||
s["meta"] = metadata
|
||||
|
||||
scan.append(dict(s))
|
||||
|
||||
else:
|
||||
@ -206,9 +206,7 @@ def parse_1D(fileobj, data_type):
|
||||
else:
|
||||
s["indices"] = "real"
|
||||
|
||||
metadata["data_type"] = data_type
|
||||
|
||||
return {"meta": metadata, "scan": scan}
|
||||
return scan
|
||||
|
||||
|
||||
def export_1D(data, path, area_method=AREA_METHODS[0], lorentz=False, hkl_precision=2):
|
||||
@ -217,10 +215,10 @@ def export_1D(data, path, area_method=AREA_METHODS[0], lorentz=False, hkl_precis
|
||||
Scans with integer/real hkl values are saved in .comm/.incomm files correspondingly. If no scans
|
||||
are present for a particular output format, that file won't be created.
|
||||
"""
|
||||
zebra_mode = data["meta"]["zebra_mode"]
|
||||
zebra_mode = data[0]["meta"]["zebra_mode"]
|
||||
file_content = {".comm": [], ".incomm": []}
|
||||
|
||||
for scan in data["scan"]:
|
||||
for scan in data:
|
||||
if "fit" not in scan:
|
||||
continue
|
||||
|
||||
|
@ -13,14 +13,14 @@ def create_tuples(x, y, y_err):
|
||||
|
||||
|
||||
def normalize_all(dictionary, monitor=100000):
|
||||
for scan in dictionary["scan"]:
|
||||
for scan in dictionary:
|
||||
counts = np.array(scan["Counts"])
|
||||
sigma = np.sqrt(counts) if "sigma" not in scan else scan["sigma"]
|
||||
monitor_ratio = monitor / scan["monitor"]
|
||||
scan["Counts"] = counts * monitor_ratio
|
||||
scan["sigma"] = np.array(sigma) * monitor_ratio
|
||||
scan["monitor"] = monitor
|
||||
print("Normalized %d scans to monitor %d" % (len(dictionary["scan"]), monitor))
|
||||
print("Normalized %d scans to monitor %d" % (len(dictionary), monitor))
|
||||
|
||||
|
||||
def merge(scan1, scan2):
|
||||
@ -77,11 +77,11 @@ def merge(scan1, scan2):
|
||||
|
||||
|
||||
def check_UB(dict1, dict2, precision=0.01):
|
||||
return np.max(np.abs(dict1["meta"]["ub"] - dict2["meta"]["ub"])) < precision
|
||||
return np.max(np.abs(dict1[0]["meta"]["ub"] - dict2[0]["meta"]["ub"])) < precision
|
||||
|
||||
|
||||
def check_zebramode(dict1, dict2):
|
||||
if dict1["meta"]["zebra_mode"] == dict2["meta"]["zebra_mode"]:
|
||||
if dict1[0]["meta"]["zebra_mode"] == dict2[0]["meta"]["zebra_mode"]:
|
||||
return True
|
||||
else:
|
||||
return False
|
||||
@ -128,12 +128,12 @@ def check_temp_mag(scan1, scan2):
|
||||
|
||||
def merge_dups(dictionary):
|
||||
|
||||
if dictionary["meta"]["data_type"] == "dat":
|
||||
if dictionary[0]["meta"]["data_type"] == "dat":
|
||||
return
|
||||
|
||||
if dictionary["meta"]["zebra_mode"] == "bi":
|
||||
if dictionary[0]["meta"]["zebra_mode"] == "bi":
|
||||
angles = ["twotheta", "omega", "chi", "phi"]
|
||||
elif dictionary["meta"]["zebra_mode"] == "nb":
|
||||
elif dictionary[0]["meta"]["zebra_mode"] == "nb":
|
||||
angles = ["gamma", "omega", "nu"]
|
||||
|
||||
precision = {
|
||||
@ -145,19 +145,19 @@ def merge_dups(dictionary):
|
||||
"gamma": 0.05,
|
||||
}
|
||||
|
||||
for i in range(len(dictionary["scan"])):
|
||||
for j in range(len(dictionary["scan"])):
|
||||
for i in range(len(dictionary)):
|
||||
for j in range(len(dictionary)):
|
||||
if i == j:
|
||||
continue
|
||||
else:
|
||||
# print(i, j)
|
||||
if check_angles(
|
||||
dictionary["scan"][i], dictionary["scan"][j], angles, precision
|
||||
) and check_temp_mag(dictionary["scan"][i], dictionary["scan"][j]):
|
||||
merge(dictionary["scan"][i], dictionary["scan"][j])
|
||||
if check_angles(dictionary[i], dictionary[j], angles, precision) and check_temp_mag(
|
||||
dictionary[i], dictionary[j]
|
||||
):
|
||||
merge(dictionary[i], dictionary[j])
|
||||
print("merged %d with %d within the dictionary" % (i, j))
|
||||
|
||||
del dictionary["scan"][j]
|
||||
del dictionary[j]
|
||||
merge_dups(dictionary)
|
||||
break
|
||||
else:
|
||||
@ -166,29 +166,24 @@ def merge_dups(dictionary):
|
||||
|
||||
|
||||
def add_scan(dict1, dict2, scan_to_add):
|
||||
max_scan = len(dict1["scan"])
|
||||
dict1["scan"].append(dict2["scan"][scan_to_add])
|
||||
if dict1.get("extra_meta") is None:
|
||||
dict1["extra_meta"] = {}
|
||||
dict1["extra_meta"][max_scan + 1] = dict2["meta"]
|
||||
del dict2["scan"][scan_to_add]
|
||||
dict1.append(dict2[scan_to_add])
|
||||
del dict2[scan_to_add]
|
||||
|
||||
|
||||
def process(dict1, dict2, angles, precision):
|
||||
# stop when the second dict is empty
|
||||
# print(dict2["scan"])
|
||||
if dict2["scan"]:
|
||||
if dict2:
|
||||
# check UB matrixes
|
||||
if check_UB(dict1, dict2):
|
||||
# iterate over second dict and check for matches
|
||||
for i in range(len(dict2["scan"])):
|
||||
for j in range(len(dict1["scan"])):
|
||||
if check_angles(dict1["scan"][j], dict2["scan"][i], angles, precision):
|
||||
for i in range(len(dict2)):
|
||||
for j in range(len(dict1)):
|
||||
if check_angles(dict1[j], dict2[i], angles, precision):
|
||||
# angles good, see the mag and temp
|
||||
if check_temp_mag(dict1["scan"][j], dict2["scan"][i]):
|
||||
merge(dict1["scan"][j], dict2["scan"][i])
|
||||
if check_temp_mag(dict1[j], dict2[i]):
|
||||
merge(dict1[j], dict2[i])
|
||||
print("merged %d with %d from different dictionaries" % (i, j))
|
||||
del dict2["scan"][i]
|
||||
del dict2[i]
|
||||
process(dict1, dict2, angles, precision)
|
||||
break
|
||||
else:
|
||||
@ -225,9 +220,9 @@ def unified_merge(dict1, dict2):
|
||||
return
|
||||
|
||||
# decide angles
|
||||
if dict1["meta"]["zebra_mode"] == "bi":
|
||||
if dict1[0]["meta"]["zebra_mode"] == "bi":
|
||||
angles = ["twotheta", "omega", "chi", "phi"]
|
||||
elif dict1["meta"]["zebra_mode"] == "nb":
|
||||
elif dict1[0]["meta"]["zebra_mode"] == "nb":
|
||||
angles = ["gamma", "omega", "nu"]
|
||||
|
||||
# precision of angles to check
|
||||
@ -239,7 +234,7 @@ def unified_merge(dict1, dict2):
|
||||
"omega": 5,
|
||||
"gamma": 0.1,
|
||||
}
|
||||
if (dict1["meta"]["data_type"] == "ccl") and (dict2["meta"]["data_type"] == "ccl"):
|
||||
if (dict1[0]["meta"]["data_type"] == "ccl") and (dict2[0]["meta"]["data_type"] == "ccl"):
|
||||
precision["omega"] = 0.05
|
||||
|
||||
process(dict1, dict2, angles, precision)
|
||||
@ -254,33 +249,20 @@ def add_dict(dict1, dict2):
|
||||
Note: dict1 must be made from ccl, otherwise we would have to change the structure of loaded
|
||||
dat file"""
|
||||
try:
|
||||
if dict1["meta"]["zebra_mode"] != dict2["meta"]["zebra_mode"]:
|
||||
if dict1[0]["meta"]["zebra_mode"] != dict2[0]["meta"]["zebra_mode"]:
|
||||
print("You are trying to add scans measured with different zebra modes")
|
||||
return
|
||||
# this is for the qscan case
|
||||
except KeyError:
|
||||
print("Zebra mode not specified")
|
||||
max_measurement_dict1 = len(dict1["scan"])
|
||||
new_filenames = np.arange(
|
||||
max_measurement_dict1 + 1, max_measurement_dict1 + 1 + len(dict2["scan"])
|
||||
)
|
||||
|
||||
if dict1.get("extra_meta") is None:
|
||||
dict1["extra_meta"] = {}
|
||||
for s in dict2:
|
||||
if s not in dict1:
|
||||
dict1.append(s)
|
||||
|
||||
new_meta_name = "meta" + str(dict2["meta"]["original_filename"])
|
||||
if new_meta_name not in dict1:
|
||||
for keys, name in zip(range(len(dict2["scan"])), new_filenames):
|
||||
dict2["scan"][keys]["file_of_origin"] = str(dict2["meta"]["original_filename"])
|
||||
dict1["scan"].append(dict2["scan"][keys])
|
||||
dict1["extra_meta"][name] = dict2["meta"]
|
||||
|
||||
dict1[new_meta_name] = dict2["meta"]
|
||||
else:
|
||||
raise KeyError(
|
||||
str(
|
||||
else:
|
||||
print(
|
||||
"The file %s has alredy been added to %s"
|
||||
% (dict2["meta"]["original_filename"], dict1["meta"]["original_filename"])
|
||||
% (dict2[0]["meta"]["original_filename"], dict1[0]["meta"]["original_filename"])
|
||||
)
|
||||
)
|
||||
return dict1
|
||||
|
@ -10,6 +10,7 @@ from mpl_toolkits.mplot3d import Axes3D # dont delete, otherwise waterfall wont
|
||||
import collections
|
||||
|
||||
from .ccl_io import load_1D
|
||||
from .merge_function import add_dict
|
||||
|
||||
|
||||
def create_tuples(x, y, y_err):
|
||||
@ -52,10 +53,10 @@ def load_dats(filepath):
|
||||
else:
|
||||
|
||||
dict1 = add_dict(dict1, load_1D(file_list[i]))
|
||||
dict1["scan"].append({})
|
||||
dict1.append({})
|
||||
if data_type == "txt":
|
||||
for x in range(len(col_names) - 1):
|
||||
dict1["scan"][i + 1]["params"][col_names[x + 1]] = float(file_list[i][x + 1])
|
||||
dict1[i + 1]["params"][col_names[x + 1]] = float(file_list[i][x + 1])
|
||||
return dict1
|
||||
|
||||
|
||||
@ -77,18 +78,15 @@ def create_dataframe(dict1, variables):
|
||||
print(keys)
|
||||
|
||||
# populate the dict
|
||||
for keys in range(len(dict1["scan"])):
|
||||
if "file_of_origin" in dict1["scan"][keys]:
|
||||
pull_dict["filenames"].append(dict1["scan"][keys]["file_of_origin"].split("/")[-1])
|
||||
else:
|
||||
pull_dict["filenames"].append(dict1["meta"]["original_filename"].split("/")[-1])
|
||||
for keys in range(len(dict1)):
|
||||
pull_dict["filenames"].append(dict1[0]["meta"]["original_filename"].split("/")[-1])
|
||||
|
||||
pull_dict["fit_area"].append(dict1["scan"][keys]["fit"]["fit_area"])
|
||||
pull_dict["int_area"].append(dict1["scan"][keys]["fit"]["int_area"])
|
||||
pull_dict["Counts"].append(dict1["scan"][keys]["Counts"])
|
||||
pull_dict["fit_area"].append(dict1[keys]["fit"]["fit_area"])
|
||||
pull_dict["int_area"].append(dict1[keys]["fit"]["int_area"])
|
||||
pull_dict["Counts"].append(dict1[keys]["Counts"])
|
||||
for key in variables:
|
||||
for i in variables[key]:
|
||||
pull_dict[i].append(_finditem(dict1["scan"][keys], i))
|
||||
pull_dict[i].append(_finditem(dict1[keys], i))
|
||||
|
||||
return pd.DataFrame(data=pull_dict)
|
||||
|
||||
@ -284,42 +282,6 @@ def merge(scan1, scan2, keep=True, monitor=100000):
|
||||
print("merging done")
|
||||
|
||||
|
||||
def add_dict(dict1, dict2):
|
||||
"""adds two dictionaries, meta of the new is saved as meata+original_filename and
|
||||
measurements are shifted to continue with numbering of first dict
|
||||
:arg dict1 : dictionarry to add to
|
||||
:arg dict2 : dictionarry from which to take the measurements
|
||||
:return dict1 : combined dictionary
|
||||
Note: dict1 must be made from ccl, otherwise we would have to change the structure of loaded
|
||||
dat file"""
|
||||
try:
|
||||
if dict1["meta"]["zebra_mode"] != dict2["meta"]["zebra_mode"]:
|
||||
print("You are trying to add scans measured with different zebra modes")
|
||||
return
|
||||
# this is for the qscan case
|
||||
except KeyError:
|
||||
print("Zebra mode not specified")
|
||||
max_measurement_dict1 = len(dict1["scan"])
|
||||
new_filenames = np.arange(
|
||||
max_measurement_dict1 + 1, max_measurement_dict1 + 1 + len(dict2["scan"])
|
||||
)
|
||||
new_meta_name = "meta" + str(dict2["meta"]["original_filename"])
|
||||
if new_meta_name not in dict1:
|
||||
for keys, name in zip(dict2["scan"], new_filenames):
|
||||
dict2["scan"][keys]["file_of_origin"] = str(dict2["meta"]["original_filename"])
|
||||
dict1["scan"][name] = dict2["scan"][keys]
|
||||
|
||||
dict1[new_meta_name] = dict2["meta"]
|
||||
else:
|
||||
raise KeyError(
|
||||
str(
|
||||
"The file %s has alredy been added to %s"
|
||||
% (dict2["meta"]["original_filename"], dict1["meta"]["original_filename"])
|
||||
)
|
||||
)
|
||||
return dict1
|
||||
|
||||
|
||||
def auto(dict):
|
||||
"""takes just unique tuples from all tuples in dictionary returend by scan_dict
|
||||
intendet for automatic merge if you doesent want to specify what scans to merge together
|
||||
@ -344,31 +306,31 @@ def scan_dict(dict, precision=0.5):
|
||||
note: can be checked by "not d", true if empty
|
||||
"""
|
||||
|
||||
if dict["meta"]["zebra_mode"] == "bi":
|
||||
if dict[0]["meta"]["zebra_mode"] == "bi":
|
||||
angles = ["twotheta", "omega", "chi", "phi"]
|
||||
elif dict["meta"]["zebra_mode"] == "nb":
|
||||
elif dict[0]["meta"]["zebra_mode"] == "nb":
|
||||
angles = ["gamma", "omega", "nu"]
|
||||
else:
|
||||
print("Unknown zebra mode")
|
||||
return
|
||||
|
||||
d = {}
|
||||
for i in range(len(dict["scan"])):
|
||||
for j in range(len(dict["scan"])):
|
||||
if dict["scan"][i] != dict["scan"][j]:
|
||||
for i in range(len(dict)):
|
||||
for j in range(len(dict)):
|
||||
if dict[i] != dict[j]:
|
||||
itup = list()
|
||||
for k in angles:
|
||||
itup.append(abs(abs(dict["scan"][i][k]) - abs(dict["scan"][j][k])))
|
||||
itup.append(abs(abs(dict[i][k]) - abs(dict[j][k])))
|
||||
|
||||
if all(i <= precision for i in itup):
|
||||
print(itup)
|
||||
print([dict["scan"][i][k] for k in angles])
|
||||
print([dict["scan"][j][k] for k in angles])
|
||||
if str([np.around(dict["scan"][i][k], 0) for k in angles]) not in d:
|
||||
d[str([np.around(dict["scan"][i][k], 0) for k in angles])] = list()
|
||||
d[str([np.around(dict["scan"][i][k], 0) for k in angles])].append((i, j))
|
||||
print([dict[i][k] for k in angles])
|
||||
print([dict[j][k] for k in angles])
|
||||
if str([np.around(dict[i][k], 0) for k in angles]) not in d:
|
||||
d[str([np.around(dict[i][k], 0) for k in angles])] = list()
|
||||
d[str([np.around(dict[i][k], 0) for k in angles])].append((i, j))
|
||||
else:
|
||||
d[str([np.around(dict["scan"][i][k], 0) for k in angles])].append((i, j))
|
||||
d[str([np.around(dict[i][k], 0) for k in angles])].append((i, j))
|
||||
|
||||
else:
|
||||
pass
|
||||
@ -400,15 +362,15 @@ def variables(dictionary):
|
||||
# find all variables that are in all scans
|
||||
stdev_precision = 0.05
|
||||
all_vars = list()
|
||||
for keys in range(len(dictionary["scan"])):
|
||||
all_vars.append([key for key in dictionary["scan"][keys] if key != "params"])
|
||||
if dictionary["scan"][keys]["params"]:
|
||||
all_vars.append(key for key in dictionary["scan"][keys]["params"])
|
||||
for keys in range(len(dictionary)):
|
||||
all_vars.append([key for key in dictionary[keys] if key != "params"])
|
||||
if dictionary[keys]["params"]:
|
||||
all_vars.append(key for key in dictionary[keys]["params"])
|
||||
|
||||
all_vars = [i for sublist in all_vars for i in sublist]
|
||||
# get the ones that are in all scans
|
||||
b = collections.Counter(all_vars)
|
||||
inall = [key for key in b if b[key] == len(dictionary["scan"])]
|
||||
inall = [key for key in b if b[key] == len(dictionary)]
|
||||
# delete those that are obviously wrong
|
||||
wrong = [
|
||||
"NP",
|
||||
@ -433,15 +395,15 @@ def variables(dictionary):
|
||||
# check for primary variable, needs to be list, we dont suspect the
|
||||
# primary variable be as a parameter (be in scan[params])
|
||||
primary_candidates = list()
|
||||
for key in range(len(dictionary["scan"])):
|
||||
for key in range(len(dictionary)):
|
||||
for i in inall_red:
|
||||
if isinstance(_finditem(dictionary["scan"][key], i), list):
|
||||
if np.std(_finditem(dictionary["scan"][key], i)) > stdev_precision:
|
||||
if isinstance(_finditem(dictionary[key], i), list):
|
||||
if np.std(_finditem(dictionary[key], i)) > stdev_precision:
|
||||
primary_candidates.append(i)
|
||||
# check which of the primary are in every scan
|
||||
primary_candidates = collections.Counter(primary_candidates)
|
||||
second_round_primary_candidates = [
|
||||
key for key in primary_candidates if primary_candidates[key] == len(dictionary["scan"])
|
||||
key for key in primary_candidates if primary_candidates[key] == len(dictionary)
|
||||
]
|
||||
|
||||
if len(second_round_primary_candidates) == 1:
|
||||
@ -455,29 +417,29 @@ def variables(dictionary):
|
||||
# print("secondary candidates", secondary_candidates)
|
||||
# select arrays and floats and ints
|
||||
second_round_secondary_candidates = list()
|
||||
for key in range(len(dictionary["scan"])):
|
||||
for key in range(len(dictionary)):
|
||||
for i in secondary_candidates:
|
||||
if isinstance(_finditem(dictionary["scan"][key], i), float):
|
||||
if isinstance(_finditem(dictionary[key], i), float):
|
||||
second_round_secondary_candidates.append(i)
|
||||
elif isinstance(_finditem(dictionary["scan"][key], i), int):
|
||||
elif isinstance(_finditem(dictionary[key], i), int):
|
||||
second_round_secondary_candidates.append(i)
|
||||
elif isinstance(_finditem(dictionary["scan"][key], i), list):
|
||||
if np.std(_finditem(dictionary["scan"][key], i)) < stdev_precision:
|
||||
elif isinstance(_finditem(dictionary[key], i), list):
|
||||
if np.std(_finditem(dictionary[key], i)) < stdev_precision:
|
||||
second_round_secondary_candidates.append(i)
|
||||
|
||||
second_round_secondary_candidates = collections.Counter(second_round_secondary_candidates)
|
||||
second_round_secondary_candidates = [
|
||||
key
|
||||
for key in second_round_secondary_candidates
|
||||
if second_round_secondary_candidates[key] == len(dictionary["scan"])
|
||||
if second_round_secondary_candidates[key] == len(dictionary)
|
||||
]
|
||||
# print("secondary candidates after second round", second_round_secondary_candidates)
|
||||
# now we check if they vary between the scans
|
||||
third_round_sec_candidates = list()
|
||||
for i in second_round_secondary_candidates:
|
||||
check_array = list()
|
||||
for keys in range(len(dictionary["scan"])):
|
||||
check_array.append(np.average(_finditem(dictionary["scan"][keys], i)))
|
||||
for keys in range(len(dictionary)):
|
||||
check_array.append(np.average(_finditem(dictionary[keys], i)))
|
||||
# print(i, check_array, np.std(check_array))
|
||||
if np.std(check_array) > stdev_precision:
|
||||
third_round_sec_candidates.append(i)
|
||||
|
Loading…
x
Reference in New Issue
Block a user