Data from DanMAX is stored centrally at MAX IV and is available for further processing using MAX IV resources – or for download using either the Globus connect service or sftp. Detailed information about downloading the data can be found at the following link.
Data handling guides
All scans at DanMAX results in a number of files written in the HDF5 format (*.h5). The files are written to be NeXuS and/or DXchange compatible.
The raw files are stored in:
while reduced files are stored in:
Ipython notebookes for processing data will be placed in:
Separate experiments/samples may be stored in their own folder. The
raw folder is write protected to protect the raw data. The
scripts, and other folders are not protected, so it is possible to edit/overwrite files in these
Each scan will be saved with the name
#### is a sequential number. It is important to keep track of this number and the corresponding sample/experiment, e.g. by using the Elogy logbook at elogy.maxiv.lu.se.
scan-####.h5 file is a master file that contains the command issued to perform the scan, the scanned parameters (e.g. motor positions), experimental channels (e.g. beam intensity, position, temperature) a snapshot of the instrument at the beginning of the scan and links to the detector data files.
The data from the detectors are located in separate files, e.g.
scan-####_orca.h5. The link from the master file to the detector data is relative, and thus all files must be located in the same folder.
The tomography data is also saved in the DXchange format for straightforward processing using, e.g., TomoPy. This file is located in the
raw folder with the name
scan-####_dxchange.h5. This file still links to the detector data located in the separate files.
To visualize the data and the file structure, we recommend using silx view, which is available on the control computers. The Silx installation files can be found on silx.org
As the PXRD area detector data is compressed using a bit-shuffle algorithm, it is necessary to install hdf5plugin, which can be found on pypi.org
A description of the experimental snapshot parameters is given in the next tab.
Azimuthal integration pipeline
An automated azimuthal integration pipeline is available based on the MATFRAIA algorithm developed by the Birkedal group at Aarhus University (Publication: https://scripts.iucr.org/cgi-bin/paper?S1600577522008232, Python source code). To prepare the pipeline, a
.poni file and a detector mask (
.npy) are needed.
pyFAI-calib2 is installed on the control computers and is commonly used to generate these files. More information about the pipeline can be found on the beamline wiki, under Pilatus 2M CdTe
The integrated data will be available as saved in the following folder, mirroring the folder name and scan number in the /raw/ folder.
It is also possible to process the data in Jupyter notebooks using the MAX IV Jupytherhub, where you need to log in using your DUO credentials. To have access to the regular environment with HDF5, silx, and pyFAI tools, you should select the “HDF5 / Simple Analysis /GPU” kernel.
An automatic tomographic reconstruction pipeline is being developed, and this page will be updated as work progresses.
To prepare the pipeline, a
.tori (Tomographic Reconstruction Information) file is needed. This file is prepared in the ‘Reconstructor’ GUI available at the beamline.
The reconstructed volumes will be available as saved in the following folder, mirroring the folder name and scan number in the /raw/ folder.
It is also possible to process the data in Jupyter notebooks, using either the normal MAX IV Jupyterhub (~8 cores, 50 GB of memory) or the high-performance HPC Jupyterhub (Up to 32 cores, 384 GB of memory and possibility for Nvidia V100 GPU). You will need a VPN connection and DUO credentials to log in to either system.
Select the “Tomography” kernel for access to the correct packages for tomographic reconstruction and analysis.
The experimental snapshot is a recording of the essential instrument parameters at the beginning of the scan. The snap shot is contained in the master file for each scan, and is located under
|Is the undulator gap
|Energy of the Si111 hDCM based on encoder position
|Energy of the hMLM based on encoder position
|Is the horizontal/vertical openings and offsets of the slits
|Is referring to positions of the equipment on the Beam Conditioning Unit
|Is referring to positions of the axis of the Tomograph
|Is referring to positions of the axis of the optical bench and the small gantry
|Is referring to the positions of the equipment on the PXRD2D instrument
|Is referring to the position of the Large Gantry
|Is the X and Y position of the X-ray beam position monitor on the BCU
|The tomography sample to microscope distance. Caution: May not reflect the actual distance if a calibration has not been performed after changing the configuration.
|Is the X and Y position of the PXRD2D pinhole
|Is the position and angle of the PXRD2D hexapod
|Is the position of the optional rotation table at the PXRD2D instrument
|Was the PXRD2D sample spinner spinning at the beginning of the scan?
|Is the location of the PXRD2D beamstop
|PXRD2D sample to detector distance (Pilatus 2M on the large gantry) of it is not tilted. Caution: May not reflect the actual distance if a calibration has not been performed after changing the configuration.
|Is the energy setting on the Pilatus responsible for setting the flatfield of the detector
|Is the threshold energy of the Pilatus responsible for energy discrimination
|PXRD2D sample to XRF detector distance. Caution: May not reflect the actual distance if a calibration has not been performed after changing the configuration.
Some names are self-explanatory and not listed here.