Skip to content

Commit

Permalink
Merge pull request #13 from materialdigital/main
Browse files Browse the repository at this point in the history
merge main
  • Loading branch information
jan-janssen authored Apr 30, 2024
2 parents fdef12f + 6b21203 commit 248fdba
Show file tree
Hide file tree
Showing 54 changed files with 221 additions and 88,970 deletions.
28 changes: 28 additions & 0 deletions .github/workflows/book.yml
Original file line number Diff line number Diff line change
@@ -0,0 +1,28 @@
name: Jupyterbook

on:
pull_request:
branches: [ main ]

jobs:
build:
runs-on: ubuntu-latest
steps:
- uses: actions/checkout@v2
- uses: conda-incubator/setup-miniconda@v2
with:
auto-update-conda: true
python-version: "3.11"
environment-file: book/environment.yml
auto-activate-base: false
- name: Install Jupyterbook
shell: bash -l {0}
run: |
cp book/logo_dark.png .
cp book/_config.yml .
cp book/_toc.yml .
cp book/*.md .
cp -r book/images .
jupyter-book build . --path-output public
- run: mv public/_build/html public_html
- run: touch public_html/.nojekyll
35 changes: 35 additions & 0 deletions .github/workflows/delpoy.yml
Original file line number Diff line number Diff line change
@@ -0,0 +1,35 @@
name: Deploy

on:
push:
branches: [ main ]

jobs:
build:
runs-on: ubuntu-latest
steps:
- uses: actions/checkout@v2
- uses: conda-incubator/setup-miniconda@v2
with:
auto-update-conda: true
python-version: "3.11"
environment-file: book/environment.yml
auto-activate-base: false
- name: Install Jupyterbook
shell: bash -l {0}
run: |
cp book/logo_dark.png .
cp book/_config.yml .
cp book/_toc.yml .
cp book/*.md .
cp -r book/images .
jupyter-book build . --path-output public
- run: mv public/_build/html public_html
- run: touch public_html/.nojekyll
- name: Deploy 🚀
uses: JamesIves/[email protected]
with:
GITHUB_TOKEN: ${{ secrets.GITHUB_TOKEN }}
BRANCH: gh-pages # The branch the action should deploy to.
FOLDER: public_html # The folder the action should deploy.
CLEAN: true
16 changes: 14 additions & 2 deletions README.md
Original file line number Diff line number Diff line change
@@ -1,2 +1,14 @@
# ADIS2023
ADIS 2023
# Quantum Espresso Workflow
The structure optimization of an Aluminium bulk structure followed by the calculation of the bulk modulus by computing the energy for different volumes is implemented in three different workflow frameworks, [Aiida](https://aiida.net), [jobflow](https://materialsproject.github.io/jobflow/) and [pyiron_base](https://pyiron.org).

Test it on: [![Binder](https://mybinder.org/badge_logo.svg)](https://mybinder.org/v2/gh/materialdigital/ADIS2023/HEAD)

## Explanation
* `adis_tools` - quantum espresso parser independent of the workflow frameworks.
* `aiida_qe_basic` - simple Aiida interface.
* `espresso/pseudo/Al.pbe-n-kjpaw_psl.1.0.0.UPF` - This is the pseudo potential for qunatum espresso. By placing it in `~/espresso/pseudo`, it is automatically detected by quantum espresso.
* `aiida.ipynb` - workflow implemented in [Aiida](https://aiida.net)
* `environment.yml` - Conda environment to define the dependencies.
* `jobflow.ipynb` - workflow implemented in [jobflow](https://materialsproject.github.io/jobflow/)
* `postBuild` - script to install `aiida_qe_basic` in the mybinder environment.
* `pyiron_base.ipynb` - workflow implemented in [pyiron_base](https://pyiron.org)
File renamed without changes.
File renamed without changes.
File renamed without changes.
File renamed without changes.
1 change: 1 addition & 0 deletions aiida.ipynb

Large diffs are not rendered by default.

File renamed without changes.
File renamed without changes.
File renamed without changes.
17 changes: 17 additions & 0 deletions book/_config.yml
Original file line number Diff line number Diff line change
@@ -0,0 +1,17 @@
title: ADIS2023
author: pyiron
logo: logo_dark.png

execute:
execute_notebooks : off

html:
extra_navbar : Powered by <a href="https://pyiron.org">pyiron</a>

repository:
url : https://github.com/materialdigital/ADIS2023
path_to_book : ""

launch_buttons:
notebook_interface : jupyterlab
binderhub_url : https://mybinder.org
9 changes: 9 additions & 0 deletions book/_toc.yml
Original file line number Diff line number Diff line change
@@ -0,0 +1,9 @@
format: jb-book
root: README
chapters:
- file: introduction.md
- file: challenges.md
- file: aiida.ipynb
- file: jobflow.ipynb
- file: pyiron_base.ipynb
- file: conclusion.md
92 changes: 92 additions & 0 deletions book/challenges.md
Original file line number Diff line number Diff line change
@@ -0,0 +1,92 @@
# Challenges for workflow frameworks
Over the recent years a number of simulation frameworks have been developed which address various aspects of the development of simulation protocols. The general challenges are:

* Interface with the simulation codes**: While some modern simulation codes already provide Python bindings, the majority require simulation code-specific input files, have specific variable names and the internal unit system used by a given simulation code might differ depending on the community it was developed for.
* **Access to high-performance computing (HPC) resources**: Traditional queuing systems like the Simple Linux Utility for Resource Management (SLURM) handle the execution of compute-intensive tasks. Still in the context of up-scaling simulation protocols or parameter studies the tracking of which parameters lead to a successful calculation and which calculations failed remains a manual task.
* **Efficient Data storage**: The input and output files of many simulation codes in the scientific community were intended to be human readable first and the machine readability was only a secondary concern. This resulted in a wide range of different formats of plain text files which are inefficient when the number of calculations increases.

In the following, we compare the implementation of the same workflow in four different simulation frameworks, namely AiiDA, jobflow, pyiron and Simstack. All four of these frameworks are released as open-source software and developed in the Python programming language

## Example Workflow
![workflow](images/workflow.png)
The comparison is focused on highlighting the implementation of a new simulation code and a simulation workflow in the four workflow frameworks. The calculation of a structure optimization followed by the calculation of the bulk modulus from fitting an energy volume curve with the quantum espresso open-source density functional theory (DFT) simulation code is chosen as an example workflow. The workflow consists of the following four steps:

* Create a face-centred-cubic Aluminium supercell with 4 atoms and a lattice constant of 4.15A.
* Use the quantum espresso simulation code to optimize both the lattice constant and the positions in the crystal structure.
* Apply five strains ranging from -10% to +10% on the optimized structure to generate five strained structures.
* Evaluate these five structures with Quantum Espresso to calculate the energy.
* Plot the resulting energy volume curve.

This workflow covers both a serial dependence of tasks as well as parallel execution of tasks. Rather than using the already existing framework-specific parsers to write the input files and parse the output files for the quantum espresso simulation code the same parser is implemented in all frameworks. This also highlights how new parsers can be developed in a general way to simplify the integration in existing simulation frameworks and enhance the parser's transferability.

## Implementation
The challenge is to develop a workflow in a way that is can be easily integrated in a number of workflow frameworks, to give the users the option to choose the workflow framework which best suits their needs.

### Python Function
For the case of python functions, integrating the function should ideally be as simple as setting a python decorator:
```python
@job_decorator
def my_function(*args, **kwargs):
...
```

The generation of the strained structures can be represented as such a python function. It takes a single structure and a list of strains as an input and returns a list of strained structures as an output:
```python
def generate_structures(structure, strain_lst):
structure_lst = []
for strain in strain_lst:
structure_strain = structure.copy()
structure_strain.set_cell(
structure_strain.cell * strain**(1/3),
scale_atoms=True
)
structure_lst.append(structure_strain)
return structure_lst
```

In the same way, the plotting of the resulting energy volume curve can be represented as python functions:
```python
def plot_energy_volume_curve(volume_lst, energy_lst):
plt.plot(volume_lst, energy_lst)
plt.xlabel("Volume")
plt.ylabel("Energy")
plt.savefig("evcurve.png")
```

### External Executable
For the case of interfacing with an external executable three steps are required. The writing of the input files, the calling of the executable and the parsing of the output files. For interfacing with the quantum espresso DFT simulation code, the function to write the input files could be written as:
```python
def write_input(input_dict, working_directory="."):
filename = os.path.join(working_directory, 'input.pwi')
os.makedirs(working_directory, exist_ok=True)
write(
filename=filename,
images=input_dict["structure"],
Crystal=True,
kpts=input_dict["kpts"],
input_data={
'calculation': input_dict["calculation"],
'occupations': 'smearing',
'degauss': input_dict["smearing"],
},
pseudopotentials=input_dict["pseudopotentials"],
tstress=True,
tprnfor=True
)
```

Here all quantum espresso-specific inputs are defined by the `input_dict` and the additional `working_directory` specifies the directory in which the input should be written. In analogy, the function to collect the output of the quantum espresso calculation also receives the `working_directory` as input and returns the output as a python dictionary:
```python
def collect_output(working_directory="."):
output = parse_pw(os.path.join(working_directory, "pwscf.xml"))
return {
"structure": output["ase_structure"],
"energy": output["energy"],
"volume": output["ase_structure"].get_volume(),
}
```

By choosing to define both the input as well as the output as python dictionary, the interface is very flexible as it can be extended by additional elements.

### Universal Interface
By following this recommendation of using python functions as well as external executables which are interfaced by one function to write the input and another function to parse the output any simulation workflow can be integrated in any of the four frameworks. This is demonstrated below for the workflow of calculating the energy volume curve with quantum espresso.
2 changes: 2 additions & 0 deletions book/conclusion.md
Original file line number Diff line number Diff line change
@@ -0,0 +1,2 @@
# Conclusion
The same workflow was implemented in four different simulation frameworks. On the one hand this simplifies the selection of a simulation framework for users who look for a simulation framework for their own code. On the other hand it also highlights how to develop transferable simulation protocols to improve the portability of a given simulation protocol from one framework to the next.
4 changes: 4 additions & 0 deletions book/environment.yml
Original file line number Diff line number Diff line change
@@ -0,0 +1,4 @@
channels:
- conda-forge
dependencies:
- jupyter-book
Binary file added book/images/workflow.png
Loading
Sorry, something went wrong. Reload?
Sorry, we cannot display this file.
Sorry, this file is invalid so it cannot be displayed.
2 changes: 2 additions & 0 deletions book/introduction.md
Original file line number Diff line number Diff line change
@@ -0,0 +1,2 @@
# Introduction
Why do we need workflow frameworks? For tenth to hundreds of simulations commonly used command line utilities and file-based storage are sufficient but when it comes to thousands of calculations or the coupling with modern machine learning frameworks, these traditional simulation environments are limited.
Binary file added book/logo_dark.png
Loading
Sorry, something went wrong. Reload?
Sorry, we cannot display this file.
Sorry, this file is invalid so it cannot be displayed.
14 changes: 14 additions & 0 deletions environment.yml
Original file line number Diff line number Diff line change
@@ -0,0 +1,14 @@
channels:
- conda-forge
dependencies:
- python=3.11
- pyiron_base=0.8.0
- qe=7.2
- qe-tools=2.0.0
- ase=3.22.1
- matplotlib=3.8.3
- xmlschema=3.2.0
- jobflow=0.1.17
- pymatgen=2024.3.1
- aiida-core=2.5.1
- aiida-core.services
12 changes: 0 additions & 12 deletions eos/README.md

This file was deleted.

28 changes: 0 additions & 28 deletions eos/Si.cif

This file was deleted.

1 change: 0 additions & 1 deletion eos/aiida/.gitignore

This file was deleted.

43 changes: 0 additions & 43 deletions eos/aiida/README.md

This file was deleted.

10 changes: 0 additions & 10 deletions eos/aiida/environment.yml

This file was deleted.

Loading

0 comments on commit 248fdba

Please sign in to comment.