Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Getting huge memory file crest_rotamer.xyz . #316

Open
anand5ky opened this issue Jul 5, 2024 · 1 comment
Open

Getting huge memory file crest_rotamer.xyz . #316

anand5ky opened this issue Jul 5, 2024 · 1 comment

Comments

@anand5ky
Copy link

anand5ky commented Jul 5, 2024

I put the system with 85 atoms ,322 electrons on NCI mode conformational search with the following Slurm job script.
#!/bin/sh
#SBATCH --job-name=bbphenol
#SBATCH --account=Chemistry
#SBATCH --nodes=1
#SBATCH --time=7-00:00:00
#SBATCH --error=job.%J.err
#SBATCH --output=job.%J.out
#SBATCH --ntasks-per-node=10
#SBATCH --mem=20GB

module load anaconda/Anaconda3-2024.2

crest bluebox_phenol_crest.xyz --nci --chrg 4 --T 10 > out

after running this script, I am getting a file named crest_rotamer.xyz of 30TB. Please help me to solve this issue.

@pprcht
Copy link
Contributor

pprcht commented Jul 5, 2024

Producing such a large file should be impossible at 85 atoms. Even summing up all snapshots from all metadynamics would never result in this amount of structures.
There must be some issue with either the file system or some special character ending up in the file. See if head -87 crest_rotamers.xyz shows you something is wrong here.
Otherwise the only advice I can give is check that you are using the most current version (3.0.1) and potentially reduce settings to reduce the number of structures (lower ewin, lower MD lengths). And use -T instead of --T please.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants