CryoSPARC-Sapelo2: Difference between revisions
No edit summary |
No edit summary |
||
Line 55: | Line 55: | ||
===== User Login ===== | ===== User Login ===== | ||
User needs to establish a SSH tunnel to expose the port 39000 from the master node to a local computer. If you are using a Linux or Apple desktop or laptop, you can use the following command in Terminal to establish the ssh tunnel: <blockquote>'''ssh -N -L 39000:128.192.75.59:39000 <username>@ss-cryo.gacrc.uga.edu'''</blockquote>If you are using a Windows desktop or laptop, please download the '''[https://the.earth.li/~sgtatham/putty/latest/x86/plink.exe plink program]''' to use in place of the ssh client:<blockquote>'''plink -ssh -N -L 39000:128.192.75.59:39000 <username>@ss-cryo.gacrc.uga.edu'''</blockquote> | User needs to establish a SSH tunnel to expose the port 39000 from the master node to a local computer. | ||
If you are using a Linux or Apple desktop or laptop, you can use the following command in Terminal to establish the ssh tunnel: <blockquote>'''ssh -N -L 39000:128.192.75.59:39000 <username>@ss-cryo.gacrc.uga.edu'''</blockquote>If you are using a Windows desktop or laptop, please download the '''[https://the.earth.li/~sgtatham/putty/latest/x86/plink.exe plink program]''' to use in place of the ssh client:<blockquote>'''plink -ssh -N -L 39000:128.192.75.59:39000 <username>@ss-cryo.gacrc.uga.edu'''</blockquote> | |||
'''Note:''' Please put the plink.exe in the current directory where you have a command window open. | '''Note:''' Please put the plink.exe in the current directory where you have a command window open. | ||
Line 62: | Line 64: | ||
Once you established the ssh tunnel by running the above command, you can open a browser (Chrome) on the local machine and navigate to<code><nowiki>http://localhost:39000</nowiki></code>. The cryoSPARC User Interface should be presented with the cryoSPARC login page. | Once you established the ssh tunnel by running the above command, you can open a browser (Chrome) on the local machine and navigate to<code><nowiki>http://localhost:39000</nowiki></code>. The cryoSPARC User Interface should be presented with the cryoSPARC login page. | ||
Please refer to https://guide.cryosparc.com/setup-configuration-and-management/how-to-download-install-and-configure/accessing-cryosparc | |||
===== Computing Resource for cryoSPARC ===== | |||
Revision as of 16:05, 16 December 2021
Category
Engineering
Program On
Sapelo2
Version
3.3.1
Author / Distributor
See https://guide.cryosparc.com/
Description
"CryoSPARC (Cryo-EM Single Particle Ab-Initio Reconstruction and Classification) is a state of the art HPC software solution for complete processing of single-particle cryo-electron microscopy (cryo-EM) data. CryoSPARC is useful for solving cryo-EM structures of membrane proteins, viruses, complexes, flexible molecules, small particles, phase plate data and negative stain data." For more information, please see https://guide.cryosparc.com/.
NOTE: Users are required to be added into GACRC cryosparc group before being allowed to run this software. Please fill out the GACRC General Support form to request. We will reach out to you once we received your request.
Configurations
Master node VM:
- Host name: ss-cryo.gacrc.uga.edu
- Intel Xeon processors (8 cores and 24GB of RAM)
- mongodb is installed
Worker nodes:
- Two NVIDIA Tesla K40m nodes, Intel Xeon processors (16 cores and 128GB of RAM) and 8 NVIDIA K40m GPU cards per node.
- cryoSPARC recommends to use SSD or caching particle data. /lscratch/gacrc-cryo was set up on worker nodes for this purpose.
- The amount of space that cryoSPARC can use in /lscratch/gacrc-cryo is capped at 100GB.
cryoSPARC group: cryosparc
cryoSPARC service account: gacrc-cryo
- gacrc-cryo is the service user account that will run the cryoSPARC workflow jobs for all regular cryoSPARC users on the master node and each worker node that will be used for computation.
- Some tasks can only be performed by gacrc-cryo, like start or stop cryosparcm from the master node, user management, connect or update worker nodes to master, etc.
- Regular cryoSPARC users can still run cryosparcm on the master node to check status of the master and its database, using cryosparcm status and cryosparcm checkdb
cryoSPARC group space: /work/cryosparc/ . There are 6 sub-directories in /work/cryosparc/ :
- cryosparc_master/ , cryosparc_worker/ : Master and worker installation folders
- database/ : cryoSPARC database folder
- users/ : cryoSPARC user project folder
- cryosparc_cluster/ : The folder storing cluster integration scripts
- testdataset/ : The folder storing cryoSPARC test data
- src_v3.3.1/ : The folder storing cryoSPARC sources v3.3.1
Running cryoSPARC from Sapelo2
User Login
User needs to establish a SSH tunnel to expose the port 39000 from the master node to a local computer.
If you are using a Linux or Apple desktop or laptop, you can use the following command in Terminal to establish the ssh tunnel:
ssh -N -L 39000:128.192.75.59:39000 <username>@ss-cryo.gacrc.uga.edu
If you are using a Windows desktop or laptop, please download the plink program to use in place of the ssh client:
plink -ssh -N -L 39000:128.192.75.59:39000 <username>@ss-cryo.gacrc.uga.edu
Note: Please put the plink.exe in the current directory where you have a command window open.
Unless you have SSH public key configured, you will be prompted for your MyID password and for Archpass Duo authentication. Once authentication is established, this session prompt will hang and you are ready to go to access the cryoSPARC User Interface.
Once you established the ssh tunnel by running the above command, you can open a browser (Chrome) on the local machine and navigate tohttp://localhost:39000
. The cryoSPARC User Interface should be presented with the cryoSPARC login page.
Please refer to https://guide.cryosparc.com/setup-configuration-and-management/how-to-download-install-and-configure/accessing-cryosparc
Computing Resource for cryoSPARC
Documentation
About cryoSPARC: https://guide.cryosparc.com/
User Interface and Usage Guide: https://guide.cryosparc.com/processing-data/user-interface-and-usage-guide
Accessing the cryoSPARC User Interface https://guide.cryosparc.com/setup-configuration-and-management/how-to-download-install-and-configure/accessing-cryosparc
All Job Types in cryoSPARC: https://guide.cryosparc.com/processing-data/all-job-types-in-cryosparc
Management and Monitoring: https://guide.cryosparc.com/setup-configuration-and-management/management-and-monitoring
Cluster (Slurm) integration: https://guide.cryosparc.com/setup-configuration-and-management/how-to-download-install-and-configure/downloading-and-installing-cryosparc#connect-a-cluster-to-cryosparc
Introductory Tutorial: https://guide.cryosparc.com/processing-data/cryo-em-data-processing-in-cryosparc-introductory-tutorial
Tutorials and Usage Guides: https://guide.cryosparc.com/processing-data/tutorials-and-case-studies
Installation
- Version 3.3.1 master is installed on the master node (ss-cryo.gacrc.uga.edu). Source codes are downloaded in /work/cryosparc/cryosparc_master on the master node.
- Version 3.3.1 workers are installed on two worker nodes (NVIDIA Tesla K40m GPU nodes rb6-[3-4]). Source codes are downloaded in /work/cryosparc/cryosparc_worker on the master ndoe.
System
64-bit Linux