-
Notifications
You must be signed in to change notification settings - Fork 2
Getting access to BioHPC Linux nodes
You may be asked to compute on BioHPC's ECCO Linux nodes (f.i., for very long-running jobs, or for very large memory).
Go to the BioHPC account request page, requesting specifically joining the ECCO group and lv39 (Lars') "lab"
- Go to BioHPC Reservations page, choose "Restricted", and reserve a node:
- cbsuecco02: up to 7 days
- all others: up to 3 days
- in both cases, renewable
See Getting Started Guide and Remote Access. SSH is the best path (if you don't need graphical applications).
Note that, for off-campus access, you will need to use Cornell VPN. Instructions can be found here.
Your default home directory (/home/NETID
) is not shared among group users (same as on CISER). Use /home/ecco_lv39
instead.
- The BioHPC instructions for using FileZilla are great for moving data from your personal workspace to BioHPC.
- In the event that you need to transfer data from CISER to BioHPC (CISER does not have FileZilla installed):
- First, open up a bash shell in the directory that holds the folder which you want to transfer to BioHPC.
- SFTP into BioHPC
sftp [email protected]
. Your password is the same that you use to login to the cbsuecco02 node (or the login node). -
cd
into the desired directory on the BioHPC node. - You need to first create the directory on BioHPC:
mkdir data
. - Use the put command to place the desired folder (i.e. "data") on BioHPC:
put -r data/
. - If you run into an error along the lines of
Can't find request for ID 31425
, try zipping up the files and just transferring the zip file. Once transferred, you can unzip on BioHPC (if you run into issues with the "unzip" command, try using 7z: i.e./programs/bin/util/7z x (ZIPFILE)
) - Give Lars access to my "workdir/mjd443" with
chmod -R a+rwX /workdir/netid
(this command is not permanent and should be run again after any edits to the directory).
-
To use Stata version 16, execute one of the following commands before running
stata-mp
:/usr/local/stata16/stata-mp
orexport PATH=/usr/local/stata16:$PATH
-
Ensure that the tmp directory being used is running on the BioHPC /workdir space, by running the following commands before executing your program(s):
export STATATMP=/workdir/netid/tmp mkdir $STATATMP
-
Don't run Stata interactively via SSH. Instead, execute the program by the following:
stata-mp -b do master.do
To execute a Matlab program (example - program called main.m
) from the command line:
matlab -nodisplay -r "addpath(genpath('.')); main"
Cheatsheet: https://gist.github.com/MohamedAlaa/2961058
- Login via SSH
- Launch tmux with a session name that makes sense, e.g.
tmux new -s AEAREP-xxxx
- Launch your Matlab, Stata, etc job
- Disconnect from tmux:
ctrl-b d
. You don't need to press this both Keyboard shortcut at a time. First press "Ctrl+b" and then press "d". - Log out of SSH
Next time:
- Login via SSH
- Reconnect to your tmux session:
tmux a -t AEAREP-xxxx
- If you forgot what session,
tmux ls
Note: When logged into the compute node, you can call ps ux
to see all your running jobs.
The BioHPC docker
command is docker1
, see here for more details. All files that are shared via the -v
option must reside on /workdir/NETID
and cannot be shared across nodes. To get the files to /workdir/NETID
, the following commands can be used, assuming that your files are in /home/ecco_lv39/Workspace/aearep-$AEAREP
:
- Sync to
workdir
:
AEAREP=12345
[[ -d /workdir/$(id -nu) ]] || mkdir /workdir/$(id -nu)
rsync -auv /home/ecco_lv39/Workspace/aearep-$AEAREP/ /workdir/$(id -nu)/aearep-$AEAREP/
- Sync back to shared drive (once computations are done, or at any time
AEAREP=12345
rsync -auv /workdir/$(id -nu)/aearep-$AEAREP/ /home/ecco_lv39/Workspace/aearep-$AEAREP/
-
Training
-
Tips for authors
-
Tips for replicators
-
Questionnaires
-
Definitions
-
Generic workflow
-
Post-publication replications
-
Technical issues
-
Appendix