Difference between revisions of "Getting Started"

From UFRC
Jump to navigation Jump to search
(48 intermediate revisions by 6 users not shown)
Line 1: Line 1:
[[Category:Docs]]
+
[[Category:Basics]]
 
{|align=right
 
{|align=right
 
   |__TOC__
 
   |__TOC__
 
   |}
 
   |}
Welcome to UF Research Computing! This page is intended to help new users understand and use UFRC resources. Be sure to check out [https://wiki.rc.ufl.edu/doc/Training our training schedule] if you'd like help getting started in person!
+
Welcome to UF Research Computing! This page is intended to help new clients get started on HiPerGator.
  
==From Zero to HiPerGator==
+
=From Zero to HiPerGator=
There are many approaches and resources a researcher can use to analyze their data. Some analyses can be performed on a laptop or a desktop in the lab. Some analyses that rely on proprietary technologies can be done at least partially with resources and applications provided by the vendor. Some researchers are able to 'game' the cloud spot instance market to find relatively cheap resources to perform their analyses on commercial computational resources and are able to install, manage, and use all applications and workflows they need on their own. However, study after study shows that local HPC resources subsidized by the institution still provide the best cost and level of service to academic researchers. Our own [https://www.rc.ufl.edu/services/cost-comparison/ cloud vs. HPC cost comparison] supports this notion. At the University of Florida researchers arguably have access to the best of both worlds - low cost resources and top-notch systems, application, and research facilitation support.  
+
==Initial Consult==
 +
If a face-to-face discussion about the group's needs is needed you can [https://www.rc.ufl.edu/get-support/walk-in-support/ meet one of the UF Research Computing Facilitators] face-to-face or virtually or [https://support.rc.ufl.edu/ submit a support request to start the conversation].  
  
Once analyses move beyond what an individual computer can handle and a UF research group decides to seek additional computational resources the process usually starts from the following steps:
+
==HiPerGator Accounts==
* If a face-to-face discussion about the group's needs is needed one of the group's members or the group's sponsor (usually a UF faculty member) can [https://www.rc.ufl.edu/help/walk-in-support/ meet one of the UF Research Computing Facilitators] or [https://support.rc.ufl.edu/ submit a support request to start the conversation].
+
Group's sponsor has to be the first person to [https://www.rc.ufl.edu/access/account-request/ request a HiPerGator account] indicating that they are a new sponsor. In the process we will create their sponsored group.
* Once a decision is made to try HiPerGator group members should [https://www.rc.ufl.edu/access/request-account/ submit HiPerGator account requests] at which point a group and an account for the group's Faculty or Staff sponsor will be created.
 
* Group's sponsor (usually a UF faculty member) should [https://www.rc.ufl.edu/services/hipergator/request-trial-allocation/ request a trial allocation] on HiPerGator.
 
* Group members can now use HiPerGator for 3 months. They should [https://support.rc.ufl.edu/ submit support requests] if a new application or help with an error is needed or there is a question about applications, workflows, scheduler usage or anything else related to HiPerGator. Note that there are [[Applications|over a thousand installed applications]] on HiPerGator already. See below on how to use environment modules to search for and enable them.
 
* Once the trial is over and the group has a rough idea of what computational and storage resources it needs the group's sponsor should [https://www.rc.ufl.edu/services/purchase-request/ submit a purchase request] to invest into the resources the group needs. Some groups may have access to shared departmental allocations in which case group members should [https://support.rc.ufl.edu/ submit support requests] to be added to the appropriate account. Some examples of departments with shared allocations include the [http://ufgi.ufl.edu/ Genetics Institute], [http://epi.ufl.edu/ Emerging Pathogens Institute], [https://stat.ufl.edu/ Statistics Department], [http://biostat.ufl.edu/ Biostatistics Department], [https://www.eng.ufl.edu/ccmt/ Center for Compressible Multiphase Turbulence (CCMT)], [https://chp.phhp.ufl.edu/research/affiliated-centers/center-for-cognitive-aging-memory-cam/ Cognitive Aging and Memory Clinical Translational Research Program (CAMCTRP)], [https://efrc.ufl.edu/ Center for Molecular Magnetic Quantum Materials], [https://www.phys.ufl.edu/ Physics Department], and [https://plantpath.ifas.ufl.edu/ Plant Pathology Department]. In addition, several research groups working on collaborative projects have shared allocations accessible to members of those projects.
 
  
;Note<nowiki>:</nowiki> a computational allocation is mandatory! No analyses can be run on HiPerGator without a computational allocation.
+
Afterwards, group members will be able to [https://www.rc.ufl.edu/access/account-request/ submit HiPerGator account requests] indicating their PI as the sponsor. Once approved, their linux accounts will be created.
  
==Creating an Account==
+
==Trial Allocation==
To be able to do ''anything'' on HiPerGator you need a UF Research Computing account. To create an account with UF Research Computing, you must read the [https://www.rc.ufl.edu/about/policies/account/ UFRC Account Policy]. After you have reviewed the policy, go to [https://www.rc.ufl.edu/help/account-request/ our website to submit an account request]. You will have to tell us the name of the Principal Investigator (PI) who sponsors the access of the group. Once the PI approves your access to their allocation your account will be created; the username and password will be the same as your GatorLink username and password.  
+
We recommend that the group's sponsor [https://gravity.rc.ufl.edu/access/request-trial-allocation/ requests a '''free''' trial allocation] for storage and computational resources to get the group started on HiPerGator. Group members can then use HiPerGator for the 3 month duration of the trial allocation to figure out what resources and applications they really need.
  
If you are a new Principal Investigator you will need to indicate this on the request form, so that we can create a new group for you. Please note that to do useful work, your group will have to [https://www.rc.ufl.edu/services/rates/ invest into computational resources], or you will have to join a departmental group with a shared allocation.
+
==Purchasing Resources==
 +
After or while the group uses a trial allocation to determine the computational and storage resources it needs the group's sponsor can submit a purchase request for [https://gravity.rc.ufl.edu/access/purchase-request/hpg-hardware/ hardware (5-years)] or [https://gravity.rc.ufl.edu/access/purchase-request/hpg-service/ services (3-months to longer)] to invest into the resources to cover the group's HiPerGator use.
  
==Connecting to HiPerGator==
+
Some groups may have access to shared departmental allocations. In this case, instead of purchasing resources, group members can [https://support.rc.ufl.edu/ request] to be added to the departmental group to gain access to the shared resources.  
To work on HiPerGator you will have to connect to it from your local computer either via SSH (terminal session) or via one of the web/application interfaces we provide such as [[Galaxy]] or Matlab (for Matlab distributed computing toolbox pilot project we are testing).
 
  
''Note about using this guide:'' for any given command, <code><username></code> should be replaced with the UFRC username (same as your GatorLink username).
+
Some examples of departments with shared allocations include the [http://ufgi.ufl.edu/ Genetics Institute], [http://epi.ufl.edu/ Emerging Pathogens Institute], [https://stat.ufl.edu/ Statistics Department], [http://biostat.ufl.edu/ Biostatistics Department], [https://www.eng.ufl.edu/ccmt/ Center for Compressible Multiphase Turbulence (CCMT)], [https://chp.phhp.ufl.edu/research/affiliated-centers/center-for-cognitive-aging-memory-cam/ Cognitive Aging and Memory Clinical Translational Research Program (CAMCTRP)], [https://efrc.ufl.edu/ Center for Molecular Magnetic Quantum Materials], [https://www.phys.ufl.edu/ Physics Department], and [https://plantpath.ifas.ufl.edu/ Plant Pathology Department]. In addition, several research groups working on collaborative projects have shared allocations accessible to members of those projects.
  
For example, if you are referencing the command <code>ssh <username>@hpg.rc.ufl.edu</code> and your Gatorlink username is '''smith''', you would use the command:
+
'''At this point a group is established on HiPerGator and can continue their computational work. See below for more details on the basic use.'''
<pre style="width:75%;">ssh smith@hpg.rc.ufl.edu</pre>
 
  
===Connecting from Windows===
+
=Introduction to Using HiPerGator=
 +
 
 +
;Note: see a short [[Quick Start Guide]] for some hints on getting going and avoiding common pitfalls.
 +
 
 +
To use HiPerGator or HiPerGator-AI you need three basic parts
 +
* Interfaces
 +
You use Interfaces to interact with the system, manage data, initialize computation, and view the results. The main categories of interfaces 'Command-Line' also known as Terminal, Graphical User Interfaces, and Web Interfaces or applications for more specialized use. Some distinctions here are blurred because, for example, you can open a Terminal while using a Web Interface like [[JupyterHub]] or [[Open OnDemand]], but mostly you use a command-line Terminal interface through SSH connections (see below).
 +
* Data Management
 +
To perform research analyses you need to [[Transfer_Data|upload]] and [[Storage|manage]] data. Note that misuse of the storage systems is the second main reason for account suspension after running analyses on login nodes.
 +
*Computation
 +
'''Warning:''' do not run full-scale (normal) analyses on login nodes. [[Development and Testing]] is required reading. The main approach to run computational analyses is through writing [[Sample SLURM Scripts|job scripts]] and sending them to the [[SLURM_Commands|scheduler]] to run. Some interfaces like [[Open OnDemand]], [[JupyterHub], and [[Galaxy]] can manage job scheduling behind the scenes and may be more convenient than job submission from the command-line when appropriate.
 +
 
 +
==Interfaces==
 +
===Connecting to a HiPerGator Terminal via SSH===
 +
To work on HiPerGator you will have to connect to it from your local computer either via SSH (terminal session) or via one of the web/application interfaces we provide such as [[Galaxy]], [[Open_OnDemand|Open OnDemand]], or [[JupyterHub]].
 +
 
 +
Note that the UF security requirements include the use of [[Authentication_With_MFA|Multi-Factor Authentication]] when connecting to HiPerGator. See the [[SSH Multiplexing]] help if you are connecting from Linux or MacOS to reduce the frequency of MFA use.
 +
 
 +
For any given command below, <code><username></code> should be replaced with the UFRC username (same as your GatorLink username).
 +
 
 +
====Connecting from Windows====
 
<div class="mw-collapsible mw-collapsed" style="width:70%; padding: 5px; border: 1px solid gray;">
 
<div class="mw-collapsible mw-collapsed" style="width:70%; padding: 5px; border: 1px solid gray;">
 
''Expand this section to view instructions for logging in with Windows.''
 
''Expand this section to view instructions for logging in with Windows.''
Line 36: Line 52:
 
Since Microsoft Windows does not come with a built-in SSH client, you must download a client from the web.  
 
Since Microsoft Windows does not come with a built-in SSH client, you must download a client from the web.  
  
UFRC recommends using [http://www.chiark.greenend.org.uk/~sgtatham/putty PuTTY] or [http://mobaxterm.mobatek.net/ MobaXterm].  
+
For University-managed computers [http://www.chiark.greenend.org.uk/~sgtatham/putty PuTTY], [https://tabby.sh/ Tabby], and [https://gitforwindows.org/ Git Bash] are approved for 'fast track' installations.
  
 
'''PuTTY'''
 
'''PuTTY'''
* [http://www.chiark.greenend.org.uk/~sgtatham/putty Download PuTTY] to your local machine and start the program .
+
* [http://www.chiark.greenend.org.uk/~sgtatham/putty Download PuTTY] to your local machine and start the program.
 
* Connect to hpg.rc.ufl.edu.
 
* Connect to hpg.rc.ufl.edu.
 
* At the login prompt, enter your username (this should be the same as your GatorLink username)
 
* At the login prompt, enter your username (this should be the same as your GatorLink username)
 
* Enter your password when prompted. You are now connected and ready to work!
 
* Enter your password when prompted. You are now connected and ready to work!
  
'''MobaXterm'''
+
'''Tabby'''
* [http://mobaxterm.mobatek.net/ Download MobaXterm] to your local machine and start the program.
+
* [https://github.com/Eugeny/tabby/releases/latest Download Tabby] to your local machine: tabby-version#-setup.exe or tabby-version#-portable.zip for a portable version.
* Click "Start local terminal"  
+
* Star the program and click Settings > Profiles > +New profile > SSH connection
* Open a terminal and run <code>ssh <username>@hpg.rc.ufl.edu</code>
+
  Name: HiPerGator
* Enter your password when prompted. You are now connected and ready to work!
+
  Host: hpg.rc.ufl.edu
 +
  Username: <username>
 +
  Password: "Set password" or "Add a private key"
 +
* Click "Save"
 +
* Click on the window icon "New tab with profile" and select "HiPerGator hpg.rc.ufl.edu"
 +
* You are now connected and ready to work!
 
</div>
 
</div>
 
</div>
 
</div>
  
===Connecting from Linux===
+
====Connecting from Linux and MacOS====
 
<div class="mw-collapsible mw-collapsed" style="width:70%; padding: 5px; border: 1px solid gray;">
 
<div class="mw-collapsible mw-collapsed" style="width:70%; padding: 5px; border: 1px solid gray;">
''Expand this section to view instructions for logging in with Linux/Unix.''
+
''Expand to view instructions for connecting from Linux or MacOS.''
 
<div class="mw-collapsible-content" style="padding: 5px;">
 
<div class="mw-collapsible-content" style="padding: 5px;">
 
Open a terminal and run
 
Open a terminal and run
Line 62: Line 83:
 
</div>
 
</div>
 
</div>
 
</div>
 +
====Invalid Password====
 +
If you have forgotten your password or need to change it, use the links below to set a new one:
  
===Connecting from MacOS X===
+
[https://my.ufl.edu/ps/signon.html Change or Reset Password]
<div class="mw-collapsible mw-collapsed" style="width:70%; padding: 5px; border: 1px solid gray;">
 
''Expand this section to view instructions for logging in with OS X.''
 
<div class="mw-collapsible-content" style="padding: 5px;">
 
For OS X users, connection instructions are very similar to those for Linux users.
 
  
'''Terminal''', the terminal emulation application for OS X, is located in the folder Applications/Utilities.
+
'''NOTE''': Research Computing now uses Gatorlink Authentication for all passwords. Please use the UF Gatorlink password system for all password issues.
  
Start the Terminal application and run
+
==Data Management==
ssh <username>@hpg.rc.ufl.edu
 
  
Enter your password when the prompt appears. You are now connected and ready to work!
+
===Transferring Data===
</div>
 
</div>
 
 
 
==Transferring Data==
 
 
If you need to transfer datasets to or from HiPerGator and your local computer or another external location you have to pick the appropriate transfer mechanism.
 
If you need to transfer datasets to or from HiPerGator and your local computer or another external location you have to pick the appropriate transfer mechanism.
  
===Samba===
+
====SFTP====
Samba service, also known as a '<code>network share</code>' or '<code>mapped drive</code>' provides you with an ability to connect to some HiPerGator filesystems as locally mapped drives (or mount points on Linux or MacOS X). Once you connected to a share this mechanism provides you with a file transfer option that allows you to use your client computer's native file manager to access and manage your files. Samba works best for moving smaller files, like job scripts, to and from the system. You must be connected to the UF network (either on-campus or through the [[VPN]]) to connect to Samba shares.
+
SFTP, or secure file transfer, works well for small to medium data transfers and is appropriate for both small and large data files.
 
 
* [[Samba_Access|See the page on accessing Samba]] for setup information specific to your computer's operating system.
 
 
 
===SFTP===
 
SFTP, or secure file transfer, works well for small to mediaum data transfers and is appropriate for both small and large data files.
 
  
If you would like to use a Graphical User Interface ecure file transfer client we recommend:  
+
If you would like to use a Graphical User Interface secure file transfer client we recommend:  
* [https://filezilla-project.org/download.php?show_all=1 FileZilla] for Windows or MacOS X.
+
* <s>[https://filezilla-project.org/download.php?show_all=1 FileZilla] for Windows or MacOS X.</s>
 +
: The FileZilla installer contains adware/malware and probably should be avoided.
 
* [http://winscp.net/eng/index.php WinSCP] for Windows.
 
* [http://winscp.net/eng/index.php WinSCP] for Windows.
* [http://cyberduck.ch/ Cyberduck] for MacOS X.
+
* [http://cyberduck.io/ Cyberduck] for MacOS X and Windows.
  
 
After you have chosen and downloaded a client, configure the client to connect to <code>hpg.rc.ufl.edu</code>, specifying port number 22. Use your username and password to log in.
 
After you have chosen and downloaded a client, configure the client to connect to <code>hpg.rc.ufl.edu</code>, specifying port number 22. Use your username and password to log in.
  
===Rsync===
+
====Rsync====
 
If you prefer to use the command-line or to get maximum efficiency from your data transfers Rsync, which is an incremental file transfer utility that minimizes network usage, is a good choice. It does so by transmitting only the differences between local and remote files rather than transmitting complete files every time a sync is run as SFTP does. Rsync is best used for tasks like synchronizing files stored across multiple subdirectories, or updating large data sets. It works well both for small and large files. [[Rsync|See the Rsync page]] for instructions on using rsync.
 
If you prefer to use the command-line or to get maximum efficiency from your data transfers Rsync, which is an incremental file transfer utility that minimizes network usage, is a good choice. It does so by transmitting only the differences between local and remote files rather than transmitting complete files every time a sync is run as SFTP does. Rsync is best used for tasks like synchronizing files stored across multiple subdirectories, or updating large data sets. It works well both for small and large files. [[Rsync|See the Rsync page]] for instructions on using rsync.
  
 
===Globus===
 
===Globus===
 
Globus is a high-performance mechanism for file transfer. Globus works especially well for transferring large files or data sets  
 
Globus is a high-performance mechanism for file transfer. Globus works especially well for transferring large files or data sets  
 +
* [[Globus|See the Globus page]] for setup and configuration information.
 +
 +
===Samba===
 +
Samba service, also known as a '<code>network share</code>' or '<code>mapped drive</code>' provides you with an ability to connect to some HiPerGator filesystems as locally mapped drives (or mount points on Linux or MacOS X).Once you connected to a share this mechanism provides you with a file transfer option that allows you to use your client computer's native file manager to access and manage your files.  UFRC Samba setup does not provide high performance, so try to use it sparingly and for smaller files, like job scripts or analysis reports to be copied to  your local system. You must be connected to the UF network (either on-campus or through the [[VPN]]) to connect to Samba shares.
  
* [[Globus|See the Globus page]] for setup and configuration information.
+
* [[Samba_Access|See the page on accessing Samba]] for setup information specific to your computer's operating system.
  
 
===Automounted Paths===
 
===Automounted Paths===
Line 118: Line 132:
 
You can also use your favorite file editor on your local machine, and then transfer the files to the cluster afterward. A caveat to this is that files created on Windows machines usually contain unprintable characters, which may be misinterpreted by Linux command interpreters (shells). If this happens, there is a utility called <code>dos2unix</code> that you can use to convert the text file from DOS/Windows formatting to Linux formatting.
 
You can also use your favorite file editor on your local machine, and then transfer the files to the cluster afterward. A caveat to this is that files created on Windows machines usually contain unprintable characters, which may be misinterpreted by Linux command interpreters (shells). If this happens, there is a utility called <code>dos2unix</code> that you can use to convert the text file from DOS/Windows formatting to Linux formatting.
  
==Using installed software==
+
==Computation==
 +
===Using installed software===
 
The full list of software available for use can be viewed on the [[Installed_Software|Installed Software]] page. Access to installed software is provided through [[Modules|Environment Modules]].
 
The full list of software available for use can be viewed on the [[Installed_Software|Installed Software]] page. Access to installed software is provided through [[Modules|Environment Modules]].
  
Line 135: Line 150:
 
There are some useful commands and utilities in a [[UFRC_environment_module|'ufrc' environment module]] in addition to installed applications.
 
There are some useful commands and utilities in a [[UFRC_environment_module|'ufrc' environment module]] in addition to installed applications.
  
==Doing Interactive Testing or Development==
+
===Interactive Testing or Development===
 
You don't always have to use the SLURM scheduler. When all you need is a quick shell session to run a command or two, write and/or test a job script, or compile some code use [[Development_and_Testing|SLURM Dev Sessions]].
 
You don't always have to use the SLURM scheduler. When all you need is a quick shell session to run a command or two, write and/or test a job script, or compile some code use [[Development_and_Testing|SLURM Dev Sessions]].
  
==Running Graphical Programs==
+
===Running Graphical Programs===
 
It is possible to run programs that use a graphical user interface (GUI) on the system. However, doing so requires an installation of and configuration of additional software on the client computer.  
 
It is possible to run programs that use a graphical user interface (GUI) on the system. However, doing so requires an installation of and configuration of additional software on the client computer.  
  
 
Please see the [[GUI_Programs|GUI Programs]] page for information on running graphical user interface applications at UFRC.
 
Please see the [[GUI_Programs|GUI Programs]] page for information on running graphical user interface applications at UFRC.
  
==Scheduling jobs using SLURM==
+
===Scheduling computational jobs===
 
UFRC uses the Simple Linux Utility for Resource Management, or '''SLURM''', to allocate resources and schedule jobs. Users can create SLURM job scripts to submit jobs to the system. These scripts can, and should, be modified in order to control several aspects of your job, like resource allocation, email notifications, or an output destination.
 
UFRC uses the Simple Linux Utility for Resource Management, or '''SLURM''', to allocate resources and schedule jobs. Users can create SLURM job scripts to submit jobs to the system. These scripts can, and should, be modified in order to control several aspects of your job, like resource allocation, email notifications, or an output destination.
  
Line 150: Line 165:
  
  
To submit a job script from one of login nodes via hpg.rc.ufl.edu, use the following command:
+
To submit a job script from one of the login nodes accessed via hpg.rc.ufl.edu, use the following command:
 
<pre>
 
<pre>
 
$ sbatch <your_job_script>
 
$ sbatch <your_job_script>
Line 161: Line 176:
 
View [[SLURM_Commands]] for more useful SLURM commands.
 
View [[SLURM_Commands]] for more useful SLURM commands.
  
==Managing Cores and Memory==
+
====Managing Cores and Memory====
The amount of resources within an investment is calculated in NCU (Normalized Computing Units), which correspond to 1 CPU core and about 3.5GB of memory for each NCU. Resources can be consumed either by using more CPU cores or by more memory. System will automatically calculate the usage according to the <code>1 CPU core to 3.5GB memory</code> ratio. This ratio is determined by the form factor of the compute nodes on HiPerGator. Most nodes have 32 or 64 cores and 128 or 256 GB of RAM. The [[Large-Memory SMP Servers|bigmem]] nodes and the newer Skylake nodes have a higher ratio of 16 GB/core and 6 GB/core, respectively. The majority of HiPerGator nodes, old and new, have the same ratio of about 4 GB or RAM per core, which, after accounting of the operating systems and system services, leaves about 3.5 GB usable for jobs. See [[Available_Node_Features]] for the exact data on resources available on all types of nodes on HiPerGator.
+
See [[Account and QOS limits under SLURM]] for the main documentation on efficient management of computational resources.
  
You must specify both the number of cores and the amount of RAM needed in the job script for SLURM with the pmem or mem-mer-cpu option. Otherwise, the job will be assigned the default 600mb of memory.
+
The amount of resources within an investment is calculated in NCU (Normalized Computing Units), which correspond to 1 CPU core and about 3.5GB of memory for each NCU purchased. CPUs (cores) and RAM are allocated to jobs independently as requested by your job script.
  
For example, if you need to run with 4 cores and pmem=6GB of RAM, then SLURM will provide 4 cores subtract 4 NCUs from the allocation in the QOS you are using investment because the job RAM limit is below the 3.5 GB/core.
+
Your group's investment can run out of **cores** (SLURM may show <code>QOSGrpCpuLimit</code> in the reason a job is pending) OR **memory** (SLURM may show <code>QOSGrpMemLimit</code> in the reason a job is pending) depending on current use by running jobs.
 +
The majority of HiPerGator nodes have the same ratio of about 4 GB of RAM per core, which, after accounting for the operating system and system services, leaves about 3.5 GB usable for jobs; hence the ratio of 1 core and 3.5GB RAM per NCU.
  
If, for example, you need to use 1 core and pmem=42GB to have more memory available to the application you are running then SLURM will allocate 42GB of memory on the node which it assigns to your job and will subtract 12 NCUs out of the QOS allocation even though you are using only one core, because you are using the amount of memory associated with 12 NCUs.
+
Most HiPerGator nodes have 32 cores and 128 GB RAM (~30,000 cores in the newer part of the cluster) or 64 cores and 256 GB RAM (~16,000 cores in the older part of the cluster). The [[Large-Memory SMP Servers|bigmem]] nodes and the newer Skylake nodes have a higher ratio of 16 GB/core and 6 GB/core, respectively. See [[Available_Node_Features]] for the exact data on resources available on all types of nodes on HiPerGator.
 +
 
 +
You must specify both the number of cores and the amount of RAM needed in the job script for SLURM with the <code>--mem</code> (total job memory) or <code>--mem-per-cpu</code> (per-core memory) options. Otherwise, the job will be assigned the default 600mb of memory.
 
   
 
   
If you need more than 128 GB or RAM, you can only run on the older nodes, which have 256 GB of RAM, or on the bigmem nodes, which have up to 1.5 TB of RAM.
+
If you need more than 128 GB of RAM, you can only run on the older nodes, which have 256 GB of RAM, or on the bigmem nodes, which have up to 1.5 TB of RAM.
  
 
See [[Account and QOS limits under SLURM]] for an extensive explanation of QOS and SLURM account use.
 
See [[Account and QOS limits under SLURM]] for an extensive explanation of QOS and SLURM account use.
Line 179: Line 197:
 
==Visual Overview==
 
==Visual Overview==
 
The diagram below shown a high-level overview of HiPerGator use. We will go over each part in sections below
 
The diagram below shown a high-level overview of HiPerGator use. We will go over each part in sections below
 +
 
[[file:HiPerGator.png|800px]]
 
[[file:HiPerGator.png|800px]]

Revision as of 17:52, 6 May 2022

Welcome to UF Research Computing! This page is intended to help new clients get started on HiPerGator.

From Zero to HiPerGator

Initial Consult

If a face-to-face discussion about the group's needs is needed you can meet one of the UF Research Computing Facilitators face-to-face or virtually or submit a support request to start the conversation.

HiPerGator Accounts

Group's sponsor has to be the first person to request a HiPerGator account indicating that they are a new sponsor. In the process we will create their sponsored group.

Afterwards, group members will be able to submit HiPerGator account requests indicating their PI as the sponsor. Once approved, their linux accounts will be created.

Trial Allocation

We recommend that the group's sponsor requests a free trial allocation for storage and computational resources to get the group started on HiPerGator. Group members can then use HiPerGator for the 3 month duration of the trial allocation to figure out what resources and applications they really need.

Purchasing Resources

After or while the group uses a trial allocation to determine the computational and storage resources it needs the group's sponsor can submit a purchase request for hardware (5-years) or services (3-months to longer) to invest into the resources to cover the group's HiPerGator use.

Some groups may have access to shared departmental allocations. In this case, instead of purchasing resources, group members can request to be added to the departmental group to gain access to the shared resources.

Some examples of departments with shared allocations include the Genetics Institute, Emerging Pathogens Institute, Statistics Department, Biostatistics Department, Center for Compressible Multiphase Turbulence (CCMT), Cognitive Aging and Memory Clinical Translational Research Program (CAMCTRP), Center for Molecular Magnetic Quantum Materials, Physics Department, and Plant Pathology Department. In addition, several research groups working on collaborative projects have shared allocations accessible to members of those projects.

At this point a group is established on HiPerGator and can continue their computational work. See below for more details on the basic use.

Introduction to Using HiPerGator

Note
see a short Quick Start Guide for some hints on getting going and avoiding common pitfalls.

To use HiPerGator or HiPerGator-AI you need three basic parts

  • Interfaces

You use Interfaces to interact with the system, manage data, initialize computation, and view the results. The main categories of interfaces 'Command-Line' also known as Terminal, Graphical User Interfaces, and Web Interfaces or applications for more specialized use. Some distinctions here are blurred because, for example, you can open a Terminal while using a Web Interface like JupyterHub or Open OnDemand, but mostly you use a command-line Terminal interface through SSH connections (see below).

  • Data Management

To perform research analyses you need to upload and manage data. Note that misuse of the storage systems is the second main reason for account suspension after running analyses on login nodes.

  • Computation

Warning: do not run full-scale (normal) analyses on login nodes. Development and Testing is required reading. The main approach to run computational analyses is through writing job scripts and sending them to the scheduler to run. Some interfaces like Open OnDemand, [[JupyterHub], and Galaxy can manage job scheduling behind the scenes and may be more convenient than job submission from the command-line when appropriate.

Interfaces

Connecting to a HiPerGator Terminal via SSH

To work on HiPerGator you will have to connect to it from your local computer either via SSH (terminal session) or via one of the web/application interfaces we provide such as Galaxy, Open OnDemand, or JupyterHub.

Note that the UF security requirements include the use of Multi-Factor Authentication when connecting to HiPerGator. See the SSH Multiplexing help if you are connecting from Linux or MacOS to reduce the frequency of MFA use.

For any given command below, <username> should be replaced with the UFRC username (same as your GatorLink username).

Connecting from Windows

Expand this section to view instructions for logging in with Windows.

Since Microsoft Windows does not come with a built-in SSH client, you must download a client from the web.

For University-managed computers PuTTY, Tabby, and Git Bash are approved for 'fast track' installations.

PuTTY

  • Download PuTTY to your local machine and start the program.
  • Connect to hpg.rc.ufl.edu.
  • At the login prompt, enter your username (this should be the same as your GatorLink username)
  • Enter your password when prompted. You are now connected and ready to work!

Tabby

  • Download Tabby to your local machine: tabby-version#-setup.exe or tabby-version#-portable.zip for a portable version.
  • Star the program and click Settings > Profiles > +New profile > SSH connection
 Name: HiPerGator
 Host: hpg.rc.ufl.edu
 Username: <username>
 Password: "Set password" or "Add a private key"
  • Click "Save"
  • Click on the window icon "New tab with profile" and select "HiPerGator hpg.rc.ufl.edu"
  • You are now connected and ready to work!

Connecting from Linux and MacOS

Expand to view instructions for connecting from Linux or MacOS.

Open a terminal and run

ssh <username>@hpg.rc.ufl.edu

Enter your password when the prompt appears. You are now connected and ready to work!

Invalid Password

If you have forgotten your password or need to change it, use the links below to set a new one:

Change or Reset Password

NOTE: Research Computing now uses Gatorlink Authentication for all passwords. Please use the UF Gatorlink password system for all password issues.

Data Management

Transferring Data

If you need to transfer datasets to or from HiPerGator and your local computer or another external location you have to pick the appropriate transfer mechanism.

SFTP

SFTP, or secure file transfer, works well for small to medium data transfers and is appropriate for both small and large data files.

If you would like to use a Graphical User Interface secure file transfer client we recommend:

The FileZilla installer contains adware/malware and probably should be avoided.

After you have chosen and downloaded a client, configure the client to connect to hpg.rc.ufl.edu, specifying port number 22. Use your username and password to log in.

Rsync

If you prefer to use the command-line or to get maximum efficiency from your data transfers Rsync, which is an incremental file transfer utility that minimizes network usage, is a good choice. It does so by transmitting only the differences between local and remote files rather than transmitting complete files every time a sync is run as SFTP does. Rsync is best used for tasks like synchronizing files stored across multiple subdirectories, or updating large data sets. It works well both for small and large files. See the Rsync page for instructions on using rsync.

Globus

Globus is a high-performance mechanism for file transfer. Globus works especially well for transferring large files or data sets

Samba

Samba service, also known as a 'network share' or 'mapped drive' provides you with an ability to connect to some HiPerGator filesystems as locally mapped drives (or mount points on Linux or MacOS X).Once you connected to a share this mechanism provides you with a file transfer option that allows you to use your client computer's native file manager to access and manage your files. UFRC Samba setup does not provide high performance, so try to use it sparingly and for smaller files, like job scripts or analysis reports to be copied to your local system. You must be connected to the UF network (either on-campus or through the VPN) to connect to Samba shares.

Automounted Paths

Note: NFS-based storage on our systems are typically automounted, which means they are dynamically mounted only when users are actually accessing them. For example if you have an invested folder as /orange/smith, to access it you will have to specifically type in the full path of "/orange/smith" to be able to see the contents and access them. Directly browsing /orange will not show the smith sub-folder unless someone else is using it coincidentally. Automounted folders are pretty common on the systems, they include /orange, /bio, /rlts and even /home etc.

Editing your files

Several methods exist for editing your files on the cluster.

Native Editors

  • vi - The visual editor (vi) is the traditional Unix editor; however, it is not necessarily the most intuitive editor. View a tutorial for using vi
  • emacs - Emacs is a much heavier duty editor, but again has the problem of having commands that are non-intuitive. View a tutorial for using emacs
  • pico - While pico is not installed on the system, nano is installed, and is a pico work-a-like.
  • nano - Nano has a good bit of on-screen help to make it easier to use.

External Editors

You can also use your favorite file editor on your local machine, and then transfer the files to the cluster afterward. A caveat to this is that files created on Windows machines usually contain unprintable characters, which may be misinterpreted by Linux command interpreters (shells). If this happens, there is a utility called dos2unix that you can use to convert the text file from DOS/Windows formatting to Linux formatting.

Computation

Using installed software

The full list of software available for use can be viewed on the Installed Software page. Access to installed software is provided through Environment Modules.

The following command can be used to browse the full list of available modules, along with short descriptions of the applications they make available:

module spider

To load a module, use the following command:

module load <module_name>

For more information on loading modules to access software, view the page on the basic usage of environment modules.

There are some useful commands and utilities in a 'ufrc' environment module in addition to installed applications.

Interactive Testing or Development

You don't always have to use the SLURM scheduler. When all you need is a quick shell session to run a command or two, write and/or test a job script, or compile some code use SLURM Dev Sessions.

Running Graphical Programs

It is possible to run programs that use a graphical user interface (GUI) on the system. However, doing so requires an installation of and configuration of additional software on the client computer.

Please see the GUI Programs page for information on running graphical user interface applications at UFRC.

Scheduling computational jobs

UFRC uses the Simple Linux Utility for Resource Management, or SLURM, to allocate resources and schedule jobs. Users can create SLURM job scripts to submit jobs to the system. These scripts can, and should, be modified in order to control several aspects of your job, like resource allocation, email notifications, or an output destination.


To submit a job script from one of the login nodes accessed via hpg.rc.ufl.edu, use the following command:

$ sbatch <your_job_script>

To check the status of submitted jobs, use the following command:

$ squeue -u <username>

View SLURM_Commands for more useful SLURM commands.

Managing Cores and Memory

See Account and QOS limits under SLURM for the main documentation on efficient management of computational resources.

The amount of resources within an investment is calculated in NCU (Normalized Computing Units), which correspond to 1 CPU core and about 3.5GB of memory for each NCU purchased. CPUs (cores) and RAM are allocated to jobs independently as requested by your job script.

Your group's investment can run out of **cores** (SLURM may show QOSGrpCpuLimit in the reason a job is pending) OR **memory** (SLURM may show QOSGrpMemLimit in the reason a job is pending) depending on current use by running jobs. The majority of HiPerGator nodes have the same ratio of about 4 GB of RAM per core, which, after accounting for the operating system and system services, leaves about 3.5 GB usable for jobs; hence the ratio of 1 core and 3.5GB RAM per NCU.

Most HiPerGator nodes have 32 cores and 128 GB RAM (~30,000 cores in the newer part of the cluster) or 64 cores and 256 GB RAM (~16,000 cores in the older part of the cluster). The bigmem nodes and the newer Skylake nodes have a higher ratio of 16 GB/core and 6 GB/core, respectively. See Available_Node_Features for the exact data on resources available on all types of nodes on HiPerGator.

You must specify both the number of cores and the amount of RAM needed in the job script for SLURM with the --mem (total job memory) or --mem-per-cpu (per-core memory) options. Otherwise, the job will be assigned the default 600mb of memory.

If you need more than 128 GB of RAM, you can only run on the older nodes, which have 256 GB of RAM, or on the bigmem nodes, which have up to 1.5 TB of RAM.

See Account and QOS limits under SLURM for an extensive explanation of QOS and SLURM account use.

Getting help

If you are having problems using the UFRC system, please let our staff know by submitting a support request.

Visual Overview

The diagram below shown a high-level overview of HiPerGator use. We will go over each part in sections below

HiPerGator.png