- View
hpc-toolset-tutorial/coldfront/install.sh
to see how ColdFront is installed - View
hpc-toolset-tutorial/coldfront/local_settings.py
to see how ColdFront is configured - This is where you'd enable or disable any plugins and set variables for your local installation
NOTE: you can also install ColdFront using pip: https://pypi.org/project/coldfront/
- URL https://localhost:2443/
- You'll need to login as some of the users for this tutorial to get things started:
- Login locally as user
hpcadmin
password:ilovelinux
- Logout
- Login locally as user
cgray
password:test123
- Logout
- Login locally as
admin:admin
- Go to Admin interface, Users
- Click on the hpcadmin user
- Make this user a 'superuser' by checking the boxes next to "Staff Status" and "Superuser Status" - SAVE
- Go to Admin interface, User Profiles
- Click on
cgray
check"Is pi"
SAVE - Go back to Admin interface, Click on Resources
- Add a resource:
cluster, cluster name=hpc, attribute: slurm_cluster=hpc
- Logout
- Login locally as PI
cgray:test123
- Create a new project
- Request an allocation for resource: hpc
- Logout
- Login locally as
hpcadmin
- Activate the allocation and set the appropriate allocation attributes:
slurm_account:cgray, slurm_specs:Fairshare=100, slurm_user_specs:Fairshare=parent
-
Login to the frontend container first, then to the coldfront container:
ssh -p 6222 hpcadmin@localhost
ssh coldfront
cd /srv/www
source venv/bin/activate
cd coldfront
-
Let's see what slurm access cgray currently has:
sacctmgr show user cgray -s list
-
Now dump the slurm account/association info from ColdFront's active allocations:
coldfront slurm_dump -c hpc -o ~/slurm_dump
-
Let's see what was created:
ls -al ~/slurm_dump
cat ~/slurm_dump/hpc.cfg
-
Load the slurm dump into slurm database:
sacctmgr load file=~/slurm_dump/hpc.cfg
Type 'Y'
to add the new account & associations for cgray -
Let's look at cgray's slurm account again:
sacctmgr show user cgray -s list
ssh -p 6222 hpcadmin@localhost
- check slurm associations for cgray again: they should now show access to the hpc cluster
sacctmgr show user cgray -s list
su - cgray
sbatch --wrap "sleep 600"
squeue
(the job should be running on a node)
ssh
to the allocated node
ps -ef |grep cgray
- Login to Open OnDemand as
cgray (test123)
https://localhost:3443/ - Go to Active Jobs and click on your running job
- Delete (cancel) the job
- Submit a job using job template
- Launch an interactive Job
- Login to Open XDMoD https://localhost:4443/
-- Under the section "Sign in with tutorial:" Click on "Login Here" and entercgray:test123
- There is currently no data in XDMoD
ssh hpcadmin@xdmod
- In order to see the job data just generated in slurm, we need to ingest the data into Open XDMoD and aggregate it. This is normally done once a day on a typical system but for the purposes of this demo, we have created a script that you can run now:
sudo -u xdmod /srv/xdmod/scripts/shred-ingest-aggregate-all.sh
Note: More information about this script in the Open XDMoD portion of this tutorial
- Login to Open XDMoD https://localhost:4443/
-- Under the section "Sign in with tutorial:" Click on "Login Here" and entercgray:test123
- You should see the data from the job you just ran
NOTE: There won't be much info except that we ran a few jobs. More will be presented in the XDMoD portion of the tutorial
- Login to ColdFront as local user
csimmons (ilovelinux)
https://localhost:2443/
NOTE: You'll need to use a separate browser or incognito window as you'll already be logged in using the single sign on with XDMoD and OnDemand. - Notice there are no projects and no allocations. Logout
- Login locally as user
cgray (test123)
- Click on project
- Click on Add User - search for
csimmons
- Add to allocation
- Login to coldfront container and re-run slurm plug-in commands to add csimmons to slurm associations
ssh coldfront
- Let's see what slurm access csimmons currently has:
sacctmgr show user csimmons -s list
- Now dump the slurm account/association info from ColdFront's active allocations:
coldfront slurm_dump -c hpc -o /tmp/slurm_dump
- Let's see what was created:
ls -al /tmp/slurm_dump
cat /tmp/slurm_dump/hpc.cfg
- Load the slurm dump into slurm database:
sacctmgr load file=/tmp/slurm_dump/hpc.cfg
Type 'Y'
to add the new association for csimmons - Let's look at csimmons's slurm account again:
sacctmgr show user csimmons -s list
Next - Open XDMoD
Previous Step - Accessing the Applications
Docker Tips
Back to Start