Skip to main content

Nebari and Nebari Slurm Comparison

At a high level QHub is focused on a Kubernetes and container based deployment of all of its components. Many of the advantages of a container based deployment allow for better security, scalability of components and compute nodes.

QHub-HPC is focused on bringing many of the same features but within a bare metal installation allowing users to fully take advantage of their hardware for performance. Additionally these installations tend to be easier to manage and debug when issues arise (traditional linux sys-admin experience works well here). Due to this approach QHub-HPC lacks containers but achieves workflows and scheduling of compute via Slurm and keeping services available.

Questions to help determine which solution may be best for you:

  1. Are you deploying to the cloud e.g. AWS, GCP, Azure, or Digital Ocean?

QHub is likely your best option. The auto-scalability of QHub compute allows for cost effective usage of the cloud while taking advantage of a managed Kubernetes.

  1. Are you deploying to a bare metal cluster?

QHub-HPC may be your best option since deployment does not require the complexity of managing a kubernetes cluster. If you do have a devops or IT team to help manage kubernetes on bare metal QHub could be a great option. But be advised that managing Kubernetes comes with quite a lot of complexity which the cloud providers handle for us.

  1. Are you concerned about absolute best performance?

QHub-HPC is likely your best option. But note when we say absolute performance we mean your software is able to fully take advantage of your networks Infiniband hardware, uses MPI, and SIMD instructions. Few users fall into this camp and should rarely be a reason to chose QHub-HPC (unless you know why you are making this choice).

Feature Matrix

CoreQHubQHub-HPC
SchedulerKubernetesSystemD and Slurm
User IsolationContainers (cgroups and namespaces)Slurm (cgroups)
Auto-scaling compute nodesX
Cost efficient compute support (Spot/Premptible)X
Static compute nodesX
User ServicesQHubQHub-HPC
Dask GatewayXX
JupyterHubXX
JupyterHub-sshXX
CDSDashboardsXX
Conda-Store environment managementXX
ipyparallelX
Native MPI supportX
Core ServicesQHubQHub-HPC
Monitoring via Grafana and PrometheusXX
Auth integration (OAuth2, OpenID, ldap, kerberos)XX
Role based authorization on JupyterHub, Grafana, Dask-GatewayXX
Configurable user groupsXX
Shared folders for each user's groupXX
Traefik proxyXX
Automated Let's Encrypt and manual TLS certificatesXX
Forward authentication ensuring all endpoints authenticatedX
Backups via ResticX
IntegrationsQHubQHub-HPC
ClearMLX
PrefectX
BodoX