Tutorial: Deploying GPU Workloads With Ori and Google Cloud

Learn how to leverage Ori to deploy GPU workloads on Google Cloud.

In this tutorial we will look at setting up clusters with GPUs on public clouds and adding them to OGC. We will then deploy a couple of Applications with OGC, making use of those GPUs for acceleration.

GPU K8s Clusters on Google Cloud Platform

Adding Clusters with GPUs on Google Cloud platform starts with logging into to your Google Cloud account and accessing GKE.

createcluster-gcp-image1

Select Create on the Kubernetes Engine page. In the pop up screen chose "Standard: You manage your cluster" > Configure

On the Cluster create form, provide an appropriate name or let the default provided, chose Nodepool and select 1 single node (or add more if you are testing a production like environment).

nodepool-gcp-image2

Chose 'Nodes' and select GPUs and then the appropriate GPU type for your use. Tick Enable GPU Sharing and tick I understand the limitations.

nodegpu-gcp-image3

Now select 4vCPU and 15GB of memory for node size - this will depend on your use case, but ML applications usually need a lot of RAM! Below, select spot VMs instance to minimise the cost of your cluster.

nodesize-gcp-image4

Now click Create and then wait until the cluster is available (roughly 5-10 mins is enough). You will now have a cluster with GPUs ready to use.

Onboarding and preparing our GPU cluster

Log-on to your Ori organisation, go to Clusters and select Clusters

clustersgpu-gcp-image5

Select + Add cluster

cluster-add-image7

Fill in the form and click Create

attachcluster-gcp-image8

Now, + Add Labels to your Cluster.

addlabels-gcp-image9

Copy your Cluster Helm install command and run it on your K8s cluster CLI.

helminstall-gcp-image11

helmcli-gcp-image12

The cluster will start attaching to OGC. This will take 5 mins or so.

clusterinstall-gcp-image13

Meanwhile, install the NVIDIA Cuda driver on the K8s cluster CLI.

kubectl apply -f https://raw.githubusercontent.com/GoogleCloudPlatform/container-engine-accelerators/master/nvidia-driver-installer/cos/daemonset-preloaded-latest.yaml

nvidia-gcp-image14

Note the instructions on these pages for further information:

Finally your cluster should be ready for running your deployments.

clusterup-gcp-image15

Testing CUDA deployment

We are now going to test the GPU K8s cluster by running a Cuda test container.

Start by creating a package called CUDA test. Fill in the Application form as per below:

cudatest-package-image16

Click Save and continue and then Skip the Networking and Storage form tabs.

We now need a placement rule on our CUDA test package to ensure it only uses clusters with GPUs. Add the following placement rule to the package.

cuda-placement-image17

Now Save the new placement rule.

Also on the CUDA test package select the Applications tab and then the container.

cuda-container-image18

Click the Resources tab and in the form + Add a new resource as shown below. This will ensure the container has access to the nvidia/gpu resource.

cuda-resources-image19

Finally Save the form.

We are now ready to deploy the CUDA test package! Click Deploy and then Continue.

cuda-deploy-image20

To verify this is running correctly you need to reach out to the cluster CLI again and check the logs of the pod. You should obtain the same output from your running container logs as shown below.

cuda-running-imaeg21

The CUDA test deployment will be either Synchronising or Running. You can Stop it now from the deployment details page or the package details page.

Testing Stable Diffusion Deployment

To test running a Stable Diffusion model follow the steps below.

You will need to create a K8s cluster as above, but with a compatible GPU. I suggest the following choice, but things are always evolving so other GPUs may also work. Dont forget to select Spot VMs to lower the cost.

sdgpu-gcp-image22

sdspot-gcp-image23

Now we can build a Stable diffusion package to deploy it.

Start by selecting + Add package from the Packages menu.

Name your package for example Stable diffusion and fill in the Application and Container form as per below.

sd-package-image24

In the Networking (inbound) form fill it in as per below.

sd-networking-image25

Skip the outbound and storage tabs.

Now back on the Package details page, select Placement rules and add a rule as per below.

sd-placement-image26

Now select the Applications tab again and select the container.

sd-container-imaeg27

And as a last step add the the gpu resource.

sd-resources-image28

Now deploy your package!

After a while (about 5 mins) you should be able to connect to the stable diffusion UI and you can try the text to image feature. Look for the FQDN on the Deployment details page, cut and paste to your browser and don't forget to add the port number 8080.

sd-output-image29

You can also use your cluster's CLI to look into the logs of the stable-diffusion pod. Eventually if all works it will finish with the following message logs.

sdlogs-gcp-image30

Now you have have a Stable Diffusion model running that you can use to generate all the images you want!


Ori simplifies the process of deploying and serving models 

Ori abstracts the infrastructure complexities and makes it simple to deploy and serve models on the edge, hybrid and multi-cloud.

Here we explored how to do it with Google Cloud and we will further explore in other tutorials how to do it with other cloud providers. Once you have your Kubernetes clusters up and running the process of taking models to market becomes very simple with Ori, no matter which cloud or environment your workloads call home.

Subscribe for more news and insights

 

Similar posts

Join the new class of AI infrastructure! 

Build a modern GPU cloud with Ori to accelerate your AI workloads at scale.