Skip to content

Commit

Permalink
Gcp example (#7)
Browse files Browse the repository at this point in the history
* wip

* Update docs/source/how-to/cloud/gcp.mdx

Co-authored-by: Alvaro Bartolome <[email protected]>

* Update docs/source/how-to/cloud/gcp.mdx

Co-authored-by: Alvaro Bartolome <[email protected]>

* Update docs/source/how-to/cloud/gcp.mdx

Co-authored-by: Alvaro Bartolome <[email protected]>

* Update docs/source/how-to/cloud/gcp.mdx

Co-authored-by: Alvaro Bartolome <[email protected]>

* Update docs/source/how-to/cloud/gcp.mdx

Co-authored-by: Alvaro Bartolome <[email protected]>

* gcp example

* removed soon

---------

Co-authored-by: Alvaro Bartolome <[email protected]>
  • Loading branch information
philschmid and alvarobartt authored Oct 31, 2024
1 parent 50335ce commit 092e96e
Show file tree
Hide file tree
Showing 4 changed files with 134 additions and 4 deletions.
2 changes: 1 addition & 1 deletion docs/source/faq.mdx
Original file line number Diff line number Diff line change
Expand Up @@ -22,7 +22,7 @@ Yes, HUGS is available on major cloud platforms. For specific instructions, chec

- [AWS](./how-to/cloud/aws)
- [DigitalOcean](./how-to/cloud/digital-ocean)
- [Google Cloud](./how-to/cloud/gcp) (coming soon)
- [Google Cloud](./how-to/cloud/gcp)
- [Microsoft Azure](./how-to/cloud/azure) (coming soon)

## How does HUGS pricing work?
Expand Down
132 changes: 131 additions & 1 deletion docs/source/how-to/cloud/gcp.mdx
Original file line number Diff line number Diff line change
@@ -1,3 +1,133 @@
# HUGS on Google Cloud

Coming soon!
The Hugging Face Generative AI Services, also known as HUGS, can be deployed in Google Cloud via the Google Cloud Marketplace offering.

This collaboration brings Hugging Face's extensive library of pre-trained models and their Text Generation Inference (TGI) solution to Google Cloud customers, enabling seamless integration of state-of-the-art Large Language Models (LLMs) within the Google Cloud infrastructure.

HUGS provides access to a hand-picked and manually benchmarked collection of the most performant and latest open LLMs hosted in the Hugging Face Hub to TGI-optimized container applications, allowing users to deploy third-party Kubernetes applications on AWS or on-premises environments.

With HUGS, developers can easily find, subscribe to, and deploy Hugging Face models using AWS infrastructure, leveraging the power of NVIDIA GPUs on optimized, zero-configuration TGI containers.

## Subscribe to HUGS on AWS Marketplace

1. Go to [HUGS Google Cloud Marketplace listing](https://console.cloud.google.com/marketplace/product/huggingface-public/hugs)

![HUGS on Google Cloud Marketplace](https://huggingface.co/datasets/huggingface/documentation-images/resolve/main/hugs/gcp/hugs-marketplace-listing.png)

2. Subscribe to the product in Google Cloud by following the instructions on the page. At the time of writing (October 2024), the steps are to:

1. Click `Purchase`, then go to the next page.
2. Configure the order by selecting the right plan, billing account, and confirming the terms. Then click `Subscribe`.

![HUGS Configuration on Google Cloud Marketplace](https://huggingface.co/datasets/huggingface/documentation-images/resolve/main/hugs/gcp/hugs-configuration.png)

3. You should see a "Your order request has been sent to Hugging Face" message. With a button "Go to Product Page". Click on it.

![HUGS Confirmation on Google Cloud Marketplace](https://huggingface.co/datasets/huggingface/documentation-images/resolve/main/hugs/gcp/hugs-confirmation.png)

<Tip>

To know whether you are subscribed or not, you can either see if the "Purchase" button or "Configure" button is enabled on the product page, meaning that either you or someone else from your organization has already requested access for your account.

</Tip>


## Deploy HUGS on Google Cloud GKE

This example showcases how to deploy a HUGS container and model on Google Cloud GKE.

<Tip>

This example assumes that you have an Google Cloud Account, that you have [installed and setup the Google Cloud CLI](https://cloud.google.com/sdk/docs/install), and that you are logged in into your account with the necessary permissions to subscribe to offerings in the Google Cloud Marketplace, and create and manage IAM permissions and resources such as Google Kubernetes Engine (GKE).

</Tip>

When deploying HUGS on Google Cloud through the UI you can either select an existing GKE cluster or create a new one. If you want to create a new one, you can follow the instructions [here](https://cloud.google.com/kubernetes-engine/docs/how-to/creating-a-cluster). Additionally you need to define:

* Namespace: The namespace to deploy the HUGS container and model.
* App Instance Name: The name of the HUGS container.
* Hugs Model Id: Select the model you want to deploy from the Hugging Face Hub. You can find all supported model [here](../models)
* GPU Number: The number of GPUs you have available and want to use for the deployment, make sure to check the [supported model matrix](../../models) to know which model requires GPUs.
* GPU Type: The type of GPU you have available inside your GKE cluster.
* Reporting Service Account: The service account to use for reporting.

![HUGS Deployment Configuration](https://huggingface.co/datasets/huggingface/documentation-images/resolve/main/hugs/gcp/hugs-deploy.png)

Next you click on `Deploy` and wait for the deployment to finish. This takes around 10-15 minutes.

<Tip>

If you want to better understand the different deployment options you have, e.g. 1x NVIDIA L4 GPU for Meta Llama 3.1 8B Instruct, you can checkout the [supported model matrix](../../models).

</Tip>

## Send request to the HUGS application

Every HUGS application includes instructions on how to retrieve the Ingress IP address and port to send requests to the application. A HUGS deployment is a deployment of a HELM chart that includes our model container, marketplace agent (sidecar), a volume and a ingress load balancer to make the application accessible from outside the cluster.

![HUGS Ingress](https://huggingface.co/datasets/huggingface/documentation-images/resolve/main/hugs/gcp/applicaiton-instructions.png)


Alternatively, you can also use the Messages API via `openai`. Learn more about inference [here](../guides/inference).


## Create a GPU GKE Cluster for HUGS

To deploy HUGS on Google Cloud, you'll need a GKE cluster with GPU support. Here's a step-by-step guide to create one:

1. Ensure you have the [Google Cloud CLI installed and configured](https://cloud.google.com/sdk/docs/install-sdk).

2. Set up environment variables for your cluster configuration:

```bash
export PROJECT_ID="your-project-id" # Your Google Cloud Project ID which is subscribed to HUGS
export CLUSTER_NAME="hugs-cluster" # The name of the GKE cluster
export LOCATION="us-central1" # The location of the GKE cluster
export MACHINE_TYPE="g2-standard-12" # The machine type of the GKE cluster
export GPU_TYPE="nvidia-l4" # The type of GPU to use
export GPU_COUNT=1 # The number of GPUs to use
```

3. Create the GKE cluster:

```bash
gcloud container clusters create $CLUSTER_NAME \
--project=$PROJECT_ID \
--zone=$LOCATION \
--release-channel=stable \
--cluster-version=1.29 \
--machine-type=$MACHINE_TYPE \
--num-nodes=1 \
--no-enable-autoprovisioning
```

4. Add a GPU node pool to the cluster:

```bash
gcloud container node-pools create gpu-pool \
--cluster=$CLUSTER_NAME \
--zone=$LOCATION \
--machine-type=$MACHINE_TYPE \
--accelerator type=$GPU_TYPE,count=$GPU_COUNT,gpu-driver-version=default \
--num-nodes=1 \
--enable-autoscaling \
--min-nodes=1 \
--max-nodes=1 \
--spot \
--disk-type=pd-ssd \
--disk-size=100GB
```

5. Configure kubectl to use the new cluster:

```bash
gcloud container clusters get-credentials $CLUSTER_NAME --zone=$LOCATION
```

Your GKE cluster with GPU support is now ready for HUGS deployment. You can proceed to deploy HUGS using the Google Cloud Marketplace as described in the previous section.

<Tip>

For more detailed information on creating and managing GKE clusters, refer to the [official Google Kubernetes Engine documentation](https://cloud.google.com/kubernetes-engine/docs) or [run GPUs in GKE Standard node pools](https://cloud.google.com/kubernetes-engine/docs/how-to/gpus).

</Tip>
2 changes: 1 addition & 1 deletion docs/source/index.mdx
Original file line number Diff line number Diff line change
Expand Up @@ -42,7 +42,7 @@ For detailed instructions on deployment and usage:
- [Hugging Face Enterprise](https://huggingface.co/enterprise)
- [Amazon Web Services (AWS)](./how-to/cloud/aws)
- [DigitalOcean](./how-to/cloud/digital-ocean)
- [Google Cloud Platform (GCP)](./how-to/cloud/gcp) (coming soon)
- [Google Cloud Platform (GCP)](./how-to/cloud/gcp)
- [Microsoft Azure](./how-to/cloud/azure) (coming soon)

## More Resources
Expand Down
2 changes: 1 addition & 1 deletion docs/source/pricing.mdx
Original file line number Diff line number Diff line change
Expand Up @@ -7,7 +7,7 @@ HUGS (Hugging Face Generative AI Services) offers a on-demand pricing based on t
For deployments on major cloud platforms, HUGS is available through their respective marketplaces:

- **AWS Marketplace**: $1 per hour per container
- (Soon) **Google Cloud Platform (GCP) Marketplace**: $1 per hour per container
- **Google Cloud Platform (GCP) Marketplace**: $1 per hour per container
- (Soon) **Microsoft Azure**

This pricing model is based on the uptime of each container, allowing you to scale your usage according to your needs.
Expand Down

0 comments on commit 092e96e

Please sign in to comment.