Deploy on Google Cloud Platform¶
You can run Roboflow Inference on machines hosted on Google Cloud Platform (GCP).
This is ideal if you want to benefit from all of the features Inference has to offer but also want to manage your own cloud infrastructure.
Set up a Google Cloud Compute VM¶
To get started, you will need a cloud compute instance running on GCP.
For deploying instances, we recommend SkyPilot, a tool designed to help you set up cloud instances for AI projects.
To get started, run the following command on your own machine:
pip install inference "skypilot[gcp]"
Follow the SkyPilot GCP documentation to authenticate with AWS
Then, run:
inference cloud deploy --provider gcp --compute-type gpu
This will provision a GPU-capable instance in GCP.
The latest version of Roboflow Inference will be automatically installed on the machine.
When the command has run, you should see a message like:
Deployed Roboflow Inference to gcp on gpu, deployment name is ...
To get a list of your deployments: inference status
To delete your deployment: inference undeploy ...
To ssh into the deployed server: ssh ...
The Roboflow Inference Server is running at http://34.66.116.66:9001
You can then use the API endpoint for your server for use in running models.
You can run any model that Inference supports, including object detection, segmentation, classification, and keypoint models that you have available on Roboflow, and foundation models like CLIP, PaliGemma, SAM-2, and more.
Next Steps¶
Once you've decided on a deployment method and have a server running, interfacing with it is easy.
Fill in your api_url
with the IP address of your VM and the port (9001
by default) your server is running on.