Deploy serverless GPU applications on your AWS account
3.7
to 3.11
. Make sure that your virtual environment is set up with one of these versions.I consent, Let's Deploy
button.
login
button provided. Please note that you will be redirected to the AWS console. After you are done with the login and can see your AWS console, click on the Continue
button.
AWS account ID
. You can find that in the top right corner of your AWS console.
Grant Permissions
button.
Create Stack
button.
Create Stack
button.
After this, we will automatically create some granular roles with minimal access to create and manage resources on your behalf.
TensorkubeAccessStack
and TensorkubeGranularPermissionsStack
permission stacks are created successfully, you will be able to create your cluster.
Create Cluster
button.
3.7
to 3.11
. Make sure that your virtual environment is set up with one of these versions.3.7
to 3.11
. If you use a different version, there’s a chance that things will not work as expected./readiness
endpoint configured in your FastAPI app.
Tensorkube uses this endpoint to check the health of your deployments. Given below is a simple FastAPI app that you can deploy:
env
directories inside the project folder which then interferes with the build process.
If you have other things in the folder, make sure that you make a .dockerignore
file and put the irrelevant stuff in that file.--gpu-type
argument supports all the GPU types that are available on AWS. You can find the list of supported GPU types here.