Deploy serverless GPU applications on your AWS account
tensorkube deploy --gpus 1 --gpu-type a10g --cpu 2500 --memory 12000 --min-scale 1 --max-scale 10 --env staging --secret aws-secret --secret huggingface-secret
--cpu
is in millicores, 1000 = 1vCPU--memory
is in MB, 1024 = 1GBtensorkube list deployments --all
to get the details of your deployed apps including their names, status and their HTTP endpoints
.
Use tensorkube list deployments --env <env_name>
for apps in an environment.
tensorkube deployment logs <deployment-name> --env <env-name>
to get the logs of a deployment/pod. This will show you the logs of the pod and help you debug any issues that might be present in the pod.
tensorkube deployment ssh <deployment-name> --env <env-name>
to ssh into a pod. This will help you debug any issues that might be present in the pod.
tensorkube deployment describe <deployment-name> --env <env-name>
kubectl get pods -n <env-name>
to get the details of all the running pods across all your deployments in an environment.
kubectl describe pod <pod-name>
to get the status of a pod before it is created. This will show you all the steps that Tensorkube takes to create a pod. It will also give you
the time each of those steps take and whether there is any scope of optimisation or not.