Deploy your jobs and queue them programmatically with different parameters
--name <job-name>
: The name of the job.--gpus <number-of-gpus>
: The number of GPUs required for the job. [Default 0]--gpu-type <gpu-type>
: The type of GPU required.--max-scale <max-scale>
: The maximum scale for the job. [Default 3]--cpu <cpu-units>
: The amount of CPU units required. Used only if GPUs are 0. Specified in milliCPUs [Default 100]--memory <memory-size>
: The amount of memory required. Specified in MB [Default 200]--secret <secret-key>
: The name of the secret required by the job. Can be used multiple times to attach multiple secrets.job-name
matches the job you deployed.
--job-name <job-name>
: The name of the job to be queued.--job-id <job-id>
: The unique identifier for the job.--payload <payload>
: The parameters or data to be passed to the job. Data Type: String.tensorkube
package in your Docker image and add the following snippet to your code.
--job-name <job-name>
: The name of the job to be polled.--job-id <job-id>
: The unique identifier for the job whose status you want to check.hf-transfer
.get_queued_message()
from tensorkube
requirements.txt
/ Dockerfile
file.Dockerfile
if you are deploying your service in a container.
aws configure
and enter your ACCESS_KEY_ID
, SECRET_ACCESS_KEY
, SESSION_TOKEN
(only for Identity Center User) and REGION
values as you are prompted. You can also directly modify your ~/.aws/credentials
file for backend deployments. Read more about configuring your AWS CLI here https://docs.aws.amazon.com/cli/latest/userguide/cli-chap-configure.html