Run serverless GPUs on

Run serverless GPUs on

Run serverless GPUs on

your

your

Deploy and auto-scale generative AI models on your own infra. Pay for what you use, no idle costs.

Deploy and auto-scale generative AI models on your own infra. Pay for what you use, no idle costs.

Deploy and auto-scale generative AI models on your own infra. Pay for what you use, no idle costs.

Trusted by

  • The Forecasting Company

    T

    F

    C

  • Lumina

  • Haystack

  • The Forecasting Company

    T

    F

    C

  • Lumina

  • Haystack

  • The Forecasting Company

    T

    F

    C

  • Lumina

  • Haystack

  • The Forecasting Company

    T

    F

    C

  • Lumina

  • Haystack

  • The Forecasting Company

    T

    F

    C

  • Lumina

  • Haystack

Ship fast.

Ship fast.

Leave heavy lifting to us.

Leave heavy lifting to us.

Connect

Connect your cloud account (AWS, GCP or Azure) and Tensorfuse will automatically provision the resources to manage your infra.

Deploy

Deploy ML models to your own cloud via the Tensorfuse SDK.

Data never leaves your cloud and you can start using an OpenAI compatible API.

import tensorkube


image = tensorkube.Image.from_registry(

"nvidia/cuda" ).add_python(version='3.9')

.apt_install([ 'git','git-lfs' ])

.pip_install([ 'transformers', 'torch', 'torchvision', 'tensorrt', ])

.env( { 'SOME-RANDOM-SECRET-KEY': 'xxx-xyz-1234-abc-5678', } )

.run_custom_function( download_and_quantize_model, )


@tensorkube.entrypoint(image, gpu = 'A10G')

def load_model_on_gpu():

import transformers

model = transformers.BertModel.from_pretrained('bert-base-uncased')

model.to('cuda')

tensorkube.pass_reference(model, 'model')


@tensorkube.function(image)

def infer(input: str):

model = tensorkube.get_reference('model')

# test the model on input

response = model(input)

return response



Scale

Tensorfuse automatically scales in response to the amount of traffic your app receives.

Fast cold boots with our optimized container system

Ease and speed of serverless.

Flexibility and control of your own infra.

Ease and speed of serverless.

Flexibility and control of your own infra.

Customize your environment

Describe container images and hardware specifications in simple Python. No YAML

import tensorkube


image = tensorkube.Image.from_registry(

"nvidia/cuda" ).add_python(version='3.9')

.apt_install([ 'git','git-lfs' ])

.pip_install([ 'transformers', 'torch', 'torchvision', 'tensorrt', ])

.env( { 'SOME-RANDOM-SECRET-KEY': 'xxx-xyz-1234-abc-5678', } )

.run_custom_function( download_and_quantize_model, )


@tensorkube.use_image(image)

def infer():

print('Your inference code goes Here!')

Private by default

Your model and data live within your private cloud


Scale at will

Meet user demand in real time by scaling GPU workers from zero to hundreds in seconds.

Cost effective

Reduce egress charges by using model inference within your cloud environment.


OpenAI compatible

Start using your deployment on an OpenAI compatible endpoint.

Compute utilization

Easily utilize compute resources across multiple cloud providers.

Pricing for every team's size

Bill monthly

Bill annually (15% off)

Hacker

Free

100 MGHs

1 seat included

Community support

Starter

$599 per month

5k MGHs

$0.1 per MGH after that

5 seats included

Access to managed dashboard

Logging and monitoring

Support via private slack

14 days free trial

Growth

$999 per month

10k MGHs

$0.1 per MGH after that

10 seats included

Access to managed

dashboard

Access to dev containers

Logging and monitoring

Support via private slack

14 days free trial

Recommended

Enterprise

Custom

Custom MGHs

Volume discount

Everything in Growth. plus

Custom seats

Single sign-on

Security (SOC2)

14 days free trial

Bill monthly

Bill annually (15% off)

Hacker

Free

100 MGHs

1 seat included

Community support

Starter

$599 per month

5k MGHs

$0.1 per MGH after that

5 seats included

Access to managed dashboard

Logging and monitoring

Support via private slack

14 days free trial

Growth

$999 per month

10k MGHs

$0.1 per MGH after that

10 seats included

Access to managed dashboard

Access to dev containers

Logging and monitoring

Support via private slack

14 days free trial

Recommended

Enterprise

Custom

Custom MGHs

Volume discount

Everything in Growth. plus

Custom seats

Single sign-on

Security (SOC2)

14 days free trial

Bill monthly

Bill annually (15% off)

Hacker

Free

100 MGHs

1 seat included

Community support

Starter

$599 per month

5k MGHs

$0.1 per MGH after that

5 seats included

Access to managed dashboard

Logging and monitoring

Support via private slack

14 days free trial

Growth

$999 per month

10k MGHs

$0.1 per MGH after that

10 seats included

Access to managed dashboard

Access to dev containers

Logging and monitoring

Support via private slack

14 days free trial

Recommended

Enterprise

Custom

Custom MGHs

Volume discount

Everything in Growth. plus

Custom seats

Single sign-on

Security (SOC2)

14 days free trial

Bill monthly

Bill annually (15% off)

Hacker

Free

100 MGHs

1 seat included

Community support

Starter

$599 per month

5k MGHs

$0.1 per MGH after that

5 seats included

Access to managed dashboard

Logging and monitoring

Support via private slack

14 days free trial

Growth

$999 per month

10k MGHs

$0.1 per MGH after that

10 seats included

Access to managed dashboard

Access to dev containers

Logging and monitoring

Support via private slack

14 days free trial

Recommended

Enterprise

Custom

Custom MGHs

Volume discount

Everything in Growth. plus

Custom seats

Single sign-on

Security (SOC2)

14 days free trial

Bill monthly

Bill annually (15% off)

Hacker

Free

100 MGHs

1 seat included

Community support

Starter

$599 per month

5k MGHs

$0.1 per MGH after that

5 seats included

Access to managed dashboard

Logging and monitoring

Support via private slack

14 days free trial

Growth

$999 per month

10k MGHs

$0.1 per MGH after that

10 seats included

Access to managed

dashboard

Access to dev containers

Logging and monitoring

Support via private slack

14 days free trial

Recommended

Enterprise

Custom

Custom MGHs

Volume discount

Everything in Growth. plus

Custom seats

Single sign-on

Security (SOC2)

14 days free trial

Get started with Tensorfuse today.

Get started with Tensorfuse today.

Deploy in minutes, scale in seconds.

import tensorkube


image = tensorkube.Image.from_registry(

"nvidia/cuda" ).add_python(version='3.9')

.apt_install([ 'git','git-lfs' ])

.pip_install([ 'transformers', 'torch', 'torchvision', 'tensorrt', ])

.env( { 'SOME-RANDOM-SECRET-KEY': 'xxx-xyz-1234-abc-5678', } )

.run_custom_function( download_and_quantize_model, )


@tensorkube.entrypoint(image, gpu = 'A10G')

def load_model_on_gpu():

import transformers

model = transformers.BertModel.from_pretrained('bert-base-uncased')

model.to('cuda')

tensorkube.pass_reference(model, 'model')


@tensorkube.function(image)

def infer(input: str):

model = tensorkube.get_reference('model')

# test the model on input

response = model(input)

return response



Get started with Tensorfuse today.

Deploy in minutes, scale in seconds.

import tensorkube


image = tensorkube.Image.from_registry(

"nvidia/cuda" ).add_python(version='3.9')

.apt_install([ 'git','git-lfs' ])

.pip_install([ 'transformers', 'torch', 'torchvision', 'tensorrt', ])

.env( { 'SOME-RANDOM-SECRET-KEY': 'xxx-xyz-1234-abc-5678', } )

.run_custom_function( download_and_quantize_model, )


@tensorkube.entrypoint(image, gpu = 'A10G')

def load_model_on_gpu():

import transformers

model = transformers.BertModel.from_pretrained('bert-base-uncased')

model.to('cuda')

tensorkube.pass_reference(model, 'model')


@tensorkube.function(image)

def infer(input: str):

model = tensorkube.get_reference('model')

# test the model on input

response = model(input)

return response



© 2024. All rights reserved.

Privacy Policy