Instance Sizes & ML Credits

Instance sizes enable the simple selection of the best compute and memory resources when building and deploying models.

On this page, you will find detailed information about the different instance sizes available on JFrog ML, helping you choose the optimal instance size to suit your needs.

📘

Note

Please note that as of February 2025, we've updated our data cluster sizes and ML Credits to reflect upgrades to next-gen instances, providing faster runtimes and greater efficiency.

Select an instance size from a wide variety of options

Build & Deploy Models

JFrog ML offers a wide range of instance size to build and deploy models.

📘

Note

Instance configuration for building and deploying models may still be customized individually.

General Purpose Instances

Our general-purpose instances provide varying levels of CPU and memory resources, allowing you to optimize efficiency and performance.

Select the instance size that best matches your requirements from the table below:

Instance IDdisplay_namedisplay_orderCPUsMemory (GB)EnabledGPUcluster_type
promptPrompt10.51true0.125SAAS
tinyTiny21.02true0.25SAAS
smallSmall32.04true0.5SAAS
mediumMedium44.08true1.0SAAS
largeLarge58.016true2.0SAAS
xlargeXLarge616.032true4.0SAAS
2xlarge2XLarge732.064true8.0SAAS
4xlarge4XLarge864.0128true16.0SAAS

GPU Instances

Build and deploy models on GPU-based machines from the selection available in the table below (scroll right for more data):

Instance IDDisplay NameDisplay OrderCPUMemory Amount (GB)GPU AmountGPU TypeAWS SupportedGCP SupportedqpuEnabledCluster TypeAzure Supported
gpu.azure.m60.4xlM60 4XLarge347.04434NVIDIA_M60falsefalse22.8trueSAAStrue
gpu.azure.m60.2xlM60 2XLarge223.02192NVIDIA_M60falsefalse11.4trueSAAStrue
gpu.azure.m60.xlM60 XLarge111.01071NVIDIA_M60falsefalse5.7trueSAAStrue
gpu.azure.a10.xlA10 XLarge1171.08751NVIDIA_A10falsefalse32.6trueSAAStrue
gpu.azure.a10.largeA10 Large1035.04351NVIDIA_A10falsefalse16.0trueSAAStrue
gpu.azure.a10.mediumA10 Medium917.02151NVIDIA_A10falsefalse8.0trueSAAStrue
gpu.azure.a10.smallA10 Small811.01051NVIDIA_A10falsefalse4.54trueSAAStrue
gpu.azure.t4.8xlT4 8XLarge764.04354NVIDIA_T4falsefalse21.76trueSAAStrue
gpu.azure.t4.4xlT4 4XLarge615.01051NVIDIA_T4falsefalse6.02trueSAAStrue
gpu.azure.t4.2xlT4 2XLarge57.0511NVIDIA_T4falsefalse3.76trueSAAStrue
gpu.azure.t4.xlT4 XLarge43.0231NVIDIA_T4falsefalse2.63trueSAAStrue
gpu.a10.12xlA10 12Xlarge1547.01894NVIDIA_A10Gtruefalse28.3600006trueSAASfalse
gpu.a10.2xlA10 2Xlarge27.0281NVIDIA_A10Gtruefalse6.05999994trueSAASfalse
gpu.a10.4xlA10 4Xlarge315.0591NVIDIA_A10Gtruefalse8.11999989trueSAASfalse
gpu.a10.8xlA10 8Xlarge431.01231NVIDIA_A10Gtruefalse12.2399998trueSAASfalse
gpu.a10.xlA10 Xlarge13.0141NVIDIA_A10Gtruefalse5.03000021trueSAASfalse
gpu.a100.8xlA100 8Xlarge895.010728NVIDIA_A100truefalse163.199997trueSAASfalse
gpu.gcp.a100.8xlA100 8Xlarge895.010728NVIDIA_A100_80GB_8_96_1360falsetrue163.199997trueSAASfalse
gpu.gcp.t4.2xlT4 2Xlarge67.0251NVIDIA_T4_1_8_30falsetrue3.31999993trueSAASfalse
gpu.gcp.t4.4xlT4 4Xlarge715.0521NVIDIA_T4_1_16_60falsetrue5.57999992trueSAASfalse
gpu.gcp.t4.xlT4 Xlarge53.0111NVIDIA_T4_1_4_15falsetrue2.19000006trueSAASfalse
gpu.l4.xlL4 Xlarge173.0121NVIDIA_L4truefalse3.52999997trueSAASfalse
gpu.t4.2xlT4 2Xlarge67.0281NVIDIA_T4truefalse3.31999993trueSAASfalse
gpu.t4.4xlT4 4Xlarge715.0591NVIDIA_T4truefalse5.57999992trueSAASfalse
gpu.t4.xlT4 Xlarge53.0141NVIDIA_T4truefalse2.19000006trueSAASfalse
gpu.v100.4xlV100 4Xlarge1031.02274NVIDIA_V100truefalse63.5999985trueSAASfalse
gpu.v100.8xlV100 8Xlarge1163.04548NVIDIA_V100truefalse127.199997trueSAASfalse
gpu.a100.xlA100 Xlarge1610.0751NVIDIA_A100truefalse15.8999996trueSAASfalse
gpu.v100.xlV100 Xlarge97.0531NVIDIA_V100truefalse15.8999996trueSAASfalse
gpu.gcp.v100.xlV100 Xlarge97.0521NVIDIA_V100_1_8_52falsetrue15.8999996trueSAASfalse
gpu.gcp.v100.4xlV100 4Xlarge1031.02084NVIDIA_V100_4_32_208falsetrue63.5999985trueSAASfalse
📘

Note

Instance specifications are based on AWS standards. Actual resource allocation may vary slightly depending on your cloud provider (AWS, GCP, or Azure), but will consistently meet the performance tier requirements.

Feature Store

Data Cluster Sizes

Our Feature Store offers a variety of data cluster sizes to accommodate your needs. Select the appropriate size to ensure scalability and efficiency in handling your data ingestion jobs.

The table below explores the available data cluster sizes:

SizeML Credits (per hour)Notes
Nano4Available for Streaming features
Small8
Medium15
Large30
X-Large60
2X-Large120

Instance Sizes in flogml-cli

Using the frogml-cli provides you with flexibility in choosing instance sizes for building and deploying models.

See the examples below to understand how to specify the required instance size.

Build Models on CPU Instances

frogml models build --model-id "example-model-id" --instance medium .

Build Models on GPU Instances

frogml models build --model-id "example-model-id" --instance "gpu.t4.xl" .

Deploy Models on CPU Instances

frogml models deploy realtime --model-id "example-model-id" --instance large

Deploy Models on GPU Instances

frogml models deploy realtime --model-id "example-model-id" --instance "gpu.a10.4xl"
📘

Note

Existing resource configuration flags are also supported: --memory, --cpus, --gpu-type, --gpu-amount.

Instance Sizes in the UI

In the JFrog ML UI, you can easily select and configure instance sizes for your models. Whether you need CPU or GPU instances, the JFrog ML UI offers intuitive options to choose the correct size for your workload.

During the deployment process, use the dropdown to specify the instance size for optimal performance.

The instance size dropdown offers a wide selection of available instances

Setting Custom Configuration

JFrog ML enables you to manually set custom instance configuration sizes for building and deploying your models, regardless of the default instance type options.

Custom instance type configuration is currently available for CPU deployments only.

Set custom instance configuration for CPU deployments