Instance Sizes & ML Credits

Instance sizes enable the simple selection of the best compute and memory resources when building and deploying models.

On this page, you will find detailed information about the different instance sizes available on JFrog ML, helping you choose the optimal instance size to suit your needs.

📘

Note

Please note that as of February 2025, we've updated our data cluster sizes and ML Credits to reflect upgrades to next-gen instances, providing faster runtimes and greater efficiency.

Build & Deploy Models

JFrog ML offers a wide range of instance size to build and deploy models.

📘

Note

Instance configuration for building and deploying models may still be customized individually.

General Purpose Instances

Our general-purpose instances provide varying levels of CPU and memory resources, allowing you to optimize efficiency and performance.

Select the instance size that best matches your requirements from the table below:

Instance IDNameCPUsMemory (GB)ML Credits
promptPrompt0.510.125
tinyTiny1.020.25
smallSmall2.040.5
mediumMedium4.081.0
largeLarge8.0162.0
xlargeXLarge16.0324.0
2xlarge2XLarge32.0648.0
4xlarge4XLarge64.012816.0

GPU Instances

Build and deploy models on GPU-based machines from the selection available in the table below (scroll right for more data):

Instance IDDisplay NameCPUMemory Amount (GB)GPU AmountGPU TypeAWS SupportedGCP SupportedML CreditsAzure Supported
gpu.azure.m60.4xlM60 4XLarge47.04434NVIDIA_M60falsefalse22.8true
gpu.azure.m60.2xlM60 2XLarge23.02192NVIDIA_M60falsefalse11.4true
gpu.azure.m60.xlM60 XLarge11.01071NVIDIA_M60falsefalse5.7true
gpu.azure.a10.xlA10 XLarge71.08751NVIDIA_A10falsefalse32.6true
gpu.azure.a10.largeA10 Large35.04351NVIDIA_A10falsefalse16.0true
gpu.azure.a10.mediumA10 Medium17.02151NVIDIA_A10falsefalse8.0true
gpu.azure.a10.smallA10 Small11.01051NVIDIA_A10falsefalse4.54true
gpu.azure.t4.8xlT4 8XLarge64.04354NVIDIA_T4falsefalse21.76true
gpu.azure.t4.4xlT4 4XLarge15.01051NVIDIA_T4falsefalse6.02true
gpu.azure.t4.2xlT4 2XLarge7.0511NVIDIA_T4falsefalse3.76true
gpu.azure.t4.xlT4 XLarge3.0231NVIDIA_T4falsefalse2.63true
gpu.a10.12xlA10 12Xlarge47.01894NVIDIA_A10Gtruefalse28.3600006false
gpu.a10.2xlA10 2Xlarge7.0281NVIDIA_A10Gtruefalse6.05999994false
gpu.a10.4xlA10 4Xlarge15.0591NVIDIA_A10Gtruefalse8.11999989false
gpu.a10.8xlA10 8Xlarge31.01231NVIDIA_A10Gtruefalse12.2399998false
gpu.a10.xlA10 Xlarge3.0141NVIDIA_A10Gtruefalse5.03000021false
gpu.a100.8xlA100 8Xlarge95.010728NVIDIA_A100truefalse163.199997false
gpu.gcp.a100.8xlA100 8Xlarge95.010728NVIDIA_A100_80GB_8_96_1360falsetrue163.199997false
gpu.gcp.t4.2xlT4 2Xlarge7.0251VIDIA_T4_1_8_30falsetrue3.31999993false
gpu.gcp.t4.4xlT4 4Xlarge15.0521VIDIA_T4_1_16_60falsetrue5.57999992false
gpu.gcp.t4.xlT4 Xlarge3.0111VIDIA_T4_1_4_15falsetrue2.19000006false
gpu.l4.xlL4 Xlarge3.0121NVIDIA_L4truefalse3.52999997false
gpu.t4.2xlT4 2Xlarge7.0281NVIDIA_T4truefalse3.31999993false
gpu.t4.4xlT4 4Xlarge15.0591NVIDIA_T4truefalse5.57999992false
gpu.t4.xlT4 Xlarge3.0141NVIDIA_T4truefalse2.19000006false
gpu.v100.4xlV100 4Xlarge31.02274NVIDIA_V100truefalse63.5999985false
gpu.v100.8xlV100 8Xlarge63.04548NVIDIA_V100truefalse127.199997false
gpu.a100.xlA100 Xlarge10.0751NVIDIA_A100truefalse15.8999996false
gpu.v100.xlV100 Xlarge7.0531NVIDIA_V100truefalse15.8999996false
gpu.gcp.v100.xlV100 Xlarge7.0521NVIDIA_V100_1_8_52falsetrue15.8999996false
gpu.gcp.v100.4xlV100 4Xlarge31.02084NVIDIA_V100_4_32_208falsetrue63.5999985false
📘

Note

Instance specifications are based on AWS standards. Actual resource allocation may vary slightly depending on your cloud provider (AWS, GCP, or Azure), but will consistently meet the performance tier requirements.

Feature Store

Data Cluster Sizes

Our Feature Store offers a variety of data cluster sizes to accommodate your needs. Select the appropriate size to ensure scalability and efficiency in handling your data ingestion jobs.

The table below explores the available data cluster sizes:

SizeML Credits (per hour)Notes
Nano4Available for Streaming features
Small8
Medium15
Large30
X-Large60
2X-Large120

Instance Sizes in flogml-cli

Using the frogml-cli provides you with flexibility in choosing instance sizes for building and deploying models.

See the examples below to understand how to specify the required instance size.

Build Models on CPU Instances

frogml models build --model-id "example-model-id" --instance medium .

Build Models on GPU Instances

frogml models build --model-id "example-model-id" --instance "gpu.t4.xl" .

Deploy Models on CPU Instances

frogml models deploy realtime --model-id "example-model-id" --instance large

Deploy Models on GPU Instances

frogml models deploy realtime --model-id "example-model-id" --instance "gpu.a10.4xl"
📘

Note

Existing resource configuration flags are also supported: --memory, --cpus, --gpu-type, --gpu-amount.

Instance Sizes in the UI

In the JFrog ML UI, you can easily select and configure instance sizes for your models. Whether you need CPU or GPU instances, the JFrog ML UI offers intuitive options to choose the correct size for your workload.

During the deployment process, use the dropdown to specify the instance size for optimal performance.

Setting Custom Configuration

JFrog ML enables you to manually set custom instance configuration sizes for building and deploying your models, regardless of the default instance type options.

Custom instance type configuration is currently available for CPU deployments only.