Private Cloud / On-Prem Installation Using Open-Weight AI Models

Tabnine AI code assistant: System requirements

You can also power Tabnine by supporting open-weight models that are installed on-premises or on one of the private clouds mentioned above.

Models

For Self-Hosted (SH) customers, your hardware needs depend on whether or not you already have any open-weight models within your infrastructure.

Tabnine-Supported Open-Weight Models

Devstral-Small-2-24B-Instruct-2512

Devstral-2-123B-Instruct-2512**

MiniMax-M2.5

GLM-4.7

Qwen-3-Coder-480B-A35B-Instruct

Qwen-3-30B (Chat only)

If not, we will install one of the following models on-premises for you:

Open-Weight Models that Tabnine Offers to Install On-Prem
circle-info

**Devstral 2 (123B parameters) is operating under a modified MIT license. If your organization's global consolidated monthly revenue is exceeding $20 million, utilizing this model requires Devstral's permission.Hardware Requirements.

Hardware Requirements

There are different installation requirements, aimed to make sure users have the optimal experience when using Tabnine. Those requirements will be different for Agentic workflows or Chat.

Agent + Chat

Agent + Chat
≤100 Users — Recommended
≤100 Users — Minimal
101-500 Users — Recommended
101-500 Users — Minimal
501-1000 Users — Recommended
501-1000 Users — Minimal
1001-2000 Users — Recommended
1001-2000 Users — Minimal

Devstral-Small-2-24B-Instruct-2512

2 B200

2 H100

2 B200

3 H100

4 B200

6 H100

8 B200

12 H100

Devstral-2-123B-Instruct-2512

4 B200

4 H100

8 B200

8 H100

16 B200

8 B200

24 B200

16 B200

MiniMax-M2.5

2 B200

2 H200

4 B200

4 H200

8 B200

8 H200

16 B200

16 H200

GLM-4.7

2 B200

8 H100

4 B200

2 B200

8 B200

4 B200

16 B200

8 B200

Qwen-3-Coder-480B-A35B-Instruct

2 B200

8 H100

4 B200

2 B200

8 B200

4 B200

16 B200

8 B200

Chat Only

Chat Only
≤100 Users — Recommended
≤100 Users — Minimal
101-500 Users — Recommended
101-500 Users — Minimal
501-1000 Users — Recommended
501-1000 Users — Minimal
1001-2000 Users — Recommended
1001-2000 Users — Minimal

Devstral-Small-2-24B-Instruct-2512

2 B200

2 H100

2 B200

2 H100

2 B200

2 H100

2 B200

4 H100

Devstral-2-123B-Instruct-2512

2 B200

4 H100

2 B200

4 H100

4 B200

8 H100

8 B200

16 H100

MiniMax-M2.5

2 B200

2 H 200

2 B200

2 H200 /

4 H100

2 B200

4 H200 /

8 H100

3 B200

8 H200

GLM-4.7

2 B200

8 H100

2 B200

2 B200

4 B200

2 B200

6 B200

4 B200

Qwen-3-Coder-480B-A35B-Instruct

2 B200

8 H100

2 B200

8 H100

4 B200

4 B200

8 B200

8 B200

Qwen-3-30B

2 B200

2 H100

2 B200

2 H100

2 B200

2 H100

2 B200

2 H100

GPU Availability by Cloud Provider

GPU
AWS
Azure
GCP

H100

p5.4xlarge (H100 80GB)

NC40ads_H100_v5 (H100 94GB)

a3-highgpu-1g (H100 80GB)

H200

p5en.48xlarge (8×H200 141GB)

ND96isr_H200_v5 (8×H200 141GB)

a3-ultragpu-8g (8×H200 141GB)

B200

p6-b200.48xlarge (8×B200 HBM3e)

ND128isr_NDR_GB200_v6 (4×Blackwell 192GB)

a4-highgpu-8g (8×B200 HBM3e)

circle-info

If you don’t have an open-weight model that is not on the list, contact us and our team will work with you.

Open-Weight Model Installation

Devstral-2-123B-Instruct-2512

Execution Script:

Devstral-Small-2-24B-Instruct-2512

Execution Script:

MiniMax-M2.5

Automated Startup Script:

Verification Commands

Last updated

Was this helpful?