Based on the best stability in Korea, it provides the optimal computing resources conveniently and flexibly according to the purpose of use.
This is the multi-page printable view of this section. Click here to print.
Compute
- 1: Virtual Server
- 1.1: Overview
- 1.1.1: Server Type
- 1.1.2: Monitoring Metrics
- 1.1.3: ServiceWatch Metrics
- 1.2: How-to guides
- 1.2.1: Image
- 1.2.2: Keypair
- 1.2.3: Server Group
- 1.2.4: IP Change
- 1.2.5: Linux NTP Setting
- 1.2.6: Setting up RHEL Repo and WKMS
- 1.2.7: Installing ServiceWatch Agent
- 1.3: API Reference
- 1.4: CLI Reference
- 1.5: Release Note
- 2: Virtual Server Auto-Scaling
- 2.1: Overview
- 2.1.1: Monitoring Metrics
- 2.1.2: ServiceWatch 지표
- 2.2: How-to guides
- 2.2.1: Launch Configuration
- 2.2.2: Managing Policies
- 2.2.3: Managing Schedules
- 2.2.4: Managing Notifications
- 2.3: API Reference
- 2.4: CLI Reference
- 2.5: Release Note
- 3: GPU Server
- 3.1: Overview
- 3.1.1: Server Type
- 3.1.2: Monitoring Metrics
- 3.1.3: ServiceWatch Metrics
- 3.2: How-to guides
- 3.2.1: Image Management
- 3.2.2: Using Multi-instance GPU in GPU Server
- 3.2.3: Using NVSwitch on GPU Server
- 3.2.4: Keypair Management
- 3.2.5: ServiceWatch Agent Installation
- 3.3: API Reference
- 3.4: CLI Reference
- 3.5: Release Note
- 4: Bare Metal Server
- 4.1: Overview
- 4.1.1: Server Type
- 4.1.2: Monitoring Metrics
- 4.2: How-to guides
- 4.2.1: Installing ServiceWatch Agent
- 4.2.2: RHEL Repo 설정하기
- 4.3: API Reference
- 4.4: CLI Reference
- 4.5: Release Note
- 5: Multi-node GPU Cluster
- 5.1: Overview
- 5.1.1: Server Type
- 5.1.2: Monitoring Metrics
- 5.2: How-to guides
- 5.2.1: Cluster Fabric Management
- 5.2.2: Installing ServiceWatch Agent
- 5.2.3: Multi-node GPU Cluster Service Scope and Inspection Guide
- 5.3: Release Note
- 6: Cloud Functions
- 6.1: Overview
- 6.1.1: ServiceWatch Metrics
- 6.2: How-to guides
- 6.2.1: Set Trigger
- 6.2.2: AIOS Connect
- 6.2.3: Blueprint Detailed Guide
- 6.2.4: PrivateLink Service Integration
- 6.2.5: 리소스 기반 정책 가이드
- 6.3: Release Note
- 7: Virtual Server DR
- 7.1: Overview
- 7.2: Release Note
- 8: Block Storage
- 8.1: Overview
- 8.1.1: Monitoring Metrics
- 8.2: How-to guides
- 8.2.1: Connecting to the Server
- 8.2.2: Using Snapshots
- 8.2.3: Move Volume
- 8.3: API Reference
- 8.4: CLI Reference
- 8.5: Release Note
1 - Virtual Server
1.1 - Overview
Service Overview
Virtual Server is a virtual server optimized for cloud computing that allows you to freely allocate and use as much as you need at the necessary time without purchasing infrastructure resources such as CPU, Memory, etc. provided by the server individually. You can use resources with optimized performance according to your computing usage purposes such as development, testing, and application execution in a cloud environment.
Key Features
Easy and convenient computing environment configuration: Through the web-based Console, users can easily use Self Service from Virtual Server provisioning to resource management and cost management. If you need to change the capacity of major resources such as CPU or Memory while using Virtual Server, you can easily expand or reduce without operator intervention.
Provision of various types of services: Provides virtualized vCore/Memory resources according to predefined server types (1~128 vCore).
- General Virtual Server: Provides Computing Specs generally used (maximum 16vCore, 256GB)
- High Capacity Virtual Server: Provides when large capacity resources larger than General Virtual Server Spec are needed
Strong security application: Protects servers safely by controlling Inbound/Outbound traffic communicating with external internet or other VPC (Virtual Private Cloud) through Security Group service. In addition, you can stably operate computing resources through real-time monitoring.
Service Architecture
Provided Functions
Virtual Server provides the following functions.
- Automatic Provisioning and Management: Provides functions from Virtual Server provisioning to resource management and cost management through web-based Console. If you need to change the capacity of major resources such as CPU or Memory while using Virtual Server, you can change immediately using the server type modification function.
- Standard server type and Image provision: Provides virtualized vCore/Memory resources according to standard server types, and provides standard OS Image.
- Storage connection: Provides additional connection storage in addition to OS disk. You can use by additionally connecting Block Storage, File Storage, and Object Storage.
- Network connection: You can connect Virtual Server’s general subnet/IP and Public NAT IP. Provides local subnet connection for communication between servers. This task can be modified on the detail page.
- Security Group application: Protects servers safely by controlling Inbound/Outbound traffic communicating with external internet or other VPC through Security Group service.
- Monitoring: You can check monitoring information such as CPU, Memory, Disk, etc. corresponding to computing resources through Cloud Monitoring service.
- Backup and recovery: You can backup and recover Virtual Server Image through Backup service.
- Cost management: You can create, stop, and terminate servers as needed, and since billing is based on actual usage time, you can check costs according to usage.
- ServiceWatch service integration provision: You can monitor data through ServiceWatch service.
Components
Virtual Server provides standard server types and standard OS Image. Users can select and use them according to the desired service scale.
Image
You can create and manage Image. Main functions are as follows.
- Image creation: You can create the configuration of the Virtual Server in use as an Image, and you can create an Image by uploading the user’s Image file to Object Storage.
- Shared Image creation: You can create a Shared Image that can be shared from an Image with Private Visibility.
- Share to other Account: You can share an Image to another Account.
- For Image creation and usage methods, refer to How-to guides > Image document.
Keypair
To provide more secure OS access, we strengthen security by providing Key Pair instead of ID/Password input method. Main functions are as follows.
- Keypair creation: Creates user authentication for connecting to Virtual Server.
- Import public key: You can import a public key by file upload or direct input of public key.
- For Keypair creation and usage methods, refer to How-to guides > Keypair document.
Server Group
Through Server Group settings, you can place Virtual Server and Block Storage added when creating Virtual Server close to or distributed across racks and hosts. Main functions are as follows.
- Server Group creation: You can set Virtual Servers belonging to the same Server Group as Anti-Affinity (distributed placement), Affinity (close placement), Partition (Virtual Server and Block Storage distributed placement).
- For Server Group creation and usage methods, refer to How-to guides > Server Group document.
OS Image Provided Versions
The OS Images provided by Virtual Server are as follows
| OS Image Version | EoS Date |
|---|---|
| Alma Linux 8.10 | 2029-05-31 |
| Alma Linux 9.6 | 2025-11-17 |
| Oracle Linux 8.10 | 2029-07-31 |
| Oracle Linux 9.6 | 2025-11-25 |
| RHEL 8.10 | 2029-05-31 |
| RHEL 9.4 | 2026-04-30 |
| RHEL 9.6 | 2027-05-31 |
| Rocky Linux 8.10 | 2029-05-31 |
| Rocky Linux 9.6 | 2025-11-30 |
| Ubuntu 22.04 | 2027-06-30 |
| Ubuntu 24.04 | 2029-06-30 |
| Windows 2019 | 2029-01-09 |
| Windows 2022 | 2031-10-14 |
| Windows 2016 | 2027-01-12 |
- Linux operating systems such as Alma Linux and Rocky Linux only provide even Minor versions, except for the last release version of the Major version. This is a policy to ensure the stability and consistency of the SCP system. Please check the EOS (End of Support) and EOL (End of Life) dates of the operating system and apply new or additional individual packages as needed to maintain a stable environment.
Server Type
The server types supported by Virtual Server are as follows. For details on server types, refer to Virtual Server Server Type.
Standard s1v2m4
Category | Example | Detailed Description |
|---|---|---|
| Server Type | Standard | Classification of provided server types
|
| Server Spec | s1 | Classification of provided server types and generation
|
| Server Spec | v2 | Number of vCores
|
| Server Spec | m4 | Memory capacity
|
Constraints
- If you create Virtual Server with Rocky Linux or Oracle Linux, additional settings are required for time synchronization (NTP: Network Time Protocol). For other Images, it is automatically set and no separate settings are required.
For details, refer to Configure Linux NTP. - If you create RHEL and Windows Server before August 2025, you need to modify RHEL Repository and WKMS (Windows Key Management Service) settings.
For details, refer to Configure RHEL Repo and WKMS.
Prerequisite Services
This is a list of services that need to be configured in advance before creating this service. For details, please prepare in advance by referring to the guide provided for each service.
| Service Category | Service | Detailed Description |
|---|---|---|
| Networking | VPC | Service that provides independent virtual network in cloud environment |
| Networking | Security Group | Virtual firewall that controls server traffic |
1.1.1 - Server Type
Virtual Server server type
Virtual Server provides a server type suitable for the purpose of use. The server type consists of various combinations such as CPU, Memory, Network Bandwidth, etc. The host server used by the Virtual Server is determined by the server type selected when creating the Virtual Server. Please select a server type according to the specifications of the application you want to run on the Virtual Server.
The server types supported by Virtual Server are as follows.
Standard s1v2m4
Classification | Example | Detailed Description |
|---|---|---|
| Server Type | Standard | Provided server type distinction
|
| Server Specification | s1 | Type of server provided and generation distinction
|
| Server Specification | v2 | Number of vCores
|
| Server Specification | m4 | Memory Capacity
|
s1 server type
The s1 server type of Virtual Server is provided with standard specifications (vCPU, Memory) and is suitable for various applications.
- Samsung Cloud Platform v2’s 1st generation: up to 3.3Ghz Intel 3rd generation (Ice Lake) Xeon Gold 6342 Processor
- Supports up to 16 vCPUs and 256 GB of memory
- Up to 12.5 Gbps networking speed
| Division | Server Type | vCPU | Memory | Network Bandwidth |
|---|---|---|---|---|
| Standard | s1v1m2 | 1 vCore | 2 GB | up to 10 Gbps |
| Standard | s1v2m4 | 2 vCore | 4 GB | Up to 10 Gbps |
| Standard | s1v2m8 | 2 vCore | 8 GB | Up to 10 Gbps |
| Standard | s1v2m16 | 2 vCore | 16 GB | Up to 10 Gbps |
| Standard | s1v2m24 | 2 vCore | 24 GB | Up to 10 Gbps |
| Standard | s1v2m32 | 2 vCore | 32 GB | Up to 10 Gbps |
| Standard | s1v4m8 | 4 vCore | 8 GB | Up to 10 Gbps |
| Standard | s1v4m16 | 4 vCore | 16 GB | Up to 10 Gbps |
| Standard | s1v4m32 | 4 vCore | 32 GB | Up to 10 Gbps |
| Standard | s1v4m48 | 4 vCore | 48 GB | Up to 10 Gbps |
| Standard | s1v4m64 | 4 vCore | 64 GB | Up to 10 Gbps |
| Standard | s1v6m12 | 6 vCore | 12 GB | Up to 10 Gbps |
| Standard | s1v6m24 | 6 vCore | 24 GB | Up to 10 Gbps |
| Standard | s1v6m48 | 6 vCore | 48 GB | Up to 10 Gbps |
| Standard | s1v6m72 | 6 vCore | 72 GB | Up to 10 Gbps |
| Standard | s1v6m96 | 6 vCore | 96 GB | Up to 10 Gbps |
| Standard | s1v8m16 | 8 vCore | 16 GB | Up to 10 Gbps |
| Standard | s1v8m32 | 8 vCore | 32 GB | Up to 10 Gbps |
| Standard | s1v8m64 | 8 vCore | 64 GB | Up to 10 Gbps |
| Standard | s1v8m96 | 8 vCore | 96 GB | Up to 10 Gbps |
| Standard | s1v8m128 | 8 vCore | 128 GB | Up to 10 Gbps |
| Standard | s1v10m20 | 10 vCore | 20 GB | up to 10 Gbps |
| Standard | s1v10m40 | 10 vCore | 40 GB | Up to 10 Gbps |
| Standard | s1v10m80 | 10 vCore | 80 GB | up to 10 Gbps |
| Standard | s1v10m120 | 10 vCore | 120 GB | up to 10 Gbps |
| Standard | s1v10m160 | 10 vCore | 160 GB | Up to 10 Gbps |
| Standard | s1v12m24 | 12 vCore | 24 GB | Up to 12.5 Gbps |
| Standard | s1v12m48 | 12 vCore | 48 GB | Up to 12.5 Gbps |
| Standard | s1v12m96 | 12 vCore | 96 GB | Up to 12.5 Gbps |
| Standard | s1v12m144 | 12 vCore | 144 GB | Up to 12.5 Gbps |
| Standard | s1v12m192 | 12 vCore | 192 GB | up to 12.5 Gbps |
| Standard | s1v14m28 | 14 vCore | 28 GB | Up to 12.5 Gbps |
| Standard | s1v14m56 | 14 vCore | 56 GB | Up to 12.5 Gbps |
| Standard | s1v14m112 | 14 vCore | 112 GB | Up to 12.5 Gbps |
| Standard | s1v14m168 | 14 vCore | 168 GB | Up to 12.5 Gbps |
| Standard | s1v14m224 | 14 vCore | 224 GB | Up to 12.5 Gbps |
| Standard | s1v16m32 | 16 vCore | 32 GB | up to 12.5 Gbps |
| Standard | s1v16m64 | 16 vCore | 64 GB | Up to 12.5 Gbps |
| Standard | s1v16m128 | 16 vCore | 128 GB | up to 12.5 Gbps |
| Standard | s1v16m192 | 16 vCore | 192 GB | Up to 12.5 Gbps |
| Standard | s1v16m256 | 16 vCore | 256 GB | up to 12.5 Gbps |
S2 Server Type
Virtual Server s2 server type is provided with standard specifications (vCPU, Memory) and is suitable for various applications.
- Samsung Cloud Platform v2’s 2nd generation: up to 3.2GHz Intel 4th generation (Sapphire Rapids) Xeon Gold 6448H Processor
- Supports up to 16 vCPUs and 256 GB of memory
- Up to 12.5 Gbps networking speed
| Classification | Server Type | CPU vCore | Memory | Network Bandwidth(Gbps) |
|---|---|---|---|---|
| Standard | s2v1m2 | 1 vCore | 2 GB | Up to 10 Gbps |
| Standard | s2v2m4 | 2 vCore | 4 GB | Up to 10 Gbps |
| Standard | s2v2m8 | 2 vCore | 8 GB | Up to 10 Gbps |
| Standard | s2v2m16 | 2 vCore | 16 GB | Up to 10 Gbps |
| Standard | s2v2m24 | 2 vCore | 24 GB | Up to 10 Gbps |
| Standard | s2v2m32 | 2 vCore | 32 GB | Up to 10 Gbps |
| Standard | s2v4m8 | 4 vCore | 8 GB | Up to 10 Gbps |
| Standard | s2v4m16 | 4 vCore | 16 GB | Up to 10 Gbps |
| Standard | s2v4m32 | 4 vCore | 32 GB | Up to 10 Gbps |
| Standard | s2v4m48 | 4 vCore | 48 GB | Up to 10 Gbps |
| Standard | s2v4m64 | 4 vCore | 64 GB | Up to 10 Gbps |
| Standard | s2v6m12 | 6 vCore | 12 GB | Up to 10 Gbps |
| Standard | s2v6m24 | 6 vCore | 24 GB | Up to 10 Gbps |
| Standard | s2v6m48 | 6 vCore | 48 GB | Up to 10 Gbps |
| Standard | s2v6m72 | 6 vCore | 72 GB | Up to 10 Gbps |
| Standard | s2v6m96 | 6 vCore | 96 GB | Up to 10 Gbps |
| Standard | s2v8m16 | 8 vCore | 16 GB | Up to 10 Gbps |
| Standard | s2v8m32 | 8 vCore | 32 GB | Up to 10 Gbps |
| Standard | s2v8m64 | 8 vCore | 64 GB | Up to 10 Gbps |
| Standard | s2v8m96 | 8 vCore | 96 GB | Up to 10 Gbps |
| Standard | s2v8m128 | 8 vCore | 128 GB | Up to 10 Gbps |
| Standard | s2v10m20 | 10 vCore | 20 GB | Up to 10 Gbps |
| Standard | s2v10m40 | 10 vCore | 40 GB | Up to 10 Gbps |
| Standard | s2v10m80 | 10 vCore | 80 GB | Up to 10 Gbps |
| Standard | s2v10m120 | 10 vCore | 120 GB | Up to 10 Gbps |
| Standard | s2v10m160 | 10 vCore | 160 GB | Up to 10 Gbps |
| Standard | s2v12m24 | 12 vCore | 24 GB | Up to 12.5 Gbps |
| Standard | s2v12m48 | 12 vCore | 48 GB | Up to 12.5 Gbps |
| Standard | s2v12m96 | 12 vCore | 96 GB | Up to 12.5 Gbps |
| Standard | s2v12m144 | 12 vCore | 144 GB | Up to 12.5 Gbps |
| Standard | s2v12m192 | 12 vCore | 192 GB | Up to 12.5 Gbps |
| Standard | s2v14m28 | 14 vCore | 28 GB | Up to 12.5 Gbps |
| Standard | s2v14m56 | 14 vCore | 56 GB | Up to 12.5 Gbps |
| Standard | s2v14m112 | 14 vCore | 112 GB | Up to 12.5 Gbps |
| Standard | s2v14m168 | 14 vCore | 168 GB | up to 12.5 Gbps |
| Standard | s2v14m224 | 14 vCore | 224 GB | Up to 12.5 Gbps |
| Standard | s2v16m32 | 16 vCore | 32 GB | up to 12.5 Gbps |
| Standard | s2v16m64 | 16 vCore | 64 GB | Up to 12.5 Gbps |
| Standard | s2v16m128 | 16 vCore | 128 GB | up to 12.5 Gbps |
| Standard | s2v16m192 | 16 vCore | 192 GB | up to 12.5 Gbps |
| Standard | s2v16m256 | 16 vCore | 256 GB | up to 12.5 Gbps |
h2 Server Type
The h2 server type of Virtual Server is provided with large-capacity server specifications and is suitable for applications for large-scale data processing.
- Samsung Cloud Platform v2’s 2nd generation: up to 3.2GHz Intel 4th generation (Sapphire Rapids) Xeon Gold 6448H Processor Supports up to 128 vCPUs and 1,536 GB of memory
- up to 25Gbps networking speed
| Division | Server Type | vCPU | Memory | Network Bandwidth |
|---|---|---|---|---|
| High Capacity | h2v24m48 | 24 vCore | 48 GB | Up to 25 Gbps |
| High Capacity | h2v24m96 | 24 vCore | 96 GB | Up to 25 Gbps |
| High Capacity | h2v24m192 | 24 vCore | 192 GB | Up to 25 Gbps |
| High Capacity | h2v24m288 | 24 vCore | 288 GB | Up to 25 Gbps |
| High Capacity | h2v32m64 | 32 vCore | 64 GB | Up to 25 Gbps |
| High Capacity | h2v32m128 | 32 vCore | 128 GB | Up to 25 Gbps |
| High Capacity | h2v32m256 | 32 vCore | 256 GB | Up to 25 Gbps |
| High Capacity | h2v32m384 | 32 vCore | 384 GB | Up to 25 Gbps |
| High Capacity | h2v48m96 | 48 vCore | 96 GB | up to 25 Gbps |
| High Capacity | h2v48m192 | 48 vCore | 192 GB | Up to 25 Gbps |
| High Capacity | h2v48m384 | 48 vCore | 384 GB | Up to 25 Gbps |
| High Capacity | h2v48m576 | 48 vCore | 576 GB | Up to 25 Gbps |
| High Capacity | h2v64m128 | 64 vCore | 128 GB | Up to 25 Gbps |
| High Capacity | h2v64m256 | 64 vCore | 256 GB | Up to 25 Gbps |
| High Capacity | h2v64m512 | 64 vCore | 512 GB | Up to 25 Gbps |
| High Capacity | h2v64m768 | 64 vCore | 768 GB | Up to 25 Gbps |
| High Capacity | h2v72m144 | 72 vCore | 144 GB | Up to 25 Gbps |
| High Capacity | h2v72m288 | 72 vCore | 288 GB | Up to 25 Gbps |
| High Capacity | h2v72m576 | 72 vCore | 576 GB | Up to 25 Gbps |
| High Capacity | h2v72m864 | 72 vCore | 864 GB | Up to 25 Gbps |
| High Capacity | h2v96m192 | 96 vCore | 192 GB | Up to 25 Gbps |
| High Capacity | h2v96m384 | 96 vCore | 384 GB | Up to 25 Gbps |
| High Capacity | h2v96m768 | 96 vCore | 768 GB | Up to 25 Gbps |
| High Capacity | h2v96m1152 | 96 vCore | 1152 GB | Up to 25 Gbps |
| High Capacity | h2v128m256 | 128 vCore | 256 GB | Up to 25 Gbps |
| High Capacity | h2v128m512 | 128 vCore | 512 GB | Up to 25 Gbps |
| High Capacity | h2v128m1024 | 128 vCore | 1024 GB | Up to 25 Gbps |
| High Capacity | h2v128m1536 | 128 vCore | 1536 GB | Up to 25 Gbps |
1.1.2 - Monitoring Metrics
Virtual Server monitoring metrics
The following table shows the monitoring metrics of Virtual Server that can be checked through Cloud Monitoring. For more information on how to use Cloud Monitoring, please refer to the Cloud Monitoring guide.
You can get basic monitoring metrics without installing Agent, and please check the metrics below in Table. Virtual Server Monitoring Metrics (Basic). In addition, you can check the metrics that can be retrieved by installing Agent in Table. Virtual Server Additional Monitoring Metrics (Agent Installation Required).
For Windows OS, memory-related metrics can only be retrieved if the Agent is installed.
| Performance Item | Detailed Description | Unit |
|---|---|---|
| Memory Total [Basic] | Available memory bytes | bytes |
| Memory Used [Basic] | Currently used memory bytes | bytes |
| Memory Swap In [Basic] | Swapped memory bytes | bytes |
| Memory Swap Out [Basic] | bytes of swapped memory | bytes |
| Memory Free [Basic] | Unused memory bytes | bytes |
| Disk Read Bytes [Basic] | Read bytes | bytes |
| Disk Read Requests [Basic] | Number of Read Requests | cnt |
| Disk Write Bytes [Basic] | Write bytes | bytes |
| Disk Write Requests [Basic] | Number of Write Requests | cnt |
| CPU Usage [Basic] | 1-minute average system CPU usage rate | % |
| Instance State [Basic] | Instance Status | state |
| Network In Bytes [Basic] | Received bytes | bytes |
| Network In Dropped [Basic] | Receive Packet Drop | cnt |
| Network In Packets [Basic] | Received Packet Count | cnt |
| Network Out Bytes [Basic] | Transmission bytes | bytes |
| Network Out Dropped [Basic] | Transmission Packet Drop | cnt |
| Network Out Packets [Basic] | Transmission packet count | cnt |
| Performance Item | Detailed Description | Unit |
|---|---|---|
| Core Usage [IO Wait] | The ratio of CPU time spent in waiting state (disk waiting) | % |
| Core Usage [System] | The ratio of CPU time spent in kernel space | % |
| Core Usage [User] | The ratio of CPU time spent in user space | % |
| CPU Cores | The number of CPU cores on the host | cnt |
| CPU Usage [Active] | Idle and IOWait status excluding the percentage of CPU time used | % |
| CPU Usage [Idle] | The ratio of CPU time spent in idle state. | % |
| CPU Usage [IO Wait] | the ratio of CPU time spent in a waiting state (disk waiting) | % |
| CPU Usage [System] | The percentage of CPU time used by the kernel | % |
| CPU Usage [User] | The percentage of CPU time used in the user area | % |
| CPU Usage/Core [Active] | Idle and IOWait status excluding the percentage of CPU time used | % |
| CPU Usage/Core [Idle] | The ratio of CPU time spent in idle state. | % |
| CPU Usage/Core [IO Wait] | the ratio of CPU time spent in waiting state (disk waiting) | % |
| CPU Usage/Core [System] | The percentage of CPU time used by the kernel | % |
| CPU Usage/Core [User] | The percentage of CPU time used in the user area | % |
| DiskCPU Usage [IO Request] | The ratio of CPU time spent executing input/output requests for the device | % |
| Disk Queue Size [Avg] | The average queue length of requests executed for the device. | num |
| Disk Read Bytes | The number of bytes read from the device per second. | bytes |
| Disk Read Bytes [Delta Avg] | Average of system.diskio.read.bytes_delta for each Disk | bytes |
| Disk Read Bytes [Delta Max] | Individual Disks’ system.diskio.read.bytes_delta maximum | bytes |
| Disk Read Bytes [Delta Min] | Individual Disks’ system.diskio.read.bytes_delta minimum | bytes |
| Disk Read Bytes [Delta Sum] | Individual Disks’ sum of system.diskio.read.bytes_delta | bytes |
| Disk Read Bytes [Delta] | Individual Disk’s system.diskio.read.bytes value delta | bytes |
| Disk Read Bytes [Success] | Total bytes read successfully | bytes |
| Disk Read Requests | Number of read requests for the disk device during 1 second | cnt |
| Disk Read Requests [Delta Avg] | Individual Disks’ average of system.diskio.read.count_delta | cnt |
| Disk Read Requests [Delta Max] | Individual Disks’ system.diskio.read.count_delta maximum | cnt |
| Disk Read Requests [Delta Min] | Individual Disks’ minimum of system.diskio.read.count_delta | cnt |
| Disk Read Requests [Delta Sum] | The sum of system.diskio.read.count_delta of individual Disks | cnt |
| Disk Read Requests [Success Delta] | Individual Disk’s system.diskio.read.count delta | cnt |
| Disk Read Requests [Success] | Successfully completed total read count | cnt |
| Disk Request Size [Avg] | The average size of requests executed for the device (unit: sector) | num |
| Disk Service Time [Avg] | The average service time (in milliseconds) for the input requests executed on the device. | ms |
| Disk Wait Time [Avg] | The average time spent on requests executed for supported devices. | ms |
| Disk Wait Time [Read] | Disk Average Wait Time | ms |
| Disk Wait Time [Write] | Disk Average Wait Time | ms |
| Disk Write Bytes [Delta Avg] | Individual Disks’ average of system.diskio.write.bytes_delta | bytes |
| Disk Write Bytes [Delta Max] | Individual Disks’ system.diskio.write.bytes_delta maximum | bytes |
| Disk Write Bytes [Delta Min] | Individual Disks’ system.diskio.write.bytes_delta minimum | bytes |
| Disk Write Bytes [Delta Sum] | Individual Disks’ sum of system.diskio.write.bytes_delta | bytes |
| Disk Write Bytes [Delta] | Individual Disk’s system.diskio.write.bytes value delta | bytes |
| Disk Write Bytes [Success] | Total bytes written successfully | bytes |
| Disk Write Requests | Number of write requests to the disk device for 1 second | cnt |
| Disk Write Requests [Delta Avg] | Individual Disks’ average of system.diskio.write.count_delta | cnt |
| Disk Write Requests [Delta Max] | Individual Disks’ system.diskio.write.count_delta maximum | cnt |
| Disk Write Requests [Delta Min] | Minimum of system.diskio.write.count_delta for each Disk | cnt |
| Disk Write Requests [Delta Sum] | Sum of system.diskio.write.count_delta of individual Disks | cnt |
| Disk Write Requests [Success Delta] | Individual Disk’s system.diskio.write.count Delta | cnt |
| Disk Write Requests [Success] | Total number of writes completed successfully | cnt |
| Disk Writes Bytes | The number of bytes written to the device per second | bytes |
| Filesystem Hang Check | filesystem(local/NFS) hang check(Normal:1, Abnormal:0) | status |
| Filesystem Nodes | The total number of file nodes in the file system. | cnt |
| Filesystem Nodes [Free] | The total number of available file nodes in the file system. | cnt |
| Filesystem Size [Available] | Disk space (bytes) available for use by unauthorized users | bytes |
| Filesystem Size [Free] | Available disk space(bytes) | bytes |
| Filesystem Size [Total] | Total disk space (bytes) | bytes |
| Filesystem Usage | Used disk space percentage | % |
| Filesystem Usage [Avg] | Average of individual filesystem.used.pct | % |
| Filesystem Usage [Inode] | iNode usage rate | % |
| Filesystem Usage [Max] | Individual filesystem used percentage Max | % |
| Filesystem Usage [Min] | Individual minimum of filesystem.used.pct | % |
| Filesystem Usage [Total] | - | % |
| Filesystem Used | Used Disk Space (bytes) | bytes |
| Filesystem Used [Inode] | iNode usage | bytes |
| Memory Free | Total available memory amount (bytes) | bytes |
| Memory Free [Actual] | Actually available Memory(bytes) | bytes |
| Memory Free [Swap] | Available Swap memory | bytes |
| Memory Total | Total Memory | bytes |
| Memory Total [Swap] | Total Swap memory. | bytes |
| Memory Usage | used Memory percentage | % |
| Memory Usage [Actual] | Actual used Memory percentage | % |
| Memory Usage [Cache Swap] | cached swap usage rate | % |
| Memory Usage [Swap] | Used Swap memory percentage | % |
| Memory Used | Used Memory | bytes |
| Memory Used [Actual] | Actually used Memory(bytes) | bytes |
| Memory Used [Swap] | Used Swap memory | bytes |
| Collisions | Network Collision | cnt |
| Network In Bytes | Received byte count | bytes |
| Network In Bytes [Delta Avg] | Individual Networks’ average of system.network.in.bytes_delta | bytes |
| Network In Bytes [Delta Max] | Individual Network’s system.network.in.bytes_delta maximum | bytes |
| Network In Bytes [Delta Min] | Individual Networks’ system.network.in.bytes_delta minimum values | bytes |
| Network In Bytes [Delta Sum] | Individual networks’ sum of system.network.in.bytes_delta | bytes |
| Network In Bytes [Delta] | Received byte count delta | bytes |
| Network In Dropped | Number of packets dropped among incoming packets | cnt |
| Network In Errors | Number of errors during reception | cnt |
| Network In Packets | Received packet count | cnt |
| Network In Packets [Delta Avg] | Individual Networks’ average of system.network.in.packets_delta | cnt |
| Network In Packets [Delta Max] | Individual Network’s system.network.in.packets_delta maximum value | cnt |
| Network In Packets [Delta Min] | Individual Network’s system.network.in.packets_delta minimum value | cnt |
| Network In Packets [Delta Sum] | The sum of system.network.in.packets_delta of individual Networks | cnt |
| Network In Packets [Delta] | Received packet count delta | cnt |
| Network Out Bytes | Sent byte count | bytes |
| Network Out Bytes [Delta Avg] | Individual Networks’ average of system.network.out.bytes_delta | bytes |
| Network Out Bytes [Delta Max] | Individual Networks’ system.network.out.bytes_delta maximum | bytes |
| Network Out Bytes [Delta Min] | Individual Networks’ system.network.out.bytes_delta minimum value | bytes |
| Network Out Bytes [Delta Sum] | Individual Network’s system.network.out.bytes_delta sum | bytes |
| Network Out Bytes [Delta] | Sent byte count delta | bytes |
| Network Out Dropped | number of packets dropped among outgoing packets | cnt |
| Network Out Errors | Number of errors during transmission | cnt |
| Network Out Packets | Transmitted packet count | cnt |
| Network Out Packets [Delta Avg] | Average of system.network.out.packets_delta for individual Networks | cnt |
| Network Out Packets [Delta Max] | Individual Networks’ system.network.out.packets_delta maximum values | cnt |
| Network Out Packets [Delta Min] | Individual Network’s system.network.out.packets_delta minimum value | cnt |
| Network Out Packets [Delta Sum] | The sum of system.network.out.packets_delta of individual Networks | cnt |
| Network Out Packets [Delta] | Sent packet count delta | cnt |
| Open Connections [TCP] | All open TCP connections | cnt |
| Open Connections [UDP] | All open UDP connections | cnt |
| Port Usage | Accessible port usage rate | % |
| SYN Sent Sockets | Number of sockets in SYN_SENT state (when connecting from local to remote) | cnt |
| Kernel PID Max | kernel.pid_max value | count |
| Kernel Thread Max | kernel threads maximum value | count |
| Process CPU Usage | The percentage of CPU time consumed by the process after the last update | % |
| Process CPU Usage/Core | The percentage of CPU time used by the process since the last event | % |
| Process Memory Usage | main memory(RAM) where the process occupies a ratio | % |
| Process Memory Used | Resident Set size. The amount of memory a process occupies in RAM | bytes |
| Process PID | Process pid | pid |
| Process PPID | Parent process’s pid | pid |
| Processes [Dead] | number of dead processes | cnt |
| Processes [Idle] | idle Number of Processes | cnt |
| Processes [Running] | running Number of Processes | count |
| Processes [Sleeping] | sleeping processes count | cnt |
| Processes [Stopped] | stopped processes count | cnt |
| Processes [Total] | Total number of processes | cnt |
| Processes [Unknown] | The status cannot be searched or the number of unknown processes | cnt |
| Processes [Zombie] | Number of zombie processes | cnt |
| Running Process Usage | Process Usage Rate | % |
| Running Processes | number of running processes | count |
| Running Thread Usage | thread usage rate | % |
| Running Threads | running processes where the total number of threads being executed | cnt |
| Context Switches | number of context switches (per second) | cnt |
| Load/Core [1 min] | The value divided by the number of cores for the last 1 minute load | cnt |
| Load/Core [15 min] | The value of load divided by the number of cores for the last 15 minutes | cnt |
| Load/Core [5 min] | The value divided by the number of cores for the last 5 minutes | cnt |
| Multipaths [Active] | External storage connection path status = active count | cnt |
| Multipaths [Failed] | External storage connection path status = failed count | cnt |
| Multipaths [Faulty] | External storage connection path status = faulty count | cnt |
| NTP Offset last | sample’s measured offset (time difference between NTP server and local environment) | num |
| Run Queue Length | Execution Waiting Queue Length | num |
| Uptime | OS operation time(uptime) (milliseconds) | ms |
| Context Switchies CPU | number of context switches (per second) | cnt |
| Disk Read Bytes [Sec] | bytes read from the Windows logical disk in 1 second
| cnt |
| Disk Read Time [Avg] | Data Read Average Time (sec)
| sec |
| Disk Transfer Time [Avg] | Disk average wait time (seconds)
| sec |
| Disk Write Bytes [Sec] | The number of bytes written to the Windows logical disk in 1 second
| cnt |
| Disk Write Time [Avg] | Data write average time (seconds)
| sec |
| Pagingfile Usage | Paging file usage rate
| % |
| Pool Used [Non Paged] | _KERNEL MEMORY among Nonpaged Pool usage
| bytes |
| Pool Used [Paged] | Kernel memory Paged Pool usage among kernel memory
| bytes |
| Process [Running] | The number of processes currently running
| cnt |
| Threads [Running] | The number of threads currently running
| cnt |
| Threads [Waiting] | The number of threads waiting for processor time
| cnt |
1.1.3 - ServiceWatch Metrics
Virtual Server sends metrics to ServiceWatch. The metrics provided by basic monitoring are data collected at 5-minute intervals. When detailed monitoring is enabled, you can view data collected at 1-minute intervals.
For information on how to enable detailed monitoring for Virtual Server, refer to How-to guides > Enable ServiceWatch detailed monitoring.
Basic Metrics
The following are basic metrics for the namespace Virtual Server.
In the table below, metrics with metric names marked in bold are selected as key metrics among the basic metrics provided by Virtual Server. Key metrics are used to configure service dashboards that are automatically built for each service in ServiceWatch. You can also check key metrics on the monitoring tab of the Virtual Server detail page.
Each metric guides you on which statistic value is meaningful when querying that metric through the user guide, and the statistic value marked in bold among the meaningful statistics is the key statistic value. In the service dashboard or monitoring tab, you can view key metrics through key statistic values.
| Performance Item (Metric Name) | Detailed Description | Unit | Meaningful Statistics | |
|---|---|---|---|---|
| Instance State | Instance state display
| None |
| |
| CPU Usage | CPU usage | Percent |
| |
| Disk Read Bytes | Amount read from block device (bytes) | Bytes |
| |
| Disk Read Requests | Number of read requests from block device | Count |
| |
| Disk Write Bytes | Amount written to block device (bytes) | Bytes |
| |
| Disk Write Requests | Number of write requests to block device | Count |
| |
| Network In Bytes | Amount received on network interface (bytes) | Bytes |
| |
| Network In Dropped | Number of received packets dropped on network interface | Count |
| |
| Network In Packets | Number of received packets on network interface | Count |
| |
| Network Out Bytes | Amount transmitted on network interface (bytes) | Bytes |
| |
| Network Out Dropped | Number of transmitted packets dropped on network interface | Count |
| |
| Network Out Packets | Number of transmitted packets on network interface | Count |
|
1.2 - How-to guides
Users can create Virtual Server services by entering required information and selecting detailed options through Samsung Cloud Platform Console.
Create Virtual Server
You can create and use Virtual Server service in Samsung Cloud Platform Console.
To create a Virtual Server, follow the procedure below.
Click All Services > Compute > Virtual Server menu. You will be moved to the Service Home page of Virtual Server.
On the Service Home page, click the Create Virtual Server button. You will be moved to the Create Virtual Server page.
On the Create Virtual Server page, enter information required for service creation and select detailed options.
- In the Image and version selection area, select the required information.
Category RequiredDetailed Description Image Required Select the type of Image provided - Standard: Samsung Cloud Platform standard provided Image
- Alma Linux, Oracle Linux, RHEL, Rocky Linux, Ubuntu, Windows
- Custom: User created Image
- Kubernetes: Image for Kubernetes
- RHEL, Ubuntu
- Marketplace: Image subscribed from Marketplace
Image Version Required Select version of selected Image - Provides version list of provided server Image
Table. Virtual Server Image and Version Selection Input Items - Standard: Samsung Cloud Platform standard provided Image
- In the Service information input area, enter or select the required information.
Category RequiredDetailed Description Server Count Required Number of servers to create simultaneously - Only numbers can be entered, enter value between 1 and 100
Service Type > Server Type Required Virtual Server server type - Standard: Standard specs generally used
- High Capacity: Large capacity server specs larger than Standard
- For details on server types provided by Virtual Server, refer to Virtual Server Server Type
Service Type > Planned Compute Required Resource status with Planned Compute set - In Use: Number in use among resources with Planned Compute set
- Set: Number of resources with Planned Compute set
- Coverage Preview: Amount applied as Planned Compute per resource
- Apply for Planned Compute Service: Move to Planned Compute service creation page
- For details, refer to Apply for Planned Compute
Block Storage Required Configure Block Storage used by server according to purpose - Default OS: Area where OS is installed and used
- Capacity is entered in Units, minimum capacity varies depending on OS Image type
- Alma Linux: Enter value between 2 and 1,536
- Oracle Linux: Enter value between 7 and 1,536
- RHEL: Enter value between 2 and 1,536
- Rocky Linux: Enter value between 2 and 1,536
- Ubuntu: Enter value between 2 and 1,536
- Windows: Enter value between 4 and 1,536
- SSD: High performance general volume
- HDD: General volume
- SSD/HDD_KMS: Additional encryption volume using Samsung Cloud Platform KMS (Key Management Service) encryption key
- Encryption application is only possible at initial creation and cannot be changed after creation
- Performance degradation occurs when using SSD_KMS disk type
- SSD_Provisioned: SSD volume with configurable IOPS and Throughput
- Capacity is entered in Units, minimum capacity varies depending on OS Image type
- Additional: Use when additional user space is needed outside OS area
- Select Use and then enter storage type and capacity
- To add storage, click + button, to delete, click x button (can add up to 25)
- Capacity is entered in Units, enter value between 1 and 1,536
- 1 Unit is 8GB, so 8 to 12,288GB is created
- SSD: High performance general volume
- HDD: General volume
- SSD/HDD_KMS: Additional encryption volume using Samsung Cloud Platform KMS (Key Management Service) encryption key
- Encryption application is only possible at initial creation and cannot be changed after creation
- Performance degradation may occur when using SSD_KMS disk type
- SSD/HDD_MultiAttach: Volume that can connect to 2 or more servers
- SSD_Provisioned: SSD volume with configurable IOPS and Throughput
- For details by Block Storage type, refer to Create Block Storage
- Delete on termination: When Delete on Termination is selected, the Volume is also terminated when the server is terminated
- Volumes with snapshots are not deleted even when Delete on termination is Use
- Multi attach volumes are only deleted when the server to be deleted is the last remaining server connected to the volume
- Max IOPS: Enter IOPS maximum value within 5,000 to 20,000
- Cannot set when disk type is HDD, HDD_KMS, HDD_MultiAttach
Server Group Optional Set servers belonging to the same Server Group to Anti-Affinity (distributed placement), Affinity (close placement), Partition (Virtual Server and Block Storage distributed placement) - Select Use and then select Server Group
- Select Create New to create Server Group
- Servers belonging to the same Server Group are placed in Best Effort manner according to selected policy
- Select policy from Anti-Affinity (distributed placement), Affinity (close placement), Partition (Virtual Server and Block Storage distributed placement)
Table. Virtual Server Service Information Input ItemsCautionWhen using Partition (Virtual Server and Block Storage distributed placement) policy among Server Group policies, Block Storage Volume cannot be additionally allocated after Virtual Server creation, so create all necessary Block Storage at the Virtual Server creation stage. - In the Required information input area, enter or select the required information.
Category RequiredDetailed Description Server Name Required Enter name for server identification when selected server count is 1 - Set hostname with entered server name
- Enter within 63 characters using English, numbers, spaces, and special characters (
-,_)
Network Settings > Create New Network Port Required Set network where Virtual Server will be installed - VPC Name: Select pre-created VPC
- General Subnet: Select pre-created general Subnet
- IP can select Auto Create or user input, if Input is selected, user can directly enter IP
- NAT: Can only use when server count is 1 and Internet Gateway is connected to VPC. Check Use to select NAT IP
- NAT IP: Select NAT IP
- If there is no NAT IP to select, click Create New button to create Public IP
- Click Refresh button to check and select created Public IP
- When Public IP is created, fees are charged according to Public IP fee standard
- Local Subnet (Optional): Select Use for Local Subnet
- Not required for service creation
- Must select pre-created Local Subnet
- IP can select Auto Create or user input, if Input is selected, user can directly enter IP
- Security Group: Settings required to connect to server
- Select: Select pre-created Security Group
- Create New: If there is no Security Group to apply, can create separately in Security Group service
- Can select up to 5
- If Security Group is not set, all connections are blocked by default
- Security Group must be set to allow necessary connections
Network Settings > Specify Existing Network Port Required Set network where Virtual Server will be installed - VPC: Select pre-created VPC
- General Subnet: Select pre-created general Subnet and Port
- NAT: Can only use when server count is 1 and Internet Gateway is connected to VPC. Check Use to select NAT IP
- NAT IP: Select NAT IP
- If there is no NAT IP to select, click Create New button to create Public IP
- Click Refresh button to check and select created Public IP
- Local Subnet (Optional): Select Use for Local Subnet
- Select pre-created Local Subnet and Port
Keypair Required User authentication method to use when connecting to the server - Create New: Create a new one when a new Keypair is needed
- For the new Keypair creation method, refer to Create Keypair
- Default login account list by OS
- Alma Linux: almalinux
- Oracle Linux: cloud-user
- RHEL: cloud-user
- Rocky Linux: rocky
- Ubuntu: ubuntu
- Windows: sysadmin
</ul
Table. Virtual Server Required Information Input Items - In the Additional information input area, enter or select the required information.
Category RequiredDetailed Description Lock Optional Set whether to use Lock - When Lock is used, prevents operations such as server termination, start, stop to prevent accidental malfunction
Init Script Optional Script to execute when server starts - Init script must be written as Batch script for Windows or Shell script or cloud-init for Linux depending on Image type.
- Can enter up to 45,000 bytes
Tag Optional Add tags - Can add up to 50 per resource
- Click Add Tag button and then enter or select Key, Value values
Table. Virtual Server Additional Information Input Items
- In the Image and version selection area, select the required information.
On the Summary panel, check the created detailed information and estimated billing amount, then click the Create button.
- When creation is complete, check the created resource on the Virtual Server List page.
- When entering server name, if spaces and special characters (
_) are used, OS hostname is set with spaces and special characters (_) changed to special character (-). Please note when setting OS hostname.- Example: If server name is ‘server name_01’, OS hostname is set to ‘server-name-01’.
- If you need to manage server names uniquely, create with different server names (Prefix).
- When creating servers, numbers do not automatically increase based on server name (Prefix), so Virtual Servers with the same name may be created.
- Example: If you first create 2 Virtual Servers using ’test’ as server name (Prefix), ’test-1’, ’test-2’ are created. Even if you create 2 Virtual Servers later using ’test’ as Prefix again, ’test-1’, ’test-2’ are created.
- If you create Virtual Server with Rocky Linux or Oracle Linux, additional settings are required for time synchronization (NTP: Network Time Protocol). For details, refer to Configure Linux NTP.
- If you create RHEL and Windows Server before July 2025, you need to modify RHEL Repository and WKMS (Windows Key Management Service) settings. For details, refer to Configure RHEL Repo and WKMS.
Check Virtual Server Details
Virtual Server service can check and modify overall resource list and detailed information. Virtual Server Detail page is composed of Detail Information, Monitoring, Tags, Operation History tabs.
To check detailed information of Virtual Server service, follow the procedure below.
- Click All Services > Compute > Virtual Server menu. You will be moved to the Service Home page of Virtual Server.
- On the Service Home page, click the Virtual Server menu. You will be moved to the Virtual Server List page.
- On the Virtual Server List page, click the resource to check detailed information for. You will be moved to the Virtual Server Detail page.
- Virtual Server Detail page displays status information and additional function information, and is composed of Detail Information, Monitoring, Tags, Operation History tabs.
- For details on Virtual Server status, refer to Check Virtual Server Detail Information.
Category Detailed Description Virtual Server Status Status of Virtual Server created by user - Build: State where Build command is transmitted
- Building: Build in progress
- Networking: Process in progress during server creation
- Scheduling: Process in progress during server creation
- Block_Device_Mapping: Connecting Block Storage during server creation
- Spawning: State where server creation process is in progress
- Active: Usable state
- Powering_off: State when stop is requested
- Deleting: Server deletion in progress
- Reboot_Started: State where Reboot is in progress
- Error: Error state
- Migrating: State where server is being migrated to another host
- Reboot: State where Reboot command is transmitted
- Rebooting: Restart in progress
- Rebuild: State where Rebuild command is transmitted
- Rebuilding: State when Rebuild is requested
- Rebuild_Spawning: State where Rebuild process is in progress
- Resize: State where Resize command is transmitted
- Resizing: Resize in progress
- Resize_Prep: State when server type modification is requested
- Resize_Migrating: State where server is moving to another host while Resize is in progress
- Resize_Migrated: State where server has completed moving to another host while Resize is in progress
- Resize_Finish: Resize completed
- Revert_Resize: Server Resize or migration failed for some reason. Target server is cleaned and original source server is restarted
- Shutoff: State when Powering off is completed
- Verity_ Resize: State where server type confirmation/server type reversion is selectable after Resize_Prep progress according to server type modification request
- Resize_Reverting: State when server type reversion is requested
- Resize_Confirming: State where server Resize request is being confirmed
Server Control Buttons to change server status - Start: Start stopped server
- Stop: Stop running server
- Restart: Restart running server
Create Image Create a user image from the current server image - For detailed image creation method, refer to Create Image
Console Log Check current server console log - Can check console log output from current server. For details, refer to Check Console Log
Create Dump Create current server Dump - Dump file is created inside Virtual Server
- For detailed Dump creation method, refer to Create Dump
Rebuild Delete existing Virtual Server OS area data and settings, and configure by Rebuilding as new server - For details, refer to Perform Rebuild
Terminate Service Button to terminate service Table. Virtual Server Status Information and Additional Functions
Detail Information
On the Virtual Server List page, you can check detailed information of selected resource and modify information if necessary.
| Category | Detailed Description |
|---|---|
| Service | Service name |
| Resource Type | Resource type |
| SRN | Unique resource ID in Samsung Cloud Platform
|
| Resource Name | Resource name
|
| Resource ID | Unique resource ID in service |
| Creator | User who created service |
| Creation Date | Date and time when service was created |
| Modifier | User who modified service information |
| Modification Date | Date and time when service information was modified |
| Server Name | Server name
|
| Server Type | Display vCPU, memory information
|
| Image Name | Server OS Image and version
|
| Lock | Display whether Lock is used/not used
|
| Server Group | Server group name to which server belongs
|
| Keypair Name | Server authentication information set by user
|
| Planned Compute | Resource status with Planned Compute set
|
| LLM Endpoint | URL for using LLM
|
| ServiceWatch Detailed Monitoring | Display whether ServiceWatch detailed monitoring is enabled |
- To **enable** ServiceWatch detailed monitoring, click **Modify** button to set
- For details, refer to [Enable ServiceWatch Detailed Monitoring](#servicewatch-세부-모니터링-활성화하기)
- For details on ServiceWatch service, refer to [ServiceWatch Overview](/userguide/management/service_watch/overview/_index.md)
- Not provided for Auto-Scaling Group, Virtual Server created from Marketplace
- VPC, General Subnet, IP and status, Public NAT IP and status, Private NAT IP and status, Virtual IP, Security Group
- If IP change is needed, click **Modify** button to set
- Can only modify when Virtual Server status is **Active**, **Shutoff**
- For Default port, **Default** is displayed next to IP, and cannot detach
- If Security Group change is needed, click **Modify** button to set
- If Virtual IP change is needed, can modify on **Virtual IP Management** tab of **Networking > VPC > Subnet Detail** page
- **Add as New Network Port**: Select General Subnet and IP
- Can select other General Subnet within same VPC
- IP can select auto create or user input, if input is selected, user can directly enter IP
- **Add as Existing Network Port**: Select pre-created General Subnet and port
- Local Subnet, Local Subnet IP, Security Group name, Virtual IP
- If Security Group change is needed, click **Modify** button to set
- **Add as New Network Port**: Select Local Subnet and IP
- Can select other General Subnet within same VPC
- **IP** can select **Auto Create** or user input, if **Input** is selected, user can directly enter IP
- **Add as Existing Network Port**: Select pre-created Local Subnet and port
- Volume ID, Volume name, Disk type, Capacity, Connection information, Type, Delete on termination, Status
- **Add**: Connect additional Block Storage when needed
- **Modify Delete on termination**: Modify Delete on termination value of selected Block Storage from list
- **More > Disconnect**: Disconnect Block Storage connection of selected Block Storage from list
- Cannot disconnect for OS default Storage
Monitoring
On the Virtual Server List page, you can monitor ServiceWatch metrics of selected resource. On the Monitoring tab, you can view monitoring charts for Virtual Server, and each chart is based on available Service Watch metrics.
| Category | Detailed Description |
|---|---|
| Period Setting Area | Select period to apply to chart
|
| Timezone Setting Area | Select timezone to apply to chart |
| Reset Button | Reset all manipulations or settings made on chart |
| Refresh Setting Area | Select refresh period of chart
|
| Move to Service Dashboard | Move to ServiceWatch dashboard list screen |
| More | Display additional work items for managing chart
|
| Graph Area | Data graph collected during period applied to chart
|
- Metrics provided by basic monitoring are data collected at 5-minute intervals.
- For details on Virtual Server ServiceWatch metrics, refer to Virtual Server ServiceWatch Metrics.
Tags
On the Virtual Server List page, you can check tag information of selected resource, and add, modify, or delete tags.
| Category | Detailed Description |
|---|---|
| Tag List | Tag list
|
Operation History
On the Virtual Server List page, you can check operation history of selected resource.
| Category | Detailed Description |
|---|---|
| Operation History List | Resource change history
|
Control Virtual Server Operation
If operation control of created Virtual Server resource is needed, you can perform tasks on Virtual Server List or Virtual Server Detail page. You can start, stop, and restart running servers.
Start Virtual Server
You can start stopped (Shutoff) Virtual Server. To start Virtual Server, follow the procedure below.
- Click All Services > Compute > Virtual Server menu. You will be moved to the Service Home page of Virtual Server.
- On the Service Home page, click the Virtual Server menu. You will be moved to the Virtual Server List page.
- On the Virtual Server List page, click the resource to start among stopped (Shutoff) servers to move to the Virtual Server Detail page.
- On the Virtual Server List page, you can Start through More button on the right for each resource.
- After selecting multiple servers with checkbox, you can control multiple servers simultaneously through Start button at the top.
- On the Virtual Server Detail page, click Start button at the top to start the server. Check changed server status in Status Display item.
- When Virtual Server start is completed, server status changes from Shutoff to Active.
- For details on Virtual Server status, refer to Check Virtual Server Detail Information.
Stop Virtual Server
You can stop running (Active) Virtual Server. To stop Virtual Server, follow the procedure below.
- Click All Services > Compute > Virtual Server menu. You will be moved to the Service Home page of Virtual Server.
- On the Service Home page, click the Virtual Server menu. You will be moved to the Virtual Server List page.
- On the Virtual Server List page, click the resource to stop among running (Active) servers to move to the Virtual Server Detail page.
- On the Virtual Server List page, you can Stop through More button on the right for each resource.
- After selecting multiple servers with checkbox, you can control multiple servers simultaneously through Stop button at the top.
- On the Virtual Server Detail page, click Stop button at the top to start the server. Check changed server status in Status Display item.
- When Virtual Server stop is completed, server status changes from Active to Shutoff.
- For details on Virtual Server status, refer to Check Virtual Server Detail Information.
Restart Virtual Server
You can restart created Virtual Server. To restart Virtual Server, follow the procedure below.
- Click All Services > Compute > Virtual Server menu. You will be moved to the Service Home page of Virtual Server.
- On the Service Home page, click the Virtual Server menu. You will be moved to the Virtual Server List page.
- On the Virtual Server List page, click the resource to restart to move to the Virtual Server Detail page.
- On the Virtual Server List page, you can Restart through More button on the right for each resource.
- After selecting multiple servers with checkbox, you can control multiple servers simultaneously through Restart button at the top.
- On the Virtual Server Detail page, click Restart button at the top to start the server. Check changed server status in Status Display item.
- During Virtual Server restart, server status changes to Rebooting and finally to Active.
- For details on Virtual Server status, refer to Check Virtual Server Detail Information.
Manage Virtual Server Resources
If server control and management functions of created Virtual Server resource are needed, you can perform tasks on Virtual Server List or Virtual Server Detail page.
Create Image
You can create Image of running Virtual Server.
This content guides how to create user Image from running Virtual Server.
- Create user Image by clicking Create Image button on Virtual Server List or Virtual Server Detail page.
- For method of creating Image by uploading user’s owned Image file, refer to Create Image in Image Detail Guide.
To create Virtual Server Image, follow the procedure below.
Click All Services > Compute > Virtual Server menu. You will be moved to the Service Home page of Virtual Server.
On the Service Home page, click the Virtual Server menu. You will be moved to the Virtual Server List page.
On the Virtual Server List page, click the resource to create Image. You will be moved to the Virtual Server Detail page.
On the Virtual Server Detail page, click the Create Image button. You will be moved to the Create Image page.
- In the Service information input area, enter the required information.
Category RequiredDetailed Description Image Name Required Name of Image to create - Enter within 200 characters using English, numbers, spaces, and special characters (
-,_)
Table. Image Service Information Input Items - Enter within 200 characters using English, numbers, spaces, and special characters (
- In the Service information input area, enter the required information.
Check input information and click Complete button.
- When creation is complete, check the created resource on All Services > Compute > Virtual Server > Image List page.
- When Image is created, it is saved to Object Storage used as internal storage. Therefore, usage fee is charged for Image storage.
- File system of Image created from Active Virtual Server cannot guarantee integrity, so Image creation after server stop is recommended.
Modify Server Type
You can modify server type of Virtual Server server.
To modify Virtual Server server type, follow the procedure below.
- Click All Services > Compute > Virtual Server menu. You will be moved to the Service Home page of Virtual Server.
- On the Service Home page, click the Virtual Server menu. You will be moved to the Virtual Server List page.
- On the Virtual Server List page, click the resource to control operation. You will be moved to the Virtual Server Detail page.
- On the Virtual Server Detail page, check server status and click server type Modify button. Modify Server Type popup window opens.
- On the Modify Server Type popup window, change server type and then click Confirm button.
- When Virtual Server server type is modified, Virtual Server status changes to Resize execution related status.
- For details on Virtual Server status, refer to Check Virtual Server Detail Information.
Change IP
For method of changing IP, refer to Change IP.
- If IP change is performed, can no longer communicate with that IP, and IP change cannot be cancelled during progress.
- Server is rebooted to apply changed IP.
- If server is in Load Balancer service, must delete existing IP from LB server group and directly add changed IP as member of LB server group.
- For servers using Public NAT/Private NAT, must disable use of Public NAT/Private NAT and then set again after IP change.
- If Public NAT/Private NAT is in use, first disable use of Public NAT/Private NAT and after completing IP change, set again.
- Whether Public NAT/Private NAT is used can be changed by clicking Modify button of Public NAT IP/Private NAT IP on Virtual Server Detail page.
Enable ServiceWatch Detailed Monitoring
By default, Virtual Server is linked with ServiceWatch for basic monitoring. If needed, you can enable detailed monitoring to more quickly identify and respond to operational issues. For details on ServiceWatch, refer to ServiceWatch Overview.
To enable Virtual Server ServiceWatch detailed monitoring, follow the procedure below.
- Click All Services > Compute > Virtual Server menu. You will be moved to the Service Home page of Virtual Server.
- On the Service Home page, click the Virtual Server menu. You will be moved to the Virtual Server List page.
- On the Virtual Server List page, click the resource to enable ServiceWatch detailed monitoring. You will be moved to the Virtual Server Detail page.
- On the Virtual Server Detail page, click ServiceWatch detailed monitoring Modify button. You will be moved to Modify ServiceWatch Detailed Monitoring popup window.
- On the Modify ServiceWatch Detailed Monitoring popup window, select Enable, check guide text and then click Confirm button.
- Check ServiceWatch detailed monitoring item on Virtual Server Detail page.
Disable ServiceWatch Detailed Monitoring
To disable Virtual Server ServiceWatch detailed monitoring, follow the procedure below.
- Click All Services > Compute > Virtual Server menu. You will be moved to the Service Home page of Virtual Server.
- On the Service Home page, click the Virtual Server menu. You will be moved to the Virtual Server List page.
- On the Virtual Server List page, click the resource to disable ServiceWatch detailed monitoring. You will be moved to the Virtual Server Detail page.
- On the Virtual Server Detail page, click ServiceWatch detailed monitoring Modify button. You will be moved to Modify ServiceWatch Detailed Monitoring popup window.
- On the Modify ServiceWatch Detailed Monitoring popup window, deselect Enable, check guide text and then click Confirm button.
- Check ServiceWatch detailed monitoring item on Virtual Server Detail page.
Virtual Server Management Additional Functions
For Virtual Server server management, you can check console log, create Dump, and perform Rebuild. To check Virtual Server console log, create Dump, and perform Rebuild, follow the procedure below.
Check Console Log
You can check current console log of Virtual Server.
To check Virtual Server console log, follow the procedure below.
- Click All Services > Compute > Virtual Server menu. You will be moved to the Service Home page of Virtual Server.
- On the Service Home page, click the Virtual Server menu. You will be moved to the Virtual Server List page.
- On the Virtual Server List page, click the resource to check console log. You will be moved to the Virtual Server Detail page.
- On the Virtual Server Detail page, click Console Log button. You will be moved to Console Log popup window.
- Check console log output on Console Log popup window.
Create Dump
To create Virtual Server Dump file, follow the procedure below.
- Click All Services > Compute > Virtual Server menu. You will be moved to the Service Home page of Virtual Server.
- On the Service Home page, click the Virtual Server menu. You will be moved to the Virtual Server List page.
- On the Virtual Server List page, click the resource to check detailed information. You will be moved to the Virtual Server Detail page.
- On the Virtual Server Detail page, click Create Dump button.
- Dump file is created inside Virtual Server.
Perform Rebuild
You can delete existing Virtual Server server OS area data and settings, and configure by Rebuilding as new server.
To perform Virtual Server Rebuild, follow the procedure below.
- Click All Services > Compute > Virtual Server menu. You will be moved to the Service Home page of Virtual Server.
- On the Service Home page, click the Virtual Server menu. You will be moved to the Virtual Server List page.
- On the Virtual Server List page, click the resource to perform Rebuild. You will be moved to the Virtual Server Detail page.
- On the Virtual Server Detail page, click Rebuild button.
- During Virtual Server Rebuild, server status changes to Rebuilding and when Rebuild is completed, returns to state before Rebuild execution.
- For details on Virtual Server status, refer to Check Virtual Server Detail Information.
Terminate Virtual Server
Terminating unused Virtual Server can reduce operating costs. However, since Virtual Server termination can immediately stop running services, you must fully consider the impact of service interruption before proceeding with termination.
To terminate Virtual Server, follow the procedure below.
- Click All Services > Compute > Virtual Server menu. You will be moved to the Service Home page of Virtual Server.
- On the Service Home page, click the Virtual Server menu. You will be moved to the Virtual Server List page.
- On the Virtual Server List page, select the resource to terminate and click Terminate Service button.
- Termination of connected storage varies depending on the Delete on termination setting; see Termination Constraints.
- When termination is complete, check if resource is terminated on Virtual Server List page.
Termination Constraints
If termination is not possible when Virtual Server termination is requested, you will be guided through a popup window. Refer to the cases below.
- When File Storage is connected: First disconnect File Storage connection.
- When LB server group is connected: First disconnect LB server group Pool connection.
- When Lock is set: Change Lock setting to not used and then try again.
- When Backup is connected: First disconnect Backup connection.
- When Auto-Scaling Group connected to Virtual Server is not In Service state: Change state of connected Auto-Scaling Group and then try again.
Termination of connected storage varies depending on Delete on termination setting, so please refer.
- Volume deletion varies depending on whether Delete on termination is set.
- When Delete on termination is not set: Volume is not deleted even if Virtual Server is terminated.
- When Delete on termination is set: Volume is deleted when Virtual Server is terminated.
- Volumes with snapshots are not deleted even if Delete on termination is set.
- Multi attach volumes are deleted only when the server to be deleted is the last remaining server connected to the volume.
1.2.1 - Image
The user can enter the required information for the Image service within the Virtual Server service and select detailed options through the Samsung Cloud Platform Console to create the service.
Image generation
You can create and use the Image service while using the Virtual Server service on the Samsung Cloud Platform Console.
To create an Image, follow the steps below.
Click the All Services > Compute > Virtual Server menu. Go to the Virtual Server’s Service Home page.
Click the Image menu on the Service Home page. Go to the Image List page.
Click the Image Create button on the Image List page. It navigates to the Image Create page.
- Service Information Input Enter or select the required information in the area.
Category RequiredDetailed description Image name Required Name of the Image to create - Enter within 255 characters using English letters, numbers, spaces, and special characters (
-,_)
Image file > URL Required Enter URL after uploading Image file to Object Storage - Object Storage Details page allows copying URL
- The Bucket of Object Storage where the Image file is uploaded must be in the same zone as the server to be created
- Image file can only have .qcow2 extension
- Upload a secure Image file to minimize security risks.
OS type Required OS type of the uploaded Image file - Select from Alma Linux, CentOS, Oracle Linux, RHEL, Rocky Linux, Ubuntu
Minimum Disk Required Minimum disk size (GB) for the Image to be created - Enter a value between
0and12,288GB
Minimum RAM Required Minimum RAM capacity (GB) of the image to be created - Enter a value between
0and2,097,151GB
Visibility Required Indicates access permissions for the Image - Private: Can be used only within the Account
- Shared: Can be shared between Accounts
Protected Select Select whether Image deletion is prohibited - Checking Use prevents accidental deletion of the Image
- This setting can be changed after Image creation
Table. Image Service Information Input Items - Enter within 255 characters using English letters, numbers, spaces, and special characters (
- Additional Information Input Enter or select the required information in the area.
Category RequiredDetailed description Tag Select Add Tag - Up to 50 can be added per resource
- After clicking the Add Tag button, enter or select Key, Value values
Table. Image additional information input items
- Service Information Input Enter or select the required information in the area.
Summary Check the detailed information and estimated billing amount generated in the panel, and click the Complete button.
- When creation is complete, check the created resources on the Image List page.
Image Check detailed information
Image service can view and edit the full resource list and detailed information. Image detail page consists of detailed information, tags, operation history tabs.
To view detailed information of the Image service, follow the steps below.
- All Services > Compute > Virtual Server Click the menu. Go to the Service Home page of Virtual Server.
- Click the Image menu on the Service Home page. Go to the Image list page.
- Image List page, click the resource to view detailed information. Image Details page will be opened.
- Image Details page displays status information and additional feature information, and consists of Detail Information, Tag, Work History tabs.
Category Detailed description Image status Status of the Image created by the user - Active: Available
- Queued: When an Image creation request is made, the Image is uploaded and waiting for processing
- Importing: When an Image creation request is made, the Image is uploaded and being processed
Create shared Image Create Image to share with another Account - Can be created only when the Image’s Visibility is private and the Image has snapshot information
Share with another Account Image can be shared with another Account - If the Image’s Visibility is Shared, it can be shared with another Account
- Only displayed for Images created by Create shared Image or by uploading a qcow2 file
Image Delete Button to delete the Image - If the Image is deleted, it cannot be recovered
Table. Image status information and additional functions
- Image Details page displays status information and additional feature information, and consists of Detail Information, Tag, Work History tabs.
Detailed Information
Image list page allows you to view detailed information of the selected resource and edit the information if needed.
| Category | Detailed description |
|---|---|
| Service | Service Name |
| Resource Type | Resource Type |
| SRN | Unique resource ID in Samsung Cloud Platform
|
| Resource Name | Image Name |
| Resource ID | Image’s unique resource ID |
| Creator | User who created the Image |
| Creation time | Image creation time |
| Editor | User who modified the Image information |
| Modification Date | Date/Time when Image information was modified |
| Image name | Image name |
| Minimum Disk | Image’s minimum disk capacity (GB)
|
| Minimum RAM | Minimum RAM capacity of the Image (GB) |
| OS type | Image’s OS type
|
| OS hash algorithm | OS hash algorithm method |
| Visibility | Displays access permissions for the Image
|
| Protected | Select whether image deletion is prohibited
|
| Image size | Image size
|
| Image Type | Classification by Image creation method
|
| Image file URL | Image file URL uploaded to Object Storage when creating an Image
|
| Sharing Status | Status of sharing images with other Accounts
|
Tag
Image list page, you can view the tag information of the selected resource, and you can add, modify, or delete it.
| Category | Detailed description |
|---|---|
| Tag List | Tag List
|
Work History
You can view the operation history of the selected resource on the Image list page.
| Category | Detailed description |
|---|---|
| Work History List | Resource Change History
|
Image Resource Management
Describes the control and management functions of the generated Image.
Create Image for Sharing
Create an Image to share with another Account.
- Image’s Visibility is private and only when the Image has snapshot information can a shared Image be created.
- Shared Image includes only one OS area disk volume as the imaging target. Additionally, connected data volumes are not included in the Image, so if needed, please copy the data to a separate volume and use the volume migration function.
To create an image for sharing, follow the steps below.
- Log in to the shared Account and click the All Services > Compute > Virtual Server menu. Go to the Virtual Server’s Service Home page.
- Click the Image menu on the Service Home page. Navigate to the Image List page.
- Click the Image to create a shared Image on the Image List page. You will be taken to the Image Details page.
- Create Shared Image Click the button. A popup window notifying the creation of a shared Image will open.
- After checking the notification content, click the Complete button.
Share Image to another Account
Create an image to share with another Account.
- .qcow2 extension file uploaded to create, or only Images created via Image Details page with Create Shared Image can be shared with other Accounts.
- The Image to be shared must have Visibility set to Shared.
To share the Image with another Account, follow these steps.
Log in to the shared Account and click the All Services > Compute > Virtual Server menu. Navigate to the Virtual Server’s Service Home page.
Click the Image menu on the Service Home page. It navigates to the Image List page.
On the Image List page, click the Image you want to share with another Account. It moves to the Image Details page.
Click the Share to another Account button. A popup window notifying Image sharing opens.
After checking the notification content, click the Confirm button. It moves to the Share Image with another Account page.
Share Image with another Account on the page, enter Share Account ID, and click the Complete button. A popup notifying Image sharing opens.
Category RequiredDetailed description Image name - Name of the Image to share - Input not allowed
Image ID - Image ID to share - Input not allowed
Shared Account ID Required Enter another Account ID to share - Enter within 64 characters using English letters, numbers, and special character
-
Table. Image sharing items to another AccountAfter checking the notification content, click the Confirm button. You can check the information in the sharing status of the Image Details page.
- When first requested, the status is Pending, and when approval is completed by the Account to be shared, it changes to Accepted, and if approval is denied, it changes to Rejected.
Receive shared Image from another Account
To receive an Image shared from another Account, follow the steps below.
Log in to the account to be shared and click the All Services > Compute > Virtual Server menu. Go to the Service Home page of the Virtual Server.
Click the Image menu on the Service Home page. It navigates to the Image List page.
Image List on the page More > Image Share Request List click the button. Image Share Request List popup opens.
Image Sharing Request List In the popup window, click the Approve or Reject button for the Image to be shared.
Category Detailed description Image name shared Image name OS type OS type of shared Image Owner Account ID Owner Account ID of shared Image Creation time Creation time of shared Image Approval Approve the shared Image Reject Reject processing of the shared Image Table. Image sharing request list itemAfter checking the notification content, click the Confirm button. You can check the shared Image in the Image list.
Image Delete
You can delete unused images. However, once an image is deleted it cannot be recovered, so you should fully consider the impact before proceeding with the deletion.
To delete Image, follow the steps below.
- All Services > Compute > Virtual Server Click the menu. Navigate to the Service Home page of Virtual Server.
- Click the Image menu on the Service Home page. Go to the Image list page.
- On the Image list page, select the resource to delete and click the Delete button.
- Image list page, select multiple Image check boxes, and click the Delete button at the top of the resource list.
- When deletion is complete, check on the Image List page whether the resource has been deleted.
1.2.2 - Keypair
Users can create a Keypair within the Virtual Server service by entering the required information and selecting detailed options through the Samsung Cloud Platform Console.
Creating a Keypair
You can create and use a Keypair while using the Virtual Server service in the Samsung Cloud Platform Console.
Follow these steps to create a Keypair.
All Services > Compute > Virtual Server – Click to go to the Virtual Server Service Home page.
On the Service Home page, click the Keypair menu to go to the Keypair List page.
On the Keypair List page, click the Create Keypair button to go to the Create Keypair page.
- In the Service Information Input section, enter the required information.
Item Required Description Keypair Name Required Name of the Keypair to create. Use English letters, numbers, spaces, and special characters ( -,_) up to 255 characters.Keypair Type Required ssh Table. Keypair service information input items - In the Additional Information Input section, enter or select the required information.
Item Required Description Tags Optional Add tags (up to 50 per resource). Click the Add Tag button and enter/select Key and Value. Table. Additional Keypair information input items
- In the Service Information Input section, enter the required information.
Review the entered information and click the Create button.
- After creation, the new Keypair will appear on the Keypair List page.
- After creation, you can download the Private Key only once. It cannot be reissued, so ensure you have downloaded it.
- Store the downloaded Private Key in a secure location.
Viewing Keypair Details
The Keypair service allows you to view and edit the resource list and detailed information. The Keypair Details page consists of Details, Tags, and Activity History tabs.
To view the details of a Keypair, follow these steps.
- All Services > Compute > Virtual Server – Click to go to the Virtual Server Service Home page.
- On the Service Home page, click the Keypair menu to go to the Keypair List page.
- On the Keypair List page, click the resource you want to view details for. You will be taken to the Keypair Details page.
- The Keypair Details page displays status information and additional feature information, organized into Details, Tags, and Activity History tabs.
Details
You can view the detailed information of the selected resource from the Keypair List page, and modify the information if needed.
| Item | Description |
|---|---|
| Service | Service name |
| Resource Type | Resource type |
| SRN | Unique resource ID in Samsung Cloud Platform (for Keypair, refers to Keypair SRN) |
| Resource Name | Keypair name |
| Resource ID | Unique resource ID of the Keypair |
| Creator | User who created the Keypair |
| Creation Time | Timestamp when the Keypair was created |
| Modifier | User who modified the Keypair information |
| Modification Time | Timestamp when the Keypair information was modified |
| Keypair Name | Name of the Keypair |
| Fingerprint | Unique value to identify the key |
| User ID | User ID of the Keypair creator |
| Public Key | Public key information |
Tags
You can view, add, modify, or delete the tags of a selected resource from the Keypair List page.
| Item | Description |
|---|---|
| Tag List | List of tags. You can view the Key and Value of each tag. Up to 50 tags can be added per resource. When entering a tag, you can search the existing Key and Value lists to select them. |
Activity History
You can view the activity history of a selected resource from the Keypair List page.
| Item | Description |
|---|---|
| Activity History List | Resource change history, including operation time, resource ID, resource name, operation details, event topic, operation result, and operator information. |
Managing Keypair Resources
This section describes the control and management functions for Keypair.
Retrieving the Public Key
Follow these steps to retrieve the public key.
All Services > Compute > Virtual Server – Click to go to the Virtual Server Service Home page.
On the Service Home page, click the Keypair menu to go to the Keypair List page.
On the Keypair List page, click the More button at the top and then click Retrieve Public Key. You will be taken to the Retrieve Public Key page.
- In the Required Information Input section, enter or select the required information.
Item Required Description Keypair Name Required Name of the Keypair to retrieve Keypair Type Required ssh Public Key Required Enter the public key. - File Upload: Click the Attach File button to attach a public key file (only .pem files are allowed).
- Public Key Input: Paste the copied public key value (you can copy the public key value from the Keypair Details page).
Table. Required fields for retrieving a public key
- In the Required Information Input section, enter or select the required information.
Review the entered information and click the Complete button.
- After completion, the newly created resource will appear on the Keypair List page.
Deleting a Keypair
You can delete an unused Keypair. However, once a Keypair is deleted, it cannot be recovered, so please review the impact carefully before deletion.
To delete a Keypair, follow these steps.
- All Services > Compute > Virtual Server – Click to go to the Virtual Server Service Home page.
- On the Service Home page, click the Keypair menu to go to the Keypair List page.
- On the Keypair List page, select the resource you want to delete and click the Delete button.
- You can select multiple Keypairs using checkboxes and click the Delete button at the top of the resource list.
- After deletion is complete, verify that the resource has been removed from the Keypair List page.
1.2.3 - Server Group
Users can enter the required information for a Server Group within the Virtual Server service and select detailed options through the Samsung Cloud Platform Console to create the service.
Server Group Create
You can create and use the Server Group service while using the Virtual Server service in the Samsung Cloud Platform Console.
To create a Server Group, follow the steps below.
- Click the All Services > Compute > Virtual Server menu. Go to the Service Home page of Virtual Server.
- Click the Server Group menu on the Server Group page. Go to the Server Group list page.
- Server Group List on the page, click the Server Group Create button. Navigate to the Server Group Create page.
- Service Information Input area, enter or select the required information.
Category RequiredDetailed description Server Group name Required Name of the Server Group to create - Enter within 255 characters using English letters, numbers, spaces, and special characters (
-,_)
Policy Required Set Anti-Affinity (distributed placement), Affinity (proximate placement), Partition (distributed placement of Virtual Server and Block Storage) for Virtual Servers belonging to the same Server Group - Anti-Affinity (distributed placement) and Affinity (proximate placement) policies place Virtual Servers belonging to the same Server Group based on the selected policy in a Best Effort manner, but are not absolutely guaranteed.
- Anti-Affinity (distributed placement): A policy that places servers belonging to a Server Group on different racks and hosts as much as possible
- Affinity (proximate placement): A policy that places servers belonging to a Server Group close together within the same rack and host as much as possible
- Partition (distributed placement of Virtual Server and Block Storage): A policy that places Virtual Servers belonging to a Server Group and the Block Storage connected to those servers in different distribution units (Partitions)
- The Partition (distributed placement of Virtual Server and Block Storage) policy displays the Partition number together so that it is clear which Partition each Virtual Server and its associated Block Storage belong to.
- Partition numbers are assigned based on the Partition Size (up to 3) set for the Server Group.
Table. Server Group Service Information Input Items - Enter within 255 characters using English letters, numbers, spaces, and special characters (
- Add Information Input area, enter or select the required information.
Category RequiredDetailed description Tag Select Add Tag - Up to 50 can be added per resource
- After clicking the Add Tag button, enter or select Key, Value values
Table. Server Group Additional Information Input Items
- Service Information Input area, enter or select the required information.
- Check the input information and click the Complete button.
- When creation is complete, check the created resources on the Server Group List page.
Server Group View detailed information
Server Group service can view and edit the full resource list and detailed information. Server Group Details page consists of Details, Tags, Activity Log tabs.
To view detailed information of the Server Group, follow the steps below.
- Click the All Services > Compute > Virtual Server menu. Go to the Service Home page of Virtual Server.
- Click the Server Group menu on the Service Home page. You will be taken to the Server Group List page.
- Click the resource to view detailed information on the Server Group List page. It navigates to the Server Group Details page.
- Server Group Details page displays status information and additional feature information, and consists of Details, Tags, Activity History tabs.
Detailed Information
On the Server Group List page, you can view detailed information of the selected resource and, if necessary, edit the information.
| Category | Detailed description |
|---|---|
| Service | Service Name |
| Resource Type | Resource Type |
| SRN | Unique resource ID in Samsung Cloud Platform
|
| Resource Name | Server Group Name |
| Resource ID | Unique resource ID of Server Group |
| Creator | User who created the Server Group |
| Creation time | Server Group creation time |
| Server Group name | Server Group name |
| Policy | Anti-Affinity(distributed placement), Affinity(proximal placement), Partition(distributed placement of Virtual Server and Block Storage) |
| Server Group Member | List of Virtual Servers belonging to the Server Group
|
Tag
Server Group List page you can view the tag information of the selected resource, and you can add, modify, or delete it.
| Category | Detailed description |
|---|---|
| Tag List | Tag List
|
Work History
You can view the operation history of the selected resource on the Server Group List page.
| Category | Detailed description |
|---|---|
| Work History List | Resource Change History
|
Server Group Delete
You can delete unused Server Groups. However, once a Server Group is deleted it cannot be recovered, so please review the impact thoroughly in advance before proceeding with deletion.
To delete a Server Group, follow these steps.
- All Services > Compute > Virtual Server menu, click it. Go to the Virtual Server’s Service Home page.
- On the Service Home page, click the Server Group menu. Navigate to the Server Group List page.
- Server Group list On the page, select the resource to delete, and click the Delete button.
- Server Group list on the page select multiple Server Group check boxes, and click the Delete button at the top of the resource list.
- When deletion is complete, check whether the resource has been deleted on the Server Group list page.
1.2.4 - IP Change
You can change the IP of the Virtual Server and add network ports to the Virtual Server to set the IP.
IP Change
You can change the IP of the Virtual Server.
- If you proceed with changing the IP, you will no longer be able to communicate with that IP, and you cannot cancel the IP change while it is in progress.
- The server will be rebooted to apply the changed IP.
- If the server is running the Load Balancer service, you must delete the existing IP from the LB server group and directly add the changed IP as a member of the LB server group.
- Servers using Public NAT/Private NAT must disable and reconfigure Public NAT/Private NAT after changing the IP.
- If you are using Public NAT/Private NAT, first disable the use of Public NAT/Private NAT, complete the IP change, and then set it again.
- Whether to use Public NAT/Private NAT can be changed by clicking the Edit button of Public NAT IP/Private NAT IP on the Virtual Server Details page.
To change the IP, follow the steps below.
- All Services > Compute > Virtual Server Click the menu. Navigate to the Service Home page of Virtual Server.
- Click the Virtual Server menu on the Service Home page. Move to the Virtual Server List page.
- Virtual Server List Click the resource to change the IP on the page. Navigate to the Virtual Server Details page.
- Virtual Server Details page, click the Edit button of the IP item to change the IP. The IP Edit popup opens.
- Edit IP In the popup window, after selecting Subnet, set the IP to change.
- Input: Enter the IP to be changed directly.
- Automatic Generation: Automatically generate the IP and apply it.
- When the settings are complete, click the Confirm button.
- When the popup notifying IP modification opens, click the Confirm button.
Setting IP on the server after adding network ports
If you create a Virtual Server with Ubuntu Linux, after adding a network port on Samsung Cloud Platform, additional IP configuration is required on the server.
As the root user of the Virtual Server’s OS, use the ip command to check the assigned network interface name.
Color modeip aip aCode block. ip command - network interface check command - If there is an added interface, the following result is displayed.
Color mode[root@scp-test-vm-01 ~] # ip a 3: ens7: <BROADCAST,MULTICAST> mtu 9000 qdisc noop state DOWN group default qlen 1000 link/ether fa:16:3e:98:b6:64 brd ff:ff:ff:ff:ff:ff altname enp0s7[root@scp-test-vm-01 ~] # ip a 3: ens7: <BROADCAST,MULTICAST> mtu 9000 qdisc noop state DOWN group default qlen 1000 link/ether fa:16:3e:98:b6:64 brd ff:ff:ff:ff:ff:ff altname enp0s7Code block. ip command - Network interface check result Use a text editor (e.g., vim) to open the /etc/netplan/50-cloud-init.yaml file.
Add the following content to the /etc/netplan/50-cloud-init.yaml file and save it.
Color modenetwork: version: 2 ethernets: ens7: match: macaddress: "fa:16:3e:98:b6:64" dhcp4: true set-name: "ens7" mtu: 9000network: version: 2 ethernets: ens7: match: macaddress: "fa:16:3e:98:b6:64" dhcp4: true set-name: "ens7" mtu: 9000Code block. Edit YAML file
Set the IP on the added network DEVICE using the netplan command.
Color modenetplan --debug applynetplan --debug applyCode block. netplan applied Use the ip command to verify that the IP is set correctly.
Color mode[root@scp-test-vm-01 ~] # ip a 3: ens7: <BROADCAST,MULTICAST> mtu 9000 qdisc noop state DOWN group default qlen 1000 link/ether fa:16:3e:98:b6:64 brd ff:ff:ff:ff:ff:ff altname enp0s7 inet 10.10.10.10/24 metric 100 brd 10.10.10.255 scope global dynamic ens7 valid_lft 43197sec preferred_lft 43197sec inet6 fe80::f816:3eff:fe0a:96bf/64 scope link valid_lft forever preferred_lft forever[root@scp-test-vm-01 ~] # ip a 3: ens7: <BROADCAST,MULTICAST> mtu 9000 qdisc noop state DOWN group default qlen 1000 link/ether fa:16:3e:98:b6:64 brd ff:ff:ff:ff:ff:ff altname enp0s7 inet 10.10.10.10/24 metric 100 brd 10.10.10.255 scope global dynamic ens7 valid_lft 43197sec preferred_lft 43197sec inet6 fe80::f816:3eff:fe0a:96bf/64 scope link valid_lft forever preferred_lft foreverCode block. Check IP settings
1.2.5 - Linux NTP Setting
If a user creates a Virtual Server with Rocky Linux or Oracle Linux via the Samsung Cloud Platform Console, additional configuration is required for time synchronization (NTP: Network Time Protocol). For other OS standard Linux images (RHEL, Alma Linux, Ubuntu), NTP is already configured, so no additional setup is needed.
Install NTP Daemon
You can install the chrony daemon to configure NTP. To install the chrony daemon, follow the steps below.
Check whether the chrony package is installed using the dnf command as the root user of the OS of the Virtual Server.
Color modednf list chronydnf list chronyCode block. dnf command - chrony package installation verification command - chrony If the chrony package is installed, the following result is displayed.
Color mode[root@scp-test-vm-01 ~] # dnf list chrony Last metadata expiration check: 1:47:29 ago on Wed 19 Feb 2025 05:55:57 PM KST. Installed Packages chrony.x86_64 3.5-1.0.1.el8 @anaconda[root@scp-test-vm-01 ~] # dnf list chrony Last metadata expiration check: 1:47:29 ago on Wed 19 Feb 2025 05:55:57 PM KST. Installed Packages chrony.x86_64 3.5-1.0.1.el8 @anacondaCode block. dnf command - chrony package installation verification result If the chrony package is not installed, use the dnf command to install the chrony package.
Color modednf install chrony -ydnf install chrony -yCode block. dnf command - chrony package installation verification command
NTP Daemon Setup
To set up the chrony daemon, follow these steps.
Load the /etc/chrony.conf file using a text editor (e.g., vim).
Add the following content to the /etc/chrony.conf file and save.
Color modeserver 198.19.0.54 iburstserver 198.19.0.54 iburstCode block. /etc/chrony.conf edit Set it to automatically start the chrony daemon using the systemctl command.
Color modesystemctl enable chronydsystemctl enable chronydCode block. systemctl command - chrony daemon auto start setting Restart the chrony daemon using the systemctl command.
Color modesystemctl restart chronydsystemctl restart chronydCode block. systemctl command - restart chrony daemon Run the chronyc sources command with the “v” option (display detailed information) to check the IP address of the configured NTP server and verify whether synchronization is in progress.
Color modechronyc sources -vchronyc sources -vCode block. chronyc sources command - NTP synchronization check - When you run the chronyc sources command, the following result is displayed.
Color mode[root@scp-test-vm-01 ~] # chronyc sources -v 210 Number of sources = 1 .-- Source mode '^' = server, '=' = peer, '#' = local clock. / .- Source state '*' = current synced, '+' = combined , '-' = not combined, | / '?' = unreachable, 'x' = time may be in error, '~' = time too variable. || .- xxxx [ yyyy ] +/- zzzz || Reachability register (octal) -. | xxxx = adjusted offset, || Log2(Polling interval) --. | | yyyy = measured offset, || \ | | zzzz = estimated error. || | | \ MS Name/IP address Stratum Poll Reach LastRx Last sample ========================================================================= ^* 198.19.0.54 2 6 377 52 -129us[ -128us] +/- 14ms[root@scp-test-vm-01 ~] # chronyc sources -v 210 Number of sources = 1 .-- Source mode '^' = server, '=' = peer, '#' = local clock. / .- Source state '*' = current synced, '+' = combined , '-' = not combined, | / '?' = unreachable, 'x' = time may be in error, '~' = time too variable. || .- xxxx [ yyyy ] +/- zzzz || Reachability register (octal) -. | xxxx = adjusted offset, || Log2(Polling interval) --. | | yyyy = measured offset, || \ | | zzzz = estimated error. || | | \ MS Name/IP address Stratum Poll Reach LastRx Last sample ========================================================================= ^* 198.19.0.54 2 6 377 52 -129us[ -128us] +/- 14msCode block. chronyc sources command - NTP synchronization check Run the chronyc tracking command to check the synchronization metrics.
Color mode[root@scp-test-vm-01 ~] # chronyc tracking Reference ID : A9FEA9FE (198.19.0.54) Stratum : 3 Ref time (UTC) : Wed Feb 19 18:48:41 2025 System time : 0.000000039 seconds fast of NTP time Last offset : -0.000084246 seconds RMS offset : 0.000084246 seconds Frequency : 21.667 ppm slow Residual freq : +4.723 ppm Skew : 0.410 ppm Root delay : 0.000564836 seconds Root dispersion : 0.027399288 seconds Update interval : 2.0 seconds Leap status : Normal[root@scp-test-vm-01 ~] # chronyc tracking Reference ID : A9FEA9FE (198.19.0.54) Stratum : 3 Ref time (UTC) : Wed Feb 19 18:48:41 2025 System time : 0.000000039 seconds fast of NTP time Last offset : -0.000084246 seconds RMS offset : 0.000084246 seconds Frequency : 21.667 ppm slow Residual freq : +4.723 ppm Skew : 0.410 ppm Root delay : 0.000564836 seconds Root dispersion : 0.027399288 seconds Update interval : 2.0 seconds Leap status : NormalCode block. chronyc tracking command - NTP synchronization metric
1.2.6 - Setting up RHEL Repo and WKMS
- If the user created RHEL and Windows Server prior to August 2025 via the Samsung Cloud Platform Console, they need to modify the RHEL Repository and WKMS (Windows Key Management Service) settings.
- The SCP RHEL Repository is a repository provided by SCP to support user environments such as VPC Private Subnet where external access is restricted.
Since the SCP RHEL Repository synchronizes with each Region Local Repository according to the internal schedule, it is recommended to switch to an external Public Mirror site to quickly apply the latest patches.
RHEL Repository Configuration Guide
In Samsung Cloud Platform, when using RHEL, you can install and download the same packages as the official RHEL Repository by utilizing the RHEL Repository provided by SCP.
SCP provides the latest version of the repository for the given major version by default. To set up the RHEL repository, follow the steps below.
Using the root user of the Virtual Server’s OS, use the cat command to check the
/etc/yum.repos.d/scp.rhel8.repoor/etc/yum.repos.d/scp.rhel9.reposettings.Color modecat /etc/yum.repos.d/scp.rhel8.repocat /etc/yum.repos.d/scp.rhel8.repoCode block. repo configuration check (RHEL8) Color modecat /etc/yum.repos.d/scp.rhel9.repocat /etc/yum.repos.d/scp.rhel9.repoCode block. repo configuration check (RHEL9) - When checking the configuration file, the following result is displayed.Color mode
[rhel-8-baseos] name=rhel-8-baseos gpgcheck=0 enabled=1 baseurl=http://scp-rhel8-ip/rhel/8/baseos [rhel-8-baseos-debug] name=rhel-8-baseos-debug gpgcheck=0 enabled=1 baseurl=http://scp-rhel8-ip/rhel/8/baseos-debug [rhel-8-appstream] name=rhel-8-appstream gpgcheck=0 enabled=1 baseurl=http://scp-rhel8-ip/rhel/8/appstream[rhel-8-baseos] name=rhel-8-baseos gpgcheck=0 enabled=1 baseurl=http://scp-rhel8-ip/rhel/8/baseos [rhel-8-baseos-debug] name=rhel-8-baseos-debug gpgcheck=0 enabled=1 baseurl=http://scp-rhel8-ip/rhel/8/baseos-debug [rhel-8-appstream] name=rhel-8-appstream gpgcheck=0 enabled=1 baseurl=http://scp-rhel8-ip/rhel/8/appstreamCode block. Check repo settings (RHEL8) Color mode[rhel-9-for-x86_64-baseos-rpms] name=rhel-9-for-x86_64-baseos-rpms baseurl=http://scp-rhel9-ip/rhel/$releasever/x86_64/baseos gpgcheck=0 enabled=1 [rhel-9-for-x86_64-appstream-rpms] name=rhel-9-for-x86_64-appstream-rpms baseurl=http://scp-rhel9-ip/rhel/$releasever/x86_64/appstream gpgcheck=0 enabled=1 [codeready-builder-for-rhel-9-x86_64-rpms] name=codeready-builder-for-rhel-9-x86_64-rpms baseurl=http://scp-rhel9-ip/rhel/$releasever/x86_64/codeready-builder gpgcheck=0 enabled=1 [rhel-9-for-x86_64-highavailability-rpms] name=rhel-9-for-x86_64-highavailability-rpms baseurl=http://scp-rhel9-ip/rhel/$releasever/x86_64/ha gpgcheck=0 enabled=1 [rhel-9-for-x86_64-supplementary-rpms] name=rhel-9-for-x86_64-supplementary-rpms baseurl=http://scp-rhel9-ip/rhel/$releasever/x86_64/supplementary gpgcheck=0 enabled=1[rhel-9-for-x86_64-baseos-rpms] name=rhel-9-for-x86_64-baseos-rpms baseurl=http://scp-rhel9-ip/rhel/$releasever/x86_64/baseos gpgcheck=0 enabled=1 [rhel-9-for-x86_64-appstream-rpms] name=rhel-9-for-x86_64-appstream-rpms baseurl=http://scp-rhel9-ip/rhel/$releasever/x86_64/appstream gpgcheck=0 enabled=1 [codeready-builder-for-rhel-9-x86_64-rpms] name=codeready-builder-for-rhel-9-x86_64-rpms baseurl=http://scp-rhel9-ip/rhel/$releasever/x86_64/codeready-builder gpgcheck=0 enabled=1 [rhel-9-for-x86_64-highavailability-rpms] name=rhel-9-for-x86_64-highavailability-rpms baseurl=http://scp-rhel9-ip/rhel/$releasever/x86_64/ha gpgcheck=0 enabled=1 [rhel-9-for-x86_64-supplementary-rpms] name=rhel-9-for-x86_64-supplementary-rpms baseurl=http://scp-rhel9-ip/rhel/$releasever/x86_64/supplementary gpgcheck=0 enabled=1Code block. Check repo settings (RHEL9)
- When checking the configuration file, the following result is displayed.
Use a text editor (e.g., vim) to open the
/etc/hostsfile./etc/hostsModify the file with the content below and save.Color mode198.19.2.13 scp-rhel8-ip scp-rhel9-ip scp-rhel-ip198.19.2.13 scp-rhel8-ip scp-rhel9-ip scp-rhel-ipCode block. /etc/hosts file setting change Verify the RHEL Repository connection configured on the server using the yum command.
Color modeyum repolist –vyum repolist –vCode block. repository connection settings check - If the RHEL Repository is successfully connected, you can check the Repository list.Color mode
Repo-id : rhel-8-appstream Repo-name : rhel-8-appstream Repo-revision : 1718903734 Repo-updated : Fri 21 Jun 2024 02:15:34 AM KST Repo-pkgs : 38,260 Repo-available-pkgs: 25,799 Repo-size : 122 G Repo-baseurl : http://scp-rhel8-ip/rhel/8/appstream Repo-expire : 172,800 second(s) (last: Thu 08 Aug 2024 07:27:57 AM KST) Repo-filename : /etc/yum.repos.d/scp.rhel8.repo Repo-id : rhel-8-baseos Repo-name : rhel-8-baseos Repo-revision : 1718029433 Repo-updated : Mon 10 Jun 2024 11:23:52 PM KST Repo-pkgs : 17,487 Repo-available-pkgs: 17,487 Repo-size : 32 G Repo-baseurl : http://scp-rhel8-ip/rhel/8/baseos Repo-expire : 172,800 second(s) (last: Thu 08 Aug 2024 07:27:57 AM KST) Repo-filename : /etc/yum.repos.d/scp.rhel8.repo Repo-id : rhel-8-baseos-debug Repo-name : rhel-8-baseos-debug Repo-revision : 1717662461 Repo-updated : Thu 06 Jun 2024 05:27:41 PM KST Repo-pkgs : 17,078 Repo-available-pkgs: 17,078 Repo-size : 100 G Repo-baseurl : http://scp-rhel8-ip/rhel/8/baseos-debug Repo-expire : 172,800 second(s) (last: Thu 08 Aug 2024 07:27:57 AM KST) Repo-filename : /etc/yum.repos.d/scp.rhel8.repoRepo-id : rhel-8-appstream Repo-name : rhel-8-appstream Repo-revision : 1718903734 Repo-updated : Fri 21 Jun 2024 02:15:34 AM KST Repo-pkgs : 38,260 Repo-available-pkgs: 25,799 Repo-size : 122 G Repo-baseurl : http://scp-rhel8-ip/rhel/8/appstream Repo-expire : 172,800 second(s) (last: Thu 08 Aug 2024 07:27:57 AM KST) Repo-filename : /etc/yum.repos.d/scp.rhel8.repo Repo-id : rhel-8-baseos Repo-name : rhel-8-baseos Repo-revision : 1718029433 Repo-updated : Mon 10 Jun 2024 11:23:52 PM KST Repo-pkgs : 17,487 Repo-available-pkgs: 17,487 Repo-size : 32 G Repo-baseurl : http://scp-rhel8-ip/rhel/8/baseos Repo-expire : 172,800 second(s) (last: Thu 08 Aug 2024 07:27:57 AM KST) Repo-filename : /etc/yum.repos.d/scp.rhel8.repo Repo-id : rhel-8-baseos-debug Repo-name : rhel-8-baseos-debug Repo-revision : 1717662461 Repo-updated : Thu 06 Jun 2024 05:27:41 PM KST Repo-pkgs : 17,078 Repo-available-pkgs: 17,078 Repo-size : 100 G Repo-baseurl : http://scp-rhel8-ip/rhel/8/baseos-debug Repo-expire : 172,800 second(s) (last: Thu 08 Aug 2024 07:27:57 AM KST) Repo-filename : /etc/yum.repos.d/scp.rhel8.repoCode block. Repository list check
- If the RHEL Repository is successfully connected, you can check the Repository list.
Windows Key Management Service Configuration Guide
In Samsung Cloud Platform, when using Windows Server, you can authenticate genuine products by using the Key Management Service provided by SCP. Follow the steps below.
After right-clicking the Windows Start icon, please run cmd from Windows PowerShell (Administrator) or the Windows Run menu.
Windows PowerShell (administrator) or in cmd, please run the command below to register the KMS Server.
Color modeslmgr /skms 198.19.2.23:1688slmgr /skms 198.19.2.23:1688Code block. WKMS Settings After executing the KMS Server registration command, check the notification popup indicating successful registration, then click OK.
Figure. WKMS setting check Windows PowerShell (Administrator) or in cmd, please execute the command below to perform product activation.
Color modeslmgr /atoslmgr /atoCode block. Windows Server activation settings After confirming the notification popup that the product activation was successful, click OK.
Figure. Windows Server genuine activation verification Windows PowerShell (Administrator) or cmd, run the command below to check if it has been activated.
Color modeslmgr /dlvslmgr /dlvCode block. Windows Server genuine activation verification After confirming the notification popup that the product activation was successfully performed, click OK.
Figure. Windows Server genuine activation verification
1.2.7 - Installing ServiceWatch Agent
Users can install the ServiceWatch Agent on a Virtual Server to collect custom metrics and logs.
ServiceWatch Agent
The agents required to collect custom metrics and logs for ServiceWatch on a Virtual Server can be broadly divided into two types: Prometheus Exporter and Open Telemetry Collector.
| Category | Description | |
|---|---|---|
| Prometheus Exporter | Provides metrics of a specific application or service in a format that Prometheus can scrape.
| |
| Open Telemetry Collector | Acts as a centralized collector that gathers telemetry data such as metrics and logs from distributed systems, processes them (filtering, sampling, etc.), and exports them to multiple backends (e.g., Prometheus, Jaeger, Elasticsearch).
|
Prerequisites for Using ServiceWatch Agent
To use the ServiceWatch Agent, please refer to Prerequisite environment configuration for ServiceWatch Agent and prepare the necessary settings.
Installing Prometheus Exporter for Virtual Server (Linux)
Install the Prometheus Exporter on a Linux server following the steps below.
Installing Node Exporter
Install Node Exporter according to the steps below.
Create Node Exporter User
Create a dedicated user to securely isolate the Node Exporter process.
sudo useradd --no-create-home --shell /bin/false node_exportersudo useradd --no-create-home --shell /bin/false node_exporterNode Exporter Configuration
Download Node Exporter for installation. The guide uses the following version:
- Download path: /tmp
- Version: 1.7.0Color mode
cd /tmp wget https://github.com/prometheus/node_exporter/releases/download/v1.7.0/node_exporter-1.7.0.linux-amd64.tar.gzcd /tmp wget https://github.com/prometheus/node_exporter/releases/download/v1.7.0/node_exporter-1.7.0.linux-amd64.tar.gzCode block. Node Exporter download command NoteThe latest Node Exporter version can be found at Node Exporter > Releases > Latest, and specific versions are available at Node Exporter > Releases.
Install the downloaded Node Exporter and set permissions on the executable.
Color modecd /tmp sudo tar -xvf node_exporter-1.7.0.linux-amd64.tar.gz -C /usr/local/bin --strip-components=1 node_exporter-1.7.0.linux-amd64/node_exportercd /tmp sudo tar -xvf node_exporter-1.7.0.linux-amd64.tar.gz -C /usr/local/bin --strip-components=1 node_exporter-1.7.0.linux-amd64/node_exporterCode block. Node Exporter installation command Color modesudo chown node_exporter:node_exporter /usr/local/bin/node_exportersudo chown node_exporter:node_exporter /usr/local/bin/node_exporterCode block. Node Exporter permission setting command Create a service file. Configure Node Exporter to collect memory (meminfo) or block storage (filesystem) metrics.
Color modesudo vi /etc/systemd/system/node_exporter.servicesudo vi /etc/systemd/system/node_exporter.serviceCode block. Open Node Exporter service file Color mode[Unit] Description=Prometheus Node Exporter (meminfo only) Wants=network-online.target After=network-online.target [Service] User=node_exporter Group=node_exporter Type=simple ExecStart=/usr/local/bin/node_exporter \ --collector.disable-defaults \ # Disable default collectors --collector.meminfo \ # Enable memory metrics --collector.filesystem # Enable block storage filesystem metrics Restart=on-failure [Install] WantedBy=multi-user.target[Unit] Description=Prometheus Node Exporter (meminfo only) Wants=network-online.target After=network-online.target [Service] User=node_exporter Group=node_exporter Type=simple ExecStart=/usr/local/bin/node_exporter \ --collector.disable-defaults \ # Disable default collectors --collector.meminfo \ # Enable memory metrics --collector.filesystem # Enable block storage filesystem metrics Restart=on-failure [Install] WantedBy=multi-user.targetCode block. Node Exporter service file content NoteCollectors can be enabled or disabled using flags.
<code>--collector.{name}</code>: Enables a specific metric collector.<code>--no-collector.{name}</code>: Disables a specific metric collector.- To disable all default metrics and enable only specific collectors, use
<code>--collector.disable-defaults --collector.{name} ...</code>.
Below is a description of the main collectors.
Collector Description Labels meminfo Provides memory statistics - filesystem Provides filesystem statistics such as used disk space device: Physical or virtual device path where the filesystem is located (e.g.,/dev/sda1)
fstype: Filesystem type (e.g.,ext4,xfs,nfs,tmpfs)
mountpoint: Path where the filesystem is mounted on the host OS; serves as an intuitive way to distinguish disks (e.g.,/,/var/lib/docker,/mnt/data)
Table. Description of main Node Exporter collectors- For detailed information on available metrics and configuration, see the Node Exporter > Collector page.
- Available metrics may vary depending on the version of Node Exporter you use. See the Node Exporter repository.
- Enable and start the service.
Register the Node Exporter service and verify the registered service and configured metrics.Color mode
sudo systemctl daemon-reload sudo systemctl enable --now node_exportersudo systemctl daemon-reload sudo systemctl enable --now node_exporterCode block. Enable and start Node Exporter service Color modesudo systemctl status node_exportersudo systemctl status node_exporterCode block. Check Node Exporter service status Color modecurl http://localhost:9100/metrics | grep node_memorycurl http://localhost:9100/metrics | grep node_memoryCode block. Verify Node Exporter metrics
See ServiceWatch > Using ServiceWatch Agent for details.
Installing Prometheus Exporter for Virtual Server (Windows)
Install the Prometheus Exporter on a Windows server following the steps below.
Installing Windows Exporter
Install the Windows Exporter according to the steps below.
Windows Exporter Configuration
Download the Windows Exporter installation file.
- Download path: C:\Temp
- Test version: 0.31.3Color mode
$ mkdir /Temp $ Invoke-WebRequest -Uri "https://github.com/prometheus-community/windows_exporter/releases/download/v0.31.3/windows_exporter-0.31.3-amd64.exe" -OutFile "C:\Temp\windows_exporter-0.31.3-amd64.exe"$ mkdir /Temp $ Invoke-WebRequest -Uri "https://github.com/prometheus-community/windows_exporter/releases/download/v0.31.3/windows_exporter-0.31.3-amd64.exe" -OutFile "C:\Temp\windows_exporter-0.31.3-amd64.exe"Code block. Windows Exporter download NoteWindows Exporter versions and installation files can be found at Windows Exporter > Releases.
Test the Windows Exporter execution. By default, Windows Exporter enables all collectors, but to collect only desired metrics, enable the following collectors:
- Memory metrics: memory
- Block storage metrics: logical_disk
- Host name: osColor mode
$ cd C:\Temp $ .\windows_exporter-0.31.3-amd64.exe --collectors.enabled memory,logical_disk,os$ cd C:\Temp $ .\windows_exporter-0.31.3-amd64.exe --collectors.enabled memory,logical_disk,osCode block. Test Windows Exporter execution
Collectors can be enabled using flags.
<code>--collectors.enabled "[defaults]"</code>: Enables the default provided metrics.<code>--collector.enabled {name},{name},{name}...</code>: Enables specific metrics.
Below is a description of the main collectors.
| Collector | Description | Labels |
|---|---|---|
| memory | Provides memory statistics | |
| logical_disk | Collects performance and status metrics of local logical disks (e.g., C:, D: drives) |
|
- For detailed information on available metrics and configuration, see the Windows Exporter > Collector page.
- Available metrics may vary depending on the version of Windows Exporter you use. See the Windows Exporter repository.
Register the service and verify.
Color mode$ sc.exe create windows_exporter binPath= "C:\Temp\windows_exporter-0.31.3-amd64.exe --collectors.enabled memory,logical_disk,os" DisplayName= "Prometheus Windows Exporter" start= auto $ Start-Service windows_exporter$ sc.exe create windows_exporter binPath= "C:\Temp\windows_exporter-0.31.3-amd64.exe --collectors.enabled memory,logical_disk,os" DisplayName= "Prometheus Windows Exporter" start= auto $ Start-Service windows_exporterCode block. Register Windows Exporter service Color mode# Verify service $ Get-Service windows_exporter # Verify metrics $ Invoke-WebRequest -Uri "http://localhost:9182/metrics" | Select-String memory# Verify service $ Get-Service windows_exporter # Verify metrics $ Invoke-WebRequest -Uri "http://localhost:9182/metrics" | Select-String memoryCode block. Verify Windows Exporter service Set configuration file.
- Use the
--config.fileoption to specify a YAML configuration file.Color mode$ .\windows_exporter.exe --config.file=config.yml $ .\windows_exporter.exe --config.file="C:\Program Files\windows_exporter\config.yml" # When using an absolute path, wrap it in quotes$ .\windows_exporter.exe --config.file=config.yml $ .\windows_exporter.exe --config.file="C:\Program Files\windows_exporter\config.yml" # When using an absolute path, wrap it in quotesCode block. Run Windows Exporter with config file Color modecollectors: enabled: cpu,net,service collector: service: include: windows_exporter log: level: warncollectors: enabled: cpu,net,service collector: service: include: windows_exporter log: level: warnCode block. Sample Windows Exporter config
- Use the
- Refer to the official example config file at Windows Exporter > Example config file.
$ sc.exe create windows_exporter binPath= "C:\Temp\windows_exporter-0.31.3-amd64.exe --config.file=C:\Temp\config.yml" DisplayName= "Prometheus Windows Exporter" start= auto
$ Start-Service windows_exporter$ sc.exe create windows_exporter binPath= "C:\Temp\windows_exporter-0.31.3-amd64.exe --config.file=C:\Temp\config.yml" DisplayName= "Prometheus Windows Exporter" start= auto
$ Start-Service windows_exporterSee ServiceWatch > Using ServiceWatch Agent for details.
Node Exporter Metrics
Main Node Exporter metrics
The following are the collector and metric information available through Node Exporter. Collectors can be enabled, and specific metrics can be activated.
| Category | Collector | Metric | Description |
|---|---|---|---|
| Memory | meminfo | node_memory_MemTotal_bytes | Total memory |
| Memory | meminfo | node_memory_MemAvailable_bytes | Available memory (used for determining memory shortage) |
| Memory | meminfo | node_memory_MemFree_bytes | Free memory |
| Memory | meminfo | node_memory_Buffers_bytes | IO buffers |
| Memory | meminfo | node_memory_Cached_bytes | Page cache |
| Memory | meminfo | node_memory_SwapTotal_bytes | Total swap |
| Memory | meminfo | node_memory_SwapFree_bytes | Remaining swap |
| Filesystem | filesystem | node_filesystem_size_bytes | Total filesystem size |
| Filesystem | filesystem | node_filesystem_free_bytes | Total free space |
| Filesystem | filesystem | node_filesystem_avail_bytes | Space actually available to unprivileged users |
Node Exporter collector and metric collection settings
Node Exporter enables most collectors by default, but you can enable or disable specific collectors as needed.
Enable specific collectors only
- When you want to use only memory and filesystem collectors:Color mode
./node_exporter \ --collector.meminfo # Enable memory collector --collector.filesystem # Enable filesystem collector./node_exporter \ --collector.meminfo # Enable memory collector --collector.filesystem # Enable filesystem collectorCode block. Enable specific Node Exporter collectors - When you want to disable all default collectors and use only memory and filesystem collectors:Color mode
./node_exporter \ --collector.disable-defaults # Disable default metrics --collector.meminfo # Enable memory collector --collector.filesystem # Enable filesystem collector./node_exporter \ --collector.disable-defaults # Disable default metrics --collector.meminfo # Enable memory collector --collector.filesystem # Enable filesystem collectorCode block. Enable specific Node Exporter collectors (disable defaults) - Enable filesystem collector for specific mount points:Color mode
./node_exporter \ --collector.disable-defaults \ --collector.filesystem.mount-points-include="/|/data" # Enable filesystem collector for / (root) and /data mount points./node_exporter \ --collector.disable-defaults \ --collector.filesystem.mount-points-include="/|/data" # Enable filesystem collector for / (root) and /data mount pointsCode block. Enable filesystem collector for specific mount points - Enable filesystem collector excluding specific mount points:Color mode
./node_exporter \ --collector.disable-defaults \ --collector.filesystem.mount-points-exclude="/boot|/var/log" # Exclude /boot and /var/log mount points./node_exporter \ --collector.disable-defaults \ --collector.filesystem.mount-points-exclude="/boot|/var/log" # Exclude /boot and /var/log mount pointsCode block. Exclude specific mount points from filesystem collector
Disable specific collectors (no-collector)
When you do not want to use the filesystem collector:
./node_exporter --no-collector.filesystem./node_exporter --no-collector.filesystemConfigure collector as a systemd service (recommended)
[Unit]
Description=Node Exporter
After=network-online.target
[Service]
User=nodeexp
ExecStart=/usr/local/bin/node_exporter \
--collector.disable-defaults # Disable all default metric collectors
--collector.meminfo
--collector.filesystem
[Install]
WantedBy=multi-user.target[Unit]
Description=Node Exporter
After=network-online.target
[Service]
User=nodeexp
ExecStart=/usr/local/bin/node_exporter \
--collector.disable-defaults # Disable all default metric collectors
--collector.meminfo
--collector.filesystem
[Install]
WantedBy=multi-user.targetFilter specific metrics
Using the Open Telemetry Collector configuration, you can select only the required metrics collected by Node Exporter.
For guidance on pre‑configuring the Open Telemetry Collector for ServiceWatch, see Prerequisite Open Telemetry Collector configuration for ServiceWatch.
Windows Exporter Metrics
Main Windows Exporter metrics
The following are the collector and metric information available through Windows Exporter. Collectors can be enabled, and specific metrics can be activated.
| Category | Collector | Metric | Description |
|---|---|---|---|
| Memory | memory | windows_memory_available_bytes | Available memory |
| Memory | memory | windows_memory_cache_bytes | Cached memory |
| Memory | memory | windows_memory_committed_bytes | Committed memory |
| Memory | memory | windows_memory_commit_limit | Commit limit |
| Memory | memory | windows_memory_pool_paged_bytes | Paged pool |
| Memory | memory | windows_memory_pool_nonpaged_bytes | Non‑paged pool |
| Disk | logical_disk | windows_logical_disk_free_bytes | Free space |
| Disk | logical_disk | windows_logical_disk_size_bytes | Total capacity |
| Disk | logical_disk | windows_logical_disk_read_bytes_total | Total read bytes |
| Disk | logical_disk | windows_logical_disk_write_bytes_total | Total written bytes |
| Disk | logical_disk | windows_logical_disk_read_seconds_total | Read latency |
| Disk | logical_disk | windows_logical_disk_write_seconds_total | Write latency |
| Disk | logical_disk | windows_logical_disk_idle_seconds_total | Idle time |
Windows Exporter collector and metric collection settings
Windows Exporter enables most collectors by default, but you can configure only the desired collectors.
Enable specific collectors only
If you want to use only CPU, memory, and logical disk collectors:
# The --collector.enabled option disables defaults and enables only the listed collectors
.\windows_exporter.exe --collectors.enabled="memory,logical_disk"# The --collector.enabled option disables defaults and enables only the listed collectors
.\windows_exporter.exe --collectors.enabled="memory,logical_disk"--collector.enabled will collect only the collectors specified in the option.Configure collector as a service (recommended)
# Register windows_exporter as a service
sc.exe create windows_exporter binPath= "C:\Temp\windows_exporter-0.31.3-amd64.exe --config.file=C:\Temp\config.yml" DisplayName= "Prometheus Windows Exporter" start= auto
# Start the service
Start-Service windows_exporter# Register windows_exporter as a service
sc.exe create windows_exporter binPath= "C:\Temp\windows_exporter-0.31.3-amd64.exe --config.file=C:\Temp\config.yml" DisplayName= "Prometheus Windows Exporter" start= auto
# Start the service
Start-Service windows_exporter# Note this is not an exhaustive list of all configuration values
collectors:
enabled: logical_disk,memory # Set collectors to enable
collector:
service:
include: "windows_exporter"
scheduled_task:
include: /Microsoft/.+
log:
level: debug
scrape:
timeout-margin: 0.5
telemetry:
path: /metrics
web:
listen-address: ":9182"# Note this is not an exhaustive list of all configuration values
collectors:
enabled: logical_disk,memory # Set collectors to enable
collector:
service:
include: "windows_exporter"
scheduled_task:
include: /Microsoft/.+
log:
level: debug
scrape:
timeout-margin: 0.5
telemetry:
path: /metrics
web:
listen-address: ":9182"Filter specific metrics
Using the Open Telemetry Collector configuration, you can select only the required metrics collected by Windows Exporter.
For guidance on pre‑configuring the Open Telemetry Collector for ServiceWatch, see Prerequisite Open Telemetry Collector configuration for ServiceWatch.
1.3 - API Reference
1.4 - CLI Reference
1.5 - Release Note
Virtual Server
- OS Image addition provision
- Standard Image has been added. (Window server 2016)
- SSD volume with configurable IOPS and Throughput has been added.
- You can select SSD_Provisioned disk type when creating Block Storage.
- You can set IOPS and Throughput maximum values.
- You can view Virtual Server ServiceWatch metric monitoring graphs on the detail page.
- OS Image addition provision
- Standard Image has been added. (Alma Linux 9.6, Oracle Linux 9.6, RHEL 9.6, Rocky Linux 9.6)
- New Server Group policy addition
- Partition (Virtual Server and Block Storage distributed placement) policy has been added.
- You can collect custom metrics and logs by installing Virtual Server ServiceWatch Agent.
- You can change the server name on the Virtual Server detail page of Samsung Cloud Platform Console.
- When changing the server name, only the information in Samsung Cloud Platform Console is changed, not the OS’s Hostname.
- ServiceWatch service integration provision
- You can monitor data through ServiceWatch service.
- Virtual Server feature addition
- IP, Public NAT IP, Private NAT IP configuration feature has been added.
- LLM Endpoint for using LLM is provided.
- You can select OS Image subscribed from Marketplace when creating Virtual Server.
- 2nd generation server type has been added.
- 2nd generation (s2) server type based on Intel 4th generation (Sapphire Rapids) Processor has been added. For details, refer to Virtual Server Server Type
- Image sharing method between Accounts has been changed.
- You can share by creating a new qcow2 Image or Image for sharing.
- Virtual Server feature addition
- NAT configuration feature has been added in Virtual Server.
- OS Image addition provision
- Standard Image has been added. (Alma Linux 8.10, Alma Linux 9.4, Oracle Linux 8.10, Oracle Linux 9.4, RHEL 8.10, RHEL 9.4, Rocky Linux 8.10, Rocky Linux 9.4, Ubuntu 24.04)
- Image for Kubernetes has been added. You can create Kubernetes Engine using Image for Kubernetes.
- 2nd generation server type addition
- 2nd generation (h2) server type based on Intel 4th generation (Sapphire Rapids) Processor has been added. For details, refer to Virtual Server Server Type
- Samsung Cloud Platform common function change
- Common CX changes such as Account, IAM and Service Home, tags have been reflected.
- Virtual Server service official release.
- We have released a virtualization server that allows you to freely allocate and use as much as you need at the necessary time without purchasing infrastructure resources individually.
- We have released a virtualization server that allows you to freely allocate and use as much as you need at the necessary time without purchasing infrastructure resources individually.
2 - Virtual Server Auto-Scaling
2.1 - Overview
Service Overview
Virtual Server Auto-Scaling is a service that automatically scales resources up or down according to demand. You can add or terminate the number of servers running the application according to predefined conditions or schedule.
Auto-Scaling Group uses a pre-created Launch Configuration as a pre-configuration template to create servers, and can adjust and manage the number of servers. It adjusts so that the number does not fall below the specified minimum number of servers or exceed the maximum number of servers.
If you register a schedule with Auto-Scaling Group, you can set the number of servers according to the predetermined schedule. If you register a policy, you can increase or decrease the number of servers based on predefined conditions.
Features
Easy and convenient computing environment configuration: Through the web-based Console, users can easily configure the required computing environment themselves via Self Service, from creating Launch Configurations to creating/modifying/deleting Auto-Scaling Groups.
Elastic Resource Usage: You can elastically use computing resources according to the service’s load and usage. Users can schedule resource usage for predictable specific time periods, and can adjust resource usage to prepare for temporary connections from an unspecified large number of users.
Availability Improvement: Virtual Server Auto-Scaling adjusts resources to match variable demand so that the traffic required by the user can always be processed. Through this, users can achieve improved application performance and availability.
Maximizing Cost Reduction Effect: By using resources only as needed according to demand fluctuations, unnecessary costs can be reduced. Through flexible resource usage according to traffic increases or decreases at specific times such as night, weekends, and month-end, the cost reduction effect can be maximized.
Service Architecture Diagram
Provided Features
Virtual Server Auto-Scaling provides the following features.
- Launch Configuration: It is a configuration template used to create a Virtual Server in an Auto-Scaling Group. When creating a Launch Configuration, you set information about the Virtual Server such as image, server type, key pair, block storage, etc.
- Server Count Adjustment: Provides several ways to adjust the number of servers. Using policies, you can add a Virtual Server when the load exceeds a threshold and release the Virtual Server when demand is low, maintaining application availability and reducing costs. You can add and release Virtual Servers according to a schedule, and you can also manually adjust the number of servers in an Auto-Scaling Group as needed.
- Load Balancer integration: You can use a Load Balancer to evenly distribute application traffic to Virtual Server. Whenever a Virtual Server is added or removed, it is automatically registered and deregistered with the Load Balancer.
- Network Connection: You can connect the general subnet of the Auto-Scaling Group, automatic IP allocation, and a Public NAT IP. It provides a local subnet connection for inter-server communication.
- Security Group applied: Security Group is a virtual logical firewall that controls inbound/outbound traffic generated on a Virtual Server. Inbound rules control incoming traffic to the Virtual Server, and Outbound rules control outgoing traffic from the Virtual Server.
- Monitoring: You can view monitoring information such as CPU, Memory, Disk of Virtual Servers created in the Auto-Scaling Group through the Cloud Monitoring service. Based on the monitoring information, you can use Auto-Scaling policies to set thresholds for load, and when the threshold is exceeded, you can add or remove servers.
Components
Virtual Server Auto-Scaling creates an Auto-Scaling Group through Launch Configuration and checks and manages the server.
Launch Configuration
This is a Configuration template used to create a Virtual Server in an Auto-Scaling Group. The main features are as follows.
- Image: Provides OS standard images and user-created custom images. Users can select and use them according to the service they want to configure.
- Keypair: Provides the Keypair method for a secure OS access method.
- Init script: The user can define a script to be executed when the Virtual Server starts.
- For more details, please refer to Launch Configuration Creating.
Auto-Scaling Group
Use Launch Configuration as a pre-configuration template for server creation. You can create an Auto-Scaling Group to adjust and manage the number of servers. The main features are as follows.
- Launch Configuration: It is a configuration template used to create a Virtual Server in an Auto-Scaling Group.
- Server Count Settings: Virtual Server Auto-Scaling provides several ways to adjust the number of servers in an Auto-Scaling Group.
- Fixed Server Count Method: When creating an Auto-Scaling Group, this method keeps the default settings by maintaining the configured number of servers without any added schedules or policies. Refer to Create Auto-Scaling Group and set the Min, Desired, Max server counts.
- Manual Server Count Adjustment Method: In an Auto-Scaling Group, this method increases or decreases the number of servers by modifying the server count to the desired amount. You can choose whether to manually set the desired number of servers. Refer to Modify Server Count.
- Schedule Reservation Method: You can schedule daily, weekly, monthly, or one-time, and set the desired number of servers at a specified time. This is useful when you can predict when to reduce or increase the number of servers. If you use the schedule method, refer to Manage Schedule to add and manage schedules.
- Policy Method: You can use a policy as a way to dynamically adjust servers. When the set threshold based on monitoring metrics is exceeded, it adjusts the number of servers. At this time, you can choose one of three ways to adjust the server count. Increase or decrease the number of servers by a specified number, increase or decrease by a specified ratio, or fix the number of servers to an entered value. When servers start and terminate due to the policy, the monitoring metric CPU usage may temporarily exceed the threshold registered in the policy. However, because this is a temporary moment, a cooldown period is set to avoid judging it as an abnormal situation. If you want to use the policy method, refer to Manage Policies.
- Load Balancer: Whenever a Virtual Server is added or terminated, it automatically connects to and disconnects from the Load Balancer registered in the Auto-Scaling Group.
- Auto-Scaling Group’s Load Balancer related detailed information, please refer to Check Auto-Scaling Group detailed information.
Constraints
The constraints of Virtual Server Auto-Scaling are as follows.
| Category | Description |
|---|---|
| Number of Virtual Servers per Auto-Scaling Group | 50 or less |
| Number of policies per Auto-Scaling Group | 12 or fewer |
| Number of schedules per Auto-Scaling Group | 20 or fewer |
| Number of LB server groups and ports per Auto-Scaling Group | 3 or less |
- If the Image you are using is a discontinued standard Image, Scale-out will not work.
If the Image you are using is a Custom Image, Scale out will continue to work properly even after the version is no longer provided. - Before the end of support for the Image you are using, we recommend replacing the Launch Configuration with the latest version of the Image or a Custom Image.
- For detailed information about the OS Image provided by Virtual Server, see OS Image Provided Versions.
Preliminary Service
This is a list of services that must be pre-configured before creating the service. Please refer to the guide provided for each service for details and prepare in advance.
| Service Category | Service | Detailed Description |
|---|---|---|
| Networking | VPC | A service that provides an independent virtual network in a cloud environment |
| Networking | Security Group | Virtual firewall that controls server traffic |
2.1.1 - Monitoring Metrics
Virtual Server Auto-Scaling is a service provided for Virtual Server targets, providing individual Virtual Server monitoring metrics and monitoring metrics provided by Cloud Monitoring-based policies.
Virtual Server monitoring metrics
The following table shows the monitoring metrics of Virtual Server that can be checked through Cloud Monitoring. For detailed usage of Cloud Monitoring, please refer to the Cloud Monitoring guide.
For Windows OS, memory-related metrics are not provided.
| Performance Item | Detailed Description | Unit |
|---|---|---|
| Memory Total [Basic] | Available memory bytes | bytes |
| Memory Used [Basic] | Currently used memory bytes | bytes |
| Memory Swap In [Basic] | Replaced memory bytes | bytes |
| Memory Swap Out [Basic] | Replaced memory bytes | bytes |
| Memory Free [Basic] | Unused memory bytes | bytes |
| Disk Read Bytes [Basic] | Read bytes | bytes |
| Disk Read Requests [Basic] | Number of Read Requests | cnt |
| Disk Write Bytes [Basic] | Write bytes | bytes |
| Disk Write Requests [Basic] | Write Request Count | cnt |
| CPU Usage [Basic] | 1-minute average system CPU usage rate | % |
| Instance State [Basic] | Instance Status | state |
| Network In Bytes [Basic] | Received bytes | bytes |
| Network In Dropped [Basic] | Received Packet Drop | cnt |
| Network In Packets [Basic] | Received Packet Count | cnt |
| Network Out Bytes [Basic] | Transmission bytes | bytes |
| Network Out Dropped [Basic] | Transmission Packet Drop | cnt |
| Network Out Packets [Basic] | Transmission Packet Count | cnt |
Monitoring metrics provided by Cloud Monitoring-based policies
The following table shows the monitoring metrics provided by the policy of Cloud Monitoring-based Auto-Scaling Group. For more information on policy settings, see Managing Policies.
| Performance Item | Detailed Description | Unit |
|---|---|---|
| CPU Usage [Basic] | 1-minute average system CPU usage rate | % |
| Memory Used [Basic] | Currently used memory bytes | bytes |
| Network In Bytes [Basic] | Received bytes | bytes |
| Network In Packets [Basic] | Number of Received Packets | cnt |
| Network Out Bytes [Basic] | Transmission bytes | bytes |
| Network Out Packets [Basic] | Transmission Packet Count | cnt |
2.1.2 - ServiceWatch 지표
Virtual Server Auto-Scaling is a service offered for Virtual Servers that provides individual Virtual Server monitoring metrics as well as monitoring metrics supplied by ServiceWatch‑based policies.
Virtual Server Monitoring Metrics
The basic metrics provided by Virtual Server can be found in ServiceWatch > Virtual Server Basic Metrics.
ServiceWatch monitoring metrics provided by the Auto-Scaling Group policy
The table below shows the ServiceWatch monitoring metrics provided by the Auto-Scaling Group policy. For detailed information on Auto-Scaling Group policy configuration, see Managing Policies.
| Performance Items | Detailed description | unit |
|---|---|---|
| CPU Usage | CPU usage | Percent |
| Network In Bytes | Bytes received on the network interface | Bytes |
| Network In Packets | Number of packets received on the network interface | Count |
| Network Out Bytes | Data transmitted on the network interface (bytes) | Bytes |
| Network Out Packets | Number of packets transmitted on the network interface | Count |
2.2 - How-to guides
Users can create an Auto-Scaling Group service by entering the required information and selecting detailed options through the Samsung Cloud Platform Console.
Creating an Auto-Scaling Group
You can create an Auto-Scaling Group service through the Samsung Cloud Platform Console.
To create an Auto-Scaling Group, follow these steps:
Click All Services > Compute > Virtual Server menu. It will move to the Virtual Server’s Service Home page.
Click the Auto-Scaling Group menu. It will move to the Auto-Scaling Group list page.
On the Auto-Scaling Group list page, click the Create Auto-Scaling Group button. It will move to the Create Auto-Scaling Group page.
On the Create Auto-Scaling Group page, enter the information required to create the service.
- In the Launch Configuration section, select a Launch Configuration.
- You can create a new Launch Configuration by clicking the Create Launch Configuration button.
- In the Service Information Input section, enter or select the required information.
Category RequiredDetailed Description Auto-Scaling Group Name Required Auto-Scaling Group name - Manage servers of the same type and purpose as a group
Server Name Required Server name within the Auto-Scaling Group - An identifier to distinguish servers created within the Auto-Scaling Group, automatically assigned based on the input server name and sequence
Number of Servers Required Number of servers to create in the Auto-Scaling Group - Enter a value between 0 and 20 (Min≤Desired≤Max)
- Min: Set the minimum number of servers for the Auto-Scaling Group to maintain
- Desired: Set the target number of servers within the Auto-Scaling Group, also meaning the initial number of servers created when the Auto-Scaling Group is created
- Max: Set the maximum number of servers that the Auto-Scaling Group can maintain
- After creating the Auto-Scaling Group, you can modify the settings using the Modify button. For more information, refer to Modifying the Number of Servers
Manual Desired Server Count Setting Optional Choose whether to manually change the Desired server count - After creating the Auto-Scaling Group, you can modify the settings using the Modify button. For more information, refer to Modifying Manual Desired Server Count Setting
Network Settings > Network Settings Required Network settings for the Auto-Scaling Group - Select the desired VPC and general Subnet
- IP can only be automatically generated
- If you select a local Subnet, you can choose the desired local Subnet, and IP can only be automatically generated
Network Settings > Security Group Optional Set a Security Group to allow necessary access - If you don’t set a Security Group, it will follow the default rule (Any/Deny) and block all inbound and outbound traffic
- For Linux servers, allow SSH traffic
- For Windows servers, allow RDP traffic
- After creating the Auto-Scaling Group, you can modify the settings using the Modify button. For more information, refer to Setting Security Group
Load Balancer Optional Connect the Auto-Scaling Group to a Load Balancer - Register the servers in the Auto-Scaling Group as members of the LB server group
- LB server group: Select an existing LB server group in the chosen VPC
- Port: Enter a value between 1 and 65,534
- Click the + button to add an LB server group (up to 3 LB server groups and ports can be added)
- Weighted Round Robin or Weighted Least Connection load balancing LB server groups cannot be selected
- Draining Timeout value: If Draining Timeout is checked as used, set the Draining Timeout value
- Draining Timeout: The time to wait before disconnecting the server from the Load Balancer
- This allows for safe session cleanup, as sessions connected to the server may still exist
- If Load Balancer is not used, Draining Timeout setting is not available
- The default value is 300 seconds, and you can enter a value between 1 second and 3,600 seconds
- Draining Timeout: The time to wait before disconnecting the server from the Load Balancer
- After creating the Auto-Scaling Group, you can modify the settings, and for more information, refer to Using Load Balancer in Auto-Scaling Group
Table. Auto-Scaling Group Service Information Input Items - In the Scaling Policy Settings section, set the scaling policy.
- For more information on policy settings, refer to Adding a Policy.
Category RequiredDetailed Description Set Now Optional Set the scaling policy now - Click the Add Policy button to display the policy information input items
Set Later Optional Set the policy after creating the Auto-Scaling Group, on the detailed information page Table. Auto-Scaling Group Scaling Policy Settings Items
- For more information on policy settings, refer to Adding a Policy.
- In the Notification Settings section, set the notification recipient and method.
- For more information on notification settings, refer to Adding a Notification.
Category RequiredDetailed Description Set Now Optional Set the notification recipient and method now - Click the Add Notification button to open the Add Notification popup window
- For more information on notification settings, refer to the detailed information
- Click the Modify button in the notification recipient list to change the notification information
Set Later Optional Set the notification recipient and method after creating the Auto-Scaling Group, on the detailed information page Table. Auto-Scaling Group Notification Settings Items
- For more information on notification settings, refer to Adding a Notification.
- In the Additional Information Input section, enter or select the required information.
Category RequiredDetailed Description Tag Optional Add a tag - Up to 50 tags can be added per resource
- Click the Add Tag button, then enter or select the Key and Value
Table. Auto-Scaling Group Additional Information Input Items
- In the Launch Configuration section, select a Launch Configuration.
In the Summary panel, review the created details and estimated billing amount, then click the Complete button.
- After creation is complete, you can find the created Auto-Scaling Group on the Auto-Scaling Group list page.
Checking Auto-Scaling Group Details
The Auto-Scaling Group service allows you to view and modify the overall resource list and detailed information. The Auto-Scaling Group details page consists of Details, Policy, Schedule, Virtual Server, Load Balancer, Tag, and Work History tabs.
To check the Auto-Scaling Group details, follow these steps:
- Click All Services > Compute > Virtual Server menu. It will move to the Virtual Server’s Service Home page.
- Click the Auto-Scaling Group menu. It will move to the Auto-Scaling Group list page.
- On the Auto-Scaling Group list page, click the resource you want to check the details for. It will move to the Auto-Scaling Group details page.
- The Auto-Scaling Group details page displays status information and additional feature information, and consists of Details, Policy, Schedule, Virtual Server, Load Balancer, Tag, and Work History tabs.
Category Detailed Description Auto-Scaling Group Status The status of the Auto-Scaling Group created by the user - Creating: Auto-Scaling Group creation in progress
- In Service: Serviceable state
- Scale In: Scale In in progress
- Scale Out: Scale Out in progress
- Cool Down: Cool-down wait in progress
- Terminating: Auto-Scaling Group deletion in progress
- Attach to LB: Connecting to Load Balancer in progress
- Detach from LB: Detaching from Load Balancer in progress
Auto-Scaling Group Deletion Button to delete the Auto-Scaling Group Table. Auto-Scaling Group Status Information and Additional Features
- The Auto-Scaling Group details page displays status information and additional feature information, and consists of Details, Policy, Schedule, Virtual Server, Load Balancer, Tag, and Work History tabs.
Details
… (rest of the content remains the same) Auto-Scaling Group Details page where you can check the detailed information of the selected resource and modify the information if necessary.
| Category | Detailed Description |
|---|---|
| Service | Service name |
| Resource Type | Resource type |
| SRN | Unique resource ID in Samsung Cloud Platform
|
| Resource Name | Resource name
|
| Resource ID | Unique resource ID in the service |
| Creator | User who created the service |
| Creation Time | Time when the service was created |
| Modifier | User who modified the service information |
| Modification Time | Time when the service information was modified |
| Auto-Scaling Group Name | Auto-Scaling Group name |
| Launch Configuration Name | Launch Configuration name selected when creating the Auto-Scaling Group
|
| Number of Servers | Current number of servers in the Auto-Scaling Group and set Min, Desired, Max server numbers
|
| Manual Setting of Desired Server Number | Use or do not use manual setting of Desired server number
|
| VPC | VPC information of the Auto-Scaling Group |
| General Subnet | General Subnet and NAT IP usage information of the Auto-Scaling Group |
| Local Subnet | Local Subnet information of the Auto-Scaling Group |
| Security Group | Security Group information of the Auto-Scaling Group
|
Policy
On the Auto-Scaling Group List page, you can check the policy list of the selected resource and add or manage policies if necessary.
| Category | Detailed Description |
|---|---|
| Category | Policy category
|
| Policy Name | Policy name |
| Execution Condition | Condition for executing the policy
|
| Execution Unit | Method of executing the policy
|
| Cool-down | Waiting time (in seconds) after the policy starts or stops a server
|
| More > Modify | Modify the policy information
|
| More > Activate | Activate the policy
|
| More > Deactivate | Deactivate the policy
|
For more information on policy management and policy examples, see Policy Management.
Schedule
On the Auto-Scaling Group List page, you can check the schedule list of the selected resource and add or manage schedules if necessary.
| Category | Detailed Description |
|---|---|
| Name | Schedule name |
| Min | Minimum server number set in the schedule |
| Desired | Target server number set in the schedule |
| Max | Maximum server number set in the schedule |
| Period | Schedule execution period
|
| Date/Day of the Week | Schedule execution date or day of the week
|
| Execution Time | Schedule execution time |
| Time Zone | Schedule execution time zone |
| Status | Schedule status |
| More > Modify | Modify the schedule information
|
| More > Activate | Activate the schedule
|
| More > Deactivate | Deactivate the schedule
|
For more information on schedule management, see Adding Schedule and Deleting Schedule.
Virtual Server
On the Auto-Scaling Group List page, you can check the Virtual Server list of the selected resource.
| Category | Detailed Description |
|---|---|
| Server Name | Name of the server created in the Auto-Scaling Group
|
| IP | IP assigned to the server |
| Creation Time | Date and time when the server was created |
| Status | Server status
|
| Load Balancer Connection Status | Load Balancer connection status
|
Load Balancer
On the Auto-Scaling Group List page, you can check the Load Balancer list of the selected resource and add or manage Load Balancers if necessary.
| Category | Detailed Description |
|---|---|
| Draining Timeout | Draining Timeout usage
|
| Load Balancer | Load Balancer usage
|
| Load Balancer > Load Balancer Name | Load Balancer name to connect to the Auto-Scaling Group |
| Load Balancer > LB Server Group | Load Balancer’s LB server group
|
| Load Balancer > Port | Port registered as a member of the LB server group |
- The LB server group member information set in the Auto-Scaling Group can also be checked in the Connected Resources of the LB Server Group.
- If manual connection/disconnection between the server and Load Balancer is required, refer to Adding an LB Server Group Member.
Notification
You can check the notification recipient information and notification method for the selected resource on the Auto-Scaling Group List page.
| Category | Detailed Description |
|---|---|
| Notification Recipient | Name of the notification recipient |
| Email of the notification recipient | |
| Server Creation | Whether to send a notification when a server creation-related notification occurs
|
| Server Termination | Whether to send a notification when a server termination-related notification occurs
|
| Policy Execution Condition | Whether to send a notification when the policy execution condition is met |
| Status | Notification activation status
|
| More > Edit | Edit the notification information |
| More > Activate | Activate the notification information
|
| More > Deactivate | Deactivate the notification information
|
For more information on notification settings, refer to Managing Notifications.
Tag
You can check the tag information of the selected resource on the Auto-Scaling Group List page and add, modify, or delete tags.
| Category | Detailed Description |
|---|---|
| Tag List | Tag list
|
Work History
You can check the work history of the selected resource on the Auto-Scaling Group List page.
| Category | Detailed Description |
|---|---|
| Work History List | Resource change history
|
Managing Auto-Scaling Group Resources
If you need to manage the created Auto-Scaling Group, you can perform tasks on the Auto-Scaling Group Details page.
Modifying Launch Configuration
You can modify the Launch Configuration of the Auto-Scaling Group.
To modify the Launch Configuration of the Auto-Scaling Group, follow these steps:
Click All Services > Compute > Virtual Server. The Virtual Server Service Home page opens.
Click Auto-Scaling Group. The Auto-Scaling Group List page opens.
On the Auto-Scaling Group List page, click the resource for which you want to modify the Launch Configuration. The Auto-Scaling Group Details page opens.
Click the Modify button next to the Launch Configuration name. The Modify Launch Configuration popup window opens, where you can view the list of available Launch Configurations.
CategoryDetailed Description Launch Configuration Name Launch Configuration name Image Launch Configuration OS image Server Type Launch Configuration server type Block Storage Launch Configuration Block Storage settings Auto-Scaling Group Count Number of Auto-Scaling Groups to which the Launch Configuration is applied Detailed View Button to view detailed Launch Configuration information Table. Launch Configuration List ItemsIn the Modify Launch Configuration popup window, select the Launch Configuration you want to modify and click OK. The Launch Configuration Modification Notification popup window opens. Check the message and click OK.
Modifying Server Count
You can modify the server count of the Auto-Scaling Group.
To modify the server count of the Auto-Scaling Group, follow these steps:
- Click All Services > Compute > Virtual Server. The Virtual Server Service Home page opens.
- Click Auto-Scaling Group. The Auto-Scaling Group List page opens.
- On the Auto-Scaling Group List page, click the resource for which you want to modify the server count. The Auto-Scaling Group Details page opens.
- Click the Edit Server Count button. The Edit Server Count popup window opens.
- In the Edit Server Count popup window, enter the required items and click the Confirm button.
Classification RequiredDetailed Description Server Count > Min Required Modify the minimum number of servers - Set the minimum number of servers that the Auto-Scaling Group will maintain
Server Count > Desired Required Modify the target server count - Set the target server count in the Auto-Scaling Group
- If Desired Server Count Manual Setting is Not Used, you cannot modify the Desired server count. To modify the Desired server count, refer to Modifying Desired Server Count Manual Setting
Server Count > Max Required Modify the maximum server count - Set the maximum number of servers that the Auto-Scaling Group can maintain
Table. Auto-Scaling Group Server Count Modification Items
Canceling a Virtual Server Created in an Auto-Scaling Group
A Virtual Server created in an Auto-Scaling Group can be canceled by reducing the desired number of servers.
To cancel a Virtual Server created in an Auto-Scaling Group, follow these steps:
- Click All Services > Compute > Virtual Server. You will be taken to the Virtual Server’s Service Home page.
- Click Auto-Scaling Group. You will be taken to the Auto-Scaling Group List page.
- On the Auto-Scaling Group List page, click the resource you want to cancel. You will be taken to the Auto-Scaling Group Details page.
- Click the Edit button in the Server Count section. The Edit Server Count popup window will open.
- In the Edit Server Count popup window, reduce the Desired count and click the Confirm button. The Desired server count will be adjusted, and the Virtual Server will be canceled.
Modifying Desired Server Count Manual Setting
You can change the Desired server count manual setting of the Auto-Scaling Group.
To modify the Desired server count manual setting of the Auto-Scaling Group, follow these steps:
- Click the All Services > Compute > Virtual Server menu. The Virtual Server Service Home page opens.
- Click the Auto-Scaling Group menu. The Auto-Scaling Group List page opens.
- On the Auto-Scaling Group List page, click the resource for which you want to change the Desired server count manual setting. The Auto-Scaling Group Details page opens.
- Click the Edit button for the server count. The Desired Server Count Manual Setting popup window opens.
- In the Desired Server Count Manual Setting popup window, select whether to use it and click the Confirm button.
Setting Security Group
You can set the Security Group for the Auto-Scaling Group.
To set the Security Group for the Auto-Scaling Group, follow these steps:
Click the All Services > Compute > Virtual Server menu. The Virtual Server Service Home page opens.
Click the Auto-Scaling Group menu. The Auto-Scaling Group List page opens.
On the Auto-Scaling Group List page, click the resource for which you want to set the Security Group. The Auto-Scaling Group Details page opens.
Click the Edit button for the Security Group. The Security Group Modification popup window opens, where you can view the list of available Security Groups.
Classification Detailed Description Security Group Name Security Group name Table. Security Group List ItemsIn the Security Group Modification popup window, select the Security Group and click the Confirm button. The Security Group Modification Notification popup window opens. Check the message in the notification popup window and click the Confirm button.
Managing Additional Auto-Scaling Group Information
You can set the Load Balancer to use and select the LB server group for the Auto-Scaling Group. For an Auto-Scaling Group that is using a Load Balancer, you can change it to not use it.
Modifying Load Balancer Draining Timeout
You can set the Load Balancer Draining Timeout for the Auto-Scaling Group.
Draining Timeout is the time to wait before disconnecting the server from the Load Balancer.
- You can set the Draining Timeout to safely clean up sessions, as there may be remaining sessions connected to the server.
- If the Load Balancer is Not Used, the Draining Timeout cannot be set.
- The default value is 300 seconds, and you can set it to a minimum of 1 second and a maximum of 3,600 seconds.
To set the Load Balancer Draining Timeout for the Auto-Scaling Group, follow these steps:
- Click the All Services > Compute > Virtual Server menu. The Virtual Server Service Home page opens.
- Click the Auto-Scaling Group menu. The Auto-Scaling Group List page opens.
- On the Auto-Scaling Group List page, click the resource for which you want to set the Load Balancer Draining Timeout. The Auto-Scaling Group Details page opens.
- Click the Load Balancer tab. The Load Balancer list page opens.
- Click the Edit button for the Draining Timeout. The Draining Timeout Modification popup window opens.
- In the Draining Timeout Modification popup window, select whether to use the Draining Timeout and enter the Draining Timeout time (in seconds).
- In the Draining Timeout Modification popup window, check the input values and click the Confirm button. The Draining Timeout Modification Notification popup window opens. Check the message in the notification popup window and click the Confirm button.
Using Load Balancer
You can modify the Load Balancer for the Auto-Scaling Group. To set the Load Balancer for the Auto-Scaling Group, follow these steps:
- When the Auto-Scaling Group’s server is created, it is automatically connected to the selected Load Balancer’s LB server group as a member, and when the server is terminated, it is disconnected from the LB server group.
- If the Draining Timeout is Used, the server is disconnected from the LB server group after waiting for the Draining Timeout (in seconds).
- For Load Balancer modification, the member is detached from the LB server group and waits in the Detach from LB state. For Scale In, the member is disconnected from the LB server group and waits in the Scale In state.
- Click the All Services > Compute > Virtual Server menu. The Virtual Server Service Home page opens.
- Click the Auto-Scaling Group menu. The Auto-Scaling Group List page opens.
- On the Auto-Scaling Group List page, click the resource for which you want to set the Load Balancer. The Auto-Scaling Group Details page opens.
- Click the Load Balancer tab. The Load Balancer list page opens.
- Click the Edit button for the Load Balancer. The Load Balancer Modification popup window opens.
- In the Load Balancer Modification popup window, select whether to use it. If you select Use, you can select the Load Balancer.
Classification Detailed Description LB Server Group LB server group name - Select the LB server group created in the selected VPC
- LB server groups using Weighted Round Robin or Weighted Least Connection load balancing cannot be selected
Port LB server group port information - Enter the port information required for registering the LB server group member
- Enter a value between 1 and 65,534
Table. Load Balancer List Items- You can add an LB server group by clicking the + button. Up to 3 can be added. You can remove the added Load Balancer by clicking the X button.
- Check the Load Balancer list and click the Confirm button. The Load Balancer Modification Notification popup window opens. Check the message in the notification popup window and click the Confirm button.Caution
- Be cautious when detaching/attaching servers from Load Balancer, as it may affect the service.
- If Draining Timeout is in use, setting Load Balancer to not in use or removing some connected Load Balancers using the X button will not immediately detach the server. The server will be detached from Load Balancer after waiting for the Draining Timeout (seconds). At this time, Auto-Scaling Group will be in Detach from LB state.
- The LB server group member information set in Auto-Scaling Group can also be checked in LB server group connected resources.
- If manual connection/disconnection between server and Load Balancer is required, refer to Registering LB server group members.
Not using Load Balancer
You can modify the Load Balancer of Auto-Scaling Group to not in use. To set Load Balancer to not in use in Auto-Scaling Group, follow the procedure below.
- Be cautious when detaching/attaching servers from Load Balancer, as it may affect the service.
- If Draining Timeout is in use, setting Load Balancer to not in use or removing some connected Load Balancers using the X button will not immediately detach the server. The server will be detached from Load Balancer after waiting for the Draining Timeout (seconds). At this time, Auto-Scaling Group will be in Detach from LB state.
- Click All Services > Compute > Virtual Server menu. Move to the Service Home page of Virtual Server.
- Click Auto-Scaling Group menu. Move to the Auto-Scaling Group list page.
- Click the resource to set Load Balancer in the Auto-Scaling Group list page. Move to the Auto-Scaling Group details page.
- Click the Load Balancer tab. Move to the Load Balancer list page.
- Click the Modify button of Load Balancer. The Load Balancer modification popup window opens.
- Select whether to use Load Balancer in the Load Balancer modification popup window. If you deselect Use, Load Balancer will not be used.
- Confirm the deselection of Use and click the OK button. The Load Balancer modification notification popup window opens. Check the message in the notification popup window and click the OK button.
Deleting Auto-Scaling Group
Deleting unused Auto-Scaling Groups can reduce operating costs. However, deleting an Auto-Scaling Group may immediately stop the service in operation, so you must consider the impact of service termination before proceeding with the deletion.
To delete an Auto-Scaling Group, follow the procedure below.
- Click All Services > Compute > Virtual Server menu. Move to the Service Home page of Virtual Server.
- Click Auto-Scaling Group menu. Move to the Auto-Scaling Group list page.
- Click the resource to delete in the Auto-Scaling Group list page. Move to the Auto-Scaling Group details page.
- Click the Delete Auto-Scaling Group button.
- After deletion is complete, check if the resource has been deleted in the Auto-Scaling Group list page.
2.2.1 - Launch Configuration
To create an Auto-Scaling Group, you need to create a Launch Configuration in advance.
Creating a Launch Configuration
You can create a Launch Configuration service on the Samsung Cloud Platform Console and use it.
To create a Launch Configuration, follow these steps:
Click All Services > Compute > Virtual Server menu. It moves to the Service Home page of Virtual Server.
Click the Launch Configuration menu. It moves to the Launch Configuration list page.
Click the Create Launch Configuration button on the Launch Configuration list page. It moves to the Create Launch Configuration page.
Select the required information in the Image and Version Selection section of the Create Launch Configuration page and click the Next button.
NoteThe images that can be selected in the Launch Configuration are as follows: Virtual Server OS Image Provided Version.Enter the required information in the Service Information Input section of the Create Launch Configuration page.
Category RequiredDescription Launch Configuration Name Required The name of the Launch Configuration - A name to distinguish the Launch Configuration
Service Type > Server Type Required The server type of the Launch Configuration - Standard: Standard specifications commonly used
- High Capacity: Server specifications with higher capacity than Standard
- Refer to Virtual Server Server Type for the available server types
Block Storage Required Block Storage settings according to the purpose of the Launch Configuration - Basic OS: The area where the OS is installed and used
- The capacity is entered in Units, and the minimum capacity varies depending on the OS image type
- Alma Linux: Enter a value between 2 and 1,536
- Oracle Linux: Enter a value between 5 and 1,536
- RHEL: Enter a value between 2 and 1,536
- Rocky Linux: Enter a value between 2 and 1,536
- Ubuntu: Enter a value between 1 and 1,536
- Windows: Enter a value between 4 and 1,536
- SSD: High-performance general volume
- HDD: General volume
- SSD/HDD_KMS: Additional encrypted volume using Samsung Cloud Platform KMS (Key Management System) encryption key
- Encryption can only be applied when created, and cannot be changed after creation
- Using the SSD_KMS disk type may cause performance degradation
- The capacity is entered in Units, and the minimum capacity varies depending on the OS image type
- Add: Additional user space outside the OS area
- Select Use and enter the storage type and capacity
- Click the + button to add storage, and click the x button to delete (up to 25 can be added)
- Capacity is entered in Units, and enter a value between 1 and 1,536
- 1 Unit is 8GB, so 8 to 12,288GB is created
- SSD: High-performance general volume
- HDD: General volume
- SSD/HDD_KMS: Additional encrypted volume using Samsung Cloud Platform KMS (Key Management System) encryption key
- Encryption can only be applied when created, and cannot be changed after creation
- Using the SSD_KMS disk type may cause performance degradation
- SSD_MultiAttach: Volume that can be connected to two or more servers
- For more information on Block Storage types, refer to Creating Block Storage
Keypair Required Select the authentication method for the Launch Configuration - Server authentication information to access the server created by creating an Auto-Scaling Group with the Launch Configuration
- Create New: Create a new Keypair if needed
- Refer to Creating a Keypair for how to create a new Keypair
- Default access account list by OS
- Alma Linux: almalinux
- RHEL: cloud-user
- Rocky Linux: rocky
- Ubuntu: ubuntu
- Windows: sysadmin
Table. Launch Configuration Service Information Input ItemsEnter the information in the Additional Information Input section of the Create Launch Configuration page and click the Next button.
Category RequiredDescription Init Script Optional Script that runs when the server starts using the Launch Configuration - Enter within 45,000 bytes
- The Init Script must be a batch script for Windows or a shell script or cloud-init for Linux, depending on the selected image.
Tag Optional Add a tag - Up to 50 tags can be added per resource
- Click the Add Tag button, enter the Key and Value, or select them
Table. Launch Configuration Additional Information Input ItemsCheck the input information and estimated cost on the Create Information Confirmation page, and click the Complete button.
- After creation is complete, check the created Launch Configuration on the Launch Configuration list page.
Checking Launch Configuration Details
The Launch Configuration service allows you to check the overall resource list and detailed information, and modify it. The Launch Configuration details page consists of Details, Tags, and Work History tabs.
To check the Launch Configuration details, follow these steps:
- Click All Services > Compute > Virtual Server menu. It moves to the Service Home page of Virtual Server.
- Click the Launch Configuration menu. It moves to the Launch Configuration list page.
- Click the resource to check the details on the Launch Configuration list page. It moves to the Launch Configuration details page.
- The top of the Launch Configuration details page displays status information and additional feature information, and consists of Details, Tags, and Work History tabs.
| Category | Description |
|---|---|
| Launch Configuration Status | The status of the Launch Configuration created by the user
|
| Launch Configuration Deletion | Button to delete the Launch Configuration |
Details
You can check and modify the detailed information of the selected resource on the Launch Configuration list page.
| Category | Description |
|---|---|
| Service | Service category |
| Resource Type | Service name |
| SRN | Unique resource ID in Samsung Cloud Platform
|
| Resource Name | Resource name
|
| Resource ID | Unique resource ID in the service |
| Creator | The user who created the service |
| Creation Time | The time when the service was created |
| Modifier | The user who modified the service information |
| Modification Time | The time when the service information was modified |
| Launch Configuration Name | Launch Configuration name |
| Image | The image name selected when creating the Launch Configuration
|
| Number of Auto-Scaling Groups | The number of Auto-Scaling Groups using the Launch Configuration |
| Server Type | The server type set in the Launch Configuration |
| Block Storage | Block Storage information set in the Launch Configuration
|
| Keypair | Server authentication information set in the Launch Configuration
|
| Init Script | Init Script set in the Launch Configuration
|
Tags
You can check the tag information of the selected resource on the Launch Configuration list page, and add, change, or delete it.
| Category | Description |
|---|---|
| Tag List | Tag list
|
Work History
You can check the work history of the selected resource on the Launch Configuration list page.
| Category | Description |
|---|---|
| Work History List | Resource change history
|
Deleting a Launch Configuration
You can reduce operating costs by deleting unused Launch Configurations. However, deleting a Launch Configuration may immediately stop the operating service, so you should consider the impact of stopping the service before proceeding with the deletion.
To delete a Launch Configuration, follow these steps:
- Click All Services > Compute > Virtual Server menu. It moves to the Service Home page of Virtual Server.
- Click the Launch Configuration menu. It moves to the Launch Configuration list page.
- Click the resource to delete on the Launch Configuration list page. It moves to the Launch Configuration details page.
- Click the Delete Launch Configuration button.
- After deletion is complete, check that the resource has been deleted on the Launch Configuration list page.
2.2.2 - Managing Policies
The number of servers in an Auto-Scaling Group can be dynamically adjusted based on monitoring metrics. When the set threshold is exceeded based on the monitoring metrics, the number of servers is adjusted. At this time, you can choose one of three ways to adjust the number of servers: increase or decrease the number of servers by a specified number, increase or decrease the number of servers by a specified ratio, or fix the number of servers to a specified value. When a server is started or terminated due to a policy, the monitoring metric, such as CPU usage, may temporarily exceed the threshold set in the policy. However, since this is a temporary moment, a cooldown time is set so that it is not judged as an abnormal situation. You can add and manage policies for an Auto-Scaling Group created in the Samsung Cloud Platform Console.
Adding a Policy
You can add a policy to an Auto-Scaling Group. To add a policy to an Auto-Scaling Group, follow these steps:
All Services > Compute > Virtual Server menu, click. Move to the Service Home page of Virtual Server.
Click the Auto-Scaling Group menu. Move to the Auto-Scaling Group List page.
On the Auto-Scaling Group List page, click the resource to view detailed information. Move to the Auto-Scaling Group Details page.
Click the Policy Tab. Move to the Policy Tab page.
Click the Add Policy button. The Add Policy popup window opens.
Classification RequiredDetailed Description Classification Required Policy classification - Scale In: Server reduction
- Scale Out: Server increase
Policy Name Required Policy name for distinction Execution Condition Required Condition for executing the policy - Statistic: Method of calculating the metric type
- Average: Average of servers in the Auto-Scaling Group
- Min: Minimum value among servers in the Auto-Scaling Group
- Max: Maximum value among servers in the Auto-Scaling Group
- Metric Type: CPU Usage, Memory Usage, Network In(bytes), Network Out(bytes), Network In(Packets), Network Out(Packets)
- Note: Memory usage policy is not available for Windows servers
- Operator:
>=><=<
- Threshold: Threshold for the metric type
- Period: Continuous time to trigger the execution condition (N minutes)
Execution Unit Required Method of executing the policy - Policy Type: Select the type of policy to execute.
- Increase or decrease the number of servers by a specified number: Increase or decrease the target value
- Increase or decrease the number of servers by a specified ratio: Increase or decrease the target value ratio
- Fix the number of servers to a specified value: Fix the target value
- Target Value: Number or ratio to execute the selected Policy Type
Cooldown Required Time to wait (in seconds) when a server is started or terminated due to a policy - Default value is 300 seconds, and it can be set between 60 seconds and 3,600 seconds.
Table. Add Policy Popup ItemsNotePolicy > Cooldown Setting
- When a server is started or terminated due to a policy, wait for the cooldown time set. The monitoring metric, such as CPU usage, may temporarily exceed the threshold set in the policy. However, since this is a temporary moment and not a condition for adjusting the number of servers, the cooldown time is set to wait.
GuidePolicy execution operates within the set minimum and maximum number of servers.
- Even if the number of servers is increased or decreased, or fixed, beyond the minimum and maximum number of servers, it operates within the set minimum and maximum number of servers.
- Example: If the minimum number of servers is 3, even if the number of servers is fixed to 1, the number of servers will not decrease to 1, but will be maintained at the minimum number of servers, which is 3.
In the Add Policy popup window, enter the required values and click the Confirm button. The added policy can be checked in the Policy List.
Policy Creation Example
The following is an explanation of the policy example. Refer to it when creating a policy.
| Classification | Execution Condition | Execution Unit | Cooldown |
|---|---|---|---|
| Scale Out | Average CPU Usage >= 60% for 1 minute | Increase the number of servers by a specified number, 1 server | 300 seconds |
- If the average CPU usage of the servers in the Auto-Scaling Group is 60% or higher for 1 minute, 1 server is added.
- When a server is added, the cooldown time is 300 seconds. During the cooldown time, no additional servers are added or terminated due to the policy.
- After the cooldown time ends, the policy execution condition is checked again.
| Classification | Execution Condition | Execution Unit | Cooldown |
|---|---|---|---|
| Scale In | Min CPU Usage <= 5% for 1 minute | Decrease the number of servers by a specified ratio, 50% | 300 seconds |
- If the minimum CPU usage of the servers in the Auto-Scaling Group is 5% or lower for 1 minute, 50% of the current number of servers are terminated.
- When a server is terminated, the cooldown time is 300 seconds. During the cooldown time, no additional servers are added or terminated due to the policy.
- After the cooldown time ends, the policy execution condition is checked again.
| Classification | Execution Condition | Execution Unit | Cooldown |
|---|---|---|---|
| Scale Out | Max CPU Usage >= 90% for 1 minute | Fix the number of servers to a specified value, 5 servers | 300 seconds |
- If the maximum CPU usage of the servers in the Auto-Scaling Group is 90% or higher for 1 minute, the number of servers is fixed to 5.
- During the server creation, the cooldown time is 300 seconds. During the cooldown time, no additional servers are added or terminated due to the policy.
- After the cooldown time ends, the policy execution condition is checked again.
Modifying a Policy
You can modify a policy of an Auto-Scaling Group. To modify a policy of an Auto-Scaling Group, follow these steps:
All Services > Compute > Virtual Server menu, click. Move to the Service Home page of Virtual Server.
Auto-Scaling Group menu should be clicked. It moves to the Auto-Scaling Group list page.
In the Auto-Scaling Group list page, click on the resource to check the detailed information. It moves to the Auto-Scaling Group details page.
Click on the Policy tab. It moves to the Policy tab page.
Click on the More > Edit button of the policy to be modified. The Policy modification popup opens.
Classification RequiredDetailed Description Classification Required Policy classification - Scale In: Server count return
- Scale Out: Server count increase
Policy Name Required Policy name for distinction Execution Condition Required Condition for executing the policy - Statistic: Method of calculating Metric Type
- Average: Average of servers in Auto-Scaling Group
- Min: Minimum value among servers in Auto-Scaling Group
- Max: Maximum value among servers in Auto-Scaling Group
- Metric Type: CPU Usage, Memory Usage, Network In(bytes), Network Out(bytes), Network In(Packets), Network Out(Packets)
- Note: Memory usage policy cannot be set for Windows servers
- Operator:
>=><=<
- Threshold: Threshold corresponding to Metric Type
- Period: Continuous time (N minutes) to trigger the execution condition
Execution Unit Required Method of executing the policy - Policy Type: Select the type of policy to be executed.
- Increase or decrease the server count by a specified number: Increase or decrease the server count by the target value
- Increase or decrease the server count by a specified ratio: Increase or decrease the server count by the target value ratio
- Fix the server count to the input value: Fix the server count to the target value
- Target Value: The number or ratio of the selected Policy Type to be executed
Cooldown Required Waiting time (in seconds) when a server is started or terminated due to a policy - Default value is 300 seconds, and it can be set from 1 second to a maximum of 3,600 seconds
Table. Policy modification popup itemsClick the Confirm button after entering the required values in the Policy modification popup window.
Policy Addition and Modification Restrictions
There are restrictions when adding or modifying policies, depending on the policy classification, execution condition, and execution condition range. Refer to the examples of restrictions below and add or modify policies accordingly.
Example 1 - Check for duplicate registration of policy classification and execution condition
Duplicate registration is not allowed when the policy classification (Scale Out or Scale In) and execution condition (Metric type) are the same.
| Policy Classification | Policy Name | Execution Condition (Statistic) | Execution Condition (Metric Type) | Execution Condition Range |
|---|---|---|---|---|
Scale Out | ScaleOutPolicy | Average | CPU Usage | >= 60% |
If a policy is already registered as shown above, it is not possible to add or modify a policy with the same classification (Scale Out) and execution condition (Metric type = CPU Usage).
Example 2 - Check the execution condition range for the same execution condition (Metric type) and execution condition (Statistic)
When the policy distinction (Scale Out or Scale In) is different, the execution condition range (Comparison operator + Threshold) cannot be duplicated for the same execution condition (Metric type) and execution condition (Statistic).
| Policy Distinction | Policy Name | Execution Condition (Statistic) | Execution Condition (Metric type) | Execution Condition Range |
|---|---|---|---|---|
Scale Out | ScaleOutPolicy | Average | CPU Usage | >= 60% |
In the case where a policy is registered as above, it is not possible to add a policy or modify it as follows: If the CPU Usage is 60% or higher on average, since the Scale Out policy is already registered, it is not possible to register a Scale In policy for CPU Usage average of 60% or lower, as the 60% case would be a duplicate of the same execution condition.
| Policy Distinction | Policy Name | Execution Condition (Statistic) | Execution Condition (Metric type) | Execution Condition Range |
|---|---|---|---|---|
Scale In | AddUpdatePolicy | Average | CPU Usage | <= 60% |
If a policy is already registered as shown above, it is not possible to add or modify a policy with the same execution condition (Metric type = CPU Usage) and execution condition (Statistic = Average), and an execution condition range that overlaps with the existing policy.
Example 3 - Check the execution condition range for the same execution condition (Metric type) and execution condition (Statistic)
When the policy distinction (Scale Out or Scale In) is different, the execution condition range (Comparison operator + Threshold) cannot be duplicated for the same execution condition (Metric type) and execution condition (Statistic).
| Policy Distinction | Policy Name | Execution Condition (Statistic) | Execution Condition (Metric type) | Execution Condition Range |
|---|---|---|---|---|
Scale In | ScaleInPolicy | Average | CPU Usage | <= 10% |
In the case where a policy is registered as above, it is not possible to add or modify a policy as follows: Since the Scale In policy is already registered when the CPU usage is 10% or less on average, it is not possible to register a Scale Out policy when the CPU usage is less than 60%, less than or equal to 60%, greater than 10%, or greater than 9%.
| Policy Distinction | Policy Name | Execution Condition (Statistic) | Execution Condition (Metric type) | Execution Condition Range |
|---|---|---|---|---|
Scale Out | AddUpdatePolicy1 | Average | CPU Usage | < 60% |
Scale Out | AddUpdatePolicy2 | Average | CPU Usage | <= 60% |
Scale Out | AddUpdatePolicy3 | Average | CPU Usage | >= 10% |
Scale Out | AddUpdatePolicy4 | Average | CPU Usage | > 9% |
Example 4 - Registration is possible when the execution condition range does not overlap
When the policy distinction (Scale Out or Scale In) is different, it is possible to register even if the execution condition (Statistic) is different or the execution condition range (Comparison operator + Threshold) does not overlap for the same execution condition (Metric type).
| Policy Distinction | Policy Name | Execution Condition (Statistic) | Execution Condition (Metric type) | Execution Condition Range |
|---|---|---|---|---|
Scale Out | ScaleOutPolicy | Average | CPU Usage | >= 60% |
In the case where a policy is registered as above, it is possible to add or modify a policy as follows. If the execution condition range does not overlap or the execution condition (Statistic) is different, registration is possible.
| Policy Distinction | Policy Name | Execution Condition (Statistic) | Execution Condition (Metric type) | Execution Condition Range |
|---|---|---|---|---|
Scale In | AddUpdatePolicy1 | Average | CPU Usage | <= 10% |
Scale In | AddUpdatePolicy2 | Min | CPU Usage | <= 60% |
Deleting a Policy
It is possible to delete a policy from an Auto-Scaling Group. To delete a policy, follow the procedure below.
- All Services > Compute > Virtual Server menu, click. Move to the Service Home page of Virtual Server.
- Auto-Scaling Group menu, click. Move to the Auto-Scaling Group List page.
- On the Auto-Scaling Group List page, click the resource to check the detailed information. Move to the Auto-Scaling Group Details page.
- Click the Policy Tab. Move to the Policy Tab page.
- Select the policy to delete and click the Delete button. The Policy Delete Confirmation popup window opens.
- Confirm the Policy Delete Confirmation popup window and click the Confirm button.
2.2.3 - Managing Schedules
You can schedule daily, weekly, monthly, or one-time, and set the desired number of servers at a fixed time. This is useful when it is possible to predict when to reduce or increase the number of servers.
Add schedule
You can add a schedule to the Auto-Scaling Group. To add a schedule to the Auto-Scaling Group, follow these steps.
- Click All Services > Compute > Virtual Server menu. It moves to the Service Home page of Virtual Server.
- Auto-Scaling Group menu is clicked. It moves to the Auto-Scaling Group list page.
- Auto-Scaling Group list page, click the resource to check the detailed information. It moves to the Auto-Scaling Group details page.
- Click the Schedule Tab. It moves to the Schedule Tab page.
- Click the Add Schedule button. The Add Schedule popup window opens.
Classification RequiredDetailed Description Schedule Name Required Schedule-specific distinguishing name Server count selection Required When performing a schedule, select the number of servers to adjust - Min: The minimum number of servers that the Auto-Scailg Group will maintain
- Desired: The target number of servers within the Auto-Scailg Group
- Max: The maximum number of servers that the Auto-Scailg Group can maintain
Enter the number of servers Required Enter the value of the selected server number - Min value: Please enter a value between 0 and 50. (Min≤Desired≤Max)
- Desired value: Please enter a value between 0 and 50. (Min≤Desired≤Max)
- Max value: Please enter a value between 0 and 50. (Min≤Desired≤Max)
Period Required Schedule execution period - Daily: You can set the start date and end date, and permanent settings for daily schedule execution. You can also set time and time zone
- Weekly: You can set start date and end date, permanent settings, and time and time zone settings. You can also select the day of the week for weekly schedule execution
- Monthly: You can set start date and end date, permanent settings, and time and time zone settings. You can also enter the date for monthly schedule execution
- Once: You can set time and time zone settings. You can also set the date for one-time schedule execution
Start Date Select Set schedule start date - Cannot be set to a date prior to the current date. The default is the current date.
End Date Select Set schedule end date - Cannot be set to a date prior to the current date. The default is the current date + 7.
Permanent Select If permanent is set, the schedule end date is set to 9999-12-31 Time Required Schedule execution time setting - Can be set in 30-minute units. Time faster than the current date and time cannot be set
Time Zone Required Time zone corresponding to the schedule execution time (e.g., Asia/Seoul (GMT +09:00)) Day of the week Required If you select cycle as every week, select the day of the week to perform the schedule Date Essential - Cycle is selected as every month, enter the Date to perform the schedule
- Please enter one or more between -31 and 31, excluding 0. (Example: 3,4,5)
- Cycle is selected as once, set the Date to perform the schedule
- It cannot be set before the current date. The default value is the current date.
Table. Schedule Add Popup Item - In the Add Schedule popup window, enter the required values and click the OK button.
- Check the message in the Add Schedule Confirmation popup window, then click the Confirm button.
If you select monthly for the schedule cycle, you must enter the schedule execution date, which is Date. Please refer to the following contents to register the schedule.
- If you enter a number greater than 0, it means the date of the month.
- Example: If you enter 1, it will be August 1, September 1, …, December 1
- If you enter a number less than 0, it will be calculated from the last day of the month.
- Entering -1 means the last day of the month.
- Example: August 31, September 30, …, December 31
- If -2 is entered, it means the day before the last day of the month.
- Example: August 30, September 29, …, December 30
- Since the last day of each month is different, such as 31st, 30th, 29th, 28th, to handle cases where a schedule should be executed on the last day of each month, negative numbers are used to calculate from the last day, as shown above.
- When the schedule is executed, if the minimum number of servers set in the schedule is greater than the desired number of servers, or the maximum number of servers is less than the desired number of servers, the desired number of servers is also modified.
- If there are schedules with overlapping execution times, they may not run normally. Please try to avoid overlapping execution times if possible.
Modify Schedule
You can modify the schedule of the Auto-Scaling Group. To modify the schedule of the Auto-Scaling Group, follow these steps.
Click All services > Compute > Virtual Server menu. It moves to the Service Home page of Virtual Server.
Auto-Scaling Group menu is clicked. It moves to the Auto-Scaling Group list page.
Auto-Scaling Group list page, click the resource to check the detailed information. Move to the Auto-Scaling Group details page.
Click the Schedule Tab. It moves to the Schedule Tab page.
Click the More > Edit button of the schedule you want to modify. The Edit Schedule popup window will open.
Classification RequiredDetailed Description Schedule Name Required Schedule name to distinguish by schedule Number of servers to select Required When performing a schedule, select the number of servers to adjust - Min: The minimum number of servers that the Auto-Scailg Group will maintain
- Desired: The target number of servers within the Auto-Scailg Group
- Max: The maximum number of servers that the Auto-Scailg Group can maintain
Enter the number of servers Required Enter the value of the selected server number - Min value: Please enter a value between 0 and 50. (Min≤Desired≤Max)
- Desired value: Please enter a value between 0 and 50. (Min≤Desired≤Max)
- Max value: Please enter a value between 0 and 50. (Min≤Desired≤Max)
Period Required Schedule execution period - Daily: You can set the start date and end date, and permanent settings for daily schedule execution. You can also set time and time zone
- Weekly: You can set start date and end date, permanent settings, and time and time zone. You can also select the day of the week for weekly schedule execution
- Monthly: You can set start date and end date, permanent settings, and time and time zone. You can also enter the date for monthly schedule execution
- Once: You can set time and time zone. You can also set the date for one-time schedule execution
Start Date Select Set schedule start date - Cannot be set to a date prior to the current date. The default is the current date.
End Date Select Set schedule end date - Cannot be set to a date prior to the current date. The default is the current date + 7.
Permanent Select If permanent is set, the schedule end date is set to 9999-12-31 Time Required Schedule execution time setting - Can be set in 30-minute units. Time faster than the current date and time cannot be set
Time Zone Required Time zone corresponding to the schedule execution time (e.g. Asia/Seoul (GMT +09:00)) Day Required If you select Weekly as the Cycle, select the day of the week to perform the schedule Date Essential - When Cycle is set to Every Month, enter the Date to perform the schedule
- Please enter one or more values from -31 to 31, excluding 0. (Example: 3,4,5)
- When Cycle is set to Once, set the Date to perform the schedule
- Cannot be set to a date before the current date. The default value is the current date.
Fig. Schedule modification popup itemIn the Modify Schedule popup window, enter the required values and click the Confirm button.
Schedule Modification Confirmation Check the message in the popup window and click the Confirm button.
Delete Schedule
You can delete the schedule of the Auto-Scaling Group. To delete the schedule of the Auto-Scaling Group, follow the next procedure.
- Click All Services > Compute > Virtual Server menu. It moves to the Service Home page of Virtual Server.
- Auto-Scaling Group menu is clicked. It moves to the Auto-Scaling Group list page.
- Auto-Scaling Group list page, click the resource to check the detailed information. Move to the Auto-Scaling Group details page.
- Click the Schedule Tab. It moves to the Schedule Tab page.
- Select the schedule to be deleted and click the Delete button. The Schedule Deletion Confirmation popup window will open.
- Schedule deletion confirmation popup window, confirm and click the Confirm button.
2.2.4 - Managing Notifications
You can specify the notification recipient to send a notification message via E-mail or SMS for a specific situation.
- Notification method (E-mail or SMS) can be set by selecting Notification target as Service > Virtual Server Auto-Scaling on the Notification settings page.
- For more information about modifying alert settings, see Modifying Alert Settings.
Add Alert
You can add notifications to the Auto-Scaling Group. To add notifications to the Auto-Scaling Group, follow these steps.
- All services > Compute > Virtual Server menu, click. It moves to the Service Home page of Virtual Server.
- Auto-Scaling Group menu is clicked. It moves to the Auto-Scaling Group list page.
- Click the resource to add notification information on the Auto-Scaling Group list page. It moves to the Auto-Scaling Group details page.
- Click the Notification Tab. You will be taken to the Notification Tab page.
- Click the Add Notification button. The Add Notification popup window opens.
- In the Add Notification popup window, enter the required values and click the Confirm button.
Division Detailed Description Notification Point Notification point when Auto-Scaling Group alert occurs - Server creation, Server termination, Server creation failure, Server termination failure, When policy execution conditions are met
- Multiple selections are possible
Notification Recipient User to receive notification when notification occurs - Add Notification Recipient button to select user
- Only Samsung Cloud Platform users can be selected as recipients
Table. Notification Items
- Check the message in the Add Alert Confirmation popup window, then click the Confirm button.
Modify Alert
You can modify the notification information of the Auto-Scaling Group. To modify the notification information of the Auto-Scaling Group, follow the procedure below.
- Click All Services > Compute > Virtual Server menu. It moves to the Service Home page of Virtual Server.
- Auto-Scaling Group menu will be clicked. It moves to the Auto-Scaling Group list page.
- Click the resource to modify the notification information on the Auto-Scaling Group list page. It moves to the Auto-Scaling Group details page.
- Click the Notification Tab. It moves to the Notification Tab page.
- Click the More > Edit button for the notification information you want to modify in the notification list. The Edit Notification popup window opens.
- Modify Notification In the notification modification popup window, modify the notification information and click the Confirm button.
Classification Detailed Description Notification Point Notification point when Auto-Scaling Group alert occurs - Server creation, Server termination, Server creation failure, Server termination failure, When policy execution conditions are met
- Multiple selections are possible
Table. Notification Modification Items - Check the message in the Notification Modification Confirmation popup window, then click the Confirm button.
Delete Notification
You can delete the notification of Auto-Scaling Group. To delete the notification of Auto-Scaling Group, follow the procedure below.
- All Services > Compute > Virtual Server menu should be clicked. It moves to the Service Home page of Virtual Server.
- Auto-Scaling Group menu is clicked. It moves to the Auto-Scaling Group list page.
- Auto-Scaling Group list page, click the resource to modify the notification information. Move to the Auto-Scaling Group details page.
- Click the Notification Tab. It moves to the Notification Tab page.
- Select the notification to be deleted from the notification list, then click the Delete button. The Delete Notification Confirmation popup window will open.
- Notification deletion confirmation popup window, confirm and click the Confirm button.
2.3 - API Reference
2.4 - CLI Reference
2.5 - Release Note
Virtual Server Auto-Scaling
- Added notification feature to Virtual Server Auto-Scaling.
- You can add notification settings in the Auto-Scaling Group creation or detail screen.
- You can set the scaling policy when creating an Auto-Scaling Group.
- Added Metric Type of Auto-Scaling Group policy.
- Added: Memory Usage, Network In(bytes), Network Out(bytes), Network In(Packets), Network Out(Packets)
- You can set the Draining Timeout when connecting to the Load Balancer.
- In an Auto-Scaling Group, a Virtual Server can be connected to up to 50 instances, and an LB server group and port can be connected up to 3 instances.
- Virtual Server Auto-Scaling feature change
- It will be released in conjunction with the Load Balancer service to be released in February 2025.
- NAT setting feature has been added to Auto-Scaling Group.
- Samsung Cloud Platform common feature changes
- Account, IAM and Service Home, tags, etc. reflected common CX changes.
- Virtual Server Auto-Scaling creates and manages Auto-Scaling Group through Launch Configuration and checks and manages the server.
- It provides a schedule method that can set the desired number of servers at a fixed time and a policy method that adjusts the number of servers based on CPU usage rate.
3 - GPU Server
3.1 - Overview
Service Overview
GPU Server is a virtualized computing service that allows you to freely allocate and use as much infrastructure resources provided by the server such as CPU, GPU, and memory as needed at the desired time without having to purchase them individually. It is suitable for tasks that require fast computing speed such as AI model experimentation, prediction, and inference in a cloud environment, and allows you to flexibly select and use resources with optimized performance according to task type and scale. GPU Server provides the following features:
Provided Features
- GPU Server Management: Users can directly manage creation, deletion, and changes as a Self Service from GPU Server provisioning to monitoring and billing through a web-based Console.
- Provisioning by GPU Quantity: You can configure a virtual server by freely selecting the quantity of H100/A100 GPUs according to project purpose and scale.
- High Performance GPU Provision: Provides high-performance GPU servers at the physical server level using the Pass-through method.
- Storage Connection: Provides additional connected storage besides OS disks. You can connect and use Block Storage, File Storage, and Object Storage.
- Strong Security Application: Protects servers safely by controlling Inbound/Outbound traffic exchanged with external internet or other VPCs (Virtual Private Cloud) through the Security Group service.
- Monitoring: You can check monitoring information such as CPU, Memory, Disk, and GPU status corresponding to computing resources through the Cloud Monitoring service.
- Network Setting Management: The server’s subnet/IP can be easily changed from the values set at initial creation. Provides management functionality that allows you to set use/terminate NAT IP as needed.
- Key Pair Method: Provides a Key Pair method instead of ID/PW access for secure OS access.
- Image Management: You can create and manage Custom Images and provides sharing functionality between projects.
- ServiceWatch Service Integration Provision: You can monitor data through the ServiceWatch service.
Components
GPU Server provides GPU, NVSwitch, and NVLink on top of virtualized computing resources.
- NVSwitch can be activated and used only for instance types that allocate 8 GPUs to a single GPU Server.
Specifications by GPU Type
GPU (Graphic Processing Unit) plays the role of performing calculations necessary to create images that make up the computer screen and is specialized for parallel processing, enabling it to process large amounts of data quickly, handling large-scale parallel operations such as artificial intelligence (AI) and data analysis.
The following are the specifications of GPU types provided by the GPU Server service.
| Item | A100 Type | H100 Type |
|---|---|---|
| Service Provision Method | Pass-through | Pass-through |
| GPU Architecture | NVIDIA Ampere | NVIDIA Hopper |
| GPU Memory | 80 GB | 80 GB |
| GPU Transistors | 54 billion 7N TSMC | 80 billion 4N TSMC |
| FP16 Tensor Core (Dense) | 312 TFLOPs | 989 TFLOPs |
| FP8 Tensor Core (Dense) | Not supported | 1,979 TFLOPs |
| FP4 Tensor Core (Dense) | Not supported | Not supported |
| GPU Memory Bandwidth | 2,039 GB/s HBM2e | 3,352 GB/s HBM3 |
| NVLink Performance | NVLink 3 | NVLink 4 |
| NVLink Signaling Rate | 25 GB/s (x12) | 25 GB/s (x18) |
| NVSwitch GPU-to-GPU Bandwidth | 600 GB/s | 900 GB/s |
| Total NVSwitch Aggregate Bandwidth | 4.8 TB/s | 7.2 TB/s |
Server Type
The server types provided by GPU Server are as follows. For a detailed description of the server types provided by GPU Server, see GPU Server Server Types.
| Item | Server Type | CPU vCore | Memory(GB) | GPU Quantity |
|---|---|---|---|---|
| GPU-A100-1 | g1v16a1 | 16 | 234 | 1 |
| GPU-A100-1 | g1v32a2 | 32 | 468 | 2 |
| GPU-A100-1 | g1v64a4 | 64 | 936 | 4 |
| GPU-A100-1 | g1v128a8 | 128 | 1872 | 8 |
| GPU-H100-2 | g2v12h1 | 12 | 234 | 1 |
| GPU-H100-2 | g2v24h2 | 24 | 468 | 2 |
| GPU-H100-2 | g2v48h4 | 48 | 936 | 4 |
| GPU-H100-2 | g2v96h8 | 96 | 1872 | 8 |
OS and GPU Driver Version
The operating systems (OS) supported by GPU Server are as follows:
| OS | OS Version | GPU Driver Version |
|---|---|---|
| Ubuntu | 22.04 | 535.183.06 |
| Ubuntu | 24.04 | 570.195.03 |
| RHEL | 8.10 | 535.183.06 |
Prerequisite Services
This is a service that must be pre-installed before creating this service. Please prepare by referring to the user guide provided in advance.
| Service Category | Service | Description |
|---|---|---|
| Networking | VPC | Service that provides independent virtual networks in cloud environment |
| Networking | Security Group | Virtual firewall that controls server traffic |
3.1.1 - Server Type
GPU Server Server Type
GPU Server is classified according to the GPU Type provided, and the GPU used in the GPU Server is determined by the server type selected when creating the GPU Server. Please select the server type according to the specifications of the application you want to run on the GPU Server.
The server types supported by the GPU Server are as follows.
GPU-H100-2 g2v12h1
Category | Example | Detailed description |
|---|---|---|
| Server Type | GPU-H100-2 | Provided server type classification
|
| Server specifications | g2 | Provided server type classification and generation
|
| Server specifications | v12 | Number of vCores
|
| Server specifications | h1 | GPU type and quantity
|
g1 server type
The g1 server type is a GPU Server that uses NVIDIA A100 Tensor Core GPU, suitable for high-performance applications.
- Provides up to 8 NVIDIA A100 Tensor Core GPUs
- Equipped with 6,912 CUDA cores and 432 Tensor cores per GPU
- Supports up to 128 vCPUs and 1,920 GB of memory
- Maximum 40 Gbps networking speed
- 600GB/s GPU and NVIDIA NVSwitch P2P communication
| Category | Server Type | GPU | CPU | Memory | GPU Memory | Network Bandwidth |
|---|---|---|---|---|---|---|
| GPU-A100-1 | g1v16a1 | 1 | 16 vCore | 234 GB | 80 GB | up to 20 Gbps |
| GPU-A100-1 | g1v32a2 | 2 | 32 vCore | 468 GB | 160 GB | up to 20 Gbps |
| GPU-A100-1 | g1v64a4 | 4 | 64 vCore | 936 GB | 320 GB | up to 40 Gbps |
| GPU-A100-1 | g1v128a8 | 8 | 128 vCore | 1872 GB | 640 GB | Maximum 40 Gbps |
g2 server type
The g2 server type is a GPU Server that uses NVIDIA H100 Tensor Core GPU, suitable for high-performance applications.
- Up to 8 NVIDIA H100 Tensor Core GPUs provided
- Equipped with 16,896 CUDA cores and 528 Tensor cores per GPU
- Supports up to 96 vCPUs and 1,920 GB of memory
- Maximum networking speed of 40Gbps
- 900GB/s GPU and NVIDIA NVSwitch P2P communication
| Category | Server Type | GPU | CPU | Memory | GPU Memory | Network Bandwidth |
|---|---|---|---|---|---|---|
| GPU-H100-2 | g2v12h1 | 1 | 12 vCore | 234 GB | 80 GB | up to 20 Gbps |
| GPU-H100-2 | g2v24h2 | 2 | 24 vCore | 468 GB | 160 GB | up to 20 Gbps |
| GPU-H100-2 | g2v48h4 | 4 | 48 vCore | 936 GB | 320 GB | Maximum 40 Gbps |
| GPU-H100-2 | g2v96h8 | 8 | 96 vCore | 1872 GB | 640 GB | up to 40 Gbps |
3.1.2 - Monitoring Metrics
GPU Server Monitoring Metrics
The following table shows the monitoring metrics of the GPU Server that can be checked through Cloud Monitoring.
Even without installing an Agent, basic monitoring metrics are provided. Please check the Table. GPU Server Monitoring Metrics (Basic) below. Additionally, metrics that can be retrieved by installing an Agent are referenced in the Table. GPU Server Additional Monitoring Metrics (Agent Installation Required) below.
For detailed Cloud Monitoring usage, please refer to the Cloud Monitoring guide.
| Performance Item Name | Description | Unit |
|---|---|---|
| Memory Total [Basic] | Total available memory in bytes | bytes |
| Memory Used [Basic] | Currently used memory in bytes | bytes |
| Memory Swap In [Basic] | Swapped memory in bytes | bytes |
| Memory Swap Out [Basic] | Swapped memory in bytes | bytes |
| Memory Free [Basic] | Unused memory in bytes | bytes |
| Disk Read Bytes [Basic] | Read bytes | bytes |
| Disk Read Requests [Basic] | Number of read requests | cnt |
| Disk Write Bytes [Basic] | Written bytes | bytes |
| Disk Write Requests [Basic] | Number of write requests | cnt |
| CPU Usage [Basic] | Average system CPU usage over 1 minute | % |
| Instance State [Basic] | Instance state | state |
| Network In Bytes [Basic] | Received bytes | bytes |
| Network In Dropped [Basic] | Dropped received packets | cnt |
| Network In Packets [Basic] | Received packets | cnt |
| Network Out Bytes [Basic] | Sent bytes | bytes |
| Network Out Dropped [Basic] | Dropped sent packets | cnt |
| Network Out Packets [Basic] | Sent packets | cnt |
| Performance Item Name | Description | Unit |
|---|---|---|
| GPU Count | Number of GPUs | cnt |
| GPU Memory Usage | GPU memory usage rate | % |
| GPU Memory Used | Used GPU memory | MB |
| GPU Temperature | GPU temperature | ℃ |
| GPU Usage | GPU utilization | % |
| GPU Usage [Avg] | Average GPU usage rate | % |
| GPU Power Cap | Maximum power capacity of the GPU | W |
| GPU Power Usage | Current power usage of the GPU | W |
| GPU Memory Usage [Avg] | Average GPU memory usage rate | % |
| GPU Count in use | Number of GPUs in use by jobs on the node | cnt |
| Execution Status for nvidia-smi | Execution result of the nvidia-smi command | status |
| Core Usage [IO Wait] | CPU time spent in IO wait state | % |
| Core Usage [System] | CPU time spent in system space | % |
| Core Usage [User] | CPU time spent in user space | % |
| CPU Cores | Number of CPU cores on the host | cnt |
| CPU Usage [Active] | CPU time used, excluding idle and IO wait states | % |
| CPU Usage [Idle] | CPU time spent in idle state | % |
| CPU Usage [IO Wait] | CPU time spent in IO wait state | % |
| CPU Usage [System] | CPU time used by the kernel | % |
| CPU Usage [User] | CPU time used by user space | % |
| CPU Usage/Core [Active] | CPU time used per core, excluding idle and IO wait states | % |
| CPU Usage/Core [Idle] | CPU time spent in idle state per core | % |
| CPU Usage/Core [IO Wait] | CPU time spent in IO wait state per core | % |
| CPU Usage/Core [System] | CPU time used by the kernel per core | % |
| CPU Usage/Core [User] | CPU time used by user space per core | % |
| Disk CPU Usage [IO Request] | CPU time spent on IO requests | % |
| Disk Queue Size [Avg] | Average queue length of requests | num |
| Disk Read Bytes | Bytes read from the device per second | bytes |
| Disk Read Bytes [Delta Avg] | Average delta of bytes read from the device | bytes |
| Disk Read Bytes [Delta Max] | Maximum delta of bytes read from the device | bytes |
| Disk Read Bytes [Delta Min] | Minimum delta of bytes read from the device | bytes |
| Disk Read Bytes [Delta Sum] | Sum of delta of bytes read from the device | bytes |
| Disk Read Bytes [Delta] | Delta of bytes read from the device | bytes |
| Disk Read Bytes [Success] | Total bytes successfully read | bytes |
| Disk Read Requests | Number of read requests to the device per second | cnt |
| Disk Read Requests [Delta Avg] | Average delta of read requests to the device | cnt |
| Disk Read Requests [Delta Max] | Maximum delta of read requests to the device | cnt |
| Disk Read Requests [Delta Min] | Minimum delta of read requests to the device | cnt |
| Disk Read Requests [Delta Sum] | Sum of delta of read requests to the device | cnt |
| Disk Read Requests [Success Delta] | Delta of successful read requests to the device | cnt |
| Disk Read Requests [Success] | Total successful read requests | cnt |
| Disk Request Size [Avg] | Average size of requests to the device | num |
| Disk Service Time [Avg] | Average service time of requests to the device | ms |
| Disk Wait Time [Avg] | Average wait time of requests to the device | ms |
| Disk Wait Time [Read] | Average read wait time of the device | ms |
| Disk Wait Time [Write] | Average write wait time of the device | ms |
| Disk Write Bytes [Delta Avg] | Average delta of bytes written to the device | bytes |
| Disk Write Bytes [Delta Max] | Maximum delta of bytes written to the device | bytes |
| Disk Write Bytes [Delta Min] | Minimum delta of bytes written to the device | bytes |
| Disk Write Bytes [Delta Sum] | Sum of delta of bytes written to the device | bytes |
| Disk Write Bytes [Delta] | Delta of bytes written to the device | bytes |
| Disk Write Bytes [Success] | Total bytes successfully written | bytes |
| Disk Write Requests | Number of write requests to the device per second | cnt |
| Disk Write Requests [Delta Avg] | Average delta of write requests to the device | cnt |
| Disk Write Requests [Delta Max] | Maximum delta of write requests to the device | cnt |
| Disk Write Requests [Delta Min] | Minimum delta of write requests to the device | cnt |
| Disk Write Requests [Delta Sum] | Sum of delta of write requests to the device | cnt |
| Disk Write Requests [Success Delta] | Delta of successful write requests to the device | cnt |
| Disk Write Requests [Success] | Total successful write requests | cnt |
| Disk Writes Bytes | Bytes written to the device per second | bytes |
| Filesystem Hang Check | Filesystem hang check (normal: 1, abnormal: 0) | status |
| Filesystem Nodes | Total number of filesystem nodes | cnt |
| Filesystem Nodes [Free] | Total number of available filesystem nodes | cnt |
| Filesystem Size [Available] | Available disk space in bytes | bytes |
| Filesystem Size [Free] | Free disk space in bytes | bytes |
| Filesystem Size [Total] | Total disk space in bytes | bytes |
| Filesystem Usage | Disk space usage rate | % |
| Filesystem Usage [Avg] | Average disk space usage rate | % |
| Filesystem Usage [Inode] | Inode usage rate | % |
| Filesystem Usage [Max] | Maximum disk space usage rate | % |
| Filesystem Usage [Min] | Minimum disk space usage rate | % |
| Filesystem Usage [Total] | Total disk space usage rate | % |
| Filesystem Used | Used disk space in bytes | bytes |
| Filesystem Used [Inode] | Used inode space in bytes | bytes |
| Memory Free | Total available memory in bytes | bytes |
| Memory Free [Actual] | Actual available memory in bytes | bytes |
| Memory Free [Swap] | Available swap memory in bytes | bytes |
| Memory Total | Total memory in bytes | bytes |
| Memory Total [Swap] | Total swap memory in bytes | bytes |
| Memory Usage | Memory usage rate | % |
| Memory Usage [Actual] | Actual memory usage rate | % |
| Memory Usage [Cache Swap] | Cache swap usage rate | % |
| Memory Usage [Swap] | Swap memory usage rate | % |
| Memory Used | Used memory in bytes | bytes |
| Memory Used [Actual] | Actual used memory in bytes | bytes |
| Memory Used [Swap] | Used swap memory in bytes | bytes |
| Collisions | Network collisions | cnt |
| Network In Bytes | Received bytes | bytes |
| Network In Bytes [Delta Avg] | Average delta of received bytes | bytes |
| Network In Bytes [Delta Max] | Maximum delta of received bytes | bytes |
| Network In Bytes [Delta Min] | Minimum delta of received bytes | bytes |
| Network In Bytes [Delta Sum] | Sum of delta of received bytes | bytes |
| Network In Bytes [Delta] | Delta of received bytes | bytes |
| Network In Dropped | Dropped received packets | cnt |
| Network In Errors | Received errors | cnt |
| Network In Packets | Received packets | cnt |
| Network In Packets [Delta Avg] | Average delta of received packets | cnt |
| Network In Packets [Delta Max] | Maximum delta of received packets | cnt |
| Network In Packets [Delta Min] | Minimum delta of received packets | cnt |
| Network In Packets [Delta Sum] | Sum of delta of received packets | cnt |
| Network In Packets [Delta] | Delta of received packets | cnt |
| Network Out Bytes | Sent bytes | bytes |
| Network Out Bytes [Delta Avg] | Average delta of sent bytes | bytes |
| Network Out Bytes [Delta Max] | Maximum delta of sent bytes | bytes |
| Network Out Bytes [Delta Min] | Minimum delta of sent bytes | bytes |
| Network Out Bytes [Delta Sum] | Sum of delta of sent bytes | bytes |
| Network Out Bytes [Delta] | Delta of sent bytes | bytes |
| Network Out Dropped | Dropped sent packets | cnt |
| Network Out Errors | Sent errors | cnt |
| Network Out Packets | Sent packets | cnt |
| Network Out Packets [Delta Avg] | Average delta of sent packets | cnt |
| Network Out Packets [Delta Max] | Maximum delta of sent packets | cnt |
| Network Out Packets [Delta Min] | Minimum delta of sent packets | cnt |
| Network Out Packets [Delta Sum] | Sum of delta of sent packets | cnt |
| Network Out Packets [Delta] | Delta of sent packets | cnt |
| Open Connections [TCP] | Open TCP connections | cnt |
| Open Connections [UDP] | Open UDP connections | cnt |
| Port Usage | Port usage rate | % |
| SYN Sent Sockets | Number of sockets in SYN_SENT state | cnt |
| Kernel PID Max | Maximum PID value | cnt |
| Kernel Thread Max | Maximum thread value | cnt |
| Process CPU Usage | CPU time used by the process | % |
| Process CPU Usage/Core | CPU time used by the process per core | % |
| Process Memory Usage | Resident Set size | % |
| Process Memory Used | Used memory by the process | bytes |
| Process PID | Process ID | PID |
| Process PPID | Parent process ID | PID |
| Processes [Dead] | Number of dead processes | cnt |
| Processes [Idle] | Number of idle processes | cnt |
| Processes [Running] | Number of running processes | cnt |
| Processes [Sleeping] | Number of sleeping processes | cnt |
| Processes [Stopped] | Number of stopped processes | cnt |
| Processes [Total] | Total number of processes | cnt |
| Processes [Unknown] | Number of unknown processes | cnt |
| Processes [Zombie] | Number of zombie processes | cnt |
| Running Process Usage | Process usage rate | % |
| Running Processes | Number of running processes | cnt |
| Running Thread Usage | Thread usage rate | % |
| Running Threads | Number of running threads | cnt |
| Context Switches | Context switches per second | cnt |
| Load/Core [1 min] | Load per core over 1 minute | cnt |
| Load/Core [15 min] | Load per core over 15 minutes | cnt |
| Load/Core [5 min] | Load per core over 5 minutes | cnt |
| Multipaths [Active] | Number of active multipath connections | cnt |
| Multipaths [Failed] | Number of failed multipath connections | cnt |
| Multipaths [Faulty] | Number of faulty multipath connections | cnt |
| NTP Offset | Measured offset from the NTP server | num |
| Run Queue Length | Run queue length | num |
| Uptime | System uptime in milliseconds | ms |
| Context Switchies | Context switches per second | cnt |
| Disk Read Bytes [Sec] | Bytes read from the device per second | cnt |
| Disk Read Time [Avg] | Average read time from the device | sec |
| Disk Transfer Time [Avg] | Average disk transfer time | sec |
| Disk Usage | Disk usage rate | % |
| Disk Write Bytes [Sec] | Bytes written to the device per second | cnt |
| Disk Write Time [Avg] | Average write time to the device | sec |
| Pagingfile Usage | Paging file usage rate | % |
| Pool Used [Non Paged] | Non-paged pool usage | bytes |
| Pool Used [Paged] | Paged pool usage | bytes |
| Process [Running] | Number of running processes | cnt |
| Threads [Running] | Number of running threads | cnt |
| Threads [Waiting] | Number of waiting threads | cnt |
3.1.3 - ServiceWatch Metrics
GPU Server sends metrics to ServiceWatch. The metrics provided by basic monitoring are data collected at 5-minute intervals. When detailed monitoring is enabled, you can view data collected at 1-minute intervals.
- GPU Server’s basic monitoring and detailed monitoring are provided with the same metrics as Virtual Server, and the namespace is also provided as Virtual Server.
- GPU-related metrics are provided through ServiceWatch Agent. For information on how to collect metrics using ServiceWatch Agent, refer to the ServiceWatch Agent guide.
For information on how to enable detailed monitoring for GPU Server, refer to How-to guides > Enable ServiceWatch detailed monitoring.
Basic Metrics
The following are basic metrics for the namespace Virtual Server.
In the table below, metrics with metric names marked in bold are selected as key metrics among the basic metrics provided by Virtual Server. Key metrics are used to configure service dashboards that are automatically built for each service in ServiceWatch.
Each metric guides you on which statistic value is meaningful when querying that metric through the user guide, and the statistic value marked in bold among the meaningful statistics is the key statistic value. In the service dashboard, you can view key metrics through key statistic values.
| Performance Item | Detailed Description | Unit | Meaningful Statistics |
|---|---|---|---|
| Instance State | Instance state display
| None |
|
| CPU Usage | CPU usage | Percent |
|
| Disk Read Bytes | Amount read from block device (bytes) | Bytes |
|
| Disk Read Requests | Number of read requests from block device | Count |
|
| Disk Write Bytes | Amount written to block device (bytes) | Bytes |
|
| Disk Write Requests | Number of write requests to block device | Count |
|
| Network In Bytes | Amount received on network interface (bytes) | Bytes |
|
| Network In Dropped | Number of received packets dropped on network interface | Count |
|
| Network In Packets | Number of received packets on network interface | Count |
|
| Network Out Bytes | Amount transmitted on network interface (bytes) | Bytes |
|
| Network Out Dropped | Number of transmitted packets dropped on network interface | Count |
|
| Network Out Packets | Amount transmitted on network interface | Count |
|
3.2 - How-to guides
The user can enter the required information of the GPU Server through the Samsung Cloud Platform Console, select detailed options, and create the service.
Create GPU Server
You can create and use GPU Server services from the Samsung Cloud Platform Console.
If you want to create a GPU Server, follow the steps below.
- All Services > Compute > GPU Server Please click the menu. Navigate to the GPU Server’s Service Home page.
- Service Home on the page, click the Create GPU Server button. Navigate to the Create GPU Server page.
- GPU Server creation On the page, enter the information required to create the service, and select detailed options.
- Image and version selection Select the required information in the area.
Category Required or notDetailed description Image Required Select provided image type - Ubuntu
Image version Required Select version of the chosen image - Provides a list of server image versions offered
Table. GPU Server image and version selection input items - Service Information Input Enter or select the required information in the area.
Category Required or notDetailed description Server count Required Number of GPU Server servers to create simultaneously - Only numbers can be entered, and input a value between 1 and 100
Service Type > Server Type Required GPU Server Server Type - Indicates the server specifications of GPU type, select a server that includes 1, 2, 4, or 8 GPUs
- For detailed information about the server types provided by GPU Server, refer to GPU Server Server Type
Service Type > Planned Compute Select Resource status with Planned Compute set - In Use: Number of resources with Planned Compute set that are currently in use
- Configured: Number of resources with Planned Compute set
- Coverage Preview: Amount applied per resource by Planned Compute
- Planned Compute Service Application: Go to the Planned Compute service application page
- For more details, refer to Apply for Planned Compute
Block Storage Required Set the Block Storage used by the GPU Server according to its purpose - Basic: Area where the OS is installed and used
- Capacity can be entered in Units (minimum capacity varies depending on the OS image type)
- RHEL: values between 3 and 1,536 can be entered
- Ubuntu: values between 3 and 1,536 can be entered
- SSD: high-performance general volume
- HDD: general volume
- SSD/HDD_KMS: additional encrypted volume using Samsung Cloud Platform KMS (Key Management System) encryption key
- Encryption can only be applied at initial creation (cannot be changed after creation)
- Performance degradation occurs when using SSD_KMS disk type
- Capacity can be entered in Units (minimum capacity varies depending on the OS image type)
- Additional: Used when additional user space is needed outside the OS area
- After selecting Use, enter the storage type and capacity
- To add storage, click the + button (up to 25 can be added); to delete, click the x button
- Capacity can be entered in Units, values between 1 and 1,536
- Since 1 Unit equals 8 GB, 8 to 12,288 GB can be created
- SSD: high-performance general volume
- HDD: general volume
- SSD/HDD_KMS: additional encrypted volume using Samsung Cloud Platform KMS (Key Management System) encryption key
- Encryption can only be applied at initial creation (cannot be changed after creation)
- Performance degradation may occur when using SSD_KMS disk type
- For detailed information on each Block Storage type, refer to Create Block Storage
- Delete on termination: If Delete on Termination is set to Enabled, the volume will be terminated together when the server is terminated
- Volumes with snapshots are not deleted even if Delete on termination is set to Enabled
- A multi-attach volume is deleted only when the server being terminated is the last remaining server attached to the volume
Table. GPU Server Service Configuration Items - Required Information Input Please enter or select the required information in the area.
Category RequiredDetailed description Server Name Required Enter a name to distinguish the server when the selected server count is 1 - Set hostname with the entered server name
- Use English letters, numbers, spaces, and special characters (
-_) within 63 characters
Server name Prefix Required Enter Prefix to distinguish each server generated when the selected number of servers is 2 or more - Automatically generated as user input value (prefix) + ‘
-#’ format
- Enter within 59 characters using English letters, numbers, spaces, and special characters (
-,_)
Network Settings > Create New Network Port Required Set the network where the GPU Server will be installed - Select a pre-created VPC.
- General Subnet: Select a pre-created General Subnet
- IP can be set to auto-generate or user input; if input is selected, the user can directly enter the IP
- NAT: Can be used only if there is one server and the VPC has an Internet Gateway attached. Use checked allows selection of NAT IP
- NAT IP: Select NAT IP
- If there is no NAT IP to select, click the Create New button to generate a Public IP
- Refresh button click to view and select the created Public IP
- Creating a Public IP incurs charges according to the Public IP pricing policy
- Local Subnet (Optional): Select Local Subnet Use
- Not a required element for creating the service
- A pre-created Local Subnet must be selected
- IP can be set to auto-generate or user input; input selected allows user to directly enter IP
- Security Group: Settings required to access the server
- Select: Choose a pre-created Security Group
- Create New: If there is no applicable Security Group, it can be created separately in the Security Group service
- Up to 5 can be selected
- If no Security Group is set, all access is blocked by default
- A Security Group must be set to allow required access
Network Settings > Existing Network Port Specification Required Set the network where the GPU Server will be installed - Select a pre-created VPC
- General Subnet: Select a pre-created General Subnet and Port
- NAT: Can be used only if there is one server and the VPC has an Internet Gateway attached; checking use allows selection of a NAT IP.
- NAT IP: Select NAT IP
- If there is no NAT IP to select, click the New Creation button to generate a Public IP
- Click the Refresh button to view and select the created Public IP
- Local Subnet (Optional): Select Use of Local Subnet
- Select a pre-created Local Subnet and Port
Keypair Required User authentication method to use when connecting to the server - New Creation: If a new Keypair is needed, create a new one
- Refer to Create Keypair for how to create a new Keypair
- List of default login accounts by OS
- RHEL: cloud-user
- Ubuntu: ubuntu
</ul
Table. GPU Server required information input items - Additional Information Input Enter or select the required information in the area.
Category RequiredDetailed description Lock Select Set whether to use Lock - Using Lock prevents actions such as server termination, start, stop from being executed, preventing malfunctions caused by mistakes
Init script Select Script executed when the server starts - The Init script must be written as a Batch script for Windows, a Shell script or cloud‑init for Linux, depending on the image type.
- Up to 45,000 bytes can be entered
Tag Select Add Tag - Up to 50 can be added per resource
- After clicking the Add Tag button, enter or select Key, Value values
Table. GPU Server Additional Information Input Items
- Image and version selection Select the required information in the area.
- Summary Check the detailed information and estimated billing amount generated in the panel, and click the Complete button.
- When creation is complete, check the created resources on the GPU Server list page.
GPU Server Check detailed information
GPU Server service can view and edit the full resource list and detailed information. GPU Server Detail page consists of Detail Information, Tags, Job History tabs.
To view detailed information about the GPU Server service, follow the steps below.
- All Services > Compute > GPU Server Click the menu. Go to the GPU Server’s Service Home page.
- Click the GPU Server menu on the Service Home page. Navigate to the GPU Server list page.
- GPU Server List Click the resource to view detailed information on the page. GPU Server Details You will be taken to the page.
- GPU Server Details page displays status information and additional feature information, and consists of Details, Tags, Job History tabs.
- For detailed information about the GPU Server status, please refer to Check GPU Server detailed information.
Category Detailed description GPU Server status Status of GPU Server created by the user - Build: State where Build command has been delivered
- Building: Build in progress
- Networking: Server creation in‑progress process
- Scheduling: Server creation in‑progress process
- Block_Device_Mapping: Connecting Block Storage during server creation
- Spawning: State where server creation process is ongoing
- Active: Usable state
- Powering_off: State when stop request is made
- Deleting: Server deletion in progress
- Reboot_Started: Reboot in progress state
- Error: Error state
- Migrating: State where server is migrating to another host
- Reboot: State where Reboot command has been delivered
- Rebooting: Restart in progress
- Rebuild: State where Rebuild command has been delivered
- Rebuilding: State when Rebuild is requested
- Rebuild_Spawning: State where Rebuild process is ongoing
- Resize: State where Resize command has been delivered
- Resizing: Resize in progress
- Resize_Prep: State when server type modification is requested
- Resize_Migrating: State where server is moving to another host while Resize is in progress
- Resize_Migrated: State where server has completed moving to another host while Resize is in progress
- Resize_Finish: Resize completed
- Revert_Resize: Resize or migration of the server failed for some reason. The target server is cleaned up and the original server restarts
- Shutoff: State when Powering off is completed
- Verity_Resize: After Resize_Prep due to server type modification request, state where server type is confirmed / can choose to revert server type
- Resize_Reverting: State when server type revert is requested
- Resize_Confirming: State where server’s Resize request is being confirmed
Server Control Button to change server status - Start: Start a stopped server
- Stop: Stop a running server
- Restart: Restart a running server
Image Generation Create user custom image using the current server’s image Console Log View current server’s console log - You can check the console log output from the current server. For more details, refer to Check Console Log.
Dump creation Create a dump of the current server - The dump file is created inside the GPU Server
- For detailed dump creation method, refer to [Create Dump](#dump-생성하기)
- For details, refer to [Perform Rebuild](#rebuild-수행하기).
Table. GPU Server status information and additional features
Detailed Information
GPU Server list page, you can view the detailed information of the selected resource and, if necessary, edit the information.
| Category | Detailed description |
|---|---|
| service | service name |
| Resource Type | Resource Type |
| SRN | Unique resource ID in Samsung Cloud Platform
|
| Resource Name | Resource Name
|
| Resource ID | Unique resource ID in the service |
| Creator | User who created the service |
| Creation time | Service creation time |
| Editor | User who edited the service information |
| Modification Date | Date Service Information Was Modified |
| Server name | Server name |
| Server Type | vCPU, Memory, GPU Information Display
|
| Image Name | Service OS Image and Version |
| Lock | Display Lock usage status
|
| Server Group | Server group name the server belongs to |
| Keypair name | Server authentication information set by the user |
| Planned Compute | Resource status with Planned Compute set
|
| Network | Network information of GPU Server
|
| Local Subnet | GPU Server’s Local Subnet Information
|
| Block Storage | Information of Block Storage connected to the server
|
tag
On the GPU Server List page, you can view the tag information of the selected resource, and you can add, modify, or delete it.
| Category | Detailed description |
|---|---|
| Tag List | Tag List
|
Work History
You can view the job history of the selected resource on the GPU Server List page.
| Category | Detailed description |
|---|---|
| Work History List | Resource Change History
|
GPU Server Operation Control
If you need to control the operation of the generated GPU Server resources, you can perform the task on the GPU Server List or GPU Server Details page. You can start, stop, and restart a running server.
GPU Server Start
You can start a shutoff GPU Server. To start the GPU Server, follow the steps below.
- All Services > Compute > GPU Server Please click the menu. Navigate to the GPU Server’s Service Home page.
- Click the GPU Server menu on the Service Home page. Go to the GPU Server List page.
- GPU Server List page, click the resource to start among the shutoff servers, and go to the GPU Server Details page.
- GPU Server list page, you can start each resource via the right more button.
- After selecting multiple servers with the check box, you can control multiple servers simultaneously through the Start button at the top.
- GPU Server Details On the page, click the Start button at the top to start the server. Check the changed server status in the Status Display item.
- When the GPU Server start is completed, the server status changes from Shutoff to Active.
- For detailed information about the GPU Server status, please refer to Check GPU Server detailed information.
GPU Server Stop
You can stop a GPU Server that is active. To stop the GPU Server, follow the steps below.
- All Services > Compute > GPU Server Click the menu. Go to the Service Home page of GPU Server.
- Service Home page, click the GPU Server menu. Navigate to the GPU Server List page.
- GPU Server List page, click the resource to stop among the servers that are active (Active), and go to the GPU Server Details page.
- GPU Server List on the page, you can stop each resource via the right More button.
- After selecting multiple servers with the checkbox, you can control multiple servers simultaneously through the Stop button at the top.
- GPU Server Details page, click the top Stop button to start the server. Check the changed server status in the Status Display item.
- When GPU Server shutdown is completed, the server status changes from Active to Shutoff.
- For detailed information about the GPU Server status, please refer to Check GPU Server detailed information.
GPU Server Restart
You can restart the generated GPU Server. To restart the GPU Server, follow the steps below.
- All Services > Compute > GPU Server Please click the menu. Navigate to the GPU Server’s Service Home page.
- Click the GPU Server menu on the Service Home page. It navigates to the GPU Server List page.
- GPU Server List page, click the resource to restart, and navigate to the GPU Server Details page.
- GPU Server List page, you can restart each resource by using the right More button.
- After selecting multiple servers with the checkbox, you can control multiple servers simultaneously through the Restart button at the top.
- GPU Server Details on the page, click the Restart button at the top to start the server. Check the changed server status in the Status Display item.
- GPU Server during restart, the server status goes through Rebooting and finally changes to Active.
- For detailed information about the GPU Server status, please refer to Check GPU Server detailed information.
GPU Server Resource Management
If you need server control and management functions for the generated GPU Server resources, you can perform the work on the GPU Server Resource List or GPU Server Details page.
Image Create
You can create an image of a running GPU server.
This content provides instructions on how to create a user custom image using the image of a running GPU server.
- GPU Server list or GPU Server details page, click the Create Image button to create a user Custom Image.
To create an Image of the GPU Server, follow the steps below.
All Services > Compute > GPU Server Click the menu. Navigate to the GPU Server’s Service Home page.
Service Home page, click the GPU Server menu. Go to the GPU Server list page.
Click the resource to create an Image on the GPU Server List page. It navigates to the GPU Server Details page.
GPU Server Details on the page, click the Image Generation button. Navigate to the Image Generation page.
- Service Information Input area, please enter the required information.
Category RequiredDetailed description Image Name Required Image name to be generated - Enter within 200 characters using English letters, numbers, spaces, and special characters (
-_)
Table. Image Service Information Input Items - Enter within 200 characters using English letters, numbers, spaces, and special characters (
- Service Information Input area, please enter the required information.
Check the input information and click the Complete button.
- When creation is complete, check the created resources on the All Services > Compute > GPU Server > Image List page.
- If you create an Image, the generated Image is stored in the Object Storage used as internal storage. Therefore, Object Storage usage fees are charged.
- Active state GPU Server-generated image’s file system cannot guarantee integrity, so image creation after server shutdown is recommended.
ServiceWatch Enable Detailed Monitoring
Basically, the GPU Server is linked with the basic monitoring of ServiceWatch and the Virtual Server namespace. You can enable detailed monitoring as needed to more quickly identify and address operational issues. For more information about ServiceWatch, see the ServiceWatch Overview (/userguide/management/service_watch/overview/).
To enable detailed monitoring of ServiceWatch on the GPU Server, follow these steps.
- All Services > Compute > GPU Server Please click the menu. Navigate to the GPU Server’s Service Home page.
- Click the GPU Server menu on the Service Home page. Navigate to the GPU Server list page.
- On the GPU Server List page, click the resource to enable ServiceWatch detailed monitoring. You will be taken to the GPU Server Details page.
- GPU Server Details page, click the ServiceWatch detailed monitoring Edit button. You will be taken to the ServiceWatch Detailed Monitoring Edit popup.
- ServiceWatch Detailed Monitoring Modification In the popup window, after selecting Enable, check the guidance text and click the Confirm button.
- GPU Server Details page, check the ServiceWatch detailed monitoring items.
ServiceWatch Disable detailed monitoring
To disable the detailed monitoring of ServiceWatch on the GPU Server, follow the steps below.
- All Services > Compute > GPU Server Please click the menu. Navigate to the GPU Server’s Service Home page.
- Service Home page, click the GPU Server menu. Navigate to the GPU Server list page.
- Click the resource to disable ServiceWatch detailed monitoring on the GPU Server List page. It navigates to the GPU Server Details page.
- GPU Server Details page, click the ServiceWatch detailed monitoring Edit button. It moves to the ServiceWatch Detailed Monitoring Edit popup.
- ServiceWatch Detailed Monitoring Edit in the popup window, after deselecting Enable, check the guide text and click the Confirm button.
- GPU Server Details page, check the ServiceWatch detailed monitoring items.
GPU Server Management Additional Features
For GPU Server management, you can view Console logs, generate Dump, and Rebuild. To view Console logs, generate Dump, and Rebuild the GPU Server, follow the steps below.
Check console log
You can view the current console log of the GPU Server.
To check the console logs of the GPU Server, follow the steps below.
- All Services > Compute > GPU Server Please click the menu. Navigate to the GPU Server’s Service Home page.
- Click the GPU Server menu on the Service Home page. Navigate to the GPU Server List page.
- On the GPU Server List page, click the resource to view the console log. Navigate to the GPU Server Details page.
- GPU Server Details on the page, click the Console Log button. It will move to the Console Log popup window.
- Console Log Check the console log displayed in the popup window.
Create Dump
To create a Dump file of the GPU Server, follow the steps below.
- All Services > Compute > GPU Server Click the menu. Navigate to the GPU Server’s Service Home page.
- Click the GPU Server menu on the Service Home page. Navigate to the GPU Server List page.
- GPU Server List Click the resource to view detailed information on the page. GPU Server Details Navigate to the page.
- GPU Server Details on the page Create Dump Click the button.
- The dump file is created inside the GPU Server.
Rebuild perform
You can delete all data and settings of the existing GPU Server and rebuild it on a new server.
To perform the Rebuild of the GPU Server, follow the steps below.
- All Services > Compute > GPU Server Click the menu. Navigate to the Service Home page of GPU Server.
- Click the GPU Server menu on the Service Home page. It navigates to the GPU Server List page.
- GPU Server List page, click the resource to perform Rebuild. GPU Server Details page will be opened.
- GPU Server Details on the page click the Rebuild button.
- During GPU Server Rebuild, the server status changes to Rebuilding, and when the Rebuild is completed, it returns to the state before the Rebuild.
- For detailed information about the GPU Server status, please refer to Check GPU Server detailed information.
GPU Server Cancel
If you cancel an unused GPU Server, you can reduce operating costs. However, if you cancel a GPU Server, the service currently running may be stopped immediately, so you should proceed with the cancellation after fully considering the impact that may occur when the service is interrupted.
To cancel the GPU Server, follow the steps below.
- All Services > Compute > GPU Server Click the menu. Navigate to the GPU Server’s Service Home page.
- Click the GPU Server menu on the Service Home page. Navigate to the GPU Server List page.
- GPU Server List on the page, select the resource to cancel, and click the Cancel Service button.
- The termination of connected storage depends on the Delete on termination setting, so please refer to Termination Constraints.
- When termination is completed, check on the GPU Server List page whether the resource has been terminated.
Termination Constraints
If the termination request for GPU Server cannot be processed, we will guide you with a popup window. Please refer to the cases below.
- If File Storage is connected, please disconnect the File Storage connection first.
- If LB Pool is connected please disconnect the LB Pool connection first.
- When Lock is set please change the Lock setting to unused and try again.
The termination of attached storage depends on the Delete on termination setting.
- Delete on termination Whether the volume deletion also varies depending on the setting.
- Delete on termination If not set: Even if you terminate the GPU Server, the volume will not be deleted.
- Delete on termination when set: If you terminate the GPU Server, the volume will be deleted.
- Volumes with a Snapshot will not be deleted even if Delete on termination is set.
- Multi attach volume is deleted only when the server you are trying to delete is the last remaining server attached to the volume.
3.2.1 - Image Management
Users can enter the required information for the Image service within the GPU Server service and select detailed options through the Samsung Cloud Platform Console to create the respective service.
Image Generation
You can create an image of a running GPU Server. To create an image of a GPU Server, please refer to Create Image.
Image Check detailed information
Image service allows you to view and edit the full resource list and detailed information. Image detail page consists of detail information, tags, operation history tabs.
To view detailed information of the Image service, follow the steps below.
- All Services > Compute > GPU Server Click the menu. Navigate to the Service Home page of GPU Server.
- Click the Image menu on the Service Home page. Go to the Image List page.
- Image List page, click the resource to view detailed information. You will be taken to the Image Detail page.
- Image Details page displays status information and additional feature information, and consists of Detail Information, Tag, Work History tabs.
Category Detailed description Image Status User-generated Image’s status - Active: Available state
- Queued: When creating Image, Image is uploaded and waiting for processing
- Importing: When creating Image, Image is uploaded and being processed
Share to another Account Image can be shared to another Account - Image’s Visibility must be in Shared state to be able to share to another Account
Delete Image Button to delete the Image - If the Image is deleted, it cannot be recovered
Table. GPU Server Image status information and additional features
- Image Details page displays status information and additional feature information, and consists of Detail Information, Tag, Work History tabs.
Detailed Information
Image list page allows you to view detailed information of the selected resource and edit the information if necessary.
| Category | Detailed description |
|---|---|
| Service | Service Name |
| Resource Type | Resource Type |
| SRN | Unique resource ID in Samsung Cloud Platform
|
| Resource Name | Image Name |
| Resource ID | Image ID |
| Creator | User who created the Image |
| Creation date/time | Date/time when the image was created |
| Editor | User who edited the Image |
| Edit date/time | Date and time the image was edited |
| Image name | Image name |
| Minimum Disk | Image’s minimum disk capacity (GB)
|
| Minimum RAM | Image’s minimum RAM capacity (GB) |
| OS type | Image’s OS type |
| OS hash algorithm | OS hash algorithm method |
| Visibility | Displays access permissions for images
|
| Protected | Select whether image deletion is prohibited
|
| Image File URL | Image file URL uploaded when generating image
|
| Sharing Status | Status of sharing images with other Accounts
|
Tag
On the Image list page, you can view the tag information of the selected resource, and you can add, modify, or delete it.
| Category | Detailed description |
|---|---|
| Tag List | Tag List
|
Work History
You can view the operation history of the selected resource on the Image list page.
| Category | Detailed description |
|---|---|
| Work History List | Resource Change History
|
Image Resource Management
Describes the control and management functions of the generated image.
Share to another account
To share the Image with another Account, follow the steps below.
- Log in to the shared Account and click the All Services > Compute > GPU Server menu. Navigate to the GPU Server’s Service Home page.
- Click the Image menu on the Service Home page. Go to the Image list page.
- Click the Image to control on the Image List page. It moves to the Image Detail page.
- Click the Share to another Account button. It navigates to the Share image to another Account page.
- Share to another Account feature allows you to share the Image with another Account. To share the Image with another Account, the Image’s Visibility must be Shared.
- Share images to another Account On the page, enter the required information and click the Complete button.
Category Required or notDetailed description Image Name - Name of the image to share - Input not allowed
Image ID - Image ID to share - Input not allowed
Shared Account ID Required Enter another Account ID to share - Enter within 64 characters using English letters, numbers, special characters
-
Table. Required input items for sharing images to another Account - Image Details page’s sharing status can be checked for information.
- At the initial request, the status is Pending, and when approval is completed in the account to be shared, it changes to Accepted.
Receive sharing from another account
To receive an Image shared from another Account, follow the steps below.
- Log in to the account to be shared and click the All Services > Compute > GPU Server menu. Navigate to the Service Home page of the GPU Server.
- Click the Image menu on the Service Home page. Go to the Image list page.
- Image List on the page click the Get Image Share button. Go to the Get Image Share popup window.
- Receive Image Share In the popup window, enter the resource ID of the Image you want to receive, and click the Confirm button.
- When image sharing is completed, you can check the shared Image in the Image list.
Image Delete
You can delete unused Images. However, once an Image is deleted it cannot be recovered, so you should fully consider the impact before proceeding with the deletion.
To delete the image, follow the steps below.
- All Services > Compute > GPU Server Click the menu. Navigate to the GPU Server’s Service Home page.
- Click the Image menu on the Service Home page. Go to the Image List page.
- On the Image list page, select the resource to delete and click the Delete button.
- Image list page, select multiple Image check boxes, and click the Delete button at the top of the resource list.
- When deletion is complete, check on the Image List page whether the resource has been deleted.
3.2.2 - Using Multi-instance GPU in GPU Server
After creating a GPU Server, you can enable the MIG (Multi-instance GPU) feature on the GPU Server’s VM (Guest OS) and create an instance to use it.
Multi-instance GPU (NVIDIA A100) Overview
NVIDIA A100 is a Multi-instance GPU (MIG) based on the NVIDIA Ampere architecture, which can be securely divided into up to 7 independent GPU instances to operate CUDA (Compute Unified Device Architecture) applications. The NVIDIA A100 provides independent GPU resources to multiple users by allocating computing resources in a way optimized for GPU usage while utilizing high-bandwidth memory (HBM) and cache. Users can maximize GPU utilization by utilizing workloads that have not reached the maximum computing capacity of the GPU through parallel execution of each workload.
Using Multi-instance GPU Feature
To use the multi-instance GPU feature, you must create a GPU Server service on the Samsung Cloud Platform and then create a VM Instance (GuestOS) with an A100 GPU assigned. After completing the GPU Server creation, you can follow the MIG application order and MIG release order below to apply it.
- The system requirements for using the MIG feature are as follows (refer to NVIDIA - Supported GPUs).
- CUDA toolkit 11, NVIDIA driver 450.80.02 or later version
- Linux distribution operating system supporting CUDA toolkit 11
- When operating a container or Kubernetes service, the requirements for using the MIG feature are as follows.
- NVIDIA Container Toolkit(nvidia-docker2) v 2.5.0 or later version
- NVIDIA K8s Device Plugin v 0.7.0 or later version
- NVIDIA gpu-feature-discovery v 0.2.0 or later version
MIG Application and Usage
To activate MIG and create an instance to assign a task, follow these steps.
MIG Activation
Check the GPU status on the VM Instance (GuestOS) before applying MIG.
- MIG mode is Disabled status, please check.Color mode
$ nvidia-smi Mon Sep 27 08:37:08 2021 +-----------------------------------------------------------------------------+ | NVIDIA-SMI 470.57.02 Driver Version: 470.57.02 CUDA Version: 11.4 | |-------------------------------+----------------------+----------------------| | GPU Name Persistence-M| Bus-Id Disp.A | Volatile Uncorr. ECC | | Fan Temp Perf Pwr:Usage/Cap| Memory-Usage | GPU-Util Compute M. | | | | MIG M. | |===============================+======================+======================| | 0 NVDIA A100-SXM... Off | 00000000:05:00.0 Off | 0 | | N/A 32C P0 59W / 400W | 0MiB / 81251MiB | 0% Default | | | | Disabled | +-------------------------------+----------------------+----------------------+ +-----------------------------------------------------------------------------+ | Processes: | | GPU GI CI PID Type Process name GPU Memory | | ID ID Usage | |=============================================================================| | No running processes found | +-----------------------------------------------------------------------------+$ nvidia-smi Mon Sep 27 08:37:08 2021 +-----------------------------------------------------------------------------+ | NVIDIA-SMI 470.57.02 Driver Version: 470.57.02 CUDA Version: 11.4 | |-------------------------------+----------------------+----------------------| | GPU Name Persistence-M| Bus-Id Disp.A | Volatile Uncorr. ECC | | Fan Temp Perf Pwr:Usage/Cap| Memory-Usage | GPU-Util Compute M. | | | | MIG M. | |===============================+======================+======================| | 0 NVDIA A100-SXM... Off | 00000000:05:00.0 Off | 0 | | N/A 32C P0 59W / 400W | 0MiB / 81251MiB | 0% Default | | | | Disabled | +-------------------------------+----------------------+----------------------+ +-----------------------------------------------------------------------------+ | Processes: | | GPU GI CI PID Type Process name GPU Memory | | ID ID Usage | |=============================================================================| | No running processes found | +-----------------------------------------------------------------------------+Code block. nvidia-smi command - Check GPU inactive state (1) Color mode$ nvidia-smi –L GPU 0: NVIDIA A100-SXM-80GB (UUID: GPU-c956838f-494a-92b2-6818-56eb28fe25e0)$ nvidia-smi –L GPU 0: NVIDIA A100-SXM-80GB (UUID: GPU-c956838f-494a-92b2-6818-56eb28fe25e0)Code block. nvidia-smi command - Check GPU inactive state (2)
- MIG mode is Disabled status, please check.
In the VM Instance(GuestOS), enable MIG for each GPU and reboot the VM Instance.
Color mode$ nvidia-smi –I 0 –mig 1 Enabled MIG mode for GPU 00000000:05:00.0 All done. # reboot$ nvidia-smi –I 0 –mig 1 Enabled MIG mode for GPU 00000000:05:00.0 All done. # rebootCode Block. nvidia-smi Command - MIG Activation
If the GPU monitoring agent displays the following warning message, stop the nvsm and dcgm services before enabling MIG.
Warning: MIG mode is in pending enable state for GPU 00000000:05:00.0: In use by another client. 00000000:05:00.0 is currently being used by one or more other processes (e.g. CUDA application or a monitoring application such as another instance of nvidia-smi).
# systemctl stop nvsm
# systemctl stop dcgm
- After completing the MIG work, restart the nvsm and dcgm services.
- Check the GPU status after applying MIG in the VM Instance(GuestOS).
- MIG mode must be in Enabled state.Color mode
$ nvidia-smi Mon Sep 27 09:44:33 2021 +-----------------------------------------------------------------------------+ | NVIDIA-SMI 470.57.02 Driver Version: 470.57.02 CUDA Version: 11.4 | |-------------------------------+----------------------+----------------------| | GPU Name Persistence-M| Bus-Id Disp.A | Volatile Uncorr. ECC | | Fan Temp Perf Pwr:Usage/Cap| Memory-Usage | GPU-Util Compute M. | | | | MIG M. | |===============================+======================+======================| | 0 NVDIA A100-SXM... Off | 00000000:05:00.0 Off | On | | N/A 32C P0 59W / 400W | 0MiB / 81251MiB | 0% Default | | | | Enabled | +-------------------------------+----------------------+----------------------+ +-----------------------------------------------------------------------------+ | MIG devices: | +-----------------------------------------------------------------------------+ | GPU GI CI MIG | Memory-Usage | Vol| Shared | | ID ID Dev | BAR1-Usage | SM Unc| CE ENC DEC OFA JPG| | | | ECC| | |=============================================================================| | No MIG devices found | +-----------------------------------------------------------------------------+ +-----------------------------------------------------------------------------+ | Processes: | | GPU GI CI PID Type Process name GPU Memory | | ID ID Usage | |=============================================================================| | No running processes found | +-----------------------------------------------------------------------------+$ nvidia-smi Mon Sep 27 09:44:33 2021 +-----------------------------------------------------------------------------+ | NVIDIA-SMI 470.57.02 Driver Version: 470.57.02 CUDA Version: 11.4 | |-------------------------------+----------------------+----------------------| | GPU Name Persistence-M| Bus-Id Disp.A | Volatile Uncorr. ECC | | Fan Temp Perf Pwr:Usage/Cap| Memory-Usage | GPU-Util Compute M. | | | | MIG M. | |===============================+======================+======================| | 0 NVDIA A100-SXM... Off | 00000000:05:00.0 Off | On | | N/A 32C P0 59W / 400W | 0MiB / 81251MiB | 0% Default | | | | Enabled | +-------------------------------+----------------------+----------------------+ +-----------------------------------------------------------------------------+ | MIG devices: | +-----------------------------------------------------------------------------+ | GPU GI CI MIG | Memory-Usage | Vol| Shared | | ID ID Dev | BAR1-Usage | SM Unc| CE ENC DEC OFA JPG| | | | ECC| | |=============================================================================| | No MIG devices found | +-----------------------------------------------------------------------------+ +-----------------------------------------------------------------------------+ | Processes: | | GPU GI CI PID Type Process name GPU Memory | | ID ID Usage | |=============================================================================| | No running processes found | +-----------------------------------------------------------------------------+Code block. nvidia-smi command - Check GPU activation status (1) Color mode$ nvidia-smi –L GPU 0: NVIDIA A100-SXM-80GB (UUID: GPU-c956838f-494a-92b2-6818-56eb28fe25e0)$ nvidia-smi –L GPU 0: NVIDIA A100-SXM-80GB (UUID: GPU-c956838f-494a-92b2-6818-56eb28fe25e0)Code block. nvidia-smi command - Check GPU activation status (2)
- MIG mode must be in Enabled state.
GPU Instance Creation
After activating MIG and checking the status, you can create a GPU Instance.
Check the list of MIG GPU instance profiles that can be created.
Color mode$ nvidia-smi mig -i [GPU ID] -lgip$ nvidia-smi mig -i [GPU ID] -lgipCode block. nvidia-smi command - MIG GPU Instance profile list check Color mode$ nvidia-smi mig -i 0 -lgip +-----------------------------------------------------------------------------+ | GPU instance profiles: | | GPU Name ID Instances Memory P2P SM DEC ENC | | Free/Total GiB CE JPEG OFA | |=============================================================================| | 0 MIG 1g.10gb 19 7/7 9.50 No 14 0 0 | | 1 0 0 | +-----------------------------------------------------------------------------+ | 0 MIG 1g.10gb+me 20 1/1 9.50 No 14 0 0 | | 1 1 1 | +-----------------------------------------------------------------------------+ | 0 MIG 2g.20gb 14 3/3 19.50 No 28 1 0 | | 2 0 0 | +-----------------------------------------------------------------------------+ | 0 MIG 3g.40gb 9 2/2 39.50 No 42 2 0 | | 3 0 0 | +-----------------------------------------------------------------------------+ | 0 MIG 4g.40gb 5 1/1 39.50 No 56 2 0 | | 4 0 0 | +-----------------------------------------------------------------------------+ | 0 MIG 7g.80gb 0 1/1 79.25 No 98 0 0 | | 7 1 1 | +-----------------------------------------------------------------------------+$ nvidia-smi mig -i 0 -lgip +-----------------------------------------------------------------------------+ | GPU instance profiles: | | GPU Name ID Instances Memory P2P SM DEC ENC | | Free/Total GiB CE JPEG OFA | |=============================================================================| | 0 MIG 1g.10gb 19 7/7 9.50 No 14 0 0 | | 1 0 0 | +-----------------------------------------------------------------------------+ | 0 MIG 1g.10gb+me 20 1/1 9.50 No 14 0 0 | | 1 1 1 | +-----------------------------------------------------------------------------+ | 0 MIG 2g.20gb 14 3/3 19.50 No 28 1 0 | | 2 0 0 | +-----------------------------------------------------------------------------+ | 0 MIG 3g.40gb 9 2/2 39.50 No 42 2 0 | | 3 0 0 | +-----------------------------------------------------------------------------+ | 0 MIG 4g.40gb 5 1/1 39.50 No 56 2 0 | | 4 0 0 | +-----------------------------------------------------------------------------+ | 0 MIG 7g.80gb 0 1/1 79.25 No 98 0 0 | | 7 1 1 | +-----------------------------------------------------------------------------+Code Block. MIG GPU Instance Profile List
| Profile Name | Fraction of Memory | Fraction of SMs | Hardware Units | L2 Cache Size | Number of Instances Available |
|---|---|---|---|---|---|
| MIG 1g.10gb | 1/8 | 1/7 | 0 NVDECs /0 JPEG /0 OFA | 1/8 | 7 |
| MIG 1g.10gb+me | 1/8 | 1/7 | 1 NVDEC /1 JPEG /1 OFA | 1/8 | 1 (A single 1g profile can include media extensions) |
| MIG 2g.20gb | 2/8 | 2/7 | 1 NVDECs /0 JPEG /0 OFA | 2/8 | 3 |
| MIG 3g.40gb | 4/8 | 3/7 | 2 NVDECs /0 JPEG /0 OFA | 4/8 | 2 |
| MIG 4g.40gb | 4/8 | 4/7 | 2 NVDECs /0 JPEG /0 OFA | 4/8 | 1 |
| MIG 7g.80gb | Full | 7/7 | 5 NVDECs /1 JPEG /1 OFA | Full | 1 |
- Check after creating the MIG GPU Instance.
GPU Instance creation
Color mode$ nvidia-smi mig -i [GPU ID] -cgi [Profile ID]$ nvidia-smi mig -i [GPU ID] -cgi [Profile ID]Code Block. nvidia-smi command - GPU Instance creation Color mode$ nvidia-smi mig -i 0 -cgi 0 Successfully created GPU instance ID 0 on GPU 0 using profile MIG 7g.80gb (ID 0)$ nvidia-smi mig -i 0 -cgi 0 Successfully created GPU instance ID 0 on GPU 0 using profile MIG 7g.80gb (ID 0)Code block. nvidia-smi command - GPU Instance creation example GPU Instance check
Color mode$ nvidia-smi mig -i [GPU ID] -lgi$ nvidia-smi mig -i [GPU ID] -lgiCode Block. nvidia-smi Command - GPU Instance Check Color mode$ nvidia-smi mig -i 0 -lgi +--------------------------------------------------------+ | GPU instances: | | GPU Name Profile Instance Placement | | ID ID Start:Size | |========================================================| | 0 MIG 7g.80gb 0 0 0:8 | +--------------------------------------------------------+$ nvidia-smi mig -i 0 -lgi +--------------------------------------------------------+ | GPU instances: | | GPU Name Profile Instance Placement | | ID ID Start:Size | |========================================================| | 0 MIG 7g.80gb 0 0 0:8 | +--------------------------------------------------------+Code block. nvidia-smi command - GPU Instance check example
Compute Instance Creation
If you have created a GPU Instance, you can create a Compute Instance.
Check the MIG Compute Instance profile that can be created.
Color mode$ nvidia-smi mig -i [GPU ID] -gi [GPU Instance ID] -lcip$ nvidia-smi mig -i [GPU ID] -gi [GPU Instance ID] -lcipCode Block. nvidia-smi command - MIG Compute Instance profile check Color mode$ nvidia-smi mig -i 0 -gi 0 -lcip +---------------------------------------------------------------------------------+ | Compute instance profiles: | | GPU GPU Name Profile Instances Exclusive Shared | | GPU Instance ID Free/Total SM DEC ENC OFA | | ID CE JPEG | |=================================================================================| | 0 0 MIG 1c.7g.80gb 0 7/7 14 5 0 1 | | 7 1 | +---------------------------------------------------------------------------------+ | 0 0 MIG 2c.7g.80gb 1 3/3 28 5 0 1 | | 7 1 | +---------------------------------------------------------------------------------+ | 0 0 MIG 3c.7g.80gb 2 2/2 42 5 0 1 | | 7 1 | +---------------------------------------------------------------------------------+ | 0 0 MIG 4c.7g.80gb 3 1/1 56 5 0 1 | | 7 1 | +---------------------------------------------------------------------------------+ | 0 0 MIG 7g.80gb 4* 1/1 98 5 0 1 | | 7 1 | +---------------------------------------------------------------------------------+$ nvidia-smi mig -i 0 -gi 0 -lcip +---------------------------------------------------------------------------------+ | Compute instance profiles: | | GPU GPU Name Profile Instances Exclusive Shared | | GPU Instance ID Free/Total SM DEC ENC OFA | | ID CE JPEG | |=================================================================================| | 0 0 MIG 1c.7g.80gb 0 7/7 14 5 0 1 | | 7 1 | +---------------------------------------------------------------------------------+ | 0 0 MIG 2c.7g.80gb 1 3/3 28 5 0 1 | | 7 1 | +---------------------------------------------------------------------------------+ | 0 0 MIG 3c.7g.80gb 2 2/2 42 5 0 1 | | 7 1 | +---------------------------------------------------------------------------------+ | 0 0 MIG 4c.7g.80gb 3 1/1 56 5 0 1 | | 7 1 | +---------------------------------------------------------------------------------+ | 0 0 MIG 7g.80gb 4* 1/1 98 5 0 1 | | 7 1 | +---------------------------------------------------------------------------------+Code block. MIG Compute Instance profile list example Create and check the MIG Compute Instance.
- MIG Compute Instance creationColor mode
$ nvidia-smi mig -i [GPU ID] -gi [GPU Instance ID] -cci [Compute Profile ID]$ nvidia-smi mig -i [GPU ID] -gi [GPU Instance ID] -cci [Compute Profile ID]Code Block. nvidia-smi command - MIG Compute Instance creation Color mode$ nvidia-smi mig -i 0 -gi 0 -cci 4 Successfully created compute instance ID 0 on GPU instance ID 0 using profile MIG 7g.80gb(ID 4)$ nvidia-smi mig -i 0 -gi 0 -cci 4 Successfully created compute instance ID 0 on GPU instance ID 0 using profile MIG 7g.80gb(ID 4)Code block. nvidia-smi command - MIG Compute Instance creation example - MIG Compute Instance checkColor mode
$ nvidia-smi mig -i [GPU ID] -gi [GPU Instance ID] –lci$ nvidia-smi mig -i [GPU ID] -gi [GPU Instance ID] –lciCode block. nvidia-smi command - MIG Compute Instance check Color mode$ nvidia-smi mig -i 0 -gi 0 –lci +-----------------------------------------------------------------+ | Compute instance profiles: | | GPU GPU Name Profile Instances Placement | | GPU Instance ID ID Start:Size | | ID | |=================================================================| | 0 0 MIG 7g.80gb 4 0 0:7 | +-----------------------------------------------------------------+$ nvidia-smi mig -i 0 -gi 0 –lci +-----------------------------------------------------------------+ | Compute instance profiles: | | GPU GPU Name Profile Instances Placement | | GPU Instance ID ID Start:Size | | ID | |=================================================================| | 0 0 MIG 7g.80gb 4 0 0:7 | +-----------------------------------------------------------------+Code block. MIG Compute Instance confirmation example Color mode$ nvidia-smi –L GPU 0: NVIDIA A100-SXM-80GB (UUID: GPU-c956838f-494a-92b2-6818-56eb28fe25e0) MIG 7g.80gb Device 0: (UUID: MIG-53e20040-758b-5ecb-948e-c626d03a9a32)$ nvidia-smi –L GPU 0: NVIDIA A100-SXM-80GB (UUID: GPU-c956838f-494a-92b2-6818-56eb28fe25e0) MIG 7g.80gb Device 0: (UUID: MIG-53e20040-758b-5ecb-948e-c626d03a9a32)Code block. nvidia-smi command - Check GPU status (1) Color mode$ nvidia-smi Mon Sep 27 09:52:17 2021 +-----------------------------------------------------------------------------+ | NVIDIA-SMI 470.57.02 Driver Version: 470.57.02 CUDA Version: 11.4 | |-------------------------------+----------------------+----------------------| | GPU Name Persistence-M| Bus-Id Disp.A | Volatile Uncorr. ECC | | Fan Temp Perf Pwr:Usage/Cap| Memory-Usage | GPU-Util Compute M. | | | | MIG M. | |===============================+======================+======================| | 0 NVDIA A100-SXM... Off | 00000000:05:00.0 Off | On | | N/A 32C P0 49W / 400W | 0MiB / 81251MiB | N/A Default | | | | Enabled | +-------------------------------+----------------------+----------------------+ +-----------------------------------------------------------------------------+ | MIG devices: | +-----------------------------------------------------------------------------+ | GPU GI CI MIG | Memory-Usage | Vol| Shared | | ID ID Dev | BAR1-Usage | SM Unc| CE ENC DEC OFA JPG| | | | ECC| | |=============================================================================| | 0 0 0 0 | 0MiB / 81251MiB | 98 0 | 7 0 5 1 1 | | | 1MiB / 13107... | | | +-----------------------------------------------------------------------------+ +-----------------------------------------------------------------------------+ | Processes: | | GPU GI CI PID Type Process name GPU Memory | | ID ID Usage | |=============================================================================| | No running processes found | +-----------------------------------------------------------------------------+$ nvidia-smi Mon Sep 27 09:52:17 2021 +-----------------------------------------------------------------------------+ | NVIDIA-SMI 470.57.02 Driver Version: 470.57.02 CUDA Version: 11.4 | |-------------------------------+----------------------+----------------------| | GPU Name Persistence-M| Bus-Id Disp.A | Volatile Uncorr. ECC | | Fan Temp Perf Pwr:Usage/Cap| Memory-Usage | GPU-Util Compute M. | | | | MIG M. | |===============================+======================+======================| | 0 NVDIA A100-SXM... Off | 00000000:05:00.0 Off | On | | N/A 32C P0 49W / 400W | 0MiB / 81251MiB | N/A Default | | | | Enabled | +-------------------------------+----------------------+----------------------+ +-----------------------------------------------------------------------------+ | MIG devices: | +-----------------------------------------------------------------------------+ | GPU GI CI MIG | Memory-Usage | Vol| Shared | | ID ID Dev | BAR1-Usage | SM Unc| CE ENC DEC OFA JPG| | | | ECC| | |=============================================================================| | 0 0 0 0 | 0MiB / 81251MiB | 98 0 | 7 0 5 1 1 | | | 1MiB / 13107... | | | +-----------------------------------------------------------------------------+ +-----------------------------------------------------------------------------+ | Processes: | | GPU GI CI PID Type Process name GPU Memory | | ID ID Usage | |=============================================================================| | No running processes found | +-----------------------------------------------------------------------------+Code block. nvidia-smi command - Check GPU status (2)
- MIG Compute Instance creation
Using MIG
- Use the MIG Instance to perform the Job.
- Work execution exampleYou can check an example of how to perform the task as follows.Color mode
$ docker run --gpus '"device=[GPU ID]:[MIG ID]"' -rm nvcr.io/nvidia/cuda nvidia-smi$ docker run --gpus '"device=[GPU ID]:[MIG ID]"' -rm nvcr.io/nvidia/cuda nvidia-smiCode Block. Work Execution Example Color mode$ docker run --gpus '"device=0:0"' -rm -it --network=host --shm-size=1g --ipc=host -v /root/.ssh/:/root/.ssh ================ == TensorFlow == ================ NVIDIA Release 21.08-tf1 (build 26012104) TensorFlow Version 1.15.5 Container image Copyright (c) 2021, NVIDIA CORPORATION. All right reserved. ... # Python process execution root@d622a93c9281:/workspace# python /workspace/nvidia-examples/cnn/resnet.py --num_iter 100 ... PY 3.8.10 (default, Jun 2 2021, 10:49:15) [GCC 9.4.0] TF 1.15.5 ...$ docker run --gpus '"device=0:0"' -rm -it --network=host --shm-size=1g --ipc=host -v /root/.ssh/:/root/.ssh ================ == TensorFlow == ================ NVIDIA Release 21.08-tf1 (build 26012104) TensorFlow Version 1.15.5 Container image Copyright (c) 2021, NVIDIA CORPORATION. All right reserved. ... # Python process execution root@d622a93c9281:/workspace# python /workspace/nvidia-examples/cnn/resnet.py --num_iter 100 ... PY 3.8.10 (default, Jun 2 2021, 10:49:15) [GCC 9.4.0] TF 1.15.5 ...Code Block. Work Result
- Work execution example
- Check the GPU usage rate. (Creating a JOB process)
- You can see that when the Job is driven, the process is assigned to the MIG device and the usage rate increases.You can check the GPU usage rate as follows.Color mode
$ nvidia-smi mig -i [GPU ID] -gi [GPU Instance ID] -lcip$ nvidia-smi mig -i [GPU ID] -gi [GPU Instance ID] -lcipCode Block. nvidia-smi command - Check GPU usage Color mode+-----------------------------------------------------------------------------+ | MIG devices: | +-----------------------------------------------------------------------------+ | GPU GI CI MIG | Memory-Usage | Vol| Shared | | ID ID Dev | BAR1-Usage | SM Unc| CE ENC DEC OFA JPG| | | | ECC| | |=============================================================================| | 0 0 0 0 | 66562MiB / 81251MiB | 98 0 | 7 0 5 1 1 | | | 5MiB / 13107... | | | +-----------------------------------------------------------------------------+ +-----------------------------------------------------------------------------+ | Processes: | | GPU GI CI PID Type Process name GPU Memory | | ID ID Usage | |=============================================================================| | 0 0 0 17483 C python 66559MiB | +-----------------------------------------------------------------------------++-----------------------------------------------------------------------------+ | MIG devices: | +-----------------------------------------------------------------------------+ | GPU GI CI MIG | Memory-Usage | Vol| Shared | | ID ID Dev | BAR1-Usage | SM Unc| CE ENC DEC OFA JPG| | | | ECC| | |=============================================================================| | 0 0 0 0 | 66562MiB / 81251MiB | 98 0 | 7 0 5 1 1 | | | 5MiB / 13107... | | | +-----------------------------------------------------------------------------+ +-----------------------------------------------------------------------------+ | Processes: | | GPU GI CI PID Type Process name GPU Memory | | ID ID Usage | |=============================================================================| | 0 0 0 17483 C python 66559MiB | +-----------------------------------------------------------------------------+Code block. Example of checking GPU usage
- You can see that when the Job is driven, the process is assigned to the MIG device and the usage rate increases.
MIG Instance deletion and release
To delete a MIG instance and release the MIG, follow these procedures.
Compute Instance deletion
- Delete the Compute Instance.Color mode
$ nvidia-smi mig -i [GPU ID] -gi [GPU Instance ID] –dci $ nvidia-smi mig -i [GPU ID] -gi [GPU Instance ID] -ci [Compute Instance] –dci$ nvidia-smi mig -i [GPU ID] -gi [GPU Instance ID] –dci $ nvidia-smi mig -i [GPU ID] -gi [GPU Instance ID] -ci [Compute Instance] –dciCode Block. nvidia-smi command - Compute Instance deletion Color mode$ nvidia-smi mig -i 0 -gi 0 –lci +-----------------------------------------------------------------+ | Compute instance profiles: | | GPU GPU Name Profile Instances Placement | | GPU Instance ID ID Start:Size | | ID | |=================================================================| | 0 0 MIG 7g.80gb 4 0 0:7 | +-----------------------------------------------------------------+$ nvidia-smi mig -i 0 -gi 0 –lci +-----------------------------------------------------------------+ | Compute instance profiles: | | GPU GPU Name Profile Instances Placement | | GPU Instance ID ID Start:Size | | ID | |=================================================================| | 0 0 MIG 7g.80gb 4 0 0:7 | +-----------------------------------------------------------------+Code Block. MIG Compute Instance Check Example Color mode$ nvidia-smi mig -i 0 -gi 0 –dci Successfully destroyed compute instance ID 0 from GPU instance ID 0$ nvidia-smi mig -i 0 -gi 0 –dci Successfully destroyed compute instance ID 0 from GPU instance ID 0Code Block. Compute Instance deletion example Color mode$ nvidia-smi mig -i 0 -gi 0 –lci No compute instances found: Not found$ nvidia-smi mig -i 0 -gi 0 –lci No compute instances found: Not foundCode Block. Compute Instance deletion confirmation
GPU Instance deletion
- Delete the GPU Instance.Color mode
$ nvidia-smi mig -i [GPU ID] –dgi $ nvidia-smi mig -i [GPU ID] -gi [GPU Instance ID] –dgi$ nvidia-smi mig -i [GPU ID] –dgi $ nvidia-smi mig -i [GPU ID] -gi [GPU Instance ID] –dgiCode block. nvidia-smi command - GPU Instance deletion Color mode$ nvidia-smi mig -i 0 -lgi +--------------------------------------------------------+ | GPU instances: | | GPU Name Profile Instance Placement | | ID ID Start:Size | |========================================================| | 0 MIG 7g.80gb 0 0 0:8 | +--------------------------------------------------------+$ nvidia-smi mig -i 0 -lgi +--------------------------------------------------------+ | GPU instances: | | GPU Name Profile Instance Placement | | ID ID Start:Size | |========================================================| | 0 MIG 7g.80gb 0 0 0:8 | +--------------------------------------------------------+Code block. nvidia-smi command - GPU Instance check example Color mode$ nvidia-smi mig -i 0 -dgi Successfully destroyed GPU instance ID 0 from GPU 0$ nvidia-smi mig -i 0 -dgi Successfully destroyed GPU instance ID 0 from GPU 0Code block. nvidia-smi command - GPU Instance deletion example Color mode$ nvidia-smi mig -i 0 -lgi No GPU instances found: Not found$ nvidia-smi mig -i 0 -lgi No GPU instances found: Not foundCode block. nvidia-smi command - GPU Instance deletion example
MIG Function Disablement (Deactivation)
- Disable MIG and then reboot.Color mode
$ nvidia-smi -mig 0 Disabled MIG Mode for GPU 00000000:05:00.0 All done.$ nvidia-smi -mig 0 Disabled MIG Mode for GPU 00000000:05:00.0 All done.Code Block. nvidia-smi command - MIG disable Color mode$ nvidia-smi Mon Sep 30 05:18:28 2021 +-----------------------------------------------------------------------------+ | NVIDIA-SMI 470.57.02 Driver Version: 470.57.02 CUDA Version: 11.4 | |-------------------------------+----------------------+----------------------| | GPU Name Persistence-M| Bus-Id Disp.A | Volatile Uncorr. ECC | | Fan Temp Perf Pwr:Usage/Cap| Memory-Usage | GPU-Util Compute M. | | | | MIG M. | |===============================+======================+======================| | 0 NVDIA A100-SXM... Off | 00000000:05:00.0 Off | 0 | | N/A 33C P0 60W / 400W | 0MiB / 81251MiB | 0% Default | | | | Disabled | +-------------------------------+----------------------+----------------------+ +-----------------------------------------------------------------------------+ | MIG devices: | +-----------------------------------------------------------------------------+ | GPU GI CI MIG | Memory-Usage | Vol| Shared | | ID ID Dev | BAR1-Usage | SM Unc| CE ENC DEC OFA JPG| | | | ECC| | |=============================================================================| | No MIG devices found | +-----------------------------------------------------------------------------+ +-----------------------------------------------------------------------------+ | Processes: | | GPU GI CI PID Type Process name GPU Memory | | ID ID Usage | |=============================================================================| | No running processes found | +-----------------------------------------------------------------------------+$ nvidia-smi Mon Sep 30 05:18:28 2021 +-----------------------------------------------------------------------------+ | NVIDIA-SMI 470.57.02 Driver Version: 470.57.02 CUDA Version: 11.4 | |-------------------------------+----------------------+----------------------| | GPU Name Persistence-M| Bus-Id Disp.A | Volatile Uncorr. ECC | | Fan Temp Perf Pwr:Usage/Cap| Memory-Usage | GPU-Util Compute M. | | | | MIG M. | |===============================+======================+======================| | 0 NVDIA A100-SXM... Off | 00000000:05:00.0 Off | 0 | | N/A 33C P0 60W / 400W | 0MiB / 81251MiB | 0% Default | | | | Disabled | +-------------------------------+----------------------+----------------------+ +-----------------------------------------------------------------------------+ | MIG devices: | +-----------------------------------------------------------------------------+ | GPU GI CI MIG | Memory-Usage | Vol| Shared | | ID ID Dev | BAR1-Usage | SM Unc| CE ENC DEC OFA JPG| | | | ECC| | |=============================================================================| | No MIG devices found | +-----------------------------------------------------------------------------+ +-----------------------------------------------------------------------------+ | Processes: | | GPU GI CI PID Type Process name GPU Memory | | ID ID Usage | |=============================================================================| | No running processes found | +-----------------------------------------------------------------------------+Code Block. nvidia-smi command - Check GPU status
3.2.3 - Using NVSwitch on GPU Server
After creating a GPU Server, you can enable the NVSwitch feature in the GPU Server’s VM (Guest OS) and quickly use P2P (GPU to GPU) communication between GPUs.
Exploring NVIDIA NVSwitch for Multi GPU
NVIDIA A100 GPU server is a multi-GPU based on the NVIDIA Ampere architecture, with 8 Ampere 80 GB GPUs installed on the baseboard. The GPUs installed on the baseboard are connected to 6 NVSwitches via NVLink ports. Communication between GPUs on the baseboard is done using the full 600 GBps bandwidth. For this reason, the 8 GPUs installed on the A100 GPU server can be connected and operated like one, thereby maximizing GPU-to-GPU usage.
- NVLink(25 GBps) 12 Lane 8 GPU configuration
- NVSwitch(600 GBps) 6 units 8 GPU configuration diagram
Create GPU NVSwitch
To use the GPU NVSwitch feature, create a GPU Server service on the Samsung Cloud Platform, create a VM Instance (GuestOS) with 8 A100 GPUs assigned, and activate the Fabricmanager.
- NVSwitch can only be activated and used for products with 8 A100 GPUs assigned to a single GPU server (g1v128a8 (vCPU 128 | Memory 1920G | A100(80GB)*8)).
- Currently, GPU Server created with Windows OS does not support NVSwitch (Fabricmanager).
NVSwitch Installation and Operation Check (Fabric Manager Activation)
To operate NVSwitch, install Fabricmanager on the GPU Instance and follow the next procedure.
Install NVIDIA GPU Driver (470.52.02 Version) on the GPU server.
Color mode$ add-apt-repository ppa:graphics-drivers/ppa $ apt-get update $ apt-get install nvidia-driver-470-server$ add-apt-repository ppa:graphics-drivers/ppa $ apt-get update $ apt-get install nvidia-driver-470-serverCode Block. NVIDIA GPU Driver Installation Install and run NVIDIA Fabric Manager (470 Version) on the GPU server (For NVSwitch).
Color mode$ apt-get install cuda-drivers-fabricmanager-470 $ systemctl enable nvidia-fabricmanager $ systemctl start nvidia-fabricmanager$ apt-get install cuda-drivers-fabricmanager-470 $ systemctl enable nvidia-fabricmanager $ systemctl start nvidia-fabricmanagerCode Block. NVIDIA Fabric Manager Installation and Operation Check the status of NVIDIA Fabric Manager running on the GPU server.
- Normal operation indicates active (running)Color mode
$ systemctl status nvidia-fabricmanager$ systemctl status nvidia-fabricmanagerCode Block. Check NVIDIA Fabric Manager Operation Status
- Normal operation indicates active (running)
- Check the NVSwitch operation status on the GPU server.
- Normal operation indicates NV12Color mode
$ nvidia-smi topo --matrix$ nvidia-smi topo --matrixCode block. NVSwitch operation status check
- Normal operation indicates NV12
3.2.4 - Keypair Management
The user can enter the required information for the Keypair within the GPU Server service through the Samsung Cloud Platform Console, select detailed options, and create the service.
Keypair Create
You can create and use the Keypair service while using the GPU Server service on the Samsung Cloud Platform Console.
To create a keypair, follow these steps.
- All Services > Compute > GPU Server Click the menu. Navigate to the Service Home page of GPU Server.
- Click the Keypair menu on the Service Home page. It will go to the Keypair List page.
- Click the Keypair List page’s Create Keypair button. It navigates to the Create Keypair page.
- Service Information Input Enter the required information in the area.
Category Required or notDetailed description Keypair name Required Name of the Keypair to create - Enter within 255 characters using English letters, numbers, spaces, and special characters (
-,_)
Keypair type Required ssh Table. Keypair Service Information Input Items - Enter within 255 characters using English letters, numbers, spaces, and special characters (
- Additional Information Input Enter or select the required information in the area.
Category Required or notDetailed description Tag Select Add Tag - Up to 50 can be added per resource
- After clicking the Add Tag button, enter or select Key, Value values
Table. Keypair additional information input itemsCaution- After creation is complete, you can download the Key only once for the first time. Since reissuance is not possible, make sure it has been downloaded.
- Save the downloaded Private Key in a safe place.
- Service Information Input Enter the required information in the area.
- Check the input information and click the Complete button.
- When creation is complete, check the created resource on the Keypair List page.
Keypair Check detailed information
Keypair service can view and edit the full resource list and detailed information. Keypair Details page consists of Details, Tags, Activity History tabs.
To view detailed information about the Keypair, follow the steps below.
- All Services > Compute > GPU Server Click the menu. Navigate to the GPU Server’s Service Home page.
- Service Home page, click the Keypair menu. Navigate to the Keypair list page.
- Keypair List Click the resource to view detailed information on the page. Go to the Keypair Detail page.
- Keypair Details page displays status information and additional feature information, and consists of Details, Tags, Activity History tabs.
Detailed Information
On the Keypair List page, you can view detailed information of the selected resource and, if necessary, edit the information.
| Category | Detailed description |
|---|---|
| Service | Service Name |
| Resource Type | Resource Type |
| SRN | Unique resource ID in Samsung Cloud Platform
|
| Resource Name | Keypair Name |
| Resource ID | Keypair’s unique resource ID |
| Creator | User who created the Keypair |
| Creation time | Time when the keypair was created |
| Editor | User who modified the Keypair information |
| Modification Date/Time | Timestamp of Keypair information modification |
| Keypair name | Keypair name |
| Fingerprint | Unique value for identifying a Key |
| User ID | User ID of the user who created the Keypair |
| Public Key | Public Key Information |
Tag
Keypair list page allows you to view the tag information of selected resources, and you can add, modify, or delete them.
| Category | Detailed description |
|---|---|
| Tag List | Tag List
|
Work History
On the Keypair list page, you can view the operation history of the selected resource.
| Category | Detailed description |
|---|---|
| Work History List | Resource Change History
|
Keypair Resource Management
Describes the control and management functions of the keypair.
Get public key
To retrieve the public key, follow the steps below.
Click the All Services > Compute > GPU Server menu. Go to the GPU Server’s Service Home page.
Service Home page, click the Keypair menu. Navigate to the Keypair list page.
In the Keypair list page, click the More button at the top, then click the Import Public Key button. You will be taken to the Import Public Key page.
- Required Information Input Enter or select the required information in the area.
Category RequiredDetailed description Keypair name required Name of the Keypair to create Keypair type required ssh Public Key Required Enter Public Key - Load File: Attach File button to select and attach the public key file
- Only files with the following extension (.pem) can be attached
- Enter Public Key: Paste the copied public key value
- Public key value can be copied from the Keypair Details page
Table. Required input items for fetching public key - Load File: Attach File button to select and attach the public key file
- Required Information Input Enter or select the required information in the area.
Check the entered information and click the Complete button.
- Once creation is complete, check the created resource on the Keypair List page.
Delete Keypair
You can delete unused Keypairs. However, once a Keypair is deleted it cannot be recovered, so please review the impact thoroughly in advance before proceeding with deletion.
To delete a keypair, follow the steps below.
- All Services > Compute > GPU Server Click the menu. Navigate to the GPU Server’s Service Home page.
- Click the Keypair menu on the Service Home page. It moves to the Keypair List page.
- On the Keypair list page, select the resource to delete and click the Delete button.
- On the Keypair list page, select multiple Keypair check boxes, and click the Delete button at the top of the resource list.
- After deletion is complete, check on the Keypair List page whether the resource has been deleted.
3.2.5 - ServiceWatch Agent Installation
Users can install ServiceWatch Agent on GPU Server to collect custom metrics and logs.
ServiceWatch Agent
There are two main types of Agents that must be installed to collect ServiceWatch custom metrics and logs on GPU Server. Prometheus Exporter and Open Telemetry Collector.
| Category | Detailed Description | |
|---|---|---|
| Prometheus Exporter | Provides metrics of specific applications or services in a format that Prometheus can scrape
| |
| Open Telemetry Collector | A centralized collector that collects telemetry data such as metrics and logs from distributed systems, processes (filtering, sampling, etc.), and sends them to multiple backends (e.g., Prometheus, Jaeger, Elasticsearch, etc.)
|
If Kubernetes Engine is configured on GPU Server, please check GPU metrics through the metrics provided by Kubernetes Engine.
- If you install DCGM Exporter on a GPU Server with Kubernetes Engine configured, it may not work properly.
Prerequisites for Using ServiceWatch Agent
Step 1. Check NVIDIA Driver installation
Install Prometheus Exporter for GPU Metrics (for Ubuntu)
Install the Prometheus Exporter to collect GPU Server metrics in the following order.
Check NVDIA Driver Installation
- Check the installed NVDIA Driver.Color mode
nvidia-smi --query-gpu driver_version --format csvnvidia-smi --query-gpu driver_version --format csvCode block. Check NVDIA Driver version command Color modedriver_version 535.183.06 ... 535.183.06driver_version 535.183.06 ... 535.183.06Code block. Check NVDIA Driver version example
Install NVSwitch Configuration and Query (NSCQ) Library
Install cuda-keyring.
Color modewget https://developer.download.nvidia.com/compute/cuda/repos/<distro>/<arch>/cuda-keyring_1.1-1_all.debwget https://developer.download.nvidia.com/compute/cuda/repos/<distro>/<arch>/cuda-keyring_1.1-1_all.debCode block. NSCQ library download command Color modesudo dpkg -i cuda-keyring_1.1-1_all.deb apt updatesudo dpkg -i cuda-keyring_1.1-1_all.deb apt updateCode block. NSCQ library installation command Color modenvidia-smi --query-gpu driver_version --format csvnvidia-smi --query-gpu driver_version --format csvCode block. Check NVDIA Driver version command Color modedriver_version 535.183.06 ... 535.183.06driver_version 535.183.06 ... 535.183.06Code block. Check NVDIA Driver version example Install libnvidia-nscq.
Color modeapt-cache policy libnvidia-nscq-535apt-cache policy libnvidia-nscq-535Code block. NSCQ library apt-cache command Color modelibnvidia-nscq-535: Installed: (none) Candidate: 535.247.01-1 Version table: 535.247.01-1 600 600 https://developer.download.nvidia.com/compute/cuda/repos/ubuntu2204/x86_64 Packages ... 600 https://developer.download.nvidia.com/compute/cuda/repos/ubuntu2204/x86_64 Packages 535.216.01-1 600 600 https://developer.download.nvidia.com/compute/cuda/repos/ubuntu2204/x86_64 Packages 535.183.06-1 600 # Install the version that matches the Driver 600 https://developer.download.nvidia.com/compute/cuda/repos/ubuntu2204/x86_64 Packages 535.183.01-1 600 ... 600 https://developer.download.nvidia.com/compute/cuda/repos/ubuntu2204/x86_64 Packages 535.54.03-1 600 600 https://developer.download.nvidia.com/compute/cuda/repos/ubuntu2204/x86_64 Packageslibnvidia-nscq-535: Installed: (none) Candidate: 535.247.01-1 Version table: 535.247.01-1 600 600 https://developer.download.nvidia.com/compute/cuda/repos/ubuntu2204/x86_64 Packages ... 600 https://developer.download.nvidia.com/compute/cuda/repos/ubuntu2204/x86_64 Packages 535.216.01-1 600 600 https://developer.download.nvidia.com/compute/cuda/repos/ubuntu2204/x86_64 Packages 535.183.06-1 600 # Install the version that matches the Driver 600 https://developer.download.nvidia.com/compute/cuda/repos/ubuntu2204/x86_64 Packages 535.183.01-1 600 ... 600 https://developer.download.nvidia.com/compute/cuda/repos/ubuntu2204/x86_64 Packages 535.54.03-1 600 600 https://developer.download.nvidia.com/compute/cuda/repos/ubuntu2204/x86_64 PackagesCode block. NSCQ library apt-cache command result Color modeapt install libnvidia-nscq-535=535.183.06-1apt install libnvidia-nscq-535=535.183.06-1Code block. NSCQ library installation command NoticeYou must install the same version as the NVDIA Driver version.
- Example) driver version: 535.183.06, libnvdia-nscq version: 535.183.06-1
Install NVSwitch Device Monitoring API(NVSDM) Library
- Install NVSDM library.Color mode
apt-cache policy libnvsdmapt-cache policy libnvsdmCode block. NVSDM library apt-cache command Color modelibnvsdm: Installed: (none) Candidate: 580.105.08-1 Version table: 580.105.08-1 600 600 https://developer.download.nvidia.com/compute/cuda/repos/ubuntu2204/x86_64 Packages 580.95.05-1 600 600 https://developer.download.nvidia.com/compute/cuda/repos/ubuntu2204/x86_64 Packages 580.82.07-1 600 600 https://developer.download.nvidia.com/compute/cuda/repos/ubuntu2204/x86_64 Packages 580.65.06-1 600 600 https://developer.download.nvidia.com/compute/cuda/repos/ubuntu2204/x86_64 Packageslibnvsdm: Installed: (none) Candidate: 580.105.08-1 Version table: 580.105.08-1 600 600 https://developer.download.nvidia.com/compute/cuda/repos/ubuntu2204/x86_64 Packages 580.95.05-1 600 600 https://developer.download.nvidia.com/compute/cuda/repos/ubuntu2204/x86_64 Packages 580.82.07-1 600 600 https://developer.download.nvidia.com/compute/cuda/repos/ubuntu2204/x86_64 Packages 580.65.06-1 600 600 https://developer.download.nvidia.com/compute/cuda/repos/ubuntu2204/x86_64 PackagesCode block. NVSDM library apt-cache command result Color modeapt install libnvsdm=580.105.08-1apt install libnvsdm=580.105.08-1Code block. NVSDM library installation
Install NVIDIA DCGM (for Ubuntu)
Install DCGM Exporter in the following order.
- Install datacenter-gpu-manager-exporter
- Enable and start DCGM service
- For metrics that can be collected with GPU DCGM Exporter and configuration methods, refer to DCGM Exporter Metrics.
Install DCGM(datacenter-gpu-manager)
Refers to a specific version of NVIDIA’s Data Center GPU Manager (DCGM) tool, which is a package for managing and monitoring NVIDIA data center GPUs. In particular, cuda12 indicates that this management tool is installed for CUDA 12 version, and datacenter-gpu-manager-4 means DCGM version 4.x. This tool provides various functions including GPU status monitoring, diagnostics, alert systems, and power/clock management.
- Check CUDA version.Color mode
nvidia-smi | grep CUDAnvidia-smi | grep CUDACode block. Check CUDA version Color mode| NVIDIA-SMI 535.183.06 Driver Version: 535.183.06 CUDA Version: 12.2 || NVIDIA-SMI 535.183.06 Driver Version: 535.183.06 CUDA Version: 12.2 |Code block. Check CUDA version result example Color modeCUDA_VERSION=12CUDA_VERSION=12Code block. Set CUDA version command - Install datacenter-gpu-manager-cuda.Color mode
apt install datacenter-gpu-manager-4-cuda${CUDA_VERSION}apt install datacenter-gpu-manager-4-cuda${CUDA_VERSION}Code block. datacenter-gpu-manager-cuda installation command
Install datacenter-gpu-manager-exporter
This is a tool that collects various GPU metrics such as GPU usage, memory usage, temperature, and power consumption based on NVIDIA Data Center GPU Manager (DCGM) and exposes them for use in monitoring systems like Prometheus.
- Install datacenter-gpu-manager-exporter.Color mode
apt install datacenter-gpu-manager-exporterapt install datacenter-gpu-manager-exporterCode block. datacenter-gpu-manager-exporter installation command - Check DCGM Exporter configuration file.Color mode
cat /usr/lib/systemd/system/nvidia-dcgm-exporter.service | grep ExecStartcat /usr/lib/systemd/system/nvidia-dcgm-exporter.service | grep ExecStartCode block. Check datacenter-gpu-manager-exporter configuration file command Color modeExecStart=/usr/bin/dcgm-exporter -f /etc/dcgm-exporter/default-counters.csvExecStart=/usr/bin/dcgm-exporter -f /etc/dcgm-exporter/default-counters.csvCode block. Check datacenter-gpu-manager-exporter configuration file result example - Check the configuration provided at DCGM Exporter installation, remove
#for necessary metrics, and add#for unnecessary metrics.Color modevi /etc/dcgm-exporter/default-counters.csv ## Example ## ... DCGM_FI_PROF_PIPE_TENSOR_ACTIVE, gauge, Ratio of cycles the tensor (HMMA) pipe is active. DCGM_FI_PROF_DRAM_ACTIVE, gauge, Ratio of cycles the device memory interface is active sending or receiving data. # DCGM_FI_PROF_PIPE_FP64_ACTIVE, gauge, Ratio of cycles the fp64 pipes are active. # DCGM_FI_PROF_PIPE_FP32_ACTIVE, gauge, Ratio of cycles the fp32 pipes are active. ...vi /etc/dcgm-exporter/default-counters.csv ## Example ## ... DCGM_FI_PROF_PIPE_TENSOR_ACTIVE, gauge, Ratio of cycles the tensor (HMMA) pipe is active. DCGM_FI_PROF_DRAM_ACTIVE, gauge, Ratio of cycles the device memory interface is active sending or receiving data. # DCGM_FI_PROF_PIPE_FP64_ACTIVE, gauge, Ratio of cycles the fp64 pipes are active. # DCGM_FI_PROF_PIPE_FP32_ACTIVE, gauge, Ratio of cycles the fp32 pipes are active. ...Code block. datacenter-gpu-manager-exporter metric configuration example ReferenceStep 1. Check NVIDIA Driver installation<brCautionMetric collection through ServiceWatch Agent is classified as custom metrics, and unlike basic metrics collected from each service, fees are charged. Therefore, unnecessary metric collection should be removed or disabled to avoid excessive charges.
Enable and Start DCGM Service
Enable and start nvdia-dcgm service.
Color modesystemctl enable --now nvidia-dcgmsystemctl enable --now nvidia-dcgmCode block. Enable and start nvdia-dcgm service command Enable and start nvdia-dcgm-exporter service.
Color modesystemctl enable --now nvidia-dcgm-exportersystemctl enable --now nvidia-dcgm-exporterCode block. Enable and start nvdia-dcgm-exporter service command
For details, refer to ServiceWatch > Use ServiceWatch Agent.
Install Prometheus Exporter for GPU Metrics (for RHEL)
Install ServiceWatch Agent to collect GPU Server metrics in the following order.
Check NVDIA Driver Installation (for RHEL)
- Check the installed NVDIA Driver.Color mode
nvidia-smi --query-gpu driver_version --format csvnvidia-smi --query-gpu driver_version --format csvCode block. Check NVDIA Driver version command Color modedriver_version 535.183.06 ... 535.183.06driver_version 535.183.06 ... 535.183.06Code block. Check NVDIA Driver version example
Install NVSwitch Configuration and Query (NSCQ) Library (for RHEL)
NVSwitch Configuration and Query (NSCQ) Library is required for Hopper or earlier Generation GPUs.
- For RHEL, check if libnvdia-nscq is installed and then install it.
Check libnvdia-nscq package.
Color moderpm -qa | grep libnvidia-nscq libnvidia-nscq-535-535.183.06-1.x86_64rpm -qa | grep libnvidia-nscq libnvidia-nscq-535-535.183.06-1.x86_64Code block. Check NSCQ library package Add CUDA Repository to DNF.
Color modednf config-manager --add-repo https://developer.download.nvidia.com/compute/cuda/repos/rhel8/x86_64/cuda-rhel8.repodnf config-manager --add-repo https://developer.download.nvidia.com/compute/cuda/repos/rhel8/x86_64/cuda-rhel8.repoCode block. Add DNF Repository Reset NVDIA Driver status
Color modednf module reset nvidia-driverdnf module reset nvidia-driverCode block. Reset NVIDIA Driver DNF module status Color modeUpdating Subscription Management repositories. Last metadata expiration check: 0:03:15 ago on Wed 19 Nov 2025 01:23:48 AM EST. Dependencies resolved. ============================================= Package Architecture Version Repository Size ============================================= Disabling module profiles: nvidia-driver/default nvidia-driver/fm Resetting modules: nvidia-driver Transaction Summary ============================================= Is this ok [y/N]: yUpdating Subscription Management repositories. Last metadata expiration check: 0:03:15 ago on Wed 19 Nov 2025 01:23:48 AM EST. Dependencies resolved. ============================================= Package Architecture Version Repository Size ============================================= Disabling module profiles: nvidia-driver/default nvidia-driver/fm Resetting modules: nvidia-driver Transaction Summary ============================================= Is this ok [y/N]: yCode block. Reset NVIDIA Driver DNF module status result example Enable NVDIA Driver module.
Color modednf module enable nvidia-driver:535-opendnf module enable nvidia-driver:535-openCode block. Enable NVDIA Driver module Color modeUpdating Subscription Management repositories. Last metadata expiration check: 0:04:22 ago on Wed 19 Nov 2025 01:23:48 AM EST. Dependencies resolved. ============================================= Package Architecture Version Repository Size ============================================= Enabling module streams: nvidia-driver 535-open Transaction Summary ============================================= Is this ok [y/N]: yUpdating Subscription Management repositories. Last metadata expiration check: 0:04:22 ago on Wed 19 Nov 2025 01:23:48 AM EST. Dependencies resolved. ============================================= Package Architecture Version Repository Size ============================================= Enabling module streams: nvidia-driver 535-open Transaction Summary ============================================= Is this ok [y/N]: yCode block. Enable NVDIA Driver module result example Check libnvdia-nscq module list.
Color modednf list libnvidia-nscq-535 --showduplicatesdnf list libnvidia-nscq-535 --showduplicatesCode block. Check libnvdia-nscq module list Install libnvdia-nscq.
Color modednf install libnvidia-nscq-535-535.183.06-1dnf install libnvidia-nscq-535-535.183.06-1Code block. libnvdia-nscq installation command
Install NVSwitch Device Monitoring API(NVSDM) Library (for RHEL)
Check NVSDM library module list.
Color modednf list libnvsdm --showduplicatesdnf list libnvsdm --showduplicatesCode block. Check NVSDM library module list Color modelibnvsdm: Installed: (none) Candidate: 580.105.08-1 Version table: 580.105.08-1 600 600 https://developer.download.nvidia.com/compute/cuda/repos/ubuntu2204/x86_64 Packages 580.95.05-1 600 600 https://developer.download.nvidia.com/compute/cuda/repos/ubuntu2204/x86_64 Packages 580.82.07-1 600 600 https://developer.download.nvidia.com/compute/cuda/repos/ubuntu2204/x86_64 Packages 580.65.06-1 600 600 https://developer.download.nvidia.com/compute/cuda/repos/ubuntu2204/x86_64 Packageslibnvsdm: Installed: (none) Candidate: 580.105.08-1 Version table: 580.105.08-1 600 600 https://developer.download.nvidia.com/compute/cuda/repos/ubuntu2204/x86_64 Packages 580.95.05-1 600 600 https://developer.download.nvidia.com/compute/cuda/repos/ubuntu2204/x86_64 Packages 580.82.07-1 600 600 https://developer.download.nvidia.com/compute/cuda/repos/ubuntu2204/x86_64 Packages 580.65.06-1 600 600 https://developer.download.nvidia.com/compute/cuda/repos/ubuntu2204/x86_64 PackagesCode block. Check NVSDM library module list result example Install libnvsdm.
Color modednf install libnvsdm-580.105.08-1dnf install libnvsdm-580.105.08-1Code block. NVSDM library installation Color modeUpdating Subscription Management repositories. Last metadata expiration check: 0:08:18 ago on Wed 19 Nov 2025 01:05:28 AM EST. Dependencies resolved. ======================================================================== Package Architecture Version Repository Size ======================================================================== Installing: libnvsdm x86_64 580.105.08-1 cuda-rhel8-x86_64 675 k Installing dependencies: infiniband-diags x86_64 48.0-1.el8 rhel-8-for-x86_64-baseos-rpms 323 k libibumad x86_64 48.0-1.el8 rhel-8-for-x86_64-baseos-rpms 34 k Transaction Summary ======================================================================== Install 3 Packages Total download size: 1.0 M Installed size: 3.2 M Is this ok [y/N]: yUpdating Subscription Management repositories. Last metadata expiration check: 0:08:18 ago on Wed 19 Nov 2025 01:05:28 AM EST. Dependencies resolved. ======================================================================== Package Architecture Version Repository Size ======================================================================== Installing: libnvsdm x86_64 580.105.08-1 cuda-rhel8-x86_64 675 k Installing dependencies: infiniband-diags x86_64 48.0-1.el8 rhel-8-for-x86_64-baseos-rpms 323 k libibumad x86_64 48.0-1.el8 rhel-8-for-x86_64-baseos-rpms 34 k Transaction Summary ======================================================================== Install 3 Packages Total download size: 1.0 M Installed size: 3.2 M Is this ok [y/N]: yCode block. NVSDM library installation command result example
Install NVIDIA DCGM (for RHEL)
Install Node Exporter in the following order.
- Install datacenter-gpu-manager-exporter
- Enable and start DCGM service
- For metrics that can be collected with GPU DCGM Exporter and configuration methods, refer to DCGM Exporter Metrics.
Install DCGM(datacenter-gpu-manager) (for RHEL)
Refers to a specific version of NVIDIA’s Data Center GPU Manager (DCGM) tool, which is a package for managing and monitoring NVIDIA data center GPUs. In particular, cuda12 indicates that this management tool is installed for CUDA 12 version, and datacenter-gpu-manager-4 means DCGM version 4.x. This tool provides various functions including GPU status monitoring, diagnostics, alert systems, and power/clock management.
- Add CUDA Repository to DNF.Color mode
dnf config-manager --add-repo https://developer.download.nvidia.com/compute/cuda/repos/rhel8/x86_64/cuda-rhel8.repodnf config-manager --add-repo https://developer.download.nvidia.com/compute/cuda/repos/rhel8/x86_64/cuda-rhel8.repoCode block. Add DNF Repository - Check CUDA version.Color mode
nvidia-smi | grep CUDAnvidia-smi | grep CUDACode block. Check CUDA version Color mode| NVIDIA-SMI 535.183.06 Driver Version: 535.183.06 CUDA Version: 12.2 || NVIDIA-SMI 535.183.06 Driver Version: 535.183.06 CUDA Version: 12.2 |Code block. Check CUDA version result example Color modeCUDA_VERSION=12CUDA_VERSION=12Code block. Set CUDA version command - Check datacenter-gpu-manager-cuda module list.Color mode
dnf list datacenter-gpu-manager-4-cuda${CUDA_VERSION} --showduplicatesdnf list datacenter-gpu-manager-4-cuda${CUDA_VERSION} --showduplicatesCode block. Check datacenter-gpu-manager-cuda module list Color modeUpdating Subscription Management repositories. Unable to read consumer identity This system is not registered with an entitlement server. You can use subscription-manager to register. Last metadata expiration check: 0:00:34 ago on Wed 19 Nov 2025 12:26:56 AM EST. Available Packages datacenter-gpu-manager-4-cuda12.x86_64 1:4.0.0-1 cuda-rhel8-x86_64 datacenter-gpu-manager-4-cuda12.x86_64 1:4.1.0-1 cuda-rhel8-x86_64 datacenter-gpu-manager-4-cuda12.x86_64 1:4.1.1-1 cuda-rhel8-x86_64 datacenter-gpu-manager-4-cuda12.x86_64 1:4.2.0-1 cuda-rhel8-x86_64 datacenter-gpu-manager-4-cuda12.x86_64 1:4.2.2-1 cuda-rhel8-x86_64 datacenter-gpu-manager-4-cuda12.x86_64 1:4.2.3-1 cuda-rhel8-x86_64 datacenter-gpu-manager-4-cuda12.x86_64 1:4.2.3-2 cuda-rhel8-x86_64 datacenter-gpu-manager-4-cuda12.x86_64 1:4.3.0-1 cuda-rhel8-x86_64 datacenter-gpu-manager-4-cuda12.x86_64 1:4.3.1-1 cuda-rhel8-x86_64 datacenter-gpu-manager-4-cuda12.x86_64 1:4.4.0-1 cuda-rhel8-x86_64 datacenter-gpu-manager-4-cuda12.x86_64 1:4.4.1-1 cuda-rhel8-x86_64 datacenter-gpu-manager-4-cuda12.x86_64 1:4.4.2-1 cuda-rhel8-x86_64Updating Subscription Management repositories. Unable to read consumer identity This system is not registered with an entitlement server. You can use subscription-manager to register. Last metadata expiration check: 0:00:34 ago on Wed 19 Nov 2025 12:26:56 AM EST. Available Packages datacenter-gpu-manager-4-cuda12.x86_64 1:4.0.0-1 cuda-rhel8-x86_64 datacenter-gpu-manager-4-cuda12.x86_64 1:4.1.0-1 cuda-rhel8-x86_64 datacenter-gpu-manager-4-cuda12.x86_64 1:4.1.1-1 cuda-rhel8-x86_64 datacenter-gpu-manager-4-cuda12.x86_64 1:4.2.0-1 cuda-rhel8-x86_64 datacenter-gpu-manager-4-cuda12.x86_64 1:4.2.2-1 cuda-rhel8-x86_64 datacenter-gpu-manager-4-cuda12.x86_64 1:4.2.3-1 cuda-rhel8-x86_64 datacenter-gpu-manager-4-cuda12.x86_64 1:4.2.3-2 cuda-rhel8-x86_64 datacenter-gpu-manager-4-cuda12.x86_64 1:4.3.0-1 cuda-rhel8-x86_64 datacenter-gpu-manager-4-cuda12.x86_64 1:4.3.1-1 cuda-rhel8-x86_64 datacenter-gpu-manager-4-cuda12.x86_64 1:4.4.0-1 cuda-rhel8-x86_64 datacenter-gpu-manager-4-cuda12.x86_64 1:4.4.1-1 cuda-rhel8-x86_64 datacenter-gpu-manager-4-cuda12.x86_64 1:4.4.2-1 cuda-rhel8-x86_64Code block. Check datacenter-gpu-manager-cuda module list result example - Install datacenter-gpu-manager-cuda.Color mode
dnf install datacenter-gpu-manager-4-cuda${CUDA_VERSION}dnf install datacenter-gpu-manager-4-cuda${CUDA_VERSION}Code block. datacenter-gpu-manager-cuda installation Color modeUpdating Subscription Management repositories. Unable to read consumer identity This system is not registered with an entitlement server. You can use subscription-manager to register. Last metadata expiration check: 0:07:12 ago on Wed 19 Nov 2025 12:26:56 AM EST. Dependencies resolved. =================================================================================================== Package Architecture Version Repository Size =================================================================================================== Installing: datacenter-gpu-manager-4-cuda12 x86_64 1:4.4.2-1 cuda-rhel8-x86_64 554 M Installing dependencies: datacenter-gpu-manager-4-core x86_64 1:4.4.2-1 cuda-rhel8-x86_64 9.9 M Installing weak dependencies: datacenter-gpu-manager-4-proprietary x86_64 1:4.4.2-1 cuda-rhel8-x86_64 5.3 M datacenter-gpu-manager-4-proprietary-cuda12 x86_64 1:4.4.2-1 cuda-rhel8-x86_64 289 M Transaction Summary ==================================================================================================== Install 4 Packages ... Is this ok [y/N]: yUpdating Subscription Management repositories. Unable to read consumer identity This system is not registered with an entitlement server. You can use subscription-manager to register. Last metadata expiration check: 0:07:12 ago on Wed 19 Nov 2025 12:26:56 AM EST. Dependencies resolved. =================================================================================================== Package Architecture Version Repository Size =================================================================================================== Installing: datacenter-gpu-manager-4-cuda12 x86_64 1:4.4.2-1 cuda-rhel8-x86_64 554 M Installing dependencies: datacenter-gpu-manager-4-core x86_64 1:4.4.2-1 cuda-rhel8-x86_64 9.9 M Installing weak dependencies: datacenter-gpu-manager-4-proprietary x86_64 1:4.4.2-1 cuda-rhel8-x86_64 5.3 M datacenter-gpu-manager-4-proprietary-cuda12 x86_64 1:4.4.2-1 cuda-rhel8-x86_64 289 M Transaction Summary ==================================================================================================== Install 4 Packages ... Is this ok [y/N]: yCode block. datacenter-gpu-manager-cuda installation result example
Install datacenter-gpu-manager-exporter (for RHEL)
This is a tool that collects various GPU metrics such as GPU usage, memory usage, temperature, and power consumption based on NVIDIA Data Center GPU Manager (DCGM) and exposes them for use in monitoring systems like Prometheus.
Add CUDA Repository to DNF. (If you have already run this command, proceed to the next step.)
Color modednf config-manager --add-repo https://developer.download.nvidia.com/compute/cuda/repos/rhel8/x86_64/cuda-rhel8.repodnf config-manager --add-repo https://developer.download.nvidia.com/compute/cuda/repos/rhel8/x86_64/cuda-rhel8.repoCode block. Add DNF Repository Check CUDA version. (If you have already run this command, proceed to the next step.)
Color modenvidia-smi | grep CUDAnvidia-smi | grep CUDACode block. Check CUDA version Color mode| NVIDIA-SMI 535.183.06 Driver Version: 535.183.06 CUDA Version: 12.2 || NVIDIA-SMI 535.183.06 Driver Version: 535.183.06 CUDA Version: 12.2 |Code block. Check CUDA version result example Color modeCUDA_VERSION=12CUDA_VERSION=12Code block. Set CUDA version command Check datacenter-gpu-manager-exporter module list.
Color modednf list datacenter-gpu-manager-exporter --showduplicatesdnf list datacenter-gpu-manager-exporter --showduplicatesCode block. Check datacenter-gpu-manager-exporter module list Color modeUpdating Subscription Management repositories. Unable to read consumer identity This system is not registered with an entitlement server. You can use subscription-manager to register. Last metadata expiration check: 0:02:11 ago on Wed 19 Nov 2025 12:26:56 AM EST. Available Packages datacenter-gpu-manager-exporter.x86_64 4.0.1-1 cuda-rhel8-x86_64 datacenter-gpu-manager-exporter.x86_64 4.1.0-1 cuda-rhel8-x86_64 datacenter-gpu-manager-exporter.x86_64 4.1.1-1 cuda-rhel8-x86_64 datacenter-gpu-manager-exporter.x86_64 4.1.3-1 cuda-rhel8-x86_64 datacenter-gpu-manager-exporter.x86_64 4.5.0-1 cuda-rhel8-x86_64 datacenter-gpu-manager-exporter.x86_64 4.5.1-1 cuda-rhel8-x86_64 datacenter-gpu-manager-exporter.x86_64 4.5.2-1 cuda-rhel8-x86_64 datacenter-gpu-manager-exporter.x86_64 4.6.0-1 cuda-rhel8-x86_64Updating Subscription Management repositories. Unable to read consumer identity This system is not registered with an entitlement server. You can use subscription-manager to register. Last metadata expiration check: 0:02:11 ago on Wed 19 Nov 2025 12:26:56 AM EST. Available Packages datacenter-gpu-manager-exporter.x86_64 4.0.1-1 cuda-rhel8-x86_64 datacenter-gpu-manager-exporter.x86_64 4.1.0-1 cuda-rhel8-x86_64 datacenter-gpu-manager-exporter.x86_64 4.1.1-1 cuda-rhel8-x86_64 datacenter-gpu-manager-exporter.x86_64 4.1.3-1 cuda-rhel8-x86_64 datacenter-gpu-manager-exporter.x86_64 4.5.0-1 cuda-rhel8-x86_64 datacenter-gpu-manager-exporter.x86_64 4.5.1-1 cuda-rhel8-x86_64 datacenter-gpu-manager-exporter.x86_64 4.5.2-1 cuda-rhel8-x86_64 datacenter-gpu-manager-exporter.x86_64 4.6.0-1 cuda-rhel8-x86_64Code block. Check datacenter-gpu-manager-exporter module list result example Install datacenter-gpu-manager-cuda. dcgm-exporter 4.5.X requires glibc 2.34 or higher, but RHEL9 provides glibc 2.34, so specify version 4.1.3-1 to install.
Color modednf install datacenter-gpu-manager-exporter-4.1.3-1dnf install datacenter-gpu-manager-exporter-4.1.3-1Code block. datacenter-gpu-manager-cuda installation Color modeUpdating Subscription Management repositories. Unable to read consumer identity This system is not registered with an entitlement server. You can use subscription-manager to register. Last metadata expiration check: 0:07:12 ago on Wed 19 Nov 2025 12:26:56 AM EST. Dependencies resolved. ==================================================================================================== Package Architecture Version Repository Size ==================================================================================================== Installing: datacenter-gpu-manager-exporter x86_64 4.1.3-1 cuda-rhel8-x86_64 26 M ... Is this ok [y/N]: yUpdating Subscription Management repositories. Unable to read consumer identity This system is not registered with an entitlement server. You can use subscription-manager to register. Last metadata expiration check: 0:07:12 ago on Wed 19 Nov 2025 12:26:56 AM EST. Dependencies resolved. ==================================================================================================== Package Architecture Version Repository Size ==================================================================================================== Installing: datacenter-gpu-manager-exporter x86_64 4.1.3-1 cuda-rhel8-x86_64 26 M ... Is this ok [y/N]: yCode block. datacenter-gpu-manager-cuda installation result example Color modecat /usr/lib/systemd/system/nvidia-dcgm-exporter.service | grep ExecStartcat /usr/lib/systemd/system/nvidia-dcgm-exporter.service | grep ExecStartCode block. datacenter-gpu-manager-exporter configuration file Color modeExecStart=/usr/bin/dcgm-exporter -f /etc/dcgm-exporter/default-counters.csvExecStart=/usr/bin/dcgm-exporter -f /etc/dcgm-exporter/default-counters.csvCode block. Check datacenter-gpu-manager-exporter configuration file result example Check the configuration provided at DCGM Exporter installation, remove
#for necessary metrics, and add#for unnecessary metrics.Color modevi /etc/dcgm-exporter/default-counters.csv ## Example ## ... DCGM_FI_PROF_PIPE_TENSOR_ACTIVE, gauge, Ratio of cycles the tensor (HMMA) pipe is active. DCGM_FI_PROF_DRAM_ACTIVE, gauge, Ratio of cycles the device memory interface is active sending or receiving data. # DCGM_FI_PROF_PIPE_FP64_ACTIVE, gauge, Ratio of cycles the fp64 pipes are active. # DCGM_FI_PROF_PIPE_FP32_ACTIVE, gauge, Ratio of cycles the fp32 pipes are active. ...vi /etc/dcgm-exporter/default-counters.csv ## Example ## ... DCGM_FI_PROF_PIPE_TENSOR_ACTIVE, gauge, Ratio of cycles the tensor (HMMA) pipe is active. DCGM_FI_PROF_DRAM_ACTIVE, gauge, Ratio of cycles the device memory interface is active sending or receiving data. # DCGM_FI_PROF_PIPE_FP64_ACTIVE, gauge, Ratio of cycles the fp64 pipes are active. # DCGM_FI_PROF_PIPE_FP32_ACTIVE, gauge, Ratio of cycles the fp32 pipes are active. ...Code block. datacenter-gpu-manager-exporter metric configuration example ReferenceFor metrics that can be collected with the GPU DCGM Exporter and configuration methods, refer to DCGM Exporter Metrics.CautionMetric collection through ServiceWatch Agent is classified as custom metrics, and unlike basic metrics collected from each service, fees are charged. Therefore, unnecessary metric collection should be removed or disabled to avoid excessive charges.
Enable and Start DCGM Service (for RHEL)
Enable and start nvdia-dcgm service.
Color modesystemctl enable --now nvidia-dcgmsystemctl enable --now nvidia-dcgmCode block. Enable and start nvdia-dcgm service command Enable and start nvdia-dcgm-exporter service.
Color modesystemctl enable --now nvidia-dcgm-exportersystemctl enable --now nvidia-dcgm-exporterCode block. Enable and start nvdia-dcgm-exporter service command
For details, refer to ServiceWatch > Use ServiceWatch Agent.
DCGM Exporter Metrics
DCGM Exporter Key Metrics
The key GPU metrics provided by DCGM Exporter are as follows.
| Category | DCGM Field | Prometheus Metric Type | Summary | |
|---|---|---|---|---|
| Clocks | DCGM_FI_DEV_SM_CLOCK | gauge | SM clock frequency (in MHz) | |
| Clocks | DCGM_FI_DEV_MEM_CLOCK | gauge | Memory clock frequency (in MHz) | |
| Temperature | DCGM_FI_DEV_GPU_TEMP | gauge | GPU temperature (in C) | |
| Power | DCGM_FI_DEV_POWER_USAGE | gauge | Power draw (in W) | |
| Utilization | DCGM_FI_DEV_GPU_UTIL | gauge | GPU utilization (in %) | |
| Utilization | DCGM_FI_DEV_MEM_COPY_UTIL | gauge | Memory utilization (in %) | |
| Memory Usage | DCGM_FI_DEV_FB_FREE | gauge | Frame buffer memory free (in MiB) | |
| Memory Usage | DCGM_FI_DEV_FB_USED | gauge | Frame buffer memory used (in MiB) | |
| Nvlink | DCGM_FI_DEV_NVLINK_BANDWIDTH_TOTAL(8 GPU only) | counter | Total number of NVLink bandwidth counters for all lanes |
DCGM Exporter Metrics Collection Configuration
For the metrics configured by default in DCGM Exporter, refer to DCGM Exporter > Default Metrics.
- For additional metrics to configure beyond the default settings, remove
#in default-counters.csv. - For metrics you do not want to collect among the default configured metrics, add
#or delete the item.
# Format
# If line starts with a '#' it is considered a comment
# DCGM FIELD, Prometheus metric type, help message
# Clocks
DCGM_FI_DEV_SM_CLOCK, gauge, SM clock frequency (in MHz).
DCGM_FI_DEV_MEM_CLOCK, gauge, Memory clock frequency (in MHz).
# Temperature
DCGM_FI_DEV_MEMORY_TEMP, gauge, Memory temperature (in C).
DCGM_FI_DEV_GPU_TEMP, gauge, GPU temperature (in C).
# Power
DCGM_FI_DEV_POWER_USAGE, gauge, Power draw (in W).
DCGM_FI_DEV_TOTAL_ENERGY_CONSUMPTION, counter, Total energy consumption since boot (in mJ).
# PCIE
# DCGM_FI_PROF_PCIE_TX_BYTES, counter, Total number of bytes transmitted through PCIe TX via NVML.
# DCGM_FI_PROF_PCIE_RX_BYTES, counter, Total number of bytes received through PCIe RX via NVML.
...# Format
# If line starts with a '#' it is considered a comment
# DCGM FIELD, Prometheus metric type, help message
# Clocks
DCGM_FI_DEV_SM_CLOCK, gauge, SM clock frequency (in MHz).
DCGM_FI_DEV_MEM_CLOCK, gauge, Memory clock frequency (in MHz).
# Temperature
DCGM_FI_DEV_MEMORY_TEMP, gauge, Memory temperature (in C).
DCGM_FI_DEV_GPU_TEMP, gauge, GPU temperature (in C).
# Power
DCGM_FI_DEV_POWER_USAGE, gauge, Power draw (in W).
DCGM_FI_DEV_TOTAL_ENERGY_CONSUMPTION, counter, Total energy consumption since boot (in mJ).
# PCIE
# DCGM_FI_PROF_PCIE_TX_BYTES, counter, Total number of bytes transmitted through PCIe TX via NVML.
# DCGM_FI_PROF_PCIE_RX_BYTES, counter, Total number of bytes received through PCIe RX via NVML.
...3.3 - API Reference
3.4 - CLI Reference
3.5 - Release Note
GPU Server
- ServiceWatch service integration provision
- You can monitor data through the ServiceWatch service.
- You can select a RHEL image when creating a GPU Server.
- Keypair management feature has been added.
- You can create a keypair to use, or retrieve a public key and apply it.
- GPU Server feature addition
- IP, Public NAT IP, Private NAT IP configuration feature has been added.
- LLM Endpoint is provided for LLM usage.
- The method of sharing images between accounts has been changed.
- You can create a new shared Image and share it.
- Add GPU Server usage guide
- Using Multi-instance GPU on GPU Server and Using NVSwitch on GPU Server added the guides.
- GPU Server RHEL OS and GPU driver version have been added.
- GPU Server feature addition
- NAT setting feature has been added to GPU Server.
- Samsung Cloud Platform Common Feature Change
- Account, IAM and Service Home, tags, etc. have been reflected in common CX changes.
- GPU Server service has been officially launched.
- CPU, GPU, memory, etc., we have launched a virtualization computing service that allows you to allocate and use the infrastructure resources provided by the server as needed at the required time without having to purchase them individually.
4 - Bare Metal Server
4.1 - Overview
Service Overview
Bare Metal Server does not use virtualization technology and a high-performance cloud computing service that can allocate and use physically separated computing resources such as CPU and memory individually. Since it is not affected by other cloud users, you can reliably operate performance-sensitive services.
Features
Easy and convenient computing environment setup: Through the web-based console, you can easily use everything from Bare Metal Server provisioning to resource management and cost management. You can receive a server with standard specs (CPU, Memory, Disk) allocated exclusively and use it immediately.
Providing High-Performance Computing Environment: We provide servers suitable for workloads that require large capacity and high performance, such as real-time (Real-Time) systems, HPC (High Performance Computing), and servers that demand excessive I/O usage, in a physically isolated environment.
Efficient Service Provision: We ensure performance and stability through optimal server selection and in-house testing. Customers can select the optimal resources for their service environment through the various specifications of Bare Metal Servers offered by Samsung Cloud Platform.
Service Diagram
Provided Features
Bare Metal Server provides the following features.
- Auto Provisioning and Management: Through the web-based console, you can easily use everything from Bare Metal Server provisioning to resource management and cost management.
- Providing various types of server types and OS images: Provides CPU/Memory/Disk resources of standard server types, and offers various standard OS images.
- Storage Connection: Provides additional connected storage besides the OS disk. You can connect and use Block Storage, File Storage, and Object Storage.
- Network Connection: You can connect the general subnet/IP settings of the Bare Metal Server and the Public NAT IP. Provides a local subnet connection for communication between servers. This operation can be modified on the detail page.
- Monitoring: You can view monitoring information such as CPU, Memory, Disk, which are computing resources, through Cloud Monitoring. To use the Cloud Monitoring service of Bare Metal Server, you need to install the Agent. Please be sure to install the Agent for stable Bare Metal Server service usage. For more details, refer to Bare Metal Server Monitoring Metrics.
- Backup and Recovery: Bare Metal Server’s Filesystem backup and recovery can be used through the Backup service.
- Efficient Cost Management: You can easily create/terminate servers as needed, and because billing is based on actual usage time, you can use it cost-effectively in various unpredictable situations.
- Local disk partition creation You can create and use up to 10 local disk partitions.
- Terraform provision: Provides an IaC environment through Terraform.
Components
Bare Metal Server provides various OS standard images and standard server types. Users can select and use them according to the scale of the service they want to configure.
OS Image Provided Version
The OS images supported by Bare Metal Server are as follows
| OS Image Version | EoS Date |
|---|---|
| Oracle Linux 9.6 | 2032-06-30 |
| RHEL 8.10 | 2029-05-31 |
| RHEL 9.4 | 2026-04-30 |
| RHEL 9.6 | 2027-05-31 |
| Rocky Linux 8.10 | 2029-05-31 |
| Rocky Linux 9.6 | 2025-11-30 |
| Ubuntu 22.04 | 2027-06-30 |
| Ubuntu 24.04 | 2029-06-30 |
| Windows 2019 | 2029-01-09 |
| Windows 2022 | 2031-10-14 |
Server Type
The server types supported by Bare Metal Server are as follows. For more details about the server types supported by Bare Metal Server, see Bare Metal Server Server Type.
s3v16m64_metal
| Category | Example | Detailed description |
|---|---|---|
| Server Generation | s3 | Provided server classification and generation
|
| CPU vCore | v16 | vCore count
|
| Memory | m64 | Memory Capacity
|
Preceding Service
This is a list of services that must be pre-configured before creating the service. Please refer to the guide provided for each service for details and prepare in advance.
| Service Category | Service | Detailed Description |
|---|---|---|
| Networking | VPC | A service that provides an independent virtual network in a cloud environment |
4.1.1 - Server Type
Bare Metal Server Server Types
Bare Metal Server provides server types according to usage purposes. Server types are configured with various combinations of CPU, Memory, etc. The server used for Bare Metal Server is determined by the server type selected when creating Bare Metal Server. Please select a server type according to the specifications of the application you want to run on Bare Metal Server.
The server types supported by Bare Metal Server follow the format below.
s3v16m64_metal
Category | Example | Description |
|---|---|---|
| Server Generation | s3 | Server category and generation provided
|
| CPU vCore | v16 | Number of vCores
|
| Memory | m64 | Memory capacity
|
s4/h4 Server Types
Bare Metal Server s4 server type is provided with standard specifications (vCPU, Memory) and is suitable for high-performance applications because it receives physically isolated resources for use. Also, Bare Metal Server h4 server type is provided with high-capacity server specifications and is suitable for high-performance applications for large-scale data processing.
- Supports 5 types of vCPU in total (16, 32, 64, 96, 128 vCore)
- Intel 6th generation (Granite Rapids) Processor
- Supports up to 64 physical cores, 128 vCPUs, and 2,048 GB of memory
- Provides up to 2 x 1.92 TB Internal Disks (for OS)
| Server Type | Physical CPU | vCPU | Memory | CPU Type | Internal Disk(OS) | ||
|---|---|---|---|---|---|---|---|
| s4v16m64_metal | 8 Core | 16 vCore | 64 GB | Intel Xeon 6507P up to 4.3GHz | 480GB * 2EA | ||
| s4v16m128_metal | 8 Core | 16 vCore | 128 GB | Intel Xeon 6507P up to 4.3GHz | 480GB * 2EA | ||
| s4v16m256_metal | 8 Core | 16 vCore | 256 GB | Intel Xeon 6507P up to 4.3GHz | 480GB * 2EA | ||
| h4v32m128_metal | 16 Core | 32 vCore | 128 GB | Intel Xeon 6517P up to 4.0GHz | 960GB * 2EA | ||
| h4v32m256_metal | 16 Core | 32 vCore | 256 GB | Intel Xeon 6517P up to 4.0GHz | 960GB * 2EA | ||
| h4v32m512_metal | 16 Core | 32 vCore | 512 GB | Intel Xeon 6517P up to 4.0GHz | 960GB * 2EA | ||
| h4v64m256_metal | 32 Core | 64 vCore | 256 GB | Intel Xeon 6737P up to 4.0GHz | 1.92TB * 2EA | ||
| h4v64m512_metal | 32 Core | 64 vCore | 512 GB | Intel Xeon 6737P up to 4.0GHz | 1.92TB * 2EA | ||
| h4v64m1024_metal | 32 Core | 64 vCore | 1024 GB | Intel Xeon 6737P up to 4.0GHz | 1.92TB * 2EA | ||
| h4v96m512_metal | 48 Core | 96 vCore | 512 GB | Intel Xeon 6520P up to 3.4GHz | 1.92TB * 2EA | ||
| h4v96m768_metal | 48 Core | 96 vCore | 768 GB | Intel Xeon 6520P up to 3.4GHz | 1.92TB * 2EA | ||
| h4v96m2048_metal | 48 Core | 96 vCore | 2048 GB | Intel Xeon 6520P up to 3.4GHz | 1.92TB * 2EA | ||
| h4v128m512_metal | 64 Core | 128 vCore | 512 GB | Intel Xeon 6737P up to 4.0GHz | 1.92TB * 2EA | ||
| h4v128m1024_metal | 64 Core | 128 vCore | 1024 GB | Intel Xeon 6737P up to 4.0GHz | 1.92TB * 2EA | ||
| h4v128m2048_metal | 64 Core | 128 vCore | 2048 GB | Intel Xeon 6737P up to 4.0GHz | 1.92TB * 2EA |
s3/h3 Server Types
Bare Metal Server s3 server type is provided with standard specifications (vCPU, Memory) and is suitable for high-performance applications because it receives physically isolated resources for use. Also, Bare Metal Server h3 server type is provided with high-capacity server specifications and is suitable for high-performance applications for large-scale data processing.
- Supports 5 types of vCPU in total (16, 32, 64, 96, 128 vCore)
- Intel 4th generation (Sapphire Rapids) Processor
- Supports up to 64 physical cores, 128 vCPUs, and 2,048 GB of memory
- Provides up to 2 x 1.92 TB Internal Disks (for OS)
| Server Type | Physical CPU | vCPU | Memory | CPU Type | Internal Disk(OS) | ||
|---|---|---|---|---|---|---|---|
| s3v16m64_metal | 8 Core | 16 vCore | 64 GB | Intel Xeon Gold 6434 up to 4.1GHz | 480 GB * 2EA | ||
| s3v16m128_metal | 8 Core | 16 vCore | 128 GB | Intel Xeon Gold 6434 up to 4.1GHz | 480 GB * 2EA | ||
| s3v16m256_metal | 8 Core | 16 vCore | 256 GB | Intel Xeon Gold 6434 up to 4.1GHz | 480 GB * 2EA | ||
| h3v32m128_metal | 16 Core | 32 vCore | 128 GB | Intel Xeon Gold 6444Y up to 4.0GHz | 960 GB * 2EA | ||
| h3v32m256_metal | 16 Core | 32 vCore | 256 GB | Intel Xeon Gold 6444Y up to 4.0GHz | 960 GB * 2EA | ||
| h3v32m512_metal | 16 Core | 32 vCore | 512 GB | Intel Xeon Gold 6444Y up to 4.0GHz | 960 GB * 2EA | ||
| h3v64m256_metal | 32 Core | 64 vCore | 256 GB | Intel Xeon Gold 6448H up to 3.2GHz | 1.92 TB * 2EA | ||
| h3v64m512_metal | 32 Core | 64 vCore | 512 GB | Intel Xeon Gold 6448H up to 3.2GHz | 1.92 TB * 2EA | ||
| h3v64m1024_metal | 32 Core | 64 vCore | 1024 GB | Intel Xeon Gold 6448H up to 3.2GHz | 1.92 TB * 2EA | ||
| h3v96m384_metal | 48 Core | 96 vCore | 384 GB | Intel Xeon Gold 6442Y up to 3.3GHz | 1.92 TB * 2EA | ||
| h3v96m768_metal | 48 Core | 96 vCore | 768 GB | Intel Xeon Gold 6442Y up to 3.3GHz | 1.92 TB * 2EA | ||
| h3v96m1536_metal | 48 Core | 96 vCore | 1536 GB | Intel Xeon Gold 6442Y up to 3.3GHz | 1.92 TB * 2EA | ||
| h3v128m512_metal | 64 Core | 128 vCore | 512 GB | Intel Xeon Gold 6448H up to 3.2GHz | 1.92 TB * 2EA | ||
| h3v128m1024_metal | 64 Core | 128 vCore | 1024 GB | Intel Xeon Gold 6448H up to 3.2GHz | 1.92 TB * 2EA | ||
| h3v128m2048_metal | 64 Core | 128 vCore | 2048 GB | Intel Xeon Gold 6448H up to 3.2GHz | 1.92 TB * 2EA |
s2/h2 Server Types
Bare Metal Server s2 server type is provided with standard specifications (vCPU, Memory) and is suitable for high-performance applications because it receives physically isolated resources for use.
Also, Bare Metal Server h2 server type is provided with high-capacity server specifications and is suitable for high-performance applications for large-scale data processing.
- Supports 5 types of vCPU in total (16, 24, 32, 72, 96 vCore)
- Intel 3rd generation (Ice Lake) Processor
- Supports up to 48 physical cores, 96 vCPUs, and 1,024 GB of memory
- Provides up to 2 x 1.92 TB Internal Disks (for OS)
| Server Type | Physical CPU | vCPU | Memory | CPU Type | Internal Disk(OS) | |
|---|---|---|---|---|---|---|
| s2v16m64_metal | 8 Core | 16 vCore | 64 GB | Intel Xeon Gold 6334 up to 3.6GHz | 480 GB * 2EA | |
| s2v16m128_metal | 8 Core | 16 vCore | 128 GB | Intel Xeon Gold 6334 up to 3.6GHz | 480 GB * 2EA | |
| s2v16m256_metal | 8 Core | 16 vCore | 256 GB | Intel Xeon Gold 6334 up to 3.6GHz | 480 GB * 2EA | |
| h2v24m96_metal | 12 Core | 24 vCore | 96 GB | Intel Xeon Gold 5317 up to 3.4GHz | 960 GB * 2EA | |
| h2v24m192_metal | 12 Core | 24 vCore | 192 GB | Intel Xeon Gold 5317 up to 3.4GHz | 960 GB * 2EA | |
| h2v24m384_metal | 12 Core | 24 vCore | 384 GB | Intel Xeon Gold 5317 up to 3.4GHz | 960 GB * 2EA | |
| h2v32m128_metal | 16 Core | 32 vCore | 128 GB | Intel Xeon Gold 6346 up to 3.6GHz | 960 GB * 2EA | |
| h2v32m256_metal | 16 Core | 32 vCore | 256 GB | Intel Xeon Gold 6346 up to 3.6GHz | 960 GB * 2EA | |
| h2v32m512_metal | 16 Core | 32 vCore | 512 GB | Intel Xeon Gold 6346 up to 3.6GHz | 960 GB * 2EA | |
| h2v72m256_metal | 36 Core | 72 vCore | 256 GB | Intel Xeon Gold 6354 up to 3.6GHz | 1.92 TB * 2EA | |
| h2v72m512_metal | 36 Core | 72 vCore | 512 GB | Intel Xeon Gold 6354 up to 3.6GHz | 1.92 TB * 2EA | |
| h2v72m1024_metal | 36 Core | 72 vCore | 1024 GB | Intel Xeon Gold 6354 up to 3.6GHz | 1.92 TB * 2EA | |
| h2v96m384_metal | 48 Core | 96 vCore | 384 GB | Intel Xeon Gold 6342 up to 3.3GHz | 1.92 TB * 2EA | |
| h2v96m768_metal | 48 Core | 96 vCore | 768 GB | Intel Xeon Gold 6342 up to 3.3GHz | 1.92 TB * 2EA |
4.1.2 - Monitoring Metrics
Bare Metal Server Monitoring Metrics
The following table shows the monitoring metrics for Bare Metal Server that can be checked through Cloud Monitoring.
| Performance Item | Detailed Description | Unit |
|---|---|---|
| Core Usage [IO Wait] | The ratio of CPU time spent in a waiting state (disk wait) | % |
| Core Usage [System] | The ratio of CPU time spent in kernel space | % |
| Core Usage [User] | The ratio of CPU time spent in user space | % |
| CPU Cores | The number of CPU cores on the host | cnt |
| CPU Usage [Active] | The percentage of CPU time used, excluding idle and IOWait states | % |
| CPU Usage [Idle] | The ratio of CPU time spent in an idle state | % |
| CPU Usage [IO Wait] | The ratio of CPU time spent in a waiting state (disk wait) | % |
| CPU Usage [System] | The percentage of CPU time used by the kernel | % |
| CPU Usage [User] | The percentage of CPU time used by the user area | % |
| CPU Usage/Core [Active] | The percentage of CPU time used, excluding idle and IOWait states | % |
| CPU Usage/Core [Idle] | The ratio of CPU time spent in an idle state | % |
| CPU Usage/Core [IO Wait] | The ratio of CPU time spent in a waiting state (disk wait) | % |
| CPU Usage/Core [System] | The percentage of CPU time used by the kernel | % |
| CPU Usage/Core [User] | The percentage of CPU time used by the user area | % |
| Disk CPU Usage [IO Request] | The ratio of CPU time spent executing I/O requests for the device | % |
| Disk Queue Size [Avg] | The average queue length of requests executed for the device | num |
| Disk Read Bytes | The number of bytes read from the device per second | bytes |
| Disk Read Bytes [Delta Avg] | The average of system.diskio.read.bytes_delta for individual disks | bytes |
| Disk Read Bytes [Delta Max] | The maximum of system.diskio.read.bytes_delta for individual disks | bytes |
| Disk Read Bytes [Delta Min] | The minimum of system.diskio.read.bytes_delta for individual disks | bytes |
| Disk Read Bytes [Delta Sum] | The sum of system.diskio.read.bytes_delta for individual disks | bytes |
| Disk Read Bytes [Delta] | The delta of system.diskio.read.bytes for individual disks | bytes |
| Disk Read Bytes [Success] | The total number of bytes read successfully | bytes |
| Disk Read Requests | The number of read requests for the disk device per second | cnt |
| Disk Read Requests [Delta Avg] | The average of system.diskio.read.count_delta for individual disks | cnt |
| Disk Read Requests [Delta Max] | The maximum of system.diskio.read.count_delta for individual disks | cnt |
| Disk Read Requests [Delta Min] | The minimum of system.diskio.read.count_delta for individual disks | cnt |
| Disk Read Requests [Delta Sum] | The sum of system.diskio.read.count_delta for individual disks | cnt |
| Disk Read Requests [Success Delta] | The delta of system.diskio.read.count for individual disks | cnt |
| Disk Read Requests [Success] | The total number of successful reads | cnt |
| Disk Request Size [Avg] | The average size of requests executed for the device (in sectors) | num |
| Disk Service Time [Avg] | The average service time for I/O requests executed for the device (in milliseconds) | ms |
| Disk Wait Time [Avg] | The average time spent waiting for I/O requests executed for the device | ms |
| Disk Wait Time [Read] | The average disk read wait time | ms |
| Disk Wait Time [Write] | The average disk write wait time | ms |
| Disk Write Bytes [Delta Avg] | The average of system.diskio.write.bytes_delta for individual disks | bytes |
| Disk Write Bytes [Delta Max] | The maximum of system.diskio.write.bytes_delta for individual disks | bytes |
| Disk Write Bytes [Delta Min] | The minimum of system.diskio.write.bytes_delta for individual disks | bytes |
| Disk Write Bytes [Delta Sum] | The sum of system.diskio.write.bytes_delta for individual disks | bytes |
| Disk Write Bytes [Delta] | The delta of system.diskio.write.bytes for individual disks | bytes |
| Disk Write Bytes [Success] | The total number of bytes written successfully | bytes |
| Disk Write Requests | The number of write requests for the disk device per second | cnt |
| Disk Write Requests [Delta Avg] | The average of system.diskio.write.count_delta for individual disks | cnt |
| Disk Write Requests [Delta Max] | The maximum of system.diskio.write.count_delta for individual disks | cnt |
| Disk Write Requests [Delta Min] | The minimum of system.diskio.write.count_delta for individual disks | cnt |
| Disk Write Requests [Delta Sum] | The sum of system.diskio.write.count_delta for individual disks | cnt |
| Disk Write Requests [Success Delta] | The delta of system.diskio.write.count for individual disks | cnt |
| Disk Write Requests [Success] | The total number of successful writes | cnt |
| Disk Writes Bytes | The number of bytes written to the device per second | bytes |
| Filesystem Hang Check | Filesystem (local/NFS) hang check (normal: 1, abnormal: 0) | status |
| Filesystem Nodes | The total number of file nodes in the filesystem | cnt |
| Filesystem Nodes [Free] | The total number of available file nodes in the filesystem | cnt |
| Filesystem Size [Available] | The available disk space for non-privileged users (in bytes) | bytes |
| Filesystem Size [Free] | The available disk space (in bytes) | bytes |
| Filesystem Size [Total] | The total disk space (in bytes) | bytes |
| Filesystem Usage | The percentage of used disk space | % |
| Filesystem Usage [Avg] | The average of filesystem.used.pct for individual filesystems | % |
| Filesystem Usage [Inode] | The inode usage rate | % |
| Filesystem Usage [Max] | The maximum of filesystem.used.pct for individual filesystems | % |
| Filesystem Usage [Min] | The minimum of filesystem.used.pct for individual filesystems | % |
| Filesystem Usage [Total] | The total filesystem usage | % |
| Filesystem Used | The used disk space (in bytes) | bytes |
| Filesystem Used [Inode] | The inode usage | bytes |
| Memory Free | The total available memory (in bytes), excluding system cache and buffer memory | bytes |
| Memory Free [Actual] | The actual available memory (in bytes) | bytes |
| Memory Free [Swap] | The available swap memory | bytes |
| Memory Total | The total memory | bytes |
| Memory Total [Swap] | The total swap memory | bytes |
| Memory Usage | The percentage of used memory | % |
| Memory Usage [Actual] | The percentage of actual used memory | % |
| Memory Usage [Cache Swap] | The cache swap usage rate | % |
| Memory Usage [Swap] | The percentage of used swap memory | % |
| Memory Used | The used memory | bytes |
| Memory Used [Actual] | The actual used memory (in bytes), subtracted from the total memory | bytes |
| Memory Used [Swap] | The used swap memory | bytes |
| Collisions | Network collisions | cnt |
| Network In Bytes | The number of received bytes | bytes |
| Network In Bytes [Delta Avg] | The average of system.network.in.bytes_delta for individual networks | bytes |
| Network In Bytes [Delta Max] | The maximum of system.network.in.bytes_delta for individual networks | bytes |
| Network In Bytes [Delta Min] | The minimum of system.network.in.bytes_delta for individual networks | bytes |
| Network In Bytes [Delta Sum] | The sum of system.network.in.bytes_delta for individual networks | bytes |
| Network In Bytes [Delta] | The delta of received bytes | bytes |
| Network In Dropped | The number of dropped incoming packets | cnt |
| Network In Errors | The number of errors during reception | cnt |
| Network In Packets | The number of received packets | cnt |
| Network In Packets [Delta Avg] | The average of system.network.in.packets_delta for individual networks | cnt |
| Network In Packets [Delta Max] | The maximum of system.network.in.packets_delta for individual networks | cnt |
| Network In Packets [Delta Min] | The minimum of system.network.in.packets_delta for individual networks | cnt |
| Network In Packets [Delta Sum] | The sum of system.network.in.packets_delta for individual networks | cnt |
| Network In Packets [Delta] | The delta of received packets | cnt |
| Network Out Bytes | The number of transmitted bytes | bytes |
| Network Out Bytes [Delta Avg] | The average of system.network.out.bytes_delta for individual networks | bytes |
| Network Out Bytes [Delta Max] | The maximum of system.network.out.bytes_delta for individual networks | bytes |
| Network Out Bytes [Delta Min] | The minimum of system.network.out.bytes_delta for individual networks | bytes |
| Network Out Bytes [Delta Sum] | The sum of system.network.out.bytes_delta for individual networks | bytes |
| Network Out Bytes [Delta] | The delta of transmitted bytes | bytes |
| Network Out Dropped | The number of dropped outgoing packets | cnt |
| Network Out Errors | The number of errors during transmission | cnt |
| Network Out Packets | The number of transmitted packets | cnt |
| Network Out Packets [Delta Avg] | The average of system.network.out.packets_delta for individual networks | cnt |
| Network Out Packets [Delta Max] | The maximum of system.network.out.packets_delta for individual networks | cnt |
| Network Out Packets [Delta Min] | The minimum of system.network.out.packets_delta for individual networks | cnt |
| Network Out Packets [Delta Sum] | The sum of system.network.out.packets_delta for individual networks | cnt |
| Network Out Packets [Delta] | The delta of transmitted packets | cnt |
| Open Connections [TCP] | The number of open TCP connections | cnt |
| Open Connections [UDP] | The number of open UDP connections | cnt |
| Port Usage | The port usage rate | % |
| SYN Sent Sockets | The number of sockets in the SYN_SENT state (when connecting from local to remote) | cnt |
| Kernel PID Max | The value of kernel.pid_max | cnt |
| Kernel Thread Max | The value of kernel.threads-max | cnt |
| Process CPU Usage | The percentage of CPU time consumed by the process since the last update | % |
| Process CPU Usage/Core | The percentage of CPU time used by the process since the last event | % |
| Process Memory Usage | The percentage of main memory (RAM) used by the process | % |
| Process Memory Used | The resident set size, which is the amount of memory used by the process in RAM | bytes |
| Process PID | The process ID | pid |
| Process PPID | The parent process ID | pid |
| Processes [Dead] | The number of dead processes | cnt |
| Processes [Idle] | The number of idle processes | cnt |
| Processes [Running] | The number of running processes | cnt |
| Processes [Sleeping] | The number of sleeping processes | cnt |
| Processes [Stopped] | The number of stopped processes | cnt |
| Processes [Total] | The total number of processes | cnt |
| Processes [Unknown] | The number of processes with unknown or unsearchable status | cnt |
| Processes [Zombie] | The number of zombie processes | cnt |
| Running Process Usage | The process usage rate | % |
| Running Processes | The number of running processes | cnt |
| Running Thread Usage | The thread usage rate | % |
| Running Threads | The total number of threads running in running processes | cnt |
| Context Switches | The number of context switches (per second) | cnt |
| Load/Core [1 min] | The load averaged over the last 1 minute, divided by the number of cores | cnt |
| Load/Core [15 min] | The load averaged over the last 15 minutes, divided by the number of cores | cnt |
| Load/Core [5 min] | The load averaged over the last 5 minutes, divided by the number of cores | cnt |
| Multipaths [Active] | The number of external storage connection paths with status = active | cnt |
| Multipaths [Failed] | The number of external storage connection paths with status = failed | cnt |
| Multipaths [Faulty] | The number of external storage connection paths with status = faulty | cnt |
| NTP Offset | The measured offset (time difference between the NTP server and the local environment) of the last sample | num |
| Run Queue Length | The length of the run queue | num |
| Uptime | The OS uptime (in milliseconds) | ms |
| Context Switchies | The number of CPU context switches (per second) | cnt |
| Disk Read Bytes [Sec] | The number of bytes read from the Windows logical disk per second | cnt |
| Disk Read Time [Avg] | The average data read time (in seconds) | sec |
| Disk Transfer Time [Avg] | The average disk wait time | sec |
| Disk Usage | The disk usage rate | % |
| Disk Write Bytes [Sec] | The number of bytes written to the Windows logical disk per second | cnt |
| Disk Write Time [Avg] | The average data write time (in seconds) | sec |
| Pagingfile Usage | The paging file usage rate | % |
| Pool Used [Non Paged] | The non-paged pool usage in kernel memory | bytes |
| Pool Used [Paged] | The paged pool usage in kernel memory | bytes |
| Process [Running] | The number of currently running processes | cnt |
| Threads [Running] | The number of currently running threads | cnt |
| Threads [Waiting] | The number of threads waiting for processor time | cnt |
4.2 - How-to guides
The user can input required information for a Bare Metal Server through the Samsung Cloud Platform Console, select detailed options, and create the service.
Bare Metal Server Create
You can create and use the Bare Metal Server service from the Samsung Cloud Platform Console.
To create a Bare Metal Server, follow the steps below.
- All Services > Compute > Bare Metal Server Click the menu. Go to the Service Home page of Bare Metal Server.
- Click the Bare Metal Server Create button on the Service Home page. It navigates to the Bare Metal Server Create page.
- Bare Metal Server creation On the page, enter the information required to create the service, and select detailed options.
- Image and Version Selection area, select the required information.
Category Required or notDetailed description Image Required Select the type of image provided - RHEL
- Rocky Linux
- Ubuntu
- Windows
Image Version Required Select version of the chosen image - Provides a list of versions of the provided server images
Table. Bare Metal Server Image and Version Input Items - Service Information Input area, please input or select the required information.
Category Required or notDetailed description Server count Required Number of Bare Metal Server servers to create simultaneously - Only numbers can be entered, and must be between 1 and 5
Service Type > Server Type Required Bare Metal Server Server Type - Select desired vCPU, Memory, Disk specifications
- For detailed information about server types provided by Bare Metal Server, refer to Bare Metal Server Server Type
Service Type > Planned Compute Required Status of resources with Planned Compute set - In Use: Number of resources with Planned Compute set that are in use
- Configured: Number of resources with Planned Compute set
- Coverage Preview: Amount applied by Planned Compute per resource
- Apply for Planned Compute Service: Go to Planned Compute application page
- For details, refer to Apply for Planned Compute
Automation Account Required Automatically creates an account to provide automation functions after Bare Metal Server creation - The account is used only for inter-system interface purposes
- Password is encrypted and cannot be accessed outside the system
- If the account is deleted, network changes and some automation functions will be restricted
Table. Bare Metal Server Service Information Input Items - In the Required Information Input area, enter or select the required information.
Category Required or notDetailed description Administrator Account Required Set the administrator account and password to be used when connecting to the server - RHEL, Ubuntu OS are provided fixed as root
- For Windows OS, enter using lowercase English letters and numbers, 5~20 characters
Administratornot allowed
Server Name Required Enter a name to distinguish the Bare Metal Server when the selected number of servers is 1 - Set the hostname to the entered server name
- Start with a lowercase English letter, and use lowercase letters, numbers, and special character (
-) to enter within 3 to 15 characters
- Must not end with a special character (
-)
Server Name Prefix Required Input Prefix to distinguish each Bare Metal Server generated when the selected number of servers is 2 or more - Automatically generated in the form of user input value (prefix) + ‘
-###’
- Must start with a lowercase English letter, and be entered using lowercase letters, numbers, and special characters (
-) within 3 to 15 characters
- Must not end with a special character (
-)
Network Settings Required Set the network where the Bare Metal Server will be installed - Select a pre-created VPC
- General Subnet: Select a pre-created general Subnet
- IP can be set to Auto-generated or User input; if Input is selected, the user enters the IP directly
- NAT: Available only when there is a single server and the VPC has an Internet Gateway attached
- When checked, a NAT IP can be selected
- NAT IP: Select a NAT IP
- If there is no NAT IP to select, click the Create New button to generate a Public IP
- Click the Refresh button to view and select the created Public IP
- Creating a Public IP incurs charges according to the Public IP pricing policy
- Local Subnet (Optional): Choose to use a local Subnet
- Not a required element for creating the service
- Select a pre-created local Subnet
- IP can be set to Auto-generated or User input; if Input is selected, the user enters the IP directly
Table. Bare Metal Server Required Information Input Items
- Image and Version Selection area, select the required information.
Please use a firewall etc. to control traffic access for Bare Metal Server. Security Group is not provided.
The firewall of the Bare Metal Server can only be used for traffic control between the Bare Metal Server and the Virtual Server. To use the Bare Metal Server’s firewall, follow the steps below.
- Separate the VPC of the Bare Metal Server: Separate them so that the Bare Metal Server and Virtual Server do not use the same VPC.
- Create Transit Gateway: Please create the Transit Gateway.
- The integration between the VPC of Virtual Server and the VPC of Bare Metal Server uses a Transit Gateway.
- When creating a Transit Gateway integration in the VPC of a Bare Metal Server, you must also create the Bare Metal Server’s firewall.
- Firewall Rule registration: Register a rule in the Firewall of the Bare Metal Server.
- Bare Metal Server Creation on the page Additional Information Input area, enter or select the required information.
Category Required or notDetailed description Local disk partition Select Set whether to use local disk partition - Up to 10 can be created, including the root partition
- Up to 90% of total capacity can be used
- After checking Use, partition information can be set
- Root partition information setting
- Partition type: flat, lvm selectable
- Partition name: enter partition name
- Can be entered only when partition type is lvm
- Enter within 15 characters, starting with a letter and including letters, numbers, and special characters (
-)
- Partition size: enter at least 50 GB
- Filesystem type: select according to the used image
- For RHEL, Rocky Linux: xfs, ext4
- For Ubuntu: ext4, xfs, btrfs
- For SLES: btrfs, xfs, ext4
- Mount point: start with special character
/and enter within 15 characters, including letters, numbers, and special characters (-)- If Filesystem type is swap, entry not allowed
- Available capacity: 90% of the default disk capacity provided when selecting a server
- When setting partition size, the remaining capacity is automatically calculated and displayed
- Total partition disk amount cannot exceed available capacity
- Additional partition information setting
- Partition type: flat, lvm selectable
- Partition name: enter partition name
- Can be entered only when partition type is lvm
- Enter within 15 characters, starting with a letter and including letters, numbers, and special characters (
-)
- Partition size: enter at least 1 GB
- Filesystem type: select according to the used image>
- For RHEL, Rocky Linux: xfs, ext4, swap
- For Ubuntu: ext4, xfs, btrfs, swap
- For SLES: btrfs, xfs, ext4, swap
- Mount point: start with special character
/and enter within 15 characters, including letters, numbers, and special characters (-)- If Filesystem type is swap, entry not allowed
- Available capacity: 90% of the default disk capacity provided when selecting a server
- When setting partition size, the remaining capacity is automatically calculated and displayed
- Total partition disk amount cannot exceed available capacity
Placement Group Select Servers belonging to the same Placement group are distributed across different racks - Provides distributed placement for up to 2 servers belonging to the same Placement group
- For distribution of 3 or more servers, add additional Placement groups
- Applicable only at initial creation; cannot be modified after creation
- If you terminate the last server belonging to a Placement group, that Placement group is automatically deleted
Lock Select Using Lock prevents actions caused by mistakes, preventing the server from being terminated, started, or stopped Hyper Threading Select Set logical cores to operate at twice the number of physical cores - Uncheck the box to turn off Hyper Threading
- Cannot be changed after server creation
Init Script Select Script to run when the server starts - Init Script must be selected differently depending on the image type
- For Windows: Select Batch Script
- For Linux: Shell Script
Table. Bare Metal Server Additional Information Input Items - Summary Check the detailed information and estimated billing amount generated in the panel, and click the Complete button.
- When creation is complete, check the created resources on the Bare Metal Server List page.
Bare Metal Server Check detailed information
The Bare Metal Server service can view and edit the full resource list and detailed information. Bare Metal Server Details page consists of Details, Tags, Activity History tabs.
Bare Metal Server If you want to view detailed information, follow the steps below.
- All Services > Compute > Bare Metal Server Click the menu. Navigate to the Service Home page of Bare Metal Server.
- Click the Bare Metal Server menu on the Service Home page. Go to the Bare Metal Server List page.
- Bare Metal Server List page, click the resource to view detailed information. Bare Metal Server Details page moves.
- Bare Metal Server Detail page displays status information and additional feature information, and consists of Detail Information, Tag, Operation History tabs.
Category Detailed description Bare Metal Server status Status of the Bare Metal Server created by the user - Creating: server is being created
- Running:: creation complete and usable
- Editing:: IP is being changed
- Unknown: error state
- Starting: server is starting
- Stopping: server is stopping
- Stopped: server has stopped
- Terminating: termination in progress
- Terminated: termination complete
Server Control Button to change server status - Start: Start a stopped server
- Stop: Stop a running server
Service termination Button to cancel the service Table. Bare Metal Server status information and additional functions
Detailed Information
Bare Metal Server List page allows you to view detailed information of the selected resource and, if necessary, edit the information.
| Category | Detailed description |
|---|---|
| Service | Service Name |
| Resource Type | Resource Type |
| SRN | Unique resource ID in Samsung Cloud Platform
|
| Resource Name | Resource Name
|
| Resource ID | Unique resource ID in the service |
| Creator | User who created the service |
| Creation time | Service creation time |
| Editor | User who edited the service information |
| Modification DateTime | Date and time when service information was modified |
| Server name | Server name |
| Image/Version | Server’s OS image and version |
| Server Type | vCPU, memory information display |
| Planned Compute | Resource status with Planned Compute set
|
| Lock | Display lock usage status
|
| Hyper Threading | Hyper Threading usage/not usage indication
|
| Network | Network information of Bare Metal Server
|
| Local Subnet | Local Subnet information of Bare Metal Server
|
| Block Storage | Block Storage information connected to the server
|
| Init Script | View the Init Script entered when creating the server |
Tag
Bare Metal Server List page, you can view the tag information of the selected resource, and add, modify, or delete it.
| Category | Detailed description |
|---|---|
| Tag List | Tag List
|
Work History
You can view the operation history of the selected resources on the Bare Metal Server List page.
| Category | Detailed description |
|---|---|
| Work History List | Resource Change History
|
Bare Metal Server Resource Management
If you need server control and management functions for the created Bare Metal Server resources, you can perform the tasks on the Bare Metal Server List or Bare Metal Server Details page.
Bare Metal Server Operation Control
You can start, stop, and restart a running Bare Metal Server.
To control the operation of Bare Metal Server, follow the steps below.
- All Services > Compute > Bare Metal Server Click the menu. Navigate to the Service Home page of Bare Metal Server.
- Click the Bare Metal Server menu on the Service Home page. Navigate to the Bare Metal Server List page.
- Bare Metal Server list On the page, after selecting multiple servers, you can control multiple servers simultaneously using the start and stop buttons at the top of the table.
- Bare Metal Server Details page also allows you to start and stop the server.
- Bare Metal Server List on the page click the resource to control operation and navigate to the Bare Metal Server Detail page.
- Check the server status and complete the changes using each Server Management button.
- Start: Start the stopped server.
- Stop: Stops the running server.
When a Bare Metal Server is stopped, the server’s power turns off.
- Since it may affect applications or storage in use, we recommend shutting down the OS and then stopping.
- After shutting down the OS, be sure to also stop in the Console.
- Bare Metal Server If you cannot start when requesting a start, see below.
- When Lock is set: After changing the Lock setting to disabled, try again.
- If the Bare Metal Server’s status is not Stopped: Change the Bare Metal Server’s status to Stopped, then try again.
- If stopping a Bare Metal Server request is not possible, refer to the following.
- If Lock is set: Change the Lock setting to disabled, then try again.
- If the Bare Metal Server’s status is not Running: Change the Bare Metal Server’s status to Running, then try again.
Add Block Storage
You can add Block Storage to a Bare Metal Server.
To add Block Storage, follow the steps below.
- All Services > Compute > Bare Metal Server Click the menu. Navigate to the Service Home page of Bare Metal Server.
- Click the Bare Metal Server menu on the Service Home page. Go to the Bare Metal Server list page.
- On the Bare Metal Server List page, click the server to add Block Storage. You will be taken to the Bare Metal Server Details page.
- Click the Add button in the Block Storage item on the Bare Metal Server Details page.
- If the popup window confirming Block Storage addition opens, click the Confirm button. Move to the Block Storage (BM) Creation page.
- Block Storage(BM) Creation on the page, enter the information required to create the service and create the Block Storage.
- For detailed information about creating Block Storage(BM), please refer to Block Storage(BM) Create.
- Navigate to the Bare Metal Server Details page after adding Block Storage and verify that Block Storage has been added.
Bare Metal Server Termination
If you terminate an unused Bare Metal Server, you can reduce operating costs. However, terminating a Bare Metal Server may cause the running service to stop immediately, so you should consider the impact of service interruption sufficiently before proceeding with the termination.
To cancel the Bare Metal Server, follow the steps below.
- All Services > Compute > Bare Metal Server Click the menu. Navigate to the Service Home page of Bare Metal Server.
- Click the Bare Metal Server menu on the Service Home page. Go to the Bare Metal Server list page.
- Bare Metal Server List page, select the resource to cancel, and click the Cancel Service button.
- You can select multiple resources and delete them simultaneously.
- You can also delete by clicking the Service Termination button on the Bare metal Server details page of the resource to be terminated.
- When termination is complete, check on the Bare Metal Server List page whether the resource has been terminated.
Termination Constraints
Bare Metal Server when a termination request cannot be processed, we will guide with a popup window. Please refer to the cases below.
Block Storage(BM) is connected (simultaneous termination of 2 or more servers): Disconnect the Block Storage(BM) first.
- For detailed information on how to cancel, please refer to Block Storage(BM) Cancel.
If File Storage is connected: Please disconnect the File Storage first.
- For detailed information on how to cancel, please refer to File Storage Cancel.
If Lock is set: After changing the Lock setting to disabled, try again.
If there is a Backup Agent or Load Balancer connection resource: Terminate the connection of that resource first.
If resource management tasks for Bare Metal Server are in progress on the same account: After the Bare Metal Server resource management tasks are completed, please try again.
If the Bare Metal Server’s status is not Running or Stopped: Change the Bare Metal Server’s status to Running or Stopped, then try again.
If the server that cannot be terminated simultaneously is included: Please select only the resources that can be terminated and try again.
Local Subnet Setup
After completing the creation of a Bare Metal Server, if you add a local Subnet on the Bare Metal Server Details page, you must configure the network settings of the local Subnet yourself.
First Connection(kr-west)
There is no local subnet connected to the Bare Metal Server, and if you are adding the first connection, proceed according to the guide below.
This guide applies to kr-west (Korea West) when adding the first local Subnet connection to the server.
- kr-south(South Korea) guide, refer to the First Connection(kr-south) chapter.
Linux - Setting up Subnet on Ubuntu
On the Ubuntu operating system, add a local Subnet, and to proceed with network configuration, follow the steps below.
On the Bare Metal Server Details page, check the Interface Name.
View the network configuration information.
Color mode[root@localhost ~]# vi /etc/netplan/50-cloud-init.yaml network: ethernets: ens2f1: match: macaddress: 68:05:ca:d4:09:91 mtu: 1500 set-name: ens2f1 ens4f1: match: macaddress: 68:05:ca:d4:09:01 mtu: 1500 set-name: ens4f1[root@localhost ~]# vi /etc/netplan/50-cloud-init.yaml network: ethernets: ens2f1: match: macaddress: 68:05:ca:d4:09:91 mtu: 1500 set-name: ens2f1 ens4f1: match: macaddress: 68:05:ca:d4:09:01 mtu: 1500 set-name: ens4f1Code block. Network configuration file lookup After adding a new VLAN, set the IP for the Bonding configuration.
- Change the ID and IP in the example code to the assigned ID and IP.Color mode
[root@localhost ~]# vi /etc/netplan/50-cloud-init.yaml network: bonds: bond-mgt: interfaces: - ens2f1 // **Bare Metal Server Details** page, enter the Interface Name you confirmed. - ens4f1 // **Bare Metal Server detailed** page, enter the Interface Name you verified. mtu: 1500 parameters: mii-monitor-interval: 100 mode: active-backup transmit-hash-policy: layer2 ethernets: ens2f1: match: macaddress: 68:05:ca:d4:09:91 mtu: 1500 set-name: ens2f1 ens4f1: match: macaddress: 68:05:ca:d4:09:01 mtu: 1500 set-name: ens4f1 vlans: bond-mgt.20: // Enter the Vlan ID confirmed in the SCP Console instead of 20. addresses: - 192.168.0.10/24 // Set it to the local Subnet IP confirmed in the SCP Console. id: 20 // Set it to the VLAN ID confirmed in the SCP Console. link: bond-mgt mtu: 1500[root@localhost ~]# vi /etc/netplan/50-cloud-init.yaml network: bonds: bond-mgt: interfaces: - ens2f1 // **Bare Metal Server Details** page, enter the Interface Name you confirmed. - ens4f1 // **Bare Metal Server detailed** page, enter the Interface Name you verified. mtu: 1500 parameters: mii-monitor-interval: 100 mode: active-backup transmit-hash-policy: layer2 ethernets: ens2f1: match: macaddress: 68:05:ca:d4:09:91 mtu: 1500 set-name: ens2f1 ens4f1: match: macaddress: 68:05:ca:d4:09:01 mtu: 1500 set-name: ens4f1 vlans: bond-mgt.20: // Enter the Vlan ID confirmed in the SCP Console instead of 20. addresses: - 192.168.0.10/24 // Set it to the local Subnet IP confirmed in the SCP Console. id: 20 // Set it to the VLAN ID confirmed in the SCP Console. link: bond-mgt mtu: 1500Code block. IP Settings
- Change the ID and IP in the example code to the assigned ID and IP.
Apply the changes to the system.
Color mode# netplan apply# netplan applyCode block. Reflect changes Check the interface status.
Color mode# ip a or # bash /usr/local/bin/ip.sh# ip a or # bash /usr/local/bin/ip.shCode block. Interface lookup
Linux – Setting up Subnet on CentOS/Red Hat
After adding a local Subnet on the CentOS/Red Hat operating system, follow the steps below to configure the network.
On the Bare Metal Server Details page, check the Interface Name.
Modify the following command and execute.
Color mode#!/usr/bin/bash IP_ADDR="10.1.1.3/24" // Set the local Subnet IP that you checked from the Console. VLAN_ID="7" // Set the Vlan ID confirmed in the Console. BOND_NAME="bond-mgt" BOND_IF_name1="ens2f1" // Enter the Interface Name you verified on the **Bare Metal Server Details** page. BOND_IF_name2="ens4f0" // **Bare Metal Server Details** Enter the Interface Name you verified on the page. # Delete Connection nmcli con down "Bond ${BOND_NAME}" nmcli con del "Bond ${BOND_NAME}" nmcli con down "System ${BOND_IF_name1}" nmcli con down "System ${BOND_IF_name2}" nmcli con del "System ${BOND_IF_name1}" nmcli con del "System ${BOND_IF_name2}" # Create Bonding nmcli con add con-name ${BOND_NAME} type bond ifname ${BOND_NAME} nmcli conn mod ${BOND_NAME} con-name "Bond ${BOND_NAME}" nmcli conn mod "Bond ${BOND_NAME}" ipv4.method disabled nmcli conn mod "Bond ${BOND_NAME}" ipv6.method ignore nmcli conn mod "Bond ${BOND_NAME}" connect.autoconnect yes nmcli conn mod "Bond ${BOND_NAME}" +bond.options mode=active-backup \ +bond.options xmit_hash_policy=layer2 \ +bond.options miimon=100 \ +bond.options num_grat_arp=1 \ +bond.options downdelay=0 \ +bond.options updelay=0 # Assign bond-slave nmcli conn add type bond-slave ifname ${BOND_IF_name1} con-name "${BOND_IF_name1}" master ${BOND_NAME} nmcli conn mod ${BOND_IF_name1} con-name "System ${BOND_IF_name1}" nmcli conn add type bond-slave ifname ${BOND_IF_name2} con-name "${BOND_IF_name2}" master ${BOND_NAME} nmcli conn mod ${BOND_IF_name2} con-name "System ${BOND_IF_name2}" # Connection UP nmcli conn up "Bond ${BOND_NAME}" # add vlan nmcli conn add type vlan ifname "${BOND_NAME}.${VLAN_ID}" con-name "${BOND_NAME}.${VLAN_ID}" id ${VLAN_ID} dev ${BOND_NAME} nmcli con mod ${BOND_NAME}.${VLAN_ID} con-name "Vlan ${BOND_NAME}.${VLAN_ID}" nmcli con mod "Vlan ${BOND_NAME}.${VLAN_ID}" ipv4.addresses ${IP_ADDR} nmcli con mod "Vlan ${BOND_NAME}.${VLAN_ID}" ipv4.method manual nmcli con mod "Vlan ${BOND_NAME}.${VLAN_ID}" ipv6.method "ignore" nmcli con mod "Vlan ${BOND_NAME}.${VLAN_ID}" connect.autoconnect yes nmcli con up "Vlan ${BOND_NAME}.${VLAN_ID}" nmcli device reapply ${BOND_NAME}.${VLAN_ID}#!/usr/bin/bash IP_ADDR="10.1.1.3/24" // Set the local Subnet IP that you checked from the Console. VLAN_ID="7" // Set the Vlan ID confirmed in the Console. BOND_NAME="bond-mgt" BOND_IF_name1="ens2f1" // Enter the Interface Name you verified on the **Bare Metal Server Details** page. BOND_IF_name2="ens4f0" // **Bare Metal Server Details** Enter the Interface Name you verified on the page. # Delete Connection nmcli con down "Bond ${BOND_NAME}" nmcli con del "Bond ${BOND_NAME}" nmcli con down "System ${BOND_IF_name1}" nmcli con down "System ${BOND_IF_name2}" nmcli con del "System ${BOND_IF_name1}" nmcli con del "System ${BOND_IF_name2}" # Create Bonding nmcli con add con-name ${BOND_NAME} type bond ifname ${BOND_NAME} nmcli conn mod ${BOND_NAME} con-name "Bond ${BOND_NAME}" nmcli conn mod "Bond ${BOND_NAME}" ipv4.method disabled nmcli conn mod "Bond ${BOND_NAME}" ipv6.method ignore nmcli conn mod "Bond ${BOND_NAME}" connect.autoconnect yes nmcli conn mod "Bond ${BOND_NAME}" +bond.options mode=active-backup \ +bond.options xmit_hash_policy=layer2 \ +bond.options miimon=100 \ +bond.options num_grat_arp=1 \ +bond.options downdelay=0 \ +bond.options updelay=0 # Assign bond-slave nmcli conn add type bond-slave ifname ${BOND_IF_name1} con-name "${BOND_IF_name1}" master ${BOND_NAME} nmcli conn mod ${BOND_IF_name1} con-name "System ${BOND_IF_name1}" nmcli conn add type bond-slave ifname ${BOND_IF_name2} con-name "${BOND_IF_name2}" master ${BOND_NAME} nmcli conn mod ${BOND_IF_name2} con-name "System ${BOND_IF_name2}" # Connection UP nmcli conn up "Bond ${BOND_NAME}" # add vlan nmcli conn add type vlan ifname "${BOND_NAME}.${VLAN_ID}" con-name "${BOND_NAME}.${VLAN_ID}" id ${VLAN_ID} dev ${BOND_NAME} nmcli con mod ${BOND_NAME}.${VLAN_ID} con-name "Vlan ${BOND_NAME}.${VLAN_ID}" nmcli con mod "Vlan ${BOND_NAME}.${VLAN_ID}" ipv4.addresses ${IP_ADDR} nmcli con mod "Vlan ${BOND_NAME}.${VLAN_ID}" ipv4.method manual nmcli con mod "Vlan ${BOND_NAME}.${VLAN_ID}" ipv6.method "ignore" nmcli con mod "Vlan ${BOND_NAME}.${VLAN_ID}" connect.autoconnect yes nmcli con up "Vlan ${BOND_NAME}.${VLAN_ID}" nmcli device reapply ${BOND_NAME}.${VLAN_ID}Code block. IP configuration script Check the interface status.
Color mode# ip a or # bash /usr/local/bin/ip.sh# ip a or # bash /usr/local/bin/ip.shCode block. Interface lookup
Setting up Subnet on Windows
After adding a local Subnet in the Windows operating system, follow these steps to configure the network.
Windows Start icon, right-click, then run the Windows PowerShell (Administrator) program.
Check the Interface Name on the Bare Metal Server Details page.
Run ncpa.cpl from the Windows Run menu.
Check whether the interface is activated, and if necessary, activate it.
- Bare Metal Server detail checked on the page Interface Name activate.
Create a Teaming.
Color modePS C:\> New-NetLbfoTeam –Name “bond-mgt” –TeamMembers ens2f1,ens4f1 PS C:\> Set-NetLbfoTeam –Name “bond-mgt” –LoadBalancingAlgorithm DynamicPS C:\> New-NetLbfoTeam –Name “bond-mgt” –TeamMembers ens2f1,ens4f1 PS C:\> Set-NetLbfoTeam –Name “bond-mgt” –LoadBalancingAlgorithm DynamicCode block. Teaming creation After adding a new VLAN, set the IP.
- Enter the VLAN ID and local Subnet IP confirmed on the Bare Metal Server Details page.Color mode
PS C:\> Add-NetLbfoTeamNIC -Team bond_bond-mgt -VlanID 20 -Name bond-mgt.20 PS C:\> Get-NetAdapter PS C:\> netsh interface ip set address bond-mgt.20 static “192.168.0.10/24”PS C:\> Add-NetLbfoTeamNIC -Team bond_bond-mgt -VlanID 20 -Name bond-mgt.20 PS C:\> Get-NetAdapter PS C:\> netsh interface ip set address bond-mgt.20 static “192.168.0.10/24”Code block. Windows IP Settings
- Run ncpa.cpl from the Windows Run menu to check the interface status.
First connection(kr-south)
If there is no local subnet connected to the Bare Metal Server initially, and you are adding the first connection, proceed according to the guide below.
This guide is for kr-south (Korean region) when adding the first local Subnet connection to the server.
- kr-west(Korea West) the guide that applies to First Connection(kr-west) please refer to the chapter.
Linux - Setting up Subnet on Ubuntu
To add a local Subnet on the Ubuntu operating system and proceed with network settings, follow the steps below.
After adding a new Vlan, set the IP.
- Change the ID and IP in the example code to the assigned ID and IP.Color mode
[root@localhost ~]# vi /etc/netplan/50-cloud-init.yaml network: bonds: bond-mgt: interfaces: - ens2f1 - ens4f1 mtu: 1500 parameters: mii-monitor-interval: 100 mode: active-backup transmit-hash-policy: layer2 ethernets: ens2f1: match: macaddress: 68:05:ca:d4:09:91 mtu: 1500 set-name: ens2f1 ens4f1: match: macaddress: 68:05:ca:d4:09:01 mtu: 1500 set-name: ens4f1 vlans: bond-mgt.20: addresses: - 192.168.0.10/24 id: 20 link: bond-mgt mtu: 1500 vlans: bond-mgt.21: // Enter the Vlan ID you checked on the Console instead of 21. addresses: - 192.168.0.20/24 // Set to the local Subnet IP confirmed in the Console. id: 21 // Set it to the Vlan ID verified in the Console. link: bond-mgt mtu: 1500[root@localhost ~]# vi /etc/netplan/50-cloud-init.yaml network: bonds: bond-mgt: interfaces: - ens2f1 - ens4f1 mtu: 1500 parameters: mii-monitor-interval: 100 mode: active-backup transmit-hash-policy: layer2 ethernets: ens2f1: match: macaddress: 68:05:ca:d4:09:91 mtu: 1500 set-name: ens2f1 ens4f1: match: macaddress: 68:05:ca:d4:09:01 mtu: 1500 set-name: ens4f1 vlans: bond-mgt.20: addresses: - 192.168.0.10/24 id: 20 link: bond-mgt mtu: 1500 vlans: bond-mgt.21: // Enter the Vlan ID you checked on the Console instead of 21. addresses: - 192.168.0.20/24 // Set to the local Subnet IP confirmed in the Console. id: 21 // Set it to the Vlan ID verified in the Console. link: bond-mgt mtu: 1500Code block. Vlan addition and IP setting
- Change the ID and IP in the example code to the assigned ID and IP.
Reflect the modifications in the system.
Color mode# netplan apply# netplan applyCode block. Reflect changes Check the interface status.
Color mode# ip a or # bash /usr/local/bin/ip.sh# ip a or # bash /usr/local/bin/ip.shCode block. Interface lookup
Linux – Setting up Subnet on CentOS/Red Hat
After adding a local Subnet on CentOS/Red Hat operating system, follow the steps below to configure the network.
- Check the Bond Name for local Subnet.Color mode
# sh /usr/local/bin/ip.sh# sh /usr/local/bin/ip.shCode block. Bonding check - Modify the following command and execute.Color mode
#!/usr/bin/bash IP_ADDR="10.1.1.3/24" // Set the local Subnet IP as confirmed from the Console. VLAN_ID="7" // Set the Vlan ID confirmed in the console. BOND_NAME="bond-mgt" // Set the Bond Name confirmed in step 1. # add vlan nmcli conn add type vlan ifname "${BOND_NAME}.${VLAN_ID}" con-name "${BOND_NAME}.${VLAN_ID}" id ${VLAN_ID} dev ${BOND_NAME} nmcli con mod ${BOND_NAME}.${VLAN_ID} con-name "Vlan ${BOND_NAME}.${VLAN_ID}" nmcli con mod "Vlan ${BOND_NAME}.${VLAN_ID}" ipv4.addresses ${IP_ADDR} nmcli con mod "Vlan ${BOND_NAME}.${VLAN_ID}" ipv4.method manual nmcli con mod "Vlan ${BOND_NAME}.${VLAN_ID}" ipv6.method "ignore" nmcli con mod "Vlan ${BOND_NAME}.${VLAN_ID}" connect.autoconnect yes nmcli con up "Vlan ${BOND_NAME}.${VLAN_ID}" nmcli device reapply ${BOND_NAME}.${VLAN_ID}#!/usr/bin/bash IP_ADDR="10.1.1.3/24" // Set the local Subnet IP as confirmed from the Console. VLAN_ID="7" // Set the Vlan ID confirmed in the console. BOND_NAME="bond-mgt" // Set the Bond Name confirmed in step 1. # add vlan nmcli conn add type vlan ifname "${BOND_NAME}.${VLAN_ID}" con-name "${BOND_NAME}.${VLAN_ID}" id ${VLAN_ID} dev ${BOND_NAME} nmcli con mod ${BOND_NAME}.${VLAN_ID} con-name "Vlan ${BOND_NAME}.${VLAN_ID}" nmcli con mod "Vlan ${BOND_NAME}.${VLAN_ID}" ipv4.addresses ${IP_ADDR} nmcli con mod "Vlan ${BOND_NAME}.${VLAN_ID}" ipv4.method manual nmcli con mod "Vlan ${BOND_NAME}.${VLAN_ID}" ipv6.method "ignore" nmcli con mod "Vlan ${BOND_NAME}.${VLAN_ID}" connect.autoconnect yes nmcli con up "Vlan ${BOND_NAME}.${VLAN_ID}" nmcli device reapply ${BOND_NAME}.${VLAN_ID}Code block. IP configuration script - Check the interface status.Color mode
# ip a or # bash /usr/local/bin/ip.sh# ip a or # bash /usr/local/bin/ip.shCode block. Interface lookup
Setting up Subnet on Windows
After adding a local Subnet on the Windows operating system, follow these steps to configure the network.
Windows Start icon, right-click, then run the Windows PowerShell (Administrator) program.
Check the Teaming name for local Subnet.
Color modePS C:\> Get-NetAdapterPS C:\> Get-NetAdapterCode block. Windows interface check After adding a new VLAN, set the IP.
- Enter the Teaming name confirmed in step 2, and the Vlan ID and Local Subnet IP confirmed in the Console.Color mode
PS C:\> Add-NetLbfoTeamNIC -Team bond_bond-mgt -VlanID 20 -Name bond-mgt.20 PS C:\> Get-NetAdapter PS C:\> netsh interface ip set address bond-mgt.20 static “192.168.0.10/24”PS C:\> Add-NetLbfoTeamNIC -Team bond_bond-mgt -VlanID 20 -Name bond-mgt.20 PS C:\> Get-NetAdapter PS C:\> netsh interface ip set address bond-mgt.20 static “192.168.0.10/24”Code block. Create Teaming
- Enter the Teaming name confirmed in step 2, and the Vlan ID and Local Subnet IP confirmed in the Console.
In the Windows Start menu, run ncpa.cpl to check the interface status.
Add second connection (kr-west, kr-south)
If there is a local Subnet connected to the Bare Metal Server, the guide for the second additional connection is as follows.
Because a Bonding was already created when connecting the first local Subnet, there is no procedure to create Bonding when connecting the second local Subnet.
Please refer to the details below.
Linux - Setting up Subnet on Ubuntu
To add a local Subnet on the Ubuntu operating system and proceed with network configuration, follow the steps below.
After adding a new Vlan, set the IP.
- Change the ID and IP of the example code to the assigned ID and IP.Color mode
[root@localhost ~]# vi /etc/netplan/50-cloud-init.yaml network: bonds: bond-mgt: interfaces: - ens2f1 - ens4f1 mtu: 1500 parameters: mii-monitor-interval: 100 mode: active-backup transmit-hash-policy: layer2 ethernets: ens2f1: match: macaddress: 68:05:ca:d4:09:91 mtu: 1500 set-name: ens2f1 ens4f1: match: macaddress: 68:05:ca:d4:09:01 mtu: 1500 set-name: ens4f1 vlans: bond-mgt.20: addresses: - 192.168.0.10/24 id: 20 link: bond-mgt mtu: 1500 vlans: bond-mgt.21: // Enter the Vlan ID you checked on the console instead of 21. addresses: - 192.168.0.20/24 // Set it to the local Subnet IP confirmed from the Console. id: 21 // Set it to the Vlan ID confirmed in the Console. link: bond-mgt mtu: 1500[root@localhost ~]# vi /etc/netplan/50-cloud-init.yaml network: bonds: bond-mgt: interfaces: - ens2f1 - ens4f1 mtu: 1500 parameters: mii-monitor-interval: 100 mode: active-backup transmit-hash-policy: layer2 ethernets: ens2f1: match: macaddress: 68:05:ca:d4:09:91 mtu: 1500 set-name: ens2f1 ens4f1: match: macaddress: 68:05:ca:d4:09:01 mtu: 1500 set-name: ens4f1 vlans: bond-mgt.20: addresses: - 192.168.0.10/24 id: 20 link: bond-mgt mtu: 1500 vlans: bond-mgt.21: // Enter the Vlan ID you checked on the console instead of 21. addresses: - 192.168.0.20/24 // Set it to the local Subnet IP confirmed from the Console. id: 21 // Set it to the Vlan ID confirmed in the Console. link: bond-mgt mtu: 1500Code block. Vlan addition and IP configuration
- Change the ID and IP of the example code to the assigned ID and IP.
Apply the changes to the system.
Color mode# netplan apply# netplan applyCode block. Reflect changes Check the interface status.
Color mode# ip a or # bash /usr/local/bin/ip.sh# ip a or # bash /usr/local/bin/ip.shCode block. Interface lookup
Linux – Setting up Subnet on CentOS/Red Hat
After adding a local Subnet on CentOS/Red Hat operating system, follow the steps below to configure the network.
- Check the Bond Name for the local Subnet.Color mode
# sh /usr/local/bin/ip.sh# sh /usr/local/bin/ip.shCode block. Bonding check - Modify the following command and execute.Color mode
#!/usr/bin/bash IP_ADDR="10.1.1.3/24" // Set the local Subnet IP as confirmed from the console. VLAN_ID="7" // Set the Vlan ID confirmed in the console. BOND_NAME="bond-mgt" // Set the Bond Name confirmed in step 1. # add vlan nmcli conn add type vlan ifname "${BOND_NAME}.${VLAN_ID}" con-name "${BOND_NAME}.${VLAN_ID}" id ${VLAN_ID} dev ${BOND_NAME} nmcli con mod ${BOND_NAME}.${VLAN_ID} con-name "Vlan ${BOND_NAME}.${VLAN_ID}" nmcli con mod "Vlan ${BOND_NAME}.${VLAN_ID}" ipv4.addresses ${IP_ADDR} nmcli con mod "Vlan ${BOND_NAME}.${VLAN_ID}" ipv4.method manual nmcli con mod "Vlan ${BOND_NAME}.${VLAN_ID}" ipv6.method "ignore" nmcli con mod "Vlan ${BOND_NAME}.${VLAN_ID}" connect.autoconnect yes nmcli con up "Vlan ${BOND_NAME}.${VLAN_ID}" nmcli device reapply ${BOND_NAME}.${VLAN_ID}#!/usr/bin/bash IP_ADDR="10.1.1.3/24" // Set the local Subnet IP as confirmed from the console. VLAN_ID="7" // Set the Vlan ID confirmed in the console. BOND_NAME="bond-mgt" // Set the Bond Name confirmed in step 1. # add vlan nmcli conn add type vlan ifname "${BOND_NAME}.${VLAN_ID}" con-name "${BOND_NAME}.${VLAN_ID}" id ${VLAN_ID} dev ${BOND_NAME} nmcli con mod ${BOND_NAME}.${VLAN_ID} con-name "Vlan ${BOND_NAME}.${VLAN_ID}" nmcli con mod "Vlan ${BOND_NAME}.${VLAN_ID}" ipv4.addresses ${IP_ADDR} nmcli con mod "Vlan ${BOND_NAME}.${VLAN_ID}" ipv4.method manual nmcli con mod "Vlan ${BOND_NAME}.${VLAN_ID}" ipv6.method "ignore" nmcli con mod "Vlan ${BOND_NAME}.${VLAN_ID}" connect.autoconnect yes nmcli con up "Vlan ${BOND_NAME}.${VLAN_ID}" nmcli device reapply ${BOND_NAME}.${VLAN_ID}Code block. IP configuration script - Check the interface status.Color mode
# ip a or # bash /usr/local/bin/ip.sh# ip a or # bash /usr/local/bin/ip.shCode block. Interface lookup
Setting Subnet in Windows
After adding a local Subnet in the Windows operating system, follow the steps below to set up the network.
Right-click the Windows Start icon, then run the Windows PowerShell (Administrator) program.
Check the Teaming name for local Subnet.
Color modePS C:\> Get-NetAdapterPS C:\> Get-NetAdapterCode block. Windows interface check After adding a new VLAN, set the IP.
- Enter the Teaming name confirmed in step 2, the Vlan ID and Local Subnet IP confirmed in the Console.Color mode
PS C:\> Add-NetLbfoTeamNIC -Team bond_bond-mgt -VlanID 20 -Name bond-mgt.20 PS C:\> Get-NetAdapter PS C:\> netsh interface ip set address bond-mgt.20 static “192.168.0.10/24”PS C:\> Add-NetLbfoTeamNIC -Team bond_bond-mgt -VlanID 20 -Name bond-mgt.20 PS C:\> Get-NetAdapter PS C:\> netsh interface ip set address bond-mgt.20 static “192.168.0.10/24”Code block. Teaming creation
- Enter the Teaming name confirmed in step 2, the Vlan ID and Local Subnet IP confirmed in the Console.
In the Windows Run menu, execute ncpa.cpl to check the interface status.
IP Change
IP can be changed for migration, server replacement, etc.
- If you proceed with changing the IP, you will no longer be able to communicate with that IP, and you cannot cancel the IP change while it is in progress.
- If it is a server running the Load Balancer service, you must delete the existing IP from the LB server group and directly add the changed IP as a member of the LB server group.
- Servers using Public NAT, Privat NAT must disable and reconfigure Public NAT, Privat NAT after an IP change.
- If you are using Public NAT, Privat NAT, first disable the use of Public NAT, Privat NAT, complete the IP change, and then set it again.
- Public NAT, Privat NAT usage can be changed by clicking the Edit button of Public NAT IP, Privat NAT on the Bare Metal Server Details page.
If you want to change the IP, follow the steps below.
All Services > Compute > Bare Metal Server Click the menu. Navigate to the Service Home page of Bare Metal Server.
Click the Bare Metal Server menu on the Service Home page. You will be taken to the Bare Metal Server list page.
Bare Metal Server List page, click the server to change the IP. Bare Metal Server Details page will be opened.
Click the Edit button next to the IP item on the Bare Metal Server Details page.
When the popup notifying IP modification opens, click the Confirm button. The IP Change popup opens.
IP change popup window’s Step 1, Step 2, Step 3Proceed with the tasks in order.
Guide- When changing the IP, the detailed configuration method of the IP change step varies depending on the subnet of the IP to be changed. Be sure to refer to the following example and proceed with the work for each step.
- When changing to an IP that uses the same subnet: Change to IP of the same Subnet see
- When changing to an IP that uses a different subnet: Change IP to a different Subnet see
- When each progress step is completed successfully, the task status in the upper right corner is displayed as Completed, and you can proceed to the next step.
- When performing the final check of Step 3, it is recommended to restart the server and then proceed with the inspection.
- When changing the IP, the detailed configuration method of the IP change step varies depending on the subnet of the IP to be changed. Be sure to refer to the following example and proceed with the work for each step.
After confirming that all tasks have been completed successfully, click the Confirm button.
Change to the same Subnet’s IP
Explains how to set IP per operating system when the IP to be changed uses the same subnet.
Linux – centos/redhat operating system
Step 1
Follow the next procedure and proceed with Step 1 work.
- Select the Subnet to change.
- Enter the IP to change.
- IP allocation request Click the button.
- When the popup notifying IP change confirmation opens, click the Confirm button.
- If the task completes successfully, the task status in the upper right corner will be displayed as Completed.CautionIf you proceed with the IP allocation request of Step 1, you cannot cancel or restore the IP change.
- If the task completes successfully, the task status in the upper right corner will be displayed as Completed.
Step 2
Follow the next procedure and proceed with Step 2 work.
Connect to the IP change target server using NAT IP for the IP change operation.
NoticeTo prevent situations where communication is impossible during operation, it is recommended to connect via another Virtual Server or Bare Metal Server created in the same subnet.Step 1 Enter the assigned IP and set the IP to be changed on the server.
- In the following example, replace
172.17.34.150with the assigned IP. - After checking the information of the Interface you want to change on the server, enter it instead of the example
bond-srv.9.Color mode# nmcli con mod "Vlan bond-srv.9" ipv4.addresses 172.17.34.150/24 # nmcli con mod "Vlan bond-srv.9" ipv4.method manual # nmcli device reapply bond-srv.9# nmcli con mod "Vlan bond-srv.9" ipv4.addresses 172.17.34.150/24 # nmcli con mod "Vlan bond-srv.9" ipv4.method manual # nmcli device reapply bond-srv.9Code block. IP settings to change Guide- If you set the IP, the terminal session will be disconnected.
- Step 2 After completing the task, if the task status changes to Completed, you can reconnect to the terminal.
- In the following example, replace
When all tasks are completed, select the task completion checkbox of Step 2 in the IP Change popup window.
- If the task completes successfully, the task status in the upper right corner will be displayed as Completed.Guide
- If the task status of Step 2 has been changed to Completed but there is still an issue with terminal connection, go to the All Services > Management > Support Center Contact menu and inquire.
- If the task completes successfully, the task status in the upper right corner will be displayed as Completed.
Step 3
Follow the next procedure and proceed with Step 3 work.
Connect to the target server for IP change using NAT IP and check the communication status.
- Use the following command to check again whether the previous configuration information remains and whether it has been changed correctly. If you can connect normally to the IP change target server, the changed IP is in normal communication status.Color mode
# bash /usr/local/bin/ip.sh# bash /usr/local/bin/ip.shCode block. Communication status check ReferenceNAT IP does not change.
- Use the following command to check again whether the previous configuration information remains and whether it has been changed correctly. If you can connect normally to the IP change target server, the changed IP is in normal communication status.
Once all tasks are completed, restart the server and then perform a final check.
ReferenceIt is recommended to perform the final check after restarting the server.If there are no issues in the final inspection results, select the work completion checkbox of Step 3 in the IP Change popup window.
Linux – Ubuntu Operating System
Step 1
Follow the next procedure and proceed with Step 1 work.
- Select the Subnet to change.
- Enter the IP to change.
- IP Allocation Request Click the button.
- If a popup notifying IP change confirmation opens, click the Confirm button.
- If the task completes successfully, the task status in the upper right corner will be displayed as Completed.CautionIf you proceed with the IP allocation request of Step 1, you cannot cancel or restore the IP change.
- If the task completes successfully, the task status in the upper right corner will be displayed as Completed.
Step 2
Proceed with Step 2 work following the next procedure.
To perform the IP change operation, connect to the IP change target server using a NAT IP.
GuideTo prevent situations where communication is impossible during operation, it is recommended to connect via another Virtual Server or Bare Metal Server created in the same subnet.Step 1Enter the IP assigned in Step 1 and set the IP to be changed on the server.
- In the following example, replace
172.17.34.150/24with the assigned IP. - After checking the information of the Interface you want to change on the server, enter it instead of the example
bond-srv.9.Color mode[root@localhost ~]# vi /etc/netplan/50-cloud-init.yaml network: bonds: ...................... omitted ethernets: ...................... omitted vlans: bond-srv.9: addresses: - 172.17.34.150/24 # Enter the IP assigned in Step1. gateway4: 172.17.34.2 id: 9 link: bond-srv mtu: 1500 bond-srv.350: addresses: - 172.16.87.150/24 routes: - to: 172.17.87.0/24 via: 172.16.87.1 - to: 172.17.87.0/24 via: 172.16.87.1 id: 350 link: bond-srv[root@localhost ~]# vi /etc/netplan/50-cloud-init.yaml network: bonds: ...................... omitted ethernets: ...................... omitted vlans: bond-srv.9: addresses: - 172.17.34.150/24 # Enter the IP assigned in Step1. gateway4: 172.17.34.2 id: 9 link: bond-srv mtu: 1500 bond-srv.350: addresses: - 172.16.87.150/24 routes: - to: 172.17.87.0/24 via: 172.16.87.1 - to: 172.17.87.0/24 via: 172.16.87.1 id: 350 link: bond-srvCode block. IP settings to change
- In the following example, replace
Use the Netplan apply command to apply the changes to the system.
Color mode[root@localhost ~]# netplan apply[root@localhost ~]# netplan applyCode block. Run Netplan apply Notice- If you set the IP, the terminal session will be disconnected.
- Step 2 After completing the task, if the task status changes to Completed, you can reconnect to the terminal.
When all tasks are completed, select the task completion checkbox of Step 2 in the IP Change popup window.
- If the task completes successfully, the task status in the upper right corner will be displayed as Completed.Notice
- If the task status of Step 2 has been changed to Completed but there is an issue with terminal access, go to the All Services > Management > Support Center Contact menu and inquire.
- If the task completes successfully, the task status in the upper right corner will be displayed as Completed.
Step 3
Follow the next procedure and proceed with Step 3 work.
Check the communication status by connecting to the IP change target server with NAT IP.
- Use the following command to check again whether the pre-change configuration information remains and whether it has been changed correctly. If you can connect normally to the IP change target server, the changed IP is in normal communication status.Color mode
# bash /usr/local/bin/ip.sh# bash /usr/local/bin/ip.shCode block. Communication status check ReferenceNAT IP does not change.
- Use the following command to check again whether the pre-change configuration information remains and whether it has been changed correctly. If you can connect normally to the IP change target server, the changed IP is in normal communication status.
Once all tasks are completed, restart the server and then perform a final check.
ReferenceIt is recommended to perform the final check after restarting the server.If there are no issues in the final inspection results, select the work completion checkbox of Step 3 in the IP Change popup window.
Windows Operating System
Step 1
Follow the next procedure and proceed with Step 1 work.
- Select the Subnet to change.
- Please enter the IP to change.
- IP Allocation Request Click the button.
- When the popup notifying IP change confirmation opens, click the Confirm button.
- If the task completes successfully, the task status in the upper right corner will be displayed as Completed.CautionIf you proceed with Step 1’s IP allocation request, you cannot cancel or restore the IP change.
- If the task completes successfully, the task status in the upper right corner will be displayed as Completed.
Step 2
Follow the next procedure and proceed with Step 2 work.
Connect to the target server for IP change using NAT IP for the IP change operation.
GuideTo prevent situations where communication is impossible during operation, it is recommended to connect via another Virtual Server or Bare Metal Server created in the same subnet.Right-click the Windows Start icon, then run Windows PowerShell (Administrator).
Step 1Enter the assigned IP and set the IP to be changed on the server.
- In the following example, replace
172.17.34.150with the assigned IP.Color modePS C:\> netsh interface ip set address "bond-srv.20" static 172.17.34.150 255.255.255.0PS C:\> netsh interface ip set address "bond-srv.20" static 172.17.34.150 255.255.255.0Code block. IP settings to change Notice- If you set the IP, the terminal session will be disconnected.
- Step 2 After completing the task, if the task status changes to Completed, you can reconnect to the terminal.
- In the following example, replace
When all tasks are completed, select the task completion checkbox of Step 2 in the IP change popup window.
- If the task completes successfully, the task status in the upper right corner will be displayed as Completed.Notice
- Step 2’s task status changed to Completed, and if there is an issue with terminal access, go to the All Services > Management > Support Center’s Contact Us menu and inquire.
- If the task completes successfully, the task status in the upper right corner will be displayed as Completed.
Step 3
Follow the next procedure and proceed with Step 3 work.
Connect to the server targeted for IP change using NAT IP and check the communication status.
- Use the following command to check again whether the previous configuration information remains and whether it has been changed correctly. If you can connect normally to the IP change target server, the changed IP is in normal communication status.Color mode
PS C:\> Get-NetIPAddress | Format-TablePS C:\> Get-NetIPAddress | Format-TableCode block. Communication status check ReferenceNAT IP does not change.
- Use the following command to check again whether the previous configuration information remains and whether it has been changed correctly. If you can connect normally to the IP change target server, the changed IP is in normal communication status.
Once all tasks are completed, restart the server and then perform a final check.
ReferenceIt is recommended to perform the final check after restarting the server.If there are no issues with the final inspection results, select the work completion checkbox of Step 3 in the IP change popup window.
Change to IP of another Subnet
Explains how to set IP per operating system when the IP to be changed uses a different subnet.
Linux – centos/redhat operating system
Step 1
Follow the next procedure and proceed with Step 1 work.
- Please select the Subnet to change.
- Enter the IP to change.
- Click the IP Allocation Request button.
- When the popup that notifies IP change confirmation opens, click the Confirm button.
- If the task completes successfully, Check Vlan ID, Check Default Gateway information is displayed, and the task status at the top right is shown as Completed.CautionIf you proceed with the IP allocation request of Step 1, you cannot cancel or restore the IP change.
- If the task completes successfully, Check Vlan ID, Check Default Gateway information is displayed, and the task status at the top right is shown as Completed.
Step 2
Proceed with Step 2 work following the next procedure.
Connect to the IP change target server with a NAT IP to perform the IP change operation.
GuideTo prevent situations where communication is impossible during operation, it is recommended to connect via another Virtual Server or Bare Metal Server created in the same subnet.Add a new VLAN and set the IP to add the IP to the server.
- Add VLAN: Create the interface of the Vlan ID confirmed in Step 1. In the following example, enter the assigned ID instead of
20.
- Add VLAN: Create the interface of the Vlan ID confirmed in Step 1. In the following example, enter the assigned ID instead of
- IP Settings: Enter the IP assigned in Step 1. In the following example, replace
192.168.0.10/24with the assigned IP.Color mode# nmcli conn add type vlan ifname "bond-srv.20" con-name "bond-srv.20" id 20 dev bond-srv # nmcli con mod bond-srv.20 con-name "Vlan bond-srv.20" # nmcli con mod "Vlan bond-srv.20" ipv4.addresses 192.168.0.10/24 # nmcli con mod "Vlan bond-srv.20" ipv4.method manual # nmcli con mod "Vlan bond-srv.20" ipv6.method "ignore" # nmcli con up "Vlan bond-srv.20"# nmcli conn add type vlan ifname "bond-srv.20" con-name "bond-srv.20" id 20 dev bond-srv # nmcli con mod bond-srv.20 con-name "Vlan bond-srv.20" # nmcli con mod "Vlan bond-srv.20" ipv4.addresses 192.168.0.10/24 # nmcli con mod "Vlan bond-srv.20" ipv4.method manual # nmcli con mod "Vlan bond-srv.20" ipv6.method "ignore" # nmcli con up "Vlan bond-srv.20"Code block. IP settings to change
Set the Default Gateway on the new VLAN.
- Default gateway setting: Enter the Default gateway IP assigned in Step 1. In the following example, replace
192.168.0.1with the assigned Default gateway IP.Color mode# nmcli con mod "Vlan bond-srv.20" ipv4.gateway 192.168.0.1 # nmcli device reapply bond-srv.20# nmcli con mod "Vlan bond-srv.20" ipv4.gateway 192.168.0.1 # nmcli device reapply bond-srv.20Code block. IP settings to change Guide- If you set the Default Gateway on a new VLAN, the terminal session will be disconnected.
- Step 2 After completing the task, if the task status changes to Completed, you can reconnect to the terminal.
- Default gateway setting: Enter the Default gateway IP assigned in Step 1. In the following example, replace
When all tasks are completed, select the task completion checkbox of Step 2 in the IP Change popup window.
- If the task completes successfully, the task status in the upper right corner will be displayed as Completed.Guide
- Step 2’s work status changed to Completed, but if there is still an issue with terminal access, go to the All Services > Management > Support Center Contact menu and inquire.
- If the task completes successfully, the task status in the upper right corner will be displayed as Completed.
Step 3
Follow the next procedure and proceed with Step 3 work.
Connect to the target server for IP change using NAT IP.
After checking the Default Gateway IP of the existing (pre-change) interface, delete it.
- In the following example, enter the verified IP instead of
192.168.10.1.Color mode# ip route del default via 192.168.10.1# ip route del default via 192.168.10.1Code block. Delete Default Gateway IP of existing interface
- In the following example, enter the verified IP instead of
Connect to the IP change target server with a NAT IP and check the communication status.
- Use the following command to check again whether the previous configuration information remains and whether it has been changed correctly. If you can connect normally to the IP change target server, the changed IP is in normal communication status.Color mode
# netstat –nr # bash /usr/local/bin/ip.sh# netstat –nr # bash /usr/local/bin/ip.shCode block. Communication status check ReferenceNAT IP does not change.
- Use the following command to check again whether the previous configuration information remains and whether it has been changed correctly. If you can connect normally to the IP change target server, the changed IP is in normal communication status.
After checking the VLAN information of the existing IP, delete it from the server.
- In the following example, replace
30with the ID you verified.Color mode# nmcli con delete "Vlan bond-srv.30"# nmcli con delete "Vlan bond-srv.30"Code block. Delete Vlan information of existing IP
- In the following example, replace
Once all tasks are completed, restart the server and then perform a final check.
ReferenceIt is recommended to perform the final check after restarting the server.If there is no issue with the final inspection results, select the work completion checkbox of Step 3 in the IP Change popup.
Linux – Ubuntu operating system
Step 1
Follow the next procedure and proceed with Step 1 work.
- Please select the Subnet to change.
- Enter the IP to change.
- IP allocation request Click the button.
- When the popup notifying IP change confirmation opens, click the Confirm button.
- If the task completes successfully, Check Vlan ID, Check Default Gateway information is displayed, and the task status at the top right is shown as Completed.CautionStep 1’s IP allocation request cannot be cancelled or restored once processed.
- If the task completes successfully, Check Vlan ID, Check Default Gateway information is displayed, and the task status at the top right is shown as Completed.
Step 2
Follow the next procedure and proceed with Step 2 work.
Connect to the IP change target server using a NAT IP for the IP change operation.
GuideTo prevent situations where communication is impossible during operation, it is recommended to connect via another Virtual Server or Bare Metal Server created in the same subnet.To add the IP to be changed to the server, add a new VLAN and set the IP and Default Gateway.
- This is the part where content is added below the Step 1 work description in the following example.
- In the following example, enter the assigned ID and IP instead of ID and IP.Color mode
[root@localhost ~]# vi /etc/netplan/50-cloud-init.yaml network: bonds: ...................... omitted ethernets: ...................... omitted vlans: bond-srv.9: addresses: - 172.17.34.150/24 gateway4: 172.17.34.2 id: 9 link: bond-srv mtu: 1500 bond-srv.350: addresses: - 172.16.87.150/24 routes: - to: 172.17.87.0/24 via: 172.16.87.1 - to: 172.17.87.0/24 via: 172.16.87.1 id: 350 link: bond-srv # Create the interface of the Vlan ID confirmed in Step1. # Enter the IP assigned in Step1. # Enter the Default gateway IP assigned from Step1. bond-srv.20: addresses: - 192.168.0.10/24 gateway4: 192.168.0.1 id: 20 link: bond-srv mtu: 1500[root@localhost ~]# vi /etc/netplan/50-cloud-init.yaml network: bonds: ...................... omitted ethernets: ...................... omitted vlans: bond-srv.9: addresses: - 172.17.34.150/24 gateway4: 172.17.34.2 id: 9 link: bond-srv mtu: 1500 bond-srv.350: addresses: - 172.16.87.150/24 routes: - to: 172.17.87.0/24 via: 172.16.87.1 - to: 172.17.87.0/24 via: 172.16.87.1 id: 350 link: bond-srv # Create the interface of the Vlan ID confirmed in Step1. # Enter the IP assigned in Step1. # Enter the Default gateway IP assigned from Step1. bond-srv.20: addresses: - 192.168.0.10/24 gateway4: 192.168.0.1 id: 20 link: bond-srv mtu: 1500Code block. IP settings to change
Use the Netplan apply command to apply the changes to the system.
Color mode[root@localhost ~]# netplan apply[root@localhost ~]# netplan applyCode block. Netplan apply execution Notice- If you set a new Default Gateway, the terminal session will be disconnected.
- Step 2 After completing the task, if the task status changes to Completed, you can reconnect to the terminal.
When all tasks are completed, select the task completion checkbox of Step 2 in the IP Change popup window.
- If the task completes successfully, the task status in the upper right corner will be displayed as Completed.Notice
- If the task status of Step 2 has been changed to Completed but there is still an issue with terminal access, go to the All Services > Management > Support Center Contact menu and inquire.
- If the task completes successfully, the task status in the upper right corner will be displayed as Completed.
Step 3
Follow the next procedure and proceed with Step 3 work.
Connect to the target server for IP change using NAT IP.
After checking the Default Gateway IP of the existing (pre-change) interface, delete it.
- In the following example, the Delete this line row is the part that gets deleted.Color mode
[root@localhost ~]# vi /etc/netplan/50-cloud-init.yaml network: bonds: ...................... omitted ethernets: ...................... Omitted vlans: bond-srv.9: addresses: - 172.17.34.150/24 gateway4: 172.17.34.2 # Delete this line id: 9 link: bond-srv mtu: 1500 bond-srv.350: addresses: - 172.16.87.150/24 routes: - to: 172.17.87.0/24 via: 172.16.87.1 - to: 172.17.87.0/24 via: 172.16.87.1 id: 350 link: bond-srv bond-srv.20: addresses: - 192.168.0.10/24 gateway4: 192.168.0.1 id: 20 link: bond-srv mtu: 1500[root@localhost ~]# vi /etc/netplan/50-cloud-init.yaml network: bonds: ...................... omitted ethernets: ...................... Omitted vlans: bond-srv.9: addresses: - 172.17.34.150/24 gateway4: 172.17.34.2 # Delete this line id: 9 link: bond-srv mtu: 1500 bond-srv.350: addresses: - 172.16.87.150/24 routes: - to: 172.17.87.0/24 via: 172.16.87.1 - to: 172.17.87.0/24 via: 172.16.87.1 id: 350 link: bond-srv bond-srv.20: addresses: - 192.168.0.10/24 gateway4: 192.168.0.1 id: 20 link: bond-srv mtu: 1500Code block. Delete Default Gateway IP of existing interface
- In the following example, the Delete this line row is the part that gets deleted.
Connect to the IP change target server using NAT IP and check the communication status.
- Use the following command to check again whether the previous configuration information remains and whether it has been changed correctly. If you can connect normally to the IP change target server, the changed IP is in normal communication state.Color mode
# netstat –nr # bash /usr/local/bin/ip.sh# netstat –nr # bash /usr/local/bin/ip.shCode block. Communication status check ReferenceNAT IP does not change.
- Use the following command to check again whether the previous configuration information remains and whether it has been changed correctly. If you can connect normally to the IP change target server, the changed IP is in normal communication state.
Delete the existing IP.
- In the following example, the Delete this line row is the part that gets deleted.Color mode
[root@localhost ~]# vi /etc/netplan/50-cloud-init.yaml network: bonds: ...................... omitted ethernets: ...................... omitted vlans: bond-srv.9: #Delete this line addresses: # delete this line - 172.17.34.150/24 #Delete this line gateway4: 172.17.34.2 #Delete this line id: 9 #Delete this line link: bond-srv #Delete this line mtu: 1500 #Delete this line bond-srv.350: addresses: - 172.16.87.150/24 routes: - to: 172.17.87.0/24 via: 172.16.87.1 - to: 172.17.87.0/24 via: 172.16.87.1 id: 350 link: bond-srv bond-srv.20: addresses: - 192.168.0.10/24 gateway4: 192.168.0.1 id: 20 link: bond-srv mtu: 1[root@localhost ~]# vi /etc/netplan/50-cloud-init.yaml network: bonds: ...................... omitted ethernets: ...................... omitted vlans: bond-srv.9: #Delete this line addresses: # delete this line - 172.17.34.150/24 #Delete this line gateway4: 172.17.34.2 #Delete this line id: 9 #Delete this line link: bond-srv #Delete this line mtu: 1500 #Delete this line bond-srv.350: addresses: - 172.16.87.150/24 routes: - to: 172.17.87.0/24 via: 172.16.87.1 - to: 172.17.87.0/24 via: 172.16.87.1 id: 350 link: bond-srv bond-srv.20: addresses: - 192.168.0.10/24 gateway4: 192.168.0.1 id: 20 link: bond-srv mtu: 1Code block. Delete existing IP
- In the following example, the Delete this line row is the part that gets deleted.
Apply the modified items to the system.
Color mode[root@localhost ~]# netplan apply [root@localhost ~]# ip link delete bond-srv.9 # Additional actions when VLAN is deleted[root@localhost ~]# netplan apply [root@localhost ~]# ip link delete bond-srv.9 # Additional actions when VLAN is deletedCode block. Apply changes When all tasks are completed, restart the server and then conduct a final check.
- If there are no issues in the final inspection results, select the work completion checkbox of Step 3 in the IP Change popup window.
Windows Operating System
Step 1
Follow the next procedure and proceed with Step 1 work.
- Select the Subnet to change.
- Enter the IP to change.
- IP Allocation Request Click the button.
- When the popup that notifies IP change confirmation opens, click the Confirm button.
- If the task completes successfully, Check Vlan ID, Check Default Gateway information is displayed, and the task status at the top right is shown as Completed.CautionIf you proceed with the IP allocation request of Step 1, you cannot cancel or revert the IP change.
- If the task completes successfully, Check Vlan ID, Check Default Gateway information is displayed, and the task status at the top right is shown as Completed.
Step 2
Proceed with Step 2 work following the next procedure.
Connect to the IP change target server using NAT IP for the IP change operation.
GuideTo prevent situations where communication is impossible during operation, it is recommended to connect via another Virtual Server or Bare Metal Server created in the same subnet.Windows Start icon, right-click, then run Windows PowerShell (Administrator).
Add a VLAN and set the IP and default gateway.
- Add VLAN: Create the interface for the Vlan ID identified in Step 1. In the following example, replace
20with the assigned ID. - IP setting: Enter the IP assigned in Step 1. In the following example, replace
46with the ifindex confirmed by Get-NetAdapter, and replace 192.168.0.10 with the assigned IP. - Default gateway setting: Enter the assigned Default gateway IP from Step 1. In the following example, replace
192.168.0.1with the assigned Default gateway IP.Color modePS C:\> Add-NetLbfoTeamNIC -Team bond_bond-srv -VlanID 20 -Name bond-srv.20 -Confirm:$false PS C:\> Get-NetAdapter PS C:\> New-NetIPAddress -InterfaceIndex 46 -IPAddress 192.168.0.10 -PrefixLength 24 –defaultgateway 192.168.0.1PS C:\> Add-NetLbfoTeamNIC -Team bond_bond-srv -VlanID 20 -Name bond-srv.20 -Confirm:$false PS C:\> Get-NetAdapter PS C:\> New-NetIPAddress -InterfaceIndex 46 -IPAddress 192.168.0.10 -PrefixLength 24 –defaultgateway 192.168.0.1Code block. IP settings to change Guide- If you set a new Default Gateway, the terminal session will be disconnected.
- Step 2 After completing the task, if the task status changes to Completed, you can reconnect to the terminal.
- Add VLAN: Create the interface for the Vlan ID identified in Step 1. In the following example, replace
When all tasks are completed, select the task completion checkbox of Step 2 in the IP Change popup window.
- If the task completes successfully, the task status in the upper right corner will be displayed as Completed.Guide
- If the work status of Step 2 has been changed to Completed but there is still an issue with terminal access, go to the All Services > Management > Support Center Contact menu and inquire.
- If the task completes successfully, the task status in the upper right corner will be displayed as Completed.
Step 3
Follow the next procedure and proceed with Step 3 work.
Connect to the IP change target server using NAT IP.
Run the interface index (ifindex) to check the existing Default Gateway IP.
Color modePS C:\> Get-NetAdapter Name InterfaceDescription ifIndex Status MacAddress LinkSpeed ---- -------------------- ------- ------ ---------- --------- bond-srv.9 Microsoft Network Adapter Multiple...#4 30 Up 40-A6-B7-27-96-D5 50 Gbps bond-srv Microsoft Network Adapter Multiple...#3 19 Up 40-A6-B7-27-96-D5 50 Gbps bond-iscsi Microsoft Network Adapter Multiple...#2 18 Up 40-A6-B7-27-96-D4 50 Gbps bond-backup Microsoft Network Adapter Multiplexo... 22 Up 68-05-CA-C9-EB-88 20 Gbps eno2 Intel(R) Ethernet Connection X722 fo... 12 Disabled 38-68-DD-36-A0-59 1 Gbps ens3f0 Intel(R) Ethernet Network Adapter XX... 11 Up 40-A6-B7-27-96-D4 25 Gbps ………………………………………….. omittedPS C:\> Get-NetAdapter Name InterfaceDescription ifIndex Status MacAddress LinkSpeed ---- -------------------- ------- ------ ---------- --------- bond-srv.9 Microsoft Network Adapter Multiple...#4 30 Up 40-A6-B7-27-96-D5 50 Gbps bond-srv Microsoft Network Adapter Multiple...#3 19 Up 40-A6-B7-27-96-D5 50 Gbps bond-iscsi Microsoft Network Adapter Multiple...#2 18 Up 40-A6-B7-27-96-D4 50 Gbps bond-backup Microsoft Network Adapter Multiplexo... 22 Up 68-05-CA-C9-EB-88 20 Gbps eno2 Intel(R) Ethernet Connection X722 fo... 12 Disabled 38-68-DD-36-A0-59 1 Gbps ens3f0 Intel(R) Ethernet Network Adapter XX... 11 Up 40-A6-B7-27-96-D4 25 Gbps ………………………………………….. omittedCode block. Get-NetAdapter execution Color modePS C:\> get-netroute -ifindex 30 ifIndex DestinationPrefix NextHop RouteMetric PolicyStore ------- ----------------- ------- ----------- ----------- 30 255.255.255.255/32 0.0.0.0 256 ActiveStore 30 224.0.0.0/4 0.0.0.0 256 ActiveStore 30 172.17.35.0/24 172.17.35.1 256 ActiveStore 30 172.17.34.255/32 0.0.0.0 256 ActiveStore 30 172.17.34.14/32 0.0.0.0 256 ActiveStore 30 172.17.34.0/24 0.0.0.0 256 ActiveStore 30 0.0.0.0/0 172.17.34.1 1 ActiveStorePS C:\> get-netroute -ifindex 30 ifIndex DestinationPrefix NextHop RouteMetric PolicyStore ------- ----------------- ------- ----------- ----------- 30 255.255.255.255/32 0.0.0.0 256 ActiveStore 30 224.0.0.0/4 0.0.0.0 256 ActiveStore 30 172.17.35.0/24 172.17.35.1 256 ActiveStore 30 172.17.34.255/32 0.0.0.0 256 ActiveStore 30 172.17.34.14/32 0.0.0.0 256 ActiveStore 30 172.17.34.0/24 0.0.0.0 256 ActiveStore 30 0.0.0.0/0 172.17.34.1 1 ActiveStoreCode block. -ifindex execution Delete the existing Default Gateway IP.
- In the following example, replace
30with the ifindex obtained via Get-NetAdapter, and replace172.17.34.1with the IP you verified.Color modePS C:\> Remove-NetRoute -ifIndex 30 -DestinationPrefix 0.0.0.0/0 -NextHop 172.17.34.1 -Confirm:$falsePS C:\> Remove-NetRoute -ifIndex 30 -DestinationPrefix 0.0.0.0/0 -NextHop 172.17.34.1 -Confirm:$falseCode block. Delete Default Gateway IP Color modePS C:\> get-netroute -ifindex 30 ifIndex DestinationPrefix NextHop RouteMetric PolicyStore ------- ----------------- ------- ----------- ----------- 30 255.255.255.255/32 0.0.0.0 256 ActiveStore 30 224.0.0.0/4 0.0.0.0 256 ActiveStore 30 172.17.34.255/32 0.0.0.0 256 ActiveStore 30 172.17.34.14/32 0.0.0.0 256 ActiveStore 30 172.17.34.0/24 0.0.0.0 256 ActiveStorePS C:\> get-netroute -ifindex 30 ifIndex DestinationPrefix NextHop RouteMetric PolicyStore ------- ----------------- ------- ----------- ----------- 30 255.255.255.255/32 0.0.0.0 256 ActiveStore 30 224.0.0.0/4 0.0.0.0 256 ActiveStore 30 172.17.34.255/32 0.0.0.0 256 ActiveStore 30 172.17.34.14/32 0.0.0.0 256 ActiveStore 30 172.17.34.0/24 0.0.0.0 256 ActiveStoreCode block. -ifindex execution Notice- If you delete the existing Default Gateway, the terminal session will be disconnected.
- Step 2 After completing the task, if the task status changes to Completed, you can reconnect to the terminal.
- In the following example, replace
Check the communication status by connecting to the IP change target server with a NAT IP.
- Use the following command to check again whether the previous configuration information remains and whether it has been changed correctly. If you can connect normally to the target server whose IP was changed, the changed IP is in normal communication state.Color mode
PS C:\> netstat –nr | findstr Default PS C:\> Get-NetIPAddress | Format-TablePS C:\> netstat –nr | findstr Default PS C:\> Get-NetIPAddress | Format-TableCode block. Communication status check ReferenceNAT IP does not change.
- Use the following command to check again whether the previous configuration information remains and whether it has been changed correctly. If you can connect normally to the target server whose IP was changed, the changed IP is in normal communication state.
Check the existing IP’s VLAN information in the Team information.
Color modePS C:\> Get-NetLbfoTeam Name : bond_bond-srv Members : {ens6f1, ens3f1} TeamNics : {bond-srv, bond-srv.9} TeamingMode : SwitchIndependent LoadBalancingAlgorithm : Dynamic Status : Up Name : bond_bond-iscsi Members : {ens6f0, ens3f0} TeamNics : bond-iscsi TeamingMode : SwitchIndependent LoadBalancingAlgorithm : Dynamic Status : Up Name : bond_bond-backup Members : {ens2f0, ens4f1} TeamNics : bond-backup TeamingMode : SwitchIndependent LoadBalancingAlgorithm : Dynamic Status : Up PS C:\> Get-NetAdapter Name InterfaceDescription ifIndex Status MacAddress LinkSpeed ---- -------------------- ------- ------ ---------- --------- bond-srv.9 Microsoft Network Adapter Multiple...#4 30 Up 40-A6-B7-27-96-D5 50 Gbps bond-srv Microsoft Network Adapter Multiple...#3 19 Up 40-A6-B7-27-96-D5 50 Gbps bond-iscsi Microsoft Network Adapter Multiple...#2 18 Up 40-A6-B7-27-96-D4 50 Gbps bond-backup Microsoft Network Adapter Multiplexo... 22 Up 68-05-CA-C9-EB-88 20 Gbps eno2 Intel(R) Ethernet Connection X722 fo... 12 Disabled 38-68-DD-36-A0-59 1 Gbps ens3f0 Intel(R) Ethernet Network Adapter XX... 11 Up 40-A6-B7-27-96-D4 25 Gbps ………………………………………….. omittedPS C:\> Get-NetLbfoTeam Name : bond_bond-srv Members : {ens6f1, ens3f1} TeamNics : {bond-srv, bond-srv.9} TeamingMode : SwitchIndependent LoadBalancingAlgorithm : Dynamic Status : Up Name : bond_bond-iscsi Members : {ens6f0, ens3f0} TeamNics : bond-iscsi TeamingMode : SwitchIndependent LoadBalancingAlgorithm : Dynamic Status : Up Name : bond_bond-backup Members : {ens2f0, ens4f1} TeamNics : bond-backup TeamingMode : SwitchIndependent LoadBalancingAlgorithm : Dynamic Status : Up PS C:\> Get-NetAdapter Name InterfaceDescription ifIndex Status MacAddress LinkSpeed ---- -------------------- ------- ------ ---------- --------- bond-srv.9 Microsoft Network Adapter Multiple...#4 30 Up 40-A6-B7-27-96-D5 50 Gbps bond-srv Microsoft Network Adapter Multiple...#3 19 Up 40-A6-B7-27-96-D5 50 Gbps bond-iscsi Microsoft Network Adapter Multiple...#2 18 Up 40-A6-B7-27-96-D4 50 Gbps bond-backup Microsoft Network Adapter Multiplexo... 22 Up 68-05-CA-C9-EB-88 20 Gbps eno2 Intel(R) Ethernet Connection X722 fo... 12 Disabled 38-68-DD-36-A0-59 1 Gbps ens3f0 Intel(R) Ethernet Network Adapter XX... 11 Up 40-A6-B7-27-96-D4 25 Gbps ………………………………………….. omittedCode block. Run Get-NetLbfoTeam Delete the existing IP’s VLAN information from the server.
- In the following example, enter the verified ID instead of
30.Color modePS C:\> Remove-NetLbfoTeamNic -Team bond_bond-srv -VlanID 30 -Confirm:$false PS C:\> Get-NetAdapter Name InterfaceDescription ifIndex Status MacAddress LinkSpeed ---- -------------------- ------- ------ ---------- --------- bond-srv Microsoft Network Adapter Multiple...#3 19 Up 40-A6-B7-27-96-D5 50 Gbps bond-iscsi Microsoft Network Adapter Multiple...#2 18 Up 40-A6-B7-27-96-D4 50 Gbps bond-backup Microsoft Network Adapter Multiplexo... 22 Up 68-05-CA-C9-EB-88 20 Gbps eno2 Intel(R) Ethernet Connection X722 fo... 12 Disabled 38-68-DD-36-A0-59 1 Gbps ens3f0 Intel(R) Ethernet Network Adapter XX... 11 Up 40-A6-B7-27-96-D4 25 Gbps ………………………………………….. omittedPS C:\> Remove-NetLbfoTeamNic -Team bond_bond-srv -VlanID 30 -Confirm:$false PS C:\> Get-NetAdapter Name InterfaceDescription ifIndex Status MacAddress LinkSpeed ---- -------------------- ------- ------ ---------- --------- bond-srv Microsoft Network Adapter Multiple...#3 19 Up 40-A6-B7-27-96-D5 50 Gbps bond-iscsi Microsoft Network Adapter Multiple...#2 18 Up 40-A6-B7-27-96-D4 50 Gbps bond-backup Microsoft Network Adapter Multiplexo... 22 Up 68-05-CA-C9-EB-88 20 Gbps eno2 Intel(R) Ethernet Connection X722 fo... 12 Disabled 38-68-DD-36-A0-59 1 Gbps ens3f0 Intel(R) Ethernet Network Adapter XX... 11 Up 40-A6-B7-27-96-D4 25 Gbps ………………………………………….. omittedCode block. Run Get-NetLbfoTeam
- In the following example, enter the verified ID instead of
Once all tasks are completed, restart the server and then perform a final check.
ReferenceIt is recommended to perform the final check after restarting the server.If there are no issues in the final inspection results, select the work completion checkbox of Step 3 in the IP Change popup window.
4.2.1 - Installing ServiceWatch Agent
Users can install the ServiceWatch Agent on Bare Metal Server to collect custom metrics and logs.
ServiceWatch Agent
The agents that need to be installed on Bare Metal Server for ServiceWatch’s custom metrics and log collection can be broadly divided into two types: Prometheus Exporter and Open Telemetry Collector.
| Category | Description | |
|---|---|---|
| Prometheus Exporter | Provides metrics of specific applications or services in a format that Prometheus can scrape
| |
| Open Telemetry Collector | Serves as a centralized collector that collects telemetry data such as metrics and logs from distributed systems, processes them (filtering, sampling, etc.), and sends them to multiple backends (e.g., Prometheus, Jaeger, Elasticsearch, etc.)
|
Prerequisites for Using ServiceWatch Agent
To use the ServiceWatch Agent, please prepare the prerequisites by referring to Pre-environment Setup for ServiceWatch Agent.
4.2.2 - RHEL Repo 설정하기
The Samsung Cloud Platform Console provides the SCP RHEL Repository to support user environments where external access is restricted, such as VPC Private Subnets. By using the SCP RHEL Repository, you can install and download the same packages as the official RHEL Repository.
- If the user created RHEL before August 2025 through the Samsung Cloud Platform Console, they must modify the RHEL Repository settings.
- Since the SCP RHEL Repository synchronizes with each Region Local Repository according to an internal schedule, it is recommended to switch to an external public mirror site to apply the latest patches quickly.
- Samsung Cloud Platform provides the latest repository for the specified major version.
RHEL Repository Configuration Guide
When using RHEL, you can configure the SCP RHEL Repository to install and download the same packages as the official RHEL Repository. To set up the RHEL repository, follow these steps.
On the Virtual Server, as the OS root user, use the cat command to check the
/etc/yum.repos.d/scp.rhel8.repoor/etc/yum.repos.d/scp.rhel9.repoconfiguration.Color modecat /etc/yum.repos.d/scp.rhel8.repocat /etc/yum.repos.d/scp.rhel8.repo코드블록. repo 설정 확인(RHEL8) Color modecat /etc/yum.repos.d/scp.rhel9.repocat /etc/yum.repos.d/scp.rhel9.repo코드블록. repo 설정 확인(RHEL9) - When checking the configuration file, the following result is displayed.Color mode
[rhel-8-baseos] name=rhel-8-baseos gpgcheck=0 enabled=1 baseurl=http://scp-rhel8-ip/rhel/8/baseos [rhel-8-baseos-debug] name=rhel-8-baseos-debug gpgcheck=0 enabled=1 baseurl=http://scp-rhel8-ip/rhel/8/baseos-debug [rhel-8-appstream] name=rhel-8-appstream gpgcheck=0 enabled=1 baseurl=http://scp-rhel8-ip/rhel/8/appstream[rhel-8-baseos] name=rhel-8-baseos gpgcheck=0 enabled=1 baseurl=http://scp-rhel8-ip/rhel/8/baseos [rhel-8-baseos-debug] name=rhel-8-baseos-debug gpgcheck=0 enabled=1 baseurl=http://scp-rhel8-ip/rhel/8/baseos-debug [rhel-8-appstream] name=rhel-8-appstream gpgcheck=0 enabled=1 baseurl=http://scp-rhel8-ip/rhel/8/appstream코드블록. repo 설정 확인(RHEL8) Color mode[rhel-9-for-x86_64-baseos-rpms] name=rhel-9-for-x86_64-baseos-rpms baseurl=http://scp-rhel9-ip/rhel/$releasever/x86_64/baseos gpgcheck=0 enabled=1 [rhel-9-for-x86_64-appstream-rpms] name=rhel-9-for-x86_64-appstream-rpms baseurl=http://scp-rhel9-ip/rhel/$releasever/x86_64/appstream gpgcheck=0 enabled=1 [codeready-builder-for-rhel-9-x86_64-rpms] name=codeready-builder-for-rhel-9-x86_64-rpms baseurl=http://scp-rhel9-ip/rhel/$releasever/x86_64/codeready-builder gpgcheck=0 enabled=1 [rhel-9-for-x86_64-highavailability-rpms] name=rhel-9-for-x86_64-highavailability-rpms baseurl=http://scp-rhel9-ip/rhel/$releasever/x86_64/ha gpgcheck=0 enabled=1 [rhel-9-for-x86_64-supplementary-rpms] name=rhel-9-for-x86_64-supplementary-rpms baseurl=http://scp-rhel9-ip/rhel/$releasever/x86_64/supplementary gpgcheck=0 enabled=1[rhel-9-for-x86_64-baseos-rpms] name=rhel-9-for-x86_64-baseos-rpms baseurl=http://scp-rhel9-ip/rhel/$releasever/x86_64/baseos gpgcheck=0 enabled=1 [rhel-9-for-x86_64-appstream-rpms] name=rhel-9-for-x86_64-appstream-rpms baseurl=http://scp-rhel9-ip/rhel/$releasever/x86_64/appstream gpgcheck=0 enabled=1 [codeready-builder-for-rhel-9-x86_64-rpms] name=codeready-builder-for-rhel-9-x86_64-rpms baseurl=http://scp-rhel9-ip/rhel/$releasever/x86_64/codeready-builder gpgcheck=0 enabled=1 [rhel-9-for-x86_64-highavailability-rpms] name=rhel-9-for-x86_64-highavailability-rpms baseurl=http://scp-rhel9-ip/rhel/$releasever/x86_64/ha gpgcheck=0 enabled=1 [rhel-9-for-x86_64-supplementary-rpms] name=rhel-9-for-x86_64-supplementary-rpms baseurl=http://scp-rhel9-ip/rhel/$releasever/x86_64/supplementary gpgcheck=0 enabled=1코드블록. repo 설정 확인(RHEL9)
- When checking the configuration file, the following result is displayed.
Open the
/etc/hostsfile using a text editor (e.g., vim).Edit the
/etc/hostsfile with the following content and save it.Color mode198.19.2.13 scp-rhel8-ip scp-rhel9-ip scp-rhel-ip198.19.2.13 scp-rhel8-ip scp-rhel9-ip scp-rhel-ip코드블록. /etc/hosts 파일 설정 변경 Use the yum command to verify the RHEL Repository connection configured on the server.
Color modeyum repolist –vyum repolist –v코드블록. repository 연결 설정 확인 - If the RHEL repository is successfully connected, you can view the repository list.Color mode
Repo-id : rhel-8-appstream Repo-name : rhel-8-appstream Repo-revision : 1718903734 Repo-updated : Fri 21 Jun 2024 02:15:34 AM KST Repo-pkgs : 38,260 Repo-available-pkgs: 25,799 Repo-size : 122 G Repo-baseurl : http://scp-rhel8-ip/rhel/8/appstream Repo-expire : 172,800 second(s) (last: Thu 08 Aug 2024 07:27:57 AM KST) Repo-filename : /etc/yum.repos.d/scp.rhel8.repo Repo-id : rhel-8-baseos Repo-name : rhel-8-baseos Repo-revision : 1718029433 Repo-updated : Mon 10 Jun 2024 11:23:52 PM KST Repo-pkgs : 17,487 Repo-available-pkgs: 17,487 Repo-size : 32 G Repo-baseurl : http://scp-rhel8-ip/rhel/8/baseos Repo-expire : 172,800 second(s) (last: Thu 08 Aug 2024 07:27:57 AM KST) Repo-filename : /etc/yum.repos.d/scp.rhel8.repo Repo-id : rhel-8-baseos-debug Repo-name : rhel-8-baseos-debug Repo-revision : 1717662461 Repo-updated : Thu 06 Jun 2024 05:27:41 PM KST Repo-pkgs : 17,078 Repo-available-pkgs: 17,078 Repo-size : 100 G Repo-baseurl : http://scp-rhel8-ip/rhel/8/baseos-debug Repo-expire : 172,800 second(s) (last: Thu 08 Aug 2024 07:27:57 AM KST) Repo-filename : /etc/yum.repos.d/scp.rhel8.repoRepo-id : rhel-8-appstream Repo-name : rhel-8-appstream Repo-revision : 1718903734 Repo-updated : Fri 21 Jun 2024 02:15:34 AM KST Repo-pkgs : 38,260 Repo-available-pkgs: 25,799 Repo-size : 122 G Repo-baseurl : http://scp-rhel8-ip/rhel/8/appstream Repo-expire : 172,800 second(s) (last: Thu 08 Aug 2024 07:27:57 AM KST) Repo-filename : /etc/yum.repos.d/scp.rhel8.repo Repo-id : rhel-8-baseos Repo-name : rhel-8-baseos Repo-revision : 1718029433 Repo-updated : Mon 10 Jun 2024 11:23:52 PM KST Repo-pkgs : 17,487 Repo-available-pkgs: 17,487 Repo-size : 32 G Repo-baseurl : http://scp-rhel8-ip/rhel/8/baseos Repo-expire : 172,800 second(s) (last: Thu 08 Aug 2024 07:27:57 AM KST) Repo-filename : /etc/yum.repos.d/scp.rhel8.repo Repo-id : rhel-8-baseos-debug Repo-name : rhel-8-baseos-debug Repo-revision : 1717662461 Repo-updated : Thu 06 Jun 2024 05:27:41 PM KST Repo-pkgs : 17,078 Repo-available-pkgs: 17,078 Repo-size : 100 G Repo-baseurl : http://scp-rhel8-ip/rhel/8/baseos-debug Repo-expire : 172,800 second(s) (last: Thu 08 Aug 2024 07:27:57 AM KST) Repo-filename : /etc/yum.repos.d/scp.rhel8.repo코드블록. Repository list 확인
- If the RHEL repository is successfully connected, you can view the repository list.
4.3 - API Reference
4.4 - CLI Reference
4.5 - Release Note
Bare Metal Server
- BM 4th generation based on Intel 6th generation (Granite Rapids) Processor has been released.
- Bare Metal Server s4 and h4 server types have been added.
- For details on s4/h4 server types, please refer to s3/h3 server types.
- Local disk partition feature provided
- You can now create and use up to 10 local disk partitions.
- You can release multiple resources simultaneously from the Bare Metal Server list.
- You can change the IP of a general Subnet.
- OS Images have been added.
- RHEL 8.10, Ubuntu 24.04
- Bare Metal Server features added
- Distributes servers belonging to the same Placement Group across different racks.
- OS Images added (RHEL 9.4, Rocky Linux 8.6, Rocky Linux 9.4)
- 3rd generation (s3/h3) server types based on Intel 4th generation (Sapphire Rapids) Processor added. For details, please refer to Bare Metal Server Server Types.
- Samsung Cloud Platform common feature changes
- Common CX changes for Account, IAM and Service Home, tags, etc. have been reflected.
- Bare Metal Server service has been officially released.
- Bare Metal Server service has been released, which allows customers to exclusively use physical servers without virtualization.
5 - Multi-node GPU Cluster
5.1 - Overview
Service Overview
Multi-node GPU Cluster is a service that provides physical GPU servers without virtualization for large-scale high-performance AI calculations. It can cluster multiple GPUs using two or more bare metal servers with GPUs, and can be used conveniently with Samsung Cloud Platform’s high-performance storage and networking services.
Provided Features
Multi-node GPU Cluster provides the following functions.
- Auto Provisioning and Management: Through the web-based Console, you can easily use the standard GPU Bare Metal model server with 8 GPU cards, from provisioning to resource and cost management.
- Network Connection: Two or more Bare Metal Servers can be clustered through high-speed interconnects to process multiple GPUs, and by configuring the GPU Direct RDMA (Remote Direct Memory Access) environment, direct data IO between GPU memories is possible, enabling high-speed AI/Machine Learning calculations.
- Storage Connection: It provides various additional connection storages other than OS disks. High-speed network and high-performance SSD NAS File Storage, Block Storage, and Object Storage that are directly linked can also be used in conjunction.
- Network Setting Management: The server’s subnet/IP can be easily changed with the initially set value. NAT IP provides a management function that can be used or cancelled according to needs.
- Monitoring: You can check the monitoring information of computing resources such as CPU, GPU, Memory, Disk, etc. through Cloud Monitoring. To use the Cloud Monitoring service for Multi-node GPU Cluster, you need to install the Agent. Please install the Agent for stable service use. For more information, please refer to Multi-node GPU Cluster Monitoring Metrics.
Component
Multi-node GPU Cluster provides GPU as a Bare Metal Server type with standard images and server types, and NVSwitch and NVLink are provided.
GPU(H100)
GPU (Graphic Processing Unit) is specialized in parallel calculations that can process a large amount of data quickly, enabling large-scale parallel calculation processing in fields such as artificial intelligence (AI) and data analysis.
The following are the specifications of the GPU Type provided by the Multi-node GPU Cluster service.
| Classification | H100 Type |
|---|---|
| Product Provisioning Method | Bare Metal |
| GPU Architecture | NNVIDIA Hopper |
| GPU Memory | 80GB |
| GPU Transistors | 80 billion 4N TSMC |
| GPU Tensor Performance(based on FP16) | 989.4 TFLOPs, 1,978.9 TFLOPs* |
| GPU Memory Bandwidth | 3,352 GB/sec HBM3 |
| GPU CUDA Cores | 16,896 Cores |
| GPU Tensor Cores | 528(4th Generation) |
| NVLink performance | NVLink 4 |
| Total NVLink bandwidth | 900 GB/s |
| NVLink Signaling Rate | 25 Gbps (x18) |
| NVSwitch performance | NVSwitch 3 |
| NVSwitch GPU bandwidth | 900 GB/s |
| Total NVSwitch Aggregate Bandwidth | 7.2TB/s |
- With Sparsity
OS and GPU Driver Version
The operating systems (OS) supported by Multi-node GPU Cluster are as follows.
| OS | OS version | GPU driver version |
|---|---|---|
| Ubuntu | 22.04 | 535.86.10, 535.183.06 |
Server Type
The server types provided by Multi-node GPU Cluster are as follows. For a detailed description of the server types provided by Multi-node GPU Cluster, please refer to Multi-node GPU Cluster server type.
g2c96h8_metal
| Classification | Example | Detailed Description |
|---|---|---|
| Server Generation | g2 | Provided server generation
|
| CPU | c96 | Number of Cores
|
| GPU | h8 | GPU type and quantity
|
Preceding Service
This is a list of services that must be pre-configured before creating this service. Please refer to the guide provided for each service and prepare in advance for more details.
| Service Category | Service | Detailed Description |
|---|---|---|
| Networking | VPC | A service that provides an independent virtual network in a cloud environment |
5.1.1 - Server Type
Multi-node GPU Cluster Server Type
Multi-node GPU Cluster is divided based on the provided GPU Type, and the GPU used in the Multi-node GPU Cluster is determined by the server type selected when creating a GPU Node. Please select the server type according to the specifications of the application you want to run in the Multi-node GPU Cluster.
The server types supported by Multi-node GPU Cluster are in the following format:
g2c96h8_metal
| Classification | Example | Detailed Description |
|---|---|---|
| Server Generation | g2 | Provided server generation
|
| CPU | c96 | Number of cores
|
| GPU | h8 | GPU type and quantity
|
g2 Server Type
The g2 server type is a GPU Bare Metal Server using NVIDIA H100 Tensor Core GPU, suitable for large-scale high-performance AI computing.
- Provides up to 8 NVIDIA H100 Tensor Core GPUs
- Each GPU has 16,896 CUDA cores and 528 Tensor cores
- Supports up to 96 vCPUs and 1,920 GB of memory
- Supports up to 100 Gbps networking speed
- 900GB/s GPU and NVIDIA NVSwitch P2P communication
| Server Type | GPU | GPU Memory | CPU(Core) | Memory | Disk | GPU P2P |
|---|---|---|---|---|---|---|
| g2c96h8_metal | H100 | 640 GB | 96 vCore | 2 TB | SSD(OS) 960 GB * 2, NVMeSSD 3.84 TB * 4 | 900GB/s NVSwitch |
5.1.2 - Monitoring Metrics
Multi-node GPU Cluster monitoring metrics
The following table shows the monitoring metrics of Multi-node GPU Cluster that can be checked through Cloud Monitoring.
Multi-node GPU Cluster [Cluster]
| Performance Item | Detailed Description | Unit |
|---|---|---|
| Memory Total [Basic] | Available memory bytes | bytes |
| Memory Used [Basic] | Currently used memory bytes | bytes |
| Memory Swap In [Basic] | Replaced memory bytes | bytes |
| Memory Swap Out [Basic] | Swapped out memory bytes | bytes |
| Memory Free [Basic] | Unused memory bytes | bytes |
| Disk Read Bytes [Basic] | Read bytes | bytes |
| Disk Read Requests [Basic] | Number of Read Requests | cnt |
| Disk Write Bytes [Basic] | Write bytes | bytes |
| Disk Write Requests [Basic] | Write Request Count | cnt |
| CPU Usage [Basic] | 1-minute average system CPU usage rate | % |
| Instance State [Basic] | Instance Status | state |
| Network In Bytes [Basic] | Received bytes | bytes |
| Network In Dropped [Basic] | Receive Packet Drop | cnt |
| Network In Packets [Basic] | Received Packet Count | cnt |
| Network Out Bytes [Basic] | Transmission bytes | bytes |
| Network Out Dropped [Basic] | Transmission Packet Drop | cnt |
| Network Out Packets [Basic] | Transmission Packet Count | cnt |
| Performance Item | Detailed Description | Unit |
|---|---|---|
| Cluster GPU Count | GPU Count SUM in the Cluster
| cnt |
| Cluster GPU Count In Use | Number of GPUs used by jobs in the cluster
| cnt |
| Cluster GPU Usage | Cluster internal GPU Utilization AVG
| % |
| Cluster GPU Memory Usage [Avg] | Cluster GPU Memory Utilization AVG
| % |
Multi-node GPU Cluster [Node]
| Performance Item | Detailed Description | Unit |
|---|---|---|
| Memory Total [Basic] | Available memory bytes | bytes |
| Memory Used [Basic] | bytes of memory currently being used | bytes |
| Memory Swap In [Basic] | Replaced memory bytes | bytes |
| Memory Swap Out [Basic] | Swapped out memory bytes | bytes |
| Memory Free [Basic] | unused memory bytes | bytes |
| Disk Read Bytes [Basic] | Read bytes | bytes |
| Disk Read Requests [Basic] Disk Read Requests Count | cnt | |
| Disk Write Bytes [Basic] | Write bytes | bytes |
| Disk Write Requests [Basic] | Write Request Count | cnt |
| CPU Usage [Basic] | 1-minute average system CPU usage rate | % |
| Instance State [Basic] | Instance Status | state |
| Network In Bytes [Basic] | Received bytes | bytes |
| Network In Dropped [Basic] | Received Packet Drop | cnt |
| Network In Packets [Basic] | Received Packet Count | cnt |
| Network Out Bytes [Basic] | Transmission bytes | bytes |
| Network Out Dropped [Basic] | Transmission Packet Drop | cnt |
| Network Out Packets [Basic] | Transmission packet count | cnt |
| Performance Item | Detailed Description | Unit |
|---|---|---|
| GPU Count | GPU count | cnt |
| GPU Temperature | GPU temperature | ℃ |
| GPU Usage | utilization | % |
| GPU Usage [Avg] | GPU Overall Average Usage Rate(%) | % |
| GPU Power Cap | GPU’s maximum power capacity | W |
| GPU Power Usage | GPU’s current power usage | W |
| GPU Memory Usage [Avg] | GPU Memory Uti. AVG | % |
| GPU Count in use | Node’s Job-running GPU count | cnt |
| Execution Status for nvidia-smi | nvidia-smi command execution result | status |
| Core Usage [IO Wait] | The ratio of CPU time spent in waiting state (disk waiting) | % |
| Core Usage [System] | The ratio of CPU time spent in kernel space | % |
| -Core Usage [User]- | -The ratio of CPU time spent in user space- | -%- |
| CPU Cores | The number of CPU cores on the host. The maximum value of the unnormalized ratio is 100%* of the cores. The unnormalized ratio already reflects this value, and the maximum value is 100%* of the cores. | cnt |
| CPU Usage [Active] | Percentage of CPU time used excluding Idle and IOWait states (if all 4 cores are used at 100%: 400%) | % |
| CPU Usage [Idle] | The ratio of CPU time spent in idle state. | % |
| CPU Usage [IO Wait] | The percentage of CPU time spent in waiting state (disk waiting) | % |
| CPU Usage [System] | Percentage of CPU time used by the kernel (in case of using all 4 cores 100%: 400%) | % |
| CPU Usage [User] | Percentage of CPU time used in the user area. (In case of using all 4 cores 100%, 400%) | % |
| CPU Usage/Core [Active] | Percentage of CPU time used excluding Idle and IOWait states (normalized value by number of cores, 100% if all 4 cores are used at 100%) | % |
| CPU Usage/Core [Idle] | The ratio of CPU time spent in idle state. | % |
| CPU Usage/Core [IO Wait] | The ratio of CPU time spent in waiting state (disk waiting) | % |
| CPU Usage/Core [System] | Percentage of CPU time used by the kernel (normalized value by number of cores, 100% if all 4 cores are used at 100%) | % |
| CPU Usage/Core [User] | Percentage of CPU time used in the user area. (normalized value by number of cores, 100% if all 4 cores are used at 100%) | % |
| Disk CPU Usage [IO Request] | The ratio of CPU time spent executing input/output requests for the device (device bandwidth utilization). If this value is close to 100%, the device is in a saturated state. | % |
| Disk Queue Size [Avg] | The average queue length of requests executed for the device. | num |
| Disk Read Bytes | The number of bytes read from the device per second. | bytes |
| Disk Read Bytes [Delta Avg] | Average of system.diskio.read.bytes_delta for each disk | bytes |
| Disk Read Bytes [Delta Max] | Individual disks’ system.diskio.read.bytes_delta maximum | bytes |
| Disk Read Bytes [Delta Min] | Individual disks’ minimum system.diskio.read.bytes_delta | bytes |
| Disk Read Bytes [Delta Sum] | The sum of system.diskio.read.bytes_delta of individual disks | bytes |
| Disk Read Bytes [Delta] | Delta value of system.diskio.read.bytes for each disk | bytes |
| Disk Read Bytes [Success] | The total number of bytes read successfully. In Linux, it is assumed that the sector size is 512 and the value multiplied by the number of sectors read by 512 | bytes |
| Disk Read Requests | The number of read requests for the disk device in 1 second | cnt |
| Disk Read Requests [Delta Avg] | Average of system.diskio.read.count_delta for each disk | cnt |
| Disk Read Requests [Delta Max] | Maximum of system.diskio.read.count_delta for individual disks | cnt |
| Disk Read Requests [Delta Min] | Minimum of system.diskio.read.count_delta for each disk | cnt |
| Disk Read Requests [Delta Sum] | Sum of system.diskio.read.count_delta of individual disks | cnt |
| Disk Read Requests [Success Delta] | Individual disk’s system.diskio.read.count delta | cnt |
| Disk Read Requests [Success] | Total number of successful read completions | cnt |
| Disk Request Size [Avg] | The average size of requests executed for the device (unit: sector) | num |
| Disk Service Time [Avg] | The average service time (in milliseconds) for input requests executed on the device. | ms |
| Disk Wait Time [Avg] | The average time spent on requests executed for supported devices. | ms |
| Disk Wait Time [Read] | Disk Average Wait Time | ms |
| Disk Wait Time [Write] | Disk Average Wait Time | ms |
| Disk Write Bytes [Delta Avg] | Average of system.diskio.write.bytes_delta for each disk | bytes |
| Disk Write Bytes [Delta Max] | Maximum of system.diskio.write.bytes_delta for each disk | bytes |
| Disk Write Bytes [Delta Min] | Individual disks’ minimum system.diskio.write.bytes_delta | bytes |
| Disk Write Bytes [Delta Sum] | The sum of system.diskio.write.bytes_delta of individual disks | bytes |
| Disk Write Bytes [Delta] | Delta value of system.diskio.write.bytes for each disk | bytes |
| Disk Write Bytes [Success] | The total number of bytes written successfully. In Linux, it is assumed that the sector size is 512 and the value is multiplied by 512 to the number of sectors written | bytes |
| Disk Write Requests | The number of write requests to the disk device for 1 second | cnt |
| Disk Write Requests [Delta Avg] | Average of system.diskio.write.count_delta of individual disks | cnt |
| Disk Write Requests [Delta Max] | Maximum of system.diskio.write.count_delta for each disk | cnt |
| Disk Write Requests [Delta Min] | Minimum of system.diskio.write.count_delta for individual disks | cnt |
| Disk Write Requests [Delta Sum] | Sum of system.diskio.write.count_delta of individual disks | cnt |
| Disk Write Requests [Success Delta] | Individual disk’s system.diskio.write.count delta | cnt |
| Disk Write Requests [Success] | Total number of writes completed successfully | cnt |
| Disk Writes Bytes | The number of bytes written to the device per second. | bytes |
| Filesystem Hang Check | filesystem(local/NFS) hang check (normal:1, abnormal:0) | status |
| Filesystem Nodes | The total number of file nodes in the file system. | cnt |
| Filesystem Nodes [Free] | The total number of available file nodes in the file system. | cnt |
| Filesystem Size [Available] | This is the disk space (in bytes) that can be used by unauthorized users. | bytes |
| Filesystem Size [Free] | Available disk space (bytes) | bytes |
| Filesystem Size [Total] | Total Disk Space (bytes) | bytes |
| Filesystem Usage | Used Disk Space Percentage | % |
| Filesystem Usage [Avg] | Average of individual filesystem.used.pct | % |
| Filesystem Usage [Inode] | _inode usage rate | % |
| Filesystem Usage [Max] | Maximum value among individual filesystem usage percentages | % |
| Filesystem Usage [Min] | Minimum of individual filesystem used percentages | % |
| Filesystem Usage [Total] | - | % |
| Filesystem Used | Used Disk Space (bytes) | bytes |
| Filesystem Used [Inode] | Inode usage | bytes |
| Memory Free | The total amount of available memory (bytes). It does not include memory used by system cache and buffers (see system.memory.actual.free). | bytes |
| Memory Free [Actual] | Actual available memory (bytes). The calculation method varies depending on the OS, and in Linux, it is either MemAvailable from /proc/meminfo or calculated from available memory, cache, and buffer if meminfo is not available. On OSX, it is the sum of available memory and inactive memory. On Windows, it is the same as system.memory.free. | bytes |
| Memory Free [Swap] | Available swap memory. | bytes |
| Memory Total | Total Memory | bytes |
| Memory Total [Swap] | Total swap memory. | bytes |
| Memory Usage | Used memory percentage
| % |
| Memory Usage [Actual] | The percentage of memory actually used
| % |
| Memory Usage [Cache Swap] | Cache swap usage rate | % |
| Memory Usage [Swap] | Used swap memory percentage | % |
| Memory Used | Used Memory | bytes |
| Memory Used [Actual] | Actual used memory (bytes). The value subtracted from the total memory by the used memory. The available memory is calculated differently depending on the OS (refer to system.actual.free) | bytes |
| Memory Used [Swap] | Used swap memory. | bytes |
| Collisions | Network Collisions | cnt |
| Network In Bytes | Received byte count | bytes |
| Network In Bytes [Delta Avg] | Average of system.network.in.bytes_delta for each network | bytes |
| Network In Bytes [Delta Max] | Maximum of system.network.in.bytes_delta for each network | bytes |
| Network In Bytes [Delta Min] | Minimum of system.network.in.bytes_delta for each network | bytes |
| Network In Bytes [Delta Sum] | Sum of each network’s system.network.in.bytes_delta | bytes |
| Network In Bytes [Delta] | Received byte count delta | bytes |
| Network In Dropped | The number of packets deleted among incoming packets | cnt |
| Network In Errors | Number of errors during reception | cnt |
| Network In Packets | Received packet count | cnt |
| Network In Packets [Delta Avg] | Average of system.network.in.packets_delta for each network | cnt |
| Network In Packets [Delta Max] | Individual networks’ system.network.in.packets_delta maximum | cnt |
| Network In Packets [Delta Min] | Minimum of system.network.in.packets_delta for each network | cnt |
| Network In Packets [Delta Sum] | Sum of system.network.in.packets_delta of individual networks | cnt |
| Network In Packets [Delta] | Received packet count delta | cnt |
| Network Out Bytes | Transmitted byte count | bytes |
| Network Out Bytes [Delta Avg] | Average of system.network.out.bytes_delta for each network | bytes |
| Network Out Bytes [Delta Max] | Individual networks’ system.network.out.bytes_delta maximum | bytes |
| Network Out Bytes [Delta Min] | Minimum of system.network.out.bytes_delta for each network | bytes |
| Network Out Bytes [Delta Sum] | The sum of system.network.out.bytes_delta of individual networks | bytes |
| Network Out Bytes [Delta] | Transmitted byte count delta | bytes |
| Network Out Dropped | Number of packets dropped among outgoing packets. This value is not reported by the operating system, so it is always 0 in Darwin and BSD | cnt |
| Network Out Errors | Number of errors during transmission | cnt |
| Network Out Packets | Number of transmitted packets | cnt |
| Network Out Packets [Delta Avg] | Average of system.network.out.packets_delta for each network | cnt |
| Network Out Packets [Delta Max] | Maximum of system.network.out.packets_delta for each network | cnt |
| Network Out Packets [Delta Min] | Individual networks’ minimum system.network.out.packets_delta | cnt |
| Network Out Packets [Delta Sum] | Sum of system.network.out.packets_delta of individual networks | cnt |
| Network Out Packets [Delta] | Number of transmitted packets delta | cnt |
| Open Connections [TCP] | All open TCP connections | cnt |
| Open Connections [UDP] | All open UDP connections | cnt |
| Port Usage | Port usage available for connection | % |
| SYN Sent Sockets | Number of sockets in SYN_SENT state (when connecting from local to remote) | cnt |
| Kernel PID Max | kernel.pid_max value | cnt |
| Kernel Thread Max | kernel threads-max value | cnt |
| Process CPU Usage | Percentage of CPU time consumed by the process after the last update. This value is similar to the %CPU value of the process displayed by the top command on Unix systems | % |
| Process CPU Usage/Core | Percentage of CPU time used by the process since the last event, normalized by the number of cores, with a value between 0~100% | % |
| Process Memory Usage | main memory (RAM) where the process occupies a ratio | % |
| Process Memory Used | Resident Set size. The amount of memory a process occupies in RAM. In Windows, it is the current working set size | bytes |
| Process PID | Process PID | PID |
| Process PPID | Parent process’s pid | PID |
| Processes [Dead] | .dead processes count | cnt |
| Processes [Idle] | idle process count | cnt |
| Processes [Running] | Number of running processes | cnt |
| Processes [Sleeping] | sleeping processes count | cnt |
| Processes [Stopped] | Number of stopped processes | cnt |
| Processes [Total] | Total number of processes | cnt |
| Processes [Unknown] | Cannot search or unknown number of processes | cnt |
| Processes [Zombie] | Number of zombie processes | cnt |
| Running Process Usage | process usage rate | % |
| Running Processes | Number of running processes | cnt |
| Running Thread Usage | Thread usage rate | % |
| Running Threads | number of threads running in running processes | cnt |
| Instance Status | _instance status | state |
| Context Switches | context switch count (per second) | cnt |
| Load/Core [1 min] | Load for the last 1 minute divided by the number of cores | cnt |
| Load/Core [15 min] | The value of load divided by the number of cores for the last 15 minutes | cnt |
| Load/Core [5 min] | The value of load divided by the number of cores over the last 5 minutes | cnt |
| Multipaths [Active] | External storage connection path status = active count | cnt |
| Multipaths [Failed] | External storage connection path status = failed count | cnt |
| Multipaths [Faulty] | External storage connection path status = faulty count | cnt |
| NTP Offset | last sample’s measured offset (time difference between NTP server and local environment) | num |
| Run Queue Length | Execution Waiting Queue Length | num |
| Uptime | OS operation time (uptime). (milliseconds) | ms |
| Context Switchies | CPU context switch count (per second) | cnt |
| Disk Read Bytes [Sec] | number of bytes read from the windows logical disk in 1 second | cnt |
| Disk Read Time [Avg] | Data Read Average Time (sec) | sec |
| Disk Transfer Time [Avg] | Disk average wait time | sec |
| Disk Usage | Disk Usage Rate | % |
| Disk Write Bytes [Sec] | number of bytes written to the windows logical disk in 1 second | cnt |
| Disk Write Time [Avg] | Data Write Average Time (sec) | sec |
| Pagingfile Usage | Paging file usage rate | % |
| Pool Used [Non Paged] | Kernel memory Non-paged pool usage | bytes |
| Pool Used [Paged] | Paged Pool usage among kernel memory | bytes |
| Process [Running] | The number of processes currently running | cnt |
| Threads [Running] | Number of threads currently running | cnt |
| Threads [Waiting] | The number of threads waiting for processor time | cnt |
5.2 - How-to guides
The user can enter the required information for the Multi-node GPU Cluster service through the Samsung Cloud Platform Console, select detailed options, and create the service.
Multi-node GPU Cluster Getting Started
You can create and use a Multi-node GPU Cluster service in the Samsung Cloud Platform Console.
This service consists of GPU Node and Cluster Fabric services.
GPU Node Creation
To create a Multi-node GPU Cluster, follow the steps below.
- All Services > Compute > Multi-node GPU Cluster Click the menu. Navigate to the Service Home page of Multi-node GPU Cluster.
- Click the GPU Node creation button on the Service Home page. You will be taken to the GPU Node creation page.
- GPU Node creation on the page, enter the information required to create the service, and select detailed options.
- Image and Version Selection Select the required information in the area.
Category RequiredDetailed description Image Required Select provided image type - Ubuntu
Image Version Required Select version of the chosen image - Provides a list of versions of the provided server images
Table. GPU Node image and version selection items
- Image and Version Selection Select the required information in the area.
- Enter service information area, input or select the required information.
Category RequiredDetailed description Number of servers Required Number of GPU Node servers to create simultaneously - Only numbers can be entered, and the minimum number of servers to create is 2.
- Only during the initial setup can you create 2 or more, and expansion is possible one at a time.
Service Type > Server Type Required GPU Node Server Type - Select desired CPU, Memory, GPU, Disk specifications
- For detailed information about the server types provided by GPU Node, refer to Multi-node GPU Cluster Server Type
Service Type > Planned Compute Required Status of resources with Planned Compute set - In Use: Number of resources with Planned Compute that are currently in use
- Configured: Number of resources with Planned Compute set
- Coverage Preview: Amount applied per resource by Planned Compute
- Planned Compute Service Application: Go to the Planned Compute service application page
- For details, refer to Apply for Planned Compute
Table. GPU Node Service Information Input Items- Required Information Input area, enter or select the required information.
Category Required or notDetailed description Administrator Account Required Set the administrator account and password to be used when connecting to the server - Ubuntu OS is provided fixed as root
Server Name Prefix Required Enter a Prefix to distinguish each GPU Node generated when the number of selected servers is 2 or more - Automatically generated as user input value (prefix) + ‘
-###’ format
- Start with a lowercase English letter, and use lowercase letters, numbers, and special characters (
-) within 3 to 11 characters
- Must not end with a special character (
-)
Network Settings Required Set the network where the GPU Node will be installed - VPC Name:Select a pre-created VPC
- General Subnet Name: Select a pre-created general Subnet
- IP can be set to auto-generate or user input, and if input is selected, the user enters the IP directly
- NAT: Can be used only when there is 1 server and the VPC has an Internet Gateway attached. Checking ‘use’ allows selection of a NAT IP. (When first created, it is generated only with 2 or more servers, so modify on the resource detail page)
- NAT IP: Select NAT IP
- If there is no NAT IP to select, click the Create New button to generate a Public IP
- Click the Refresh button to view and select the created Public IP
- Creating a Public IP incurs charges according to the Public IP pricing policy
Table. GPU Node required information entry items
- Cluster selection area, create or select a Cluster Fabric.
Category RequiredDetailed description Cluster Fabric Required Setting of a group of GPU Node servers that can apply GPU Direct RDMA together - Optimal GPU performance and speed can be secured only within the same Cluster Fabric
- When creating a new Cluster Fabric, select *New Input > Node pool, then enter the name of the Cluster Fabric to be created
- To add to an existing Cluster Fabric, select Existing Input > Node pool, then select the already created Cluster Fabric
Table. GPU Node Cluster Fabric selection items- Additional Information Input Enter or select the required information in the area.
Category Required or notDetailed description Lock Select Using Lock prevents accidental actions that could terminate/start/stop the server Init Script Select Script to run when the server starts - Init Script must be selected differently depending on the image type
- For Linux: Select Shell Script or cloud-init
Tag Select Add Tag - Up to 50 can be added per resource
- After clicking the Add Tag button, enter or select Key, Value values
Table. GPU Node additional information input items - Init Script must be selected differently depending on the image type
- Summary Check the detailed information and estimated billing amount generated in the panel, and click the Complete button.
- Once creation is complete, check the created resources on the GPU Node List page.
- When creating a service, the GPU MIG/ECC settings are reset. However, to apply the correct settings, perform a one-time reboot initially, verify whether the settings have been applied, and then use it.
- For detailed information about GPU MIG/ECC settings reset, please refer to GPU MIG/ECC Settings Reset Checklist Guide.
GPU Node Check Detailed Information
The Multi-node GPU Cluster service allows you to view and edit the full list of GPU Node resources and detailed information.
GPU Node Details page consists of Details, Tags, Job History tabs.
To view detailed information of the GPU Node, follow the steps below.
All Services > Compute > Multi-node GPU Cluster > GPU Node Click the menu. Navigate to the Service Home page of Multi-node GPU Cluster.
Click the GPU Node menu on the Service Home page. Navigate to the GPU Node List page.
- Resource items other than required columns can be added via the Settings button.
Category Required or notDetailed description Resource ID Select User-created GPU Node ID Cluster Fabric name required Cluster Fabric name created by the user Server name Required User-created GPU Node name Server Type Required Server type of GPU Node - User can check the number of cores, memory capacity, GPU type and count of the created resources
Image Required User-generated GPU Node image version IP Required IP of the GPU Node created by the user Status Required Status of the GPU Node created by the user Creation Time Select GPU Node creation time Table. GPU Node Resource List Items
- Resource items other than required columns can be added via the Settings button.
GPU Node List Click the resource to view detailed information. GPU Node Details You will be taken to the page.
- GPU Server Details At the top of the page, status information and descriptions of additional features are displayed.
Category Detailed description GPU Node status Status of GPU Node created by the user - Creating: State where the server is being created
- Running:: State where creation is complete and usable
- Editing:: State where IP is being changed
- Unknown: Error state
- Starting: State where the server is starting
- Stopping: State where the server is stopping
- Stopped: State where the server has stopped
- Terminating: State where termination is in progress
- Terminated: State where termination is complete
Server Control Button to change server status - Start: Start a stopped server
- Stop: Stop a running server
Service cancellation Button to cancel the service Table. GPU Node status information and additional features
- GPU Server Details At the top of the page, status information and descriptions of additional features are displayed.
Detailed Information
GPU Node List page’s Details Tab you can view the detailed information of the selected resource, and if necessary, edit the information.
| Category | Detailed description |
|---|---|
| Service | Service Name |
| Resource Type | Resource Type |
| SRN | Unique resource ID in Samsung Cloud Platform
|
| Resource Name | Resource Name
|
| Resource ID | Unique resource ID in the service |
| Creator | User who created the service |
| Creation time | Service creation time |
| Modifier | User who edited the service information |
| Modification date/time | Date and time when the service information was modified |
| Server name | Server name |
| Node pool | A group of nodes that can be grouped together with the same Cluster Fabric |
| Cluster Fabric name | Cluster Fabric name created by the user |
| Image/Version | Server’s OS image and version |
| Server Type | CPU, Memory, GPU, Information Display |
| Planned Compute | Resource status with Planned Compute set
|
| Lock | Display lock usage status
|
| Network | GPU Node network information
|
| Block Storage | Block Storage information connected to the server
|
| Init Script | View the Init Script content entered when creating the server |
Tag
GPU Node List page’s Tag Tab you can view the tag information of the selected resource, and add, modify, or delete it.
| Category | Detailed description |
|---|---|
| Tag List | Tag List
|
Work History
GPU Node List page’s Job History tab allows you to view the job history of the selected resource.
| Category | Detailed description |
|---|---|
| Work History List | Resource Change History
|
GPU Node Operation Control
If you need server control and management functions for the generated GPU Node resources, you can perform tasks on the GPU Node List or GPU Node Details page. You can start and stop the running GPU Node resources.
GPU Node Getting Started
You can start a stopped GPU Node. To start the GPU Node, follow the steps below.
- All Services > Compute > Multi-node GPU Cluster Click the menu. Navigate to the Service Home page of Multi-node GPU Cluster.
- Click the GPU Node menu on the Service Home page. You will be taken to the GPU Node List page.
- On the GPU Node List page, after selecting individual or multiple servers with the checkbox, you can Start via the More button at the top.
- GPU Node List page, click the resource. GPU Node Details page will be opened.
- GPU Node Details on the page, click the Start button at the top to start the server.
- Check the server status and complete the status change.
Stop GPU Node
You can stop a GPU Node that is active. To stop the GPU Node, follow the steps below.
- All Services > Compute > Multi-node GPU Cluster Click the menu. Move to Multi-node GPU Cluster’s Service Home page.
- Click the GPU Node menu on the Service Home page. You will be taken to the GPU Node List page.
- GPU Node List page, you can control individual or multiple servers by selecting the checkboxes and then using the Stop button at the top.
- GPU Node List page, click the resource. GPU Node Details page, navigate.
- GPU Node Details on the page, click the Stop button at the top to stop the server.
- Check the server status and complete the status change.
GPU Node Cancel
You can cancel unused GPU nodes to reduce operating costs. However, if you cancel the service, the running service may be stopped immediately, so you should consider the impact of service interruption sufficiently before proceeding with the cancellation.
To cancel the GPU Node, follow the steps below.
- All Services > Compute > Multi-node GPU Server Click the menu. Navigate to the Service Home page of the Multi-node GPU Cluster.
- Click the Cluster Fabric menu on the Service Home page. You will be taken to the Cluster Fabric List page.
- Cluster Fabric List page, select the resource to cancel, and click the Cancel Service button.
- Resources using the same Cluster Fabric can be terminated simultaneously.
- Once the termination is complete, check on the GPU Node List page whether the resources have been terminated.
The cases where GPU Node termination is not possible are as follows.
- When Block Storage (BM) is connected: Please disconnect the Block Storage (BM) connection first.
- If File Storage is connected: Please disconnect the File Storage first.
- When Lock is set: Please change the Lock setting to unused and try again.
- If the server that cannot be terminated simultaneously is included: Please re-select only the resources that can be terminated.
- If the Cluster Fabric of the server you want to terminate is different: Select only resources that use the same Cluster Fabric.
5.2.1 - Cluster Fabric Management
Cluster Fabric is a service that helps manage servers (GPU Nodes) included in a GPU Cluster. Using Cluster Fabric, you can move servers between GPU Clusters in the same Node pool and optimize the performance and speed of GPUs within the same GPU Cluster.
Creating Cluster Fabric
Cluster Fabric can be created together with a GPU Node, and it cannot be created or deleted separately. When all GPU Nodes within a Cluster Fabric are terminated, the Cluster Fabric is automatically deleted. If you haven’t created a GPU Node, please create one first. For more information, refer to Creating a GPU Node.
Checking Cluster Fabric Details
- Cluster Fabric can be created together with a GPU Node, and it cannot be created or deleted separately.
- When all GPU Nodes within a Cluster Fabric are terminated, the Cluster Fabric is automatically deleted.
- If you haven’t created a GPU Node, please create one first. For more information, refer to Creating a GPU Node.
You can check the created Cluster Fabric list and details, and move servers on the Cluster Fabric List page and Cluster Fabric Details page.
Click on All Services > Compute > Multi-node GPU Server menu. It will move to the Service Home page of the Multi-node GPU Cluster.
Click on the Cluster Fabric menu on the Service Home page. It will move to the Cluster Fabric List page.
- On the Cluster Fabric List page, you can view the list of resources of the GPU Cluster created by the user.
- Resource items other than required columns can be added through the Settings button.
Category RequiredDescription Resource ID Optional Cluster Fabric ID created by the user Cluster Fabric Name Required Cluster Fabric name created by the user Node Pool Optional A collection of nodes that can be bundled into the same Cluster Fabric Number of Servers Optional Number of GPU Nodes Server Type Optional Server type of the GPU Node - The user can check the number of cores, memory capacity, and GPU type and number of the created resource
Status Optional Status of the Cluster Fabric created by the user Creation Time Optional Time when the Cluster Fabric was created Table. Cluster Fabric resource list items
Click on the resource to check the details on the Cluster Fabric List page. It will move to the Cluster Fabric Details page.
- At the top of the Cluster Fabric Details page, status information and additional feature descriptions are displayed.
Category Description Cluster Fabric Status Status of the Cluster Fabric created by the user - Creating: Cluster creation in progress
- Active: Creation completed and available
- Editing: IP change in progress
- Deleting: Termination in progress
- Deleted: Termination completed
Add Target Server Function to move a server from another cluster to this cluster Table. Cluster Fabric status information and additional features
- At the top of the Cluster Fabric Details page, status information and additional feature descriptions are displayed.
Details
On the Details tab of the Cluster Fabric List page, you can check the details of the selected resource and bring in servers from other clusters.
| Category | Description |
|---|---|
| Service | Service category |
| Resource Type | Service name |
| SRN | Unique resource ID in Samsung Cloud Platform
|
| Resource Name | Resource name
|
| Resource ID | Unique resource ID in the service |
| Creator | User who created the service |
| Creation Time | Time when the service was created |
| Modifier | User who modified the service information |
| Modification Time | Time when the service information was modified |
| Cluster Fabric Name | Cluster Fabric name created by the user |
| Node Pool | A collection of nodes that can be bundled into the same Cluster Fabric |
| Target Server | List of GPU Nodes bound to the Cluster Fabric
|
Bringing in Cluster Fabric Servers
Using the Add Target Server feature on the Cluster Fabric Details page, you can bring in servers from other clusters and add them to the selected cluster.
- Click on All Services > Compute > Multi-node GPU Server menu. It will move to the Service Home page of the Multi-node GPU Cluster.
- Click on the Cluster Fabric menu on the Service Home page. It will move to the Cluster Fabric List page.
- Click on the resource to check the details on the Cluster Fabric List page. It will move to the Cluster Fabric Details page.
- Click the Add button on the right side of the target server on the details tab.
- The target server addition popup window opens.
- Cluster Fabric Select a cluster.
- The GPU Node bound to the selected cluster is retrieved, and you can select the GPU Node to bring in.
- The selected GPU Node is listed at the bottom with the GPU Node name.
- Click the Confirm button to complete.
- Click the Cancel button to cancel the task.
- Check if the added GPU Node is retrieved in the target server.
- The target server addition popup window opens.
Terminating Cluster Fabric
When all GPU Nodes within a Cluster Fabric are terminated, the Cluster Fabric is automatically deleted. For more information, refer to Terminating a GPU Node.
5.2.2 - Installing ServiceWatch Agent
Users can install ServiceWatch Agent on the GPU Node of Multi-node GPU Cluster to collect custom metrics and logs.
ServiceWatch Agent
The agents that need to be installed to collect ServiceWatch’s custom metrics and logs on the GPU Node of Multi-node GPU Cluster can be divided into two main types: Prometheus Exporter and Open Telemetry Collector.
| Item | Description | |
|---|---|---|
| Prometheus Exporter | Provides metrics of specific applications or services in a format that Prometheus can scrape
| |
| Open Telemetry Collector | Acts as a central collector that collects telemetry data such as metrics and logs from distributed systems, processes them (filtering, sampling, etc.), and sends them to multiple backends (e.g., Prometheus, Jaeger, Elasticsearch, etc.)
|
If Kubernetes Engine is configured on the GPU Node, please check GPU metrics through the metrics provided by Kubernetes Engine.
- If DCGM Exporter is installed on a GPU Node where Kubernetes Engine is configured, it may not operate normally.
Pre-settings for Using ServiceWatch Agent
To use ServiceWatch Agent, please prepare pre-settings by referring to Pre-environment Setup for ServiceWatch Agent.
5.2.3 - Multi-node GPU Cluster Service Scope and Inspection Guide
Multi-node GPU Cluster service scope
In the event of an IaaS HW level issue with the Multi-node GPU Cluster service, technical support can be received through the Support Center’s Contact Us. However, risks due to changes such as OS Kernel updates or application installation are the responsibility of the user, so technical support may be difficult, please be cautious when performing system updates or other tasks.
IaaS HW level problem
- IPMI(iLO) HW monitoring console where the server’s internal HW fault event occurrence message occurs
- GPU HW operation error confirmed in nvdia-smi command
- HW error messages occurring from InfiniBand HCA card or InfiniBand Switch inspection
IaaS HW Inspection Guide
After applying for the Multi-node GPU Cluster service, it is recommended to check the IaaS HW level according to the inspection guide.
OS Kernel and Package Holding
- If you do not want automatic updates of package versions, it is recommended to block package updates using the
apt-markcommand. - It is recommended to block the update of Linux kernel or IB related package versions.
To proceed with OS Kernel and Package holding, follow the procedure below.
- Use the following command to check the version of the kernel and IB-related packages.Color mode
root@bm-dev-001:~# dpkg -l | egrep -i "kernel | mlnx" root@bm-dev-001:~# dpkg -l | egrep -i "kernel | nvidia" root@bm-dev-001:~# dpkg -l | egrep -i "kernel | linux-image" ii crash 7.2.8-1ubuntu1.20.04.1 amd64 kernel debugging utility, allowing gdb like syntax ii dkms 2.8.1-5ubuntu2 all Dynamic Kernel Module Support Framework ii dmeventd 2:1.02.167-1ubuntu1 amd64 Linux Kernel Device Mapper event daemon ii dmsetup 2:1.02.167-1ubuntu1 amd64 Linux Kernel Device Mapper userspace library ii iser-dkms 5.4-OFED.5.4.3.0.1.1 all DKMS support fo iser kernel modules ii isert-dkms 5.4-OFED.5.4.3.0.1.1 all DKMS support fo isert kernel modules ii kernel-mft-dkms 4.17.2-12 all DKMS support for kernel-mft kernel modules ii kmod 27-1ubuntu2 amd64 tools for managing Linux kernel modules ii knem 1.1.4.90mlnx1-OFED.5.1.2.5.0.1 amd64 userspace tools for the KNEM kernel module ii knem-dkms 1.1.4.90mlnx1-OFED.5.1.2.5.0.1 all DKMS support for mlnx-ofed kernel modules ii libaio1:amd64 0.3.112-5 amd64 Linux kernel AIO access library - shared library ii libdevmapper-event1.02.1:amd64 2:1.02.167-1ubuntu1 amd64 Linux Kernel Device Mapper event support library ii libdevmapper1.02.1:amd64 2:1.02.167-1ubuntu1 amd64 Linux Kernel Device Mapper userspace library ii libdrm-amdgpu1:amd64 2.4.107-8ubuntu1~20.04.2 amd64 Userspace interface to amdgpu-specific kernel DRM services -- runtime ii libdrm-common 2.4.107-8ubuntu1~20.04.2 all Userspace interface to kernel DRM services -- common files ii libdrm-intel1:amd64 2.4.107-8ubuntu1~20.04.2 amd64 Userspace interface to intel-specific kernel DRM services -- runtime ii libdrm-nouveau2:amd64 2.4.107-8ubuntu1~20.04.2 amd64 Userspace interface to nouveau-specific kernel DRM services -- runtime ii libdrm-radeon1:amd64 2.4.107-8ubuntu1~20.04.2 amd64 Userspace interface to radeon-specific kernel DRM services -- runtime ii libdrm2:amd64 2.4.107-8ubuntu1~20.04.2 amd64 Userspace interface to kernel DRM services -- runtime ii linux-firmware 1.187.29 all Firmware for Linux kernel drivers hi linux-generic 5.4.0.105.109 amd64 Complete Generic Linux kernel and headers ii linux-headers-5.4.0-104 5.4.0-104.118 all Header files related to Linux kernel version 5.4.0 ii linux-headers-5.4.0-104-generic 5.4.0-104.118 amd64 Linux kernel headers for version 5.4.0 on 64 bit x86 SMP ii linux-headers-5.4.0-105 5.4.0-105.119 all Header files related to Linux kernel version 5.4.0 ii linux-headers-5.4.0-105-generic 5.4.0-105.119 amd64 Linux kernel headers for version 5.4.0 on 64 bit x86 SMP hi linux-headers-generic 5.4.0.105.109 amd64 Generic Linux kernel headers ii linux-image-5.4.0-104-generic 5.4.0-104.118 amd64 Signed kernel image generic ii linux-image-5.4.0-105-generic 5.4.0-105.119 amd64 Signed kernel image generic hi linux-image-generic 5.4.0.105.109 amd64 Generic Linux kernel image ii linux-libc-dev:amd64 5.4.0-105.119 amd64 Linux Kernel Headers for development ii linux-modules-5.4.0-104-generic 5.4.0-104.118 amd64 Linux kernel extra modules for version 5.4.0 on 64 bit x86 SMP ii linux-modules-5.4.0-105-generic 5.4.0-105.119 amd64 Linux kernel extra modules for version 5.4.0 on 64 bit x86 SMP ii linux-modules-extra-5.4.0-104-generic 5.4.0-104.118 amd64 Linux kernel extra modules for version 5.4.0 on 64 bit x86 SMP ii linux-modules-extra-5.4.0-105-generic 5.4.0-105.119 amd64 Linux kernel extra modules for version 5.4.0 on 64 bit x86 SMP ii mlnx-ofed-kernel-dkms 5.4-OFED.5.4.3.0.3.1 all DKMS support for mlnx-ofed kernel modules ii mlnx-ofed-kernel-utils 5.4-OFED.5.4.3.0.3.1 amd64 Userspace tools to restart and tune mlnx-ofed kernel modules ii mlnx-tools 5.2.0-0.54303 amd64 Userspace tools to restart and tune MLNX_OFED kernel modules ii nvidia-kernel-common-470 470.103.01-0ubuntu0.20.04.1 amd64 Shared files used with the kernel module ii nvidia-kernel-source-470 470.103.01-0ubuntu0.20.04.1 amd64 NVIDIA kernel source package ii nvidia-peer-memory 1.2-0 all nvidia peer memory kernel module. ii nvidia-peer-memory-dkms 1.2-0 all DKMS support for nvidia-peer-memory kernel modules ii rsyslog 8.2001.0-1ubuntu1.1 amd64 reliable system and kernel logging daemon ii srp-dkms 5.4-OFED.5.4.3.0.1.1 all DKMS support fo srp kernel modulesroot@bm-dev-001:~# dpkg -l | egrep -i "kernel | mlnx" root@bm-dev-001:~# dpkg -l | egrep -i "kernel | nvidia" root@bm-dev-001:~# dpkg -l | egrep -i "kernel | linux-image" ii crash 7.2.8-1ubuntu1.20.04.1 amd64 kernel debugging utility, allowing gdb like syntax ii dkms 2.8.1-5ubuntu2 all Dynamic Kernel Module Support Framework ii dmeventd 2:1.02.167-1ubuntu1 amd64 Linux Kernel Device Mapper event daemon ii dmsetup 2:1.02.167-1ubuntu1 amd64 Linux Kernel Device Mapper userspace library ii iser-dkms 5.4-OFED.5.4.3.0.1.1 all DKMS support fo iser kernel modules ii isert-dkms 5.4-OFED.5.4.3.0.1.1 all DKMS support fo isert kernel modules ii kernel-mft-dkms 4.17.2-12 all DKMS support for kernel-mft kernel modules ii kmod 27-1ubuntu2 amd64 tools for managing Linux kernel modules ii knem 1.1.4.90mlnx1-OFED.5.1.2.5.0.1 amd64 userspace tools for the KNEM kernel module ii knem-dkms 1.1.4.90mlnx1-OFED.5.1.2.5.0.1 all DKMS support for mlnx-ofed kernel modules ii libaio1:amd64 0.3.112-5 amd64 Linux kernel AIO access library - shared library ii libdevmapper-event1.02.1:amd64 2:1.02.167-1ubuntu1 amd64 Linux Kernel Device Mapper event support library ii libdevmapper1.02.1:amd64 2:1.02.167-1ubuntu1 amd64 Linux Kernel Device Mapper userspace library ii libdrm-amdgpu1:amd64 2.4.107-8ubuntu1~20.04.2 amd64 Userspace interface to amdgpu-specific kernel DRM services -- runtime ii libdrm-common 2.4.107-8ubuntu1~20.04.2 all Userspace interface to kernel DRM services -- common files ii libdrm-intel1:amd64 2.4.107-8ubuntu1~20.04.2 amd64 Userspace interface to intel-specific kernel DRM services -- runtime ii libdrm-nouveau2:amd64 2.4.107-8ubuntu1~20.04.2 amd64 Userspace interface to nouveau-specific kernel DRM services -- runtime ii libdrm-radeon1:amd64 2.4.107-8ubuntu1~20.04.2 amd64 Userspace interface to radeon-specific kernel DRM services -- runtime ii libdrm2:amd64 2.4.107-8ubuntu1~20.04.2 amd64 Userspace interface to kernel DRM services -- runtime ii linux-firmware 1.187.29 all Firmware for Linux kernel drivers hi linux-generic 5.4.0.105.109 amd64 Complete Generic Linux kernel and headers ii linux-headers-5.4.0-104 5.4.0-104.118 all Header files related to Linux kernel version 5.4.0 ii linux-headers-5.4.0-104-generic 5.4.0-104.118 amd64 Linux kernel headers for version 5.4.0 on 64 bit x86 SMP ii linux-headers-5.4.0-105 5.4.0-105.119 all Header files related to Linux kernel version 5.4.0 ii linux-headers-5.4.0-105-generic 5.4.0-105.119 amd64 Linux kernel headers for version 5.4.0 on 64 bit x86 SMP hi linux-headers-generic 5.4.0.105.109 amd64 Generic Linux kernel headers ii linux-image-5.4.0-104-generic 5.4.0-104.118 amd64 Signed kernel image generic ii linux-image-5.4.0-105-generic 5.4.0-105.119 amd64 Signed kernel image generic hi linux-image-generic 5.4.0.105.109 amd64 Generic Linux kernel image ii linux-libc-dev:amd64 5.4.0-105.119 amd64 Linux Kernel Headers for development ii linux-modules-5.4.0-104-generic 5.4.0-104.118 amd64 Linux kernel extra modules for version 5.4.0 on 64 bit x86 SMP ii linux-modules-5.4.0-105-generic 5.4.0-105.119 amd64 Linux kernel extra modules for version 5.4.0 on 64 bit x86 SMP ii linux-modules-extra-5.4.0-104-generic 5.4.0-104.118 amd64 Linux kernel extra modules for version 5.4.0 on 64 bit x86 SMP ii linux-modules-extra-5.4.0-105-generic 5.4.0-105.119 amd64 Linux kernel extra modules for version 5.4.0 on 64 bit x86 SMP ii mlnx-ofed-kernel-dkms 5.4-OFED.5.4.3.0.3.1 all DKMS support for mlnx-ofed kernel modules ii mlnx-ofed-kernel-utils 5.4-OFED.5.4.3.0.3.1 amd64 Userspace tools to restart and tune mlnx-ofed kernel modules ii mlnx-tools 5.2.0-0.54303 amd64 Userspace tools to restart and tune MLNX_OFED kernel modules ii nvidia-kernel-common-470 470.103.01-0ubuntu0.20.04.1 amd64 Shared files used with the kernel module ii nvidia-kernel-source-470 470.103.01-0ubuntu0.20.04.1 amd64 NVIDIA kernel source package ii nvidia-peer-memory 1.2-0 all nvidia peer memory kernel module. ii nvidia-peer-memory-dkms 1.2-0 all DKMS support for nvidia-peer-memory kernel modules ii rsyslog 8.2001.0-1ubuntu1.1 amd64 reliable system and kernel logging daemon ii srp-dkms 5.4-OFED.5.4.3.0.1.1 all DKMS support fo srp kernel modulesCode block. Kernel, IB related package version check - Use the apt-mark command to hold the package update.Color mode
# apt-mark hold <package name># apt-mark hold <package name>Code block. Package update hold
Intel E810 Driver Update
Check the version of the Intel E810 driver and update it to the recommended version.
- Server manufacturer Intel E810 driver recommended version: 1.15.4
- Driver (ice-1.15.4.tar.gz) download
The driver update method is as follows.
- Move the basic driver tar file to the desired directory.
Example: /home/username/ice or /usr/local/src/ice
Untar / unzip the Archiver file.
- x.x.x is the version number of the driver tar file.Color mode
tar zxf ice-x.x.x.tar.gztar zxf ice-x.x.x.tar.gzCode block. Unzip file
- x.x.x is the version number of the driver tar file.
Change to the driver src directory.
- x.x.x is the version number of the driver tar file.Color mode
cd ice-x.x.x/src/cd ice-x.x.x/src/Code block. Directory change
- x.x.x is the version number of the driver tar file.
Compile the driver module.
Color modemake installmake installCode Block. Driver Module Compile After the update is complete, check the version.
Color modelsmod | grep ice modinfo ice | grep versionlsmod | grep ice modinfo ice | grep versionCode Block. Version Check
NVIDIA driver check
nvidia-smi topo, IB nv_peer_mem status checkTo check the NVIDIA driver (nvidia-smi topo, IB nv_peer_mem status) and inspect the IaaS HW level, follow the next procedure.
Check the GPU driver and HW status.
Color modeuser@bm-dev-001:~$ nvidia-smi topo -m GPU0 GPU1 GPU2 GPU3 GPU4 GPU5 GPU6 GPU7 mlx5_0 mlx5_1 mlx5_2 mlx5_3 CPU Affinity NUMA Affinity GPU0 X NV12 NV12 NV12 NV12 NV12 NV12 NV12 SYS PXB SYS SYS 48-63 3 GPU1 NV12 X NV12 NV12 NV12 NV12 NV12 NV12 SYS PXB SYS SYS 48-63 3 GPU2 NV12 NV12 X NV12 NV12 NV12 NV12 NV12 PXB SYS SYS SYS 16-31 1 GPU3 NV12 NV12 NV12 X NV12 NV12 NV12 NV12 PXB SYS SYS SYS 16-31 1 GPU4 NV12 NV12 NV12 NV12 X NV12 NV12 NV12 SYS SYS SYS PXB 112-127 7 GPU5 NV12 NV12 NV12 NV12 NV12 X NV12 NV12 SYS SYS SYS PXB 112-127 7 GPU6 NV12 NV12 NV12 NV12 NV12 NV12 X NV12 SYS SYS PXB SYS 80-95 5 GPU7 NV12 NV12 NV12 NV12 NV12 NV12 NV12 X SYS SYS PXB SYS 80-95 5 mlx5_0 SYS SYS PXB PXB SYS SYS SYS SYS X SYS SYS SYS mlx5_1 PXB PXB SYS SYS SYS SYS SYS SYS SYS X SYS SYS mlx5_2 SYS SYS SYS SYS SYS SYS PXB PXB SYS SYS X SYS mlx5_3 SYS SYS SYS SYS PXB PXB SYS SYS SYS SYS SYS X Legend: X = Self SYS = Connection traversing PCIe as well as the SMP interconnect between NUMA nodes (e.g., QPI/UPI) NODE = Connection traversing PCIe as well as the interconnect between PCIe Host Bridges within a NUMA node PHB = Connection traversing PCIe as well as a PCIe Host Bridge (typically the CPU) PXB = Connection traversing multiple PCIe bridges (without traversing the PCIe Host Bridge) PIX = Connection traversing at most a single PCIe bridge NV# = Connection traversing a bonded set of # NVLinksuser@bm-dev-001:~$ nvidia-smi topo -m GPU0 GPU1 GPU2 GPU3 GPU4 GPU5 GPU6 GPU7 mlx5_0 mlx5_1 mlx5_2 mlx5_3 CPU Affinity NUMA Affinity GPU0 X NV12 NV12 NV12 NV12 NV12 NV12 NV12 SYS PXB SYS SYS 48-63 3 GPU1 NV12 X NV12 NV12 NV12 NV12 NV12 NV12 SYS PXB SYS SYS 48-63 3 GPU2 NV12 NV12 X NV12 NV12 NV12 NV12 NV12 PXB SYS SYS SYS 16-31 1 GPU3 NV12 NV12 NV12 X NV12 NV12 NV12 NV12 PXB SYS SYS SYS 16-31 1 GPU4 NV12 NV12 NV12 NV12 X NV12 NV12 NV12 SYS SYS SYS PXB 112-127 7 GPU5 NV12 NV12 NV12 NV12 NV12 X NV12 NV12 SYS SYS SYS PXB 112-127 7 GPU6 NV12 NV12 NV12 NV12 NV12 NV12 X NV12 SYS SYS PXB SYS 80-95 5 GPU7 NV12 NV12 NV12 NV12 NV12 NV12 NV12 X SYS SYS PXB SYS 80-95 5 mlx5_0 SYS SYS PXB PXB SYS SYS SYS SYS X SYS SYS SYS mlx5_1 PXB PXB SYS SYS SYS SYS SYS SYS SYS X SYS SYS mlx5_2 SYS SYS SYS SYS SYS SYS PXB PXB SYS SYS X SYS mlx5_3 SYS SYS SYS SYS PXB PXB SYS SYS SYS SYS SYS X Legend: X = Self SYS = Connection traversing PCIe as well as the SMP interconnect between NUMA nodes (e.g., QPI/UPI) NODE = Connection traversing PCIe as well as the interconnect between PCIe Host Bridges within a NUMA node PHB = Connection traversing PCIe as well as a PCIe Host Bridge (typically the CPU) PXB = Connection traversing multiple PCIe bridges (without traversing the PCIe Host Bridge) PIX = Connection traversing at most a single PCIe bridge NV# = Connection traversing a bonded set of # NVLinksCode Block. GPU Driver and HW Status Check Check the NVSwitch HW status.
Color modeuser@bm-dev-001:~$ nvidia-smi nvlink --status GPU 0: NVIDIA A100-SXM4-80GB (UUID: GPU-2c0d1d6b-e348-55fc-44cf-cd65a954b36c) Link 0: 25 GB/s Link 1: 25 GB/s Link 2: 25 GB/s Link 3: 25 GB/s Link 4: 25 GB/s Link 5: 25 GB/s Link 6: 25 GB/s Link 7: 25 GB/s Link 8: 25 GB/s Link 9: 25 GB/s Link 10: 25 GB/s Link 11: 25 GB/s GPU 1: NVIDIA A100-SXM4-80GB (UUID: GPU-96f429d8-893a-a9ea-deca-feffd90669e9) Link 0: 25 GB/s Link 1: 25 GB/s Link 2: 25 GB/s Link 3: 25 GB/s Link 4: 25 GB/s Link 5: 25 GB/s Link 6: 25 GB/s Link 7: 25 GB/s Link 8: 25 GB/s Link 9: 25 GB/s Link 10: 25 GB/s Link 11: 25 GB/s GPU 2: NVIDIA A100-SXM4-80GB (UUID: GPU-2e601952-b442-b757-a035-725cd320f589) Link 0: 25 GB/s Link 1: 25 GB/s Link 2: 25 GB/s Link 3: 25 GB/s Link 4: 25 GB/s Link 5: 25 GB/s Link 6: 25 GB/s Link 7: 25 GB/s Link 8: 25 GB/s Link 9: 25 GB/s Link 10: 25 GB/s Link 11: 25 GB/s GPU 3: NVIDIA A100-SXM4-80GB (UUID: GPU-bcbfd885-a9f8-ec8c-045b-c521472b4fed) Link 0: 25 GB/s Link 1: 25 GB/s Link 2: 25 GB/s Link 3: 25 GB/s Link 4: 25 GB/s Link 5: 25 GB/s Link 6: 25 GB/s Link 7: 25 GB/s Link 8: 25 GB/s Link 9: 25 GB/s Link 10: 25 GB/s Link 11: 25 GB/s GPU 4: NVIDIA A100-SXM4-80GB (UUID: GPU-30273090-2d78-fc7a-a360-ec5f871dd488) Link 0: 25 GB/s Link 1: 25 GB/s Link 2: 25 GB/s Link 3: 25 GB/s Link 4: 25 GB/s Link 5: 25 GB/s Link 6: 25 GB/s Link 7: 25 GB/s Link 8: 25 GB/s Link 9: 25 GB/s Link 10: 25 GB/s Link 11: 25 GB/s GPU 5: NVIDIA A100-SXM4-80GB (UUID: GPU-5ce7ef61-56dd-fb18-aa7c-be610c8d51c3) Link 0: 25 GB/s Link 1: 25 GB/s Link 2: 25 GB/s Link 3: 25 GB/s Link 4: 25 GB/s Link 5: 25 GB/s Link 6: 25 GB/s Link 7: 25 GB/s Link 8: 25 GB/s Link 9: 25 GB/s Link 10: 25 GB/s Link 11: 25 GB/s GPU 6: NVIDIA A100-SXM4-80GB (UUID: GPU-740a527b-b286-8b85-35eb-b6b41c0bb6d7) Link 0: 25 GB/s Link 1: 25 GB/s Link 2: 25 GB/s Link 3: 25 GB/s Link 4: 25 GB/s Link 5: 25 GB/s Link 6: 25 GB/s Link 7: 25 GB/s Link 8: 25 GB/s Link 9: 25 GB/s Link 10: 25 GB/s Link 11: 25 GB/s GPU 7: NVIDIA A100-SXM4-80GB (UUID: GPU-1fb6de95-60f6-dbf2-ffca-f7680577e37c) Link 0: 25 GB/s Link 1: 25 GB/s Link 2: 25 GB/s Link 3: 25 GB/s Link 4: 25 GB/s Link 5: 25 GB/s Link 6: 25 GB/s Link 7: 25 GB/s Link 8: 25 GB/s Link 9: 25 GB/s Link 10: 25 GB/s Link 11: 25 GB/suser@bm-dev-001:~$ nvidia-smi nvlink --status GPU 0: NVIDIA A100-SXM4-80GB (UUID: GPU-2c0d1d6b-e348-55fc-44cf-cd65a954b36c) Link 0: 25 GB/s Link 1: 25 GB/s Link 2: 25 GB/s Link 3: 25 GB/s Link 4: 25 GB/s Link 5: 25 GB/s Link 6: 25 GB/s Link 7: 25 GB/s Link 8: 25 GB/s Link 9: 25 GB/s Link 10: 25 GB/s Link 11: 25 GB/s GPU 1: NVIDIA A100-SXM4-80GB (UUID: GPU-96f429d8-893a-a9ea-deca-feffd90669e9) Link 0: 25 GB/s Link 1: 25 GB/s Link 2: 25 GB/s Link 3: 25 GB/s Link 4: 25 GB/s Link 5: 25 GB/s Link 6: 25 GB/s Link 7: 25 GB/s Link 8: 25 GB/s Link 9: 25 GB/s Link 10: 25 GB/s Link 11: 25 GB/s GPU 2: NVIDIA A100-SXM4-80GB (UUID: GPU-2e601952-b442-b757-a035-725cd320f589) Link 0: 25 GB/s Link 1: 25 GB/s Link 2: 25 GB/s Link 3: 25 GB/s Link 4: 25 GB/s Link 5: 25 GB/s Link 6: 25 GB/s Link 7: 25 GB/s Link 8: 25 GB/s Link 9: 25 GB/s Link 10: 25 GB/s Link 11: 25 GB/s GPU 3: NVIDIA A100-SXM4-80GB (UUID: GPU-bcbfd885-a9f8-ec8c-045b-c521472b4fed) Link 0: 25 GB/s Link 1: 25 GB/s Link 2: 25 GB/s Link 3: 25 GB/s Link 4: 25 GB/s Link 5: 25 GB/s Link 6: 25 GB/s Link 7: 25 GB/s Link 8: 25 GB/s Link 9: 25 GB/s Link 10: 25 GB/s Link 11: 25 GB/s GPU 4: NVIDIA A100-SXM4-80GB (UUID: GPU-30273090-2d78-fc7a-a360-ec5f871dd488) Link 0: 25 GB/s Link 1: 25 GB/s Link 2: 25 GB/s Link 3: 25 GB/s Link 4: 25 GB/s Link 5: 25 GB/s Link 6: 25 GB/s Link 7: 25 GB/s Link 8: 25 GB/s Link 9: 25 GB/s Link 10: 25 GB/s Link 11: 25 GB/s GPU 5: NVIDIA A100-SXM4-80GB (UUID: GPU-5ce7ef61-56dd-fb18-aa7c-be610c8d51c3) Link 0: 25 GB/s Link 1: 25 GB/s Link 2: 25 GB/s Link 3: 25 GB/s Link 4: 25 GB/s Link 5: 25 GB/s Link 6: 25 GB/s Link 7: 25 GB/s Link 8: 25 GB/s Link 9: 25 GB/s Link 10: 25 GB/s Link 11: 25 GB/s GPU 6: NVIDIA A100-SXM4-80GB (UUID: GPU-740a527b-b286-8b85-35eb-b6b41c0bb6d7) Link 0: 25 GB/s Link 1: 25 GB/s Link 2: 25 GB/s Link 3: 25 GB/s Link 4: 25 GB/s Link 5: 25 GB/s Link 6: 25 GB/s Link 7: 25 GB/s Link 8: 25 GB/s Link 9: 25 GB/s Link 10: 25 GB/s Link 11: 25 GB/s GPU 7: NVIDIA A100-SXM4-80GB (UUID: GPU-1fb6de95-60f6-dbf2-ffca-f7680577e37c) Link 0: 25 GB/s Link 1: 25 GB/s Link 2: 25 GB/s Link 3: 25 GB/s Link 4: 25 GB/s Link 5: 25 GB/s Link 6: 25 GB/s Link 7: 25 GB/s Link 8: 25 GB/s Link 9: 25 GB/s Link 10: 25 GB/s Link 11: 25 GB/sCode block. NVSwitch HW status check Check the InfiniBand(IB) HCA card HW status and Link.
Color modeuser@bm-dev-001:~$ ibdev2netdev -v cat: /sys/class/infiniband/mlx5_0/device/vpd: Permission denied 0000:45:00.0 mlx5_0 (MT4123 - ) fw 20.29.1016 port 1 (ACTIVE) ==> ibs18 (Down) cat: /sys/class/infiniband/mlx5_1/device/vpd: Permission denied 0000:0e:00.0 mlx5_1 (MT4123 - ) fw 20.29.1016 port 1 (ACTIVE) ==> ibs17 (Down) cat: /sys/class/infiniband/mlx5_2/device/vpd: Permission denied 0000:c5:00.0 mlx5_2 (MT4123 - ) fw 20.29.1016 port 1 (ACTIVE) ==> ibs20 (Down) cat: /sys/class/infiniband/mlx5_3/device/vpd: Permission denied 0000:85:00.0 mlx5_3 (MT4123 - ) fw 20.29.1016 port 1 (ACTIVE) ==> ibs19 (Down) user@bm-dev-001:~$ root@bm-dev-001:~# ibstat CA 'mlx5_0' CA type: MT4123 Number of ports: 1 Firmware version: 20.29.1016 Hardware version: 0 Node GUID: 0x88e9a4ffff5060ac System image GUID: 0x88e9a4ffff5060ac Port 1: State: Active Physical state: LinkUp Rate: 200 Base lid: 8 LMC: 0 SM lid: 1 Capability mask: 0x2651e848 Port GUID: 0x88e9a4ffff5060ac Link layer: InfiniBand CA 'mlx5_1' CA type: MT4123 Number of ports: 1 Firmware version: 20.29.1016 Hardware version: 0 Node GUID: 0x88e9a4ffff504080 System image GUID: 0x88e9a4ffff504080 Port 1: State: Active Physical state: LinkUp Rate: 200 Base lid: 5 LMC: 0 SM lid: 1 Capability mask: 0x2651e848 Port GUID: 0x88e9a4ffff504080 Link layer: InfiniBand CA 'mlx5_2' CA type: MT4123 Number of ports: 1 Firmware version: 20.29.1016 Hardware version: 0 Node GUID: 0x88e9a4ffff505038 System image GUID: 0x88e9a4ffff505038 Port 1: State: Active Physical state: LinkUp Rate: 200 Base lid: 2 LMC: 0 SM lid: 1 Capability mask: 0x2651e848 Port GUID: 0x88e9a4ffff505038 Link layer: InfiniBand CA 'mlx5_3' CA type: MT4123 Number of ports: 1 Firmware version: 20.29.1016 Hardware version: 0 Node GUID: 0x88e9a4ffff504094 System image GUID: 0x88e9a4ffff504094 Port 1: State: Active Physical state: LinkUp Rate: 200 Base lid: 7 LMC: 0 SM lid: 1 Capability mask: 0x2651e848 Port GUID: 0x88e9a4ffff504094 Link layer: InfiniBanduser@bm-dev-001:~$ ibdev2netdev -v cat: /sys/class/infiniband/mlx5_0/device/vpd: Permission denied 0000:45:00.0 mlx5_0 (MT4123 - ) fw 20.29.1016 port 1 (ACTIVE) ==> ibs18 (Down) cat: /sys/class/infiniband/mlx5_1/device/vpd: Permission denied 0000:0e:00.0 mlx5_1 (MT4123 - ) fw 20.29.1016 port 1 (ACTIVE) ==> ibs17 (Down) cat: /sys/class/infiniband/mlx5_2/device/vpd: Permission denied 0000:c5:00.0 mlx5_2 (MT4123 - ) fw 20.29.1016 port 1 (ACTIVE) ==> ibs20 (Down) cat: /sys/class/infiniband/mlx5_3/device/vpd: Permission denied 0000:85:00.0 mlx5_3 (MT4123 - ) fw 20.29.1016 port 1 (ACTIVE) ==> ibs19 (Down) user@bm-dev-001:~$ root@bm-dev-001:~# ibstat CA 'mlx5_0' CA type: MT4123 Number of ports: 1 Firmware version: 20.29.1016 Hardware version: 0 Node GUID: 0x88e9a4ffff5060ac System image GUID: 0x88e9a4ffff5060ac Port 1: State: Active Physical state: LinkUp Rate: 200 Base lid: 8 LMC: 0 SM lid: 1 Capability mask: 0x2651e848 Port GUID: 0x88e9a4ffff5060ac Link layer: InfiniBand CA 'mlx5_1' CA type: MT4123 Number of ports: 1 Firmware version: 20.29.1016 Hardware version: 0 Node GUID: 0x88e9a4ffff504080 System image GUID: 0x88e9a4ffff504080 Port 1: State: Active Physical state: LinkUp Rate: 200 Base lid: 5 LMC: 0 SM lid: 1 Capability mask: 0x2651e848 Port GUID: 0x88e9a4ffff504080 Link layer: InfiniBand CA 'mlx5_2' CA type: MT4123 Number of ports: 1 Firmware version: 20.29.1016 Hardware version: 0 Node GUID: 0x88e9a4ffff505038 System image GUID: 0x88e9a4ffff505038 Port 1: State: Active Physical state: LinkUp Rate: 200 Base lid: 2 LMC: 0 SM lid: 1 Capability mask: 0x2651e848 Port GUID: 0x88e9a4ffff505038 Link layer: InfiniBand CA 'mlx5_3' CA type: MT4123 Number of ports: 1 Firmware version: 20.29.1016 Hardware version: 0 Node GUID: 0x88e9a4ffff504094 System image GUID: 0x88e9a4ffff504094 Port 1: State: Active Physical state: LinkUp Rate: 200 Base lid: 7 LMC: 0 SM lid: 1 Capability mask: 0x2651e848 Port GUID: 0x88e9a4ffff504094 Link layer: InfiniBandCode block. InfiniBand(IB) HCA card HW status and Link check
IB bandwidth communication check
To check the IB bandwidth communication status (ib_send_bw) and inspect the IaaS HW level, follow these steps.
Check the name of the IB HCA interface.
Color modeuser@bm-dev-001:~$ ibdev2netdev mlx5_0 port 1 ==> ibs18 (Down) mlx5_1 port 1 ==> ibs17 (Down) mlx5_2 port 1 ==> ibs20 (Down) mlx5_3 port 1 ==> ibs19 (Down)user@bm-dev-001:~$ ibdev2netdev mlx5_0 port 1 ==> ibs18 (Down) mlx5_1 port 1 ==> ibs17 (Down) mlx5_2 port 1 ==> ibs20 (Down) mlx5_3 port 1 ==> ibs19 (Down)Code block. Check the name of IB HCA interface Check the HCA interface that can communicate with IB Switch#1.
Color modemlx5_0 port 1 ==> ibs18 (Down) mlx5_2 port 1 ==> ibs20 (Down)mlx5_0 port 1 ==> ibs18 (Down) mlx5_2 port 1 ==> ibs20 (Down)Code Block. HCA Interface Check Check the HCA interface that can communicate with IB Switch#2.
Color modemlx5_1 port 1 ==> ibs17 (Down) mlx5_3 port 1 ==> ibs19 (Down)mlx5_1 port 1 ==> ibs17 (Down) mlx5_3 port 1 ==> ibs19 (Down)Code Block. HCA Interface Check Use SERVER Side commands to check the communication status.
Client Sidecommand is entered secondarily for mutual communicationColor modeuser@bm-dev-001:~$ ib_send_bw -d mlx5_3 -i 1 –F ************************************ * Waiting for client to connect... * ************************************ --------------------------------------------------------------------------------------- Send BW Test Dual-port : OFF Device : mlx5_3 Number of qps : 1 Transport type : IB Connection type : RC Using SRQ : OFF PCIe relax order: ON ibv_wr* API : ON RX depth : 512 CQ Moderation : 1 Mtu : 4096[B] Link type : IB Max inline data : 0[B] rdma_cm QPs : OFF Data ex. method : Ethernet --------------------------------------------------------------------------------------- local address: LID 0x07 QPN 0x002e PSN 0xa86622 remote address: LID 0x0a QPN 0x002d PSN 0xfc58dd --------------------------------------------------------------------------------------- #bytes #iterations BW peak[MB/sec] BW average[MB/sec] MsgRate[Mpps] 65536 1000 0.00 19827.40 0.317238 ---------------------------------------------------------------------------------------user@bm-dev-001:~$ ib_send_bw -d mlx5_3 -i 1 –F ************************************ * Waiting for client to connect... * ************************************ --------------------------------------------------------------------------------------- Send BW Test Dual-port : OFF Device : mlx5_3 Number of qps : 1 Transport type : IB Connection type : RC Using SRQ : OFF PCIe relax order: ON ibv_wr* API : ON RX depth : 512 CQ Moderation : 1 Mtu : 4096[B] Link type : IB Max inline data : 0[B] rdma_cm QPs : OFF Data ex. method : Ethernet --------------------------------------------------------------------------------------- local address: LID 0x07 QPN 0x002e PSN 0xa86622 remote address: LID 0x0a QPN 0x002d PSN 0xfc58dd --------------------------------------------------------------------------------------- #bytes #iterations BW peak[MB/sec] BW average[MB/sec] MsgRate[Mpps] 65536 1000 0.00 19827.40 0.317238 ---------------------------------------------------------------------------------------Code Block. Communication Status Check
Use the
CLIENT Sidecommand to check the communication status.SERVER Sidecommand is entered first for mutual communicationColor moderoot@bm-dev-003:~# ib_send_bw -d mlx5_3 -i 1 -F <SERVER Side IP> --------------------------------------------------------------------------------------- Send BW Test Dual-port : OFF Device : mlx5_3 Number of qps : 1 Transport type : IB Connection type : RC Using SRQ : OFF PCIe relax order: ON ibv_wr* API : ON TX depth : 128 CQ Moderation : 1 Mtu : 4096[B] Link type : IB Max inline data : 0[B] rdma_cm QPs : OFF Data ex. method : Ethernet --------------------------------------------------------------------------------------- local address: LID 0x0a QPN 0x002a PSN 0x98a48e remote address: LID 0x07 QPN 0x002c PSN 0xe68304 --------------------------------------------------------------------------------------- #bytes #iterations BW peak[MB/sec] BW average[MB/sec] MsgRate[Mpps] 65536 1000 19008.49 19006.37 0.304102 ---------------------------------------------------------------------------------------root@bm-dev-003:~# ib_send_bw -d mlx5_3 -i 1 -F <SERVER Side IP> --------------------------------------------------------------------------------------- Send BW Test Dual-port : OFF Device : mlx5_3 Number of qps : 1 Transport type : IB Connection type : RC Using SRQ : OFF PCIe relax order: ON ibv_wr* API : ON TX depth : 128 CQ Moderation : 1 Mtu : 4096[B] Link type : IB Max inline data : 0[B] rdma_cm QPs : OFF Data ex. method : Ethernet --------------------------------------------------------------------------------------- local address: LID 0x0a QPN 0x002a PSN 0x98a48e remote address: LID 0x07 QPN 0x002c PSN 0xe68304 --------------------------------------------------------------------------------------- #bytes #iterations BW peak[MB/sec] BW average[MB/sec] MsgRate[Mpps] 65536 1000 19008.49 19006.37 0.304102 ---------------------------------------------------------------------------------------Code Block. Communication Status Check
Check IB Service Related Kernel Modules
Check the IB service-related kernel modules (lsmod) to inspect the IaaS HW level.
user@bm-dev-001:~$ lsmod | grep nv_peer_mem
nv_peer_mem 16384 0
ib_core 315392 9 rdma_cm,ib_ipoib,nv_peer_mem,iw_cm,ib_umad,rdma_ucm,ib_uverbs,mlx5_ib,ib_cm
nvidia 35315712 156 nvidia_uvm,nv_peer_mem,nvidia_modesetuser@bm-dev-001:~$ lsmod | grep nv_peer_mem
nv_peer_mem 16384 0
ib_core 315392 9 rdma_cm,ib_ipoib,nv_peer_mem,iw_cm,ib_umad,rdma_ucm,ib_uverbs,mlx5_ib,ib_cm
nvidia 35315712 156 nvidia_uvm,nv_peer_mem,nvidia_modesetuser@bm-dev-001:~$ service nv_peer_mem status
nv_peer_mem.service - LSB: Activates/Deactivates nv_peer_mem to \ start at boot time.
Loaded: loaded (/etc/init.d/nv_peer_mem; generated)
Active: active (exited) since Mon 2023-03-13 16:21:33 KST; 2 days ago
Docs: man:systemd-sysv-generator(8)
Process: 4913 ExecStart=/etc/init.d/nv_peer_mem start (code=exited, status=0/SUCCESS)user@bm-dev-001:~$ service nv_peer_mem status
nv_peer_mem.service - LSB: Activates/Deactivates nv_peer_mem to \ start at boot time.
Loaded: loaded (/etc/init.d/nv_peer_mem; generated)
Active: active (exited) since Mon 2023-03-13 16:21:33 KST; 2 days ago
Docs: man:systemd-sysv-generator(8)
Process: 4913 ExecStart=/etc/init.d/nv_peer_mem start (code=exited, status=0/SUCCESS)user@bm-dev-001:~$ lsmod | grep ib
libiscsi_tcp 32768 1 iscsi_tcp
libiscsi 57344 2 libiscsi_tcp,iscsi_tcp
scsi_transport_iscsi 110592 4 libiscsi_tcp,iscsi_tcp,libiscsi
ib_ipoib 131072 0
ib_cm 57344 2 rdma_cm,ib_ipoib
ib_umad 24576 8
mlx5_ib 380928 0
ib_uverbs 135168 18 rdma_ucm,mlx5_ib
ib_core 315392 9 rdma_cm,ib_ipoib,nv_peer_mem,iw_cm,ib_umad,rdma_ucm,ib_uverbs,mlx5_ib,ib_cm
libcrc32c 16384 2 btrfs,raid456
mlx5_core 1458176 1 mlx5_ib
auxiliary 16384 2 mlx5_ib,mlx5_core
mlx_compat 65536 12 rdma_cm,ib_ipoib,mlxdevm,iw_cm,auxiliary,ib_umad,ib_core,rdma_ucm,ib_uverbs,mlx5_ib,ib_cm,mlx5_coreuser@bm-dev-001:~$ lsmod | grep ib
libiscsi_tcp 32768 1 iscsi_tcp
libiscsi 57344 2 libiscsi_tcp,iscsi_tcp
scsi_transport_iscsi 110592 4 libiscsi_tcp,iscsi_tcp,libiscsi
ib_ipoib 131072 0
ib_cm 57344 2 rdma_cm,ib_ipoib
ib_umad 24576 8
mlx5_ib 380928 0
ib_uverbs 135168 18 rdma_ucm,mlx5_ib
ib_core 315392 9 rdma_cm,ib_ipoib,nv_peer_mem,iw_cm,ib_umad,rdma_ucm,ib_uverbs,mlx5_ib,ib_cm
libcrc32c 16384 2 btrfs,raid456
mlx5_core 1458176 1 mlx5_ib
auxiliary 16384 2 mlx5_ib,mlx5_core
mlx_compat 65536 12 rdma_cm,ib_ipoib,mlxdevm,iw_cm,auxiliary,ib_umad,ib_core,rdma_ucm,ib_uverbs,mlx5_ib,ib_cm,mlx5_coreStorage Physical Disk Resources and Multi-Path Check
Check the IaaS HW level by checking the storage physical disk resources and Multi-Path.
root@bm-dev-002:/tmp# fdisk –lroot@bm-dev-002:/tmp# fdisk –lroot@bm-dev-002:/tmp# multipath –llroot@bm-dev-002:/tmp# multipath –llMulti-node GPU Cluster new deployment after checking Service Network
Use the following command to check if the MII Status of Bonding and Slave Interface is up.
command
Color moderoot@mngc-001:~# cat /proc/net/bonding/bond-srv Ethernet Channel Bonding Driver: v5.15.0-25-genericroot@mngc-001:~# cat /proc/net/bonding/bond-srv Ethernet Channel Bonding Driver: v5.15.0-25-genericCode Block. Service Network Check Command confirmation result
Color modeBonding Mode: fault-tolerance (active-backup) Primary Slave: None Currently Active Slave: ens9f0 MII Status: up MII Polling Interval (ms): 100 Up Delay (ms): 0 Down Delay (ms): 0 Peer Notification Delay (ms): 0 Slave Interface: ens9f0 MII Status: up Speed: 100000 Mbps Duplex: full Link Failure Count: 0 Permanent HW addr: 30:3e:a7:02:35:70 Slave queue ID: 0 Slave Interface: ens11f0 MII Status: up Speed: 100000 Mbps Duplex: full Link Failure Count: 0 Permanent HW addr: 30:3e:a7:02:2f:e8 Slave queue ID: 0Bonding Mode: fault-tolerance (active-backup) Primary Slave: None Currently Active Slave: ens9f0 MII Status: up MII Polling Interval (ms): 100 Up Delay (ms): 0 Down Delay (ms): 0 Peer Notification Delay (ms): 0 Slave Interface: ens9f0 MII Status: up Speed: 100000 Mbps Duplex: full Link Failure Count: 0 Permanent HW addr: 30:3e:a7:02:35:70 Slave queue ID: 0 Slave Interface: ens11f0 MII Status: up Speed: 100000 Mbps Duplex: full Link Failure Count: 0 Permanent HW addr: 30:3e:a7:02:2f:e8 Slave queue ID: 0Code Block. Service Network Check Command Result
Multi-node GPU Cluster new deployment after checking Time Server and time synchronization
The OS image has the chrony daemon installed and set to synchronize with the SCP NTP server. Use the following command to check if there are any lines marked with ^* in the MS Name column.
command
Color moderoot@mngc-001:~# chronyc sources -Vroot@mngc-001:~# chronyc sources -VCode Block. chrony daemon installation command confirmation result
Color modeMS Name/IP address Stratum Poll Reach LastRx Last sample =============================================================================== ^+ 198.19.1.53 4 10 377 1040 -16us[ -37us] +/- 9982us ^* 198.19.1.54 4 10 377 312 -367us[ -388us] +/- 13msMS Name/IP address Stratum Poll Reach LastRx Last sample =============================================================================== ^+ 198.19.1.53 4 10 377 1040 -16us[ -37us] +/- 9982us ^* 198.19.1.54 4 10 377 312 -367us[ -388us] +/- 13msCode block. chrony daemon installation check result
GPU MIG/ECC Setting Initialization Check Guide
When applying for a multi-node GPU cluster product, the GPU MIG/ECC setting is initialized. However, to apply the exact setting value, please restart it once at the beginning, and then check and use it according to the inspection guide to see if the setting value is applied.
- MIG: Multi-Instance GPU
- ECC: Error Correction Code
MIG Setup Initialization
Refer to the following for how to check and initialize MIG settings.
Use the following command to check if the status value of MIG M is Disabled.
command
Color moderoot@bm-dev-001:~#nvidia-smiroot@bm-dev-001:~#nvidia-smiCode Block. MIG M. Initialize Settings confirmation result
Color mode+-----------------------------------------------------------------------------------------+ | NVIDIA-SMI 470.129.06 Driver version: 470.129.06 CUDA Version: 11.4 | |----------------------------------+-----------------------------+------------------------| | GPU Name Persistence-M | Bus-Id Disp.A | Volatile Uncorr. ECC | | Fan Temp Perf Pwr:Usage/Cap | Memory-Usage | GPU-Util Compute M. | | | | MIG M. | |==================================+=============================+========================| | 0 NVIDIA A100-SXM... Off | 00000000:03:00.0 Off | Off | | N/A 29C P0 57W / 400W | 0MiB / 81251MiB | 0% Default | | | | Disabled | +----------------------------------+-----------------------------+------------------------+ | 0 NVIDIA A100-SXM... Off | 00000000:0C:00.0 Off | Off | | N/A 30C P0 58W / 400W | 0MiB / 81251MiB | 18% Default | | | | Disabled | +-----------------------------------------------------------------------------------------++-----------------------------------------------------------------------------------------+ | NVIDIA-SMI 470.129.06 Driver version: 470.129.06 CUDA Version: 11.4 | |----------------------------------+-----------------------------+------------------------| | GPU Name Persistence-M | Bus-Id Disp.A | Volatile Uncorr. ECC | | Fan Temp Perf Pwr:Usage/Cap | Memory-Usage | GPU-Util Compute M. | | | | MIG M. | |==================================+=============================+========================| | 0 NVIDIA A100-SXM... Off | 00000000:03:00.0 Off | Off | | N/A 29C P0 57W / 400W | 0MiB / 81251MiB | 0% Default | | | | Disabled | +----------------------------------+-----------------------------+------------------------+ | 0 NVIDIA A100-SXM... Off | 00000000:0C:00.0 Off | Off | | N/A 30C P0 58W / 400W | 0MiB / 81251MiB | 18% Default | | | | Disabled | +-----------------------------------------------------------------------------------------+Code Block. MIG M. Initialization Setting Check Result If MIG M.’s status value is not Disabled, use the following command to initialize MIG.
Color moderoot@bm-dev-001:~# nvidia-smi -mig 0 root@bm-dev-001:~# nvidia-smi --gpu-resetroot@bm-dev-001:~# nvidia-smi -mig 0 root@bm-dev-001:~# nvidia-smi --gpu-resetCode Block. MIG M. Status Value Initialization
ECC Setting Initialization
Refer to the following for how to check and initialize the ECC settings.
Use the following command to check if the status value of Volatile Uncorr. ECC is Off.
command
Color moderoot@bm-dev-001:~#nvidia-smiroot@bm-dev-001:~#nvidia-smiCode Block. ECC Setting Command confirmation result
Color mode+-----------------------------------------------------------------------------------------+ | NVIDIA-SMI 470.129.06 Driver version: 470.129.06 CUDA Version: 11.4 | |----------------------------------+-----------------------------+------------------------| | GPU Name Persistence-M | Bus-Id Disp.A | Volatile Uncorr. ECC | | Fan Temp Perf Pwr:Usage/Cap | Memory-Usage | GPU-Util Compute M. | | | | MIG M. | |==================================+=============================+========================| | 0 NVIDIA A100-SXM... Off | 00000000:03:00.0 Off | Off | | N/A 29C P0 57W / 400W | 0MiB / 81251MiB | 0% Default | | | | Disabled | +----------------------------------+-----------------------------+------------------------+ | 0 NVIDIA A100-SXM... Off | 00000000:0C:00.0 Off | Off | | N/A 30C P0 61W / 400W | 0MiB / 81251MiB | 18% Default | | | | Disabled | +-----------------------------------------------------------------------------------------++-----------------------------------------------------------------------------------------+ | NVIDIA-SMI 470.129.06 Driver version: 470.129.06 CUDA Version: 11.4 | |----------------------------------+-----------------------------+------------------------| | GPU Name Persistence-M | Bus-Id Disp.A | Volatile Uncorr. ECC | | Fan Temp Perf Pwr:Usage/Cap | Memory-Usage | GPU-Util Compute M. | | | | MIG M. | |==================================+=============================+========================| | 0 NVIDIA A100-SXM... Off | 00000000:03:00.0 Off | Off | | N/A 29C P0 57W / 400W | 0MiB / 81251MiB | 0% Default | | | | Disabled | +----------------------------------+-----------------------------+------------------------+ | 0 NVIDIA A100-SXM... Off | 00000000:0C:00.0 Off | Off | | N/A 30C P0 61W / 400W | 0MiB / 81251MiB | 18% Default | | | | Disabled | +-----------------------------------------------------------------------------------------+Code Block. ECC Setting Check Result Volatile Uncorr. ECC’s status value is On*, please proceed with rebooting.
Volatile Uncorr. ECC status value is not On* or Off, use the following command to initialize ECC. After initialization, reboot and check if the status value is Off.
root@bm-dev-001:~# nvidia-smi --ecc-config=0root@bm-dev-001:~# nvidia-smi --ecc-config=05.3 - Release Note
Multi-node GPU Cluster
- You can cancel multiple resources at the same time from the GPU Node list.
- The nodes must use the same DataSet and Cluster Fabric.
- It has been linked with Cloud Monitoring.
- You can check major performance items in real-time in Cloud Monitoring.
- Multi-node GPU Cluster service has been launched.
- Provides a service that offers physical GPU servers without virtualization for large-scale high-performance AI computing.
6 - Cloud Functions
6.1 - Overview
Service Overview
Cloud Functions is a serverless computing-based FaaS (Function as a Service) that allows you to run applications in the form of functions without the need for server provisioning. The user does not need to manage servers or containers cumbersomely for scale adjustment, and can focus on writing and deploying code for application development.
Features
- Easy and convenient development environment: Developers can easily create Function resources connected to events in various environments using a Code Editor suitable for the chosen runtime, and can write and call code easily.
- Serverless Computing: You can use a serverless type of code execution service for development in the Samsung Cloud Platform environment. The resources required to call and execute function-type applications are allocated and managed by Samsung Cloud Platform according to the scale of execution.
- Efficient Cost Management: The called Function is charged only for the actual application runtime by aggregating usage (total number of calls, total call time). Functions with low usage are adjusted to Scale-to-zero state by Cloud Functions’ Scaler, preventing resource consumption, thus enabling efficient cost management.
Service Composition Diagram
Provided Features
Cloud Functions provides the following features.
- Code Writing Environment: Runtime-optimized Function creation, Code writing and editing
- Supported Runtime refers to Component > Runtime
- Function execution, environment management, monitoring: endpoint definition, Token management, access control setting, trigger setting, etc., definition and modification of operating environment/variables, calling/testing output for Deploy/Test, service deployment, progress status monitoring/logging
- Serverless Computing: all elements required for code writing and deployment are managed by Samsung Cloud Platform, with automatic scale adjustment according to deployment
- Sample Code Provided: Provides various sample codes through Blueprint, allowing for easy and quick start
Component
Runtime
Cloud Functions currently supports the following Runtime, and more supported Runtime will be added continuously.
| Runtime | Version |
|---|---|
| GO | 1.21, 1.23 |
| java | 17 |
| Node.js | 18, 20 |
| PHP | 8.1 |
| Python | 3.9, 3.10, 3.11 |
Regional Provision Status
Cloud Functions service is available in the following environments.
| Region | Availability |
|---|---|
| Korea West 1(kr-west1) | Provided |
| Korean East 1 (kr-east1) | Provided |
| South Korea (kr-south1) | Not provided |
| South Korea southern region 2(kr-south2) | Not provided |
| South Korea southern region 3(kr-south3) | Not provided |
Preceding Service
This is a list of services that can be configured as optional before creating the service. Please refer to the guide provided for each service and prepare in advance for more information.
| Service Category | Service | Detailed Description |
|---|---|---|
| Application Service | API Gateway | A service that easily manages and monitors APIs |
6.1.1 - ServiceWatch Metrics
Cloud Functions sends metrics to ServiceWatch. The metrics provided by default monitoring are data collected at a 1‑minute interval.
Basic Indicators
The following are the basic metrics for the Cloud Functions namespace.
| Performance Item | Detailed Description | Unit | Meaningful Statistics |
|---|
6.2 - How-to guides
The user can enter the required information for Cloud Functions through the Samsung Cloud Platform Console, select detailed options, and create the service.
Cloud Functions Create
Click the All Services > Compute > Cloud Functions menu. Navigate to the Service Home page of Cloud Functions.
Click the Create Cloud Functions button on the Service Home page. It navigates to the Create Cloud Functions page.
Create Cloud Functions page, enter the information required to create the service.
Category RequiredDetailed description Funtion name Required Enter the Funtion name to create - Start with a lowercase English letter and use lowercase English letters, numbers, and special characters (
-) to input within 3 ~ 64 characters
Runtime Required Select Runtime creation method - Create new: Create a new Runtime
- Start with Blueprint: Create using the Runtime source code provided by the service
Runtime & Version Required Select Runtime and Version - When Create New is selected
- Refer to the Runtime & Version list in Components > Runtime
- For Java runtime, UI code editing is not supported, but you can run by fetching a Jar file from Object Storage
- When Start with Blueprint is selected
- You can view source code examples by clicking the View Source Code button for the selected Runtime & Version
- Refer to the Blueprint Detailed Guide for more details on Blueprint settings
Table. Cloud Functions Service Information Input Items- Start with a lowercase English letter and use lowercase English letters, numbers, and special characters (
Summary Check the detailed information and estimated billing amount generated in the panel, and click the Complete button.
- When creation is complete, check the created resource on the Cloud Functions list page.
Cloud Functions Check Detailed Information
Cloud Functions Details page consists of Detail Information, Monitoring, Log, Code, Configuration, Trigger, Tag, Job History tabs.
To view detailed information about the Cloud Functions service, follow these steps.
- All Services > Compute > Cloud Functions Click the menu. Go to the Service Home page of Cloud Functions.
- Click the Function menu on the Service Home page. Move to the Function List page.
- Click the resource to view detailed information on the Function list page. Go to the Function detail page.
- Function Details The page displays status information and additional feature information, and consists of the Details, Monitoring, Log, Code, Configuration, Trigger, Tag, Task History tabs.
Category Detailed description Cloud Functions status Cloud Functions status information - Ready: green icon, state where normal function calls are possible
- Not Ready: gray icon, state where normal function calls are not possible
- Deploying: yellow icon, state where function is being created or changed, triggered by the following actions
- function creation and modification
- modify code with editor in the Code tab
- inspect jar file in the Code tab
- add and modify in the Trigger tab
- modify in the Configuration tab
- Running: blue icon, state where normal function calls are possible and cold start prevention policy is applied
Service cancellation Button to cancel the service Table. Cloud Functions status information and additional features
Detailed Information
Function list page, you can view detailed information of the selected resource and, if necessary, edit the information.
| Category | Detailed description |
|---|---|
| service | service name |
| Resource Type | Resource Type |
| SRN | Unique resource ID in Samsung Cloud Platform |
| Resource Name | Resource Name
|
| Resource ID | Unique resource ID of the service |
| Creator | User who created the service |
| Creation time | Date/time the service was created |
| Editor | User who modified the service |
| Modification Date and Time | Date and time the service was modified |
| Function name | Name of Cloud Function |
| Runtime | Runtime types and versions |
| LLM Endpoint | Click User Guide to view LLM Endpoint information and usage instructions |
Monitoring
You can view the Cloud Functions usage information of the selected resource on the Function List page.
| Category | Detailed description |
|---|---|
| Number of calls | Average number of times the function was called during the unit time (instances) |
| Execution Time | Average execution time (seconds) of the function during the unit time |
| Memory usage | Average memory usage (kb) used during the execution of the function per unit time |
| Current task count | If the function is called multiple times simultaneously, the average number of tasks generated per unit time for concurrent processing (count) |
| Successful call count | Average number of times (cases) the runtime code operated normally and delivered a response code per unit time when the function is called |
| Failed call count | Average number of calls with errors per unit time when the function is invoked
|
Log
You can view the Cloud Functions logs of the selected resource on the Function list page.
| Category | Detailed description |
|---|---|
| Unit period | Select the period to view Cloud Functions log information
|
| Log Message | Displayed in order from the most recent occurring function. |
Code
Function list page allows you to view and edit the Cloud Functions code of the selected resource.
The way to view and edit source code changes depending on the runtime used.
- Inline Editor: Node.js, Python, PHP, Go
- Run compressed file (.jar/.zip): Java
| Category | Detailed description |
|---|---|
| Source Code | Inline Editor Mode |
| Code Information | Display code information |
| Edit | Edit After clicking the button, you can edit the code in the inline editor |
| Category | Detailed description |
|---|---|
| Source code | Compressed file (.jar/.zip) execution method |
| Code Information | Compressed File Information Display
|
| Edit | Jar file can be changed
|
- In the case of Java Runtime, it does not provide UI code editing functionality, and you must select a compressed file (.jar/.zip) from the bucket of the Object Storage service.
- In case of users whose Object Storage service authentication key has not been generated, Import from Object Storage cannot be executed, so you must generate the authentication key in advance.
- For detailed information on creating authentication keys, refer to Create Authentication Key.
- Cloud Functions service’s Object Storage bucket’s access control must be changed to allow state.
- For detailed information about Object Storage access control, please refer to Allow Cloud Functions Service Access.
Composition
Function list page allows you to view the Cloud Functions configuration of the selected resource.
| Category | Detailed description |
|---|---|
| General Configuration | Cloud Function memory and timeout settings
|
| Environment Variable | Set runtime environment variables
|
| Function URL | Issue an HTTPS URL address that can access the function
|
| Private connection configuration | Can be used in conjunction with PrivateLink Service
|
- General Configuration’s memory allocation proportionally determines the number of CPU cores that are automatically assigned.
- General configuration’s minimum execution count of 1 or more prevents Cold Start, but costs are incurred continuously.
Trigger
Function List page allows you to view and configure trigger information of the selected resource. If you set a trigger, the Function can be automatically executed when an event occurs.
| Category | Detailed description |
|---|---|
| Cronjob | Use Cronjob as a trigger
|
| API Gateway | Use API Gateway as a trigger
|
- Deploying If in this state, cannot be edited.
- About trigger settings, please refer to Trigger Setup.
Tag
In the Tag tab, you can view the resource’s tag information, and add, modify, or delete it.
| Category | Detailed description |
|---|---|
| Tag List | Tag List
|
Work History
You can check the work history of resources on the Work History page.
| Category | Detailed description |
|---|---|
| Work History List | Resource Change History
|
Java Runtime Code Change
If you are using Java Runtime, you cannot modify the code directly, so you need to select and change the compressed file (.jar/.zip) in the bucket of the Object Storage service.
Follow the steps below to change the compressed file.
To cancel the Cloud Functions service, follow the steps below.
- Click the All Services > Compute > Cloud Functions menu. Go to the Service Home page of Cloud Functions.
- Service Home on the page click the Function menu. Navigate to the Function list page.
- Click the resource to change the compressed file within the code on the Function List page. Navigate to the Function Details page.
- Click the Edit button on the Code tab of the Function Details page. It moves to the Function Code Edit page.
- Import from Object Storage Click the button. Import from Object Storage popup opens.
| Category | Detailed description |
|---|---|
| Java Runtime | Java Runtime information |
| Handler Information | Handler Information
|
| Compressed file (.jar/.zip) | Set the compressed file to modify
|
- Object Storage URL After entering the URL information of the Object Storage from which to retrieve the compressed file, click the Confirm button. A notification popup will open.
- The URL information can be found in the Folder List tab of the detailed page of the Object Storage to be retrieved, under the File Information > Private URL item.
- Click the Confirm button. The name of the imported archive file is displayed in the Function code edit page’s Archive file name (.jar/.zip).
- Click the Save button.
- In case of a user whose authentication key has not been generated, Import from Object Storage cannot be executed.
- If the URL does not exist or the compressed file corresponds to the following, it cannot be changed.
- When using an unsupported extension
- If there is a harmful file in the compressed file
- If it exceeds the supported size
Cloud Functions Cancel
To cancel the Cloud Functions service, follow the steps below.
- All Services > Compute > Cloud Functions Click the menu. Navigate to the Service Home page of Cloud Functions.
- Click the Function menu on the Service Home page. Go to the Function List page.
- Function list page, click the resource to be terminated and click the Cancel Service button.
- When the termination is completed, check whether the resource has been terminated on the Function List page.
6.2.1 - Set Trigger
Set up trigger
- By default, all triggers can be added in Cloud Functions.
- If triggered from a specific product, it must be passed to Cloud Functions.
Cronjob Trigger Setup
To set up a Cronjob trigger, follow these steps.
- All Services > Compute > Cloud Functions Click the menu. Navigate to the Service Home page of Cloud Functions.
- Click the Function menu on the Service Home page. It moves to the Function List page.
- On the Function List page, click the resource to set a trigger. You will be taken to the Function Details page.
- After clicking the Trigger tab, click the Add Trigger button. Set it. The Add Trigger popup opens.
- Add Trigger In the popup window, select Trigger Type Cronjob. The required information input area appears at the bottom.
Category Detailed description Cronjob Settings Set the trigger’s repeat frequency - Can be set in minutes, hours, days, months, and weekdays
Timezone setting Set the trigger’s reference time zone Table. Cronjob Trigger Required Information Items - After entering the required information, click the Confirm button.
- When the pop-up window notifying addition opens, click the Confirm button.
API Gateway Trigger Setup
To set up an API Gateway trigger, follow these steps.
- All Services > Compute > Cloud Functions Click the menu. Go to the Service Home page of Cloud Functions.
- Click the Function menu on the Service Home page. Go to the Function List page.
- Function List on the page, click the resource to set the trigger. Function Details go to the page.
- After clicking the Trigger tab, click the Add Trigger button. Set it. The Add Trigger popup window opens.
- Add Trigger In the popup window, select Trigger Type API Gateway. A required information input area appears at the bottom.
Category Detailed description API name API selection - You can select an existing API or create a new one
Stage Select deployment target - You can select an existing stage or create a new one
Table. API Gateway Trigger Required Information Items - After entering the required information, click the Confirm button.
- When the popup notifying addition opens, click the Confirm button.
Setting up Multi Trigger
You can connect multiple triggers to a single function and use them.
Edit Trigger
To modify the added trigger, follow the steps below.
- All Services > Compute > Cloud Functions Click the menu. Go to the Service Home page of Cloud Functions.
- Click the Function menu on the Service Home page. Navigate to the Function List page.
- On the Function List page, click the resource to edit the trigger. It moves to the Function Details page.
- After clicking the Trigger tab, click the Edit button of the trigger whose settings you want to modify in the trigger list. The Edit Trigger popup window opens.
- Trigger Edit After modifying the setting values in the popup window, click the Confirm button.
- The setting values vary depending on the type of trigger (see Cronjob Trigger, API Gateway Trigger).
- When the popup notifying the edit opens, click Confirm.
Delete Trigger
To delete the trigger, follow the steps below.
- All Services > Compute > Cloud Functions Click the menu. Navigate to the Service Home page of Cloud Functions.
- Click the Function menu on the Service Home page. Navigate to the Function List page.
- Function List page, click the resource to set the trigger. Function Details page will be navigated.
- In the Trigger tab’s trigger list, after selecting the trigger to delete, click the Delete button.
- Click the Confirm button when the popup notifying trigger deletion opens.
6.2.2 - AIOS Connect
AIOS Linking
You can use LLM by linking Cloud Functions with AIOS.
AIOS LLM Private Endpoint
The URL of the AIOS LLM private endpoint is as follows.
- Samsung Cloud Platform for Samsung: https://aios.private.kr-west1.s.samsungsdscloud.com
- Samsung Cloud Platform for Enterprize: https://aios.private.kr-west1.e.samsungsdscloud.com
Refer to the following for detailed information on the availability and provision model of AIOS services by region.
- Availability of AIOS services by region: AIOS Service - Regional Availability Reference
- AIOS service provided LLM model: AIOS Service - LLM Provided Model reference
Blueprint Change Source Code
To integrate Cloud Functions with AIOS, you need to change the URL address in the Blueprint to match the LLM Endpoint used in each region. To change the Blueprint source code, follow the steps below.
- All Services > Compute > Cloud Functions Click the menu. Go to the Service Home page of Cloud Functions.
- Service Home on the page click the Cloud Functions menu. Navigate to the Function list page.
- On the Function List page, click the resource to be called via URL. You will be taken to the Function Detail page.
- After clicking the Code tab, click the Edit button. Navigate to the Function Code Edit page.
- After modifying the Blueprint using Python, Node.js, Go Runtime source code, click the Save button.
Python source code
Color modeimport json import requests def handle_request(params): # User writing area (Function details) url = "{AIOS LLM private endpoint}/{API}" # Destination URL data = { "model": "openai/gpt-oss-120b" , "prompt" : "Write a haiku about recursion in programming." , "temperature": 0 , "max_tokens": 100 , "stream": False } try: response = requests.post(url, json=data, verify=True) return { 'statusCode': response.status_code, 'body': json.dumps(response.text) } except requests.exceptions.RequestException as e: return str(e)import json import requests def handle_request(params): # User writing area (Function details) url = "{AIOS LLM private endpoint}/{API}" # Destination URL data = { "model": "openai/gpt-oss-120b" , "prompt" : "Write a haiku about recursion in programming." , "temperature": 0 , "max_tokens": 100 , "stream": False } try: response = requests.post(url, json=data, verify=True) return { 'statusCode': response.status_code, 'body': json.dumps(response.text) } except requests.exceptions.RequestException as e: return str(e)Python source code Node.js source code
Color modeconst request = require('request'); /** * @description User writing area (Function details) */ exports.handleRequest = async function (params) { return await sendRequest(params); }; async function sendRequest(req) { return new Promise((resolve, reject) => { url = "{AIOS LLM private endpoint}/{API}" data = { model: 'openai/gpt-oss-120b' , prompt : 'Write a haiku about recursion in programming.' , temperature: 0 , max_tokens: 100 , stream: false } const options = { uri: url, method:'POST', body: data, json: true, strictSSL: false, rejectUnauthorized: false } request(options, (error, response, body) => { if (error) { reject(error); } else { resolve({ statusCode: response.statusCode, body: JSON.stringify(body) }); } }); }); }const request = require('request'); /** * @description User writing area (Function details) */ exports.handleRequest = async function (params) { return await sendRequest(params); }; async function sendRequest(req) { return new Promise((resolve, reject) => { url = "{AIOS LLM private endpoint}/{API}" data = { model: 'openai/gpt-oss-120b' , prompt : 'Write a haiku about recursion in programming.' , temperature: 0 , max_tokens: 100 , stream: false } const options = { uri: url, method:'POST', body: data, json: true, strictSSL: false, rejectUnauthorized: false } request(options, (error, response, body) => { if (error) { reject(error); } else { resolve({ statusCode: response.statusCode, body: JSON.stringify(body) }); } }); }); }Node.js Source Code GO source code
Color modepackage gofunction import ( "bytes" "net/http" "encoding/json" "io/ioutil" ) type PostData struct { Model string `json:"model"` Prompt string `json:"prompt"` Temperature int `json:"temperature"` MaxTokens int `json:"max_tokens"` Stream bool `json:"stream"` } func HandleRequest(r *http.Request)(string, error) { url := "{AIOS LLM private endpoint}/{API}" data := PostData { Model: "openai/gpt-oss-120b", Prompt: "Write a haiku about recursion in programming.", Temperature: 0, MaxTokens: 100, Stream: false, } jsonData, err := json.Marshal(data) if err != nil { panic(err) } req, err := http.NewRequest("POST", url, bytes.NewBuffer(jsonData)) if err != nil { panic(err) } req.Header.Set("Content-Type", "application/json") client := &http.Client{} resp, err := client.Do(req) if err != nil { panic(err) } defer resp.Body.Close() // Read response body body, err := ioutil.ReadAll(resp.Body) if err != nil { panic(err) } return string(body), nil "}package gofunction import ( "bytes" "net/http" "encoding/json" "io/ioutil" ) type PostData struct { Model string `json:"model"` Prompt string `json:"prompt"` Temperature int `json:"temperature"` MaxTokens int `json:"max_tokens"` Stream bool `json:"stream"` } func HandleRequest(r *http.Request)(string, error) { url := "{AIOS LLM private endpoint}/{API}" data := PostData { Model: "openai/gpt-oss-120b", Prompt: "Write a haiku about recursion in programming.", Temperature: 0, MaxTokens: 100, Stream: false, } jsonData, err := json.Marshal(data) if err != nil { panic(err) } req, err := http.NewRequest("POST", url, bytes.NewBuffer(jsonData)) if err != nil { panic(err) } req.Header.Set("Content-Type", "application/json") client := &http.Client{} resp, err := client.Do(req) if err != nil { panic(err) } defer resp.Body.Close() // Read response body body, err := ioutil.ReadAll(resp.Body) if err != nil { panic(err) } return string(body), nil "}GO source code
6.2.3 - Blueprint Detailed Guide
Blueprint Overview
When creating Cloud Functions, you can set the Blueprint to utilize the Runtime source code provided by Cloud Functions. Refer to the following for Blueprint items provided by Cloud Functions.
| Category | Detailed Description | Remarks |
|---|---|---|
| Hello World | When the function is called, it responds with Hello Serverless World! | |
| Execution after timeout | Outputs code that should be executed after the function call time has exceeded, but does not execute. | PHP, Python not supported |
| HTTP request body | Parses the request body. | PHP not supported |
| Send HTTP requests | Send HTTP requests from Cloud functions. | PHP not supported |
| Print logs | Outputs the user’s Samsung Cloud Platform Console Request to the log. | PHP not supported |
| Throw a custom error | Enter the error logic directly to handle the error. | |
| Using Environment Variable | Configure environment variables within the Cloud function and execute. |
Hello World
Hello World explains how to set up receiving responses and an example of function call (using function URL).
Hello World Setting
To set Hello World, follow the steps below.
All Services > Compute > Cloud Functions Click the menu. Navigate to the Service Home page of Cloud Functions.
Click the Function menu on the Service Home page. Go to the Function list page.
Function list page, click the resource to be called via URL. Navigate to the Function detail page.
Click the Configuration tab, then click the Edit button of the Function URL item. The Edit Function URL popup opens.
Edit Function URL in the popup window, after setting Activation Status to Enabled, click the Confirm button.
Category Detailed description Activation status Set whether to use function URL Authentication Type Select whether to use IAM authentication when requesting the function URL Access Control Can manage by adding accessible IPs - After setting to Use, public access IP can be entered and added
Table. Required input fields when adding a triggerAfter moving to the Code tab, click the Edit button. You will be taken to the Function Code Edit page.
After adding the processing logic for success and failure cases, click the Save button.
- Node.js source codeColor mode
exports.handleRequest = async function (params) { /** * @description User writing area (Function details) */ const response = { statusCode: 200, body: JSON.stringify('Hello Serverless World!'), };" return response; };exports.handleRequest = async function (params) { /** * @description User writing area (Function details) */ const response = { statusCode: 200, body: JSON.stringify('Hello Serverless World!'), };" return response; };Hello World - Node.js source code - Python source codeColor mode
import json def handle_request(params): # User writing area (Function details) return { 'statusCode': 200, 'body': json.dumps('Hello Serverless World!') }import json def handle_request(params): # User writing area (Function details) return { 'statusCode': 200, 'body': json.dumps('Hello Serverless World!') }Hello World - Python source code - PHP source codeColor mode
<?php function handle_request() { # User writing area (Function details) $res = array( 'statusCode' => 200, 'body' => 'Hello Serverless World!', ); return $res; } ?><?php function handle_request() { # User writing area (Function details) $res = array( 'statusCode' => 200, 'body' => 'Hello Serverless World!', ); return $res; } ?>Hello World - PHP source code
- Node.js source code
Check function call
After calling the function URL in the Configuration tab of the Function Details page, verify the response.
Hello Serverless World!
# Execution after timeout
Explain the setting for execution after timeout and an example of function call (using function URL).
## Execution after timeout Setting
To set Execution after timeout, follow the steps below.
1. **All Services > Compute > Cloud Functions** Click the menu. Go to the **Service Home** page of Cloud Functions.
2. Click the **Function** menu on the **Service Home** page. Navigate to the **Function List** page.
3. **Function List** page, click the resource to set the trigger. **Function Details** page will be opened.
4. After clicking the **Trigger** tab, click the **Add Trigger** button. The **Add Trigger** popup opens.
5. **Add Trigger** In the popup window, after selecting the **Trigger Type** item, enter the required information displayed at the bottom and click the **Confirm** button.
* Required information varies depending on the trigger type.
Trigger Type
Input Item
API Gateway
- API name: You can select an existing API or create a new one
- Stage: You can select an existing stage or create a new one
Cronjob
- Refer to the example and enter the trigger’s repeat frequency (minute, hour, day, month, day of week)
- Timezone setting: select the reference time zone to apply
Table. Required input items when adding a trigger
6. **Code** after moving to the tab, click the **Edit** button. You will be taken to the **Function Code Edit** page.
7. After adding the processing logic for success and failure cases, click the **Save** button.
* Node.js source code
Color mode
exports.handleRequest = async function (params) {
/**
* @description User writing area (Function details)
*/
console.log("Hello world 3");
await delay(3000);
const response = {
statusCode: 200,
body: JSON.stringify('Hello Serverless World!'),
};;
return response;
};
const delay = (ms) => {
return new Promise(resolve=>{
setTimeout(resolve,ms)
})
}
exports.handleRequest = async function (params) {
/**
* @description User writing area (Function details)
*/
console.log("Hello world 3");
await delay(3000);
const response = {
statusCode: 200,
body: JSON.stringify('Hello Serverless World!'),
};;
return response;
};
const delay = (ms) => {
return new Promise(resolve=>{
setTimeout(resolve,ms)
})
}
Execution after timeout - Node.js source code ## Check function call
**Function Details** page's **configuration** tab call the **function URL** and after a certain amount of time, check the response.Hello Serverless World!
# HTTP request body
Explains the Request Body parsing settings and function call example (using function URL).
## Setting HTTP request body
To set the HTTP request body, follow these steps.
1. Click the **All Services > Compute > Cloud Functions** menu. Go to the **Service Home** page of Cloud Functions.
2. Click the **Function** menu on the **Service Home** page. Navigate to the **Function** list page.
3. **Function List** page, click the resource to set the trigger. **Function Details** page will be opened.
4. After clicking the **Trigger** tab, click the **Add Trigger** button. The **Add Trigger** popup opens.
5. **Add Trigger** in the popup window, after selecting the **Trigger Type** item, enter the required information displayed at the bottom and click the **Confirm** button.
* Required information varies depending on the trigger type.
Trigger Type
Input Item
API Gateway
- API name: You can select an existing API or create a new one
- Stage: You can select an existing stage or create a new one
Cronjob
- Refer to the example and enter the trigger’s repeat frequency (minute, hour, day, month, day of week)
- Timezone setting: select the time zone to apply
Table. Required input items when adding a trigger
6. After moving to the **Code** tab, click the **Edit** button. You will be taken to the **Function Code Edit** page.
7. After adding processing logic for success and failure cases, click the **Save** button.
* Node.js source code
Color mode
exports.handleRequest = async function (params) {
/**
* @description User writing area (Function details)
*/
const response = {
statusCode: 200,
body: JSON.stringify(params.body),
};;
return response;
};
exports.handleRequest = async function (params) {
/**
* @description User writing area (Function details)
*/
const response = {
statusCode: 200,
body: JSON.stringify(params.body),
};;
return response;
};
Execution after timeout - Node.js source code * Python source codeColor modeimport json
def handle_request(params):
# User writing area (Function details)
return {
'statusCode': 200,
'body': json.dumps(params.json)
}
import json
def handle_request(params):
# User writing area (Function details)
return {
'statusCode': 200,
'body': json.dumps(params.json)
}
Execution after timeout - Python source code ## Check function call
**Function Details** page's **Configuration** tab after calling the **function URL**, check the Body data, request Body value, and response Body value.
* request Body valueColor mode{
"testKey" :"cloud-001",
"testNames": ["
{
"name": "Son"
},
{
"name": "Kim"
}
],
"testCode":"test"
}
{
"testKey" :"cloud-001",
"testNames": ["
{
"name": "Son"
},
{
"name": "Kim"
}
],
"testCode":"test"
}
Request Body Value * Response Body valueColor mode{
"testKey" :"cloud-001",
"testNames": [
{
"name": "Son"
},
{
"name": "Kim"
}
],
"testCode":"test"
}
{
"testKey" :"cloud-001",
"testNames": [
{
"name": "Son"
},
{
"name": "Kim"
}
],
"testCode":"test"
}
Response Body Value # Send HTTP requests
Explains the HTTP request settings and function call example (using function URL).
## Send HTTP requests Setup
To configure Send HTTP requests, follow the steps below.
1. **All Services > Compute > Cloud Functions** Click the menu. Go to the **Service Home** page of Cloud Functions.
2. Click the **Function** menu on the **Service Home** page. Go to the **Function List** page.
3. Click the resource to set the trigger on the **Function List** page. It navigates to the **Function Details** page.
4. After clicking the **Trigger** tab, click the **Add Trigger** button. The **Add Trigger** popup opens.
5. **Add Trigger** In the popup window, after selecting the **Trigger Type** item, enter the required information displayed at the bottom and click the **Confirm** button.
* Required information varies depending on the type of trigger.Trigger Type Input Item API Gateway - API name: You can select an existing API or create a new one
- Stage: You can select an existing stage or create a new one
Cronjob - Refer to the example and enter the trigger’s repeat frequency (minute, hour, day, month, day of week)
- Timezone setting: select the reference time zone to apply
Table. Required input items when adding a trigger
6. After moving to the **Code** tab, click the **Edit** button. You will be taken to the **Function Code Edit** page.
7. After adding the processing logic for success and failure cases, click the **Save** button.
* Node.js source codeColor modeconst request = require('request');
/**
* @description User writing area (Function details)
*/
exports.handleRequest = async function (params) {
return await sendRequest(params);
};
async function sendRequest(req) {
return new Promise((resolve, reject) => {
// Port 80 and Port 443 are available
url = "https://example.com"; // Destination URL
const options = {
uri: url,
method:'GET',
json: true,
strictSSL: false,
rejectUnauthorized: false
}
request(options, (error, response, body) => {
if (error) {
reject(error);
} else {
resolve({
statusCode: response.statusCode,
body: JSON.stringify(body)
});
}
});
});
}
const request = require('request');
/**
* @description User writing area (Function details)
*/
exports.handleRequest = async function (params) {
return await sendRequest(params);
};
async function sendRequest(req) {
return new Promise((resolve, reject) => {
// Port 80 and Port 443 are available
url = "https://example.com"; // Destination URL
const options = {
uri: url,
method:'GET',
json: true,
strictSSL: false,
rejectUnauthorized: false
}
request(options, (error, response, body) => {
if (error) {
reject(error);
} else {
resolve({
statusCode: response.statusCode,
body: JSON.stringify(body)
});
}
});
});
}
Send HTTP requests - Node.js source code * Python source codeColor modeimport json
import requests
def handle_request(params):
# User writing area (Function details)
# Port 80 and Port 443 are available
url = "https://example.com" # Destination URL
try:
response = requests.get(url, verify=True)
return {
'statusCode': response.status_code,
'body': json.dumps(response.text)
}
except requests.exceptions.RequestException as e:
return str(e)
import json
import requests
def handle_request(params):
# User writing area (Function details)
# Port 80 and Port 443 are available
url = "https://example.com" # Destination URL
try:
response = requests.get(url, verify=True)
return {
'statusCode': response.status_code,
'body': json.dumps(response.text)
}
except requests.exceptions.RequestException as e:
return str(e)
Send HTTP requests - Python source code ## Check function call
**Function Details** page's **Configuration** tab, after calling the **function URL**, check the response.Color mode<!doctype html>
<html>
<head>
<title>Example Domain</title>
<meta charset="utf-8" />
<meta http-equiv="Content-type" content="text/html; charset=utf-8" />
<meta name="viewport" content="width=device-width, initial-scale=1" />
<style type="text/css">
body {
background-color: #f0f0f2;
margin: 0;
padding: 0;
font-family: -apple-system, system-ui, BlinkMacSystemFont, "Segoe UI", "Open Sans", "Helvetica Neue", Helvetica, Arial, sans-serif;
}
div {
width: 600px;
margin: 5em auto;
padding: 2em;
background-color: #fdfdff;
border-radius: 0.5em;
box-shadow: 2px 3px 7px 2px rgba(0,0,0,0.02);
}
a:link, a:visited {
color: #38488f;
text-decoration: none;
}
@media (max-width: 700px) {
div {
margin: 0 auto;
width: auto;
}
}
</style>
</head>
<body>
<div>
<h1>Example Domain</h1>
<p>This domain is for use in illustrative examples in documents. You may use this
domain in literature without prior coordination or asking for permission.</p>
<p><a href="https://www.iana.org/domains/example">More information...</a></p>
</div>
</body>
</html>
<!doctype html>
<html>
<head>
<title>Example Domain</title>
<meta charset="utf-8" />
<meta http-equiv="Content-type" content="text/html; charset=utf-8" />
<meta name="viewport" content="width=device-width, initial-scale=1" />
<style type="text/css">
body {
background-color: #f0f0f2;
margin: 0;
padding: 0;
font-family: -apple-system, system-ui, BlinkMacSystemFont, "Segoe UI", "Open Sans", "Helvetica Neue", Helvetica, Arial, sans-serif;
}
div {
width: 600px;
margin: 5em auto;
padding: 2em;
background-color: #fdfdff;
border-radius: 0.5em;
box-shadow: 2px 3px 7px 2px rgba(0,0,0,0.02);
}
a:link, a:visited {
color: #38488f;
text-decoration: none;
}
@media (max-width: 700px) {
div {
margin: 0 auto;
width: auto;
}
}
</style>
</head>
<body>
<div>
<h1>Example Domain</h1>
<p>This domain is for use in illustrative examples in documents. You may use this
domain in literature without prior coordination or asking for permission.</p>
<p><a href="https://www.iana.org/domains/example">More information...</a></p>
</div>
</body>
</html>
Check function call response # Print logs
Explains the log output settings and function call example (using function URL).
## Print logs Setup
Print logs To set up receiving responses, follow the steps below.
1. Click the **All Services > Compute > Cloud Functions** menu. Navigate to the **Service Home** page of Cloud Functions.
2. Click the **Function** menu on the **Service Home** page. Move to the **Function List** page.
3. **Function List** page, click the resource to set the trigger. **Function Details** page will be opened.
4. Click the **Trigger** tab, then click the **Add Trigger** button. The **Add Trigger** popup opens.
5. **Add Trigger** In the popup window, after selecting the **Trigger Type** item, enter the required information displayed at the bottom and click the **Confirm** button.
* Required information varies depending on the trigger type.Trigger Type Input Item API Gateway - API name: You can select an existing API or create a new one
- Stage: You can select an existing stage or create a new one
Cronjob - Refer to the example and enter the trigger’s repeat frequency (minute, hour, day, month, day of week)
- Timezone setting: select the reference time zone to apply
Table. Required input items when adding a trigger
6. After moving to the **Code** tab, click the **Edit** button. You will be taken to the **Function Code Edit** page.
7. After adding the processing logic for success and failure cases, click the **Save** button.
* Node.js source codeColor modeconst winston = require('winston');
// Log module setting
const logger = winston.createLogger({
format: winston.format.combine(
winston.format.timestamp(),
winston.format.printf(info => info.timestamp + ' ' + info.level + ': ' + info.message)
),
transports: [
new winston.transports.Console()
]
});
exports.handleRequest = async function (params) {
/**
* @description User writing area (Function details)
*/
const response = {
statusCode: 200,
body: JSON.stringify(params.body),
};"
logger.info(JSON.stringify(response, null, 2));
return response;
};
const winston = require('winston');
// Log module setting
const logger = winston.createLogger({
format: winston.format.combine(
winston.format.timestamp(),
winston.format.printf(info => info.timestamp + ' ' + info.level + ': ' + info.message)
),
transports: [
new winston.transports.Console()
]
});
exports.handleRequest = async function (params) {
/**
* @description User writing area (Function details)
*/
const response = {
statusCode: 200,
body: JSON.stringify(params.body),
};"
logger.info(JSON.stringify(response, null, 2));
return response;
};
Print logs - Node.js source code * Python source codeColor modeimport json
import logging
# Log module setting
logging.basicConfig(level=logging.INFO)
def handle_request(params):
# User writing area (Function details)
response = {
'statusCode': 200,
'body': json.dumps(params.json)
}
logging.info(response)
return response
import json
import logging
# Log module setting
logging.basicConfig(level=logging.INFO)
def handle_request(params):
# User writing area (Function details)
response = {
'statusCode': 200,
'body': json.dumps(params.json)
}
logging.info(response)
return response
Print logs - Python source code ## Check function call
**Function Details** page's **Configuration** tab after calling the **function URL**, check the log in the **Log** tab.Color mode[2023-09-07] 12:06:23] "host": "scf-xxxxxxxxxxxxxxxxxxxxx",
[2023-09-07] 12:06:23] "ce-id": "xxxxxxxxxxxxxxxxxxxxx",
[2023-09-07] 12:06:23] "ce-source": "xxxxxxxxxxxxxxxxxxxxx",
[2023-09-07] 12:06:23] "host": "scf-xxxxxxxxxxxxxxxxxxxxx",
[2023-09-07] 12:06:23] "ce-id": "xxxxxxxxxxxxxxxxxxxxx",
[2023-09-07] 12:06:23] "ce-source": "xxxxxxxxxxxxxxxxxxxxx",
Check function call response # Throw a custom error
Custom error occurrence (Throw a custom error) setting and function call example (function URL usage) is explained.
## Throw a custom error Setting
To set Throw a custom error, follow the steps below.
1. **All Services > Compute > Cloud Functions** Click the menu. Go to the **Service Home** page of Cloud Functions.
2. Click the **Function** menu on the **Service Home** page. Move to the **Function List** page.
3. **Function List** page, click the resource to set the trigger. **Function Details** page will be navigated.
4. **Trigger** tab after clicking, click the **Add Trigger** button. The **Add Trigger** popup window opens.
5. **Add Trigger** In the popup window, after selecting the **Trigger Type** item, enter the required information displayed at the bottom and click the **Confirm** button.
* Required information varies depending on the trigger type.Trigger Type Input Item API Gateway - API name: You can select an existing API or create a new one
- Stage: You can select an existing stage or create a new one
Cronjob - Refer to the example and enter the trigger’s repeat frequency (minute, hour, day, month, day of week)
- Timezone setting: select the time zone to apply
Table. Required input items when adding a trigger6. After moving to the **Code** tab, click the **Edit** button. You will be taken to the **Function Code Edit** page.
7. After adding the processing logic for success and failure cases, click the **Save** button.
* Node.js source codeColor modeclass CustomError extends Error {
constructor(message) {
super(message);
this.name = 'CustomError';
}
}
exports.handleRequest = async function (params) {
/**
* @description User writing area (Function details)
*/
throw new CustomError('This is a custom error!');
};
class CustomError extends Error {
constructor(message) {
super(message);
this.name = 'CustomError';
}
}
exports.handleRequest = async function (params) {
/**
* @description User writing area (Function details)
*/
throw new CustomError('This is a custom error!');
};
Throw a custom error - Node.js source code * Python source codeColor modeclass CustomError(Exception):
def __init__(self, message):
self.message = message
def handle_request(parmas):
raise CustomError('This is a custom error!')
class CustomError(Exception):
def __init__(self, message):
self.message = message
def handle_request(parmas):
raise CustomError('This is a custom error!')
Throw a custom error - Python source code * PHP source codeColor mode<?php
class CustomError extends Exception {
public function __construct($message) {
parent::__construct($message);
$this->message = $message;
}
}
function handle_request() {
throw new CustomError('This is a custom error!');
}
?>
<?php
class CustomError extends Exception {
public function __construct($message) {
parent::__construct($message);
$this->message = $message;
}
}
function handle_request() {
throw new CustomError('This is a custom error!');
}
?>
Throw a custom error - PHP source code ## Check function call
**Function Details** page's **Configuration** tab after calling the **function URL**, check for errors in the **Log** tab.
# Using Environment Variable
Using Environment Variable (Using Environment Variable) configuration and function call example (using function URL) is explained.
## Using Environment Variable Setup
To set Using Environment Variable, follow the steps below.
1. **All Services > Compute > Cloud Functions** Click the menu. Go to the **Service Home** page of Cloud Functions.
2. Click the **Function** menu on the **Service Home** page. Navigate to the **Function List** page.
3. **Function List** Click the resource to set the trigger on the page. **Function Details** Navigate to the page.
4. After clicking the **Trigger** tab, click the **Add Trigger** button. The **Add Trigger** popup opens.
5. **Add Trigger** In the popup window, after selecting the **Trigger Type** item, enter the required information displayed at the bottom and click the **Confirm** button.
* Required information varies depending on the type of trigger.Trigger Type Input Item API Gateway - API name: You can select an existing API or create a new one
- Stage: You can select an existing stage or create a new one
Cronjob - Refer to the example and enter the trigger’s repeat frequency (minute, hour, day, month, day of week)
- Timezone setting: select the reference time zone to apply
Table. Required input items when adding a trigger6. After moving to the **Code** tab, click the **Edit** button. You will be taken to the **Function Code Edit** page.
7. After adding the processing logic for success and failure cases, click the **Save** button.
* Node.js source codeColor modeexports.handleRequest = async function (params) {
/**
* @description User writing area (Function details)
*/
return process.env.test;
};
exports.handleRequest = async function (params) {
/**
* @description User writing area (Function details)
*/
return process.env.test;
};
Using Environment Variable - Node.js source code * Python source codeColor modeimport json
import os
def handle_request(params):
# User writing area (Function details)
return os.environ.get("test")
import json
import os
def handle_request(params):
# User writing area (Function details)
return os.environ.get("test")
Using Environment Variable - Python source code * PHP source codeColor modeimport json
def handle_request(params):
# User writing area (Function details)
return os.environ.get("test")
import json
def handle_request(params):
# User writing area (Function details)
return os.environ.get("test")
Using Environment Variable - PHP source code 9. After moving to the **Configuration** tab, click the **Edit** button in the **Environment Variable** area. The **Edit Environment Variable** popup window opens.
10. After entering the environment variable information, click the **Confirm** button.Category Detailed description Name Enter Key value value ValueEnter value
Table. Environment Variable Input Items## Check function call
**Function Details** page's **Configuration** tab, after calling the **Function URL**, check the environment variable values in the **Log** tab.6.2.4 - PrivateLink Service Integration
By linking Cloud Functions and PrivateLink services, you can connect VPCs within the Samsung Cloud Platform and VPCs to services without external internet.
The data uses only the internal network, which enhances security, and does not require a public IP, NAT, VPN, internet gateway, etc.
PrivateLink Service Integration
You can expose the function via PrivateLink Service so that it can be accessed privately from another VPC.
To integrate the PrivateLink service, follow the steps below.
- All Services > Compute > Cloud Functions Click the menu. Navigate to the Service Home page of Cloud Functions.
- Click the Cloud Functions menu on the Service Home page. You will be taken to the Function list page.
- Click the resource to associate PrivateLink on the Function List page. It moves to the Function Details page.
- Click the Configuration tab on the Function Details page.
- Click the Edit button of PrivateLink Service in Private connection configuration. Edit PrivateLink Service popup window opens.
- PrivateLink Service Edit in the popup window, after checking the Use item of Activation Status, click the Confirm button. In the Configuration tab’s Private Connection Configuration, the PrivateLink Service information is displayed.
| Category | Detailed description |
|---|---|
| Private URL | PrivateLink Service URL information |
| PrivateLink Service ID | PrivateLink Service ID information |
| Request Endpoint Management | List of PrivateLink Endpoints that requested connection to PrivateLink Service
|
PrivateLink Endpoint Create
Create an entry point to access the user’s VPC PrivateLink Service.
To create a PrivateLink Endpoint, follow these steps.
- All Services > Compute > Cloud Functions Click the menu. Go to the Service Home page of Cloud Functions.
- On the Service Home page, click the Cloud Functions menu. It navigates to the Function list page.
- Click the resource to associate with PrivateLink on the Function List page. You will be taken to the Function Details page.
- Click the Configuration tab on the Function Details page.
- Click the Add button of PrivateLink Endpoint in Configure Private Connection. Add PrivateLink Endpoint popup opens.
- PrivateLink Service Add in the popup window, after entering PrivateLink Service ID and Alias information, click the Confirm button.
- When the popup notifying creation opens, click the Confirm button. In the Configuration tab’s Private connection configuration, the PrivateLink Endpoint information is displayed.
| Category | Detailed description |
|---|---|
| PrivateLink Endpoint ID | PrivateLink Endpoint ID information |
| PrivateLink Service ID | PrivateLink Service ID information |
| Alias | Alias information |
| Status | Approval status of PrivateLink Endpoint
|
6.2.5 - 리소스 기반 정책 가이드
Overview of Resource-Based Policies
A resource-based policy for Cloud Functions is a policy attached to a resource that can allow or deny (Effect) specific actions (Action) for a given principal (Principal). You can directly define the principal that can execute (Invoke) a function using resource-based policies.
You can allow function calls by defining the following in a resource-based policy.
- User of the specified Samsung Cloud Platform account
- Specified source IP address range or CIDR block
Source policies are defined as JSON policy documents attached to an API, which control whether a specified security principal (typically an IAM role or group) can call the API.
| Category | Explanation | Example |
|---|---|---|
| Principal | Specify the function caller | specific object storage bucket, API Gateway, other Samsung Cloud Platform accounts, etc. |
| Task (Action) | Define the allowed functionality | Mostly scf:InvokeFunction |
| Condition (Condition) | Restrict to allow only in specific situations | Allow only requests from a bucket with a specific SRN. |
- Cloud Functions’ resource-based policies leverage the rules of IAM’s resource-based policies.
- Refer to the JSON Mode Usage Guide for how to create or modify policies using JSON.
Resource-based policy usage scenario
The primary use cases for resource-based policies are as follows.
Resource-based policy scenario
The resource-based policy scenarios used when a Cloud Functions function runs are as follows.
| Category | Explanation | Reference example |
|---|---|---|
| Function URL - Authentication Type None | It is required when generating a URL for a function and invoking it.
| Function URL (Auth Type None) Example |
| Function URL - Authentication Type IAM |
| Function URL (Authentication Type IAM) Example |
| API Gateway trigger | It is required when API Gateway calls Lambda to handle external API requests.
| API Gateway Trigger Example |
| PrivateLink connection | You can define a PrivateLink Service connection to allow private access to a function from another VPC.
| PrivateLink connection example |
Additional user usage scenarios
Although it is not automatically registered as a resource‑based policy for Cloud Functions, users can add and use it as needed. The scenarios that users can add and utilize are as follows.
- Cross-Account Access
- When an IAM user in account A wants to invoke a Lambda function in account B, add account A to the function policy of account B.
- Hybrid Access Control
- You can configure it so that access is allowed only when both conditions are met—a specific user and a specific IP range—rather than restricting just the account or IP alone.
Resource-based policy management for Cloud Functions
To view and configure resource-based policies for Cloud Functions, follow these steps.
- Click the All Services > Compute > Cloud Functions menu. 1. Navigate to the Service Home page of Cloud Functions.
- On the Service Home page, click the Function menu. 2. Navigate to the Function list page.
- On the Function List page, click the resource for which you want to set a policy. 3. Go to the Function Details page.
- Click the Configuration tab on the Function Details page.
- Click the Edit button for the Resource-based policy permissions item. 5. Resource Policy edit popup window opens.
- In the Resource Policy edit popup, select a Policy Template and then write the policy.
- For policy examples by template, refer to the Resource-based policy example.
- When you have finished writing, click the Confirm button.
- When you click the Delete button, the registered policy is deleted.
Resource-based policy example
Users can define additional resource-based policies as needed or modify existing policies for use.
- For some features, a resource‑based policy (or credential) must be registered to use them in Cloud Functions.
- In the resource-based policy example described in this guide, Cloud Functions automatically registers the example resource-based policy when each feature is enabled or linked.
Function URL - Authentication Type None
A policy that permits public calls when the Principal is /*.
Policy Template
"Statement": [
"Action": ["scf:InvokeFunctionUrl"],
"Condition": {
"StringEquals": {
"scf:CloudFunctionAuthType": ["NONE"]
}
},
"Effect": "Allow",
"Principal": "*"
"Resource": ["{{CloudFunctionSrn}}"],
"Sid": "InvokeFunctionURLAllowPublicAccess"
}
],
"Version": "2024-07-01"
} "Statement": [
"Action": ["scf:InvokeFunctionUrl"],
"Condition": {
"StringEquals": {
"scf:CloudFunctionAuthType": ["NONE"]
}
},
"Effect": "Allow",
"Principal": "*"
"Resource": ["{{CloudFunctionSrn}}"],
"Sid": "InvokeFunctionURLAllowPublicAccess"
}
],
"Version": "2024-07-01"
}Policy example
"Statement": [
"Action": ["scf:InvokeFunctionUrl"],
"Condition": {
"StringEquals": {
"scf:CloudFunctionAuthType": ["NONE"]
}
},
"Effect": "Allow",
"Principal": "*"
"Resource": ["srn:e::accountID:kr-west1::scf:cloud-function/functionsID"],
"Sid": "InvokeFunctionURLAllowPublicAccess"
}
],
"Version": "2024-07-01"
} "Statement": [
"Action": ["scf:InvokeFunctionUrl"],
"Condition": {
"StringEquals": {
"scf:CloudFunctionAuthType": ["NONE"]
}
},
"Effect": "Allow",
"Principal": "*"
"Resource": ["srn:e::accountID:kr-west1::scf:cloud-function/functionsID"],
"Sid": "InvokeFunctionURLAllowPublicAccess"
}
],
"Version": "2024-07-01"
}Function URL - Authentication Type IAM
This is a policy that permits specific users to invoke the public URL.
Policy Template
"Statement": [
"Action": ["scf:InvokeFunctionUrl"],
"Condition": {
"StringEquals": {
"scf:CloudFunctionAuthType": ["SCP_IAM"]
}
},
"Effect": "Allow"
"Principal": {
"scp": ["srn:{{Environment}}::{{AccountID}}:::iam:user/{{UserId}}"]
},
"Resource": ["{{CloudFunctionSrn}}"],
"Sid": "Statement1"
}
],
"Version": "2024-07-01"
} "Statement": [
"Action": ["scf:InvokeFunctionUrl"],
"Condition": {
"StringEquals": {
"scf:CloudFunctionAuthType": ["SCP_IAM"]
}
},
"Effect": "Allow"
"Principal": {
"scp": ["srn:{{Environment}}::{{AccountID}}:::iam:user/{{UserId}}"]
},
"Resource": ["{{CloudFunctionSrn}}"],
"Sid": "Statement1"
}
],
"Version": "2024-07-01"
}Policy example
"Statement": [
"Action": ["scf:InvokeFunctionUrl"],
"Condition": {
"StringEquals": {
"scf:CloudFunctionAuthType": ["SCP_IAM"]
}
},
"Effect": "Allow"
"Principal": "*"
"Resource": ["srn:e::accountID:kr-west1::scf:cloud-function/functionsID"],
"Sid": "accountID-iam-invokefunctionurl"
}
],
"Version": "2024-07-01"
} "Statement": [
"Action": ["scf:InvokeFunctionUrl"],
"Condition": {
"StringEquals": {
"scf:CloudFunctionAuthType": ["SCP_IAM"]
}
},
"Effect": "Allow"
"Principal": "*"
"Resource": ["srn:e::accountID:kr-west1::scf:cloud-function/functionsID"],
"Sid": "accountID-iam-invokefunctionurl"
}
],
"Version": "2024-07-01"
}API Gateway trigger
A policy that allows public calls with Principal set to *.
Policy Template
"Statement": [
"Action": ["scf:InvokeFunction"],
"Condition": {
"SrnLike": {
"scp:RequestAttribute/body['x-scf-request-obj-srn']": ["{{ApiGatewayMethodSrn}}"]
}
},
"Effect": "Allow",
"Principal": {
"Service": ["apigateway.samsungsdscloud.com"]
},
"Resource": ["{{CloudFunctionSrn}}"],
"Sid": "Statement1"
}
],
"Version": "2024-07-01"
} "Statement": [
"Action": ["scf:InvokeFunction"],
"Condition": {
"SrnLike": {
"scp:RequestAttribute/body['x-scf-request-obj-srn']": ["{{ApiGatewayMethodSrn}}"]
}
},
"Effect": "Allow",
"Principal": {
"Service": ["apigateway.samsungsdscloud.com"]
},
"Resource": ["{{CloudFunctionSrn}}"],
"Sid": "Statement1"
}
],
"Version": "2024-07-01"
}Policy example
{
"Statement": [
"Action": [
scf:InvokeFunction
],
"Condition": {
"SrnLike": {
"scp:RequestAttribute/body['x-scf-request-obj-srn']": [
srn:e::accountID:kr-west1::apigateway:method/MethodID/*/GET/test
]
}
},
"Effect": "Allow"
"Principal": {
"Service": [
apigateway.samsungsdscloud.com
]
},
"Resource": [
srn:e::accountID:kr-west1::scf:cloud-function/functionID
],
"Sid": "999e9a9999de4d4683c9e10c74ee999z"
}
],
"Version": "2024-07-01"
}{
"Statement": [
"Action": [
scf:InvokeFunction
],
"Condition": {
"SrnLike": {
"scp:RequestAttribute/body['x-scf-request-obj-srn']": [
srn:e::accountID:kr-west1::apigateway:method/MethodID/*/GET/test
]
}
},
"Effect": "Allow"
"Principal": {
"Service": [
apigateway.samsungsdscloud.com
]
},
"Resource": [
srn:e::accountID:kr-west1::scf:cloud-function/functionID
],
"Sid": "999e9a9999de4d4683c9e10c74ee999z"
}
],
"Version": "2024-07-01"
}PrivateLink connection
This is a policy that permits function calls via a Privatelink Endpoint for specific users.
Policy Template
"Statement": [
"Action": ["scf:InvokeFunction"],
"Condition": {
"StringEquals": {
"scf:CloudFunctionPrivatelinkServiceAuthType": ["SCP_IAM"]
}
},
"Effect": "Allow",
"Principal": {
"scp": ["srn:{{Environment}}::{{AccountID}}:::iam:user/{{UserId}}"]
},
"Resource": ["{{CloudFunctionSrn}}"],
"Sid": "Statement1"
}
],
"Version": "2024-07-01"
} "Statement": [
"Action": ["scf:InvokeFunction"],
"Condition": {
"StringEquals": {
"scf:CloudFunctionPrivatelinkServiceAuthType": ["SCP_IAM"]
}
},
"Effect": "Allow",
"Principal": {
"scp": ["srn:{{Environment}}::{{AccountID}}:::iam:user/{{UserId}}"]
},
"Resource": ["{{CloudFunctionSrn}}"],
"Sid": "Statement1"
}
],
"Version": "2024-07-01"
}Policy example
"Statement": [
"Action": [
scf:InvokeFunction
],
"Condition": {
"StringEquals": {
"scf:CloudFunctionAuthType": [
SCP_IAM
]
}
},
"Effect": "Allow"
"Principal": {
"scp": [
srn:e::accountID:::iam:user/userID
]
},
"Resource": [
srn:e::accountID:kr-west1::scf:cloud-function/functionID
],
"Sid": "rrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrr-privatelink-invokefunction"
}
],
"Version": "2024-07-01"
} "Statement": [
"Action": [
scf:InvokeFunction
],
"Condition": {
"StringEquals": {
"scf:CloudFunctionAuthType": [
SCP_IAM
]
}
},
"Effect": "Allow"
"Principal": {
"scp": [
srn:e::accountID:::iam:user/userID
]
},
"Resource": [
srn:e::accountID:kr-west1::scf:cloud-function/functionID
],
"Sid": "rrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrr-privatelink-invokefunction"
}
],
"Version": "2024-07-01"
}6.3 - Release Note
Cloud Functions
- You can use functions in conjunction with the AIOS service.
- Cloud Functions can be linked with AIOS to utilize LLM.
- You can use functions in conjunction with the PrivateLink service.
- Through Private connection (PrivateLink), you can internally connect Samsung Cloud Platform’s VPC to VPC, and VPC to services without going through the Internet.
- The feature to upload Java Runtime executable files has been added.
- You can fetch and configure a Java Runtime executable archive file (.jar/.zip) to Object Storage.
- Cloud Functions service has been officially launched.
- It is a serverless computing-based FaaS (Function as a Service) that easily runs function-style applications without the need for server provisioning.
7 - Virtual Server DR
In the event of a system disruption due to various disasters and risk factors, the Block Storage connected to the Virtual Server in another region can be replicated to restore normal operating conditions in a short period of time.
7.1 - Overview
Service Overview
Virtual Server DR is a service that can quickly recover the system by replicating Virtual Server and connected Block Storage in a different region from the currently used region. Even in the event of various disasters and unexpected situations that interrupt the system, Virtual Server DR can be used to quickly recover to a normal operating state.
- Virtual Server DR service can be configured through a partner solution sold in the Samsung Cloud Platform’s Marketplace.
- For more information about using Marketplace, please refer to Marketplace.
- When purchasing and using services sold on the Marketplace, a separate contract with the Marketplace software supplier is issued in accordance with a separate tax invoice.
- If you have applied for a partner solution product for Virtual Server DR on the Marketplace, the application information will be sent to the person in charge by email. Please coordinate the product details and schedule with the person in charge. The software installation and cost will be charged based on the confirmed date.
- Services sold in the Samsung Cloud Platform’s Marketplace are services sold by individual sellers, and SamsungSDS is an intermediary of electronic commerce and is not a party to the electronic commerce. Therefore, SamsungSDS does not guarantee or take responsibility for the service information and transactions sold by individual sellers.
Features
- Easy DR Environment Configuration: You can easily configure a Virtual Server for DR configuration through partner solutions in Samsung Cloud Platform’s Marketplace.
- Various Environment Configuration: Using partner solutions, you can configure various environments such as physical to virtual environment (P2V), virtual to virtual environment (V2V), and support multiple operating systems (Windows, Linux).
Service Composition Diagram
Provided Features
The main feature is to refer to the product catalog details page of the partner solution being sold in the Samsung Cloud Platform’s Marketplace.
Preceding Service
This is a list of services that must be pre-configured before creating this service. Please refer to the guide provided for each service and prepare in advance for more details.
| Service Category | Service | Detailed Description |
|---|---|---|
| Networking | VPC | A service that provides an independent virtual network in a cloud environment |
| Networking | Security Group | A virtual firewall that controls the server’s traffic |
| Compute | Virtual Server | Cloud computing-optimized virtual server |
7.2 - Release Note
Virtual Server DR
- Virtual Server DR service has been officially released.
- The system can be restored to normal operating conditions in a short period of time when it is interrupted by various disasters and risk factors.
8 - Block Storage
8.1 - Overview
Service Overview
Block Storage is a high-performance storage that stores data in block units arranged in a certain size and array.
It is suitable for large-capacity, high-performance requirements such as databases and mail servers, and users can directly assign volumes to servers for use.
Key Features
- Large-capacity Volume Provision: Volumes for OS configuration are created with at least the minimum capacity per image and can be expanded up to 12TB, and volumes for data storage other than OS can be created and expanded with capacities from a minimum of 8GB to a maximum of 12TB. Capacity expansion is performed stably in an online state.
- High Performance Based on Full SSD: Provides high durability and availability based on redundant Controllers and Disk Array Raid. Since Full SSD disks are provided by default, it is suitable for high-speed data processing tasks such as database workloads.
- Snapshot Backup: Recovery of changed and deleted data is possible through the image snapshot function. Users can select a snapshot created at the desired time point from the list to perform recovery.
Service Architecture
Provided Features
Block Storage provides the following features:
- Volume Name: Users can set or modify names for each volume.
- Capacity: Volumes can be created with capacities from a minimum of 8GB to a maximum of 12TB, and can be expanded during use. OS basic volumes can be created with at least the minimum capacity per image.
- Connected Server: You can connect or disconnect by selecting a Virtual Server.
- Multi Attach: Connects to 2 or more servers, no limit on the number of connected servers per volume, and Virtual Server can connect up to 26 volumes
- Encryption: AES-256 algorithm encryption is applied by default to all volumes of Block Storage, and when the volume is HDD/SSD_KMS disk type, it additionally provides transfer encryption between the instance and the Block Storage section connected to the instance.
- Snapshot: Recovery of changed and deleted data is possible through the image snapshot function. Users select a snapshot created at the desired time point from the list to recover.
- Volume Transfer: You can transfer volumes to another Account through the volume transfer function.
- Monitoring: You can check monitoring information such as IOPS, Latency, Throughput, etc. through the Cloud Monitoring service.
Components
You can create a volume by entering capacity according to your service scale and performance requirements and selecting a disk type. When using the snapshot function, you can recover data to the desired time point.
Volume
Volume is the basic creation unit of the Block Storage service and is used as data storage space. Users create a volume by selecting a name, capacity, and disk type, then connect it to a Virtual Server for use.
The volume name creation rules are as follows:
-, _).Snapshot
Snapshot is an image backup of a volume at a specific point in time. Users can check the snapshot name and creation date/time from the snapshot list and select the snapshot they want to recover, and can recover changed or deleted data through that snapshot.
Notes for using snapshots are as follows:
- Snapshot creation date/time is based on Asia/Seoul(GMT +09:00).
- You can recover a Block Storage volume to the latest snapshot by selecting the snapshot recovery button.
- When selecting a specific snapshot from the snapshot list, you can recover by creating a new snapshot-based volume.
- Snapshots are charged according to the size of the original Block Storage, so please delete unnecessary snapshots.
Prerequisite Services
This is a list of services that must be pre-configured before creating this service. Please prepare in advance by referring to the guides provided for each service.
| Service Category | Service | Description |
|---|---|---|
| Compute | Virtual Server | Virtual server optimized for cloud computing |
8.1.1 - Monitoring Metrics
Block Storage monitoring metrics
The table below shows the monitoring metrics of Block Storage that can be checked through Cloud Monitoring. For detailed usage of Cloud Monitoring, refer to the Cloud Monitoring guide.
| Performance Item Name | Description | Unit |
|---|---|---|
| Volume Total | Total Bytes | bytes |
| IOPS [Read] | iops(reading) | iops |
| IOPS [Write] | iops(writing) | iops |
| Latency Time [Read] | Delay Time (Read) | usec |
| Latency Time [write] | Delay Time (write) | usec |
| Throughput [Read] | Throughput (Read) | bytes/s |
| Throughput [Write] | Throughput (Write) | bytes/s |
8.2 - How-to guides
Users can create the Block Storage service by entering required information through the Samsung Cloud Platform Console and selecting detailed options.
Creating Block Storage
You can create and use the Block Storage service through the Samsung Cloud Platform Console.
To create Block Storage, follow these steps:
Click the All Services > Compute > Virtual Server menu. This will take you to the Virtual Server Service Home page.
Click the Block Storage menu. This will take you to the Block Storage List page.
Click the Create Service button on the Block Storage page. This will take you to the Create Block Storage page.
Enter the required information for creating the service on the Create Block Storage page and select detailed options.
Item RequiredDescription Volume Name Required Volume name - Enter 255 characters using English letters, numbers, spaces, and special characters (
-,_)
Snapshot Optional Select snapshot to use when creating volume through snapshot - After checking the Use item, you can select a snapshot
- When creating service through snapshot recovery volume, provides recovery snapshot name
- If not selected, an empty volume is created
Disk Type Required Select disk type - SSD_Provisioned: SSD volume where IOPS and Throughput can be set
- SSD/HDD: General SSD/HDD volume
- SSD/HDD_KMS: Additional encryption volume
- SSD/HDD_MultiAttach: Volume that can be connected to 2 or more servers
- Cannot modify after service creation
- When creating service through snapshot recovery volume, it is set the same as the original and cannot be modified
Capacity Optional Set capacity - Can create within 8~12,228GB
- Enter number of Units provided in 8GB increments
- When creating service through snapshot recovery volume, enter capacity equal to or larger than the original
Max IOPS Required Enter maximum IOPS value between 5,000~20,000 - Can only set when Disk Type is SSD_Provisioned
Max Throughput Required Enter maximum Throughput value between 250~1,000 - Can only set when Disk Type is SSD_Provisioned
Table. Block Storage service information input items- Enter 255 characters using English letters, numbers, spaces, and special characters (
Review the detailed information and estimated charges in the Summary panel, then click the Create button.
- Once creation is complete, verify the created resource on the Block Storage List page.
- AES-256 algorithm encryption is applied by default to all volumes of Block Storage.
- Windows-based Virtual Server cannot use MultiAttach disks. Please use a separate replication method or solution.
- When the volume is HDD/SSD_KMS disk type, it additionally provides transfer encryption between the instance and the Block Storage section connected to the instance.
Viewing Block Storage Details
Block Storage service allows you to view and modify the complete resource list and detailed information. The Block Storage Details page consists of Details, Snapshot List, Tags, and Operation History tabs.
To view detailed information of Block Storage service, follow these steps:
- Click the All Services > Compute > Virtual Server menu. This will take you to the Virtual Server Service Home page.
- Click the Block Storage menu. This will take you to the Block Storage List page.
- Click the resource for which you want to view detailed information on the Block Storage List page. This will take you to the Block Storage Details page.
- The Block Storage Details page displays status information and additional feature information, and consists of Details, Snapshot List, Tags, and Operation History tabs.
Item Description Volume Status Status of the volume - Creating: Being created
- Downloading: Being created (OS image being applied)
- Available: Creation complete, can connect to server
- Reserved: Waiting for server connection
- Attaching: Connecting to server
- Detaching: Disconnecting from server
- In Use: Server connection complete
- Deleting: Service being terminated
- Awaiting Transfer: Waiting for volume transfer
- Extending: Capacity expansion
- Error Extending: Abnormal state during capacity expansion
- Backing Up: Volume being backed up
- Restoring Backup: Volume backup being recovered
- Error Backing Up: Volume backup abnormal state
- Error Restoring: Volume backup recovery abnormal state
- Error Deleting: Abnormal state during deletion
- Error Managing: Abnormal state
- Error: Abnormal state
- Maintenance: Temporary maintenance state
- Reverting: Recovering from snapshot
Volume Transfer Transfer volume to another Account - For details on volume transfer, see Volume Transfer
Create Snapshot Immediately create a snapshot at the creation point - For details on snapshot creation, see Creating a Snapshot
Snapshot Recovery Recover volume to the latest snapshot in Available state - For details on snapshot recovery, see Recovering a Snapshot
Service Termination Button to terminate the service Table. Status information and additional features
- The Block Storage Details page displays status information and additional feature information, and consists of Details, Snapshot List, Tags, and Operation History tabs.
Details
On the Block Storage List page, you can view detailed information of the selected resource and modify information if necessary.
| Item | Description |
|---|---|
| Service | Service group |
| Resource Type | Resource type |
| SRN | Unique resource ID in Samsung Cloud Platform
|
| Resource Name | Resource name
|
| Resource ID | Unique resource ID of the service |
| Creator | User who created the service |
| Creation Date | Date and time when the service was created |
| Modifier | User who modified the service |
| Modification Date | Date and time when the service was modified |
| Volume Name | Volume name
|
| Volume ID | Unique volume ID |
| Disk Type | Disk type |
| Type | Classification according to volume creation method and usage |
| Capacity | Volume capacity
|
| Max IOPS | Required |
| Max Throughput | Required |
| Connected Server | Connected Virtual Server
|
Snapshot List
On the Block Storage List page, you can view snapshots of the selected resource.
| Item | Description |
|---|---|
| Snapshot Name | Snapshot name |
| Description | Snapshot description |
| Volume Capacity | Capacity of the Block Storage original volume that is the snapshot target
|
| Creation Date | Snapshot creation date/time |
| Status | Status of the snapshot
|
| Additional Features > More | Snapshot management buttons
|
| Delete | Can delete by selecting snapshot to delete from snapshot list |
- Snapshots can affect volume capacity management. Please delete unnecessary snapshots after use.
- Snapshot recovery is possible when not connected to a server.
- Snapshot creation date/time is based on Asia/Seoul (GMT +09:00).
- When clicking the snapshot recovery button, the volume is recovered to the latest snapshot in Available status.
- When selecting create recovery volume on the snapshot list page, a new snapshot-based volume is created without changing the existing volume.
- Snapshots containing the name vProtect are temporary snapshots created when using the Backup service and are not charged.
Tags
On the Block Storage List page, you can view tag information of the selected resource, and add, modify, or delete tags.
| Item | Description |
|---|---|
| Tag List | Tag list
|
Operation History
On the Block Storage List page, you can view the operation history of the selected resource.
| Item | Description |
|---|---|
| Operation History List | Resource change history
|
Managing Block Storage Resources
If you need to modify settings of created Block Storage or add or delete connected servers, you can perform tasks on the Block Storage Details page.
Modifying Volume Name
You can modify the name of a volume. To modify the volume name, follow these steps:
- Click the All Services > Compute > Virtual Server menu. This will take you to the Virtual Server Service Home page.
- Click the Block Storage menu. This will take you to the Block Storage List page.
- Click the resource for which you want to modify the volume name on the Block Storage List page. This will take you to the Block Storage Details page.
- Click the Modify button of Volume Name. The Modify Volume Name popup window opens.
- Enter the volume name and click the Confirm button.
-, _).Expanding Capacity
You can expand the capacity of a volume. To expand capacity, follow these steps:
- Click the All Services > Compute > Virtual Server menu. This will take you to the Virtual Server Service Home page.
- Click the Block Storage menu. This will take you to the Block Storage List page.
- Click the resource for which you want to expand capacity on the Block Storage List page. This will take you to the Block Storage Details page.
- Click the Modify button of Capacity. The Modify Capacity popup window opens.
- Enter the capacity and click the Confirm button.
- Capacity reduction is not provided.
- After capacity expansion, recovery to snapshots before expansion is not possible.
- Only recovery by creating a new volume is possible with snapshots created before capacity expansion.
- Can expand to capacity larger than existing capacity within 8~12,228GB.
- Enter the number of Units provided in 8GB increments.
Modifying Connected Server
You can connect or disconnect servers. To modify connected server, follow these steps:
- Click the All Services > Compute > Virtual Server menu. This will take you to the Virtual Server Service Home page.
- Click the Block Storage menu. This will take you to the Block Storage List page.
- Click the resource for which you want to modify connected server on the Block Storage List page. This will take you to the Block Storage Details page.
- When adding Virtual Server connection, click the Add button in the Connected Server item. The Add Connected Server popup window opens.
- Select the Virtual Server you want to connect and click the Confirm button.
- When disconnecting Virtual Server connection, click the Disconnect button in the Connected Server item.
- Be sure to perform disconnection work (Umount, Disk Offline) from the server before disconnecting.
- You can connect Virtual Servers created in the same location as Block Storage.
- Virtual Servers using Partition with Server Group policy cannot be connected.
- For HDD/SSD_MultiAttach disk type, can connect with 2 or more Virtual Servers and there is no limit on the number of connections.
- Windows-based Virtual Server cannot use MultiAttach disks and must use a separate replication method or solution.
- Virtual Server can connect up to 26 volumes including OS basic.
- OS basic volume cannot modify connected server and cannot terminate service.
- When adding connected server, can use after performing connection work (Mount, Disk Online) from the server. For details on server connection, see Connecting Server.
Terminating Block Storage
You can reduce operating costs by terminating unused Block Storage. However, since terminating a service may immediately stop the operating service, you should proceed with termination after fully considering the impact of service interruption.
- Be careful as data cannot be recovered after termination.
- Block Storage volumes cannot be terminated in the following cases.
- While connected to server
- OS basic volume
- Connected to Custom Image of Virtual Server
- When volume status is not Available, Error, Error Extending, Error Restoring, Error Managing
- When terminating by selecting 2 or more volumes, only volumes that can be terminated will be terminated.
To terminate Block Storage, follow these steps:
- Click the All Services > Compute > Virtual Server menu. This will take you to the Virtual Server Service Home page.
- Click the Block Storage menu. This will take you to the Block Storage List page.
- Select the resource to terminate on the Block Storage List page and click the Terminate Service button.
- When termination is complete, verify that the resource has been terminated on the Block Storage List page.
8.2.1 - Connecting to the Server
When using a volume on a server, connection or disconnection work is required.
From the Block Storage Details page, add the connection server and then connect to the server to perform the connection work (Mount, Disk Online). After use, perform the disconnection work (Umount, Disk Offline) and then remove the connection server.
Connecting to the Server (Mount, Disk Online)
To use the volume added to the connection server, you must connect to the server and perform the connection work (Mount, Disk Online). Follow the procedure below.
Linux Operating System
- Server OS: LINUX
- Mount location: /data
- Volume capacity: 24 GB
- File system: ext3, ext4, xfs etc
- Additional allocated disk: /dev/vdb
- Click the All Services > Compute > Virtual Server menu. Move to the Service Home page of Virtual Server.
- Click the Block Storage menu. Move to the Block Storage List page.
- On the Block Storage List page, click the resource to be used by the connection server. Move to the Block Storage Details page.
- Check the server in the Connection Server section and connect to it.
- Refer to the procedure below to connect (Mount) the volume.
Switch to root privileges
$ sudo -iCheck the disk
# lsblk NAME MAJ:MIN RM SIZE RO TYPE MOUNTPOINT vda 252:0 0 24G 0 disk ├─vda1 252:1 0 23.9G 0 part [SWAP] └─vda14 252:14 0 4M 0 part / └─vda15 252:15 0 106M 0 part /boot/efi vdb 252:16 0 24G 0 diskCreate a partition
# fdisk /dev/vdb Command (m for help): n Partition type: p primary (0 primary, 0 extended, 4 free) e extended Select (default p): p Partition number (1-4, default 1): 1 First sector (2048-50331646, default 2048): Last sector, +/-sectors or +/-size{K,M,G,T,P} (2048-50331646, default 50331646): Created a new partition 1 of type 'Linux' and of size 24 GiB. Command (m for help): w The partition table has been altered! Calling ioctl() to re-read partition table. Syncing disks.Set the partition format (e.g., ext4)
# lsblk NAME MAJ:MIN RM SIZE RO TYPE MOUNTPOINT vda 252:0 0 24G 0 disk ├─vda1 252:1 0 23.9G 0 part [SWAP] └─vda14 252:14 0 4M 0 part / └─vda15 252:15 0 106M 0 part /boot/efi vdb 252:16 0 24G 0 disk └─vdb1 252:17 0 24G 0 part # mkfs.ext4 /dev/vdb1 mke2fs 1.46.5 (30-Dec-2021) ... Writing superblocks and filesystem accounting information: doneMount the volume
# mkdir /data # mount /dev/vdb1 /data # lsblk NAME MAJ:MIN RM SIZE RO TYPE MOUNTPOINT vda 252:0 0 24G 0 disk ├─vda1 252:1 0 23.9G 0 part [SWAP] └─vda14 252:14 0 4M 0 part / └─vda15 252:15 0 106M 0 part /boot/efi vdb 252:16 0 24G 0 disk └─vdb1 252:17 0 24G 0 part /data # vi /etc/fstab (add) /dev/vdb1 /data ext4 defaults 0 0
| Item | Description |
|---|---|
| cat /etc/fstab | File system information file
|
| df -h | Check the total disk usage of the mounted disk |
| fdisk -l | Check partition information
|
| Command | Description |
|---|---|
| m | Check the usage of the fdisk command |
| n | Create a new partition |
| p | Check the changed partition information |
| t | Change the system ID of the partition |
| w | Save the partition information and exit the fdisk settings |
Windows Operating System
- Click the All Services > Compute > Virtual Server menu. Move to the Service Home page of Virtual Server.
- Click the Block Storage menu. Move to the Block Storage List page.
- On the Block Storage List page, click the resource to be used by the connection server. Move to the Block Storage Details page.
- Check the server in the Connection Server section and connect to it.
- Refer to the procedure below to connect (Disk Online) the volume.
Right-click the Windows start icon and run Computer Management
In the Computer Management tree structure, select Storage > Disk Management
Check the disk
Bring the disk online
Initialize the disk
Format the partition
Check the volume
Disconnecting from the Server (Umount, Disk Offline)
Connect to the server and perform the disconnection work (Umount, Disk Offline), and then remove the connection server from the console.
Follow the procedure below.
- If you disconnect the server from the console without performing the disconnection work (Umount, Disk Offline) on the server, a server status error (Hang) may occur.
- Be sure to perform the OS work first.
- For the OS basic volume, connection server modification and service termination are not allowed.
Linux Operating System
- Click the All Services > Compute > Virtual Server menu. Move to the Service Home page of Virtual Server.
- Click the Block Storage menu. Move to the Block Storage List page.
- On the Block Storage List page, click the resource to be disconnected from the connection server. Move to the Block Storage Details page.
- Check the server in the Connection Server section and connect to it.
- Refer to the procedure below to disconnect (Umount) the volume.
- Umount the volume
# umount /dev/vdb1 /data
# lsblk
NAME MAJ:MIN RM SIZE RO TYPE MOUNTPOINT
vda 252:0 0 24G 0 disk
├─vda1 252:1 0 23.9G 0 part [SWAP]
└─vda14 252:14 0 4M 0 part /
└─vda15 252:15 0 106M 0 part /boot/efi
vdb 252:16 0 24G 0 disk
└─vdb1 252:17 0 24G 0 part
# vi /etc/fstab
(delete) /dev/vdb1 /data ext4 defaults 0 0
Windows Operating System
Click the All Services > Compute > Virtual Server menu. Move to the Service Home page of Virtual Server.
Click the Block Storage menu. Move to the Block Storage List page.
On the Block Storage List page, click the resource to be disconnected from the connection server. Move to the Block Storage Details page.
Check the server in the Connection Server section and connect to it.
Unmount the file system.
Refer to the procedure below to disconnect (Disk Offline) the volume.
Right-click the Windows start icon and run Computer Management
In the Computer Management tree structure, select Storage > Disk Management
Right-click the disk to be removed and run Offline
Check the disk status
8.2.2 - Using Snapshots
You can create, delete, or restore snapshots of the created Block Storage using snapshots. You can perform actions on the Block Storage Details page and Snapshot List page.
Create Snapshot
You can create a snapshot of the current point in time. To create a snapshot, follow these steps.
- All Services > Compute > Virtual Server Click the menu. Navigate to the Service Home page of Virtual Server.
- Click the Block Storage menu. Go to the Block Storage List page.
- Block Storage List On the page, click the resource to create a snapshot. Block Storage Details Go to the page.
- Click the Create Snapshot button. The Create Snapshot popup window opens.
- Enter Snapshot Name and Description, then click the Confirm button. It creates a snapshot of the current point in time.
- Snapshot List Click the button. Block Storage Snapshot List Navigate to the page.
- Check the generated snapshot.
Edit Snapshot
You can edit snapshot information. To edit the snapshot name or description, follow the steps below.
- All Services > Compute > Virtual Server Click the menu. Navigate to the Service Home page of Virtual Server.
- Block Storage Click the menu. Navigate to the Block Storage List page.
- Block Storage List page, click the resource to edit the snapshot information. You will be taken to the Block Storage Details page.
- Snapshot List Click the button. Block Storage Snapshot List Navigate to the page.
- After confirming the snapshot to edit, click the More button.
- Click the Edit button. The Edit Snapshot popup opens.
- Enter Snapshot name or Description and click the Confirm button.
Recover Snapshot
You can restore a Block Storage volume to the latest snapshot in Available state. To perform snapshot restoration, follow the steps below.
- All Services > Compute > Virtual Server Click the menu. Navigate to Virtual Server’s Service Home page.
- Click the Block Storage menu. Navigate to the Block Storage List page.
- Click the resource to recover from a snapshot on the Block Storage List page. You will be taken to the Block Storage Details page.
- Connection Server If there is a server added in the item, after connecting to the server, perform the disconnect operation (Umount, Disk Offline).
- For detailed information about server disconnection, please refer to Disconnect Server.
- Block Storage Details page, click the Disconnect button in the Connected Server item to remove the server. The connected server will be removed.
- For details on disconnecting the connection server, see Edit Connection Server.
- Snapshot List Click the button. Block Storage Snapshot List Navigate to the page.
- Check the latest snapshot in Available state. The volume will be restored with that snapshot.
- Click the Snapshot Recovery button. The Snapshot Recovery popup opens.
- After checking the Snapshot name and creation date/time, click the Confirm button.
- When recovery starts, it becomes Reverting, and when completed, it becomes Available.
- Detailed Information Click the page button. Block Storage Detailed Page will navigate to.
- Click the Add button of the Connected Server. Reconnect the Virtual Server.
- For detailed information about adding a connection server, please refer to Modify Connection Server.
- After connecting to the added server, perform the connection tasks (Mount, Disk Online) according to the operating system.
- For detailed information about server connection, please refer to Connect to Server.
- Snapshot recovery is possible when not connected to the server.
- If you want to recover using a snapshot that is not the latest snapshot, recovery is possible by creating a recovery volume.
- In the situation below, recovery is not possible.
- Block Storage when the volume is not in Available state
- Block Storage if there is a server connected to the volume
- If there are no recoverable snapshots
- If the latest snapshot changes during recovery creation
- If the latest snapshot is not in Available state
- If the snapshot’s volume size differs from the Block Storage volume size (when the volume has been expanded)
Create snapshot recovery volume
You can create a volume using a snapshot. To create a snapshot recovery volume, follow the steps below.
- All Services > Compute > Virtual Server Click the menu. Navigate to the Service Home page of Virtual Server.
- Click the Block Storage menu. You will be taken to the Block Storage list page.
- Click the resource to create a snapshot recovery volume on the Block Storage List page. You will be taken to the Block Storage Details page.
- Snapshot List Click the button. Block Storage Snapshot List Navigate to the page.
- After checking Snapshot name, description and Creation date/time, click the More button of the snapshot you want to restore.
- Click Create Recovery Volume. The Create Snapshot Recovery Volume popup opens.
- Click the Confirm button. You will be taken to the Create Block Storage page.
- Block Storage creation On the page, enter the information required to create the service, and select detailed options.
- Please enter the volume name and size. You can enter a size that is greater than or equal to the original volume.
- Disk type is set the same as the original and cannot be modified.
Category RequiredDetailed description Volume Name Required Volume Name - Enter up to 255 characters using English letters, numbers, spaces, and special characters (
-,_)
Disk Type Required Select Disk Type - HDD: Standard volume
- SSD: High-performance standard volume
- HDD/SSD_KMS: Volume that additionally provides transmission encryption between the instance and Block Storage
- HDD/SSD_MultiAttach: Volume that can be attached to more than one server
- Cannot be modified after service creation
- When creating a service via snapshot recovery volume creation, it is set identical to the original and cannot be modified
Capacity Select Capacity Setting - Can be created within 8~12,228GB
- Enter the number of units provided in 8GB increments
- When creating a service via snapshot recovery volume creation, input a capacity equal to or larger than the original
Recovery Snapshot Name Optional Name of the recovery snapshot used when creating the volume - Provides the recovery snapshot name when creating a service through snapshot recovery volume creation
Table. Block Storage Service Information Input Items - Enter up to 255 characters using English letters, numbers, spaces, and special characters (
- Summary Check the detailed information and estimated billing amount generated in the panel, and click the Complete button.
- Once creation is complete, check the created resource on the Block Storage List page.
Delete Snapshot
You can select a snapshot to delete. To delete a snapshot, follow these steps.
- All Services > Compute > Virtual Server Click the menu. Go to the Service Home page of Virtual Server.
- Click the Block Storage menu. Go to the Block Storage List page.
- Block Storage List on the page, click the resource to delete the snapshot. Block Storage Details navigate to the page.
- Snapshot List button, click it. Block Storage Snapshot List page, navigate to it.
- After checking the snapshot name, description and creation date/time, click the more button of the snapshot you want to delete.
- Delete Click the button. Snapshot List page, the snapshot will be removed.
8.2.3 - Move Volume
You can move the volume to a different Account, and if you move it, the volume will be removed from the existing location. You can perform volume migration from the Block Storage list or Block Storage detail page.
Volume previous
You can move a volume to a different account within the region. To move a volume, follow these steps.
- Click All services > Compute > Virtual Server menu. It moves to the Service Home page of Virtual Server.
- Block Storage menu is clicked. It moves to the Block Storage list page.
- Block Storage list page, select the resource to be moved, and then click the More > Move Volume button at the top left of the list.
- or click the Volume Migration button at the top of the Block Storage details page of the resource to be migrated.
- When the pop-up window for volume migration appears, check the volume name you want to migrate and click the Confirm button.
- When the popup window for previous completion opens, click the Confirm button. The Volume Transfer ID and Approval Key information will be downloaded as a text file.
- The volume will be changed to Awaiting Transfer status.
- Volume migration is possible within the same region.
- Volume migration is only possible when the volume is in the Available state. If it is in the In Use state, release all connected servers.
Volume rollback cancel
The volume move can be cancelled after it is created. To cancel the volume move, follow these steps.
- Click All services > Compute > Virtual Server menu. It moves to the Service Home page of Virtual Server.
- Block Storage menu is clicked. It moves to the Block Storage list page.
- Block Storage list page, click the resource to cancel the volume move. It moves to the Block Storage detail page.
- You can cancel if the volume is in the Awaiting Transfer state.
- Click the Volume Move Cancel button. The Volume Move Cancel popup window will open.
- Check the volume name you want to cancel the volume move and click the Confirm button.
- The volume will be changed to Available status.
Get previous volume
You can receive volumes from other accounts within the region. To receive a volume, follow these steps.
- Click All Services > Compute > Virtual Server menu. It moves to the Service Home page of Virtual Server.
- Block Storage menu is clicked. It moves to the Block Storage list page.
- On the Block Storage list page, click the More > Receive Volume Transfer button in the upper left corner of the list. The Receive Volume Transfer popup window opens.4. Volume Migration Enter the Volume Migration ID and Approval Key provided when creating the volume migration.
- Block Storage list page will have the volume created.
- It may take some time for the changes to be reflected.
- The account that created the volume transfer will have the transferred volume removed.
8.3 - API Reference
8.4 - CLI Reference
8.5 - Release Note
Block Storage
- The snapshot is charged based on the size of the original Block Storage.
- It has been linked with Cloud Monitoring.
- You can check IOPS, Latency, Throughput information in Cloud Monitoring.
- Block Storage feature change
- The HDD disk type has been added, and you can select the added type (HDD, HDD_MultiAttach, HDD_KMS) according to the purpose.
- Samsung Cloud Platform common feature changes
- Account, IAM and Service Home, tags, etc. reflected common CX changes.
- SSD_KMS disk type has been added.
- When SSD_KMS is selected, encryption through the KMS (Key Management Service) encryption key is added.
- Released a high-performance storage service suitable for handling large-scale data and database workloads.
- Released a high-performance storage service suitable for handling large-scale data and database workloads.






















