AI Infra Services

KQC GPUaaS – High-End AI Cloud Services Powered by H200

KQC GPUaaS

KQC GPUaaS

What is KQC AI GPUaaS?

AI GPUaaS (GPU as a Service) is a cloud-based GPU infrastructure service designed for enterprises and developers

requiring large-scale AI model training and high-performance computing.

Providing the necessary infrastructure for high-performance workloads

such as AI model training, inference, graphics processing, autonomous driving simulation andscientificcomputing,

KQC is the first in Korea to deploy and commercialize an NVIDIA H200 SXM5-based GPU Farm,

simultaneously delivering ultra-fast performance, robust security and cost efficiency.

AI GPUaaS (GPU as a Service) is a cloud-based GPU infrastructure service designed for enterprises and developers

requiring large-scale AI model training and high-performance computing.

Providing the necessary infrastructure for high-performance workloads

such as AI model training, inference, graphics processing, autonomous driving simulation andscientificcomputing,

KQC is the first in Korea to deploy and commercialize an NVIDIA H200 SXM5-based GPU Farm, simultaneously delivering ultra-fast performance, robust security and cost efficiency.

Maximize Research and Service Efficiency with
High-Performance Compute, Fast I/O, and Fully Dedicated GPU Environments

KQC AI GPUaaS Strengths

KQC’s bare-metal GPUaaS is built on dedicated servers optimized for high-performance AI computing.

It integrates NVIDIA H200 SXM5 GPUs, GDS-certified storage, and high-speed InfiniBand networking,

delivering unmatched performance and reliabilitycompared to other GPU cloud services.

KQC’s bare-metal GPUaaS is built on dedicated servers optimized for high-performance AI computing.
It integrates NVIDIA H200 SXM5 GPUs, GDS-certified storage, and high-speed InfiniBand networking,
delivering unmatched performance and reliabilitycompared to other GPU cloud services.

Top-Tier Performance

Latest NVIDIA H200 SXM5 (141GB) GPUs
Ultra-fast GPU-to-GPU communication via InfiniBand 400Gb x4
GDS-certified high-speed storage to eliminate I/O bottlenecks
Bare-metal servers with no virtualization for maximum performance utilization

Robust Security & Isolation

Dedicated environments per customer (no multi-tenancy)
Private cloud architecture maximizing security and isolation

Cost Efficiency & Transparent Pricing

No DTO charges, ideal for large-scale training and inference data transfer Simple and clear pricing enables easy budget forecasting

Technical Specifications
Comparison: KQC vs Major CSPs

Feature

Feature

KQC GPUaaS

KQC GPUaaS

A (Global CSP)

A (Global CSP)

L (Domestic CSP)

L (Domestic CSP)

GPU

GPU

H200 SXM5 141GB

H200 SXM5 141GB

A100/H100

A100/H100

A100 or L4

A100 or L4

Interconnect

Interconnect

NVLink (up to 900 GB/s)

NVLink (up to 900 GB/s)

PCIe (32~64GB/s)

PCIe (32~64GB/s)

PCIe

PCIe

Storage

Storage

GDS-certified All-Flash

GDS-certified All-Flash

Standard SSD

Standard SSD

Standard NAS

Standard NAS

Networking

Networking

InfiniBand 400Gb x4

InfiniBand
400Gb x4

200Gb

200Gb

Ethernet

Ethernet

Virtualization

Virtualization

Bare-metal architecture

Bare-metal architecture

Mostly virtualized

Mostly virtualized

Virtualized

Virtualized

Security Isolation

Security Isolation

Dedicated environment

Dedicated environment

Multi-tenancy

Multi-tenancy

Multi-tenancy

Multi-tenancy

DTO Charges

DTO Charges

None

None

Applied

Applied

Applied

Applied

Price Comparison vs Competitors

$116.60/h

A

B

C

D

E

KQC

Reasonable Pricing Structure

Reasonable Pricing Structure

Simplified billing by eliminating unnecessary costs

Simplified billing by eliminating unnecessary costs

Minimized TCO burden

Minimized TCO burden

· with charges limited to essential resources (GPU, storage)

· with charges limited to essential resources (GPU, storage)

Hybrid Integration Cost Minimization

Hybrid Integration Cost Minimization

Public Cloud

Public Cloud

· Minimize extra costs through collaboration with professional MSPs

· Minimize extra costs through collaboration with professional MSPs

On-Prem Computing/Storage

On-Prem Computing/Storage

· Flexible billing plans available

· Flexible billing plans available

Best Cost-Performance in Korea

Best Cost-Performance in Korea

Architecture optimized for maximum performance

Architecture optimized for maximum performance

Professional consulting to achieve optimal cost-efficiency

Professional consulting to achieve optimal cost-efficiency

· Flexible configuration design for diverse usage environments

· Flexible configuration design for diverse usage environments

H200 SXM5

H200 SXM5

H200 SXM5

H200 SXM5

Experience the Upgraded Cloud with H200,

the Heart of AI Supercomputing

Experience the Upgraded Cloud with H200, the Heart of AI Supercomputing

Advantages of NVIDIA H200 SXM5

Industry-leading
memory capacity
Industry-leading
memory capacity
Industry-leading
memory capacity

▸ 141 GB HBM3e enables training and inference of
ultra-large models in a single pass
▸ Ideal for large-scale models such as GPT-4 and LLaMA3

▸ 141 GB HBM3e enables training and inference of
ultra-large models in a single pass
▸ Ideal for large-scale models such as GPT-4 and LLaMA3

▸ 141 GB HBM3e enables training and inference of ultra-large models in a single pass
▸ Ideal for large-scale models such as GPT-4 and LLaMA3

High-speed HBM3e memory
High-speed HBM3e memory
High-speed HBM3e memory

▸ Accelerates large-scale parallel data processing and
AI inference

▸ Accelerates large-scale parallel data processing and
AI inference

▸ Accelerates large-scale parallel data processing and AI inference

Fast GPU-to-GPU
communication via NVLink
Fast GPU-to-GPU
communication via NVLink
Fast GPU-to-GPU
communication via NVLink

▸ Maximizes parallelism in DGX systems and
SuperPOD configurations
▸ Approximately 7× higher bandwidth than PCIe
(e.g., up to 7.2 TB/s with NVLink)

▸ Maximizes parallelism in DGX systems and
SuperPOD configurations
▸ Approximately 7× higher bandwidth than PCIe
(e.g., up to 7.2 TB/s with NVLink)

▸ Maximizes parallelism in DGX systems and SuperPOD configurations
▸ Approximately 7× higher bandwidth than PCIe (e.g., up to 7.2 TB/s with NVLink)

FP8 computation optimization
FP8 computation optimization
FP8 computation optimization

▸ Maximizes performance and throughout
while minimizing precision loss

▸Maximizes performance and throughput while minimizing precision loss

▸ Maximizes performance and throughout
while minimizing precision loss

Essential for
AI-dedicated data centers
Essential for
AI-dedicated data centers
Essential for
AI-dedicated data centers

▸ Integrated in NVIDIA DGX H200, HGX H200,
and other systems as a core component for
AI supercomputer infrastructure

▸ Integrated in NVIDIA DGX H200, HGX H200,
and other systems as a core component for
AI supercomputer infrastructure

▸ Integrated in NVIDIA DGX H200, HGX H200, and other systems as a core component for AI supercomputer infrastructure

Enhanced NVIDIA DGX SuperPOD Architecture

Enhanced NVIDIA DGX SuperPOD Architecture

Compute Fabric

NVIDIA

· InfiniBand 400Gb X 3 Line

KQC

· InfiniBand 400Gb X 4 Line

Storage Fabric

NVIDIA

· InfiniBand 200Gb X 4 Line

KQC

· InfiniBand 200Gb X 8 Line

Leaf-Spine Fabric

NVIDIA

· InfiniBand 400Gb X 2 Line

KQC

· InfiniBand 400Gb X 4 Line

Service Fabric

NVIDIA

· Int./Ext. Line Coexisting

KQC

· Int. Line 100Gb X 2 Line

· Ext. Line 10Gb X 2 Line

Key Improvements Over H100

Feature

H100 SXM

H200 SXM

Architecture

Architecture

Hopper

Hopper

Hopper

Hopper

HBM Memory

HBM Memory

80/96GB (HBM3)

80/96GB (HBM3)

141GB (HBM3e)

141GB (HBM3e)

Memory Bandwidth

Memory Bandwidth

3.35TB/s

3.35TB/s

4.8 TB/s+

4.8 TB/s+

AI Compute Performance
(FP8)

AI Compute Performance
(FP8)

~1,000 TFLOPS

~1,000 TFLOPS

Similar, but reduced memory bottlenecks improve
real-world performance

Similar, but reduced memory bottlenecks improve
real-world performance

Intended Use

Intended Use

AI training/inference

AI training/inference

Optimized for
ultra-large AI models

Optimized for
ultra-large AI models