Browser-Based High-Performance Computing For AI Research
Revolutionize your research and innovation with CR8DL AI Cloud, offering a comprehensive suite of high-performance computing domain-specific solutions engineered to meet the needs of researchers and organizations, driving breakthroughs and accelerating discovery.
Accelerate discovery
CR8DL AI Cloud
The power of supercomputing, accessible through your browser, offering unified, computationally intensive workspaces.
Origin
An intuitive web-based portal that provides simple and scalable access to high-performance computing tools and resources from a single dashboard.
Base
High-performance computing infrastructure-as-a-Service (IaaS) providing robust compute, storage, and network capabilities.
Explore
A discovery workspace with no-code, low-code, and full-code HPC tools, focused on molecular biology, quantum simulation, image processing, and more. Offering pre-configured containers for common AI frameworks to accelerate research, development, and training processes.
Scale
The AI Platform-as-a-Service (PaaS) provides on-demand, scalable high-performance computing resources ideal for AI and machine learning tasks. Accessible from anywhere, it eliminates the need for costly hardware investments, streamlining AI implementation.
Redefining AI Cloud: The CR8DL Advantage
Accessibility
BROWSER-BASED HPc PLATFORM.
Scalability
SCALE UP OR DOWN WITH EASE.
Transparency
ONLY PAY FOR WHAT YOU USE. nO HIDDEN FEES.
CR8DL AI Infrastructure
The CR8DL AI infrastructure is housed in a secure and sustainable private data center and includes high-capacity and redundant network and storage resources to ensure reliable and uninterrupted service.
CR8DL Base GPU Cluster
Node and Cluster Technical Specifications
Compute Nodes
- CPU – AMD EPYC 7713 Processor
Sockets – 2 - Cores/Threads – 64 Cores/128 Threads per socket
- L3 Cache – 256MB per socket
- Clock – 2.0Ghz boost to 3.675Ghz
- Internal Storage – 46TB over 6 x 7.7TB NVMe
- Network – 16 x 100G Ethernet/InfiniBand
Node Accelerators – GPU
- 8x NVIDIA A100/80GB RAM
- HGX backplane
- 600GB/s inter-GPU throughput/3rd Gen NVIDIA NVLink
Cluster Storage & LAN
- 8x Cluster capacity 256TB/16x100GE Interconnects
- Multi-fabric 100G Ethernet/InfiniBand
Internet Access
- Multi-homed Internet Access
- 100Gbps Fabric
Available System Benchmarks
AMD EPYC 7763, NVIDIA A100-SXM-80GB
MXNet NVIDIA Release 22.04, PyTorch NVIDIA Release 22.04, TensorFlow NVIDIA Release 22.04
Image Classification
- Imagenet/ResNet
- 28.11 Minutes
Object Detection (Heavy Weight)
- COCO/Mask R-CNN
- 43.787 Minutes
Natural Language Processing (NLP)
- Wikipedia/Bert[1]
- 19.828 Minutes
Speech Recognition
- LibriSpeech/RNN-T
- 31.291 Minutes