Exceptional Hashrate: Achieve a powerful 37MH/s, maximizing ALEO mining efficiency.
Low Power Consumption: Operates at just 360W, reducing electricity costs while maintaining high output, operating on 110v-240v input.
zkSNARK Algorithm: Specially optimized for ALEO mining, ensuring compatibility with cutting-edge blockchain technologies.
Quiet Operation: The AE Box is designed with home miners in mind, producing minimal noise without compromising performance.
Compact Design: Stylish and space-saving, it seamlessly fits into your home environment.
Power Supply Not Included: The miner is sold without a PSU. For optimal performance, select any high-quality 650W or higher modular PSU with sufficient efficiency and stability.
User-Friendly Setup: The AE Box is easy to configure, making it accessible to all levels of miners.
Affordable and Efficient: Designed for budget-conscious miners who seek high returns.
Home-Friendly Design: Quiet and compact, it’s ideal for residential settings.
Future-Proof Mining: Mines ALEO using the advanced zkSNARK algorithm, staying ahead of blockchain advancements.
Low power + Low Noise, fast ROI within 4 months ONLY (at time of writing)
Warranty: 6 months manufacturer parts or replace
This is a preorder for the March production batch, with tentative delivery at the end of March, 2025. All sales final. No returns or cancellations. For bulk inquiries, consult a live chat agent or call our toll-free number.
✓ GPU memory: 24GB HBM2
✓ GPU memory bandwidth: 933GB/s
✓ Interconnect: PCIe Gen4: 64GB/s, Third-gen NVLINK: 200GB/s**
✓ Form factor: Dual-slot, full-height, full-length (FHFL)
✓ Max thermal design power (TDP): 165W
✓ Multi-Instance GPU (MIG): 4 GPU instances @ 6GB each, 2 GPU instances @ 12GB each, 1 GPU instance @ 24GB
✓ Virtual GPU (vGPU) software support: NVIDIA AI Enterprise, NVIDIA Virtual Compute Server
✓ Warranty: 2 year manufacturer repair or replace
Ships in 7 days from payment. No returns or cancellations. All sales final. For bulk inquiries, consult a live chat agent or call our toll-free number.
Ships in 5 days from payment. All sales final. No returns or cancellations. For bulk inquiries, consult a live chat agent or call our toll-free number.
Key Features
High density 8U system for NVIDIA® HGX™ H100/H200 8-GPU
Highest GPU communication using NVIDIA® NVLINK™ + NVIDIA® NVSwitch™
8 NIC for GPU direct RDMA (1:1 GPU Ratio)
Supports up to 24 DIMM Slots; 4800 MT/s 6TB DDR5 memory (Please check System Memory Section for detail)
Up to 8 PCIe 5.0 x16 LP + 4 PCIe 5.0 x16 FHFL slots
Flexible networking options
Total of 12 Hot-swap 2.5" NVMe drive bays + 2 hot-swap 2.5" SATA drive bays
10 heavy duty fans with optimal fan speed control
Total of 6x (3+3) 3000W Redundant Titanium Level Power Supplies
(Power supply full redundancy based on configuration and application load)
✔ FP64: 34 TFLOPS
✔ FP64 Tensor Core: 67 TFLOPS
✔ FP32: 67 TFLOPS
✔ TF32 Tensor Core²: 989 TFLOPS
✔ Architecture: Blackwell
✔ BFLOAT16 Tensor Core²: 1,979 TFLOPS
✔ FP16 Tensor Core²: 1,979 TFLOPS
✔ FP8 Tensor Core²: 3,958 TFLOPS
✔ INT8 Tensor Core²: 3,958 TFLOPS
✔ GPU Memory: 141GB
✔ GPU Memory Bandwidth: 4.8TB/s
✔ Decoders: 7 NVDEC, 7 JPEG
✔ Confidential Computing: Supported
✔ Max Thermal Design Power (TDP): Up to 600W (configurable)
✔ Multi-Instance GPUs: Up to 7 MIGs @16.5GB each
✔ Form Factor: PCIe
✔ Interconnect: 2- or 4-way NVIDIA NVLink bridge: 900GB/s, PCIe Gen5: 128GB/s
✔ Server Options: NVIDIA MGX™ H200 NVL partner and NVIDIA-Certified Systems with up to 8 GPUs
✔ NVIDIA AI Enterprise: Add-on
✔ Warranty: 3 years manufacturer parts or replace
Ships in 2-3 weeks from payment. All sales final. No returns or cancellations. For bulk inquiries, consult a live chat agent or call our toll-free number.
✔ Product SKU: ARS-111GL-NHR
✔ Motherboard: Super G1SMH-G
✔ Processor: NVIDIA 72-core Grace CPU on GH200 Grace Hopper™ Superchip
✔ Memory:
✔ Chipset: System on Chip
✔ I/O Ports:
✔ BIOS Type: AMI 64MB SPI Flash EEPROM
✔ Power Management:
✔ Power Supply: 2x 2000W Redundant Titanium Level (96%)
✔ Available: Used Condition
The Stock is 3Months Used, and available in USA for Immediate Pickup. All sales final. No returns or cancellations. For bulk inquiries, consult a live chat agent or call our toll-free number.
The H100 NVL has a full 6144-bit memory interface (1024-bit for each HBM3 stack) and memory speed up to 5.1 Gbps. This means that the maximum throughput is 7.8GB/s, more than twice as much as the H100 SXM. Large Language Models require large buffers and higher bandwidth will certainly have an impact as well.
NVIDIA H100 NVL for Large Language Model Deployment is ideal for deploying massive LLMs like ChatGPT at scale. The new H100 NVL with 96GB of memory with Transformer Engine acceleration delivers up to 12x faster inference performance at GPT-3 compared to the prior generation A100 at data center scale.
Unprecedented performance, scalability, and security for every data center. Designed for deep learning and special workloads.
Part Number: 900-21010-0300-030
The SXM4 (NVLINK native soldered onto carrier boards) version of the cards are available upon request only, and are attached permanently to their respective motherboards via a complete system only, with longer lead times.
The NVIDIA RTX™ 5000 Ada Generation GPU, powered by the NVIDIA Ada Lovelace architecture, unlocks breakthroughs in generative AI and delivers the performance required to meet the challenges of today’s professional workflows. With 100 third-generation RT Cores, 400 fourth-generation Tensor Cores, 12,800 CUDA® cores, and 32GB of graphics memory, the RTX 5000 excels in rendering, AI, graphics, and compute workload performance.