Skip to main content

May 29, 2025 - NVIDIA B200 instances available

NVIDIA B200 instances are now available on CoreWeave, providing the latest in AI compute performance

Update Instances NVIDIA HGX B200 instances are now Generally Available on CoreWeave, providing next-generation AI compute capabilities. These instances deliver exceptional performance for the most demanding AI workloads, from large language model training to real-time inference.

Overview

NVIDIA HGX B200 instances are now Generally Available on CoreWeave, providing next-generation AI compute capabilities. These instances deliver exceptional performance for the most demanding AI workloads, from large language model training to real-time inference.

Key features

FeatureDescription
NVIDIA HGX B200Latest generation GPU architecture optimized for AI workloads
Next-generation AIExceptional performance for large language models and AI applications
High-performance networkingInfiniBand connectivity for distributed training
Managed infrastructureFully managed by CoreWeave with automated scaling and monitoring
Production readyGenerally Available status ensures stability and support

Performance characteristics

AI workload performance

  • Large language models: Optimized for training and fine-tuning large language models
  • Distributed training: Excellent performance for multi-GPU and multi-node training
  • Real-time inference: Low-latency inference capabilities for production workloads
  • Memory bandwidth: High memory bandwidth for memory-intensive AI operations

Compute capabilities

  • Tensor cores: Advanced tensor operations for AI workloads
  • Multi-precision support: Support for various precision formats (FP16, FP32, etc.)
  • Scalability: Efficient scaling across multiple GPUs and nodes
  • Energy efficiency: Optimized power consumption for cost-effective operations

Use cases

B200 instances are ideal for:

Use caseDescription
Large language model trainingTraining and fine-tuning large language models
Computer visionImage recognition, object detection, and video processing
Natural language processingText analysis, translation, and generation
Recommendation systemsLarge-scale recommendation and personalization
Scientific computingComputational research and data analysis
Real-time inferenceProduction inference workloads with low latency

Technical specifications

For detailed specifications of B200 instances, see B200 with InfiniBand.

Key specifications

  • GPU architecture: NVIDIA HGX B200
  • Memory: High-bandwidth memory optimized for AI workloads
  • Networking: InfiniBand connectivity for distributed workloads
  • Storage: High-performance storage options
  • Operating systems: Support for major Linux distributions

Availability and pricing

General availability

B200 instances are now Generally Available, meaning:

  • Production ready: Stable and supported for production workloads
  • Full support: Complete technical support and documentation
  • SLA guarantees: Service level agreements for reliability
  • Long-term availability: Committed to long-term availability and support

Pricing

  • Competitive pricing for next-generation AI compute
  • Flexible pricing options for different usage patterns
  • Volume discounts for large-scale deployments
  • Contact your CoreWeave account team for detailed pricing

Getting started

Deployment options

  1. CoreWeave Cloud Console: Deploy through the web interface
  2. API deployment: Use CoreWeave APIs for automated deployment
  3. Terraform: Infrastructure as code deployment with Terraform provider
  4. Kubernetes: Deploy as part of CKS clusters

Migration considerations

When migrating to B200 instances:

  • Application compatibility: Verify your applications are compatible with B200
  • Performance testing: Test performance improvements in your specific use case
  • Cost optimization: Evaluate cost vs. performance benefits
  • Gradual migration: Consider gradual migration from existing instances

Documentation

For detailed information about B200 instances and deployment options, see:

Support and resources

Technical support

  • 24/7 support: Round-the-clock technical support
  • Expert assistance: AI and GPU specialists available
  • Documentation: Comprehensive documentation and guides
  • Community: Active community for knowledge sharing

Training and resources

  • Best practices: AI workload optimization guides
  • Performance tuning: Tips for maximizing B200 performance
  • Case studies: Real-world examples and success stories
  • Webinars: Regular training sessions and updates