Examples and Tutorials
Try out Inference on CoreWeave by following our in-depth walkthrough guides
Each tutorial page includes a link to the source code for the provided example. In most cases, it will be required to clone the repository in order to follow along directly with the tutorial's walkthrough.
Input a text prompt to generate high-quality images with photorealistic capabilities
Deploy BLOOM as an InferenceService with a simple HTTP API to perform Text Generation using Transformers Accelerate
Deploy BLOOM as an InferenceService with a simple HTTP API to perform Text Generation using Transformers DeepSpeed
Step-by-step instructions are not published for this example - the deployment process is identical to the steps used in the GPT-2 Tensorflow example, save for the provided source code
Deploy EleutherAI GPT-J and GPT-NeoX on NVIDIA Triton Inference Server with the FasterTransformer backend via an InferenceService using an HTTP API to perform Text Generation
For PyTorch, use
pipto install PyTorch bundled with CUDA. Do not install CUDA separately, as doing so may result in dependency mismatches. For more information, refer to the PyTorch installation guide.