Examples and Tutorials
Try out Inference on CoreWeave by following our in-depth walkthrough guides
Each tutorial page includes a link to the source code for the provided example. In most cases, it will be required to clone the repository in order to follow along directly with the tutorial's walkthrough.
Title | Description |
---|---|
Deploy popular open source model GPT-J-6B with the click of a button |
Title | Description |
---|---|
Input a text prompt to generate high-quality images with photorealistic capabilities | |
Deploy BLOOM as an InferenceService with a simple HTTP API to perform Text Generation using Transformers Accelerate | |
Deploy BLOOM as an InferenceService with a simple HTTP API to perform Text Generation using Transformers DeepSpeed | |
Step-by-step instructions are not published for this example - the deployment process is identical to the steps used in the GPT-2 Tensorflow example, save for the provided source code |
Deploy EleutherAI GPT-J and GPT-NeoX on NVIDIA Triton Inference Server with the FasterTransformer backend via an InferenceService using an HTTP API to perform Text Generation |
Deploy an auto-scaling Inference service from a pre-existing Docker image |
Note
For PyTorch, use
pip
to install PyTorch bundled with CUDA. Do not install CUDA separately, as doing so may result in dependency mismatches. For more information, refer to the PyTorch installation guide.Last modified 1mo ago