Changelog
Follow up on the latest improvements and updates.
RSS
new
improved
fixed
Release: 3/28/24
NEW
- A100's for all: We've heard your feedback and movedalltraining jobs from A10G to A100 GPUs by default!
- Faster Training Times: In addition, we've also modified the default configurations to speed up A100 training by 2x - 5x
- New Tutorial: We've added a new “Build Your Own Lora Land” Tutorial to the homepage and docs
- Mistral-7b-instruct-v0.2 now available: You can now query Mistral-7b-instruct-v0.2 as a Serverless Endpoint as well as use it as for Dedicated Deployments & Fine-Tuning
- Deployments UI: We've added the ability to seamlessly create deployments directly from the UI
- Ability to query Stopped Models: We've changed the "Cancel" operation to "Stop" during training, allowing you to use the model from the latest saved checkpoint.
New Quickstart
Create Deployments via the UI
IMPROVED
Prompt UI Improvements
- We now show all deployments (regardless of the status) in the dropdown
- We show the the status chip next to deployment name in dropdown and in status indicator
- We also provide a "Stop" Button while a response is streaming back to UI
Pricing
- We've updated the serverless pricing bucket (Up to 13B) to include larger models (Up to 21B) at the same price ($0.25 / 1k tokens)
- We've now enabled billing for serverless inference including both the streaming and non-streaming endpoints
FIXED
- Prompt UI: We now prevent streaming of multiple responses at the same time
- Models UI: We fixed the bug where 0 values didn't show in the Learning Curves
- We've fixed the hanging behavior while using the Python SDK in Colab
- We've also improved error messages in the SDK so they're concise and readable
new
improved
fixed
Release: 3/18/24
New
- New Deployments Page in UI for both dedicated deployments and serverless endpoints
- Ability to easily use REST / SDK / CLI code snippets for prompting deployments
- Ability to delete deployments via the UI
- Ability to view live events, real-time deployment statuses and configuration
- Enabled automatic early stopping by default (validation loss over 5 consecutive checkpoints)
Improved
- Docs: Added conversion script for users with OpenAI fine-tuning datasets
- Set "My Profile" as default page on Settings
Fixed
- Improved error handling around creating model repositories with the same name
new
improved
fixed
Gemma Release
New
- Gemma-2B and Gemma-7B added as Serverless Endpoints
- Add Streaming to Prompt UI
Improved
- Show Deployment Status in Prompt UI
- Add lora_rank as configurable parameter to Python SDK