Skip to main content
Weights & Biases Documentation home page
Search...
⌘K
Log in
Sign Up
Sign Up
Search...
Navigation
Cloud Platforms
Azure OpenAI Fine-Tuning
Platform
W&B Models
W&B Weave
W&B Inference
W&B Training
Release Notes
W&B Models
W&B Quickstart
Get Started with W&B Models
Evaluate models
Guides
Experiments
Sweeps
Tables
Artifacts
Registry
Reports
Automations
W&B App UI
Integrations
Integrations
Add wandb to any library
ML Frameworks
ML Libraries
Cloud Platforms
SageMaker
Databricks
Azure OpenAI Fine-Tuning
OpenAI Fine-Tuning
Cohere fine-tuning
OpenAI API
NVIDIA NeMo Inference Microservice Deploy Job
Workflow Orchestration
Other
Tutorials
Tutorials
Fundamentals
Framework Tutorials
Advanced Tutorials
Integration tutorials
Reference
Reference
Python SDK 0.22.0
Command Line Interface
Query Expression Language
Reports and Workspaces API
Support
English
close
On this page
Introduction
Prerequisites
Workflow overview
1. Fine-tuning setup
2. Experiment tracking
3. Model evaluation
Real-world example
Additional resources
Integrations
Cloud Platforms
Azure OpenAI Fine-Tuning
Copy page
How to Fine-Tune Azure OpenAI models using W&B.
Copy page
Introduction
Fine-tuning GPT-3.5 or GPT-4 models on Microsoft Azure using W&B tracks, analyzes, and improves model performance by automatically capturing metrics and facilitating systematic evaluation through W&B’s experiment tracking and evaluation tools.
Prerequisites
Set up Azure OpenAI service according to
official Azure documentation
.
Configure a W&B account with an API key.
Workflow overview
1. Fine-tuning setup
Prepare training data according to Azure OpenAI requirements.
Configure the fine-tuning job in Azure OpenAI.
W&B automatically tracks the fine-tuning process, logging metrics and hyperparameters.
2. Experiment tracking
During fine-tuning, W&B captures:
Training and validation metrics
Model hyperparameters
Resource utilization
Training artifacts
3. Model evaluation
After fine-tuning, use
W&B Weave
to:
Evaluate model outputs against reference datasets
Compare performance across different fine-tuning runs
Analyze model behavior on specific test cases
Make data-driven decisions for model selection
Real-world example
Explore the
medical note generation demo
to see how this integration facilitates:
Systematic tracking of fine-tuning experiments
Model evaluation using domain-specific metrics
Go through an
interactive demo of fine-tuning a notebook
Additional resources
Azure OpenAI W&B Integration Guide
Azure OpenAI Fine-tuning Documentation
Was this page helpful?
Yes
No
Suggest edits
Raise issue
Databricks
Previous
OpenAI Fine-Tuning
Next
⌘I