wandb.config: Save a dictionary of hyperparameters such as learning rate or model type. The model settings you capture in config are useful later to organize and query your results.
wandb.log(): Log metrics over time in a training loop, such as accuracy and loss. By default, when you call wandb.log() it appends a new step to the history object and updates the summary object.
history: An array of dictionary-like objects that tracks metrics over time. These time series values are shown as default line plots in the UI.
summary: By default, the final value of a metric logged with wandb.log(). You can set the summary for a metric manually to capture the highest accuracy or lowest loss instead of the final value. These values are used in the table, and plots that compare runs — for example, you could visualize at the final accuracy for all runs in your project.
Artifacts: Save outputs of a run, like the model weights or a table of predictions. This lets you track not just model training, but all the pipeline steps that affect the final model.
wandb library is incredibly flexible. Here are some suggested guidelines.
Config: Track hyperparameters, architecture, dataset, and anything else you'd like to use to reproduce your model. These will show up in columns— use config columns to group, sort, and filter runs dynamically in the app.
Project: A project is a set of experiments you can compare together. Each project gets a dedicated dashboard page, and you can easily turn on and off different groups of runs to compare different model versions.
Notes: A quick commit message to yourself, the note can be set from your script and is editable in the table.
Tags: Identify baseline runs and favorite runs. You can filter runs using tags, and they're editable in the table.
import wandbconfig = dict (learning_rate = 0.01,momentum = 0.2,architecture = "CNN",dataset_id = "peds-0192",infra = "AWS",)wandb.init(project="detect-pedestrians",notes="tweak baseline",tags=["baseline", "paper1"],config=config,)
All the data logged from your script is saved locally to your machine in a wandb directory, then sync'd to the W&B cloud or your private server.
System metrics: CPU and GPU utilization, network, etc. These come from nvidia-smi and are shown in the System tab on the run page.
Command line: The stdout and stderr are picked up and show in the logs tab on the run page.
Git commit: Pick up the latest git commit and see it on the overview tab of the run page, as well as a diff.patch file if there are any uncommitted changes.
requirements.txt file will be uploaded and shown on the files tab of the run page, along with any files you save to the wandb directory for the run.
Where data and model metrics are concerned, you get to decide exactly what you want to log.
Dataset: You have to specifically log images or other dataset samples for them to stream to W&B.
PyTorch gradients: Add
wandb.watch(model) to see gradients of the weights as histograms in the UI.
Config: Log hyperparameters, a link to your dataset, or the name of the architecture you're using as config parameters, passed in like this:
wandb.log() to see metrics from your model. If you log metrics like accuracy and loss from inside your training loop, you'll get live updating graphs in the UI.