Code Monkey home page Code Monkey logo

examples's Introduction

Weights & Biases Weights & Biases

Use W&B to build better models faster. Track and visualize all the pieces of your machine learning pipeline, from datasets to production machine learning models. Get started with W&B today, sign up for a free account!

๐Ÿš€ Getting Started

Never lose your progress again.

Save everything you need to compare and reproduce models โ€” architecture, hyperparameters, weights, model predictions, GPU usage, git commits, and even datasets โ€” in 5 minutes. W&B is free for personal use and academic projects, and it's easy to get started.

Check out our libraries of example scripts and example colabs or read on for code snippets and more!

If you have any questions, please don't hesitate to ask in our Discourse forum.

๐Ÿค Simple integration with any framework

Install wandb library and login:

pip install wandb
wandb login

Flexible integration for any Python script:

import wandb

# 1. Start a W&B run
wandb.init(project='gpt3')

# 2. Save model inputs and hyperparameters
config = wandb.config
config.learning_rate = 0.01

# Model training code here ...

# 3. Log metrics over time to visualize performance
for i in range (10):
    wandb.log({"loss": loss})

If you have any questions, please don't hesitate to ask in our Discourse forum.

Explore a W&B dashboard

๐Ÿ“ˆ Track model and data pipeline hyperparameters

Set wandb.config once at the beginning of your script to save your hyperparameters, input settings (like dataset name or model type), and any other independent variables for your experiments. This is useful for analyzing your experiments and reproducing your work in the future. Setting configs also allows you to visualize the relationships between features of your model architecture or data pipeline and the model performance (as seen in the screenshot above).

wandb.init()
wandb.config.epochs = 4
wandb.config.batch_size = 32
wandb.config.learning_rate = 0.001
wandb.config.architecture = "resnet"

๐Ÿ— Use your favorite framework

Use your favorite framework with W&B. W&B integrations make it fast and easy to set up experiment tracking and data versioning inside existing projects. For more information on how to integrate W&B with the framework of your choice, see the Integrations chapter in the W&B Developer Guide.

๐Ÿ”ฅ PyTorch

Call .watch and pass in your PyTorch model to automatically log gradients and store the network topology. Next, use .log to track other metrics. The following example demonstrates an example of how to do this:

import wandb

# 1. Start a new run
run = wandb.init(project="gpt4")

# 2. Save model inputs and hyperparameters
config = run.config
config.dropout = 0.01

# 3. Log gradients and model parameters
run.watch(model)
for batch_idx, (data, target) in enumerate(train_loader):
    ...
    if batch_idx % args.log_interval == 0:
        # 4. Log metrics to visualize performance
        run.log({"loss": loss})
๐ŸŒŠ TensorFlow/Keras Use W&B Callbacks to automatically save metrics to W&B when you call `model.fit` during training.

The following code example demonstrates how your script might look like when you integrate W&B with Keras:

# This script needs these libraries to be installed:
#   tensorflow, numpy

import wandb
from wandb.keras import WandbMetricsLogger, WandbModelCheckpoint

import random
import numpy as np
import tensorflow as tf


# Start a run, tracking hyperparameters
run = wandb.init(
    # set the wandb project where this run will be logged
    project="my-awesome-project",
    # track hyperparameters and run metadata with wandb.config
    config={
        "layer_1": 512,
        "activation_1": "relu",
        "dropout": random.uniform(0.01, 0.80),
        "layer_2": 10,
        "activation_2": "softmax",
        "optimizer": "sgd",
        "loss": "sparse_categorical_crossentropy",
        "metric": "accuracy",
        "epoch": 8,
        "batch_size": 256,
    },
)

# [optional] use wandb.config as your config
config = run.config

# get the data
mnist = tf.keras.datasets.mnist
(x_train, y_train), (x_test, y_test) = mnist.load_data()
x_train, x_test = x_train / 255.0, x_test / 255.0
x_train, y_train = x_train[::5], y_train[::5]
x_test, y_test = x_test[::20], y_test[::20]
labels = [str(digit) for digit in range(np.max(y_train) + 1)]

# build a model
model = tf.keras.models.Sequential(
    [
        tf.keras.layers.Flatten(input_shape=(28, 28)),
        tf.keras.layers.Dense(config.layer_1, activation=config.activation_1),
        tf.keras.layers.Dropout(config.dropout),
        tf.keras.layers.Dense(config.layer_2, activation=config.activation_2),
    ]
)

# compile the model
model.compile(optimizer=config.optimizer, loss=config.loss, metrics=[config.metric])

# WandbMetricsLogger will log train and validation metrics to wandb
# WandbModelCheckpoint will upload model checkpoints to wandb
history = model.fit(
    x=x_train,
    y=y_train,
    epochs=config.epoch,
    batch_size=config.batch_size,
    validation_data=(x_test, y_test),
    callbacks=[
        WandbMetricsLogger(log_freq=5),
        WandbModelCheckpoint("models"),
    ],
)

# [optional] finish the wandb run, necessary in notebooks
run.finish()

Get started integrating your Keras model with W&B today:

๐Ÿค— Huggingface Transformers

Pass wandb to the report_to argument when you run a script using a HuggingFace Trainer. W&B will automatically log losses, evaluation metrics, model topology, and gradients.

Note: The environment you run your script in must have wandb installed.

The following example demonstrates how to integrate W&B with Hugging Face:

# This script needs these libraries to be installed:
#   numpy, transformers, datasets

import wandb

import os
import numpy as np
from datasets import load_dataset
from transformers import TrainingArguments, Trainer
from transformers import AutoTokenizer, AutoModelForSequenceClassification


def tokenize_function(examples):
    return tokenizer(examples["text"], padding="max_length", truncation=True)


def compute_metrics(eval_pred):
    logits, labels = eval_pred
    predictions = np.argmax(logits, axis=-1)
    return {"accuracy": np.mean(predictions == labels)}


# download prepare the data
dataset = load_dataset("yelp_review_full")
tokenizer = AutoTokenizer.from_pretrained("distilbert-base-uncased")

small_train_dataset = dataset["train"].shuffle(seed=42).select(range(1000))
small_eval_dataset = dataset["test"].shuffle(seed=42).select(range(300))

small_train_dataset = small_train_dataset.map(tokenize_function, batched=True)
small_eval_dataset = small_train_dataset.map(tokenize_function, batched=True)

# download the model
model = AutoModelForSequenceClassification.from_pretrained(
    "distilbert-base-uncased", num_labels=5
)

# set the wandb project where this run will be logged
os.environ["WANDB_PROJECT"] = "my-awesome-project"

# save your trained model checkpoint to wandb
os.environ["WANDB_LOG_MODEL"] = "true"

# turn off watch to log faster
os.environ["WANDB_WATCH"] = "false"

# pass "wandb" to the `report_to` parameter to turn on wandb logging
training_args = TrainingArguments(
    output_dir="models",
    report_to="wandb",
    logging_steps=5,
    per_device_train_batch_size=32,
    per_device_eval_batch_size=32,
    evaluation_strategy="steps",
    eval_steps=20,
    max_steps=100,
    save_steps=100,
)

# define the trainer and start training
trainer = Trainer(
    model=model,
    args=training_args,
    train_dataset=small_train_dataset,
    eval_dataset=small_eval_dataset,
    compute_metrics=compute_metrics,
)
trainer.train()

# [optional] finish the wandb run, necessary in notebooks
wandb.finish()
โšก๏ธ PyTorch Lightning

Build scalable, structured, high-performance PyTorch models with Lightning and log them with W&B.

# This script needs these libraries to be installed:
#   torch, torchvision, pytorch_lightning

import wandb

import os
from torch import optim, nn, utils
from torchvision.datasets import MNIST
from torchvision.transforms import ToTensor

import pytorch_lightning as pl
from pytorch_lightning.loggers import WandbLogger


class LitAutoEncoder(pl.LightningModule):
    def __init__(self, lr=1e-3, inp_size=28, optimizer="Adam"):
        super().__init__()

        self.encoder = nn.Sequential(
            nn.Linear(inp_size * inp_size, 64), nn.ReLU(), nn.Linear(64, 3)
        )
        self.decoder = nn.Sequential(
            nn.Linear(3, 64), nn.ReLU(), nn.Linear(64, inp_size * inp_size)
        )
        self.lr = lr

        # save hyperparameters to self.hparamsm auto-logged by wandb
        self.save_hyperparameters()

    def training_step(self, batch, batch_idx):
        x, y = batch
        x = x.view(x.size(0), -1)
        z = self.encoder(x)
        x_hat = self.decoder(z)
        loss = nn.functional.mse_loss(x_hat, x)

        # log metrics to wandb
        self.log("train_loss", loss)
        return loss

    def configure_optimizers(self):
        optimizer = optim.Adam(self.parameters(), lr=self.lr)
        return optimizer


# init the autoencoder
autoencoder = LitAutoEncoder(lr=1e-3, inp_size=28)

# setup data
batch_size = 32
dataset = MNIST(os.getcwd(), download=True, transform=ToTensor())
train_loader = utils.data.DataLoader(dataset, shuffle=True)

# initialise the wandb logger and name your wandb project
wandb_logger = WandbLogger(project="my-awesome-project")

# add your batch size to the wandb config
wandb_logger.experiment.config["batch_size"] = batch_size

# pass wandb_logger to the Trainer
trainer = pl.Trainer(limit_train_batches=750, max_epochs=5, logger=wandb_logger)

# train the model
trainer.fit(model=autoencoder, train_dataloaders=train_loader)

# [optional] finish the wandb run, necessary in notebooks
wandb.finish()
๐Ÿ’จ XGBoost Use W&B Callbacks to automatically save metrics to W&B when you call `model.fit` during training.

The following code example demonstrates how your script might look like when you integrate W&B with XGBoost:

# This script needs these libraries to be installed:
#   numpy, xgboost

import wandb
from wandb.xgboost import WandbCallback

import numpy as np
import xgboost as xgb


# setup parameters for xgboost
param = {
    "objective": "multi:softmax",
    "eta": 0.1,
    "max_depth": 6,
    "nthread": 4,
    "num_class": 6,
}

# start a new wandb run to track this script
run = wandb.init(
    # set the wandb project where this run will be logged
    project="my-awesome-project",
    # track hyperparameters and run metadata
    config=param,
)

# download data from wandb Artifacts and prep data
run.use_artifact("wandb/intro/dermatology_data:v0", type="dataset").download(".")
data = np.loadtxt(
    "./dermatology.data",
    delimiter=",",
    converters={33: lambda x: int(x == "?"), 34: lambda x: int(x) - 1},
)
sz = data.shape

train = data[: int(sz[0] * 0.7), :]
test = data[int(sz[0] * 0.7) :, :]

train_X = train[:, :33]
train_Y = train[:, 34]

test_X = test[:, :33]
test_Y = test[:, 34]

xg_train = xgb.DMatrix(train_X, label=train_Y)
xg_test = xgb.DMatrix(test_X, label=test_Y)
watchlist = [(xg_train, "train"), (xg_test, "test")]

# add another config to the wandb run
num_round = 5
run.config["num_round"] = 5
run.config["data_shape"] = sz

# pass WandbCallback to the booster to log its configs and metrics
bst = xgb.train(
    param, xg_train, num_round, evals=watchlist, callbacks=[WandbCallback()]
)

# get prediction
pred = bst.predict(xg_test)
error_rate = np.sum(pred != test_Y) / test_Y.shape[0]

# log your test metric to wandb
run.summary["Error Rate"] = error_rate

# [optional] finish the wandb run, necessary in notebooks
run.finish()
๐Ÿงฎ Sci-Kit Learn Use wandb to visualize and compare your scikit-learn models' performance:
# This script needs these libraries to be installed:
#   numpy, sklearn

import wandb
from wandb.sklearn import plot_precision_recall, plot_feature_importances
from wandb.sklearn import plot_class_proportions, plot_learning_curve, plot_roc

import numpy as np
from sklearn import datasets
from sklearn.ensemble import RandomForestClassifier
from sklearn.model_selection import train_test_split


# load and process data
wbcd = datasets.load_breast_cancer()
feature_names = wbcd.feature_names
labels = wbcd.target_names

test_size = 0.2
X_train, X_test, y_train, y_test = train_test_split(
    wbcd.data, wbcd.target, test_size=test_size
)

# train model
model = RandomForestClassifier()
model.fit(X_train, y_train)
model_params = model.get_params()

# get predictions
y_pred = model.predict(X_test)
y_probas = model.predict_proba(X_test)
importances = model.feature_importances_
indices = np.argsort(importances)[::-1]

# start a new wandb run and add your model hyperparameters
run = wandb.init(project="my-awesome-project", config=model_params)

# Add additional configs to wandb
run.config.update(
    {
        "test_size": test_size,
        "train_len": len(X_train),
        "test_len": len(X_test),
    }
)

# log additional visualisations to wandb
plot_class_proportions(y_train, y_test, labels)
plot_learning_curve(model, X_train, y_train)
plot_roc(y_test, y_probas, labels)
plot_precision_recall(y_test, y_probas, labels)
plot_feature_importances(model)

# [optional] finish the wandb run, necessary in notebooks
run.finish()

ย 

๐Ÿงน Optimize hyperparameters with Sweeps

Use Weights & Biases Sweeps to automate hyperparameter optimization and explore the space of possible models.

Benefits of using W&B Sweeps

  • Quick to setup: With just a few lines of code you can run W&B sweeps.
  • Transparent: We cite all the algorithms we're using, and our code is open source.
  • Powerful: Our sweeps are completely customizable and configurable. You can launch a sweep across dozens of machines, and it's just as easy as starting a sweep on your laptop.

Weights & Biases

Common use cases

  • Explore: Efficiently sample the space of hyperparameter combinations to discover promising regions and build an intuition about your model.
  • Optimize: Use sweeps to find a set of hyperparameters with optimal performance.
  • K-fold cross validation: Here's a brief code example of k-fold cross validation with W&B Sweeps.

Visualize Sweeps results

The hyperparameter importance plot surfaces which hyperparameters were the best predictors of, and highly correlated to desirable values for your metrics.

Weights & Biases

Parallel coordinates plots map hyperparameter values to model metrics. They're useful for honing in on combinations of hyperparameters that led to the best model performance.

Weights & Biases

๐Ÿ“œ Share insights with Reports

Reports let you organize visualizations, describe your findings, and share updates with collaborators.

Common use cases

  • Notes: Add a graph with a quick note to yourself.
  • Collaboration: Share findings with your colleagues.
  • Work log: Track what you've tried and plan next steps.

Explore reports in The Gallery โ†’ | Read the Docs

Once you have experiments in W&B, you can visualize and document results in Reports with just a few clicks. Here's a quick demo video.

๐Ÿบ Version control datasets and models with Artifacts

Git and GitHub make code version control easy, but they're not optimized for tracking the other parts of the ML pipeline: datasets, models, and other large binary files.

W&B's Artifacts are. With just a few extra lines of code, you can start tracking you and your team's outputs, all directly linked to run.

Try Artifacts in a Colab with a video tutorial

Common use cases

  • Pipeline Management: Track and visualize the inputs and outputs of your runs as a graph
  • Don't Repeat Yourselfโ„ข: Prevent the duplication of compute effort
  • Sharing Data in Teams: Collaborate on models and datasets without all the headaches

Learn about Artifacts here โ†’ | Read the Docs

Visualize and Query data with Tables

Group, sort, filter, generate calculated columns, and create charts from tabular data.

Spend more time deriving insights, and less time building charts manually.

# log my table

wandb.log({"table": my_dataframe})

Try Tables in a Colab or these examples

Explore Tables here โ†’ | Read the Docs

examples's People

Contributors

adrianbg avatar amarcodes-22 avatar anmolmann avatar ash0ts avatar ayulockin avatar bcsherma avatar borisdayma avatar charlesfrye avatar cvphelps avatar dependabot[bot] avatar dmitryduev avatar gabesmed avatar ivangrov avatar jsbroks avatar kenleejr avatar lavanyashukla avatar lukas avatar manangoel99 avatar morganmcg1 avatar ngrayluna avatar parambharat avatar raubitsj avatar sauravmaheshkar avatar scottire avatar shawnlewis avatar soumik12345 avatar staceysv avatar tcapelle avatar vanpelt avatar vwxyzjn avatar

Stargazers

 avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar

Watchers

 avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar

examples's Issues

turn off logging or using wandb when testing new code/debugging code

is there a way to fully remove using wandb when testing new code?

os.environ['WANDB_MODE'] = 'offline'
wandb.config = {
  "learning_rate": 1e-3,
  "epochs": 50,
  "batch_size": 8
}

I have this in my main.py right after import however, it still logs into my wandb account and tries to log.
https://docs.wandb.ai/guides/track/advanced/environment-variables


wandb: Currently logged in as: monajalal (use `wandb login --relogin` to force relogin)
wandb: Tracking run with wandb version 0.12.6
wandb: Syncing run firm-spaceship-87
wandb:  View project at https://wandb.ai/my-test-project/XYZ
wandb:  View run at https://wandb.ai/my-test-project/XYZ
wandb: Run data is saved locally in /XYZ
wandb: Run `wandb offline` to turn off syncing.

error by my code

wandb: Waiting for W&B process to finish, PID 30803... (failed 1). Press ctrl-c to abort syncing.
wandb:                                                                                
wandb: Synced 5 W&B file(s), 0 media file(s), 0 artifact file(s) and 0 other file(s)
wandb: Synced firm-spaceship-87: https://wandb.ai/XYZ
wandb: Find logs at: ./wandb/run-20211127_011314-3ukc1p0e/logs/debug.log
wandb: 

Can we change the Sweep Configuration after creating a sweep

Hi, wandb team, first of all, thank your team for providing such a great tool, i have a question when i am using wandb sweep, that is can we change the Sweep Configuration after creating a sweep, for example, i found the batch size of 64 is too big for my GPU that will lead to out of memory, so i want delete this value so that next run will not choose this value, but i didn't find a edit button.
image

By the way, i have a feature request, could you add a new distribution that will increase the number, something like range(start, [stop, step]), i want to use this parameter to set save directory, or the next run's result will override the previous one, or is there existing distribution can do this ?

Why isn't wandb logging at the last iteration? (when I want to log figs, ckpts etc)

I was running my experiments in pycharm and noticed that the last iteration/epoch I do is not logged for some reason (until I close the pychar debugger) + even when I close it my figure is not being logged for some reason...

The code is really trivial, and I saved the run + report for help of reproducibility:

https://wandb.ai/brando/playground/runs/wpupxvg1?workspace=user-brando
https://wandb.ai/brando/playground/reports/Untitled-Report--VmlldzoxMDEzNjAw

Results for HPO examples

Is it possible to post the results of performance of the different hyperparameter optimization experiments? Specifically, I am interested in pytorch cifar10, fashion and mnist experiments. The results could be the loss plots over time, the final accuracy received after specified budget, etc.

How to use sweep on a machine with multiple gpu

Thanks for your great work. I knew that sweep can allocate tasks to different machine. But how to use sweep on a machine with multiple gpus? I want to train model with different hyperparameters on different gpu to take fully use of them. Thanks for your help.

Single value in yaml not passed as argument

Hi,

since yesterday night when performing wandb sweep parameters with a single value such as epochs here are not passed as argument to the process.

2021-08-19 10:11:38,946 - wandb.wandb_agent - INFO - About to run command: /usr/bin/env python train.py --channels_one=18 --channels_two=28 --dropout=0.5 --learning_rate=0.005

I think this is a server-side change that has happened recently.

[Bug] Optimize Pytorch-Lightning models with Weights & Biases.ipynb is not running on colab

What happened?

  • The notebooks does not run in colab, produces an error related with deterministic behaviour being enabled.
  • The sweep also appears to have issues, it actually works but produces multiple errors:
Run w4g08tyl errored: ValueError('signal only works in main thread')
wandb: ERROR Run w4g08tyl errored: ValueError('signal only works in main thread')

Maybe this is not lightning specific, I am looking it!
@borisdayma would you mind checking this?

How to save model with estimator?

I am now following the following example guide with wandb tensorflow example.

# Init wandb
import wandb
from wandb.tensorflow import WandbHook
wandb.init(project="test", sync_tensorboard=True)
 Model instantiation code ...
# Log metrics with wandb
# If you are using tensorboard already, we will sync all the information from tensorboard and this isn't necessary.
# If you are using an estimator you can add our hook.
classifier.train(input_fn=train_input_fn, steps=100000, hooks=[WandbHook()])
# Save model to wandb
saver = tf.train.Saver()
saver.save(sess, os.path.join(wandb.run.dir, "model.ckpt"))

I found that I want to use the saver I need to get the sess variable.

But in my current training, I used tf.estimator.train_and_evaluate(estimator, train_spec, eval_spec) for training with estimator.

Now if I cannot get session, how to save the model in wandb?

Thank you.

We should incorporate the Colabs into this repository

Rationale

The Colabs are an important part of the purpose of this repo: allowing users to get a sense for how wandb works in their use case.

For that reason, we should incorporate them and control their versions. Not least so that I can stop bugging @ayulockin with issues and start bugging him with PRs.

I will open a PR shortly that demonstrates how this would work for a single Colab, but as this is a big change, I wanted to track it with an issue as well.

Process

From scratch, the process operates as follows:

  1. A Colab author writes new material in a Colab hosted on their personal Drive. They iterate until it is ready to be committed.
  2. For notebooks on which we want to preserve outputs, they restart the kernel (resetting to factory preferred, to check for errors in downloading/installation) and re-run the notebook. See below for more notes on VCS best practices with notebooks.
  3. They click "File > Save a copy in GitHub" in the toolbar.
  4. The author then sets the repository, branch, and file to commit, using the dialog box pictured below. Ideally, new branches are created and merged through PRs. New branches can be made completely within GitHub, so just click the link in the dialog box. You may need to re-open the dialog box once the branch has been created. The file name should be whatever is automatically generated by Colab based on the title (sans Copy_of_ prefix). The path should begin colabs/{identifier/}?, where the {identifer} directories help organize the Colabs, as in the examples directory. The ? indicates that one or more identifiers can be used. For Colabs, it's probably only necessary to have one, since each example only has a single file (the notebook), whereas the examples tend to have entire folders.
  5. The "include a link to Colaboratory" box should always be ticked. This results in the creation of a button at the top of the notebook stored in GitHub that can be used to directly open up the notebook in Colab. It appears below.
  6. The author then writes a commit message and clicks OK. Note that you will not be able to view a diff unless you do a PR, at which point the ReviewNB tool will pop up.
  7. If making a PR, the author then does so via the GitHub interface.

For changes to existing notebooks, we enter at step 2, by clicking the badge: Open In Colab.

Best Practices for Version Control in Notebooks

I've VC'd two large repos of notebooks: one well and one poorly. Here's what I've learned in the process:

  1. Avoid including outputs whenever possible. Outputs, especially large tables and binaries (eg media), add weight to the diffs. Only include them when they're an important part of the reader experience (eg wandb outputs). Examples: Use %%capture on installation cells. Present media as screenshots in Markdown, rather than using IPython to render them, as these are less likely to change. Use the "private outputs" setting if possible.
  2. Reduce volatility in outputs. When outputs vary from execution to execution, they add noise to the diff. Examples: Instead of logging a pseudo-random data value, log a specific value. Instead of printing a value, assert something is True.
  3. Always fix the random seed. This is a specific instance of the above, but it's important enough to bear repeating. Note that there may be more than one PRNG (eg python and TF), so make sure to seed all of the relevant PRNGs. Double-check that your outputs don't vary! Examples: random.set_seed(42); tf.random.set_seed(117). GPUs are not always deterministic; see this StackOverflow post.
  4. Design your code to be run linearly. That is, treat the code cells as though they were going into a script, to be executed in order. Otherwise, step 2 above has to be done manually, rather than just by clicking "re-run all cells". In the future, it might even be automated. Examples: Copy code and make changes, rather than directing users to make changes.
  5. Never throw an error. This is again an important special case: if a code cell throws an error, execution will stop. This is especially important for future automatibility. Examples: Use try and except rather than showing an uncaught Error.
  6. Pin versions. As versions change, DeprecationWarnings appear, behavior slightly changes, etc. If you include a pip install step for anything other than wandb, always pin the version. Unfortunately, the Colab environment is out of our control, so changes to the environment will happen anyway. Examples: %%capture\n pip install -qq numpy==1.16.
  7. Be mindful of iteration time. This reduces the burden of re-executing the notebook and often improves the reader's experience. Remember that the point of the examples isn't to achieve SOTA performance. Examples: reduce the length of Sweeps (and always set a length, due to principle 4). Subsample datasets by batch (e.g. x[::5], y[::5]) and work with the smallest reasonable dataset.

Here's a snippet that can be used to get deterministic behavior in TF:

# Ensure deterministic behavior
os.environ['TF_CUDNN_DETERMINISTIC'] = '1' 
random.seed(hash("setting random seeds") % 2**32 - 1)
np.random.seed(hash("improves reproducibility") % 2**32 - 1)
tf.random.set_seed(hash("by removing stochasticity") % 2**32 - 1)

and one in PyTorch:

# Ensure deterministic behavior
torch.backends.cudnn.deterministic = True
random.seed(hash("setting random seeds") % 2**32 - 1)
np.random.seed(hash("improves reproducibility") % 2**32 - 1)
torch.manual_seed(hash("by removing stochasticity") % 2**32 - 1)
torch.cuda.manual_seed_all(hash("so runs are repeatable") % 2**32 - 1)

Appendix: Dialog Box on Colab

image

fastai integration does not work

Hi, fastai how not work with error:


ModuleNotFoundError                       Traceback (most recent call last)
<ipython-input-8-d0cab4097689> in <module>
----> 1 import wandb.fastai

~/anaconda3/envs/ab/lib/python3.7/site-packages/wandb/fastai/__init__.py in <module>
      6 """
      7 
----> 8 from wandb.integration.fastai import WandbCallback
      9 
     10 __all__ = ['WandbCallback']

~/anaconda3/envs/ab/lib/python3.7/site-packages/wandb/integration/fastai/__init__.py in <module>
     38 import wandb
     39 import fastai
---> 40 from fastai.callbacks import TrackerCallback
     41 from pathlib import Path
     42 import random

ModuleNotFoundError: No module named 'fastai.callbacks'

They have recently released 2.0, that most probably causes problems with the callback

How to run sweep with a train.py code that receives a YAML file as argument

Hello!
I've been trying to use sweep in my project where the train.py accepts a YAML file as argument, where all the hyperparameters are defined. This is an example how I run my train.py file:
python3 train.py config_yaml.yaml

How can I configure my sweep configuration this way? If for example I have sweep_config.yaml that contains:

program: train.py
method: grid
parameters:
  hidden_layers:
    values: [1024, 2048]
  optimizer:
    values: ["adam", "sgd"]

I can't run wandb sweep sweep_config.yaml because wandb will translate this to be something like python3 train.py --hidden_layers=1024 --optimizer="adam".
Another question is that, if I define default hyperparams values in the code, can I overwrite it with the values specified in the config file?

Thank you.

Where to place my dataset (split)?

Hello! This repository helped me enormously in realizing that there is potential to use sweeps with k-Fold cross-validation (grouping each split with the same hyperparameters). However, I am very confused on how to adapt my code for this. For example, in the repository code there is no reference to where to place my dataset that will be divided into the folds. Thank you very much!

WandB โ€“ Install the W&B library

%pip install wandb -q
import wandb
from wandb.keras import WandbCallback

!pip install wandb -qq
import numpy as np
import pandas as pd
import wandb
from wandb.keras import WandbCallback
from sklearn.model_selection import train_test_split
from sklearn.preprocessing import StandardScaler
from keras.models import Sequential
from keras.layers import Dense

dataset = pd.read_excel("Dados gerados Bateria r2.xlsx", "Trial 2", index_col=None, na_values=["NA"], engine="openpyxl")
dataset = dataset.iloc[:,0:11]
dataset['STATUS_bin'] = dataset['STATUS'].apply(lambda x: 1 if x == "OK" else 0)
X = dataset.iloc[:,1:10].values
y = dataset.iloc[:,11].values

X_train, X_test, y_train, y_test = train_test_split(X, y, test_size=0.3)
sc = StandardScaler()
X_train = sc.fit_transform(X_train)
X_test = sc.transform(X_test)

Configure the sweep โ€“ specify the parameters to search through, the search strategy, the optimization metric et all.

sweep_config = {
'method': 'bayes',
'metric': {
'name': 'val_accuracy',
'goal': 'maximize'
},
'parameters': {
'epoch': {
'distribution': 'int_uniform',
'max': 250,
'min': 1
},
'activation_1': {
'distribution': 'categorical',
'values':['linear', 'relu', 'sigmoid']
},
'activation_2': {
'distribution': 'categorical',
'values':['linear', 'relu', 'sigmoid']
},
'activation_3': {
'distribution': 'categorical',
'values':['linear', 'relu', 'sigmoid']
},
'activation_4': {
'distribution': 'categorical',
'values':['linear', 'relu', 'sigmoid']
},
'activation_5': {
'distribution': 'categorical',
'values':['linear', 'relu', 'sigmoid']
},
'n_layers': {
'distribution': 'int_uniform',
'max': 5,
'min': 1
},
'layer_1': {
'distribution': 'int_uniform',
'max': 100,
'min': 1
},
'layer_2': {
'distribution': 'int_uniform',
'max': 100,
'min': 1
},
'layer_3': {
'distribution': 'int_uniform',
'max': 100,
'min': 1
},
'layer_4': {
'distribution': 'int_uniform',
'max': 100,
'min': 1
},
'layer_5': {
'distribution': 'int_uniform',
'max': 100,
'min': 1
},
'optimizer': {
'distribution': 'categorical',
'values': ['adam', 'SGD', 'Adadelta']
},

}

}

Initialize a new sweep

sweep_id = wandb.sweep(sweep_config, entity="thiagoribeiro1", project="sweeps-VentilAQUA_4_Bateria2")

def train():
config_defaults = {
"n_layers": 1,
"layer_1": 1,
"activation_1": "linear",
#"dropout": 0.3,
"layer_2": 1,
"activation_2": "linear",
"layer_3": 1,
"activation_3": "linear",
"layer_4": 1,
"activation_4": "linear",
"layer_5": 1,
"activation_5": "linear",
"optimizer": "adam",
#"loss": "binary_crossentropy",
#"metric": "accuracy",
"epoch": 100
#"batch_size": 32

}

Initialize a new wandb run

wandb.init(config=config_defaults)

Config is a variable that holds and saves hyperparameters and inputs

config = wandb.config

model = Sequential()

if config.n_layers==1:
model.add(Dense(units=config.layer_1, activation=config.activation_1))
if config.n_layers==2:
model.add(Dense(units=config.layer_1, activation=config.activation_1))
model.add(Dense(units=config.layer_2, activation=config.activation_2))
if config.n_layers==3:
model.add(Dense(units=config.layer_1, activation=config.activation_1))
model.add(Dense(units=config.layer_2, activation=config.activation_2))
model.add(Dense(units=config.layer_3, activation=config.activation_3))
if config.n_layers==4:
model.add(Dense(units=config.layer_1, activation=config.activation_1))
model.add(Dense(units=config.layer_2, activation=config.activation_2))
model.add(Dense(units=config.layer_3, activation=config.activation_3))
model.add(Dense(units=config.layer_4, activation=config.activation_4))
if config.n_layers==5:
model.add(Dense(units=config.layer_1, activation=config.activation_1))
model.add(Dense(units=config.layer_2, activation=config.activation_2))
model.add(Dense(units=config.layer_3, activation=config.activation_3))
model.add(Dense(units=config.layer_4, activation=config.activation_4))
model.add(Dense(units=config.layer_5, activation=config.activation_5))

model.add(Dense(1, activation="sigmoid"))

model.compile(loss='binary_crossentropy', optimizer=config.optimizer, metrics=['accuracy', 'AUC'])

model.fit(X_train, y_train, batch_size=64, epochs=config.epoch, validation_data=(X_test,y_test), callbacks=[WandbCallback()])

Initialize a new sweep

wandb.agent(sweep_id, train)

Cross validation

The example on cross-validation is a little tricky to understand.
These were my key takeaways:

  • Upon every iteration of the train-cross-validation.py there is a sweep_run that gets initialised.
  • After initializing the sweep_run there are num number of processes created for each fold.
  • The runs inside the processes are categorized under a Job-Type. These runs log the accuracy of the respective folds.
  • After all the processes stop (k-folds done) the sweep_run logs the average accuracy of all the individual folds.
  • Sweeps tries to optimise the parameters with the average_accuracy.

Let me know if there is something that I missed.
My issue is that, the individual runs for each process (k-folds) gets overwritten in the UI. I am not sure as to why this happens. This might be because of the process .join() or wandb.join() method. I have also tried with wandb.finish().

Using public api delete function doesnt delete all files

I am using wandb to log the output of my experiments. This unfortunately creates a large number of files on the wandb server. Since they are image files & the log files of the experiment, they soon cross the free limit of 100GB. As such I am trying to delete them using the sample provided via your documentation.

run_name = "run_name"

run = api.run("<org_name>/<project_name>/" + run_name)

files = run.files()
for file in tqdm(files):
  if file.name.endswith('.png'):
    file.delete()

All files are not deleted in a single run. I believe this is because of some time-outs in the api. Is it possible to send that as an error, so tht I can place a sleep routine & then call the api again. Right now its simply ending the loop w/o any errors.

I have shared this with your support team. Though they have tried to support my case, it has not resulted in any positive outcome. Since I still have to run the same code block number of times.

wandb_log.log

DDP example is not calling .finish in either log_all nor log with lead worker (rank0)

I was looking when to call .finish to make sure the wandb is not hanging when my code crashes etc.

When do we call wandb.finish? Since using a context manager doesn't seem to make sense to me in a DDP case...

DDP example I am refering to: https://github.com/wandb/examples/blob/master/examples/pytorch/pytorch-ddp/log-ddp.py


related:

Data hand-off

I have a private wandb project with hundreds of runs. I would like to hand off that data to another user i.e. transfer "ownership" of the project to him. Is this a supported use case? Is it supported for free accounts?

To be clear this is a one time hand-off, not an ongoing collaboration. Whether I retain a copy of the data or not is also irrelevant to me. The closes alternative I've found is downloading a csv with the data for all runs then transferring that file, but I don't think the other user can then make that into a wandb project of his own

Thanks for your time!

Save simple plain-text

Hi! I'm working on a text generation experiment.
I'm logging loss and accuracy and it's fine, I love wandb!
But once finished the training I would like to 'log' a sample o generated text and be able to see on the wandb UI.
I tried with wandb.log({"generated": txt_gen}) but on the web UI i still see "No data submitted for tag generated".

'wandb' in YOLOV4

Hi everyone!

Can anyone help me to know that How can we train custom model using 'wandb' feature in YOLOV4-tiny? We can easily use 'wandb' in YOLOV5 but how to integrate 'wandb' in YOLOV4?

Colab wandb.sweep() no attribute '_jupyter_login'

Looking at examples for running sweeps in Google Colab and it throws the error below when running the standard initialization from https://docs.wandb.com/sweeps/python-api without calling wandb.login() first. The same goes for other wandb examples in Colab. A suggestion to update the documentation to include wandb.login()

/usr/local/lib/python3.6/dist-packages/wandb/wandb_controller.py in sweep(sweep, entity, project)
    710         _api0 = InternalApi()
    711         if not _api0.api_key:
--> 712             wandb._jupyter_login(api=_api0)
    713     if entity:
    714         env.set_entity(entity)

AttributeError: module 'wandb' has no attribute '_jupyter_login'

wandb 0.10.1
Python 3.6

Java Maven artifact not found

I tried to run the java-sin example. So I cloned the examples repository, went to examples/java/java-sin and types `mvn clean packageยด only to get

[ERROR] Failed to execute goal on project wandb-sin-func: Could not resolve dependencies for project com.wandbj.sinfunc:wandb-sin-func:jar:1.0-SNAPSHOT: Could not find artifact com.wandb.client:client-ng-java:jar:1.0-SNAPSHOT

Should the artifact be there? I can't find it in the Maven repositories.

Fixes for the CV-NLP-RL Intro Colab

General

  • Bring notebook into repo for versioning
  • Periods at the end of sentences
  • Check consistency of ' and ".
  • Follow PEP8 consistently
  • Slightly expand text or convert short text to headers
  • Improve navigation experience -- hide sidebar and add links or clean sidebar
  • Americanise spelling
  • Comment lines at a consistent rate/level of detail
  • Define the garbage collection code at the top and re-use?
  • Consistently use -qqq and clear_output() and %%capture

Header

  • Fix "Create an Account" link
  • Better title and intro, fitting purpose as introductory notebook
  • Fix import order

CV

  • Shorten the iNat model training -- val/loss stops going down after 10 epochs, more like 3 minutes than 7
  • Remove promise of precise time -- Colab compute is variable
  • Better text around YOLOv5

NLP

  • Shorten the Yahoo Answers model training -- loss stops going down after one epoch

RL

  • Root-cause and fix crash in training code -- and consider removing dependence on keras-rl2, which is no longer maintained
  • test_env commented code -- can that be removed? if so, do we need all video-writing utilties?
  • Don't need gym[all]
  • Increase duration of training for cartpole example -- up to a minute or two

wandb: ERROR Can't save model, h5py returned error: Layer ModuleWrapper has arguments in `__init__` and therefore must override `get_config`.

when I sweep my model in Colab, I found a error each sweep:

`wandb: ERROR Can't save model, h5py returned error: Layer ModuleWrapper has arguments in `__init__` and therefore must override `get_config`.

there is no problem while training(param tuning) but I need to restore model.
I didn't use subclassing module. It's sequential module, so I didn't expect this error.
how can I fix it?

framework: tensorflow, keras
environment: Colab

mymodel:

with open('train.py', 'w') as f:
    f.write("""
import wandb 
from wandb.keras import WandbCallback
import tensorflow as tf
import pandas as pd
import numpy as np
from keras.models import Sequential
from keras.layers import LSTM, Dense, Dropout
from sklearn.model_selection import train_test_split

PATH = '/content/drive/MyDrive/input/'
df = pd.read_csv(PATH + 'GME_scaled.csv')

default_configs = {
    'learning_rate': 0.001,
    'dropout_rate1': 0.2,
    'hidden1': 50,
    'time_step': 60,
}
wandb.init(project='gamestop_prediction', entity='kuggle', config=default_configs)
config = wandb.config

X = df.drop('date', axis=1).values
y = df['open_price'].values
X_tr, X_te, y_tr, y_te = train_test_split(X, y, test_size=0.1, shuffle=False)

time_step = config.time_step
X_train, y_train, X_test, y_test = [], [], [], []
for i in range(time_step,len(X_tr)): # train
    X_train.append(X_tr[i-time_step:i,:])
    y_train.append(y_tr[i])
for i in range(time_step, len(X_te)): # test
    X_test.append(X_te[i-time_step:i,:])
    y_test.append(y_te[i])
X_train, y_train = np.array(X_train), np.array(y_train)
X_test, y_test = np.array(X_test), np.array(y_test)
# LSTM input shape ์กฐ๊ฑด(batch_size, timestep, feature_num)์— ๋งž๊ฒŒ reshape
X_train = np.reshape(X_train, (-1,X_train.shape[1],2))
X_test = np.reshape(X_test, (-1,X_test.shape[1],2))

model = tf.keras.Sequential([
LSTM(config.hidden1, return_sequences=True, input_shape=(config.time_step, 2)),
Dropout(config.dropout_rate1),
Dense(1),
])

opt = tf.keras.optimizers.Adam(learning_rate=config.learning_rate)

early_stopping = tf.keras.callbacks.EarlyStopping(monitor='val_loss', patience=10)

model.compile(optimizer=opt,
            loss='mse')

model.fit(X_train, y_train,
        validation_data=(X_test, y_test),
        epochs=200, callbacks=[WandbCallback(), early_stopping])
""")

restore model:

import wandb
import tensorflow as tf
from keras.layers import LSTM, Dropout, Dense



default_configs = {
    'learning_rate': 0.006008989711552188,
    'dropout_rate1': 0.39661787877288335,
    'hidden1': 15,
    'time_step': 4,
}

wandb.init(project='gamestop_prediction', entity='kuggle', config=default_configs)
config = wandb.config

model = tf.keras.Sequential([
LSTM(config.hidden1, return_sequences=True, input_shape=(config.time_step, 2)),
Dropout(config.dropout_rate1),
Dense(1),
])

# restore a model file from a specific run 
best_model = wandb.restore('model-best.h5', run_path="kuggle/gamestop_prediction/gyxqwlu4")
best_model = tf.keras.models.load_model(best_model)
best_model.summary()
---------------------------------------------------------------------------
ValueError                                Traceback (most recent call last)
<ipython-input-26-9572e4434e7f> in <module>()
      1 # restore a model file from a specific run
      2 best_model = wandb.restore('model-best.h5', run_path="kuggle/gamestop_prediction/gyxqwlu4")
----> 3 best_model = model.load_weights(best_model.name)
      4 best_model.summary()

1 frames
/usr/local/lib/python3.7/dist-packages/keras/saving/hdf5_format.py in load_weights_from_hdf5_group(f, layers)
    689                      'containing ' + str(len(layer_names)) +
    690                      ' layers into a model with ' + str(len(filtered_layers)) +
--> 691                      ' layers.')
    692 
    693   # We batch weight value assignments in a single backend call

ValueError: You are trying to load a weight file containing 0 layers into a model with 2 layers.

get_sweep_url() and get_project_url() are now protected (python wandb 0.10.4)

with wandb 0.10.4


and

gives this error
AttributeError: 'Run' object has no attribute 'get_sweep_url'

as temporary workaround one can use

    sweep_url = sweep._get_sweep_url()
    project_url = sweep._get_project_url()

Change program name

Rather than run python train.py, I want to run a shell script ./script.py train.py -d configs config1.yml config2.yml. Do I simply change the program variable in the sweep for that?

keras-cnn-fashion example fails at resume step

When running the keras-cnn-fashion example, even for the first time, it tries to resume a previous run and fails. Commenting out the "if run.resumed: resume" code (examples/examples/keras/keras-cnn-fashion/train.py lines 58-62) makes everything work fine.

python=3.8.5
wandb=0.10.9

First fix importing keras.models.load_model (PR #50), then run:

import wandb
import yaml
project_name='test'
config_yaml = 'sweep-bayes-hyperband.yaml'
with open(config_yaml) as file:
    config_dict = yaml.load(file, Loader=yaml.FullLoader)
sweep_id = wandb.sweep(config_dict, project=project_name)
wandb.agent(sweep_id, project=project_name)

Returns:

RESUMING
Traceback (most recent call last):
  File "train.py", line 61, in <module>
    model = load_model(wandb.restore("model-best.h5").name)
AttributeError: 'NoneType' object has no attribute 'name'

Simple Scikit Integration Colab - Update for Classification Block

  • Issue:
    Hi! I was surfing through the wandb quickstarts when I came across the Classification section of the Scikit Integration example. The example uses the Wine Classification dataset and tries to do a Multi-Class Classification which is not yet supported by the All-in-one: Classifier Plot (implemented using wandb.sklearn.plot_classifier()). Check out the image below:

image

  • Proposed Changes:
    I want to update the Classification section just by using the Breast Cancer Dataset which can be found under sklearn.datasets instead of the Wine Dataset so that the All-in-one: Classifier Plot section runs without any error.

Simple TensorFlow Integration colab throws errors when multiple models are trained

@ayulockin: in the TensorFlow example, attempting to run the training step more than once, e.g. after tweaking the model parameters, gives the following TensorFlow error:

ValueError: tf.function-decorated function tried to create variables on non-first call.

This discussion on GitHub includes a possible solution, but I found that just dropping the decorators removed the error. Not sure if the code remains correct, though.

Cross-validation with sweeps not working

Hi, i'm in love with wandb! In the past few days i was trying to set up a project to work with cross-validation and sweeps. I started from the dedicated example https://github.com/wandb/examples/tree/master/examples/wandb-sweeps/sweeps-cross-validation as it is.
Then, following the instructions in the Readme, I created a sweep and ran an agent. In my case:

$ wandb sweep test_sweeps.yml
wandb: Creating sweep from: test_sweeps.yml
wandb: Created sweep with ID: qmzmf6ms
wandb: View sweep at: https://wandb.ai/buckler/myproject/sweeps/qmzmf6ms
wandb: Run sweep agent with: wandb agent buckler/myproject/qmzmf6ms
$ wandb agent --count 10 buckler/myproject/qmzmf6ms

On my pc the things seems to run smoothly. After 10 run the agent stops (see the whole log below).

However what i obtain in the dashboard is not what I expected and what the example say i should obtain.
In the sweep page I obtain the log of just one run also if in the list I can see all the sweeps run (see the attached screenshot).

What I'm doing wrong? I was wondering if it could be a dashboard problem because on my side all seems to run properly.
Thanks!

Additional info:
Version of wandb i use is wandb==0.10.11. I tried the script also replacing the wandb.join() with wandb.finish()

and commenting the line sweep_run.save() to remove the deprecation warning, but this doesn't resolve the problem as expected. I got the same result.

The funny things is that I managed to get it to work deleting the project from the dashboard and restart from the beginning. It worked only once. I tried this at least 10 times with no luck.

1
2

(venv) buckler@buckler-GE62-6QD:/media/buckler/Data/Work/myproject$ wandb sweep test_sweeps.yml 
wandb: Creating sweep from: test_sweeps.yml
wandb: Created sweep with ID: qmzmf6ms
wandb: View sweep at: https://wandb.ai/buckler/myproject/sweeps/qmzmf6ms
wandb: Run sweep agent with: wandb agent buckler/myproject/qmzmf6ms
(venv) buckler@buckler-GE62-6QD:/media/buckler/Data/Work/myproject$ wandb agent --count 10 buckler/myproject/qmzmf6ms
wandb: Starting wandb agent ๐Ÿ•ต๏ธ
2020-12-02 18:44:33,734 - wandb.wandb_agent - INFO - Running runs: []
2020-12-02 18:44:34,208 - wandb.wandb_agent - INFO - Agent received command: run
2020-12-02 18:44:34,208 - wandb.wandb_agent - INFO - Agent starting run with config:
        batch_size: 36
        epochs: 200
        node_size1: 64
        node_size2: 64
        node_size3: 64
        node_size4: 64
        node_size5: 128
        num_layers: 2
        optimizer: Adam
2020-12-02 18:44:34,217 - wandb.wandb_agent - INFO - About to run command: /usr/bin/env python test_sweeps.py --batch_size=36 --epochs=200 --node_size1=64 --node_size2=64 --node_size3=64 --node_size4=64 --node_size5=128 --num_layers=2 --optimizer=Adam
2020-12-02 18:44:39,230 - wandb.wandb_agent - INFO - Running runs: ['iy0zx516']
wandb: Currently logged in as: buckler (use `wandb login --relogin` to force relogin)
wandb: Tracking run with wandb version 0.10.11
wandb: Syncing run woven-sweep-1
wandb: โญ๏ธ View project at https://wandb.ai/buckler/myproject
wandb: ๐Ÿงน View sweep at https://wandb.ai/buckler/myproject/sweeps/qmzmf6ms
wandb: ๐Ÿš€ View run at https://wandb.ai/buckler/myproject/runs/iy0zx516
wandb: Run data is saved locally in /media/buckler/Data/Work/myproject/wandb/run-20201202_184447-iy0zx516
wandb: Run `wandb offline` to turn off syncing.

wandb: WARNING Calling run.save without any arguments is deprecated.Changes to attributes are automatically persisted.
wandb: Currently logged in as: buckler (use `wandb login --relogin` to force relogin)
wandb: Tracking run with wandb version 0.10.11
wandb: Syncing run woven-sweep-1-0
wandb: โญ๏ธ View project at https://wandb.ai/buckler/myproject
wandb: ๐Ÿš€ View run at https://wandb.ai/buckler/myproject/runs/3hmddeke
wandb: Run data is saved locally in /media/buckler/Data/Work/myproject/wandb/run-20201202_184450-3hmddeke
wandb: Run `wandb offline` to turn off syncing.                                                                                                                                                                                                                                
                                                                                                                                                                                                                                                                               
                                                                                                                                                                                                                                                                               
wandb: Waiting for W&B process to finish, PID 9805                                                                                                                                                                                                                             
wandb: Program ended successfully.                                                                                                                                                                                                                                             
wandb:                                                                                                                                                                                                                                                                         
wandb: Find user logs for this run at: /media/buckler/Data/Work/myproject/wandb/run-20201202_184450-3hmddeke/logs/debug.log                                                                                                                                   
wandb: Find internal logs for this run at: /media/buckler/Data/Work/myproject/wandb/run-20201202_184450-3hmddeke/logs/debug-internal.log                                                                                                                      
wandb: Run summary:                                                                                                                                                                                                                                                            
wandb:   val_accuracy 0.00752                                                                                                                                                                                                                                                  
wandb:          _step 0                                                                                                                                                                                                                                                        
wandb:       _runtime 2                                                                                                                                                                                                                                                        
wandb:     _timestamp 1606931092                                                                                                                                                                                                                                               
wandb: Run history:                                                                                                                                                                                                                                                            
wandb:   val_accuracy โ–                                                                                                                                                                                                                                                        
wandb:          _step โ–                                                                                                                                                                                                                                                        
wandb:       _runtime โ–                                                                                                                                                                                                                                                        
wandb:     _timestamp โ–                                                                                                                                                                                                                                                        
wandb:                                                                                                                                                                                                                                                                         
wandb: Synced 5 W&B file(s), 0 media file(s), 0 artifact file(s) and 1 other file(s)                                                                                                                                                                                           
wandb:                                                                                                                                                                                                                                                                         
wandb: Synced woven-sweep-1-0: https://wandb.ai/buckler/myproject/runs/3hmddeke                                                                                                                                                                                      
wandb: Currently logged in as: buckler (use `wandb login --relogin` to force relogin)                                                                                                                                                                                          
wandb: Tracking run with wandb version 0.10.11
wandb: Syncing run woven-sweep-1-1
wandb: โญ๏ธ View project at https://wandb.ai/buckler/myproject
wandb: ๐Ÿš€ View run at https://wandb.ai/buckler/myproject/runs/2rfzlh9q
wandb: Run data is saved locally in /media/buckler/Data/Work/myproject/wandb/run-20201202_184456-2rfzlh9q
wandb: Run `wandb offline` to turn off syncing.


wandb: Waiting for W&B process to finish, PID 9986
wandb: Program ended successfully.
wandb:                                                                                
wandb: Find user logs for this run at: /media/buckler/Data/Work/myproject/wandb/run-20201202_184456-2rfzlh9q/logs/debug.log
wandb: Find internal logs for this run at: /media/buckler/Data/Work/myproject/wandb/run-20201202_184456-2rfzlh9q/logs/debug-internal.log
wandb: Run summary:
wandb:   val_accuracy 0.70615
wandb:          _step 0
wandb:       _runtime 2
wandb:     _timestamp 1606931098
wandb: Run history:
wandb:   val_accuracy โ–
wandb:          _step โ–
wandb:       _runtime โ–
wandb:     _timestamp โ–
wandb: 
wandb: Synced 5 W&B file(s), 0 media file(s), 0 artifact file(s) and 1 other file(s)
wandb: 
wandb: Synced woven-sweep-1-1: https://wandb.ai/buckler/myproject/runs/2rfzlh9q
wandb: Currently logged in as: buckler (use `wandb login --relogin` to force relogin)
wandb: Tracking run with wandb version 0.10.11
wandb: Syncing run woven-sweep-1-2
wandb: โญ๏ธ View project at https://wandb.ai/buckler/myproject
wandb: ๐Ÿš€ View run at https://wandb.ai/buckler/myproject/runs/2l6vku2o
wandb: Run data is saved locally in /media/buckler/Data/Work/myproject/wandb/run-20201202_184502-2l6vku2o
wandb: Run `wandb offline` to turn off syncing.


wandb: Waiting for W&B process to finish, PID 10044
wandb: Program ended successfully.
wandb:                                                                                
wandb: Find user logs for this run at: /media/buckler/Data/Work/myproject/wandb/run-20201202_184502-2l6vku2o/logs/debug.log
wandb: Find internal logs for this run at: /media/buckler/Data/Work/myproject/wandb/run-20201202_184502-2l6vku2o/logs/debug-internal.log
wandb: Run summary:
wandb:   val_accuracy 0.72753
wandb:          _step 0
wandb:       _runtime 2
wandb:     _timestamp 1606931104
wandb: Run history:
wandb:   val_accuracy โ–
wandb:          _step โ–
wandb:       _runtime โ–
wandb:     _timestamp โ–
wandb: 
wandb: Synced 5 W&B file(s), 0 media file(s), 0 artifact file(s) and 1 other file(s)
wandb: 
wandb: Synced woven-sweep-1-2: https://wandb.ai/buckler/myproject/runs/2l6vku2o
wandb: Currently logged in as: buckler (use `wandb login --relogin` to force relogin)
wandb: Tracking run with wandb version 0.10.11
wandb: Syncing run woven-sweep-1-3
wandb: โญ๏ธ View project at https://wandb.ai/buckler/myproject
wandb: ๐Ÿš€ View run at https://wandb.ai/buckler/myproject/runs/ji4mziqu
wandb: Run data is saved locally in /media/buckler/Data/Work/myproject/wandb/run-20201202_184508-ji4mziqu
wandb: Run `wandb offline` to turn off syncing.


wandb: Waiting for W&B process to finish, PID 10096
wandb: Program ended successfully.
wandb:                                                                                
wandb: Find user logs for this run at: /media/buckler/Data/Work/myproject/wandb/run-20201202_184508-ji4mziqu/logs/debug.log
wandb: Find internal logs for this run at: /media/buckler/Data/Work/myproject/wandb/run-20201202_184508-ji4mziqu/logs/debug-internal.log
wandb: Run summary:
wandb:   val_accuracy 0.17616
wandb:          _step 0
wandb:       _runtime 2
wandb:     _timestamp 1606931110
wandb: Run history:
wandb:   val_accuracy โ–
wandb:          _step โ–
wandb:       _runtime โ–
wandb:     _timestamp โ–
wandb: 
wandb: Synced 5 W&B file(s), 0 media file(s), 0 artifact file(s) and 1 other file(s)
wandb: 
wandb: Synced woven-sweep-1-3: https://wandb.ai/buckler/myproject/runs/ji4mziqu
wandb: Currently logged in as: buckler (use `wandb login --relogin` to force relogin)
wandb: Tracking run with wandb version 0.10.11
wandb: Syncing run woven-sweep-1-4
wandb: โญ๏ธ View project at https://wandb.ai/buckler/myproject
wandb: ๐Ÿš€ View run at https://wandb.ai/buckler/myproject/runs/24sjlmi8
wandb: Run data is saved locally in /media/buckler/Data/Work/myproject/wandb/run-20201202_184514-24sjlmi8
wandb: Run `wandb offline` to turn off syncing.


wandb: Waiting for W&B process to finish, PID 10169
wandb: Program ended successfully.
wandb:                                                                                
wandb: Find user logs for this run at: /media/buckler/Data/Work/myproject/wandb/run-20201202_184514-24sjlmi8/logs/debug.log
wandb: Find internal logs for this run at: /media/buckler/Data/Work/myproject/wandb/run-20201202_184514-24sjlmi8/logs/debug-internal.log
wandb: Run summary:
wandb:   val_accuracy 0.83827
wandb:          _step 0
wandb:       _runtime 2
wandb:     _timestamp 1606931116
wandb: Run history:
wandb:   val_accuracy โ–
wandb:          _step โ–
wandb:       _runtime โ–
wandb:     _timestamp โ–
wandb: 
wandb: Synced 5 W&B file(s), 0 media file(s), 0 artifact file(s) and 1 other file(s)
wandb: 
wandb: Synced woven-sweep-1-4: https://wandb.ai/buckler/myproject/runs/24sjlmi8

wandb: Waiting for W&B process to finish, PID 9756
wandb: Program ended successfully.
wandb:                                                                                
wandb: Find user logs for this run at: /media/buckler/Data/Work/myproject/wandb/run-20201202_184447-iy0zx516/logs/debug.log
wandb: Find internal logs for this run at: /media/buckler/Data/Work/myproject/wandb/run-20201202_184447-iy0zx516/logs/debug-internal.log
wandb: Run summary:
wandb:   val_accuracy 0.49113
wandb:          _step 0
wandb:       _runtime 33
wandb:     _timestamp 1606931120
wandb: Run history:
wandb:   val_accuracy โ–
wandb:          _step โ–
wandb:       _runtime โ–
wandb:     _timestamp โ–
wandb: 
wandb: Synced 6 W&B file(s), 0 media file(s), 0 artifact file(s) and 1 other file(s)
wandb: 
wandb: Synced woven-sweep-1: https://wandb.ai/buckler/myproject/runs/iy0zx516
****************************************
Sweep URL:        https://wandb.ai/buckler/myproject/sweeps/qmzmf6ms
Sweep Group URL:  https://wandb.ai/buckler/myproject/groups/qmzmf6ms
****************************************
2020-12-02 18:45:26,999 - wandb.wandb_agent - INFO - Cleaning up finished run: iy0zx516
2020-12-02 18:45:27,372 - wandb.wandb_agent - INFO - Agent received command: run
2020-12-02 18:45:27,372 - wandb.wandb_agent - INFO - Agent starting run with config:
        batch_size: 115
        epochs: 200
        node_size1: 128
        node_size2: 128
        node_size3: 256
        node_size4: 128
        node_size5: 64
        num_layers: 1
        optimizer: Adamax
2020-12-02 18:45:27,380 - wandb.wandb_agent - INFO - About to run command: /usr/bin/env python test_sweeps.py --batch_size=115 --epochs=200 --node_size1=128 --node_size2=128 --node_size3=256 --node_size4=128 --node_size5=64 --num_layers=1 --optimizer=Adamax
wandb: Currently logged in as: buckler (use `wandb login --relogin` to force relogin)
wandb: Tracking run with wandb version 0.10.11
wandb: Syncing run snowy-sweep-2
wandb: โญ๏ธ View project at https://wandb.ai/buckler/myproject
wandb: ๐Ÿงน View sweep at https://wandb.ai/buckler/myproject/sweeps/qmzmf6ms
wandb: ๐Ÿš€ View run at https://wandb.ai/buckler/myproject/runs/vxeefaqu
wandb: Run data is saved locally in /media/buckler/Data/Work/myproject/wandb/run-20201202_184529-vxeefaqu
wandb: Run `wandb offline` to turn off syncing.

wandb: WARNING Calling run.save without any arguments is deprecated.Changes to attributes are automatically persisted.
wandb: Currently logged in as: buckler (use `wandb login --relogin` to force relogin)
2020-12-02 18:45:32,394 - wandb.wandb_agent - INFO - Running runs: ['vxeefaqu']
wandb: Tracking run with wandb version 0.10.11
wandb: Syncing run snowy-sweep-2-0
wandb: โญ๏ธ View project at https://wandb.ai/buckler/myproject
wandb: ๐Ÿš€ View run at https://wandb.ai/buckler/myproject/runs/3qzzxs3u
wandb: Run data is saved locally in /media/buckler/Data/Work/myproject/wandb/run-20201202_184532-3qzzxs3u
wandb: Run `wandb offline` to turn off syncing.


wandb: Waiting for W&B process to finish, PID 10389
wandb: Program ended successfully.
wandb:                                                                                
wandb: Find user logs for this run at: /media/buckler/Data/Work/myproject/wandb/run-20201202_184532-3qzzxs3u/logs/debug.log
wandb: Find internal logs for this run at: /media/buckler/Data/Work/myproject/wandb/run-20201202_184532-3qzzxs3u/logs/debug-internal.log
wandb: Run summary:
wandb:   val_accuracy 0.77675
wandb:          _step 0
wandb:       _runtime 2
wandb:     _timestamp 1606931134
wandb: Run history:
wandb:   val_accuracy โ–
wandb:          _step โ–
wandb:       _runtime โ–
wandb:     _timestamp โ–
wandb: 
wandb: Synced 5 W&B file(s), 0 media file(s), 0 artifact file(s) and 1 other file(s)
wandb: 
wandb: Synced snowy-sweep-2-0: https://wandb.ai/buckler/myproject/runs/3qzzxs3u
wandb: Currently logged in as: buckler (use `wandb login --relogin` to force relogin)
wandb: Tracking run with wandb version 0.10.11
wandb: Syncing run snowy-sweep-2-1
wandb: โญ๏ธ View project at https://wandb.ai/buckler/myproject
wandb: ๐Ÿš€ View run at https://wandb.ai/buckler/myproject/runs/96h0w1qm
wandb: Run data is saved locally in /media/buckler/Data/Work/myproject/wandb/run-20201202_184538-96h0w1qm
wandb: Run `wandb offline` to turn off syncing.


wandb: Waiting for W&B process to finish, PID 10441
wandb: Program ended successfully.
wandb:                                                                                
wandb: Find user logs for this run at: /media/buckler/Data/Work/myproject/wandb/run-20201202_184538-96h0w1qm/logs/debug.log
wandb: Find internal logs for this run at: /media/buckler/Data/Work/myproject/wandb/run-20201202_184538-96h0w1qm/logs/debug-internal.log
wandb: Run summary:
wandb:   val_accuracy 0.98652
wandb:          _step 0
wandb:       _runtime 2
wandb:     _timestamp 1606931140
wandb: Run history:
wandb:   val_accuracy โ–
wandb:          _step โ–
wandb:       _runtime โ–
wandb:     _timestamp โ–
wandb: 
wandb: Synced 5 W&B file(s), 0 media file(s), 0 artifact file(s) and 1 other file(s)
wandb: 
wandb: Synced snowy-sweep-2-1: https://wandb.ai/buckler/myproject/runs/96h0w1qm
wandb: Currently logged in as: buckler (use `wandb login --relogin` to force relogin)
wandb: Tracking run with wandb version 0.10.11
wandb: Syncing run snowy-sweep-2-2
wandb: โญ๏ธ View project at https://wandb.ai/buckler/myproject
wandb: ๐Ÿš€ View run at https://wandb.ai/buckler/myproject/runs/3gzpsx8m
wandb: Run data is saved locally in /media/buckler/Data/Work/myproject/wandb/run-20201202_184544-3gzpsx8m
wandb: Run `wandb offline` to turn off syncing.


wandb: Waiting for W&B process to finish, PID 10493
wandb: Program ended successfully.
wandb:                                                                                
wandb: Find user logs for this run at: /media/buckler/Data/Work/myproject/wandb/run-20201202_184544-3gzpsx8m/logs/debug.log
wandb: Find internal logs for this run at: /media/buckler/Data/Work/myproject/wandb/run-20201202_184544-3gzpsx8m/logs/debug-internal.log
wandb: Run summary:
wandb:   val_accuracy 0.40548
wandb:          _step 0
wandb:       _runtime 2
wandb:     _timestamp 1606931146
wandb: Run history:
wandb:   val_accuracy โ–
wandb:          _step โ–
wandb:       _runtime โ–
wandb:     _timestamp โ–
wandb: 
wandb: Synced 5 W&B file(s), 0 media file(s), 0 artifact file(s) and 1 other file(s)
wandb: 
wandb: Synced snowy-sweep-2-2: https://wandb.ai/buckler/myproject/runs/3gzpsx8m
wandb: Currently logged in as: buckler (use `wandb login --relogin` to force relogin)
wandb: Tracking run with wandb version 0.10.11
wandb: Syncing run snowy-sweep-2-3
wandb: โญ๏ธ View project at https://wandb.ai/buckler/myproject
wandb: ๐Ÿš€ View run at https://wandb.ai/buckler/myproject/runs/3gk8h7vi
wandb: Run data is saved locally in /media/buckler/Data/Work/myproject/wandb/run-20201202_184550-3gk8h7vi
wandb: Run `wandb offline` to turn off syncing.


wandb: Waiting for W&B process to finish, PID 10546
wandb: Program ended successfully.
wandb:                                                                                
wandb: Find user logs for this run at: /media/buckler/Data/Work/myproject/wandb/run-20201202_184550-3gk8h7vi/logs/debug.log
wandb: Find internal logs for this run at: /media/buckler/Data/Work/myproject/wandb/run-20201202_184550-3gk8h7vi/logs/debug-internal.log
wandb: Run summary:
wandb:   val_accuracy 0.33316
wandb:          _step 0
wandb:       _runtime 2
wandb:     _timestamp 1606931152
wandb: Run history:
wandb:   val_accuracy โ–
wandb:          _step โ–
wandb:       _runtime โ–
wandb:     _timestamp โ–
wandb: 
wandb: Synced 5 W&B file(s), 0 media file(s), 0 artifact file(s) and 1 other file(s)
wandb: 
wandb: Synced snowy-sweep-2-3: https://wandb.ai/buckler/myproject/runs/3gk8h7vi
wandb: Currently logged in as: buckler (use `wandb login --relogin` to force relogin)
wandb: Tracking run with wandb version 0.10.11
wandb: Syncing run snowy-sweep-2-4
wandb: โญ๏ธ View project at https://wandb.ai/buckler/myproject
wandb: ๐Ÿš€ View run at https://wandb.ai/buckler/myproject/runs/xv9nabyg
wandb: Run data is saved locally in /media/buckler/Data/Work/myproject/wandb/run-20201202_184556-xv9nabyg
wandb: Run `wandb offline` to turn off syncing.


wandb: Waiting for W&B process to finish, PID 10599
wandb: Program ended successfully.
wandb:                                                                                
wandb: Find user logs for this run at: /media/buckler/Data/Work/myproject/wandb/run-20201202_184556-xv9nabyg/logs/debug.log
wandb: Find internal logs for this run at: /media/buckler/Data/Work/myproject/wandb/run-20201202_184556-xv9nabyg/logs/debug-internal.log
wandb: Run summary:
wandb:   val_accuracy 0.44462
wandb:          _step 0
wandb:       _runtime 2
wandb:     _timestamp 1606931158
wandb: Run history:
wandb:   val_accuracy โ–
wandb:          _step โ–
wandb:       _runtime โ–
wandb:     _timestamp โ–
wandb: 
wandb: Synced 5 W&B file(s), 0 media file(s), 0 artifact file(s) and 1 other file(s)
wandb: 
wandb: Synced snowy-sweep-2-4: https://wandb.ai/buckler/myproject/runs/xv9nabyg

wandb: Waiting for W&B process to finish, PID 10336
wandb: Program ended successfully.
wandb:                                                                                
wandb: Find user logs for this run at: /media/buckler/Data/Work/myproject/wandb/run-20201202_184529-vxeefaqu/logs/debug.log
wandb: Find internal logs for this run at: /media/buckler/Data/Work/myproject/wandb/run-20201202_184529-vxeefaqu/logs/debug-internal.log
wandb: Run summary:
wandb:   val_accuracy 0.5893
wandb:          _step 0
wandb:       _runtime 33
wandb:     _timestamp 1606931162
wandb: Run history:
wandb:   val_accuracy โ–
wandb:          _step โ–
wandb:       _runtime โ–
wandb:     _timestamp โ–
wandb: 
wandb: Synced 6 W&B file(s), 0 media file(s), 0 artifact file(s) and 1 other file(s)
wandb: 
wandb: Synced snowy-sweep-2: https://wandb.ai/buckler/myproject/runs/vxeefaqu
****************************************
Sweep URL:        https://wandb.ai/buckler/myproject/sweeps/qmzmf6ms
Sweep Group URL:  https://wandb.ai/buckler/myproject/groups/qmzmf6ms
****************************************
2020-12-02 18:46:09,038 - wandb.wandb_agent - INFO - Cleaning up finished run: vxeefaqu
2020-12-02 18:46:09,375 - wandb.wandb_agent - INFO - Agent received command: run
2020-12-02 18:46:09,376 - wandb.wandb_agent - INFO - Agent starting run with config:
        batch_size: 73
        epochs: 200
        node_size1: 64
        node_size2: 256
        node_size3: 64
        node_size4: 128
        node_size5: 128
        num_layers: 1
        optimizer: Adagrad
2020-12-02 18:46:09,384 - wandb.wandb_agent - INFO - About to run command: /usr/bin/env python test_sweeps.py --batch_size=73 --epochs=200 --node_size1=64 --node_size2=256 --node_size3=64 --node_size4=128 --node_size5=128 --num_layers=1 --optimizer=Adagrad
wandb: Currently logged in as: buckler (use `wandb login --relogin` to force relogin)
wandb: Tracking run with wandb version 0.10.11
wandb: Syncing run visionary-sweep-3
wandb: โญ๏ธ View project at https://wandb.ai/buckler/myproject
wandb: ๐Ÿงน View sweep at https://wandb.ai/buckler/myproject/sweeps/qmzmf6ms
wandb: ๐Ÿš€ View run at https://wandb.ai/buckler/myproject/runs/5j87b725
wandb: Run data is saved locally in /media/buckler/Data/Work/myproject/wandb/run-20201202_184611-5j87b725
wandb: Run `wandb offline` to turn off syncing.

wandb: WARNING Calling run.save without any arguments is deprecated.Changes to attributes are automatically persisted.
wandb: Currently logged in as: buckler (use `wandb login --relogin` to force relogin)
2020-12-02 18:46:14,398 - wandb.wandb_agent - INFO - Running runs: ['5j87b725']
wandb: Tracking run with wandb version 0.10.11
wandb: Syncing run visionary-sweep-3-0
wandb: โญ๏ธ View project at https://wandb.ai/buckler/myproject
wandb: ๐Ÿš€ View run at https://wandb.ai/buckler/myproject/runs/1sinokan
wandb: Run data is saved locally in /media/buckler/Data/Work/myproject/wandb/run-20201202_184613-1sinokan
wandb: Run `wandb offline` to turn off syncing.


wandb: Waiting for W&B process to finish, PID 10740
wandb: Program ended successfully.
wandb:                                                                                
wandb: Find user logs for this run at: /media/buckler/Data/Work/myproject/wandb/run-20201202_184613-1sinokan/logs/debug.log
wandb: Find internal logs for this run at: /media/buckler/Data/Work/myproject/wandb/run-20201202_184613-1sinokan/logs/debug-internal.log
wandb: Run summary:
wandb:   val_accuracy 0.3744
wandb:          _step 0
wandb:       _runtime 1
wandb:     _timestamp 1606931175
wandb: Run history:
wandb:   val_accuracy โ–
wandb:          _step โ–
wandb:       _runtime โ–
wandb:     _timestamp โ–
wandb: 
wandb: Synced 5 W&B file(s), 0 media file(s), 0 artifact file(s) and 1 other file(s)
wandb: 
wandb: Synced visionary-sweep-3-0: https://wandb.ai/buckler/myproject/runs/1sinokan
wandb: Currently logged in as: buckler (use `wandb login --relogin` to force relogin)
wandb: Tracking run with wandb version 0.10.11
wandb: Syncing run visionary-sweep-3-1
wandb: โญ๏ธ View project at https://wandb.ai/buckler/myproject
wandb: ๐Ÿš€ View run at https://wandb.ai/buckler/myproject/runs/3gx4ij39
wandb: Run data is saved locally in /media/buckler/Data/Work/myproject/wandb/run-20201202_184620-3gx4ij39
wandb: Run `wandb offline` to turn off syncing.


wandb: Waiting for W&B process to finish, PID 10796
wandb: Program ended successfully.
wandb:                                                                                
wandb: Find user logs for this run at: /media/buckler/Data/Work/myproject/wandb/run-20201202_184620-3gx4ij39/logs/debug.log
wandb: Find internal logs for this run at: /media/buckler/Data/Work/myproject/wandb/run-20201202_184620-3gx4ij39/logs/debug-internal.log
wandb: Run summary:
wandb:   val_accuracy 0.73789
wandb:          _step 0
wandb:       _runtime 1
wandb:     _timestamp 1606931181
wandb: Run history:
wandb:   val_accuracy โ–
wandb:          _step โ–
wandb:       _runtime โ–
wandb:     _timestamp โ–
wandb: 
wandb: Synced 5 W&B file(s), 0 media file(s), 0 artifact file(s) and 1 other file(s)
wandb: 
wandb: Synced visionary-sweep-3-1: https://wandb.ai/buckler/myproject/runs/3gx4ij39
wandb: Currently logged in as: buckler (use `wandb login --relogin` to force relogin)
wandb: Tracking run with wandb version 0.10.11
wandb: Syncing run visionary-sweep-3-2
wandb: โญ๏ธ View project at https://wandb.ai/buckler/myproject
wandb: ๐Ÿš€ View run at https://wandb.ai/buckler/myproject/runs/13wzet62
wandb: Run data is saved locally in /media/buckler/Data/Work/myproject/wandb/run-20201202_184625-13wzet62
wandb: Run `wandb offline` to turn off syncing.


wandb: Waiting for W&B process to finish, PID 10849
wandb: Program ended successfully.
wandb:                                                                                
wandb: Find user logs for this run at: /media/buckler/Data/Work/myproject/wandb/run-20201202_184625-13wzet62/logs/debug.log
wandb: Find internal logs for this run at: /media/buckler/Data/Work/myproject/wandb/run-20201202_184625-13wzet62/logs/debug-internal.log
wandb: Run summary:
wandb:   val_accuracy 0.96186
wandb:          _step 0
wandb:       _runtime 1
wandb:     _timestamp 1606931187
wandb: Run history:
wandb:   val_accuracy โ–
wandb:          _step โ–
wandb:       _runtime โ–
wandb:     _timestamp โ–
wandb: 
wandb: Synced 5 W&B file(s), 0 media file(s), 0 artifact file(s) and 1 other file(s)
wandb: 
wandb: Synced visionary-sweep-3-2: https://wandb.ai/buckler/myproject/runs/13wzet62
wandb: Currently logged in as: buckler (use `wandb login --relogin` to force relogin)
wandb: Tracking run with wandb version 0.10.11
wandb: Syncing run visionary-sweep-3-3
wandb: โญ๏ธ View project at https://wandb.ai/buckler/myproject
wandb: ๐Ÿš€ View run at https://wandb.ai/buckler/myproject/runs/1tvbyjye
wandb: Run data is saved locally in /media/buckler/Data/Work/myproject/wandb/run-20201202_184632-1tvbyjye
wandb: Run `wandb offline` to turn off syncing.


wandb: Waiting for W&B process to finish, PID 10901
wandb: Program ended successfully.
wandb:                                                                                
wandb: Find user logs for this run at: /media/buckler/Data/Work/myproject/wandb/run-20201202_184632-1tvbyjye/logs/debug.log
wandb: Find internal logs for this run at: /media/buckler/Data/Work/myproject/wandb/run-20201202_184632-1tvbyjye/logs/debug-internal.log
wandb: Run summary:
wandb:   val_accuracy 0.60187
wandb:          _step 0
wandb:       _runtime 2
wandb:     _timestamp 1606931194
wandb: Run history:
wandb:   val_accuracy โ–
wandb:          _step โ–
wandb:       _runtime โ–
wandb:     _timestamp โ–
wandb: 
wandb: Synced 5 W&B file(s), 0 media file(s), 0 artifact file(s) and 1 other file(s)
wandb: 
wandb: Synced visionary-sweep-3-3: https://wandb.ai/buckler/myproject/runs/1tvbyjye
wandb: Currently logged in as: buckler (use `wandb login --relogin` to force relogin)
wandb: Tracking run with wandb version 0.10.11
wandb: Syncing run visionary-sweep-3-4
wandb: โญ๏ธ View project at https://wandb.ai/buckler/myproject
wandb: ๐Ÿš€ View run at https://wandb.ai/buckler/myproject/runs/1kv3b9zs
wandb: Run data is saved locally in /media/buckler/Data/Work/myproject/wandb/run-20201202_184638-1kv3b9zs
wandb: Run `wandb offline` to turn off syncing.


wandb: Waiting for W&B process to finish, PID 10961
wandb: Program ended successfully.
wandb:                                                                                
wandb: Find user logs for this run at: /media/buckler/Data/Work/myproject/wandb/run-20201202_184638-1kv3b9zs/logs/debug.log
wandb: Find internal logs for this run at: /media/buckler/Data/Work/myproject/wandb/run-20201202_184638-1kv3b9zs/logs/debug-internal.log
wandb: Run summary:
wandb:   val_accuracy 0.54447
wandb:          _step 0
wandb:       _runtime 1
wandb:     _timestamp 1606931199
wandb: Run history:
wandb:   val_accuracy โ–
wandb:          _step โ–
wandb:       _runtime โ–
wandb:     _timestamp โ–
wandb: 
wandb: Synced 5 W&B file(s), 0 media file(s), 0 artifact file(s) and 1 other file(s)
wandb: 
wandb: Synced visionary-sweep-3-4: https://wandb.ai/buckler/myproject/runs/1kv3b9zs

wandb: Waiting for W&B process to finish, PID 10690
wandb: Program ended successfully.
wandb:                                                                                
wandb: Find user logs for this run at: /media/buckler/Data/Work/myproject/wandb/run-20201202_184611-5j87b725/logs/debug.log
wandb: Find internal logs for this run at: /media/buckler/Data/Work/myproject/wandb/run-20201202_184611-5j87b725/logs/debug-internal.log
wandb: Run summary:
wandb:   val_accuracy 0.6441
wandb:          _step 0
wandb:       _runtime 32
wandb:     _timestamp 1606931203
wandb: Run history:
wandb:   val_accuracy โ–
wandb:          _step โ–
wandb:       _runtime โ–
wandb:     _timestamp โ–
wandb: 
wandb: Synced 6 W&B file(s), 0 media file(s), 0 artifact file(s) and 1 other file(s)
wandb: 
wandb: Synced visionary-sweep-3: https://wandb.ai/buckler/myproject/runs/5j87b725
****************************************
Sweep URL:        https://wandb.ai/buckler/myproject/sweeps/qmzmf6ms
Sweep Group URL:  https://wandb.ai/buckler/myproject/groups/qmzmf6ms
****************************************
2020-12-02 18:46:51,043 - wandb.wandb_agent - INFO - Cleaning up finished run: 5j87b725
2020-12-02 18:46:51,384 - wandb.wandb_agent - INFO - Agent received command: run
2020-12-02 18:46:51,384 - wandb.wandb_agent - INFO - Agent starting run with config:
        batch_size: 102
        epochs: 200
        node_size1: 256
        node_size2: 64
        node_size3: 128
        node_size4: 256
        node_size5: 128
        num_layers: 3
        optimizer: Adagrad
2020-12-02 18:46:51,387 - wandb.wandb_agent - INFO - About to run command: /usr/bin/env python test_sweeps.py --batch_size=102 --epochs=200 --node_size1=256 --node_size2=64 --node_size3=128 --node_size4=256 --node_size5=128 --num_layers=3 --optimizer=Adagrad
wandb: Currently logged in as: buckler (use `wandb login --relogin` to force relogin)
wandb: Tracking run with wandb version 0.10.11
wandb: Syncing run sweepy-sweep-4
wandb: โญ๏ธ View project at https://wandb.ai/buckler/myproject
wandb: ๐Ÿงน View sweep at https://wandb.ai/buckler/myproject/sweeps/qmzmf6ms
wandb: ๐Ÿš€ View run at https://wandb.ai/buckler/myproject/runs/ghp8xus3
wandb: Run data is saved locally in /media/buckler/Data/Work/myproject/wandb/run-20201202_184653-ghp8xus3
wandb: Run `wandb offline` to turn off syncing.

wandb: WARNING Calling run.save without any arguments is deprecated.Changes to attributes are automatically persisted.
wandb: Currently logged in as: buckler (use `wandb login --relogin` to force relogin)
2020-12-02 18:46:56,394 - wandb.wandb_agent - INFO - Running runs: ['ghp8xus3']
wandb: Tracking run with wandb version 0.10.11
wandb: Syncing run sweepy-sweep-4-0
wandb: โญ๏ธ View project at https://wandb.ai/buckler/myproject
wandb: ๐Ÿš€ View run at https://wandb.ai/buckler/myproject/runs/23xd1zjw
wandb: Run data is saved locally in /media/buckler/Data/Work/myproject/wandb/run-20201202_184655-23xd1zjw
wandb: Run `wandb offline` to turn off syncing.


wandb: Waiting for W&B process to finish, PID 11105
wandb: Program ended successfully.
wandb:                                                                                
wandb: Find user logs for this run at: /media/buckler/Data/Work/myproject/wandb/run-20201202_184655-23xd1zjw/logs/debug.log
wandb: Find internal logs for this run at: /media/buckler/Data/Work/myproject/wandb/run-20201202_184655-23xd1zjw/logs/debug-internal.log
wandb: Run summary:
wandb:   val_accuracy 0.22928
wandb:          _step 0
wandb:       _runtime 2
wandb:     _timestamp 1606931217
wandb: Run history:
wandb:   val_accuracy โ–
wandb:          _step โ–
wandb:       _runtime โ–
wandb:     _timestamp โ–
wandb: 
wandb: Synced 5 W&B file(s), 0 media file(s), 0 artifact file(s) and 1 other file(s)
wandb: 
wandb: Synced sweepy-sweep-4-0: https://wandb.ai/buckler/myproject/runs/23xd1zjw
wandb: Currently logged in as: buckler (use `wandb login --relogin` to force relogin)
wandb: Tracking run with wandb version 0.10.11
wandb: Syncing run sweepy-sweep-4-1
wandb: โญ๏ธ View project at https://wandb.ai/buckler/myproject
wandb: ๐Ÿš€ View run at https://wandb.ai/buckler/myproject/runs/1qkl0ncw
wandb: Run data is saved locally in /media/buckler/Data/Work/myproject/wandb/run-20201202_184701-1qkl0ncw
wandb: Run `wandb offline` to turn off syncing.


wandb: Waiting for W&B process to finish, PID 11158
wandb: Program ended successfully.
wandb:                                                                                
wandb: Find user logs for this run at: /media/buckler/Data/Work/myproject/wandb/run-20201202_184701-1qkl0ncw/logs/debug.log
wandb: Find internal logs for this run at: /media/buckler/Data/Work/myproject/wandb/run-20201202_184701-1qkl0ncw/logs/debug-internal.log
wandb: Run summary:
wandb:   val_accuracy 0.08898
wandb:          _step 0
wandb:       _runtime 2
wandb:     _timestamp 1606931223
wandb: Run history:
wandb:   val_accuracy โ–
wandb:          _step โ–
wandb:       _runtime โ–
wandb:     _timestamp โ–
wandb: 
wandb: Synced 5 W&B file(s), 0 media file(s), 0 artifact file(s) and 1 other file(s)
wandb: 
wandb: Synced sweepy-sweep-4-1: https://wandb.ai/buckler/myproject/runs/1qkl0ncw
wandb: Currently logged in as: buckler (use `wandb login --relogin` to force relogin)
wandb: Tracking run with wandb version 0.10.11
wandb: Syncing run sweepy-sweep-4-2
wandb: โญ๏ธ View project at https://wandb.ai/buckler/myproject
wandb: ๐Ÿš€ View run at https://wandb.ai/buckler/myproject/runs/2scvk3et
wandb: Run data is saved locally in /media/buckler/Data/Work/myproject/wandb/run-20201202_184707-2scvk3et
wandb: Run `wandb offline` to turn off syncing.


wandb: Waiting for W&B process to finish, PID 11214
wandb: Program ended successfully.
wandb:                                                                                
wandb: Find user logs for this run at: /media/buckler/Data/Work/myproject/wandb/run-20201202_184707-2scvk3et/logs/debug.log
wandb: Find internal logs for this run at: /media/buckler/Data/Work/myproject/wandb/run-20201202_184707-2scvk3et/logs/debug-internal.log
wandb: Run summary:
wandb:   val_accuracy 0.55197
wandb:          _step 0
wandb:       _runtime 2
wandb:     _timestamp 1606931229
wandb: Run history:
wandb:   val_accuracy โ–
wandb:          _step โ–
wandb:       _runtime โ–
wandb:     _timestamp โ–
wandb: 
wandb: Synced 5 W&B file(s), 0 media file(s), 0 artifact file(s) and 1 other file(s)
wandb: 
wandb: Synced sweepy-sweep-4-2: https://wandb.ai/buckler/myproject/runs/2scvk3et
wandb: Currently logged in as: buckler (use `wandb login --relogin` to force relogin)
wandb: Tracking run with wandb version 0.10.11
wandb: Syncing run sweepy-sweep-4-3
wandb: โญ๏ธ View project at https://wandb.ai/buckler/myproject
wandb: ๐Ÿš€ View run at https://wandb.ai/buckler/myproject/runs/1zpp9ex8
wandb: Run data is saved locally in /media/buckler/Data/Work/myproject/wandb/run-20201202_184713-1zpp9ex8
wandb: Run `wandb offline` to turn off syncing.


wandb: Waiting for W&B process to finish, PID 11267
wandb: Program ended successfully.
wandb:                                                                                
wandb: Find user logs for this run at: /media/buckler/Data/Work/myproject/wandb/run-20201202_184713-1zpp9ex8/logs/debug.log
wandb: Find internal logs for this run at: /media/buckler/Data/Work/myproject/wandb/run-20201202_184713-1zpp9ex8/logs/debug-internal.log
wandb: Run summary:
wandb:   val_accuracy 0.84223
wandb:          _step 0
wandb:       _runtime 2
wandb:     _timestamp 1606931235
wandb: Run history:
wandb:   val_accuracy โ–
wandb:          _step โ–
wandb:       _runtime โ–
wandb:     _timestamp โ–
wandb: 
wandb: Synced 5 W&B file(s), 0 media file(s), 0 artifact file(s) and 1 other file(s)
wandb: 
wandb: Synced sweepy-sweep-4-3: https://wandb.ai/buckler/myproject/runs/1zpp9ex8
wandb: Currently logged in as: buckler (use `wandb login --relogin` to force relogin)
wandb: Tracking run with wandb version 0.10.11
wandb: Syncing run sweepy-sweep-4-4
wandb: โญ๏ธ View project at https://wandb.ai/buckler/myproject
wandb: ๐Ÿš€ View run at https://wandb.ai/buckler/myproject/runs/10bxri0v
wandb: Run data is saved locally in /media/buckler/Data/Work/myproject/wandb/run-20201202_184719-10bxri0v
wandb: Run `wandb offline` to turn off syncing.


wandb: Waiting for W&B process to finish, PID 11320
wandb: Program ended successfully.
wandb:                                                                                
wandb: Find user logs for this run at: /media/buckler/Data/Work/myproject/wandb/run-20201202_184719-10bxri0v/logs/debug.log
wandb: Find internal logs for this run at: /media/buckler/Data/Work/myproject/wandb/run-20201202_184719-10bxri0v/logs/debug-internal.log
wandb: Run summary:
wandb:   val_accuracy 0.6714
wandb:          _step 0
wandb:       _runtime 2
wandb:     _timestamp 1606931241
wandb: Run history:
wandb:   val_accuracy โ–
wandb:          _step โ–
wandb:       _runtime โ–
wandb:     _timestamp โ–
wandb: 
wandb: Synced 5 W&B file(s), 0 media file(s), 0 artifact file(s) and 1 other file(s)
wandb: 
wandb: Synced sweepy-sweep-4-4: https://wandb.ai/buckler/myproject/runs/10bxri0v

wandb: Waiting for W&B process to finish, PID 11040
wandb: Program ended successfully.
wandb:                                                                                
wandb: Find user logs for this run at: /media/buckler/Data/Work/myproject/wandb/run-20201202_184653-ghp8xus3/logs/debug.log
wandb: Find internal logs for this run at: /media/buckler/Data/Work/myproject/wandb/run-20201202_184653-ghp8xus3/logs/debug-internal.log
wandb: Run summary:
wandb:   val_accuracy 0.47677
wandb:          _step 0
wandb:       _runtime 32
wandb:     _timestamp 1606931245
wandb: Run history:
wandb:   val_accuracy โ–
wandb:          _step โ–
wandb:       _runtime โ–
wandb:     _timestamp โ–
wandb: 
wandb: Synced 6 W&B file(s), 0 media file(s), 0 artifact file(s) and 1 other file(s)
wandb: 
wandb: Synced sweepy-sweep-4: https://wandb.ai/buckler/myproject/runs/ghp8xus3
****************************************
Sweep URL:        https://wandb.ai/buckler/myproject/sweeps/qmzmf6ms
Sweep Group URL:  https://wandb.ai/buckler/myproject/groups/qmzmf6ms
****************************************
2020-12-02 18:47:33,023 - wandb.wandb_agent - INFO - Cleaning up finished run: ghp8xus3
2020-12-02 18:47:33,401 - wandb.wandb_agent - INFO - Agent received command: run
2020-12-02 18:47:33,401 - wandb.wandb_agent - INFO - Agent starting run with config:
        batch_size: 95
        epochs: 200
        node_size1: 64
        node_size2: 128
        node_size3: 256
        node_size4: 128
        node_size5: 128
        num_layers: 2
        optimizer: Adamax
2020-12-02 18:47:33,409 - wandb.wandb_agent - INFO - About to run command: /usr/bin/env python test_sweeps.py --batch_size=95 --epochs=200 --node_size1=64 --node_size2=128 --node_size3=256 --node_size4=128 --node_size5=128 --num_layers=2 --optimizer=Adamax
wandb: Currently logged in as: buckler (use `wandb login --relogin` to force relogin)
wandb: Tracking run with wandb version 0.10.11
wandb: Syncing run misunderstood-sweep-5
wandb: โญ๏ธ View project at https://wandb.ai/buckler/myproject
wandb: ๐Ÿงน View sweep at https://wandb.ai/buckler/myproject/sweeps/qmzmf6ms
wandb: ๐Ÿš€ View run at https://wandb.ai/buckler/myproject/runs/5fu8v7um
wandb: Run data is saved locally in /media/buckler/Data/Work/myproject/wandb/run-20201202_184735-5fu8v7um
wandb: Run `wandb offline` to turn off syncing.

wandb: WARNING Calling run.save without any arguments is deprecated.Changes to attributes are automatically persisted.
wandb: Currently logged in as: buckler (use `wandb login --relogin` to force relogin)
2020-12-02 18:47:38,422 - wandb.wandb_agent - INFO - Running runs: ['5fu8v7um']
wandb: Tracking run with wandb version 0.10.11
wandb: Syncing run misunderstood-sweep-5-0
wandb: โญ๏ธ View project at https://wandb.ai/buckler/myproject
wandb: ๐Ÿš€ View run at https://wandb.ai/buckler/myproject/runs/1uebb31g
wandb: Run data is saved locally in /media/buckler/Data/Work/myproject/wandb/run-20201202_184738-1uebb31g
wandb: Run `wandb offline` to turn off syncing.


wandb: Waiting for W&B process to finish, PID 11470
wandb: Program ended successfully.
wandb:                                                                                
wandb: Find user logs for this run at: /media/buckler/Data/Work/myproject/wandb/run-20201202_184738-1uebb31g/logs/debug.log
wandb: Find internal logs for this run at: /media/buckler/Data/Work/myproject/wandb/run-20201202_184738-1uebb31g/logs/debug-internal.log
wandb: Run summary:
wandb:   val_accuracy 0.5721
wandb:          _step 0
wandb:       _runtime 1
wandb:     _timestamp 1606931259
wandb: Run history:
wandb:   val_accuracy โ–
wandb:          _step โ–
wandb:       _runtime โ–
wandb:     _timestamp โ–
wandb: 
wandb: Synced 5 W&B file(s), 0 media file(s), 0 artifact file(s) and 1 other file(s)
wandb: 
wandb: Synced misunderstood-sweep-5-0: https://wandb.ai/buckler/myproject/runs/1uebb31g
wandb: Currently logged in as: buckler (use `wandb login --relogin` to force relogin)
wandb: Tracking run with wandb version 0.10.11
wandb: Syncing run misunderstood-sweep-5-1
wandb: โญ๏ธ View project at https://wandb.ai/buckler/myproject
wandb: ๐Ÿš€ View run at https://wandb.ai/buckler/myproject/runs/3rgib3uc
wandb: Run data is saved locally in /media/buckler/Data/Work/myproject/wandb/run-20201202_184743-3rgib3uc
wandb: Run `wandb offline` to turn off syncing.


wandb: Waiting for W&B process to finish, PID 11526
wandb: Program ended successfully.
wandb:                                                                                
wandb: Find user logs for this run at: /media/buckler/Data/Work/myproject/wandb/run-20201202_184743-3rgib3uc/logs/debug.log
wandb: Find internal logs for this run at: /media/buckler/Data/Work/myproject/wandb/run-20201202_184743-3rgib3uc/logs/debug-internal.log
wandb: Run summary:
wandb:   val_accuracy 0.23085
wandb:          _step 0
wandb:       _runtime 2
wandb:     _timestamp 1606931265
wandb: Run history:
wandb:   val_accuracy โ–
wandb:          _step โ–
wandb:       _runtime โ–
wandb:     _timestamp โ–
wandb: 
wandb: Synced 5 W&B file(s), 0 media file(s), 0 artifact file(s) and 1 other file(s)
wandb: 
wandb: Synced misunderstood-sweep-5-1: https://wandb.ai/buckler/myproject/runs/3rgib3uc
wandb: Currently logged in as: buckler (use `wandb login --relogin` to force relogin)
wandb: Tracking run with wandb version 0.10.11
wandb: Syncing run misunderstood-sweep-5-2
wandb: โญ๏ธ View project at https://wandb.ai/buckler/myproject
wandb: ๐Ÿš€ View run at https://wandb.ai/buckler/myproject/runs/10x360em
wandb: Run data is saved locally in /media/buckler/Data/Work/myproject/wandb/run-20201202_184749-10x360em
wandb: Run `wandb offline` to turn off syncing.


wandb: Waiting for W&B process to finish, PID 11579
wandb: Program ended successfully.
wandb:                                                                                
wandb: Find user logs for this run at: /media/buckler/Data/Work/myproject/wandb/run-20201202_184749-10x360em/logs/debug.log
wandb: Find internal logs for this run at: /media/buckler/Data/Work/myproject/wandb/run-20201202_184749-10x360em/logs/debug-internal.log
wandb: Run summary:
wandb:   val_accuracy 0.44966
wandb:          _step 0
wandb:       _runtime 2
wandb:     _timestamp 1606931272
wandb: Run history:
wandb:   val_accuracy โ–
wandb:          _step โ–
wandb:       _runtime โ–
wandb:     _timestamp โ–
wandb: 
wandb: Synced 5 W&B file(s), 0 media file(s), 0 artifact file(s) and 1 other file(s)
wandb: 
wandb: Synced misunderstood-sweep-5-2: https://wandb.ai/buckler/myproject/runs/10x360em
wandb: Currently logged in as: buckler (use `wandb login --relogin` to force relogin)
wandb: Tracking run with wandb version 0.10.11
wandb: Syncing run misunderstood-sweep-5-3
wandb: โญ๏ธ View project at https://wandb.ai/buckler/myproject
wandb: ๐Ÿš€ View run at https://wandb.ai/buckler/myproject/runs/293e9j8k
wandb: Run data is saved locally in /media/buckler/Data/Work/myproject/wandb/run-20201202_184756-293e9j8k
wandb: Run `wandb offline` to turn off syncing.


wandb: Waiting for W&B process to finish, PID 11632
wandb: Program ended successfully.
wandb:                                                                                
wandb: Find user logs for this run at: /media/buckler/Data/Work/myproject/wandb/run-20201202_184756-293e9j8k/logs/debug.log
wandb: Find internal logs for this run at: /media/buckler/Data/Work/myproject/wandb/run-20201202_184756-293e9j8k/logs/debug-internal.log
wandb: Run summary:
wandb:   val_accuracy 0.58749
wandb:          _step 0
wandb:       _runtime 2
wandb:     _timestamp 1606931278
wandb: Run history:
wandb:   val_accuracy โ–
wandb:          _step โ–
wandb:       _runtime โ–
wandb:     _timestamp โ–
wandb: 
wandb: Synced 5 W&B file(s), 0 media file(s), 0 artifact file(s) and 1 other file(s)
wandb: 
wandb: Synced misunderstood-sweep-5-3: https://wandb.ai/buckler/myproject/runs/293e9j8k
wandb: Currently logged in as: buckler (use `wandb login --relogin` to force relogin)
wandb: Tracking run with wandb version 0.10.11
wandb: Syncing run misunderstood-sweep-5-4
wandb: โญ๏ธ View project at https://wandb.ai/buckler/myproject
wandb: ๐Ÿš€ View run at https://wandb.ai/buckler/myproject/runs/1xhcuzib
wandb: Run data is saved locally in /media/buckler/Data/Work/myproject/wandb/run-20201202_184802-1xhcuzib
wandb: Run `wandb offline` to turn off syncing.


wandb: Waiting for W&B process to finish, PID 11685
wandb: Program ended successfully.
wandb:                                                                                
wandb: Find user logs for this run at: /media/buckler/Data/Work/myproject/wandb/run-20201202_184802-1xhcuzib/logs/debug.log
wandb: Find internal logs for this run at: /media/buckler/Data/Work/myproject/wandb/run-20201202_184802-1xhcuzib/logs/debug-internal.log
wandb: Run summary:
wandb:   val_accuracy 0.27434
wandb:          _step 0
wandb:       _runtime 2
wandb:     _timestamp 1606931284
wandb: Run history:
wandb:   val_accuracy โ–
wandb:          _step โ–
wandb:       _runtime โ–
wandb:     _timestamp โ–
wandb: 
wandb: Synced 5 W&B file(s), 0 media file(s), 0 artifact file(s) and 1 other file(s)
wandb: 
wandb: Synced misunderstood-sweep-5-4: https://wandb.ai/buckler/myproject/runs/1xhcuzib

wandb: Waiting for W&B process to finish, PID 11420
wandb: Program ended successfully.
wandb:                                                                                
wandb: Find user logs for this run at: /media/buckler/Data/Work/myproject/wandb/run-20201202_184735-5fu8v7um/logs/debug.log
wandb: Find internal logs for this run at: /media/buckler/Data/Work/myproject/wandb/run-20201202_184735-5fu8v7um/logs/debug-internal.log
wandb: Run summary:
wandb:   val_accuracy 0.42289
wandb:          _step 0
wandb:       _runtime 38
wandb:     _timestamp 1606931293
wandb: Run history:
wandb:   val_accuracy โ–
wandb:          _step โ–
wandb:       _runtime โ–
wandb:     _timestamp โ–
wandb: 
wandb: Synced 6 W&B file(s), 0 media file(s), 0 artifact file(s) and 1 other file(s)
wandb: 
wandb: Synced misunderstood-sweep-5: https://wandb.ai/buckler/myproject/runs/5fu8v7um
****************************************
Sweep URL:        https://wandb.ai/buckler/myproject/sweeps/qmzmf6ms
Sweep Group URL:  https://wandb.ai/buckler/myproject/groups/qmzmf6ms
****************************************
2020-12-02 18:48:20,347 - wandb.wandb_agent - INFO - Cleaning up finished run: 5fu8v7um
2020-12-02 18:48:20,691 - wandb.wandb_agent - INFO - Agent received command: run
2020-12-02 18:48:20,691 - wandb.wandb_agent - INFO - Agent starting run with config:
        batch_size: 36
        epochs: 200
        node_size1: 256
        node_size2: 256
        node_size3: 64
        node_size4: 128
        node_size5: 64
        num_layers: 1
        optimizer: Adagrad
2020-12-02 18:48:20,704 - wandb.wandb_agent - INFO - About to run command: /usr/bin/env python test_sweeps.py --batch_size=36 --epochs=200 --node_size1=256 --node_size2=256 --node_size3=64 --node_size4=128 --node_size5=64 --num_layers=1 --optimizer=Adagrad
wandb: Currently logged in as: buckler (use `wandb login --relogin` to force relogin)
wandb: Tracking run with wandb version 0.10.11
wandb: Syncing run vivid-sweep-6
wandb: โญ๏ธ View project at https://wandb.ai/buckler/myproject
wandb: ๐Ÿงน View sweep at https://wandb.ai/buckler/myproject/sweeps/qmzmf6ms
wandb: ๐Ÿš€ View run at https://wandb.ai/buckler/myproject/runs/4jqres93
wandb: Run data is saved locally in /media/buckler/Data/Work/myproject/wandb/run-20201202_184822-4jqres93
wandb: Run `wandb offline` to turn off syncing.

wandb: WARNING Calling run.save without any arguments is deprecated.Changes to attributes are automatically persisted.
wandb: Currently logged in as: buckler (use `wandb login --relogin` to force relogin)
2020-12-02 18:48:25,714 - wandb.wandb_agent - INFO - Running runs: ['4jqres93']
wandb: Tracking run with wandb version 0.10.11
wandb: Syncing run vivid-sweep-6-0
wandb: โญ๏ธ View project at https://wandb.ai/buckler/myproject
wandb: ๐Ÿš€ View run at https://wandb.ai/buckler/myproject/runs/2e0m6mik
wandb: Run data is saved locally in /media/buckler/Data/Work/myproject/wandb/run-20201202_184825-2e0m6mik
wandb: Run `wandb offline` to turn off syncing.


wandb: Waiting for W&B process to finish, PID 11862
wandb: Program ended successfully.
wandb:                                                                                
wandb: Find user logs for this run at: /media/buckler/Data/Work/myproject/wandb/run-20201202_184825-2e0m6mik/logs/debug.log
wandb: Find internal logs for this run at: /media/buckler/Data/Work/myproject/wandb/run-20201202_184825-2e0m6mik/logs/debug-internal.log
wandb: Run summary:
wandb:   val_accuracy 0.83586
wandb:          _step 0
wandb:       _runtime 2
wandb:     _timestamp 1606931307
wandb: Run history:
wandb:   val_accuracy โ–
wandb:          _step โ–
wandb:       _runtime โ–
wandb:     _timestamp โ–
wandb: 
wandb: Synced 5 W&B file(s), 0 media file(s), 0 artifact file(s) and 1 other file(s)
wandb: 
wandb: Synced vivid-sweep-6-0: https://wandb.ai/buckler/myproject/runs/2e0m6mik
wandb: Currently logged in as: buckler (use `wandb login --relogin` to force relogin)
wandb: Tracking run with wandb version 0.10.11
wandb: Syncing run vivid-sweep-6-1
wandb: โญ๏ธ View project at https://wandb.ai/buckler/myproject
wandb: ๐Ÿš€ View run at https://wandb.ai/buckler/myproject/runs/dly0tuq1
wandb: Run data is saved locally in /media/buckler/Data/Work/myproject/wandb/run-20201202_184831-dly0tuq1
wandb: Run `wandb offline` to turn off syncing.


wandb: Waiting for W&B process to finish, PID 11918
wandb: Program ended successfully.
wandb:                                                                                
wandb: Find user logs for this run at: /media/buckler/Data/Work/myproject/wandb/run-20201202_184831-dly0tuq1/logs/debug.log
wandb: Find internal logs for this run at: /media/buckler/Data/Work/myproject/wandb/run-20201202_184831-dly0tuq1/logs/debug-internal.log
wandb: Run summary:
wandb:   val_accuracy 0.13115
wandb:          _step 0
wandb:       _runtime 2
wandb:     _timestamp 1606931313
wandb: Run history:
wandb:   val_accuracy โ–
wandb:          _step โ–
wandb:       _runtime โ–
wandb:     _timestamp โ–
wandb: 
wandb: Synced 5 W&B file(s), 0 media file(s), 0 artifact file(s) and 1 other file(s)
wandb: 
wandb: Synced vivid-sweep-6-1: https://wandb.ai/buckler/myproject/runs/dly0tuq1
wandb: Currently logged in as: buckler (use `wandb login --relogin` to force relogin)
wandb: Tracking run with wandb version 0.10.11
wandb: Syncing run vivid-sweep-6-2
wandb: โญ๏ธ View project at https://wandb.ai/buckler/myproject
wandb: ๐Ÿš€ View run at https://wandb.ai/buckler/myproject/runs/14sjv6ex
wandb: Run data is saved locally in /media/buckler/Data/Work/myproject/wandb/run-20201202_184837-14sjv6ex
wandb: Run `wandb offline` to turn off syncing.


wandb: Waiting for W&B process to finish, PID 11972
wandb: Program ended successfully.
wandb:                                                                                
wandb: Find user logs for this run at: /media/buckler/Data/Work/myproject/wandb/run-20201202_184837-14sjv6ex/logs/debug.log
wandb: Find internal logs for this run at: /media/buckler/Data/Work/myproject/wandb/run-20201202_184837-14sjv6ex/logs/debug-internal.log
wandb: Run summary:
wandb:   val_accuracy 0.45778
wandb:          _step 0
wandb:       _runtime 2
wandb:     _timestamp 1606931319
wandb: Run history:
wandb:   val_accuracy โ–
wandb:          _step โ–
wandb:       _runtime โ–
wandb:     _timestamp โ–
wandb: 
wandb: Synced 5 W&B file(s), 0 media file(s), 0 artifact file(s) and 1 other file(s)
wandb: 
wandb: Synced vivid-sweep-6-2: https://wandb.ai/buckler/myproject/runs/14sjv6ex
wandb: Currently logged in as: buckler (use `wandb login --relogin` to force relogin)
wandb: Tracking run with wandb version 0.10.11
wandb: Syncing run vivid-sweep-6-3
wandb: โญ๏ธ View project at https://wandb.ai/buckler/myproject
wandb: ๐Ÿš€ View run at https://wandb.ai/buckler/myproject/runs/1qzh8cf5
wandb: Run data is saved locally in /media/buckler/Data/Work/myproject/wandb/run-20201202_184843-1qzh8cf5
wandb: Run `wandb offline` to turn off syncing.


wandb: Waiting for W&B process to finish, PID 12037
wandb: Program ended successfully.
wandb:                                                                                
wandb: Find user logs for this run at: /media/buckler/Data/Work/myproject/wandb/run-20201202_184843-1qzh8cf5/logs/debug.log
wandb: Find internal logs for this run at: /media/buckler/Data/Work/myproject/wandb/run-20201202_184843-1qzh8cf5/logs/debug-internal.log
wandb: Run summary:
wandb:   val_accuracy 0.77422
wandb:          _step 0
wandb:       _runtime 1
wandb:     _timestamp 1606931325
wandb: Run history:
wandb:   val_accuracy โ–
wandb:          _step โ–
wandb:       _runtime โ–
wandb:     _timestamp โ–
wandb: 
wandb: Synced 5 W&B file(s), 0 media file(s), 0 artifact file(s) and 1 other file(s)
wandb: 
wandb: Synced vivid-sweep-6-3: https://wandb.ai/buckler/myproject/runs/1qzh8cf5
wandb: Currently logged in as: buckler (use `wandb login --relogin` to force relogin)
wandb: Tracking run with wandb version 0.10.11
wandb: Syncing run vivid-sweep-6-4
wandb: โญ๏ธ View project at https://wandb.ai/buckler/myproject
wandb: ๐Ÿš€ View run at https://wandb.ai/buckler/myproject/runs/1cc245wg
wandb: Run data is saved locally in /media/buckler/Data/Work/myproject/wandb/run-20201202_184849-1cc245wg
wandb: Run `wandb offline` to turn off syncing.


wandb: Waiting for W&B process to finish, PID 12091
wandb: Program ended successfully.
wandb:                                                                                
wandb: Find user logs for this run at: /media/buckler/Data/Work/myproject/wandb/run-20201202_184849-1cc245wg/logs/debug.log
wandb: Find internal logs for this run at: /media/buckler/Data/Work/myproject/wandb/run-20201202_184849-1cc245wg/logs/debug-internal.log
wandb: Run summary:
wandb:   val_accuracy 0.9598
wandb:          _step 0
wandb:       _runtime 1
wandb:     _timestamp 1606931331
wandb: Run history:
wandb:   val_accuracy โ–
wandb:          _step โ–
wandb:       _runtime โ–
wandb:     _timestamp โ–
wandb: 
wandb: Synced 5 W&B file(s), 0 media file(s), 0 artifact file(s) and 1 other file(s)
wandb: 
wandb: Synced vivid-sweep-6-4: https://wandb.ai/buckler/myproject/runs/1cc245wg

wandb: Waiting for W&B process to finish, PID 11808
wandb: Program ended successfully.
wandb:                                                                                
wandb: Find user logs for this run at: /media/buckler/Data/Work/myproject/wandb/run-20201202_184822-4jqres93/logs/debug.log
wandb: Find internal logs for this run at: /media/buckler/Data/Work/myproject/wandb/run-20201202_184822-4jqres93/logs/debug-internal.log
wandb: Run summary:
wandb:   val_accuracy 0.63176
wandb:          _step 0
wandb:       _runtime 33
wandb:     _timestamp 1606931335
wandb: Run history:
wandb:   val_accuracy โ–
wandb:          _step โ–
wandb:       _runtime โ–
wandb:     _timestamp โ–
wandb: 
wandb: Synced 6 W&B file(s), 0 media file(s), 0 artifact file(s) and 1 other file(s)
wandb: 
wandb: Synced vivid-sweep-6: https://wandb.ai/buckler/myproject/runs/4jqres93
****************************************
Sweep URL:        https://wandb.ai/buckler/myproject/sweeps/qmzmf6ms
Sweep Group URL:  https://wandb.ai/buckler/myproject/groups/qmzmf6ms
****************************************
2020-12-02 18:49:02,387 - wandb.wandb_agent - INFO - Cleaning up finished run: 4jqres93
2020-12-02 18:49:02,993 - wandb.wandb_agent - INFO - Agent received command: run
2020-12-02 18:49:02,993 - wandb.wandb_agent - INFO - Agent starting run with config:
        batch_size: 118
        epochs: 200
        node_size1: 256
        node_size2: 64
        node_size3: 64
        node_size4: 128
        node_size5: 64
        num_layers: 1
        optimizer: Adamax
2020-12-02 18:49:02,996 - wandb.wandb_agent - INFO - About to run command: /usr/bin/env python test_sweeps.py --batch_size=118 --epochs=200 --node_size1=256 --node_size2=64 --node_size3=64 --node_size4=128 --node_size5=64 --num_layers=1 --optimizer=Adamax
wandb: Currently logged in as: buckler (use `wandb login --relogin` to force relogin)
wandb: Tracking run with wandb version 0.10.11
wandb: Syncing run young-sweep-7
wandb: โญ๏ธ View project at https://wandb.ai/buckler/myproject
wandb: ๐Ÿงน View sweep at https://wandb.ai/buckler/myproject/sweeps/qmzmf6ms
wandb: ๐Ÿš€ View run at https://wandb.ai/buckler/myproject/runs/6twy01ey
wandb: Run data is saved locally in /media/buckler/Data/Work/myproject/wandb/run-20201202_184904-6twy01ey
wandb: Run `wandb offline` to turn off syncing.

wandb: WARNING Calling run.save without any arguments is deprecated.Changes to attributes are automatically persisted.
wandb: Currently logged in as: buckler (use `wandb login --relogin` to force relogin)
2020-12-02 18:49:08,002 - wandb.wandb_agent - INFO - Running runs: ['6twy01ey']
wandb: Tracking run with wandb version 0.10.11
wandb: Syncing run young-sweep-7-0
wandb: โญ๏ธ View project at https://wandb.ai/buckler/myproject
wandb: ๐Ÿš€ View run at https://wandb.ai/buckler/myproject/runs/1mduow2i
wandb: Run data is saved locally in /media/buckler/Data/Work/myproject/wandb/run-20201202_184907-1mduow2i
wandb: Run `wandb offline` to turn off syncing.


wandb: Waiting for W&B process to finish, PID 12215
wandb: Program ended successfully.
wandb:                                                                                
wandb: Find user logs for this run at: /media/buckler/Data/Work/myproject/wandb/run-20201202_184907-1mduow2i/logs/debug.log
wandb: Find internal logs for this run at: /media/buckler/Data/Work/myproject/wandb/run-20201202_184907-1mduow2i/logs/debug-internal.log
wandb: Run summary:
wandb:   val_accuracy 0.41597
wandb:          _step 0
wandb:       _runtime 2
wandb:     _timestamp 1606931349
wandb: Run history:
wandb:   val_accuracy โ–
wandb:          _step โ–
wandb:       _runtime โ–
wandb:     _timestamp โ–
wandb: 
wandb: Synced 5 W&B file(s), 0 media file(s), 0 artifact file(s) and 1 other file(s)
wandb: 
wandb: Synced young-sweep-7-0: https://wandb.ai/buckler/myproject/runs/1mduow2i
wandb: Currently logged in as: buckler (use `wandb login --relogin` to force relogin)
wandb: Tracking run with wandb version 0.10.11
wandb: Syncing run young-sweep-7-1
wandb: โญ๏ธ View project at https://wandb.ai/buckler/myproject
wandb: ๐Ÿš€ View run at https://wandb.ai/buckler/myproject/runs/2rt8oxbo
wandb: Run data is saved locally in /media/buckler/Data/Work/myproject/wandb/run-20201202_184913-2rt8oxbo
wandb: Run `wandb offline` to turn off syncing.


wandb: Waiting for W&B process to finish, PID 12267
wandb: Program ended successfully.
wandb:                                                                                
wandb: Find user logs for this run at: /media/buckler/Data/Work/myproject/wandb/run-20201202_184913-2rt8oxbo/logs/debug.log
wandb: Find internal logs for this run at: /media/buckler/Data/Work/myproject/wandb/run-20201202_184913-2rt8oxbo/logs/debug-internal.log
wandb: Run summary:
wandb:   val_accuracy 0.2212
wandb:          _step 0
wandb:       _runtime 2
wandb:     _timestamp 1606931355
wandb: Run history:
wandb:   val_accuracy โ–
wandb:          _step โ–
wandb:       _runtime โ–
wandb:     _timestamp โ–
wandb: 
wandb: Synced 5 W&B file(s), 0 media file(s), 0 artifact file(s) and 1 other file(s)
wandb: 
wandb: Synced young-sweep-7-1: https://wandb.ai/buckler/myproject/runs/2rt8oxbo
wandb: Currently logged in as: buckler (use `wandb login --relogin` to force relogin)
wandb: Tracking run with wandb version 0.10.11
wandb: Syncing run young-sweep-7-2
wandb: โญ๏ธ View project at https://wandb.ai/buckler/myproject
wandb: ๐Ÿš€ View run at https://wandb.ai/buckler/myproject/runs/37oxxkxq
wandb: Run data is saved locally in /media/buckler/Data/Work/myproject/wandb/run-20201202_184919-37oxxkxq
wandb: Run `wandb offline` to turn off syncing.


wandb: Waiting for W&B process to finish, PID 12319
wandb: Program ended successfully.
wandb:                                                                                
wandb: Find user logs for this run at: /media/buckler/Data/Work/myproject/wandb/run-20201202_184919-37oxxkxq/logs/debug.log
wandb: Find internal logs for this run at: /media/buckler/Data/Work/myproject/wandb/run-20201202_184919-37oxxkxq/logs/debug-internal.log
wandb: Run summary:
wandb:   val_accuracy 0.45917
wandb:          _step 0
wandb:       _runtime 2
wandb:     _timestamp 1606931361
wandb: Run history:
wandb:   val_accuracy โ–
wandb:          _step โ–
wandb:       _runtime โ–
wandb:     _timestamp โ–
wandb: 
wandb: Synced 5 W&B file(s), 0 media file(s), 0 artifact file(s) and 1 other file(s)
wandb: 
wandb: Synced young-sweep-7-2: https://wandb.ai/buckler/myproject/runs/37oxxkxq
wandb: Currently logged in as: buckler (use `wandb login --relogin` to force relogin)
wandb: Tracking run with wandb version 0.10.11
wandb: Syncing run young-sweep-7-3
wandb: โญ๏ธ View project at https://wandb.ai/buckler/myproject
wandb: ๐Ÿš€ View run at https://wandb.ai/buckler/myproject/runs/2w39ll7j
wandb: Run data is saved locally in /media/buckler/Data/Work/myproject/wandb/run-20201202_184925-2w39ll7j
wandb: Run `wandb offline` to turn off syncing.


wandb: Waiting for W&B process to finish, PID 12371
wandb: Program ended successfully.
wandb:                                                                                
wandb: Find user logs for this run at: /media/buckler/Data/Work/myproject/wandb/run-20201202_184925-2w39ll7j/logs/debug.log
wandb: Find internal logs for this run at: /media/buckler/Data/Work/myproject/wandb/run-20201202_184925-2w39ll7j/logs/debug-internal.log
wandb: Run summary:
wandb:   val_accuracy 0.2997
wandb:          _step 0
wandb:       _runtime 2
wandb:     _timestamp 1606931367
wandb: Run history:
wandb:   val_accuracy โ–
wandb:          _step โ–
wandb:       _runtime โ–
wandb:     _timestamp โ–
wandb: 
wandb: Synced 5 W&B file(s), 0 media file(s), 0 artifact file(s) and 1 other file(s)
wandb: 
wandb: Synced young-sweep-7-3: https://wandb.ai/buckler/myproject/runs/2w39ll7j
wandb: Currently logged in as: buckler (use `wandb login --relogin` to force relogin)
wandb: Tracking run with wandb version 0.10.11
wandb: Syncing run young-sweep-7-4
wandb: โญ๏ธ View project at https://wandb.ai/buckler/myproject
wandb: ๐Ÿš€ View run at https://wandb.ai/buckler/myproject/runs/3gylfd7z
wandb: Run data is saved locally in /media/buckler/Data/Work/myproject/wandb/run-20201202_184931-3gylfd7z
wandb: Run `wandb offline` to turn off syncing.


wandb: Waiting for W&B process to finish, PID 12423
wandb: Program ended successfully.
wandb:                                                                                
wandb: Find user logs for this run at: /media/buckler/Data/Work/myproject/wandb/run-20201202_184931-3gylfd7z/logs/debug.log
wandb: Find internal logs for this run at: /media/buckler/Data/Work/myproject/wandb/run-20201202_184931-3gylfd7z/logs/debug-internal.log
wandb: Run summary:
wandb:   val_accuracy 0.41683
wandb:          _step 0
wandb:       _runtime 2
wandb:     _timestamp 1606931373
wandb: Run history:
wandb:   val_accuracy โ–
wandb:          _step โ–
wandb:       _runtime โ–
wandb:     _timestamp โ–
wandb: 
wandb: Synced 5 W&B file(s), 0 media file(s), 0 artifact file(s) and 1 other file(s)
wandb: 
wandb: Synced young-sweep-7-4: https://wandb.ai/buckler/myproject/runs/3gylfd7z

wandb: Waiting for W&B process to finish, PID 12166
wandb: Program ended successfully.
wandb:                                                                                
wandb: Find user logs for this run at: /media/buckler/Data/Work/myproject/wandb/run-20201202_184904-6twy01ey/logs/debug.log
wandb: Find internal logs for this run at: /media/buckler/Data/Work/myproject/wandb/run-20201202_184904-6twy01ey/logs/debug-internal.log
wandb: Run summary:
wandb:   val_accuracy 0.36257
wandb:          _step 0
wandb:       _runtime 32
wandb:     _timestamp 1606931376
wandb: Run history:
wandb:   val_accuracy โ–
wandb:          _step โ–
wandb:       _runtime โ–
wandb:     _timestamp โ–
wandb: 
wandb: Synced 6 W&B file(s), 0 media file(s), 0 artifact file(s) and 1 other file(s)
wandb: 
wandb: Synced young-sweep-7: https://wandb.ai/buckler/myproject/runs/6twy01ey
****************************************
Sweep URL:        https://wandb.ai/buckler/myproject/sweeps/qmzmf6ms
Sweep Group URL:  https://wandb.ai/buckler/myproject/groups/qmzmf6ms
****************************************
2020-12-02 18:49:44,658 - wandb.wandb_agent - INFO - Cleaning up finished run: 6twy01ey
2020-12-02 18:49:45,007 - wandb.wandb_agent - INFO - Agent received command: run
2020-12-02 18:49:45,008 - wandb.wandb_agent - INFO - Agent starting run with config:
        batch_size: 111
        epochs: 200
        node_size1: 64
        node_size2: 256
        node_size3: 256
        node_size4: 128
        node_size5: 64
        num_layers: 2
        optimizer: Adamax
2020-12-02 18:49:45,016 - wandb.wandb_agent - INFO - About to run command: /usr/bin/env python test_sweeps.py --batch_size=111 --epochs=200 --node_size1=64 --node_size2=256 --node_size3=256 --node_size4=128 --node_size5=64 --num_layers=2 --optimizer=Adamax
wandb: Currently logged in as: buckler (use `wandb login --relogin` to force relogin)
wandb: Tracking run with wandb version 0.10.11
wandb: Syncing run deep-sweep-8
wandb: โญ๏ธ View project at https://wandb.ai/buckler/myproject
wandb: ๐Ÿงน View sweep at https://wandb.ai/buckler/myproject/sweeps/qmzmf6ms
wandb: ๐Ÿš€ View run at https://wandb.ai/buckler/myproject/runs/i79bdos6
wandb: Run data is saved locally in /media/buckler/Data/Work/myproject/wandb/run-20201202_184946-i79bdos6
wandb: Run `wandb offline` to turn off syncing.

wandb: WARNING Calling run.save without any arguments is deprecated.Changes to attributes are automatically persisted.
wandb: Currently logged in as: buckler (use `wandb login --relogin` to force relogin)
2020-12-02 18:49:50,032 - wandb.wandb_agent - INFO - Running runs: ['i79bdos6']
wandb: Tracking run with wandb version 0.10.11
wandb: Syncing run deep-sweep-8-0
wandb: โญ๏ธ View project at https://wandb.ai/buckler/myproject
wandb: ๐Ÿš€ View run at https://wandb.ai/buckler/myproject/runs/3khd64s8
wandb: Run data is saved locally in /media/buckler/Data/Work/myproject/wandb/run-20201202_184949-3khd64s8
wandb: Run `wandb offline` to turn off syncing.


wandb: Waiting for W&B process to finish, PID 12547
wandb: Program ended successfully.
wandb:                                                                                
wandb: Find user logs for this run at: /media/buckler/Data/Work/myproject/wandb/run-20201202_184949-3khd64s8/logs/debug.log
wandb: Find internal logs for this run at: /media/buckler/Data/Work/myproject/wandb/run-20201202_184949-3khd64s8/logs/debug-internal.log
wandb: Run summary:
wandb:   val_accuracy 0.48109
wandb:          _step 0
wandb:       _runtime 2
wandb:     _timestamp 1606931391
wandb: Run history:
wandb:   val_accuracy โ–
wandb:          _step โ–
wandb:       _runtime โ–
wandb:     _timestamp โ–
wandb: 
wandb: Synced 5 W&B file(s), 0 media file(s), 0 artifact file(s) and 1 other file(s)
wandb: 
wandb: Synced deep-sweep-8-0: https://wandb.ai/buckler/myproject/runs/3khd64s8
wandb: Currently logged in as: buckler (use `wandb login --relogin` to force relogin)
wandb: Tracking run with wandb version 0.10.11
wandb: Syncing run deep-sweep-8-1
wandb: โญ๏ธ View project at https://wandb.ai/buckler/myproject
wandb: ๐Ÿš€ View run at https://wandb.ai/buckler/myproject/runs/1byk9d0l
wandb: Run data is saved locally in /media/buckler/Data/Work/myproject/wandb/run-20201202_184955-1byk9d0l
wandb: Run `wandb offline` to turn off syncing.


wandb: Waiting for W&B process to finish, PID 12601
wandb: Program ended successfully.
wandb:                                                                                
wandb: Find user logs for this run at: /media/buckler/Data/Work/myproject/wandb/run-20201202_184955-1byk9d0l/logs/debug.log
wandb: Find internal logs for this run at: /media/buckler/Data/Work/myproject/wandb/run-20201202_184955-1byk9d0l/logs/debug-internal.log
wandb: Run summary:
wandb:   val_accuracy 0.44693
wandb:          _step 0
wandb:       _runtime 2
wandb:     _timestamp 1606931397
wandb: Run history:
wandb:   val_accuracy โ–
wandb:          _step โ–
wandb:       _runtime โ–
wandb:     _timestamp โ–
wandb: 
wandb: Synced 5 W&B file(s), 0 media file(s), 0 artifact file(s) and 1 other file(s)
wandb: 
wandb: Synced deep-sweep-8-1: https://wandb.ai/buckler/myproject/runs/1byk9d0l
wandb: Currently logged in as: buckler (use `wandb login --relogin` to force relogin)
wandb: Tracking run with wandb version 0.10.11
wandb: Syncing run deep-sweep-8-2
wandb: โญ๏ธ View project at https://wandb.ai/buckler/myproject
wandb: ๐Ÿš€ View run at https://wandb.ai/buckler/myproject/runs/2li2imns
wandb: Run data is saved locally in /media/buckler/Data/Work/myproject/wandb/run-20201202_185001-2li2imns
wandb: Run `wandb offline` to turn off syncing.


wandb: Waiting for W&B process to finish, PID 12653
wandb: Program ended successfully.
wandb:                                                                                
wandb: Find user logs for this run at: /media/buckler/Data/Work/myproject/wandb/run-20201202_185001-2li2imns/logs/debug.log
wandb: Find internal logs for this run at: /media/buckler/Data/Work/myproject/wandb/run-20201202_185001-2li2imns/logs/debug-internal.log
wandb: Run summary:
wandb:   val_accuracy 0.50812
wandb:          _step 0
wandb:       _runtime 2
wandb:     _timestamp 1606931403
wandb: Run history:
wandb:   val_accuracy โ–
wandb:          _step โ–
wandb:       _runtime โ–
wandb:     _timestamp โ–
wandb: 
wandb: Synced 5 W&B file(s), 0 media file(s), 0 artifact file(s) and 1 other file(s)
wandb: 
wandb: Synced deep-sweep-8-2: https://wandb.ai/buckler/myproject/runs/2li2imns
wandb: Currently logged in as: buckler (use `wandb login --relogin` to force relogin)
wandb: Tracking run with wandb version 0.10.11
wandb: Syncing run deep-sweep-8-3
wandb: โญ๏ธ View project at https://wandb.ai/buckler/myproject
wandb: ๐Ÿš€ View run at https://wandb.ai/buckler/myproject/runs/2y7278t7
wandb: Run data is saved locally in /media/buckler/Data/Work/myproject/wandb/run-20201202_185007-2y7278t7
wandb: Run `wandb offline` to turn off syncing.


wandb: Waiting for W&B process to finish, PID 12705
wandb: Program ended successfully.
wandb:                                                                                
wandb: Find user logs for this run at: /media/buckler/Data/Work/myproject/wandb/run-20201202_185007-2y7278t7/logs/debug.log
wandb: Find internal logs for this run at: /media/buckler/Data/Work/myproject/wandb/run-20201202_185007-2y7278t7/logs/debug-internal.log
wandb: Run summary:
wandb:   val_accuracy 0.05146
wandb:          _step 0
wandb:       _runtime 2
wandb:     _timestamp 1606931409
wandb: Run history:
wandb:   val_accuracy โ–
wandb:          _step โ–
wandb:       _runtime โ–
wandb:     _timestamp โ–
wandb: 
wandb: Synced 5 W&B file(s), 0 media file(s), 0 artifact file(s) and 1 other file(s)
wandb: 
wandb: Synced deep-sweep-8-3: https://wandb.ai/buckler/myproject/runs/2y7278t7
wandb: Currently logged in as: buckler (use `wandb login --relogin` to force relogin)
wandb: Tracking run with wandb version 0.10.11
wandb: Syncing run deep-sweep-8-4
wandb: โญ๏ธ View project at https://wandb.ai/buckler/myproject
wandb: ๐Ÿš€ View run at https://wandb.ai/buckler/myproject/runs/3qripv0j
wandb: Run data is saved locally in /media/buckler/Data/Work/myproject/wandb/run-20201202_185013-3qripv0j
wandb: Run `wandb offline` to turn off syncing.


wandb: Waiting for W&B process to finish, PID 12757
wandb: Program ended successfully.
wandb:                                                                                
wandb: Find user logs for this run at: /media/buckler/Data/Work/myproject/wandb/run-20201202_185013-3qripv0j/logs/debug.log
wandb: Find internal logs for this run at: /media/buckler/Data/Work/myproject/wandb/run-20201202_185013-3qripv0j/logs/debug-internal.log
wandb: Run summary:
wandb:   val_accuracy 0.40922
wandb:          _step 0
wandb:       _runtime 2
wandb:     _timestamp 1606931415
wandb: Run history:
wandb:   val_accuracy โ–
wandb:          _step โ–
wandb:       _runtime โ–
wandb:     _timestamp โ–
wandb: 
wandb: Synced 5 W&B file(s), 0 media file(s), 0 artifact file(s) and 1 other file(s)
wandb: 
wandb: Synced deep-sweep-8-4: https://wandb.ai/buckler/myproject/runs/3qripv0j

wandb: Waiting for W&B process to finish, PID 12498
wandb: Program ended successfully.
wandb:                                                                                
wandb: Find user logs for this run at: /media/buckler/Data/Work/myproject/wandb/run-20201202_184946-i79bdos6/logs/debug.log
wandb: Find internal logs for this run at: /media/buckler/Data/Work/myproject/wandb/run-20201202_184946-i79bdos6/logs/debug-internal.log
wandb: Run summary:
wandb:   val_accuracy 0.37936
wandb:          _step 0
wandb:       _runtime 32
wandb:     _timestamp 1606931419
wandb: Run history:
wandb:   val_accuracy โ–
wandb:          _step โ–
wandb:       _runtime โ–
wandb:     _timestamp โ–
wandb: 
wandb: Synced 6 W&B file(s), 0 media file(s), 0 artifact file(s) and 1 other file(s)
wandb: 
wandb: Synced deep-sweep-8: https://wandb.ai/buckler/myproject/runs/i79bdos6
****************************************
Sweep URL:        https://wandb.ai/buckler/myproject/sweeps/qmzmf6ms
Sweep Group URL:  https://wandb.ai/buckler/myproject/groups/qmzmf6ms
****************************************
2020-12-02 18:50:26,738 - wandb.wandb_agent - INFO - Cleaning up finished run: i79bdos6
2020-12-02 18:50:27,119 - wandb.wandb_agent - INFO - Agent received command: run
2020-12-02 18:50:27,119 - wandb.wandb_agent - INFO - Agent starting run with config:
        batch_size: 54
        epochs: 200
        node_size1: 256
        node_size2: 128
        node_size3: 128
        node_size4: 64
        node_size5: 128
        num_layers: 3
        optimizer: Adagrad
2020-12-02 18:50:27,127 - wandb.wandb_agent - INFO - About to run command: /usr/bin/env python test_sweeps.py --batch_size=54 --epochs=200 --node_size1=256 --node_size2=128 --node_size3=128 --node_size4=64 --node_size5=128 --num_layers=3 --optimizer=Adagrad
wandb: Currently logged in as: buckler (use `wandb login --relogin` to force relogin)
wandb: Tracking run with wandb version 0.10.11
wandb: Syncing run misunderstood-sweep-9
wandb: โญ๏ธ View project at https://wandb.ai/buckler/myproject
wandb: ๐Ÿงน View sweep at https://wandb.ai/buckler/myproject/sweeps/qmzmf6ms
wandb: ๐Ÿš€ View run at https://wandb.ai/buckler/myproject/runs/y88eytn4
wandb: Run data is saved locally in /media/buckler/Data/Work/myproject/wandb/run-20201202_185029-y88eytn4
wandb: Run `wandb offline` to turn off syncing.

wandb: WARNING Calling run.save without any arguments is deprecated.Changes to attributes are automatically persisted.
wandb: Currently logged in as: buckler (use `wandb login --relogin` to force relogin)
2020-12-02 18:50:32,138 - wandb.wandb_agent - INFO - Running runs: ['y88eytn4']
wandb: Tracking run with wandb version 0.10.11
wandb: Syncing run misunderstood-sweep-9-0
wandb: โญ๏ธ View project at https://wandb.ai/buckler/myproject
wandb: ๐Ÿš€ View run at https://wandb.ai/buckler/myproject/runs/27jl4a8d
wandb: Run data is saved locally in /media/buckler/Data/Work/myproject/wandb/run-20201202_185031-27jl4a8d
wandb: Run `wandb offline` to turn off syncing.


wandb: Waiting for W&B process to finish, PID 12898
wandb: Program ended successfully.
wandb:                                                                                
wandb: Find user logs for this run at: /media/buckler/Data/Work/myproject/wandb/run-20201202_185031-27jl4a8d/logs/debug.log
wandb: Find internal logs for this run at: /media/buckler/Data/Work/myproject/wandb/run-20201202_185031-27jl4a8d/logs/debug-internal.log
wandb: Run summary:
wandb:   val_accuracy 0.1863
wandb:          _step 0
wandb:       _runtime 2
wandb:     _timestamp 1606931433
wandb: Run history:
wandb:   val_accuracy โ–
wandb:          _step โ–
wandb:       _runtime โ–
wandb:     _timestamp โ–
wandb: 
wandb: Synced 5 W&B file(s), 0 media file(s), 0 artifact file(s) and 1 other file(s)
wandb: 
wandb: Synced misunderstood-sweep-9-0: https://wandb.ai/buckler/myproject/runs/27jl4a8d
wandb: Currently logged in as: buckler (use `wandb login --relogin` to force relogin)
wandb: Tracking run with wandb version 0.10.11
wandb: Syncing run misunderstood-sweep-9-1
wandb: โญ๏ธ View project at https://wandb.ai/buckler/myproject
wandb: ๐Ÿš€ View run at https://wandb.ai/buckler/myproject/runs/1gkv9t4l
wandb: Run data is saved locally in /media/buckler/Data/Work/myproject/wandb/run-20201202_185038-1gkv9t4l
wandb: Run `wandb offline` to turn off syncing.


wandb: Waiting for W&B process to finish, PID 12951
wandb: Program ended successfully.
wandb:                                                                                
wandb: Find user logs for this run at: /media/buckler/Data/Work/myproject/wandb/run-20201202_185038-1gkv9t4l/logs/debug.log
wandb: Find internal logs for this run at: /media/buckler/Data/Work/myproject/wandb/run-20201202_185038-1gkv9t4l/logs/debug-internal.log
wandb: Run summary:
wandb:   val_accuracy 0.78662
wandb:          _step 0
wandb:       _runtime 1
wandb:     _timestamp 1606931439
wandb: Run history:
wandb:   val_accuracy โ–
wandb:          _step โ–
wandb:       _runtime โ–
wandb:     _timestamp โ–
wandb: 
wandb: Synced 5 W&B file(s), 0 media file(s), 0 artifact file(s) and 1 other file(s)
wandb: 
wandb: Synced misunderstood-sweep-9-1: https://wandb.ai/buckler/myproject/runs/1gkv9t4l
wandb: Currently logged in as: buckler (use `wandb login --relogin` to force relogin)
wandb: Tracking run with wandb version 0.10.11
wandb: Syncing run misunderstood-sweep-9-2
wandb: โญ๏ธ View project at https://wandb.ai/buckler/myproject
wandb: ๐Ÿš€ View run at https://wandb.ai/buckler/myproject/runs/1a45kfud
wandb: Run data is saved locally in /media/buckler/Data/Work/myproject/wandb/run-20201202_185043-1a45kfud
wandb: Run `wandb offline` to turn off syncing.


wandb: Waiting for W&B process to finish, PID 13004
wandb: Program ended successfully.
wandb:                                                                                
wandb: Find user logs for this run at: /media/buckler/Data/Work/myproject/wandb/run-20201202_185043-1a45kfud/logs/debug.log
wandb: Find internal logs for this run at: /media/buckler/Data/Work/myproject/wandb/run-20201202_185043-1a45kfud/logs/debug-internal.log
wandb: Run summary:
wandb:   val_accuracy 0.15833
wandb:          _step 0
wandb:       _runtime 1
wandb:     _timestamp 1606931445
wandb: Run history:
wandb:   val_accuracy โ–
wandb:          _step โ–
wandb:       _runtime โ–
wandb:     _timestamp โ–
wandb: 
wandb: Synced 5 W&B file(s), 0 media file(s), 0 artifact file(s) and 1 other file(s)
wandb: 
wandb: Synced misunderstood-sweep-9-2: https://wandb.ai/buckler/myproject/runs/1a45kfud
wandb: Currently logged in as: buckler (use `wandb login --relogin` to force relogin)
wandb: Tracking run with wandb version 0.10.11
wandb: Syncing run misunderstood-sweep-9-3
wandb: โญ๏ธ View project at https://wandb.ai/buckler/myproject
wandb: ๐Ÿš€ View run at https://wandb.ai/buckler/myproject/runs/gyn269lc
wandb: Run data is saved locally in /media/buckler/Data/Work/myproject/wandb/run-20201202_185049-gyn269lc
wandb: Run `wandb offline` to turn off syncing.


wandb: Waiting for W&B process to finish, PID 13072
wandb: Program ended successfully.
wandb:                                                                                
wandb: Find user logs for this run at: /media/buckler/Data/Work/myproject/wandb/run-20201202_185049-gyn269lc/logs/debug.log
wandb: Find internal logs for this run at: /media/buckler/Data/Work/myproject/wandb/run-20201202_185049-gyn269lc/logs/debug-internal.log
wandb: Run summary:
wandb:   val_accuracy 0.79449
wandb:          _step 0
wandb:       _runtime 2
wandb:     _timestamp 1606931451
wandb: Run history:
wandb:   val_accuracy โ–
wandb:          _step โ–
wandb:       _runtime โ–
wandb:     _timestamp โ–
wandb: 
wandb: Synced 5 W&B file(s), 0 media file(s), 0 artifact file(s) and 1 other file(s)
wandb: 
wandb: Synced misunderstood-sweep-9-3: https://wandb.ai/buckler/myproject/runs/gyn269lc
wandb: Currently logged in as: buckler (use `wandb login --relogin` to force relogin)
wandb: Tracking run with wandb version 0.10.11
wandb: Syncing run misunderstood-sweep-9-4
wandb: โญ๏ธ View project at https://wandb.ai/buckler/myproject
wandb: ๐Ÿš€ View run at https://wandb.ai/buckler/myproject/runs/20832hv4
wandb: Run data is saved locally in /media/buckler/Data/Work/myproject/wandb/run-20201202_185055-20832hv4
wandb: Run `wandb offline` to turn off syncing.


wandb: Waiting for W&B process to finish, PID 13128
wandb: Program ended successfully.
wandb:                                                                                
wandb: Find user logs for this run at: /media/buckler/Data/Work/myproject/wandb/run-20201202_185055-20832hv4/logs/debug.log
wandb: Find internal logs for this run at: /media/buckler/Data/Work/myproject/wandb/run-20201202_185055-20832hv4/logs/debug-internal.log
wandb: Run summary:
wandb:   val_accuracy 0.59133
wandb:          _step 0
wandb:       _runtime 2
wandb:     _timestamp 1606931457
wandb: Run history:
wandb:   val_accuracy โ–
wandb:          _step โ–
wandb:       _runtime โ–
wandb:     _timestamp โ–
wandb: 
wandb: Synced 5 W&B file(s), 0 media file(s), 0 artifact file(s) and 1 other file(s)
wandb: 
wandb: Synced misunderstood-sweep-9-4: https://wandb.ai/buckler/myproject/runs/20832hv4

wandb: Waiting for W&B process to finish, PID 12849
wandb: Program ended successfully.
wandb:                                                                                
wandb: Find user logs for this run at: /media/buckler/Data/Work/myproject/wandb/run-20201202_185029-y88eytn4/logs/debug.log
wandb: Find internal logs for this run at: /media/buckler/Data/Work/myproject/wandb/run-20201202_185029-y88eytn4/logs/debug-internal.log
wandb: Run summary:
wandb:   val_accuracy 0.50341
wandb:          _step 0
wandb:       _runtime 32
wandb:     _timestamp 1606931461
wandb: Run history:
wandb:   val_accuracy โ–
wandb:          _step โ–
wandb:       _runtime โ–
wandb:     _timestamp โ–
wandb: 
wandb: Synced 6 W&B file(s), 0 media file(s), 0 artifact file(s) and 1 other file(s)
wandb: 
wandb: Synced misunderstood-sweep-9: https://wandb.ai/buckler/myproject/runs/y88eytn4
****************************************
Sweep URL:        https://wandb.ai/buckler/myproject/sweeps/qmzmf6ms
Sweep Group URL:  https://wandb.ai/buckler/myproject/groups/qmzmf6ms
****************************************
2020-12-02 18:51:08,755 - wandb.wandb_agent - INFO - Cleaning up finished run: y88eytn4
2020-12-02 18:51:09,153 - wandb.wandb_agent - INFO - Agent received command: run
2020-12-02 18:51:09,153 - wandb.wandb_agent - INFO - Agent starting run with config:
        batch_size: 92
        epochs: 200
        node_size1: 128
        node_size2: 128
        node_size3: 64
        node_size4: 256
        node_size5: 128
        num_layers: 3
        optimizer: Adamax
2020-12-02 18:51:09,161 - wandb.wandb_agent - INFO - About to run command: /usr/bin/env python test_sweeps.py --batch_size=92 --epochs=200 --node_size1=128 --node_size2=128 --node_size3=64 --node_size4=256 --node_size5=128 --num_layers=3 --optimizer=Adamax
wandb: Currently logged in as: buckler (use `wandb login --relogin` to force relogin)
wandb: Tracking run with wandb version 0.10.11
wandb: Syncing run visionary-sweep-10
wandb: โญ๏ธ View project at https://wandb.ai/buckler/myproject
wandb: ๐Ÿงน View sweep at https://wandb.ai/buckler/myproject/sweeps/qmzmf6ms
wandb: ๐Ÿš€ View run at https://wandb.ai/buckler/myproject/runs/keofgifo
wandb: Run data is saved locally in /media/buckler/Data/Work/myproject/wandb/run-20201202_185111-keofgifo
wandb: Run `wandb offline` to turn off syncing.

wandb: WARNING Calling run.save without any arguments is deprecated.Changes to attributes are automatically persisted.
wandb: Currently logged in as: buckler (use `wandb login --relogin` to force relogin)
2020-12-02 18:51:14,174 - wandb.wandb_agent - INFO - Running runs: ['keofgifo']
wandb: Tracking run with wandb version 0.10.11
wandb: Syncing run visionary-sweep-10-0
wandb: โญ๏ธ View project at https://wandb.ai/buckler/myproject
wandb: ๐Ÿš€ View run at https://wandb.ai/buckler/myproject/runs/13ts6smg
wandb: Run data is saved locally in /media/buckler/Data/Work/myproject/wandb/run-20201202_185113-13ts6smg
wandb: Run `wandb offline` to turn off syncing.


wandb: Waiting for W&B process to finish, PID 13288
wandb: Program ended successfully.
wandb:                                                                                
wandb: Find user logs for this run at: /media/buckler/Data/Work/myproject/wandb/run-20201202_185113-13ts6smg/logs/debug.log
wandb: Find internal logs for this run at: /media/buckler/Data/Work/myproject/wandb/run-20201202_185113-13ts6smg/logs/debug-internal.log
wandb: Run summary:
wandb:   val_accuracy 0.90107
wandb:          _step 0
wandb:       _runtime 2
wandb:     _timestamp 1606931475
wandb: Run history:
wandb:   val_accuracy โ–
wandb:          _step โ–
wandb:       _runtime โ–
wandb:     _timestamp โ–
wandb: 
wandb: Synced 5 W&B file(s), 0 media file(s), 0 artifact file(s) and 1 other file(s)
wandb: 
wandb: Synced visionary-sweep-10-0: https://wandb.ai/buckler/myproject/runs/13ts6smg
wandb: Currently logged in as: buckler (use `wandb login --relogin` to force relogin)
wandb: Tracking run with wandb version 0.10.11
wandb: Syncing run visionary-sweep-10-1
wandb: โญ๏ธ View project at https://wandb.ai/buckler/myproject
wandb: ๐Ÿš€ View run at https://wandb.ai/buckler/myproject/runs/v40fbbho
wandb: Run data is saved locally in /media/buckler/Data/Work/myproject/wandb/run-20201202_185119-v40fbbho
wandb: Run `wandb offline` to turn off syncing.


wandb: Waiting for W&B process to finish, PID 13354
wandb: Program ended successfully.
wandb:                                                                                
wandb: Find user logs for this run at: /media/buckler/Data/Work/myproject/wandb/run-20201202_185119-v40fbbho/logs/debug.log
wandb: Find internal logs for this run at: /media/buckler/Data/Work/myproject/wandb/run-20201202_185119-v40fbbho/logs/debug-internal.log
wandb: Run summary:
wandb:   val_accuracy 0.24336
wandb:          _step 0
wandb:       _runtime 2
wandb:     _timestamp 1606931481
wandb: Run history:
wandb:   val_accuracy โ–
wandb:          _step โ–
wandb:       _runtime โ–
wandb:     _timestamp โ–
wandb: 
wandb: Synced 5 W&B file(s), 0 media file(s), 0 artifact file(s) and 1 other file(s)
wandb: 
wandb: Synced visionary-sweep-10-1: https://wandb.ai/buckler/myproject/runs/v40fbbho
wandb: Currently logged in as: buckler (use `wandb login --relogin` to force relogin)
wandb: Tracking run with wandb version 0.10.11
wandb: Syncing run visionary-sweep-10-2
wandb: โญ๏ธ View project at https://wandb.ai/buckler/myproject
wandb: ๐Ÿš€ View run at https://wandb.ai/buckler/myproject/runs/2e7haeft
wandb: Run data is saved locally in /media/buckler/Data/Work/myproject/wandb/run-20201202_185125-2e7haeft
wandb: Run `wandb offline` to turn off syncing.


wandb: Waiting for W&B process to finish, PID 13412
wandb: Program ended successfully.
wandb:                                                                                
wandb: Find user logs for this run at: /media/buckler/Data/Work/myproject/wandb/run-20201202_185125-2e7haeft/logs/debug.log
wandb: Find internal logs for this run at: /media/buckler/Data/Work/myproject/wandb/run-20201202_185125-2e7haeft/logs/debug-internal.log
wandb: Run summary:
wandb:   val_accuracy 0.56969
wandb:          _step 0
wandb:       _runtime 2
wandb:     _timestamp 1606931487
wandb: Run history:
wandb:   val_accuracy โ–
wandb:          _step โ–
wandb:       _runtime โ–
wandb:     _timestamp โ–
wandb: 
wandb: Synced 5 W&B file(s), 0 media file(s), 0 artifact file(s) and 1 other file(s)
wandb: 
wandb: Synced visionary-sweep-10-2: https://wandb.ai/buckler/myproject/runs/2e7haeft
wandb: Currently logged in as: buckler (use `wandb login --relogin` to force relogin)
wandb: Tracking run with wandb version 0.10.11
wandb: Syncing run visionary-sweep-10-3
wandb: โญ๏ธ View project at https://wandb.ai/buckler/myproject
wandb: ๐Ÿš€ View run at https://wandb.ai/buckler/myproject/runs/3be32jsf
wandb: Run data is saved locally in /media/buckler/Data/Work/myproject/wandb/run-20201202_185131-3be32jsf
wandb: Run `wandb offline` to turn off syncing.


wandb: Waiting for W&B process to finish, PID 13466
wandb: Program ended successfully.
wandb:                                                                                
wandb: Find user logs for this run at: /media/buckler/Data/Work/myproject/wandb/run-20201202_185131-3be32jsf/logs/debug.log
wandb: Find internal logs for this run at: /media/buckler/Data/Work/myproject/wandb/run-20201202_185131-3be32jsf/logs/debug-internal.log
wandb: Run summary:
wandb:   val_accuracy 0.54625
wandb:          _step 0
wandb:       _runtime 2
wandb:     _timestamp 1606931493
wandb: Run history:
wandb:   val_accuracy โ–
wandb:          _step โ–
wandb:       _runtime โ–
wandb:     _timestamp โ–
wandb: 
wandb: Synced 5 W&B file(s), 0 media file(s), 0 artifact file(s) and 1 other file(s)
wandb: 
wandb: Synced visionary-sweep-10-3: https://wandb.ai/buckler/myproject/runs/3be32jsf
wandb: Currently logged in as: buckler (use `wandb login --relogin` to force relogin)
wandb: Tracking run with wandb version 0.10.11
wandb: Syncing run visionary-sweep-10-4
wandb: โญ๏ธ View project at https://wandb.ai/buckler/myproject
wandb: ๐Ÿš€ View run at https://wandb.ai/buckler/myproject/runs/103qkkd3
wandb: Run data is saved locally in /media/buckler/Data/Work/myproject/wandb/run-20201202_185137-103qkkd3
wandb: Run `wandb offline` to turn off syncing.


wandb: Waiting for W&B process to finish, PID 13518
wandb: Program ended successfully.
wandb:                                                                                
wandb: Find user logs for this run at: /media/buckler/Data/Work/myproject/wandb/run-20201202_185137-103qkkd3/logs/debug.log
wandb: Find internal logs for this run at: /media/buckler/Data/Work/myproject/wandb/run-20201202_185137-103qkkd3/logs/debug-internal.log
wandb: Run summary:
wandb:   val_accuracy 0.54069
wandb:          _step 0
wandb:       _runtime 2
wandb:     _timestamp 1606931499
wandb: Run history:
wandb:   val_accuracy โ–
wandb:          _step โ–
wandb:       _runtime โ–
wandb:     _timestamp โ–
wandb: 
wandb: Synced 5 W&B file(s), 0 media file(s), 0 artifact file(s) and 1 other file(s)
wandb: 
wandb: Synced visionary-sweep-10-4: https://wandb.ai/buckler/myproject/runs/103qkkd3

wandb: Waiting for W&B process to finish, PID 13239
wandb: Program ended successfully.
wandb:                                                                                
wandb: Find user logs for this run at: /media/buckler/Data/Work/myproject/wandb/run-20201202_185111-keofgifo/logs/debug.log
wandb: Find internal logs for this run at: /media/buckler/Data/Work/myproject/wandb/run-20201202_185111-keofgifo/logs/debug-internal.log
wandb: Run summary:
wandb:   val_accuracy 0.56021
wandb:          _step 0
wandb:       _runtime 32
wandb:     _timestamp 1606931503
wandb: Run history:
wandb:   val_accuracy โ–
wandb:          _step โ–
wandb:       _runtime โ–
wandb:     _timestamp โ–
wandb: 
wandb: Synced 6 W&B file(s), 0 media file(s), 0 artifact file(s) and 1 other file(s)
wandb: 
wandb: Synced visionary-sweep-10: https://wandb.ai/buckler/myproject/runs/keofgifo
****************************************
Sweep URL:        https://wandb.ai/buckler/myproject/sweeps/qmzmf6ms
Sweep Group URL:  https://wandb.ai/buckler/myproject/groups/qmzmf6ms
****************************************
2020-12-02 18:51:50,826 - wandb.wandb_agent - INFO - Cleaning up finished run: keofgifo
wandb: Terminating and syncing runs. Press ctrl-c to kill.

`colabs/scikit/w-b-k-means-clustering.ipynb` fails to run

This nb, Produces the following error

wandb: ERROR Abnormal program exit
---------------------------------------------------------------------------
Exception                                 Traceback (most recent call last)
/usr/local/lib/python3.7/dist-packages/wandb/sdk/wandb_init.py in init(job_type, dir, config, project, entity, reinit, tags, group, name, notes, magic, config_exclude_keys, config_include_keys, anonymous, mode, allow_val_change, resume, force, tensorboard, sync_tensorboard, monitor_gym, save_code, id, settings)
    916         try:
--> 917             run = wi.init()
    918             except_exit = wi.settings._except_exit

9 frames
Exception: The wandb backend process has shutdown

The above exception was the direct cause of the following exception:

Exception                                 Traceback (most recent call last)
/usr/local/lib/python3.7/dist-packages/six.py in raise_from(value, from_value)

Exception: problem

The nb does not run.

  • specify project, entity should be None.
  • Add markdown to create a documented notebook, if not convert ->.py

Colab examples need a .finish or .join in between runs

I am making this change in the notebook examples in the ml-class repo, but wanted to sync up with y'all. We should agree on whether to use .finish or .join and then stick to it.

With the upgrade to CLIng, the behavior of wandb in a notebook has been changed sufficiently to necessitate some rewriting in notebooks with multiple runs.

For example, our Simple Keras Integration includes multiple runs. This means that when wandb.init is called for the second run, wandb.join is called on the earlier run at the same time, and both are output by the same cell.

This results in confusing output: the wandb-logged metrics and sparklines for the first run (royal-water-3) are not well-separated from the model architecture output for the second run (peach-serenity-4).

image

Recommend Projects

  • React photo React

    A declarative, efficient, and flexible JavaScript library for building user interfaces.

  • Vue.js photo Vue.js

    ๐Ÿ–– Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.

  • Typescript photo Typescript

    TypeScript is a superset of JavaScript that compiles to clean JavaScript output.

  • TensorFlow photo TensorFlow

    An Open Source Machine Learning Framework for Everyone

  • Django photo Django

    The Web framework for perfectionists with deadlines.

  • D3 photo D3

    Bring data to life with SVG, Canvas and HTML. ๐Ÿ“Š๐Ÿ“ˆ๐ŸŽ‰

Recommend Topics

  • javascript

    JavaScript (JS) is a lightweight interpreted programming language with first-class functions.

  • web

    Some thing interesting about web. New door for the world.

  • server

    A server is a program made to process requests and deliver data to clients.

  • Machine learning

    Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.

  • Game

    Some thing interesting about game, make everyone happy.

Recommend Org

  • Facebook photo Facebook

    We are working to build community through open source technology. NB: members must have two-factor auth.

  • Microsoft photo Microsoft

    Open source projects and samples from Microsoft.

  • Google photo Google

    Google โค๏ธ Open Source for everyone.

  • D3 photo D3

    Data-Driven Documents codes.