Rate this Page

Hyperparameter tuning using Ray Tune#

Created On: Aug 31, 2020 | Last Updated: Jan 08, 2026 | Last Verified: Nov 05, 2024

Author: Ricardo Decal

This tutorial shows how to integrate Ray Tune into your PyTorch training workflow to perform scalable and efficient hyperparameter tuning.

What you will learn
  • How to modify a PyTorch training loop for Ray Tune

  • How to scale a hyperparameter sweep to multiple nodes and GPUs without code changes

  • How to define a hyperparameter search space and run a sweep with tune.Tuner

  • How to use an early-stopping scheduler (ASHA) and report metrics/checkpoints

  • How to use checkpointing to resume training and load the best model

Prerequisites
  • PyTorch v2.9+ and torchvision

  • Ray Tune (ray[tune]) v2.52.1+

  • GPU(s) are optional, but recommended for faster training

Ray, a project of the PyTorch Foundation, is an open source unified framework for scaling AI and Python applications. It helps run distributed jobs by handling the complexity of distributed computing. Ray Tune is a library built on Ray for hyperparameter tuning that enables you to scale a hyperparameter sweep from your machine to a large cluster with no code changes.

This tutorial adapts the PyTorch tutorial for training a CIFAR10 classifier to run multi-GPU hyperparameter sweeps with Ray Tune.

Setup#

To run this tutorial, install the following dependencies:

pip install "ray[tune]" torchvision

Then start with the imports:

from functools import partial
import os
import tempfile
from pathlib import Path
import torch
import torch.nn as nn
import torch.nn.functional as F
import torch.optim as optim
from torch.utils.data import random_split
import torchvision
import torchvision.transforms as transforms
# New: imports for Ray Tune
import ray
from ray import tune
from ray.tune import Checkpoint
from ray.tune.schedulers import ASHAScheduler

Data loading#

Wrap the data loaders in a constructor function. In this tutorial, a global data directory is passed to the function to enable reusing the dataset across different trials. In a cluster environment, you can use shared storage, such as network file systems, to prevent each node from downloading the data separately.

def load_data(data_dir="./data"):
    # Mean and standard deviation of the CIFAR10 training subset.
    transform = transforms.Compose(
        [transforms.ToTensor(), transforms.Normalize((0.4914, 0.48216, 0.44653), (0.2022, 0.19932, 0.20086))]
    )

    trainset = torchvision.datasets.CIFAR10(
        root=data_dir, train=True, download=True, transform=transform
    )

    testset = torchvision.datasets.CIFAR10(
        root=data_dir, train=False, download=True, transform=transform
    )

    return trainset, testset

Model architecture#

This tutorial searches for the best sizes for the fully connected layers and the learning rate. To enable this, the Net class exposes the layer sizes l1 and l2 as configurable parameters that Ray Tune can search over:

class Net(nn.Module):
    def __init__(self, l1=120, l2=84):
        super().__init__()
        self.conv1 = nn.Conv2d(3, 6, 5)
        self.pool = nn.MaxPool2d(2, 2)
        self.conv2 = nn.Conv2d(6, 16, 5)
        self.fc1 = nn.Linear(16 * 5 * 5, l1)
        self.fc2 = nn.Linear(l1, l2)
        self.fc3 = nn.Linear(l2, 10)

    def forward(self, x):
        x = self.pool(F.relu(self.conv1(x)))
        x = self.pool(F.relu(self.conv2(x)))
        x = torch.flatten(x, 1)  # flatten all dimensions except batch
        x = F.relu(self.fc1(x))
        x = F.relu(self.fc2(x))
        x = self.fc3(x)
        return x

Define the search space#

Next, define the hyperparameters to tune and how Ray Tune samples them. Ray Tune offers a variety of search space distributions to suit different parameter types: loguniform, uniform, choice, randint, grid, and more. You can also express complex dependencies between parameters with conditional search spaces or sample from arbitrary functions.

Here is the search space for this tutorial:

config = {
    "l1": tune.choice([2**i for i in range(9)]),
    "l2": tune.choice([2**i for i in range(9)]),
    "lr": tune.loguniform(1e-4, 1e-1),
    "batch_size": tune.choice([2, 4, 8, 16]),
}

The tune.choice() accepts a list of values that are uniformly sampled from. In this example, the l1 and l2 parameter values are powers of 2 between 1 and 256, and the learning rate samples on a log scale between 0.0001 and 0.1. Sampling on a log scale enables exploration across a range of magnitudes on a relative scale, rather than an absolute scale.

Training function#

Ray Tune requires a training function that accepts a configuration dictionary and runs the main training loop. As Ray Tune runs different trials, it updates the configuration dictionary for each trial.

Here is the full training function, followed by explanations of the key Ray Tune integration points:

def train_cifar(config, data_dir=None):
    net = Net(config["l1"], config["l2"])
    device = config["device"]

    net = net.to(device)
    if torch.cuda.device_count() > 1:
        net = nn.DataParallel(net)

    criterion = nn.CrossEntropyLoss()
    optimizer = optim.SGD(net.parameters(), lr=config["lr"], momentum=0.9)

    # Load checkpoint if resuming training
    checkpoint = tune.get_checkpoint()
    if checkpoint:
        with checkpoint.as_directory() as checkpoint_dir:
            checkpoint_path = Path(checkpoint_dir) / "checkpoint.pt"
            checkpoint_state = torch.load(checkpoint_path)
            start_epoch = checkpoint_state["epoch"]
            net.load_state_dict(checkpoint_state["net_state_dict"])
            optimizer.load_state_dict(checkpoint_state["optimizer_state_dict"])
    else:
        start_epoch = 0

    trainset, _testset = load_data(data_dir)

    test_abs = int(len(trainset) * 0.8)
    train_subset, val_subset = random_split(
        trainset, [test_abs, len(trainset) - test_abs]
    )

    trainloader = torch.utils.data.DataLoader(
        train_subset, batch_size=int(config["batch_size"]), shuffle=True, num_workers=8
    )
    valloader = torch.utils.data.DataLoader(
        val_subset, batch_size=int(config["batch_size"]), shuffle=True, num_workers=8
    )

    for epoch in range(start_epoch, 10):  # loop over the dataset multiple times
        running_loss = 0.0
        epoch_steps = 0
        for i, data in enumerate(trainloader, 0):
            # get the inputs; data is a list of [inputs, labels]
            inputs, labels = data
            inputs, labels = inputs.to(device), labels.to(device)

            # zero the parameter gradients
            optimizer.zero_grad()

            # forward + backward + optimize
            outputs = net(inputs)
            loss = criterion(outputs, labels)
            loss.backward()
            optimizer.step()

            # print statistics
            running_loss += loss.item()
            epoch_steps += 1
            if i % 2000 == 1999:  # print every 2000 mini-batches
                print(
                    "[%d, %5d] loss: %.3f"
                    % (epoch + 1, i + 1, running_loss / epoch_steps)
                )
                running_loss = 0.0

        # Validation loss
        val_loss = 0.0
        val_steps = 0
        total = 0
        correct = 0
        for i, data in enumerate(valloader, 0):
            with torch.no_grad():
                inputs, labels = data
                inputs, labels = inputs.to(device), labels.to(device)

                outputs = net(inputs)
                _, predicted = torch.max(outputs.data, 1)
                total += labels.size(0)
                correct += (predicted == labels).sum().item()

                loss = criterion(outputs, labels)
                val_loss += loss.cpu().numpy()
                val_steps += 1

        # Save checkpoint and report metrics
        checkpoint_data = {
            "epoch": epoch,
            "net_state_dict": net.state_dict(),
            "optimizer_state_dict": optimizer.state_dict(),
        }
        with tempfile.TemporaryDirectory() as checkpoint_dir:
            checkpoint_path = Path(checkpoint_dir) / "checkpoint.pt"
            torch.save(checkpoint_data, checkpoint_path)

            checkpoint = Checkpoint.from_directory(checkpoint_dir)
            tune.report(
                {"loss": val_loss / val_steps, "accuracy": correct / total},
                checkpoint=checkpoint,
            )

    print("Finished Training")

Key integration points#

Using hyperparameters from the configuration dictionary#

Ray Tune updates the config dictionary with the hyperparameters for each trial. In this example, the model architecture and optimizer receive the hyperparameters from the config dictionary:

net = Net(config["l1"], config["l2"])
optimizer = optim.SGD(net.parameters(), lr=config["lr"], momentum=0.9)

Reporting metrics and saving checkpoints#

The most important integration is communicating with Ray Tune. Ray Tune uses the validation metrics to determine the best hyperparameter configuration and to stop underperforming trials early, saving resources.

Checkpointing enables you to later load the trained models, resume hyperparameter searches, and provides fault tolerance. It’s also required for some Ray Tune schedulers like Population Based Training that pause and resume trials during the search.

This code from the training function loads model and optimizer state at the start if a checkpoint exists:

checkpoint = tune.get_checkpoint()
if checkpoint:
    with checkpoint.as_directory() as checkpoint_dir:
        checkpoint_path = Path(checkpoint_dir) / "checkpoint.pt"
        checkpoint_state = torch.load(checkpoint_path)
        start_epoch = checkpoint_state["epoch"]
        net.load_state_dict(checkpoint_state["net_state_dict"])
        optimizer.load_state_dict(checkpoint_state["optimizer_state_dict"])

At the end of each epoch, save a checkpoint and report the validation metrics:

checkpoint_data = {
    "epoch": epoch,
    "net_state_dict": net.state_dict(),
    "optimizer_state_dict": optimizer.state_dict(),
}
with tempfile.TemporaryDirectory() as checkpoint_dir:
    checkpoint_path = Path(checkpoint_dir) / "checkpoint.pt"
    torch.save(checkpoint_data, checkpoint_path)

    checkpoint = Checkpoint.from_directory(checkpoint_dir)
    tune.report(
        {"loss": val_loss / val_steps, "accuracy": correct / total},
        checkpoint=checkpoint,
    )

Ray Tune checkpointing supports local file systems, cloud storage, and distributed file systems. For more information, see the Ray Tune storage documentation.

Multi-GPU support#

Image classification models can be greatly accelerated by using GPUs. The training function supports multi-GPU training by wrapping the model in nn.DataParallel:

This training function supports training on CPUs, a single GPU, multiple GPUs, or multiple nodes without code changes. Ray Tune automatically distributes the trials across the nodes according to the available resources. Ray Tune also supports fractional GPUs so that one GPU can be shared among multiple trials, provided that the models, optimizers, and data batches fit into the GPU memory.

Validation split#

The original CIFAR10 dataset only has train and test subsets. This is sufficient for training a single model, however for hyperparameter tuning a validation subset is required. The training function creates a validation subset by reserving 20% of the training subset. The test subset is used to evaluate the best model’s generalization error after the search completes.

Evaluation function#

After finding the optimal hyperparameters, test the model on a held-out test set to estimate the generalization error:

def test_accuracy(net, device="cpu", data_dir=None):
    _trainset, testset = load_data(data_dir)

    testloader = torch.utils.data.DataLoader(
        testset, batch_size=4, shuffle=False, num_workers=2
    )

    correct = 0
    total = 0
    with torch.no_grad():
        for data in testloader:
            image_batch, labels = data
            image_batch, labels = image_batch.to(device), labels.to(device)
            outputs = net(image_batch)
            _, predicted = torch.max(outputs.data, 1)
            total += labels.size(0)
            correct += (predicted == labels).sum().item()

    return correct / total

Configure and run Ray Tune#

With the training and evaluation functions defined, configure Ray Tune to run the hyperparameter search.

Scheduler for early stopping#

Ray Tune provides schedulers to improve the efficiency of the hyperparameter search by detecting underperforming trials and stopping them early. The ASHAScheduler uses the Asynchronous Successive Halving Algorithm (ASHA) to aggressively terminate low-performing trials:

scheduler = ASHAScheduler(
    max_t=max_num_epochs,
    grace_period=1,
    reduction_factor=2,
)

Ray Tune also provides advanced search algorithms to smartly pick the next set of hyperparameters based on previous results, instead of relying only on random or grid search. Examples include Optuna and BayesOpt.

Resource allocation#

Tell Ray Tune what resources to allocate for each trial by passing a resources dictionary to tune.with_resources:

tune.with_resources(
    partial(train_cifar, data_dir=data_dir),
    resources={"cpu": cpus_per_trial, "gpu": gpus_per_trial}
)

Ray Tune automatically manages the placement of these trials and ensures that the trials run in isolation, so you don’t need to manually assign GPUs to processes.

For example, if you are running this experiment on a cluster of 20 machines, each with 8 GPUs, you can set gpus_per_trial = 0.5 to schedule two concurrent trials per GPU. This configuration runs 320 trials in parallel across the cluster.

Note

To run this tutorial without GPUs, set gpus_per_trial=0 and expect significantly longer runtimes.

To avoid long runtimes during development, start with a small number of trials and epochs.

Creating the Tuner#

The Ray Tune API is modular and composable. Pass your configuration to the tune.Tuner class to create a tuner object, then run tuner.fit() to start training:

tuner = tune.Tuner(
    tune.with_resources(
        partial(train_cifar, data_dir=data_dir),
        resources={"cpu": cpus_per_trial, "gpu": gpus_per_trial}
    ),
    tune_config=tune.TuneConfig(
        metric="loss",
        mode="min",
        scheduler=scheduler,
        num_samples=num_trials,
    ),
    param_space=config,
)
results = tuner.fit()

After training completes, retrieve the best performing trial, load its checkpoint, and evaluate on the test set.

Putting it all together#

def main(num_trials=10, max_num_epochs=10, gpus_per_trial=0, cpus_per_trial=2):
    print("Starting hyperparameter tuning.")
    ray.init(include_dashboard=False)

    data_dir = os.path.abspath("./data")
    load_data(data_dir)  # Pre-download the dataset
    device = "cuda" if torch.cuda.is_available() else "cpu"
    config = {
        "l1": tune.choice([2**i for i in range(9)]),
        "l2": tune.choice([2**i for i in range(9)]),
        "lr": tune.loguniform(1e-4, 1e-1),
        "batch_size": tune.choice([2, 4, 8, 16]),
        "device": device,
    }
    scheduler = ASHAScheduler(
        max_t=max_num_epochs,
        grace_period=1,
        reduction_factor=2,
    )

    tuner = tune.Tuner(
        tune.with_resources(
            partial(train_cifar, data_dir=data_dir),
            resources={"cpu": cpus_per_trial, "gpu": gpus_per_trial}
        ),
        tune_config=tune.TuneConfig(
            metric="loss",
            mode="min",
            scheduler=scheduler,
            num_samples=num_trials,
        ),
        param_space=config,
    )
    results = tuner.fit()

    best_result = results.get_best_result("loss", "min")
    print(f"Best trial config: {best_result.config}")
    print(f"Best trial final validation loss: {best_result.metrics['loss']}")
    print(f"Best trial final validation accuracy: {best_result.metrics['accuracy']}")

    best_trained_model = Net(best_result.config["l1"], best_result.config["l2"])
    best_trained_model = best_trained_model.to(device)
    if gpus_per_trial > 1:
        best_trained_model = nn.DataParallel(best_trained_model)

    best_checkpoint = best_result.checkpoint
    with best_checkpoint.as_directory() as checkpoint_dir:
        checkpoint_path = Path(checkpoint_dir) / "checkpoint.pt"
        best_checkpoint_data = torch.load(checkpoint_path)

        best_trained_model.load_state_dict(best_checkpoint_data["net_state_dict"])
        test_acc = test_accuracy(best_trained_model, device, data_dir)
        print(f"Best trial test set accuracy: {test_acc}")


if __name__ == "__main__":
    # Set the number of trials, epochs, and GPUs per trial here:
    main(num_trials=10, max_num_epochs=10, gpus_per_trial=1)
Starting hyperparameter tuning.
2026-05-08 21:48:00,365 WARNING services.py:2213 -- WARNING: The object store is using /tmp/ray instead of /dev/shm because /dev/shm has only 2147471360 bytes available. This will harm performance! You may be able to free up space by deleting files in /dev/shm. If you are inside a Docker container, you can increase /dev/shm size by passing '--shm-size=10.24gb' to 'docker run' (or add it to the run_options list in a Ray cluster config). Make sure to set this to more than 30% of available RAM.
2026-05-08 21:48:02,534 INFO worker.py:2012 -- Started a local Ray instance.
/usr/local/lib/python3.10/dist-packages/ray/_private/worker.py:2051: FutureWarning: Tip: In future versions of Ray, Ray will no longer override accelerator visible devices env var if num_gpus=0 or num_gpus=None (default). To enable this behavior and turn off this error message, set RAY_ACCEL_ENV_VAR_OVERRIDE_ON_ZERO=0
  warnings.warn(

  0%|          | 0.00/170M [00:00<?, ?B/s]
  0%|          | 262k/170M [00:00<01:06, 2.57MB/s]
  0%|          | 655k/170M [00:00<00:51, 3.28MB/s]
  1%|          | 1.05M/170M [00:00<00:48, 3.51MB/s]
  1%|          | 1.47M/170M [00:00<00:46, 3.67MB/s]
  1%|          | 1.90M/170M [00:00<00:44, 3.79MB/s]
  1%|▏         | 2.33M/170M [00:00<00:43, 3.84MB/s]
  2%|▏         | 2.75M/170M [00:00<00:43, 3.88MB/s]
  2%|▏         | 3.18M/170M [00:00<00:42, 3.90MB/s]
  2%|▏         | 3.60M/170M [00:00<00:42, 3.90MB/s]
  2%|▏         | 4.00M/170M [00:01<00:42, 3.91MB/s]
  3%|▎         | 4.39M/170M [00:01<00:42, 3.88MB/s]
  3%|▎         | 4.78M/170M [00:01<00:43, 3.83MB/s]
  3%|▎         | 5.18M/170M [00:01<00:43, 3.81MB/s]
  3%|▎         | 5.57M/170M [00:01<00:43, 3.80MB/s]
  3%|▎         | 5.96M/170M [00:01<00:43, 3.81MB/s]
  4%|▎         | 6.36M/170M [00:01<00:43, 3.81MB/s]
  4%|▍         | 6.75M/170M [00:01<00:42, 3.83MB/s]
  4%|▍         | 7.14M/170M [00:01<00:42, 3.83MB/s]
  4%|▍         | 7.54M/170M [00:01<00:42, 3.81MB/s]
  5%|▍         | 7.93M/170M [00:02<00:42, 3.85MB/s]
  5%|▍         | 8.32M/170M [00:02<00:42, 3.86MB/s]
  5%|▌         | 8.72M/170M [00:02<00:42, 3.84MB/s]
  5%|▌         | 9.11M/170M [00:02<00:41, 3.87MB/s]
  6%|▌         | 9.50M/170M [00:02<00:41, 3.85MB/s]
  6%|▌         | 9.90M/170M [00:02<00:41, 3.85MB/s]
  6%|▌         | 10.3M/170M [00:02<00:41, 3.84MB/s]
  6%|▋         | 10.7M/170M [00:02<00:41, 3.82MB/s]
  6%|▋         | 11.1M/170M [00:02<00:42, 3.77MB/s]
  7%|▋         | 11.5M/170M [00:03<00:41, 3.81MB/s]
  7%|▋         | 11.9M/170M [00:03<00:41, 3.82MB/s]
  7%|▋         | 12.3M/170M [00:03<00:41, 3.84MB/s]
  7%|▋         | 12.6M/170M [00:03<00:41, 3.83MB/s]
  8%|▊         | 13.0M/170M [00:03<00:41, 3.84MB/s]
  8%|▊         | 13.5M/170M [00:03<00:40, 3.86MB/s]
  8%|▊         | 13.9M/170M [00:03<00:40, 3.86MB/s]
  8%|▊         | 14.3M/170M [00:03<00:40, 3.85MB/s]
  9%|▊         | 14.6M/170M [00:03<00:40, 3.86MB/s]
  9%|▉         | 15.0M/170M [00:03<00:40, 3.84MB/s]
  9%|▉         | 15.4M/170M [00:04<00:40, 3.84MB/s]
  9%|▉         | 15.8M/170M [00:04<00:40, 3.84MB/s]
 10%|▉         | 16.3M/170M [00:04<00:39, 3.93MB/s]
 10%|▉         | 16.7M/170M [00:04<00:39, 3.91MB/s]
 10%|█         | 17.1M/170M [00:04<00:39, 3.89MB/s]
 10%|█         | 17.5M/170M [00:04<00:39, 3.92MB/s]
 10%|█         | 17.9M/170M [00:04<00:38, 3.91MB/s]
 11%|█         | 18.3M/170M [00:04<00:38, 3.94MB/s]
 11%|█         | 18.7M/170M [00:04<00:38, 3.94MB/s]
 11%|█         | 19.2M/170M [00:04<00:38, 3.95MB/s]
 11%|█▏        | 19.6M/170M [00:05<00:38, 3.97MB/s]
 12%|█▏        | 20.0M/170M [00:05<00:37, 3.99MB/s]
 12%|█▏        | 20.4M/170M [00:05<00:37, 4.01MB/s]
 12%|█▏        | 20.9M/170M [00:05<00:37, 3.99MB/s]
 12%|█▏        | 21.3M/170M [00:05<00:37, 3.96MB/s]
 13%|█▎        | 21.7M/170M [00:05<00:37, 3.98MB/s]
 13%|█▎        | 22.2M/170M [00:05<00:36, 4.01MB/s]
 13%|█▎        | 22.6M/170M [00:05<00:36, 4.02MB/s]
 13%|█▎        | 23.0M/170M [00:05<00:36, 4.05MB/s]
 14%|█▎        | 23.4M/170M [00:06<00:36, 4.08MB/s]
 14%|█▍        | 23.9M/170M [00:06<00:35, 4.08MB/s]
 14%|█▍        | 24.3M/170M [00:06<00:35, 4.09MB/s]
 14%|█▍        | 24.7M/170M [00:06<00:35, 4.11MB/s]
 15%|█▍        | 25.1M/170M [00:06<00:35, 4.07MB/s]
 15%|█▍        | 25.6M/170M [00:06<00:35, 4.10MB/s]
 15%|█▌        | 26.0M/170M [00:06<00:35, 4.08MB/s]
 15%|█▌        | 26.4M/170M [00:06<00:35, 4.11MB/s]
 16%|█▌        | 26.8M/170M [00:06<00:35, 4.10MB/s]
 16%|█▌        | 27.3M/170M [00:06<00:34, 4.13MB/s]
 16%|█▌        | 27.7M/170M [00:07<00:34, 4.08MB/s]
 16%|█▋        | 28.1M/170M [00:07<00:34, 4.08MB/s]
 17%|█▋        | 28.5M/170M [00:07<00:34, 4.09MB/s]
 17%|█▋        | 29.0M/170M [00:07<00:34, 4.10MB/s]
 17%|█▋        | 29.4M/170M [00:07<00:34, 4.09MB/s]
 17%|█▋        | 29.8M/170M [00:07<00:34, 4.09MB/s]
 18%|█▊        | 30.2M/170M [00:07<00:34, 4.10MB/s]
 18%|█▊        | 30.7M/170M [00:07<00:33, 4.14MB/s]
 18%|█▊        | 31.1M/170M [00:07<00:33, 4.10MB/s]
 18%|█▊        | 31.5M/170M [00:08<00:33, 4.11MB/s]
 19%|█▊        | 31.9M/170M [00:08<00:33, 4.11MB/s]
 19%|█▉        | 32.4M/170M [00:08<00:33, 4.13MB/s]
 19%|█▉        | 32.8M/170M [00:08<00:33, 4.11MB/s]
 19%|█▉        | 33.2M/170M [00:08<00:33, 4.09MB/s]
 20%|█▉        | 33.7M/170M [00:08<00:33, 4.10MB/s]
 20%|█▉        | 34.1M/170M [00:08<00:33, 4.11MB/s]
 20%|██        | 34.5M/170M [00:08<00:32, 4.13MB/s]
 20%|██        | 34.9M/170M [00:08<00:32, 4.12MB/s]
 21%|██        | 35.5M/170M [00:08<00:30, 4.40MB/s]
 21%|██        | 36.0M/170M [00:09<00:28, 4.68MB/s]
 21%|██▏       | 36.6M/170M [00:09<00:27, 4.88MB/s]
 22%|██▏       | 37.1M/170M [00:09<00:26, 5.03MB/s]
 22%|██▏       | 37.7M/170M [00:09<00:25, 5.20MB/s]
 22%|██▏       | 38.3M/170M [00:09<00:24, 5.33MB/s]
 23%|██▎       | 38.9M/170M [00:09<00:24, 5.45MB/s]
 23%|██▎       | 39.5M/170M [00:09<00:23, 5.49MB/s]
 24%|██▎       | 40.1M/170M [00:09<00:23, 5.53MB/s]
 24%|██▍       | 40.7M/170M [00:09<00:23, 5.54MB/s]
 24%|██▍       | 41.3M/170M [00:10<00:23, 5.56MB/s]
 25%|██▍       | 41.8M/170M [00:10<00:23, 5.52MB/s]
 25%|██▍       | 42.4M/170M [00:10<00:23, 5.42MB/s]
 25%|██▌       | 42.9M/170M [00:10<00:23, 5.41MB/s]
 26%|██▌       | 43.5M/170M [00:10<00:23, 5.36MB/s]
 26%|██▌       | 44.0M/170M [00:10<00:23, 5.35MB/s]
 26%|██▌       | 44.6M/170M [00:10<00:23, 5.37MB/s]
 26%|██▋       | 45.2M/170M [00:10<00:23, 5.36MB/s]
 27%|██▋       | 45.7M/170M [00:10<00:23, 5.36MB/s]
 27%|██▋       | 46.3M/170M [00:10<00:23, 5.35MB/s]
 27%|██▋       | 46.8M/170M [00:11<00:23, 5.33MB/s]
 28%|██▊       | 47.4M/170M [00:11<00:23, 5.33MB/s]
 28%|██▊       | 47.9M/170M [00:11<00:23, 5.25MB/s]
 28%|██▊       | 48.5M/170M [00:11<00:23, 5.21MB/s]
 29%|██▉       | 49.1M/170M [00:11<00:23, 5.25MB/s]
 29%|██▉       | 49.6M/170M [00:11<00:22, 5.28MB/s]
 29%|██▉       | 50.2M/170M [00:11<00:22, 5.32MB/s]
 30%|██▉       | 50.7M/170M [00:11<00:22, 5.33MB/s]
 30%|███       | 51.3M/170M [00:11<00:22, 5.32MB/s]
 30%|███       | 51.8M/170M [00:11<00:22, 5.34MB/s]
 31%|███       | 52.4M/170M [00:12<00:22, 5.34MB/s]
 31%|███       | 53.0M/170M [00:12<00:22, 5.32MB/s]
 31%|███▏      | 53.5M/170M [00:12<00:21, 5.34MB/s]
 32%|███▏      | 54.1M/170M [00:12<00:21, 5.35MB/s]
 32%|███▏      | 54.6M/170M [00:12<00:21, 5.38MB/s]
 32%|███▏      | 55.2M/170M [00:12<00:21, 5.47MB/s]
 33%|███▎      | 55.8M/170M [00:12<00:20, 5.49MB/s]
 33%|███▎      | 56.4M/170M [00:12<00:20, 5.52MB/s]
 33%|███▎      | 57.0M/170M [00:12<00:20, 5.60MB/s]
 34%|███▎      | 57.5M/170M [00:13<00:20, 5.59MB/s]
 34%|███▍      | 58.1M/170M [00:13<00:20, 5.59MB/s]
 34%|███▍      | 58.7M/170M [00:13<00:20, 5.58MB/s]
 35%|███▍      | 59.3M/170M [00:13<00:19, 5.59MB/s]
 35%|███▌      | 59.9M/170M [00:13<00:20, 5.48MB/s]
 35%|███▌      | 60.5M/170M [00:13<00:20, 5.40MB/s]
 36%|███▌      | 61.0M/170M [00:13<00:20, 5.39MB/s]
 36%|███▌      | 61.6M/170M [00:13<00:20, 5.33MB/s]
 36%|███▋      | 62.1M/170M [00:13<00:20, 5.33MB/s]
 37%|███▋      | 62.7M/170M [00:13<00:20, 5.30MB/s]
 37%|███▋      | 63.2M/170M [00:14<00:20, 5.30MB/s]
 37%|███▋      | 63.8M/170M [00:14<00:20, 5.29MB/s]
 38%|███▊      | 64.4M/170M [00:14<00:20, 5.27MB/s]
 38%|███▊      | 64.9M/170M [00:14<00:20, 5.21MB/s]
 38%|███▊      | 65.4M/170M [00:14<00:20, 5.17MB/s]
 39%|███▊      | 66.0M/170M [00:14<00:20, 5.11MB/s]
 39%|███▉      | 66.5M/170M [00:14<00:20, 5.10MB/s]
 39%|███▉      | 67.0M/170M [00:14<00:20, 5.06MB/s]
 40%|███▉      | 67.5M/170M [00:14<00:20, 5.07MB/s]
 40%|███▉      | 68.1M/170M [00:15<00:20, 5.04MB/s]
 40%|████      | 68.6M/170M [00:15<00:20, 5.01MB/s]
 41%|████      | 69.1M/170M [00:15<00:20, 4.99MB/s]
 41%|████      | 69.6M/170M [00:15<00:20, 4.97MB/s]
 41%|████      | 70.2M/170M [00:15<00:20, 4.96MB/s]
 41%|████▏     | 70.7M/170M [00:15<00:20, 4.96MB/s]
 42%|████▏     | 71.2M/170M [00:15<00:20, 4.92MB/s]
 42%|████▏     | 71.7M/170M [00:15<00:20, 4.91MB/s]
 42%|████▏     | 72.3M/170M [00:15<00:19, 4.94MB/s]
 43%|████▎     | 72.8M/170M [00:15<00:19, 4.93MB/s]
 43%|████▎     | 73.3M/170M [00:16<00:19, 4.97MB/s]
 43%|████▎     | 73.8M/170M [00:16<00:19, 4.97MB/s]
 44%|████▎     | 74.4M/170M [00:16<00:19, 4.99MB/s]
 44%|████▍     | 74.9M/170M [00:16<00:19, 4.97MB/s]
 44%|████▍     | 75.4M/170M [00:16<00:19, 4.98MB/s]
 45%|████▍     | 75.9M/170M [00:16<00:19, 4.97MB/s]
 45%|████▍     | 76.4M/170M [00:16<00:18, 4.97MB/s]
 45%|████▌     | 77.0M/170M [00:16<00:18, 4.98MB/s]
 45%|████▌     | 77.5M/170M [00:16<00:18, 5.07MB/s]
 46%|████▌     | 78.1M/170M [00:17<00:18, 5.11MB/s]
 46%|████▌     | 78.6M/170M [00:17<00:18, 5.07MB/s]
 46%|████▋     | 79.1M/170M [00:17<00:18, 5.06MB/s]
 47%|████▋     | 79.6M/170M [00:17<00:17, 5.06MB/s]
 47%|████▋     | 80.2M/170M [00:17<00:17, 5.09MB/s]
 47%|████▋     | 80.7M/170M [00:17<00:17, 5.08MB/s]
 48%|████▊     | 81.2M/170M [00:17<00:17, 5.10MB/s]
 48%|████▊     | 81.7M/170M [00:17<00:17, 5.06MB/s]
 48%|████▊     | 82.2M/170M [00:17<00:17, 5.08MB/s]
 49%|████▊     | 82.8M/170M [00:17<00:17, 5.10MB/s]
 49%|████▉     | 83.3M/170M [00:18<00:17, 5.09MB/s]
 49%|████▉     | 83.9M/170M [00:18<00:16, 5.15MB/s]
 50%|████▉     | 84.4M/170M [00:18<00:16, 5.19MB/s]
 50%|████▉     | 85.0M/170M [00:18<00:16, 5.20MB/s]
 50%|█████     | 85.5M/170M [00:18<00:16, 5.22MB/s]
 50%|█████     | 86.0M/170M [00:18<00:16, 5.22MB/s]
 51%|█████     | 86.6M/170M [00:18<00:16, 5.21MB/s]
 51%|█████     | 87.1M/170M [00:18<00:16, 5.13MB/s]
 51%|█████▏    | 87.6M/170M [00:18<00:16, 5.14MB/s]
 52%|█████▏    | 88.1M/170M [00:19<00:16, 5.13MB/s]
 52%|█████▏    | 88.7M/170M [00:19<00:15, 5.12MB/s]
 52%|█████▏    | 89.2M/170M [00:19<00:15, 5.10MB/s]
 53%|█████▎    | 89.7M/170M [00:19<00:15, 5.11MB/s]
 53%|█████▎    | 90.2M/170M [00:19<00:15, 5.11MB/s]
 53%|█████▎    | 90.8M/170M [00:19<00:15, 5.11MB/s]
 54%|█████▎    | 91.3M/170M [00:19<00:15, 5.10MB/s]
 54%|█████▍    | 91.8M/170M [00:19<00:15, 5.05MB/s]
 54%|█████▍    | 92.3M/170M [00:19<00:15, 4.97MB/s]
 54%|█████▍    | 92.9M/170M [00:19<00:15, 4.92MB/s]
 55%|█████▍    | 93.4M/170M [00:20<00:15, 4.94MB/s]
 55%|█████▌    | 93.9M/170M [00:20<00:15, 4.96MB/s]
 55%|█████▌    | 94.4M/170M [00:20<00:15, 5.02MB/s]
 56%|█████▌    | 95.0M/170M [00:20<00:15, 5.03MB/s]
 56%|█████▌    | 95.5M/170M [00:20<00:14, 5.05MB/s]
 56%|█████▋    | 96.0M/170M [00:20<00:14, 5.06MB/s]
 57%|█████▋    | 96.5M/170M [00:20<00:14, 5.07MB/s]
 57%|█████▋    | 97.1M/170M [00:20<00:14, 5.07MB/s]
 57%|█████▋    | 97.6M/170M [00:20<00:14, 5.07MB/s]
 58%|█████▊    | 98.1M/170M [00:20<00:14, 5.08MB/s]
 58%|█████▊    | 98.7M/170M [00:21<00:13, 5.13MB/s]
 58%|█████▊    | 99.2M/170M [00:21<00:13, 5.23MB/s]
 59%|█████▊    | 99.8M/170M [00:21<00:13, 5.26MB/s]
 59%|█████▉    | 100M/170M [00:21<00:13, 5.36MB/s]
 59%|█████▉    | 101M/170M [00:21<00:12, 5.50MB/s]
 60%|█████▉    | 102M/170M [00:21<00:12, 5.57MB/s]
 60%|█████▉    | 102M/170M [00:21<00:12, 5.60MB/s]
 60%|██████    | 103M/170M [00:21<00:12, 5.60MB/s]
 61%|██████    | 103M/170M [00:21<00:12, 5.55MB/s]
 61%|██████    | 104M/170M [00:22<00:12, 5.54MB/s]
 61%|██████▏   | 104M/170M [00:22<00:11, 5.54MB/s]
 62%|██████▏   | 105M/170M [00:22<00:11, 5.52MB/s]
 62%|██████▏   | 106M/170M [00:22<00:11, 5.52MB/s]
 62%|██████▏   | 106M/170M [00:22<00:11, 5.55MB/s]
 63%|██████▎   | 107M/170M [00:22<00:11, 5.51MB/s]
 63%|██████▎   | 107M/170M [00:22<00:11, 5.51MB/s]
 63%|██████▎   | 108M/170M [00:22<00:11, 5.51MB/s]
 64%|██████▎   | 108M/170M [00:22<00:11, 5.50MB/s]
 64%|██████▍   | 109M/170M [00:22<00:11, 5.52MB/s]
 64%|██████▍   | 110M/170M [00:23<00:10, 5.55MB/s]
 65%|██████▍   | 110M/170M [00:23<00:10, 5.57MB/s]
 65%|██████▍   | 111M/170M [00:23<00:10, 5.59MB/s]
 65%|██████▌   | 111M/170M [00:23<00:10, 5.62MB/s]
 66%|██████▌   | 112M/170M [00:23<00:10, 5.66MB/s]
 66%|██████▌   | 112M/170M [00:23<00:10, 5.65MB/s]
 66%|██████▋   | 113M/170M [00:23<00:10, 5.67MB/s]
 67%|██████▋   | 114M/170M [00:23<00:10, 5.62MB/s]
 67%|██████▋   | 114M/170M [00:23<00:09, 5.68MB/s]
 67%|██████▋   | 115M/170M [00:23<00:09, 5.63MB/s]
 68%|██████▊   | 115M/170M [00:24<00:09, 5.57MB/s]
 68%|██████▊   | 116M/170M [00:24<00:09, 5.57MB/s]
 68%|██████▊   | 117M/170M [00:24<00:09, 5.48MB/s]
 69%|██████▊   | 117M/170M [00:24<00:09, 5.48MB/s]
 69%|██████▉   | 118M/170M [00:24<00:09, 5.48MB/s]
 69%|██████▉   | 118M/170M [00:24<00:09, 5.51MB/s]
 70%|██████▉   | 119M/170M [00:24<00:09, 5.53MB/s]
 70%|███████   | 119M/170M [00:24<00:09, 5.54MB/s]
 70%|███████   | 120M/170M [00:24<00:09, 5.56MB/s]
 71%|███████   | 121M/170M [00:25<00:08, 5.56MB/s]
 71%|███████   | 121M/170M [00:25<00:08, 5.60MB/s]
 71%|███████▏  | 122M/170M [00:25<00:08, 5.65MB/s]
 72%|███████▏  | 122M/170M [00:25<00:08, 5.69MB/s]
 72%|███████▏  | 123M/170M [00:25<00:08, 5.70MB/s]
 72%|███████▏  | 124M/170M [00:25<00:08, 5.68MB/s]
 73%|███████▎  | 124M/170M [00:25<00:08, 5.64MB/s]
 73%|███████▎  | 125M/170M [00:25<00:08, 5.44MB/s]
 73%|███████▎  | 125M/170M [00:25<00:08, 5.38MB/s]
 74%|███████▍  | 126M/170M [00:25<00:08, 5.50MB/s]
 74%|███████▍  | 126M/170M [00:26<00:08, 5.48MB/s]
 74%|███████▍  | 127M/170M [00:26<00:08, 5.37MB/s]
 75%|███████▍  | 128M/170M [00:26<00:08, 5.27MB/s]
 75%|███████▌  | 128M/170M [00:26<00:07, 5.32MB/s]
 75%|███████▌  | 129M/170M [00:26<00:07, 5.25MB/s]
 76%|███████▌  | 129M/170M [00:26<00:08, 5.11MB/s]
 76%|███████▌  | 130M/170M [00:26<00:08, 4.98MB/s]
 76%|███████▋  | 130M/170M [00:26<00:08, 4.94MB/s]
 77%|███████▋  | 131M/170M [00:26<00:08, 4.94MB/s]
 77%|███████▋  | 131M/170M [00:27<00:07, 5.12MB/s]
 77%|███████▋  | 132M/170M [00:27<00:07, 5.21MB/s]
 78%|███████▊  | 132M/170M [00:27<00:07, 5.23MB/s]
 78%|███████▊  | 133M/170M [00:27<00:07, 5.27MB/s]
 78%|███████▊  | 134M/170M [00:27<00:06, 5.35MB/s]
 79%|███████▊  | 134M/170M [00:27<00:06, 5.38MB/s]
 79%|███████▉  | 135M/170M [00:27<00:06, 5.13MB/s]
 79%|███████▉  | 135M/170M [00:27<00:07, 4.99MB/s]
 80%|███████▉  | 136M/170M [00:27<00:07, 4.95MB/s]
 80%|███████▉  | 136M/170M [00:28<00:07, 4.87MB/s]
 80%|████████  | 137M/170M [00:28<00:07, 4.78MB/s]
 81%|████████  | 137M/170M [00:28<00:07, 4.73MB/s]
 81%|████████  | 138M/170M [00:28<00:06, 4.71MB/s]
 81%|████████  | 138M/170M [00:28<00:06, 4.67MB/s]
 81%|████████▏ | 139M/170M [00:28<00:06, 4.71MB/s]
 82%|████████▏ | 139M/170M [00:28<00:06, 4.67MB/s]
 82%|████████▏ | 140M/170M [00:28<00:06, 4.68MB/s]
 82%|████████▏ | 140M/170M [00:28<00:06, 4.65MB/s]
 83%|████████▎ | 141M/170M [00:28<00:06, 4.65MB/s]
 83%|████████▎ | 141M/170M [00:29<00:06, 4.69MB/s]
 83%|████████▎ | 142M/170M [00:29<00:06, 4.67MB/s]
 83%|████████▎ | 142M/170M [00:29<00:06, 4.66MB/s]
 84%|████████▎ | 143M/170M [00:29<00:05, 4.67MB/s]
 84%|████████▍ | 143M/170M [00:29<00:05, 4.60MB/s]
 84%|████████▍ | 144M/170M [00:29<00:05, 4.73MB/s]
 85%|████████▍ | 144M/170M [00:29<00:05, 4.92MB/s]
 85%|████████▍ | 145M/170M [00:29<00:05, 5.09MB/s]
 85%|████████▌ | 145M/170M [00:29<00:04, 5.21MB/s]
 86%|████████▌ | 146M/170M [00:30<00:04, 5.25MB/s]
 86%|████████▌ | 147M/170M [00:30<00:04, 5.33MB/s]
 86%|████████▋ | 147M/170M [00:30<00:04, 5.39MB/s]
 87%|████████▋ | 148M/170M [00:30<00:04, 5.31MB/s]
 87%|████████▋ | 148M/170M [00:30<00:04, 5.28MB/s]
 87%|████████▋ | 149M/170M [00:30<00:04, 5.27MB/s]
 88%|████████▊ | 149M/170M [00:30<00:04, 5.25MB/s]
 88%|████████▊ | 150M/170M [00:30<00:03, 5.24MB/s]
 88%|████████▊ | 150M/170M [00:30<00:03, 5.25MB/s]
 89%|████████▊ | 151M/170M [00:30<00:03, 5.19MB/s]
 89%|████████▉ | 151M/170M [00:31<00:03, 5.09MB/s]
 89%|████████▉ | 152M/170M [00:31<00:03, 5.11MB/s]
 89%|████████▉ | 153M/170M [00:31<00:03, 5.18MB/s]
 90%|████████▉ | 153M/170M [00:31<00:03, 5.18MB/s]
 90%|█████████ | 154M/170M [00:31<00:03, 5.23MB/s]
 90%|█████████ | 154M/170M [00:31<00:03, 5.30MB/s]
 91%|█████████ | 155M/170M [00:31<00:02, 5.32MB/s]
 91%|█████████ | 155M/170M [00:31<00:02, 5.34MB/s]
 91%|█████████▏| 156M/170M [00:31<00:02, 5.36MB/s]
 92%|█████████▏| 156M/170M [00:32<00:02, 5.32MB/s]
 92%|█████████▏| 157M/170M [00:32<00:02, 5.36MB/s]
 92%|█████████▏| 158M/170M [00:32<00:02, 5.32MB/s]
 93%|█████████▎| 158M/170M [00:32<00:02, 5.33MB/s]
 93%|█████████▎| 159M/170M [00:32<00:02, 5.31MB/s]
 93%|█████████▎| 159M/170M [00:32<00:02, 5.30MB/s]
 94%|█████████▎| 160M/170M [00:32<00:02, 5.33MB/s]
 94%|█████████▍| 160M/170M [00:32<00:01, 5.39MB/s]
 94%|█████████▍| 161M/170M [00:32<00:01, 5.37MB/s]
 95%|█████████▍| 161M/170M [00:32<00:01, 5.40MB/s]
 95%|█████████▌| 162M/170M [00:33<00:01, 5.41MB/s]
 95%|█████████▌| 163M/170M [00:33<00:01, 5.43MB/s]
 96%|█████████▌| 163M/170M [00:33<00:01, 5.40MB/s]
 96%|█████████▌| 164M/170M [00:33<00:01, 5.38MB/s]
 96%|█████████▋| 164M/170M [00:33<00:01, 5.39MB/s]
 97%|█████████▋| 165M/170M [00:33<00:01, 5.42MB/s]
 97%|█████████▋| 165M/170M [00:33<00:00, 5.44MB/s]
 97%|█████████▋| 166M/170M [00:33<00:00, 5.42MB/s]
 98%|█████████▊| 166M/170M [00:33<00:00, 5.45MB/s]
 98%|█████████▊| 167M/170M [00:33<00:00, 5.38MB/s]
 98%|█████████▊| 168M/170M [00:34<00:00, 5.22MB/s]
 99%|█████████▊| 168M/170M [00:34<00:00, 5.06MB/s]
 99%|█████████▉| 169M/170M [00:34<00:00, 4.97MB/s]
 99%|█████████▉| 169M/170M [00:34<00:00, 4.93MB/s]
100%|█████████▉| 170M/170M [00:34<00:00, 4.89MB/s]
100%|█████████▉| 170M/170M [00:34<00:00, 4.89MB/s]
100%|██████████| 170M/170M [00:34<00:00, 4.91MB/s]
╭────────────────────────────────────────────────────────────────────╮
│ Configuration for experiment     train_cifar_2026-05-08_21-48-41   │
├────────────────────────────────────────────────────────────────────┤
│ Search algorithm                 BasicVariantGenerator             │
│ Scheduler                        AsyncHyperBandScheduler           │
│ Number of trials                 10                                │
╰────────────────────────────────────────────────────────────────────╯

View detailed results here: /var/lib/ci-user/ray_results/train_cifar_2026-05-08_21-48-41
To visualize your results with TensorBoard, run: `tensorboard --logdir /tmp/ray/session_2026-05-08_21-47-58_664157_4354/artifacts/2026-05-08_21-48-41/train_cifar_2026-05-08_21-48-41/driver_artifacts`

Trial status: 10 PENDING
Current time: 2026-05-08 21:48:41. Total running time: 0s
Logical resource usage: 2.0/16 CPUs, 1.0/1 GPUs (0.0/1.0 accelerator_type:A10G)
╭───────────────────────────────────────────────────────────────────────────────╮
│ Trial name                status       l1     l2            lr     batch_size │
├───────────────────────────────────────────────────────────────────────────────┤
│ train_cifar_ad99a_00000   PENDING      32      1   0.0123941                2 │
│ train_cifar_ad99a_00001   PENDING       4      1   0.0101213               16 │
│ train_cifar_ad99a_00002   PENDING     256      4   0.000710914              2 │
│ train_cifar_ad99a_00003   PENDING      16      1   0.00261551               8 │
│ train_cifar_ad99a_00004   PENDING     256      1   0.000823347              4 │
│ train_cifar_ad99a_00005   PENDING       8     32   0.0264297                4 │
│ train_cifar_ad99a_00006   PENDING     256    128   0.0848744               16 │
│ train_cifar_ad99a_00007   PENDING     128    128   0.0010942                2 │
│ train_cifar_ad99a_00008   PENDING       2      8   0.019285                16 │
│ train_cifar_ad99a_00009   PENDING     128     32   0.00107603               4 │
╰───────────────────────────────────────────────────────────────────────────────╯

Trial train_cifar_ad99a_00000 started with configuration:
╭──────────────────────────────────────────────────╮
│ Trial train_cifar_ad99a_00000 config             │
├──────────────────────────────────────────────────┤
│ batch_size                                     2 │
│ device                                      cuda │
│ l1                                            32 │
│ l2                                             1 │
│ lr                                       0.01239 │
╰──────────────────────────────────────────────────╯
(func pid=5565) [1,  2000] loss: 2.320
(func pid=5565) [1,  4000] loss: 1.158
(func pid=5565) [1,  6000] loss: 0.772
(func pid=5565) [1,  8000] loss: 0.579

Trial status: 1 RUNNING | 9 PENDING
Current time: 2026-05-08 21:49:11. Total running time: 30s
Logical resource usage: 2.0/16 CPUs, 1.0/1 GPUs (0.0/1.0 accelerator_type:A10G)
╭───────────────────────────────────────────────────────────────────────────────╮
│ Trial name                status       l1     l2            lr     batch_size │
├───────────────────────────────────────────────────────────────────────────────┤
│ train_cifar_ad99a_00000   RUNNING      32      1   0.0123941                2 │
│ train_cifar_ad99a_00001   PENDING       4      1   0.0101213               16 │
│ train_cifar_ad99a_00002   PENDING     256      4   0.000710914              2 │
│ train_cifar_ad99a_00003   PENDING      16      1   0.00261551               8 │
│ train_cifar_ad99a_00004   PENDING     256      1   0.000823347              4 │
│ train_cifar_ad99a_00005   PENDING       8     32   0.0264297                4 │
│ train_cifar_ad99a_00006   PENDING     256    128   0.0848744               16 │
│ train_cifar_ad99a_00007   PENDING     128    128   0.0010942                2 │
│ train_cifar_ad99a_00008   PENDING       2      8   0.019285                16 │
│ train_cifar_ad99a_00009   PENDING     128     32   0.00107603               4 │
╰───────────────────────────────────────────────────────────────────────────────╯
(func pid=5565) [1, 10000] loss: 0.463
(func pid=5565) [1, 12000] loss: 0.386
(func pid=5565) [1, 14000] loss: 0.331
(func pid=5565) [1, 16000] loss: 0.290
(func pid=5565) [1, 18000] loss: 0.257
(func pid=5565) [1, 20000] loss: 0.231
Trial status: 1 RUNNING | 9 PENDING
Current time: 2026-05-08 21:49:41. Total running time: 1min 0s
Logical resource usage: 2.0/16 CPUs, 1.0/1 GPUs (0.0/1.0 accelerator_type:A10G)
╭───────────────────────────────────────────────────────────────────────────────╮
│ Trial name                status       l1     l2            lr     batch_size │
├───────────────────────────────────────────────────────────────────────────────┤
│ train_cifar_ad99a_00000   RUNNING      32      1   0.0123941                2 │
│ train_cifar_ad99a_00001   PENDING       4      1   0.0101213               16 │
│ train_cifar_ad99a_00002   PENDING     256      4   0.000710914              2 │
│ train_cifar_ad99a_00003   PENDING      16      1   0.00261551               8 │
│ train_cifar_ad99a_00004   PENDING     256      1   0.000823347              4 │
│ train_cifar_ad99a_00005   PENDING       8     32   0.0264297                4 │
│ train_cifar_ad99a_00006   PENDING     256    128   0.0848744               16 │
│ train_cifar_ad99a_00007   PENDING     128    128   0.0010942                2 │
│ train_cifar_ad99a_00008   PENDING       2      8   0.019285                16 │
│ train_cifar_ad99a_00009   PENDING     128     32   0.00107603               4 │
╰───────────────────────────────────────────────────────────────────────────────╯
(func pid=5565) Checkpoint successfully created at: Checkpoint(filesystem=local, path=/var/lib/ci-user/ray_results/train_cifar_2026-05-08_21-48-41/train_cifar_ad99a_00000_0_batch_size=2,l1=32,l2=1,lr=0.0124_2026-05-08_21-48-41/checkpoint_000000)
(func pid=5565) [2,  2000] loss: 2.317
(func pid=5565) [2,  4000] loss: 1.159
(func pid=5565) [2,  6000] loss: 0.772
(func pid=5565) [2,  8000] loss: 0.579
Trial status: 1 RUNNING | 9 PENDING
Current time: 2026-05-08 21:50:11. Total running time: 1min 30s
Logical resource usage: 2.0/16 CPUs, 1.0/1 GPUs (0.0/1.0 accelerator_type:A10G)
Current best trial: ad99a_00000 with loss=2.31845019159317 and params={'l1': 32, 'l2': 1, 'lr': 0.012394055600315266, 'batch_size': 2, 'device': 'cuda'}
╭──────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────╮
│ Trial name                status       l1     l2            lr     batch_size     iter     total time (s)      loss     accuracy │
├──────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────┤
│ train_cifar_ad99a_00000   RUNNING      32      1   0.0123941                2        1            63.6507   2.31845       0.1025 │
│ train_cifar_ad99a_00001   PENDING       4      1   0.0101213               16                                                    │
│ train_cifar_ad99a_00002   PENDING     256      4   0.000710914              2                                                    │
│ train_cifar_ad99a_00003   PENDING      16      1   0.00261551               8                                                    │
│ train_cifar_ad99a_00004   PENDING     256      1   0.000823347              4                                                    │
│ train_cifar_ad99a_00005   PENDING       8     32   0.0264297                4                                                    │
│ train_cifar_ad99a_00006   PENDING     256    128   0.0848744               16                                                    │
│ train_cifar_ad99a_00007   PENDING     128    128   0.0010942                2                                                    │
│ train_cifar_ad99a_00008   PENDING       2      8   0.019285                16                                                    │
│ train_cifar_ad99a_00009   PENDING     128     32   0.00107603               4                                                    │
╰──────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────╯
(func pid=5565) [2, 10000] loss: 0.463
(func pid=5565) [2, 12000] loss: 0.386
(func pid=5565) [2, 14000] loss: 0.331
(func pid=5565) [2, 16000] loss: 0.290
(func pid=5565) [2, 18000] loss: 0.257
(func pid=5565) [2, 20000] loss: 0.232
Trial status: 1 RUNNING | 9 PENDING
Current time: 2026-05-08 21:50:41. Total running time: 2min 0s
Logical resource usage: 2.0/16 CPUs, 1.0/1 GPUs (0.0/1.0 accelerator_type:A10G)
Current best trial: ad99a_00000 with loss=2.31845019159317 and params={'l1': 32, 'l2': 1, 'lr': 0.012394055600315266, 'batch_size': 2, 'device': 'cuda'}
╭──────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────╮
│ Trial name                status       l1     l2            lr     batch_size     iter     total time (s)      loss     accuracy │
├──────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────┤
│ train_cifar_ad99a_00000   RUNNING      32      1   0.0123941                2        1            63.6507   2.31845       0.1025 │
│ train_cifar_ad99a_00001   PENDING       4      1   0.0101213               16                                                    │
│ train_cifar_ad99a_00002   PENDING     256      4   0.000710914              2                                                    │
│ train_cifar_ad99a_00003   PENDING      16      1   0.00261551               8                                                    │
│ train_cifar_ad99a_00004   PENDING     256      1   0.000823347              4                                                    │
│ train_cifar_ad99a_00005   PENDING       8     32   0.0264297                4                                                    │
│ train_cifar_ad99a_00006   PENDING     256    128   0.0848744               16                                                    │
│ train_cifar_ad99a_00007   PENDING     128    128   0.0010942                2                                                    │
│ train_cifar_ad99a_00008   PENDING       2      8   0.019285                16                                                    │
│ train_cifar_ad99a_00009   PENDING     128     32   0.00107603               4                                                    │
╰──────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────╯
(func pid=5565) Checkpoint successfully created at: Checkpoint(filesystem=local, path=/var/lib/ci-user/ray_results/train_cifar_2026-05-08_21-48-41/train_cifar_ad99a_00000_0_batch_size=2,l1=32,l2=1,lr=0.0124_2026-05-08_21-48-41/checkpoint_000001)
(func pid=5565) [3,  2000] loss: 2.318
(func pid=5565) [3,  4000] loss: 1.158
(func pid=5565) [3,  6000] loss: 0.772
(func pid=5565) [3,  8000] loss: 0.578
Trial status: 1 RUNNING | 9 PENDING
Current time: 2026-05-08 21:51:11. Total running time: 2min 30s
Logical resource usage: 2.0/16 CPUs, 1.0/1 GPUs (0.0/1.0 accelerator_type:A10G)
Current best trial: ad99a_00000 with loss=2.3116115387439726 and params={'l1': 32, 'l2': 1, 'lr': 0.012394055600315266, 'batch_size': 2, 'device': 'cuda'}
╭──────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────╮
│ Trial name                status       l1     l2            lr     batch_size     iter     total time (s)      loss     accuracy │
├──────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────┤
│ train_cifar_ad99a_00000   RUNNING      32      1   0.0123941                2        2            124.928   2.31161       0.1033 │
│ train_cifar_ad99a_00001   PENDING       4      1   0.0101213               16                                                    │
│ train_cifar_ad99a_00002   PENDING     256      4   0.000710914              2                                                    │
│ train_cifar_ad99a_00003   PENDING      16      1   0.00261551               8                                                    │
│ train_cifar_ad99a_00004   PENDING     256      1   0.000823347              4                                                    │
│ train_cifar_ad99a_00005   PENDING       8     32   0.0264297                4                                                    │
│ train_cifar_ad99a_00006   PENDING     256    128   0.0848744               16                                                    │
│ train_cifar_ad99a_00007   PENDING     128    128   0.0010942                2                                                    │
│ train_cifar_ad99a_00008   PENDING       2      8   0.019285                16                                                    │
│ train_cifar_ad99a_00009   PENDING     128     32   0.00107603               4                                                    │
╰──────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────╯
(func pid=5565) [3, 10000] loss: 0.463
(func pid=5565) [3, 12000] loss: 0.386
(func pid=5565) [3, 14000] loss: 0.331
(func pid=5565) [3, 16000] loss: 0.289
(func pid=5565) [3, 18000] loss: 0.257
Trial status: 1 RUNNING | 9 PENDING
Current time: 2026-05-08 21:51:41. Total running time: 3min 0s
Logical resource usage: 2.0/16 CPUs, 1.0/1 GPUs (0.0/1.0 accelerator_type:A10G)
Current best trial: ad99a_00000 with loss=2.3116115387439726 and params={'l1': 32, 'l2': 1, 'lr': 0.012394055600315266, 'batch_size': 2, 'device': 'cuda'}
╭──────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────╮
│ Trial name                status       l1     l2            lr     batch_size     iter     total time (s)      loss     accuracy │
├──────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────┤
│ train_cifar_ad99a_00000   RUNNING      32      1   0.0123941                2        2            124.928   2.31161       0.1033 │
│ train_cifar_ad99a_00001   PENDING       4      1   0.0101213               16                                                    │
│ train_cifar_ad99a_00002   PENDING     256      4   0.000710914              2                                                    │
│ train_cifar_ad99a_00003   PENDING      16      1   0.00261551               8                                                    │
│ train_cifar_ad99a_00004   PENDING     256      1   0.000823347              4                                                    │
│ train_cifar_ad99a_00005   PENDING       8     32   0.0264297                4                                                    │
│ train_cifar_ad99a_00006   PENDING     256    128   0.0848744               16                                                    │
│ train_cifar_ad99a_00007   PENDING     128    128   0.0010942                2                                                    │
│ train_cifar_ad99a_00008   PENDING       2      8   0.019285                16                                                    │
│ train_cifar_ad99a_00009   PENDING     128     32   0.00107603               4                                                    │
╰──────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────╯
(func pid=5565) [3, 20000] loss: 0.231
(func pid=5565) Checkpoint successfully created at: Checkpoint(filesystem=local, path=/var/lib/ci-user/ray_results/train_cifar_2026-05-08_21-48-41/train_cifar_ad99a_00000_0_batch_size=2,l1=32,l2=1,lr=0.0124_2026-05-08_21-48-41/checkpoint_000002)
(func pid=5565) [4,  2000] loss: 2.315
(func pid=5565) [4,  4000] loss: 1.159
(func pid=5565) [4,  6000] loss: 0.772
Trial status: 1 RUNNING | 9 PENDING
Current time: 2026-05-08 21:52:12. Total running time: 3min 30s
Logical resource usage: 2.0/16 CPUs, 1.0/1 GPUs (0.0/1.0 accelerator_type:A10G)
Current best trial: ad99a_00000 with loss=2.3287730796813966 and params={'l1': 32, 'l2': 1, 'lr': 0.012394055600315266, 'batch_size': 2, 'device': 'cuda'}
╭──────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────╮
│ Trial name                status       l1     l2            lr     batch_size     iter     total time (s)      loss     accuracy │
├──────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────┤
│ train_cifar_ad99a_00000   RUNNING      32      1   0.0123941                2        3            186.119   2.32877       0.1017 │
│ train_cifar_ad99a_00001   PENDING       4      1   0.0101213               16                                                    │
│ train_cifar_ad99a_00002   PENDING     256      4   0.000710914              2                                                    │
│ train_cifar_ad99a_00003   PENDING      16      1   0.00261551               8                                                    │
│ train_cifar_ad99a_00004   PENDING     256      1   0.000823347              4                                                    │
│ train_cifar_ad99a_00005   PENDING       8     32   0.0264297                4                                                    │
│ train_cifar_ad99a_00006   PENDING     256    128   0.0848744               16                                                    │
│ train_cifar_ad99a_00007   PENDING     128    128   0.0010942                2                                                    │
│ train_cifar_ad99a_00008   PENDING       2      8   0.019285                16                                                    │
│ train_cifar_ad99a_00009   PENDING     128     32   0.00107603               4                                                    │
╰──────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────╯
(func pid=5565) [4,  8000] loss: 0.579
(func pid=5565) [4, 10000] loss: 0.463
(func pid=5565) [4, 12000] loss: 0.386
(func pid=5565) [4, 14000] loss: 0.331
(func pid=5565) [4, 16000] loss: 0.290
(func pid=5565) [4, 18000] loss: 0.258
Trial status: 1 RUNNING | 9 PENDING
Current time: 2026-05-08 21:52:42. Total running time: 4min 0s
Logical resource usage: 2.0/16 CPUs, 1.0/1 GPUs (0.0/1.0 accelerator_type:A10G)
Current best trial: ad99a_00000 with loss=2.3287730796813966 and params={'l1': 32, 'l2': 1, 'lr': 0.012394055600315266, 'batch_size': 2, 'device': 'cuda'}
╭──────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────╮
│ Trial name                status       l1     l2            lr     batch_size     iter     total time (s)      loss     accuracy │
├──────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────┤
│ train_cifar_ad99a_00000   RUNNING      32      1   0.0123941                2        3            186.119   2.32877       0.1017 │
│ train_cifar_ad99a_00001   PENDING       4      1   0.0101213               16                                                    │
│ train_cifar_ad99a_00002   PENDING     256      4   0.000710914              2                                                    │
│ train_cifar_ad99a_00003   PENDING      16      1   0.00261551               8                                                    │
│ train_cifar_ad99a_00004   PENDING     256      1   0.000823347              4                                                    │
│ train_cifar_ad99a_00005   PENDING       8     32   0.0264297                4                                                    │
│ train_cifar_ad99a_00006   PENDING     256    128   0.0848744               16                                                    │
│ train_cifar_ad99a_00007   PENDING     128    128   0.0010942                2                                                    │
│ train_cifar_ad99a_00008   PENDING       2      8   0.019285                16                                                    │
│ train_cifar_ad99a_00009   PENDING     128     32   0.00107603               4                                                    │
╰──────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────╯
(func pid=5565) [4, 20000] loss: 0.232
(func pid=5565) Checkpoint successfully created at: Checkpoint(filesystem=local, path=/var/lib/ci-user/ray_results/train_cifar_2026-05-08_21-48-41/train_cifar_ad99a_00000_0_batch_size=2,l1=32,l2=1,lr=0.0124_2026-05-08_21-48-41/checkpoint_000003)
(func pid=5565) [5,  2000] loss: 2.317
(func pid=5565) [5,  4000] loss: 1.160
(func pid=5565) [5,  6000] loss: 0.773
Trial status: 1 RUNNING | 9 PENDING
Current time: 2026-05-08 21:53:12. Total running time: 4min 30s
Logical resource usage: 2.0/16 CPUs, 1.0/1 GPUs (0.0/1.0 accelerator_type:A10G)
Current best trial: ad99a_00000 with loss=2.314127134180069 and params={'l1': 32, 'l2': 1, 'lr': 0.012394055600315266, 'batch_size': 2, 'device': 'cuda'}
╭──────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────╮
│ Trial name                status       l1     l2            lr     batch_size     iter     total time (s)      loss     accuracy │
├──────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────┤
│ train_cifar_ad99a_00000   RUNNING      32      1   0.0123941                2        4            247.465   2.31413       0.0963 │
│ train_cifar_ad99a_00001   PENDING       4      1   0.0101213               16                                                    │
│ train_cifar_ad99a_00002   PENDING     256      4   0.000710914              2                                                    │
│ train_cifar_ad99a_00003   PENDING      16      1   0.00261551               8                                                    │
│ train_cifar_ad99a_00004   PENDING     256      1   0.000823347              4                                                    │
│ train_cifar_ad99a_00005   PENDING       8     32   0.0264297                4                                                    │
│ train_cifar_ad99a_00006   PENDING     256    128   0.0848744               16                                                    │
│ train_cifar_ad99a_00007   PENDING     128    128   0.0010942                2                                                    │
│ train_cifar_ad99a_00008   PENDING       2      8   0.019285                16                                                    │
│ train_cifar_ad99a_00009   PENDING     128     32   0.00107603               4                                                    │
╰──────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────╯
(func pid=5565) [5,  8000] loss: 0.579
(func pid=5565) [5, 10000] loss: 0.463
(func pid=5565) [5, 12000] loss: 0.386
(func pid=5565) [5, 14000] loss: 0.331
(func pid=5565) [5, 16000] loss: 0.290
(func pid=5565) [5, 18000] loss: 0.257
Trial status: 1 RUNNING | 9 PENDING
Current time: 2026-05-08 21:53:42. Total running time: 5min 0s
Logical resource usage: 2.0/16 CPUs, 1.0/1 GPUs (0.0/1.0 accelerator_type:A10G)
Current best trial: ad99a_00000 with loss=2.314127134180069 and params={'l1': 32, 'l2': 1, 'lr': 0.012394055600315266, 'batch_size': 2, 'device': 'cuda'}
╭──────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────╮
│ Trial name                status       l1     l2            lr     batch_size     iter     total time (s)      loss     accuracy │
├──────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────┤
│ train_cifar_ad99a_00000   RUNNING      32      1   0.0123941                2        4            247.465   2.31413       0.0963 │
│ train_cifar_ad99a_00001   PENDING       4      1   0.0101213               16                                                    │
│ train_cifar_ad99a_00002   PENDING     256      4   0.000710914              2                                                    │
│ train_cifar_ad99a_00003   PENDING      16      1   0.00261551               8                                                    │
│ train_cifar_ad99a_00004   PENDING     256      1   0.000823347              4                                                    │
│ train_cifar_ad99a_00005   PENDING       8     32   0.0264297                4                                                    │
│ train_cifar_ad99a_00006   PENDING     256    128   0.0848744               16                                                    │
│ train_cifar_ad99a_00007   PENDING     128    128   0.0010942                2                                                    │
│ train_cifar_ad99a_00008   PENDING       2      8   0.019285                16                                                    │
│ train_cifar_ad99a_00009   PENDING     128     32   0.00107603               4                                                    │
╰──────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────╯
(func pid=5565) [5, 20000] loss: 0.232
(func pid=5565) Checkpoint successfully created at: Checkpoint(filesystem=local, path=/var/lib/ci-user/ray_results/train_cifar_2026-05-08_21-48-41/train_cifar_ad99a_00000_0_batch_size=2,l1=32,l2=1,lr=0.0124_2026-05-08_21-48-41/checkpoint_000004)
(func pid=5565) [6,  2000] loss: 2.319
(func pid=5565) [6,  4000] loss: 1.158
(func pid=5565) [6,  6000] loss: 0.772
Trial status: 1 RUNNING | 9 PENDING
Current time: 2026-05-08 21:54:12. Total running time: 5min 30s
Logical resource usage: 2.0/16 CPUs, 1.0/1 GPUs (0.0/1.0 accelerator_type:A10G)
Current best trial: ad99a_00000 with loss=2.322457846403122 and params={'l1': 32, 'l2': 1, 'lr': 0.012394055600315266, 'batch_size': 2, 'device': 'cuda'}
╭──────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────╮
│ Trial name                status       l1     l2            lr     batch_size     iter     total time (s)      loss     accuracy │
├──────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────┤
│ train_cifar_ad99a_00000   RUNNING      32      1   0.0123941                2        5              308.4   2.32246       0.0922 │
│ train_cifar_ad99a_00001   PENDING       4      1   0.0101213               16                                                    │
│ train_cifar_ad99a_00002   PENDING     256      4   0.000710914              2                                                    │
│ train_cifar_ad99a_00003   PENDING      16      1   0.00261551               8                                                    │
│ train_cifar_ad99a_00004   PENDING     256      1   0.000823347              4                                                    │
│ train_cifar_ad99a_00005   PENDING       8     32   0.0264297                4                                                    │
│ train_cifar_ad99a_00006   PENDING     256    128   0.0848744               16                                                    │
│ train_cifar_ad99a_00007   PENDING     128    128   0.0010942                2                                                    │
│ train_cifar_ad99a_00008   PENDING       2      8   0.019285                16                                                    │
│ train_cifar_ad99a_00009   PENDING     128     32   0.00107603               4                                                    │
╰──────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────╯
(func pid=5565) [6,  8000] loss: 0.579
(func pid=5565) [6, 10000] loss: 0.464
(func pid=5565) [6, 12000] loss: 0.386
(func pid=5565) [6, 14000] loss: 0.331
(func pid=5565) [6, 16000] loss: 0.290
(func pid=5565) [6, 18000] loss: 0.257
Trial status: 1 RUNNING | 9 PENDING
Current time: 2026-05-08 21:54:42. Total running time: 6min 0s
Logical resource usage: 2.0/16 CPUs, 1.0/1 GPUs (0.0/1.0 accelerator_type:A10G)
Current best trial: ad99a_00000 with loss=2.322457846403122 and params={'l1': 32, 'l2': 1, 'lr': 0.012394055600315266, 'batch_size': 2, 'device': 'cuda'}
╭──────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────╮
│ Trial name                status       l1     l2            lr     batch_size     iter     total time (s)      loss     accuracy │
├──────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────┤
│ train_cifar_ad99a_00000   RUNNING      32      1   0.0123941                2        5              308.4   2.32246       0.0922 │
│ train_cifar_ad99a_00001   PENDING       4      1   0.0101213               16                                                    │
│ train_cifar_ad99a_00002   PENDING     256      4   0.000710914              2                                                    │
│ train_cifar_ad99a_00003   PENDING      16      1   0.00261551               8                                                    │
│ train_cifar_ad99a_00004   PENDING     256      1   0.000823347              4                                                    │
│ train_cifar_ad99a_00005   PENDING       8     32   0.0264297                4                                                    │
│ train_cifar_ad99a_00006   PENDING     256    128   0.0848744               16                                                    │
│ train_cifar_ad99a_00007   PENDING     128    128   0.0010942                2                                                    │
│ train_cifar_ad99a_00008   PENDING       2      8   0.019285                16                                                    │
│ train_cifar_ad99a_00009   PENDING     128     32   0.00107603               4                                                    │
╰──────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────╯
(func pid=5565) [6, 20000] loss: 0.232
(func pid=5565) Checkpoint successfully created at: Checkpoint(filesystem=local, path=/var/lib/ci-user/ray_results/train_cifar_2026-05-08_21-48-41/train_cifar_ad99a_00000_0_batch_size=2,l1=32,l2=1,lr=0.0124_2026-05-08_21-48-41/checkpoint_000005)
(func pid=5565) [7,  2000] loss: 2.313
(func pid=5565) [7,  4000] loss: 1.160
(func pid=5565) [7,  6000] loss: 0.772
Trial status: 1 RUNNING | 9 PENDING
Current time: 2026-05-08 21:55:12. Total running time: 6min 31s
Logical resource usage: 2.0/16 CPUs, 1.0/1 GPUs (0.0/1.0 accelerator_type:A10G)
Current best trial: ad99a_00000 with loss=2.3089891775608065 and params={'l1': 32, 'l2': 1, 'lr': 0.012394055600315266, 'batch_size': 2, 'device': 'cuda'}
╭──────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────╮
│ Trial name                status       l1     l2            lr     batch_size     iter     total time (s)      loss     accuracy │
├──────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────┤
│ train_cifar_ad99a_00000   RUNNING      32      1   0.0123941                2        6            370.055   2.30899       0.1011 │
│ train_cifar_ad99a_00001   PENDING       4      1   0.0101213               16                                                    │
│ train_cifar_ad99a_00002   PENDING     256      4   0.000710914              2                                                    │
│ train_cifar_ad99a_00003   PENDING      16      1   0.00261551               8                                                    │
│ train_cifar_ad99a_00004   PENDING     256      1   0.000823347              4                                                    │
│ train_cifar_ad99a_00005   PENDING       8     32   0.0264297                4                                                    │
│ train_cifar_ad99a_00006   PENDING     256    128   0.0848744               16                                                    │
│ train_cifar_ad99a_00007   PENDING     128    128   0.0010942                2                                                    │
│ train_cifar_ad99a_00008   PENDING       2      8   0.019285                16                                                    │
│ train_cifar_ad99a_00009   PENDING     128     32   0.00107603               4                                                    │
╰──────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────╯
(func pid=5565) [7,  8000] loss: 0.579
(func pid=5565) [7, 10000] loss: 0.464
(func pid=5565) [7, 12000] loss: 0.386
(func pid=5565) [7, 14000] loss: 0.331
(func pid=5565) [7, 16000] loss: 0.290
Trial status: 1 RUNNING | 9 PENDING
Current time: 2026-05-08 21:55:42. Total running time: 7min 1s
Logical resource usage: 2.0/16 CPUs, 1.0/1 GPUs (0.0/1.0 accelerator_type:A10G)
Current best trial: ad99a_00000 with loss=2.3089891775608065 and params={'l1': 32, 'l2': 1, 'lr': 0.012394055600315266, 'batch_size': 2, 'device': 'cuda'}
╭──────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────╮
│ Trial name                status       l1     l2            lr     batch_size     iter     total time (s)      loss     accuracy │
├──────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────┤
│ train_cifar_ad99a_00000   RUNNING      32      1   0.0123941                2        6            370.055   2.30899       0.1011 │
│ train_cifar_ad99a_00001   PENDING       4      1   0.0101213               16                                                    │
│ train_cifar_ad99a_00002   PENDING     256      4   0.000710914              2                                                    │
│ train_cifar_ad99a_00003   PENDING      16      1   0.00261551               8                                                    │
│ train_cifar_ad99a_00004   PENDING     256      1   0.000823347              4                                                    │
│ train_cifar_ad99a_00005   PENDING       8     32   0.0264297                4                                                    │
│ train_cifar_ad99a_00006   PENDING     256    128   0.0848744               16                                                    │
│ train_cifar_ad99a_00007   PENDING     128    128   0.0010942                2                                                    │
│ train_cifar_ad99a_00008   PENDING       2      8   0.019285                16                                                    │
│ train_cifar_ad99a_00009   PENDING     128     32   0.00107603               4                                                    │
╰──────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────╯
(func pid=5565) [7, 18000] loss: 0.257
(func pid=5565) [7, 20000] loss: 0.232
(func pid=5565) Checkpoint successfully created at: Checkpoint(filesystem=local, path=/var/lib/ci-user/ray_results/train_cifar_2026-05-08_21-48-41/train_cifar_ad99a_00000_0_batch_size=2,l1=32,l2=1,lr=0.0124_2026-05-08_21-48-41/checkpoint_000006)
(func pid=5565) [8,  2000] loss: 2.314
(func pid=5565) [8,  4000] loss: 1.159
Trial status: 1 RUNNING | 9 PENDING
Current time: 2026-05-08 21:56:12. Total running time: 7min 31s
Logical resource usage: 2.0/16 CPUs, 1.0/1 GPUs (0.0/1.0 accelerator_type:A10G)
Current best trial: ad99a_00000 with loss=2.3065735617160796 and params={'l1': 32, 'l2': 1, 'lr': 0.012394055600315266, 'batch_size': 2, 'device': 'cuda'}
╭──────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────╮
│ Trial name                status       l1     l2            lr     batch_size     iter     total time (s)      loss     accuracy │
├──────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────┤
│ train_cifar_ad99a_00000   RUNNING      32      1   0.0123941                2        7            431.735   2.30657       0.1034 │
│ train_cifar_ad99a_00001   PENDING       4      1   0.0101213               16                                                    │
│ train_cifar_ad99a_00002   PENDING     256      4   0.000710914              2                                                    │
│ train_cifar_ad99a_00003   PENDING      16      1   0.00261551               8                                                    │
│ train_cifar_ad99a_00004   PENDING     256      1   0.000823347              4                                                    │
│ train_cifar_ad99a_00005   PENDING       8     32   0.0264297                4                                                    │
│ train_cifar_ad99a_00006   PENDING     256    128   0.0848744               16                                                    │
│ train_cifar_ad99a_00007   PENDING     128    128   0.0010942                2                                                    │
│ train_cifar_ad99a_00008   PENDING       2      8   0.019285                16                                                    │
│ train_cifar_ad99a_00009   PENDING     128     32   0.00107603               4                                                    │
╰──────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────╯
(func pid=5565) [8,  6000] loss: 0.772
(func pid=5565) [8,  8000] loss: 0.580
(func pid=5565) [8, 10000] loss: 0.464
(func pid=5565) [8, 12000] loss: 0.386
(func pid=5565) [8, 14000] loss: 0.331
(func pid=5565) [8, 16000] loss: 0.290
Trial status: 1 RUNNING | 9 PENDING
Current time: 2026-05-08 21:56:42. Total running time: 8min 1s
Logical resource usage: 2.0/16 CPUs, 1.0/1 GPUs (0.0/1.0 accelerator_type:A10G)
Current best trial: ad99a_00000 with loss=2.3065735617160796 and params={'l1': 32, 'l2': 1, 'lr': 0.012394055600315266, 'batch_size': 2, 'device': 'cuda'}
╭──────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────╮
│ Trial name                status       l1     l2            lr     batch_size     iter     total time (s)      loss     accuracy │
├──────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────┤
│ train_cifar_ad99a_00000   RUNNING      32      1   0.0123941                2        7            431.735   2.30657       0.1034 │
│ train_cifar_ad99a_00001   PENDING       4      1   0.0101213               16                                                    │
│ train_cifar_ad99a_00002   PENDING     256      4   0.000710914              2                                                    │
│ train_cifar_ad99a_00003   PENDING      16      1   0.00261551               8                                                    │
│ train_cifar_ad99a_00004   PENDING     256      1   0.000823347              4                                                    │
│ train_cifar_ad99a_00005   PENDING       8     32   0.0264297                4                                                    │
│ train_cifar_ad99a_00006   PENDING     256    128   0.0848744               16                                                    │
│ train_cifar_ad99a_00007   PENDING     128    128   0.0010942                2                                                    │
│ train_cifar_ad99a_00008   PENDING       2      8   0.019285                16                                                    │
│ train_cifar_ad99a_00009   PENDING     128     32   0.00107603               4                                                    │
╰──────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────╯
(func pid=5565) [8, 18000] loss: 0.257
(func pid=5565) [8, 20000] loss: 0.232
(func pid=5565) Checkpoint successfully created at: Checkpoint(filesystem=local, path=/var/lib/ci-user/ray_results/train_cifar_2026-05-08_21-48-41/train_cifar_ad99a_00000_0_batch_size=2,l1=32,l2=1,lr=0.0124_2026-05-08_21-48-41/checkpoint_000007)
(func pid=5565) [9,  2000] loss: 2.317
(func pid=5565) [9,  4000] loss: 1.157
Trial status: 1 RUNNING | 9 PENDING
Current time: 2026-05-08 21:57:12. Total running time: 8min 31s
Logical resource usage: 2.0/16 CPUs, 1.0/1 GPUs (0.0/1.0 accelerator_type:A10G)
Current best trial: ad99a_00000 with loss=2.306291142511368 and params={'l1': 32, 'l2': 1, 'lr': 0.012394055600315266, 'batch_size': 2, 'device': 'cuda'}
╭──────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────╮
│ Trial name                status       l1     l2            lr     batch_size     iter     total time (s)      loss     accuracy │
├──────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────┤
│ train_cifar_ad99a_00000   RUNNING      32      1   0.0123941                2        8            492.908   2.30629       0.1025 │
│ train_cifar_ad99a_00001   PENDING       4      1   0.0101213               16                                                    │
│ train_cifar_ad99a_00002   PENDING     256      4   0.000710914              2                                                    │
│ train_cifar_ad99a_00003   PENDING      16      1   0.00261551               8                                                    │
│ train_cifar_ad99a_00004   PENDING     256      1   0.000823347              4                                                    │
│ train_cifar_ad99a_00005   PENDING       8     32   0.0264297                4                                                    │
│ train_cifar_ad99a_00006   PENDING     256    128   0.0848744               16                                                    │
│ train_cifar_ad99a_00007   PENDING     128    128   0.0010942                2                                                    │
│ train_cifar_ad99a_00008   PENDING       2      8   0.019285                16                                                    │
│ train_cifar_ad99a_00009   PENDING     128     32   0.00107603               4                                                    │
╰──────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────╯
(func pid=5565) [9,  6000] loss: 0.771
(func pid=5565) [9,  8000] loss: 0.579
(func pid=5565) [9, 10000] loss: 0.463
(func pid=5565) [9, 12000] loss: 0.386
(func pid=5565) [9, 14000] loss: 0.331
(func pid=5565) [9, 16000] loss: 0.290
Trial status: 1 RUNNING | 9 PENDING
Current time: 2026-05-08 21:57:42. Total running time: 9min 1s
Logical resource usage: 2.0/16 CPUs, 1.0/1 GPUs (0.0/1.0 accelerator_type:A10G)
Current best trial: ad99a_00000 with loss=2.306291142511368 and params={'l1': 32, 'l2': 1, 'lr': 0.012394055600315266, 'batch_size': 2, 'device': 'cuda'}
╭──────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────╮
│ Trial name                status       l1     l2            lr     batch_size     iter     total time (s)      loss     accuracy │
├──────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────┤
│ train_cifar_ad99a_00000   RUNNING      32      1   0.0123941                2        8            492.908   2.30629       0.1025 │
│ train_cifar_ad99a_00001   PENDING       4      1   0.0101213               16                                                    │
│ train_cifar_ad99a_00002   PENDING     256      4   0.000710914              2                                                    │
│ train_cifar_ad99a_00003   PENDING      16      1   0.00261551               8                                                    │
│ train_cifar_ad99a_00004   PENDING     256      1   0.000823347              4                                                    │
│ train_cifar_ad99a_00005   PENDING       8     32   0.0264297                4                                                    │
│ train_cifar_ad99a_00006   PENDING     256    128   0.0848744               16                                                    │
│ train_cifar_ad99a_00007   PENDING     128    128   0.0010942                2                                                    │
│ train_cifar_ad99a_00008   PENDING       2      8   0.019285                16                                                    │
│ train_cifar_ad99a_00009   PENDING     128     32   0.00107603               4                                                    │
╰──────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────╯
(func pid=5565) [9, 18000] loss: 0.257
(func pid=5565) [9, 20000] loss: 0.232
(func pid=5565) Checkpoint successfully created at: Checkpoint(filesystem=local, path=/var/lib/ci-user/ray_results/train_cifar_2026-05-08_21-48-41/train_cifar_ad99a_00000_0_batch_size=2,l1=32,l2=1,lr=0.0124_2026-05-08_21-48-41/checkpoint_000008)
(func pid=5565) [10,  2000] loss: 2.318
(func pid=5565) [10,  4000] loss: 1.159
Trial status: 1 RUNNING | 9 PENDING
Current time: 2026-05-08 21:58:12. Total running time: 9min 31s
Logical resource usage: 2.0/16 CPUs, 1.0/1 GPUs (0.0/1.0 accelerator_type:A10G)
Current best trial: ad99a_00000 with loss=2.3157137302160264 and params={'l1': 32, 'l2': 1, 'lr': 0.012394055600315266, 'batch_size': 2, 'device': 'cuda'}
╭──────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────╮
│ Trial name                status       l1     l2            lr     batch_size     iter     total time (s)      loss     accuracy │
├──────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────┤
│ train_cifar_ad99a_00000   RUNNING      32      1   0.0123941                2        9            553.853   2.31571       0.1025 │
│ train_cifar_ad99a_00001   PENDING       4      1   0.0101213               16                                                    │
│ train_cifar_ad99a_00002   PENDING     256      4   0.000710914              2                                                    │
│ train_cifar_ad99a_00003   PENDING      16      1   0.00261551               8                                                    │
│ train_cifar_ad99a_00004   PENDING     256      1   0.000823347              4                                                    │
│ train_cifar_ad99a_00005   PENDING       8     32   0.0264297                4                                                    │
│ train_cifar_ad99a_00006   PENDING     256    128   0.0848744               16                                                    │
│ train_cifar_ad99a_00007   PENDING     128    128   0.0010942                2                                                    │
│ train_cifar_ad99a_00008   PENDING       2      8   0.019285                16                                                    │
│ train_cifar_ad99a_00009   PENDING     128     32   0.00107603               4                                                    │
╰──────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────╯
(func pid=5565) [10,  6000] loss: 0.773
(func pid=5565) [10,  8000] loss: 0.579
(func pid=5565) [10, 10000] loss: 0.463
(func pid=5565) [10, 12000] loss: 0.386
(func pid=5565) [10, 14000] loss: 0.331
(func pid=5565) [10, 16000] loss: 0.290
Trial status: 1 RUNNING | 9 PENDING
Current time: 2026-05-08 21:58:42. Total running time: 10min 1s
Logical resource usage: 2.0/16 CPUs, 1.0/1 GPUs (0.0/1.0 accelerator_type:A10G)
Current best trial: ad99a_00000 with loss=2.3157137302160264 and params={'l1': 32, 'l2': 1, 'lr': 0.012394055600315266, 'batch_size': 2, 'device': 'cuda'}
╭──────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────╮
│ Trial name                status       l1     l2            lr     batch_size     iter     total time (s)      loss     accuracy │
├──────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────┤
│ train_cifar_ad99a_00000   RUNNING      32      1   0.0123941                2        9            553.853   2.31571       0.1025 │
│ train_cifar_ad99a_00001   PENDING       4      1   0.0101213               16                                                    │
│ train_cifar_ad99a_00002   PENDING     256      4   0.000710914              2                                                    │
│ train_cifar_ad99a_00003   PENDING      16      1   0.00261551               8                                                    │
│ train_cifar_ad99a_00004   PENDING     256      1   0.000823347              4                                                    │
│ train_cifar_ad99a_00005   PENDING       8     32   0.0264297                4                                                    │
│ train_cifar_ad99a_00006   PENDING     256    128   0.0848744               16                                                    │
│ train_cifar_ad99a_00007   PENDING     128    128   0.0010942                2                                                    │
│ train_cifar_ad99a_00008   PENDING       2      8   0.019285                16                                                    │
│ train_cifar_ad99a_00009   PENDING     128     32   0.00107603               4                                                    │
╰──────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────╯
(func pid=5565) [10, 18000] loss: 0.257
(func pid=5565) [10, 20000] loss: 0.231

Trial train_cifar_ad99a_00000 completed after 10 iterations at 2026-05-08 21:59:00. Total running time: 10min 19s
╭────────────────────────────────────────────────────────────╮
│ Trial train_cifar_ad99a_00000 result                       │
├────────────────────────────────────────────────────────────┤
│ checkpoint_dir_name                      checkpoint_000009 │
│ time_this_iter_s                                  61.31408 │
│ time_total_s                                     615.16682 │
│ training_iteration                                      10 │
│ accuracy                                            0.1017 │
│ loss                                               2.31109 │
╰────────────────────────────────────────────────────────────╯
(func pid=5565) Checkpoint successfully created at: Checkpoint(filesystem=local, path=/var/lib/ci-user/ray_results/train_cifar_2026-05-08_21-48-41/train_cifar_ad99a_00000_0_batch_size=2,l1=32,l2=1,lr=0.0124_2026-05-08_21-48-41/checkpoint_000009)

Trial train_cifar_ad99a_00001 started with configuration:
╭──────────────────────────────────────────────────╮
│ Trial train_cifar_ad99a_00001 config             │
├──────────────────────────────────────────────────┤
│ batch_size                                    16 │
│ device                                      cuda │
│ l1                                             4 │
│ l2                                             1 │
│ lr                                       0.01012 │
╰──────────────────────────────────────────────────╯

Trial status: 1 TERMINATED | 1 RUNNING | 8 PENDING
Current time: 2026-05-08 21:59:12. Total running time: 10min 31s
Logical resource usage: 2.0/16 CPUs, 1.0/1 GPUs (0.0/1.0 accelerator_type:A10G)
Current best trial: ad99a_00000 with loss=2.3110877054691317 and params={'l1': 32, 'l2': 1, 'lr': 0.012394055600315266, 'batch_size': 2, 'device': 'cuda'}
╭────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────╮
│ Trial name                status         l1     l2            lr     batch_size     iter     total time (s)      loss     accuracy │
├────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────┤
│ train_cifar_ad99a_00001   RUNNING         4      1   0.0101213               16                                                    │
│ train_cifar_ad99a_00000   TERMINATED     32      1   0.0123941                2       10            615.167   2.31109       0.1017 │
│ train_cifar_ad99a_00002   PENDING       256      4   0.000710914              2                                                    │
│ train_cifar_ad99a_00003   PENDING        16      1   0.00261551               8                                                    │
│ train_cifar_ad99a_00004   PENDING       256      1   0.000823347              4                                                    │
│ train_cifar_ad99a_00005   PENDING         8     32   0.0264297                4                                                    │
│ train_cifar_ad99a_00006   PENDING       256    128   0.0848744               16                                                    │
│ train_cifar_ad99a_00007   PENDING       128    128   0.0010942                2                                                    │
│ train_cifar_ad99a_00008   PENDING         2      8   0.019285                16                                                    │
│ train_cifar_ad99a_00009   PENDING       128     32   0.00107603               4                                                    │
╰────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────╯
(func pid=7467) [1,  2000] loss: 2.311
(func pid=7467) Checkpoint successfully created at: Checkpoint(filesystem=local, path=/var/lib/ci-user/ray_results/train_cifar_2026-05-08_21-48-41/train_cifar_ad99a_00001_1_batch_size=16,l1=4,l2=1,lr=0.0101_2026-05-08_21-48-41/checkpoint_000000)
(func pid=7467) [2,  2000] loss: 2.304
(func pid=7467) Checkpoint successfully created at: Checkpoint(filesystem=local, path=/var/lib/ci-user/ray_results/train_cifar_2026-05-08_21-48-41/train_cifar_ad99a_00001_1_batch_size=16,l1=4,l2=1,lr=0.0101_2026-05-08_21-48-41/checkpoint_000001)
(func pid=7467) [3,  2000] loss: 2.304
(func pid=7467) Checkpoint successfully created at: Checkpoint(filesystem=local, path=/var/lib/ci-user/ray_results/train_cifar_2026-05-08_21-48-41/train_cifar_ad99a_00001_1_batch_size=16,l1=4,l2=1,lr=0.0101_2026-05-08_21-48-41/checkpoint_000002)
(func pid=7467) [4,  2000] loss: 2.304
Trial status: 1 TERMINATED | 1 RUNNING | 8 PENDING
Current time: 2026-05-08 21:59:43. Total running time: 11min 1s
Logical resource usage: 2.0/16 CPUs, 1.0/1 GPUs (0.0/1.0 accelerator_type:A10G)
Current best trial: ad99a_00001 with loss=2.303117914199829 and params={'l1': 4, 'l2': 1, 'lr': 0.01012128934572376, 'batch_size': 16, 'device': 'cuda'}
╭────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────╮
│ Trial name                status         l1     l2            lr     batch_size     iter     total time (s)      loss     accuracy │
├────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────┤
│ train_cifar_ad99a_00001   RUNNING         4      1   0.0101213               16        3            28.1574   2.30312       0.1016 │
│ train_cifar_ad99a_00000   TERMINATED     32      1   0.0123941                2       10           615.167    2.31109       0.1017 │
│ train_cifar_ad99a_00002   PENDING       256      4   0.000710914              2                                                    │
│ train_cifar_ad99a_00003   PENDING        16      1   0.00261551               8                                                    │
│ train_cifar_ad99a_00004   PENDING       256      1   0.000823347              4                                                    │
│ train_cifar_ad99a_00005   PENDING         8     32   0.0264297                4                                                    │
│ train_cifar_ad99a_00006   PENDING       256    128   0.0848744               16                                                    │
│ train_cifar_ad99a_00007   PENDING       128    128   0.0010942                2                                                    │
│ train_cifar_ad99a_00008   PENDING         2      8   0.019285                16                                                    │
│ train_cifar_ad99a_00009   PENDING       128     32   0.00107603               4                                                    │
╰────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────╯
(func pid=7467) Checkpoint successfully created at: Checkpoint(filesystem=local, path=/var/lib/ci-user/ray_results/train_cifar_2026-05-08_21-48-41/train_cifar_ad99a_00001_1_batch_size=16,l1=4,l2=1,lr=0.0101_2026-05-08_21-48-41/checkpoint_000003)
(func pid=7467) [5,  2000] loss: 2.304
(func pid=7467) Checkpoint successfully created at: Checkpoint(filesystem=local, path=/var/lib/ci-user/ray_results/train_cifar_2026-05-08_21-48-41/train_cifar_ad99a_00001_1_batch_size=16,l1=4,l2=1,lr=0.0101_2026-05-08_21-48-41/checkpoint_000004)
(func pid=7467) [6,  2000] loss: 2.304
(func pid=7467) Checkpoint successfully created at: Checkpoint(filesystem=local, path=/var/lib/ci-user/ray_results/train_cifar_2026-05-08_21-48-41/train_cifar_ad99a_00001_1_batch_size=16,l1=4,l2=1,lr=0.0101_2026-05-08_21-48-41/checkpoint_000005)
(func pid=7467) [7,  2000] loss: 2.304
(func pid=7467) Checkpoint successfully created at: Checkpoint(filesystem=local, path=/var/lib/ci-user/ray_results/train_cifar_2026-05-08_21-48-41/train_cifar_ad99a_00001_1_batch_size=16,l1=4,l2=1,lr=0.0101_2026-05-08_21-48-41/checkpoint_000006)
Trial status: 1 TERMINATED | 1 RUNNING | 8 PENDING
Current time: 2026-05-08 22:00:13. Total running time: 11min 31s
Logical resource usage: 2.0/16 CPUs, 1.0/1 GPUs (0.0/1.0 accelerator_type:A10G)
Current best trial: ad99a_00001 with loss=2.3035923458099363 and params={'l1': 4, 'l2': 1, 'lr': 0.01012128934572376, 'batch_size': 16, 'device': 'cuda'}
╭────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────╮
│ Trial name                status         l1     l2            lr     batch_size     iter     total time (s)      loss     accuracy │
├────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────┤
│ train_cifar_ad99a_00001   RUNNING         4      1   0.0101213               16        7            63.6182   2.30359       0.1011 │
│ train_cifar_ad99a_00000   TERMINATED     32      1   0.0123941                2       10           615.167    2.31109       0.1017 │
│ train_cifar_ad99a_00002   PENDING       256      4   0.000710914              2                                                    │
│ train_cifar_ad99a_00003   PENDING        16      1   0.00261551               8                                                    │
│ train_cifar_ad99a_00004   PENDING       256      1   0.000823347              4                                                    │
│ train_cifar_ad99a_00005   PENDING         8     32   0.0264297                4                                                    │
│ train_cifar_ad99a_00006   PENDING       256    128   0.0848744               16                                                    │
│ train_cifar_ad99a_00007   PENDING       128    128   0.0010942                2                                                    │
│ train_cifar_ad99a_00008   PENDING         2      8   0.019285                16                                                    │
│ train_cifar_ad99a_00009   PENDING       128     32   0.00107603               4                                                    │
╰────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────╯
(func pid=7467) [8,  2000] loss: 2.304
(func pid=7467) Checkpoint successfully created at: Checkpoint(filesystem=local, path=/var/lib/ci-user/ray_results/train_cifar_2026-05-08_21-48-41/train_cifar_ad99a_00001_1_batch_size=16,l1=4,l2=1,lr=0.0101_2026-05-08_21-48-41/checkpoint_000007)
(func pid=7467) [9,  2000] loss: 2.304
(func pid=7467) Checkpoint successfully created at: Checkpoint(filesystem=local, path=/var/lib/ci-user/ray_results/train_cifar_2026-05-08_21-48-41/train_cifar_ad99a_00001_1_batch_size=16,l1=4,l2=1,lr=0.0101_2026-05-08_21-48-41/checkpoint_000008)
(func pid=7467) [10,  2000] loss: 2.304

Trial train_cifar_ad99a_00001 completed after 10 iterations at 2026-05-08 22:00:35. Total running time: 11min 54s
╭────────────────────────────────────────────────────────────╮
│ Trial train_cifar_ad99a_00001 result                       │
├────────────────────────────────────────────────────────────┤
│ checkpoint_dir_name                      checkpoint_000009 │
│ time_this_iter_s                                   8.71906 │
│ time_total_s                                      89.68737 │
│ training_iteration                                      10 │
│ accuracy                                            0.0997 │
│ loss                                               2.30332 │
╰────────────────────────────────────────────────────────────╯
(func pid=7467) Checkpoint successfully created at: Checkpoint(filesystem=local, path=/var/lib/ci-user/ray_results/train_cifar_2026-05-08_21-48-41/train_cifar_ad99a_00001_1_batch_size=16,l1=4,l2=1,lr=0.0101_2026-05-08_21-48-41/checkpoint_000009)

Trial train_cifar_ad99a_00002 started with configuration:
╭──────────────────────────────────────────────────╮
│ Trial train_cifar_ad99a_00002 config             │
├──────────────────────────────────────────────────┤
│ batch_size                                     2 │
│ device                                      cuda │
│ l1                                           256 │
│ l2                                             4 │
│ lr                                       0.00071 │
╰──────────────────────────────────────────────────╯

Trial status: 2 TERMINATED | 1 RUNNING | 7 PENDING
Current time: 2026-05-08 22:00:43. Total running time: 12min 1s
Logical resource usage: 2.0/16 CPUs, 1.0/1 GPUs (0.0/1.0 accelerator_type:A10G)
Current best trial: ad99a_00001 with loss=2.303318368148804 and params={'l1': 4, 'l2': 1, 'lr': 0.01012128934572376, 'batch_size': 16, 'device': 'cuda'}
╭────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────╮
│ Trial name                status         l1     l2            lr     batch_size     iter     total time (s)      loss     accuracy │
├────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────┤
│ train_cifar_ad99a_00002   RUNNING       256      4   0.000710914              2                                                    │
│ train_cifar_ad99a_00000   TERMINATED     32      1   0.0123941                2       10           615.167    2.31109       0.1017 │
│ train_cifar_ad99a_00001   TERMINATED      4      1   0.0101213               16       10            89.6874   2.30332       0.0997 │
│ train_cifar_ad99a_00003   PENDING        16      1   0.00261551               8                                                    │
│ train_cifar_ad99a_00004   PENDING       256      1   0.000823347              4                                                    │
│ train_cifar_ad99a_00005   PENDING         8     32   0.0264297                4                                                    │
│ train_cifar_ad99a_00006   PENDING       256    128   0.0848744               16                                                    │
│ train_cifar_ad99a_00007   PENDING       128    128   0.0010942                2                                                    │
│ train_cifar_ad99a_00008   PENDING         2      8   0.019285                16                                                    │
│ train_cifar_ad99a_00009   PENDING       128     32   0.00107603               4                                                    │
╰────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────╯
(func pid=8359) [1,  2000] loss: 2.308
(func pid=8359) [1,  4000] loss: 1.124
(func pid=8359) [1,  6000] loss: 0.699
(func pid=8359) [1,  8000] loss: 0.485
(func pid=8359) [1, 10000] loss: 0.375
(func pid=8359) [1, 12000] loss: 0.298
Trial status: 2 TERMINATED | 1 RUNNING | 7 PENDING
Current time: 2026-05-08 22:01:13. Total running time: 12min 31s
Logical resource usage: 2.0/16 CPUs, 1.0/1 GPUs (0.0/1.0 accelerator_type:A10G)
Current best trial: ad99a_00001 with loss=2.303318368148804 and params={'l1': 4, 'l2': 1, 'lr': 0.01012128934572376, 'batch_size': 16, 'device': 'cuda'}
╭────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────╮
│ Trial name                status         l1     l2            lr     batch_size     iter     total time (s)      loss     accuracy │
├────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────┤
│ train_cifar_ad99a_00002   RUNNING       256      4   0.000710914              2                                                    │
│ train_cifar_ad99a_00000   TERMINATED     32      1   0.0123941                2       10           615.167    2.31109       0.1017 │
│ train_cifar_ad99a_00001   TERMINATED      4      1   0.0101213               16       10            89.6874   2.30332       0.0997 │
│ train_cifar_ad99a_00003   PENDING        16      1   0.00261551               8                                                    │
│ train_cifar_ad99a_00004   PENDING       256      1   0.000823347              4                                                    │
│ train_cifar_ad99a_00005   PENDING         8     32   0.0264297                4                                                    │
│ train_cifar_ad99a_00006   PENDING       256    128   0.0848744               16                                                    │
│ train_cifar_ad99a_00007   PENDING       128    128   0.0010942                2                                                    │
│ train_cifar_ad99a_00008   PENDING         2      8   0.019285                16                                                    │
│ train_cifar_ad99a_00009   PENDING       128     32   0.00107603               4                                                    │
╰────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────╯
(func pid=8359) [1, 14000] loss: 0.252
(func pid=8359) [1, 16000] loss: 0.218
(func pid=8359) [1, 18000] loss: 0.191
(func pid=8359) [1, 20000] loss: 0.169
(func pid=8359) Checkpoint successfully created at: Checkpoint(filesystem=local, path=/var/lib/ci-user/ray_results/train_cifar_2026-05-08_21-48-41/train_cifar_ad99a_00002_2_batch_size=2,l1=256,l2=4,lr=0.0007_2026-05-08_21-48-41/checkpoint_000000)
Trial status: 2 TERMINATED | 1 RUNNING | 7 PENDING
Current time: 2026-05-08 22:01:43. Total running time: 13min 1s
Logical resource usage: 2.0/16 CPUs, 1.0/1 GPUs (0.0/1.0 accelerator_type:A10G)
Current best trial: ad99a_00002 with loss=1.6366215757608413 and params={'l1': 256, 'l2': 4, 'lr': 0.0007109136804996362, 'batch_size': 2, 'device': 'cuda'}
╭────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────╮
│ Trial name                status         l1     l2            lr     batch_size     iter     total time (s)      loss     accuracy │
├────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────┤
│ train_cifar_ad99a_00002   RUNNING       256      4   0.000710914              2        1            61.8357   1.63662       0.3915 │
│ train_cifar_ad99a_00000   TERMINATED     32      1   0.0123941                2       10           615.167    2.31109       0.1017 │
│ train_cifar_ad99a_00001   TERMINATED      4      1   0.0101213               16       10            89.6874   2.30332       0.0997 │
│ train_cifar_ad99a_00003   PENDING        16      1   0.00261551               8                                                    │
│ train_cifar_ad99a_00004   PENDING       256      1   0.000823347              4                                                    │
│ train_cifar_ad99a_00005   PENDING         8     32   0.0264297                4                                                    │
│ train_cifar_ad99a_00006   PENDING       256    128   0.0848744               16                                                    │
│ train_cifar_ad99a_00007   PENDING       128    128   0.0010942                2                                                    │
│ train_cifar_ad99a_00008   PENDING         2      8   0.019285                16                                                    │
│ train_cifar_ad99a_00009   PENDING       128     32   0.00107603               4                                                    │
╰────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────╯
(func pid=8359) [2,  2000] loss: 1.635
(func pid=8359) [2,  4000] loss: 0.814
(func pid=8359) [2,  6000] loss: 0.548
(func pid=8359) [2,  8000] loss: 0.397
(func pid=8359) [2, 10000] loss: 0.320
(func pid=8359) [2, 12000] loss: 0.260
Trial status: 2 TERMINATED | 1 RUNNING | 7 PENDING
Current time: 2026-05-08 22:02:13. Total running time: 13min 31s
Logical resource usage: 2.0/16 CPUs, 1.0/1 GPUs (0.0/1.0 accelerator_type:A10G)
Current best trial: ad99a_00002 with loss=1.6366215757608413 and params={'l1': 256, 'l2': 4, 'lr': 0.0007109136804996362, 'batch_size': 2, 'device': 'cuda'}
╭────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────╮
│ Trial name                status         l1     l2            lr     batch_size     iter     total time (s)      loss     accuracy │
├────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────┤
│ train_cifar_ad99a_00002   RUNNING       256      4   0.000710914              2        1            61.8357   1.63662       0.3915 │
│ train_cifar_ad99a_00000   TERMINATED     32      1   0.0123941                2       10           615.167    2.31109       0.1017 │
│ train_cifar_ad99a_00001   TERMINATED      4      1   0.0101213               16       10            89.6874   2.30332       0.0997 │
│ train_cifar_ad99a_00003   PENDING        16      1   0.00261551               8                                                    │
│ train_cifar_ad99a_00004   PENDING       256      1   0.000823347              4                                                    │
│ train_cifar_ad99a_00005   PENDING         8     32   0.0264297                4                                                    │
│ train_cifar_ad99a_00006   PENDING       256    128   0.0848744               16                                                    │
│ train_cifar_ad99a_00007   PENDING       128    128   0.0010942                2                                                    │
│ train_cifar_ad99a_00008   PENDING         2      8   0.019285                16                                                    │
│ train_cifar_ad99a_00009   PENDING       128     32   0.00107603               4                                                    │
╰────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────╯
(func pid=8359) [2, 14000] loss: 0.221
(func pid=8359) [2, 16000] loss: 0.192
(func pid=8359) [2, 18000] loss: 0.168
(func pid=8359) [2, 20000] loss: 0.151
(func pid=8359) Checkpoint successfully created at: Checkpoint(filesystem=local, path=/var/lib/ci-user/ray_results/train_cifar_2026-05-08_21-48-41/train_cifar_ad99a_00002_2_batch_size=2,l1=256,l2=4,lr=0.0007_2026-05-08_21-48-41/checkpoint_000001)
Trial status: 2 TERMINATED | 1 RUNNING | 7 PENDING
Current time: 2026-05-08 22:02:43. Total running time: 14min 2s
Logical resource usage: 2.0/16 CPUs, 1.0/1 GPUs (0.0/1.0 accelerator_type:A10G)
Current best trial: ad99a_00002 with loss=1.4878081742586569 and params={'l1': 256, 'l2': 4, 'lr': 0.0007109136804996362, 'batch_size': 2, 'device': 'cuda'}
╭────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────╮
│ Trial name                status         l1     l2            lr     batch_size     iter     total time (s)      loss     accuracy │
├────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────┤
│ train_cifar_ad99a_00002   RUNNING       256      4   0.000710914              2        2           122.079    1.48781       0.4673 │
│ train_cifar_ad99a_00000   TERMINATED     32      1   0.0123941                2       10           615.167    2.31109       0.1017 │
│ train_cifar_ad99a_00001   TERMINATED      4      1   0.0101213               16       10            89.6874   2.30332       0.0997 │
│ train_cifar_ad99a_00003   PENDING        16      1   0.00261551               8                                                    │
│ train_cifar_ad99a_00004   PENDING       256      1   0.000823347              4                                                    │
│ train_cifar_ad99a_00005   PENDING         8     32   0.0264297                4                                                    │
│ train_cifar_ad99a_00006   PENDING       256    128   0.0848744               16                                                    │
│ train_cifar_ad99a_00007   PENDING       128    128   0.0010942                2                                                    │
│ train_cifar_ad99a_00008   PENDING         2      8   0.019285                16                                                    │
│ train_cifar_ad99a_00009   PENDING       128     32   0.00107603               4                                                    │
╰────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────╯
(func pid=8359) [3,  2000] loss: 1.489
(func pid=8359) [3,  4000] loss: 0.726
(func pid=8359) [3,  6000] loss: 0.478
(func pid=8359) [3,  8000] loss: 0.362
(func pid=8359) [3, 10000] loss: 0.288
(func pid=8359) [3, 12000] loss: 0.234
Trial status: 2 TERMINATED | 1 RUNNING | 7 PENDING
Current time: 2026-05-08 22:03:13. Total running time: 14min 32s
Logical resource usage: 2.0/16 CPUs, 1.0/1 GPUs (0.0/1.0 accelerator_type:A10G)
Current best trial: ad99a_00002 with loss=1.4878081742586569 and params={'l1': 256, 'l2': 4, 'lr': 0.0007109136804996362, 'batch_size': 2, 'device': 'cuda'}
╭────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────╮
│ Trial name                status         l1     l2            lr     batch_size     iter     total time (s)      loss     accuracy │
├────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────┤
│ train_cifar_ad99a_00002   RUNNING       256      4   0.000710914              2        2           122.079    1.48781       0.4673 │
│ train_cifar_ad99a_00000   TERMINATED     32      1   0.0123941                2       10           615.167    2.31109       0.1017 │
│ train_cifar_ad99a_00001   TERMINATED      4      1   0.0101213               16       10            89.6874   2.30332       0.0997 │
│ train_cifar_ad99a_00003   PENDING        16      1   0.00261551               8                                                    │
│ train_cifar_ad99a_00004   PENDING       256      1   0.000823347              4                                                    │
│ train_cifar_ad99a_00005   PENDING         8     32   0.0264297                4                                                    │
│ train_cifar_ad99a_00006   PENDING       256    128   0.0848744               16                                                    │
│ train_cifar_ad99a_00007   PENDING       128    128   0.0010942                2                                                    │
│ train_cifar_ad99a_00008   PENDING         2      8   0.019285                16                                                    │
│ train_cifar_ad99a_00009   PENDING       128     32   0.00107603               4                                                    │
╰────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────╯
(func pid=8359) [3, 14000] loss: 0.207
(func pid=8359) [3, 16000] loss: 0.176
(func pid=8359) [3, 18000] loss: 0.157
(func pid=8359) [3, 20000] loss: 0.140
(func pid=8359) Checkpoint successfully created at: Checkpoint(filesystem=local, path=/var/lib/ci-user/ray_results/train_cifar_2026-05-08_21-48-41/train_cifar_ad99a_00002_2_batch_size=2,l1=256,l2=4,lr=0.0007_2026-05-08_21-48-41/checkpoint_000002)
Trial status: 2 TERMINATED | 1 RUNNING | 7 PENDING
Current time: 2026-05-08 22:03:43. Total running time: 15min 2s
Logical resource usage: 2.0/16 CPUs, 1.0/1 GPUs (0.0/1.0 accelerator_type:A10G)
Current best trial: ad99a_00002 with loss=1.4577347185370513 and params={'l1': 256, 'l2': 4, 'lr': 0.0007109136804996362, 'batch_size': 2, 'device': 'cuda'}
╭────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────╮
│ Trial name                status         l1     l2            lr     batch_size     iter     total time (s)      loss     accuracy │
├────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────┤
│ train_cifar_ad99a_00002   RUNNING       256      4   0.000710914              2        3           182.542    1.45773       0.5021 │
│ train_cifar_ad99a_00000   TERMINATED     32      1   0.0123941                2       10           615.167    2.31109       0.1017 │
│ train_cifar_ad99a_00001   TERMINATED      4      1   0.0101213               16       10            89.6874   2.30332       0.0997 │
│ train_cifar_ad99a_00003   PENDING        16      1   0.00261551               8                                                    │
│ train_cifar_ad99a_00004   PENDING       256      1   0.000823347              4                                                    │
│ train_cifar_ad99a_00005   PENDING         8     32   0.0264297                4                                                    │
│ train_cifar_ad99a_00006   PENDING       256    128   0.0848744               16                                                    │
│ train_cifar_ad99a_00007   PENDING       128    128   0.0010942                2                                                    │
│ train_cifar_ad99a_00008   PENDING         2      8   0.019285                16                                                    │
│ train_cifar_ad99a_00009   PENDING       128     32   0.00107603               4                                                    │
╰────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────╯
(func pid=8359) [4,  2000] loss: 1.293
(func pid=8359) [4,  4000] loss: 0.666
(func pid=8359) [4,  6000] loss: 0.449
(func pid=8359) [4,  8000] loss: 0.340
(func pid=8359) [4, 10000] loss: 0.274
(func pid=8359) [4, 12000] loss: 0.219
Trial status: 2 TERMINATED | 1 RUNNING | 7 PENDING
Current time: 2026-05-08 22:04:13. Total running time: 15min 32s
Logical resource usage: 2.0/16 CPUs, 1.0/1 GPUs (0.0/1.0 accelerator_type:A10G)
Current best trial: ad99a_00002 with loss=1.4577347185370513 and params={'l1': 256, 'l2': 4, 'lr': 0.0007109136804996362, 'batch_size': 2, 'device': 'cuda'}
╭────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────╮
│ Trial name                status         l1     l2            lr     batch_size     iter     total time (s)      loss     accuracy │
├────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────┤
│ train_cifar_ad99a_00002   RUNNING       256      4   0.000710914              2        3           182.542    1.45773       0.5021 │
│ train_cifar_ad99a_00000   TERMINATED     32      1   0.0123941                2       10           615.167    2.31109       0.1017 │
│ train_cifar_ad99a_00001   TERMINATED      4      1   0.0101213               16       10            89.6874   2.30332       0.0997 │
│ train_cifar_ad99a_00003   PENDING        16      1   0.00261551               8                                                    │
│ train_cifar_ad99a_00004   PENDING       256      1   0.000823347              4                                                    │
│ train_cifar_ad99a_00005   PENDING         8     32   0.0264297                4                                                    │
│ train_cifar_ad99a_00006   PENDING       256    128   0.0848744               16                                                    │
│ train_cifar_ad99a_00007   PENDING       128    128   0.0010942                2                                                    │
│ train_cifar_ad99a_00008   PENDING         2      8   0.019285                16                                                    │
│ train_cifar_ad99a_00009   PENDING       128     32   0.00107603               4                                                    │
╰────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────╯
(func pid=8359) [4, 14000] loss: 0.191
(func pid=8359) [4, 16000] loss: 0.167
(func pid=8359) [4, 18000] loss: 0.152
(func pid=8359) [4, 20000] loss: 0.136
(func pid=8359) Checkpoint successfully created at: Checkpoint(filesystem=local, path=/var/lib/ci-user/ray_results/train_cifar_2026-05-08_21-48-41/train_cifar_ad99a_00002_2_batch_size=2,l1=256,l2=4,lr=0.0007_2026-05-08_21-48-41/checkpoint_000003)
Trial status: 2 TERMINATED | 1 RUNNING | 7 PENDING
Current time: 2026-05-08 22:04:43. Total running time: 16min 2s
Logical resource usage: 2.0/16 CPUs, 1.0/1 GPUs (0.0/1.0 accelerator_type:A10G)
Current best trial: ad99a_00002 with loss=1.3527844864845275 and params={'l1': 256, 'l2': 4, 'lr': 0.0007109136804996362, 'batch_size': 2, 'device': 'cuda'}
╭────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────╮
│ Trial name                status         l1     l2            lr     batch_size     iter     total time (s)      loss     accuracy │
├────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────┤
│ train_cifar_ad99a_00002   RUNNING       256      4   0.000710914              2        4           242.445    1.35278       0.5298 │
│ train_cifar_ad99a_00000   TERMINATED     32      1   0.0123941                2       10           615.167    2.31109       0.1017 │
│ train_cifar_ad99a_00001   TERMINATED      4      1   0.0101213               16       10            89.6874   2.30332       0.0997 │
│ train_cifar_ad99a_00003   PENDING        16      1   0.00261551               8                                                    │
│ train_cifar_ad99a_00004   PENDING       256      1   0.000823347              4                                                    │
│ train_cifar_ad99a_00005   PENDING         8     32   0.0264297                4                                                    │
│ train_cifar_ad99a_00006   PENDING       256    128   0.0848744               16                                                    │
│ train_cifar_ad99a_00007   PENDING       128    128   0.0010942                2                                                    │
│ train_cifar_ad99a_00008   PENDING         2      8   0.019285                16                                                    │
│ train_cifar_ad99a_00009   PENDING       128     32   0.00107603               4                                                    │
╰────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────╯
(func pid=8359) [5,  2000] loss: 1.250
(func pid=8359) [5,  4000] loss: 0.627
(func pid=8359) [5,  6000] loss: 0.415
(func pid=8359) [5,  8000] loss: 0.313
(func pid=8359) [5, 10000] loss: 0.267
(func pid=8359) [5, 12000] loss: 0.212
Trial status: 2 TERMINATED | 1 RUNNING | 7 PENDING
Current time: 2026-05-08 22:05:13. Total running time: 16min 32s
Logical resource usage: 2.0/16 CPUs, 1.0/1 GPUs (0.0/1.0 accelerator_type:A10G)
Current best trial: ad99a_00002 with loss=1.3527844864845275 and params={'l1': 256, 'l2': 4, 'lr': 0.0007109136804996362, 'batch_size': 2, 'device': 'cuda'}
╭────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────╮
│ Trial name                status         l1     l2            lr     batch_size     iter     total time (s)      loss     accuracy │
├────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────┤
│ train_cifar_ad99a_00002   RUNNING       256      4   0.000710914              2        4           242.445    1.35278       0.5298 │
│ train_cifar_ad99a_00000   TERMINATED     32      1   0.0123941                2       10           615.167    2.31109       0.1017 │
│ train_cifar_ad99a_00001   TERMINATED      4      1   0.0101213               16       10            89.6874   2.30332       0.0997 │
│ train_cifar_ad99a_00003   PENDING        16      1   0.00261551               8                                                    │
│ train_cifar_ad99a_00004   PENDING       256      1   0.000823347              4                                                    │
│ train_cifar_ad99a_00005   PENDING         8     32   0.0264297                4                                                    │
│ train_cifar_ad99a_00006   PENDING       256    128   0.0848744               16                                                    │
│ train_cifar_ad99a_00007   PENDING       128    128   0.0010942                2                                                    │
│ train_cifar_ad99a_00008   PENDING         2      8   0.019285                16                                                    │
│ train_cifar_ad99a_00009   PENDING       128     32   0.00107603               4                                                    │
╰────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────╯
(func pid=8359) [5, 14000] loss: 0.187
(func pid=8359) [5, 16000] loss: 0.161
(func pid=8359) [5, 18000] loss: 0.142
(func pid=8359) [5, 20000] loss: 0.129
(func pid=8359) Checkpoint successfully created at: Checkpoint(filesystem=local, path=/var/lib/ci-user/ray_results/train_cifar_2026-05-08_21-48-41/train_cifar_ad99a_00002_2_batch_size=2,l1=256,l2=4,lr=0.0007_2026-05-08_21-48-41/checkpoint_000004)
Trial status: 2 TERMINATED | 1 RUNNING | 7 PENDING
Current time: 2026-05-08 22:05:43. Total running time: 17min 2s
Logical resource usage: 2.0/16 CPUs, 1.0/1 GPUs (0.0/1.0 accelerator_type:A10G)
Current best trial: ad99a_00002 with loss=1.3449752011217178 and params={'l1': 256, 'l2': 4, 'lr': 0.0007109136804996362, 'batch_size': 2, 'device': 'cuda'}
╭────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────╮
│ Trial name                status         l1     l2            lr     batch_size     iter     total time (s)      loss     accuracy │
├────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────┤
│ train_cifar_ad99a_00002   RUNNING       256      4   0.000710914              2        5           302.819    1.34498       0.5464 │
│ train_cifar_ad99a_00000   TERMINATED     32      1   0.0123941                2       10           615.167    2.31109       0.1017 │
│ train_cifar_ad99a_00001   TERMINATED      4      1   0.0101213               16       10            89.6874   2.30332       0.0997 │
│ train_cifar_ad99a_00003   PENDING        16      1   0.00261551               8                                                    │
│ train_cifar_ad99a_00004   PENDING       256      1   0.000823347              4                                                    │
│ train_cifar_ad99a_00005   PENDING         8     32   0.0264297                4                                                    │
│ train_cifar_ad99a_00006   PENDING       256    128   0.0848744               16                                                    │
│ train_cifar_ad99a_00007   PENDING       128    128   0.0010942                2                                                    │
│ train_cifar_ad99a_00008   PENDING         2      8   0.019285                16                                                    │
│ train_cifar_ad99a_00009   PENDING       128     32   0.00107603               4                                                    │
╰────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────╯
(func pid=8359) [6,  2000] loss: 1.161
(func pid=8359) [6,  4000] loss: 0.610
(func pid=8359) [6,  6000] loss: 0.413
(func pid=8359) [6,  8000] loss: 0.308
(func pid=8359) [6, 10000] loss: 0.237
Trial status: 2 TERMINATED | 1 RUNNING | 7 PENDING
Current time: 2026-05-08 22:06:13. Total running time: 17min 32s
Logical resource usage: 2.0/16 CPUs, 1.0/1 GPUs (0.0/1.0 accelerator_type:A10G)
Current best trial: ad99a_00002 with loss=1.3449752011217178 and params={'l1': 256, 'l2': 4, 'lr': 0.0007109136804996362, 'batch_size': 2, 'device': 'cuda'}
╭────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────╮
│ Trial name                status         l1     l2            lr     batch_size     iter     total time (s)      loss     accuracy │
├────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────┤
│ train_cifar_ad99a_00002   RUNNING       256      4   0.000710914              2        5           302.819    1.34498       0.5464 │
│ train_cifar_ad99a_00000   TERMINATED     32      1   0.0123941                2       10           615.167    2.31109       0.1017 │
│ train_cifar_ad99a_00001   TERMINATED      4      1   0.0101213               16       10            89.6874   2.30332       0.0997 │
│ train_cifar_ad99a_00003   PENDING        16      1   0.00261551               8                                                    │
│ train_cifar_ad99a_00004   PENDING       256      1   0.000823347              4                                                    │
│ train_cifar_ad99a_00005   PENDING         8     32   0.0264297                4                                                    │
│ train_cifar_ad99a_00006   PENDING       256    128   0.0848744               16                                                    │
│ train_cifar_ad99a_00007   PENDING       128    128   0.0010942                2                                                    │
│ train_cifar_ad99a_00008   PENDING         2      8   0.019285                16                                                    │
│ train_cifar_ad99a_00009   PENDING       128     32   0.00107603               4                                                    │
╰────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────╯
(func pid=8359) [6, 12000] loss: 0.201
(func pid=8359) [6, 14000] loss: 0.175
(func pid=8359) [6, 16000] loss: 0.156
(func pid=8359) [6, 18000] loss: 0.136
(func pid=8359) [6, 20000] loss: 0.126
(func pid=8359) Checkpoint successfully created at: Checkpoint(filesystem=local, path=/var/lib/ci-user/ray_results/train_cifar_2026-05-08_21-48-41/train_cifar_ad99a_00002_2_batch_size=2,l1=256,l2=4,lr=0.0007_2026-05-08_21-48-41/checkpoint_000005)
Trial status: 2 TERMINATED | 1 RUNNING | 7 PENDING
Current time: 2026-05-08 22:06:43. Total running time: 18min 2s
Logical resource usage: 2.0/16 CPUs, 1.0/1 GPUs (0.0/1.0 accelerator_type:A10G)
Current best trial: ad99a_00002 with loss=1.3899078472164692 and params={'l1': 256, 'l2': 4, 'lr': 0.0007109136804996362, 'batch_size': 2, 'device': 'cuda'}
╭────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────╮
│ Trial name                status         l1     l2            lr     batch_size     iter     total time (s)      loss     accuracy │
├────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────┤
│ train_cifar_ad99a_00002   RUNNING       256      4   0.000710914              2        6           363.304    1.38991       0.5345 │
│ train_cifar_ad99a_00000   TERMINATED     32      1   0.0123941                2       10           615.167    2.31109       0.1017 │
│ train_cifar_ad99a_00001   TERMINATED      4      1   0.0101213               16       10            89.6874   2.30332       0.0997 │
│ train_cifar_ad99a_00003   PENDING        16      1   0.00261551               8                                                    │
│ train_cifar_ad99a_00004   PENDING       256      1   0.000823347              4                                                    │
│ train_cifar_ad99a_00005   PENDING         8     32   0.0264297                4                                                    │
│ train_cifar_ad99a_00006   PENDING       256    128   0.0848744               16                                                    │
│ train_cifar_ad99a_00007   PENDING       128    128   0.0010942                2                                                    │
│ train_cifar_ad99a_00008   PENDING         2      8   0.019285                16                                                    │
│ train_cifar_ad99a_00009   PENDING       128     32   0.00107603               4                                                    │
╰────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────╯
(func pid=8359) [7,  2000] loss: 1.133
(func pid=8359) [7,  4000] loss: 0.566
(func pid=8359) [7,  6000] loss: 0.380
(func pid=8359) [7,  8000] loss: 0.291
(func pid=8359) [7, 10000] loss: 0.235
Trial status: 2 TERMINATED | 1 RUNNING | 7 PENDING
Current time: 2026-05-08 22:07:13. Total running time: 18min 32s
Logical resource usage: 2.0/16 CPUs, 1.0/1 GPUs (0.0/1.0 accelerator_type:A10G)
Current best trial: ad99a_00002 with loss=1.3899078472164692 and params={'l1': 256, 'l2': 4, 'lr': 0.0007109136804996362, 'batch_size': 2, 'device': 'cuda'}
╭────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────╮
│ Trial name                status         l1     l2            lr     batch_size     iter     total time (s)      loss     accuracy │
├────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────┤
│ train_cifar_ad99a_00002   RUNNING       256      4   0.000710914              2        6           363.304    1.38991       0.5345 │
│ train_cifar_ad99a_00000   TERMINATED     32      1   0.0123941                2       10           615.167    2.31109       0.1017 │
│ train_cifar_ad99a_00001   TERMINATED      4      1   0.0101213               16       10            89.6874   2.30332       0.0997 │
│ train_cifar_ad99a_00003   PENDING        16      1   0.00261551               8                                                    │
│ train_cifar_ad99a_00004   PENDING       256      1   0.000823347              4                                                    │
│ train_cifar_ad99a_00005   PENDING         8     32   0.0264297                4                                                    │
│ train_cifar_ad99a_00006   PENDING       256    128   0.0848744               16                                                    │
│ train_cifar_ad99a_00007   PENDING       128    128   0.0010942                2                                                    │
│ train_cifar_ad99a_00008   PENDING         2      8   0.019285                16                                                    │
│ train_cifar_ad99a_00009   PENDING       128     32   0.00107603               4                                                    │
╰────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────╯
(func pid=8359) [7, 12000] loss: 0.199
(func pid=8359) [7, 14000] loss: 0.169
(func pid=8359) [7, 16000] loss: 0.150
(func pid=8359) [7, 18000] loss: 0.135
(func pid=8359) [7, 20000] loss: 0.120
(func pid=8359) Checkpoint successfully created at: Checkpoint(filesystem=local, path=/var/lib/ci-user/ray_results/train_cifar_2026-05-08_21-48-41/train_cifar_ad99a_00002_2_batch_size=2,l1=256,l2=4,lr=0.0007_2026-05-08_21-48-41/checkpoint_000006)
Trial status: 2 TERMINATED | 1 RUNNING | 7 PENDING
Current time: 2026-05-08 22:07:44. Total running time: 19min 2s
Logical resource usage: 2.0/16 CPUs, 1.0/1 GPUs (0.0/1.0 accelerator_type:A10G)
Current best trial: ad99a_00002 with loss=1.3594130817674799 and params={'l1': 256, 'l2': 4, 'lr': 0.0007109136804996362, 'batch_size': 2, 'device': 'cuda'}
╭────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────╮
│ Trial name                status         l1     l2            lr     batch_size     iter     total time (s)      loss     accuracy │
├────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────┤
│ train_cifar_ad99a_00002   RUNNING       256      4   0.000710914              2        7           423.451    1.35941       0.54   │
│ train_cifar_ad99a_00000   TERMINATED     32      1   0.0123941                2       10           615.167    2.31109       0.1017 │
│ train_cifar_ad99a_00001   TERMINATED      4      1   0.0101213               16       10            89.6874   2.30332       0.0997 │
│ train_cifar_ad99a_00003   PENDING        16      1   0.00261551               8                                                    │
│ train_cifar_ad99a_00004   PENDING       256      1   0.000823347              4                                                    │
│ train_cifar_ad99a_00005   PENDING         8     32   0.0264297                4                                                    │
│ train_cifar_ad99a_00006   PENDING       256    128   0.0848744               16                                                    │
│ train_cifar_ad99a_00007   PENDING       128    128   0.0010942                2                                                    │
│ train_cifar_ad99a_00008   PENDING         2      8   0.019285                16                                                    │
│ train_cifar_ad99a_00009   PENDING       128     32   0.00107603               4                                                    │
╰────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────╯
(func pid=8359) [8,  2000] loss: 1.067
(func pid=8359) [8,  4000] loss: 0.550
(func pid=8359) [8,  6000] loss: 0.377
(func pid=8359) [8,  8000] loss: 0.278
(func pid=8359) [8, 10000] loss: 0.231
Trial status: 2 TERMINATED | 1 RUNNING | 7 PENDING
Current time: 2026-05-08 22:08:14. Total running time: 19min 32s
Logical resource usage: 2.0/16 CPUs, 1.0/1 GPUs (0.0/1.0 accelerator_type:A10G)
Current best trial: ad99a_00002 with loss=1.3594130817674799 and params={'l1': 256, 'l2': 4, 'lr': 0.0007109136804996362, 'batch_size': 2, 'device': 'cuda'}
╭────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────╮
│ Trial name                status         l1     l2            lr     batch_size     iter     total time (s)      loss     accuracy │
├────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────┤
│ train_cifar_ad99a_00002   RUNNING       256      4   0.000710914              2        7           423.451    1.35941       0.54   │
│ train_cifar_ad99a_00000   TERMINATED     32      1   0.0123941                2       10           615.167    2.31109       0.1017 │
│ train_cifar_ad99a_00001   TERMINATED      4      1   0.0101213               16       10            89.6874   2.30332       0.0997 │
│ train_cifar_ad99a_00003   PENDING        16      1   0.00261551               8                                                    │
│ train_cifar_ad99a_00004   PENDING       256      1   0.000823347              4                                                    │
│ train_cifar_ad99a_00005   PENDING         8     32   0.0264297                4                                                    │
│ train_cifar_ad99a_00006   PENDING       256    128   0.0848744               16                                                    │
│ train_cifar_ad99a_00007   PENDING       128    128   0.0010942                2                                                    │
│ train_cifar_ad99a_00008   PENDING         2      8   0.019285                16                                                    │
│ train_cifar_ad99a_00009   PENDING       128     32   0.00107603               4                                                    │
╰────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────╯
(func pid=8359) [8, 12000] loss: 0.188
(func pid=8359) [8, 14000] loss: 0.167
(func pid=8359) [8, 16000] loss: 0.141
(func pid=8359) [8, 18000] loss: 0.129
(func pid=8359) [8, 20000] loss: 0.116
(func pid=8359) Checkpoint successfully created at: Checkpoint(filesystem=local, path=/var/lib/ci-user/ray_results/train_cifar_2026-05-08_21-48-41/train_cifar_ad99a_00002_2_batch_size=2,l1=256,l2=4,lr=0.0007_2026-05-08_21-48-41/checkpoint_000007)
Trial status: 2 TERMINATED | 1 RUNNING | 7 PENDING
Current time: 2026-05-08 22:08:44. Total running time: 20min 2s
Logical resource usage: 2.0/16 CPUs, 1.0/1 GPUs (0.0/1.0 accelerator_type:A10G)
Current best trial: ad99a_00002 with loss=1.326467015534977 and params={'l1': 256, 'l2': 4, 'lr': 0.0007109136804996362, 'batch_size': 2, 'device': 'cuda'}
╭────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────╮
│ Trial name                status         l1     l2            lr     batch_size     iter     total time (s)      loss     accuracy │
├────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────┤
│ train_cifar_ad99a_00002   RUNNING       256      4   0.000710914              2        8           483.838    1.32647       0.566  │
│ train_cifar_ad99a_00000   TERMINATED     32      1   0.0123941                2       10           615.167    2.31109       0.1017 │
│ train_cifar_ad99a_00001   TERMINATED      4      1   0.0101213               16       10            89.6874   2.30332       0.0997 │
│ train_cifar_ad99a_00003   PENDING        16      1   0.00261551               8                                                    │
│ train_cifar_ad99a_00004   PENDING       256      1   0.000823347              4                                                    │
│ train_cifar_ad99a_00005   PENDING         8     32   0.0264297                4                                                    │
│ train_cifar_ad99a_00006   PENDING       256    128   0.0848744               16                                                    │
│ train_cifar_ad99a_00007   PENDING       128    128   0.0010942                2                                                    │
│ train_cifar_ad99a_00008   PENDING         2      8   0.019285                16                                                    │
│ train_cifar_ad99a_00009   PENDING       128     32   0.00107603               4                                                    │
╰────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────╯
(func pid=8359) [9,  2000] loss: 1.015
(func pid=8359) [9,  4000] loss: 0.527
(func pid=8359) [9,  6000] loss: 0.355
(func pid=8359) [9,  8000] loss: 0.267
(func pid=8359) [9, 10000] loss: 0.221
Trial status: 2 TERMINATED | 1 RUNNING | 7 PENDING
Current time: 2026-05-08 22:09:14. Total running time: 20min 32s
Logical resource usage: 2.0/16 CPUs, 1.0/1 GPUs (0.0/1.0 accelerator_type:A10G)
Current best trial: ad99a_00002 with loss=1.326467015534977 and params={'l1': 256, 'l2': 4, 'lr': 0.0007109136804996362, 'batch_size': 2, 'device': 'cuda'}
╭────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────╮
│ Trial name                status         l1     l2            lr     batch_size     iter     total time (s)      loss     accuracy │
├────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────┤
│ train_cifar_ad99a_00002   RUNNING       256      4   0.000710914              2        8           483.838    1.32647       0.566  │
│ train_cifar_ad99a_00000   TERMINATED     32      1   0.0123941                2       10           615.167    2.31109       0.1017 │
│ train_cifar_ad99a_00001   TERMINATED      4      1   0.0101213               16       10            89.6874   2.30332       0.0997 │
│ train_cifar_ad99a_00003   PENDING        16      1   0.00261551               8                                                    │
│ train_cifar_ad99a_00004   PENDING       256      1   0.000823347              4                                                    │
│ train_cifar_ad99a_00005   PENDING         8     32   0.0264297                4                                                    │
│ train_cifar_ad99a_00006   PENDING       256    128   0.0848744               16                                                    │
│ train_cifar_ad99a_00007   PENDING       128    128   0.0010942                2                                                    │
│ train_cifar_ad99a_00008   PENDING         2      8   0.019285                16                                                    │
│ train_cifar_ad99a_00009   PENDING       128     32   0.00107603               4                                                    │
╰────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────╯
(func pid=8359) [9, 12000] loss: 0.183
(func pid=8359) [9, 14000] loss: 0.163
(func pid=8359) [9, 16000] loss: 0.135
(func pid=8359) [9, 18000] loss: 0.125
(func pid=8359) [9, 20000] loss: 0.116
(func pid=8359) Checkpoint successfully created at: Checkpoint(filesystem=local, path=/var/lib/ci-user/ray_results/train_cifar_2026-05-08_21-48-41/train_cifar_ad99a_00002_2_batch_size=2,l1=256,l2=4,lr=0.0007_2026-05-08_21-48-41/checkpoint_000008)
Trial status: 2 TERMINATED | 1 RUNNING | 7 PENDING
Current time: 2026-05-08 22:09:44. Total running time: 21min 2s
Logical resource usage: 2.0/16 CPUs, 1.0/1 GPUs (0.0/1.0 accelerator_type:A10G)
Current best trial: ad99a_00002 with loss=1.402073765061196 and params={'l1': 256, 'l2': 4, 'lr': 0.0007109136804996362, 'batch_size': 2, 'device': 'cuda'}
╭────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────╮
│ Trial name                status         l1     l2            lr     batch_size     iter     total time (s)      loss     accuracy │
├────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────┤
│ train_cifar_ad99a_00002   RUNNING       256      4   0.000710914              2        9           543.771    1.40207       0.5574 │
│ train_cifar_ad99a_00000   TERMINATED     32      1   0.0123941                2       10           615.167    2.31109       0.1017 │
│ train_cifar_ad99a_00001   TERMINATED      4      1   0.0101213               16       10            89.6874   2.30332       0.0997 │
│ train_cifar_ad99a_00003   PENDING        16      1   0.00261551               8                                                    │
│ train_cifar_ad99a_00004   PENDING       256      1   0.000823347              4                                                    │
│ train_cifar_ad99a_00005   PENDING         8     32   0.0264297                4                                                    │
│ train_cifar_ad99a_00006   PENDING       256    128   0.0848744               16                                                    │
│ train_cifar_ad99a_00007   PENDING       128    128   0.0010942                2                                                    │
│ train_cifar_ad99a_00008   PENDING         2      8   0.019285                16                                                    │
│ train_cifar_ad99a_00009   PENDING       128     32   0.00107603               4                                                    │
╰────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────╯
(func pid=8359) [10,  2000] loss: 0.962
(func pid=8359) [10,  4000] loss: 0.520
(func pid=8359) [10,  6000] loss: 0.353
(func pid=8359) [10,  8000] loss: 0.257
(func pid=8359) [10, 10000] loss: 0.210
Trial status: 2 TERMINATED | 1 RUNNING | 7 PENDING
Current time: 2026-05-08 22:10:14. Total running time: 21min 32s
Logical resource usage: 2.0/16 CPUs, 1.0/1 GPUs (0.0/1.0 accelerator_type:A10G)
Current best trial: ad99a_00002 with loss=1.402073765061196 and params={'l1': 256, 'l2': 4, 'lr': 0.0007109136804996362, 'batch_size': 2, 'device': 'cuda'}
╭────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────╮
│ Trial name                status         l1     l2            lr     batch_size     iter     total time (s)      loss     accuracy │
├────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────┤
│ train_cifar_ad99a_00002   RUNNING       256      4   0.000710914              2        9           543.771    1.40207       0.5574 │
│ train_cifar_ad99a_00000   TERMINATED     32      1   0.0123941                2       10           615.167    2.31109       0.1017 │
│ train_cifar_ad99a_00001   TERMINATED      4      1   0.0101213               16       10            89.6874   2.30332       0.0997 │
│ train_cifar_ad99a_00003   PENDING        16      1   0.00261551               8                                                    │
│ train_cifar_ad99a_00004   PENDING       256      1   0.000823347              4                                                    │
│ train_cifar_ad99a_00005   PENDING         8     32   0.0264297                4                                                    │
│ train_cifar_ad99a_00006   PENDING       256    128   0.0848744               16                                                    │
│ train_cifar_ad99a_00007   PENDING       128    128   0.0010942                2                                                    │
│ train_cifar_ad99a_00008   PENDING         2      8   0.019285                16                                                    │
│ train_cifar_ad99a_00009   PENDING       128     32   0.00107603               4                                                    │
╰────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────╯
(func pid=8359) [10, 12000] loss: 0.175
(func pid=8359) [10, 14000] loss: 0.155
(func pid=8359) [10, 16000] loss: 0.140
(func pid=8359) [10, 18000] loss: 0.119
(func pid=8359) [10, 20000] loss: 0.111
Trial status: 2 TERMINATED | 1 RUNNING | 7 PENDING
Current time: 2026-05-08 22:10:44. Total running time: 22min 3s
Logical resource usage: 2.0/16 CPUs, 1.0/1 GPUs (0.0/1.0 accelerator_type:A10G)
Current best trial: ad99a_00002 with loss=1.402073765061196 and params={'l1': 256, 'l2': 4, 'lr': 0.0007109136804996362, 'batch_size': 2, 'device': 'cuda'}
╭────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────╮
│ Trial name                status         l1     l2            lr     batch_size     iter     total time (s)      loss     accuracy │
├────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────┤
│ train_cifar_ad99a_00002   RUNNING       256      4   0.000710914              2        9           543.771    1.40207       0.5574 │
│ train_cifar_ad99a_00000   TERMINATED     32      1   0.0123941                2       10           615.167    2.31109       0.1017 │
│ train_cifar_ad99a_00001   TERMINATED      4      1   0.0101213               16       10            89.6874   2.30332       0.0997 │
│ train_cifar_ad99a_00003   PENDING        16      1   0.00261551               8                                                    │
│ train_cifar_ad99a_00004   PENDING       256      1   0.000823347              4                                                    │
│ train_cifar_ad99a_00005   PENDING         8     32   0.0264297                4                                                    │
│ train_cifar_ad99a_00006   PENDING       256    128   0.0848744               16                                                    │
│ train_cifar_ad99a_00007   PENDING       128    128   0.0010942                2                                                    │
│ train_cifar_ad99a_00008   PENDING         2      8   0.019285                16                                                    │
│ train_cifar_ad99a_00009   PENDING       128     32   0.00107603               4                                                    │
╰────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────╯

Trial train_cifar_ad99a_00002 completed after 10 iterations at 2026-05-08 22:10:45. Total running time: 22min 3s
╭────────────────────────────────────────────────────────────╮
│ Trial train_cifar_ad99a_00002 result                       │
├────────────────────────────────────────────────────────────┤
│ checkpoint_dir_name                      checkpoint_000009 │
│ time_this_iter_s                                  61.31797 │
│ time_total_s                                     605.08908 │
│ training_iteration                                      10 │
│ accuracy                                            0.5492 │
│ loss                                               1.38372 │
╰────────────────────────────────────────────────────────────╯
(func pid=8359) Checkpoint successfully created at: Checkpoint(filesystem=local, path=/var/lib/ci-user/ray_results/train_cifar_2026-05-08_21-48-41/train_cifar_ad99a_00002_2_batch_size=2,l1=256,l2=4,lr=0.0007_2026-05-08_21-48-41/checkpoint_000009)

Trial train_cifar_ad99a_00003 started with configuration:
╭──────────────────────────────────────────────────╮
│ Trial train_cifar_ad99a_00003 config             │
├──────────────────────────────────────────────────┤
│ batch_size                                     8 │
│ device                                      cuda │
│ l1                                            16 │
│ l2                                             1 │
│ lr                                       0.00262 │
╰──────────────────────────────────────────────────╯
(func pid=10224) [1,  2000] loss: 2.177
(func pid=10224) [1,  4000] loss: 0.975
(func pid=10224) Checkpoint successfully created at: Checkpoint(filesystem=local, path=/var/lib/ci-user/ray_results/train_cifar_2026-05-08_21-48-41/train_cifar_ad99a_00003_3_batch_size=8,l1=16,l2=1,lr=0.0026_2026-05-08_21-48-41/checkpoint_000000)
(func pid=10224) [2,  2000] loss: 1.896

Trial status: 3 TERMINATED | 1 RUNNING | 6 PENDING
Current time: 2026-05-08 22:11:14. Total running time: 22min 33s
Logical resource usage: 2.0/16 CPUs, 1.0/1 GPUs (0.0/1.0 accelerator_type:A10G)
Current best trial: ad99a_00002 with loss=1.3837229575751349 and params={'l1': 256, 'l2': 4, 'lr': 0.0007109136804996362, 'batch_size': 2, 'device': 'cuda'}
╭────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────╮
│ Trial name                status         l1     l2            lr     batch_size     iter     total time (s)      loss     accuracy │
├────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────┤
│ train_cifar_ad99a_00003   RUNNING        16      1   0.00261551               8        1            18.2564   1.89973       0.2168 │
│ train_cifar_ad99a_00000   TERMINATED     32      1   0.0123941                2       10           615.167    2.31109       0.1017 │
│ train_cifar_ad99a_00001   TERMINATED      4      1   0.0101213               16       10            89.6874   2.30332       0.0997 │
│ train_cifar_ad99a_00002   TERMINATED    256      4   0.000710914              2       10           605.089    1.38372       0.5492 │
│ train_cifar_ad99a_00004   PENDING       256      1   0.000823347              4                                                    │
│ train_cifar_ad99a_00005   PENDING         8     32   0.0264297                4                                                    │
│ train_cifar_ad99a_00006   PENDING       256    128   0.0848744               16                                                    │
│ train_cifar_ad99a_00007   PENDING       128    128   0.0010942                2                                                    │
│ train_cifar_ad99a_00008   PENDING         2      8   0.019285                16                                                    │
│ train_cifar_ad99a_00009   PENDING       128     32   0.00107603               4                                                    │
╰────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────╯
(func pid=10224) [2,  4000] loss: 0.942
(func pid=10224) Checkpoint successfully created at: Checkpoint(filesystem=local, path=/var/lib/ci-user/ray_results/train_cifar_2026-05-08_21-48-41/train_cifar_ad99a_00003_3_batch_size=8,l1=16,l2=1,lr=0.0026_2026-05-08_21-48-41/checkpoint_000001)
(func pid=10224) [3,  2000] loss: 1.854
(func pid=10224) [3,  4000] loss: 0.927
(func pid=10224) Checkpoint successfully created at: Checkpoint(filesystem=local, path=/var/lib/ci-user/ray_results/train_cifar_2026-05-08_21-48-41/train_cifar_ad99a_00003_3_batch_size=8,l1=16,l2=1,lr=0.0026_2026-05-08_21-48-41/checkpoint_000002)
Trial status: 3 TERMINATED | 1 RUNNING | 6 PENDING
Current time: 2026-05-08 22:11:44. Total running time: 23min 3s
Logical resource usage: 2.0/16 CPUs, 1.0/1 GPUs (0.0/1.0 accelerator_type:A10G)
Current best trial: ad99a_00002 with loss=1.3837229575751349 and params={'l1': 256, 'l2': 4, 'lr': 0.0007109136804996362, 'batch_size': 2, 'device': 'cuda'}
╭────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────╮
│ Trial name                status         l1     l2            lr     batch_size     iter     total time (s)      loss     accuracy │
├────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────┤
│ train_cifar_ad99a_00003   RUNNING        16      1   0.00261551               8        3            51.0943   1.84168       0.2262 │
│ train_cifar_ad99a_00000   TERMINATED     32      1   0.0123941                2       10           615.167    2.31109       0.1017 │
│ train_cifar_ad99a_00001   TERMINATED      4      1   0.0101213               16       10            89.6874   2.30332       0.0997 │
│ train_cifar_ad99a_00002   TERMINATED    256      4   0.000710914              2       10           605.089    1.38372       0.5492 │
│ train_cifar_ad99a_00004   PENDING       256      1   0.000823347              4                                                    │
│ train_cifar_ad99a_00005   PENDING         8     32   0.0264297                4                                                    │
│ train_cifar_ad99a_00006   PENDING       256    128   0.0848744               16                                                    │
│ train_cifar_ad99a_00007   PENDING       128    128   0.0010942                2                                                    │
│ train_cifar_ad99a_00008   PENDING         2      8   0.019285                16                                                    │
│ train_cifar_ad99a_00009   PENDING       128     32   0.00107603               4                                                    │
╰────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────╯
(func pid=10224) [4,  2000] loss: 1.831
(func pid=10224) [4,  4000] loss: 0.918
(func pid=10224) Checkpoint successfully created at: Checkpoint(filesystem=local, path=/var/lib/ci-user/ray_results/train_cifar_2026-05-08_21-48-41/train_cifar_ad99a_00003_3_batch_size=8,l1=16,l2=1,lr=0.0026_2026-05-08_21-48-41/checkpoint_000003)
(func pid=10224) [5,  2000] loss: 1.819
(func pid=10224) [5,  4000] loss: 0.902
(func pid=10224) Checkpoint successfully created at: Checkpoint(filesystem=local, path=/var/lib/ci-user/ray_results/train_cifar_2026-05-08_21-48-41/train_cifar_ad99a_00003_3_batch_size=8,l1=16,l2=1,lr=0.0026_2026-05-08_21-48-41/checkpoint_000004)
Trial status: 3 TERMINATED | 1 RUNNING | 6 PENDING
Current time: 2026-05-08 22:12:14. Total running time: 23min 33s
Logical resource usage: 2.0/16 CPUs, 1.0/1 GPUs (0.0/1.0 accelerator_type:A10G)
Current best trial: ad99a_00002 with loss=1.3837229575751349 and params={'l1': 256, 'l2': 4, 'lr': 0.0007109136804996362, 'batch_size': 2, 'device': 'cuda'}
╭────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────╮
│ Trial name                status         l1     l2            lr     batch_size     iter     total time (s)      loss     accuracy │
├────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────┤
│ train_cifar_ad99a_00003   RUNNING        16      1   0.00261551               8        5            83.8579   1.81017       0.2571 │
│ train_cifar_ad99a_00000   TERMINATED     32      1   0.0123941                2       10           615.167    2.31109       0.1017 │
│ train_cifar_ad99a_00001   TERMINATED      4      1   0.0101213               16       10            89.6874   2.30332       0.0997 │
│ train_cifar_ad99a_00002   TERMINATED    256      4   0.000710914              2       10           605.089    1.38372       0.5492 │
│ train_cifar_ad99a_00004   PENDING       256      1   0.000823347              4                                                    │
│ train_cifar_ad99a_00005   PENDING         8     32   0.0264297                4                                                    │
│ train_cifar_ad99a_00006   PENDING       256    128   0.0848744               16                                                    │
│ train_cifar_ad99a_00007   PENDING       128    128   0.0010942                2                                                    │
│ train_cifar_ad99a_00008   PENDING         2      8   0.019285                16                                                    │
│ train_cifar_ad99a_00009   PENDING       128     32   0.00107603               4                                                    │
╰────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────╯
(func pid=10224) [6,  2000] loss: 1.786
(func pid=10224) [6,  4000] loss: 0.900
(func pid=10224) Checkpoint successfully created at: Checkpoint(filesystem=local, path=/var/lib/ci-user/ray_results/train_cifar_2026-05-08_21-48-41/train_cifar_ad99a_00003_3_batch_size=8,l1=16,l2=1,lr=0.0026_2026-05-08_21-48-41/checkpoint_000005)
(func pid=10224) [7,  2000] loss: 1.780
(func pid=10224) [7,  4000] loss: 0.894
Trial status: 3 TERMINATED | 1 RUNNING | 6 PENDING
Current time: 2026-05-08 22:12:44. Total running time: 24min 3s
Logical resource usage: 2.0/16 CPUs, 1.0/1 GPUs (0.0/1.0 accelerator_type:A10G)
Current best trial: ad99a_00002 with loss=1.3837229575751349 and params={'l1': 256, 'l2': 4, 'lr': 0.0007109136804996362, 'batch_size': 2, 'device': 'cuda'}
╭────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────╮
│ Trial name                status         l1     l2            lr     batch_size     iter     total time (s)      loss     accuracy │
├────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────┤
│ train_cifar_ad99a_00003   RUNNING        16      1   0.00261551               8        6           100.111    1.84123       0.2559 │
│ train_cifar_ad99a_00000   TERMINATED     32      1   0.0123941                2       10           615.167    2.31109       0.1017 │
│ train_cifar_ad99a_00001   TERMINATED      4      1   0.0101213               16       10            89.6874   2.30332       0.0997 │
│ train_cifar_ad99a_00002   TERMINATED    256      4   0.000710914              2       10           605.089    1.38372       0.5492 │
│ train_cifar_ad99a_00004   PENDING       256      1   0.000823347              4                                                    │
│ train_cifar_ad99a_00005   PENDING         8     32   0.0264297                4                                                    │
│ train_cifar_ad99a_00006   PENDING       256    128   0.0848744               16                                                    │
│ train_cifar_ad99a_00007   PENDING       128    128   0.0010942                2                                                    │
│ train_cifar_ad99a_00008   PENDING         2      8   0.019285                16                                                    │
│ train_cifar_ad99a_00009   PENDING       128     32   0.00107603               4                                                    │
╰────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────╯
(func pid=10224) Checkpoint successfully created at: Checkpoint(filesystem=local, path=/var/lib/ci-user/ray_results/train_cifar_2026-05-08_21-48-41/train_cifar_ad99a_00003_3_batch_size=8,l1=16,l2=1,lr=0.0026_2026-05-08_21-48-41/checkpoint_000006)
(func pid=10224) [8,  2000] loss: 1.778
(func pid=10224) [8,  4000] loss: 0.883
(func pid=10224) Checkpoint successfully created at: Checkpoint(filesystem=local, path=/var/lib/ci-user/ray_results/train_cifar_2026-05-08_21-48-41/train_cifar_ad99a_00003_3_batch_size=8,l1=16,l2=1,lr=0.0026_2026-05-08_21-48-41/checkpoint_000007)
(func pid=10224) [9,  2000] loss: 1.768
(func pid=10224) [9,  4000] loss: 0.885
Trial status: 3 TERMINATED | 1 RUNNING | 6 PENDING
Current time: 2026-05-08 22:13:14. Total running time: 24min 33s
Logical resource usage: 2.0/16 CPUs, 1.0/1 GPUs (0.0/1.0 accelerator_type:A10G)
Current best trial: ad99a_00002 with loss=1.3837229575751349 and params={'l1': 256, 'l2': 4, 'lr': 0.0007109136804996362, 'batch_size': 2, 'device': 'cuda'}
╭────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────╮
│ Trial name                status         l1     l2            lr     batch_size     iter     total time (s)      loss     accuracy │
├────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────┤
│ train_cifar_ad99a_00003   RUNNING        16      1   0.00261551               8        8           132.914    1.77903       0.2759 │
│ train_cifar_ad99a_00000   TERMINATED     32      1   0.0123941                2       10           615.167    2.31109       0.1017 │
│ train_cifar_ad99a_00001   TERMINATED      4      1   0.0101213               16       10            89.6874   2.30332       0.0997 │
│ train_cifar_ad99a_00002   TERMINATED    256      4   0.000710914              2       10           605.089    1.38372       0.5492 │
│ train_cifar_ad99a_00004   PENDING       256      1   0.000823347              4                                                    │
│ train_cifar_ad99a_00005   PENDING         8     32   0.0264297                4                                                    │
│ train_cifar_ad99a_00006   PENDING       256    128   0.0848744               16                                                    │
│ train_cifar_ad99a_00007   PENDING       128    128   0.0010942                2                                                    │
│ train_cifar_ad99a_00008   PENDING         2      8   0.019285                16                                                    │
│ train_cifar_ad99a_00009   PENDING       128     32   0.00107603               4                                                    │
╰────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────╯
(func pid=10224) Checkpoint successfully created at: Checkpoint(filesystem=local, path=/var/lib/ci-user/ray_results/train_cifar_2026-05-08_21-48-41/train_cifar_ad99a_00003_3_batch_size=8,l1=16,l2=1,lr=0.0026_2026-05-08_21-48-41/checkpoint_000008)
(func pid=10224) [10,  2000] loss: 1.753
(func pid=10224) [10,  4000] loss: 0.879

Trial train_cifar_ad99a_00003 completed after 10 iterations at 2026-05-08 22:13:35. Total running time: 24min 54s
╭────────────────────────────────────────────────────────────╮
│ Trial train_cifar_ad99a_00003 result                       │
├────────────────────────────────────────────────────────────┤
│ checkpoint_dir_name                      checkpoint_000009 │
│ time_this_iter_s                                  16.63904 │
│ time_total_s                                     165.84065 │
│ training_iteration                                      10 │
│ accuracy                                             0.262 │
│ loss                                               1.77842 │
╰────────────────────────────────────────────────────────────╯
(func pid=10224) Checkpoint successfully created at: Checkpoint(filesystem=local, path=/var/lib/ci-user/ray_results/train_cifar_2026-05-08_21-48-41/train_cifar_ad99a_00003_3_batch_size=8,l1=16,l2=1,lr=0.0026_2026-05-08_21-48-41/checkpoint_000009)

Trial train_cifar_ad99a_00004 started with configuration:
╭──────────────────────────────────────────────────╮
│ Trial train_cifar_ad99a_00004 config             │
├──────────────────────────────────────────────────┤
│ batch_size                                     4 │
│ device                                      cuda │
│ l1                                           256 │
│ l2                                             1 │
│ lr                                       0.00082 │
╰──────────────────────────────────────────────────╯

Trial status: 4 TERMINATED | 1 RUNNING | 5 PENDING
Current time: 2026-05-08 22:13:44. Total running time: 25min 3s
Logical resource usage: 2.0/16 CPUs, 1.0/1 GPUs (0.0/1.0 accelerator_type:A10G)
Current best trial: ad99a_00002 with loss=1.3837229575751349 and params={'l1': 256, 'l2': 4, 'lr': 0.0007109136804996362, 'batch_size': 2, 'device': 'cuda'}
╭────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────╮
│ Trial name                status         l1     l2            lr     batch_size     iter     total time (s)      loss     accuracy │
├────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────┤
│ train_cifar_ad99a_00004   RUNNING       256      1   0.000823347              4                                                    │
│ train_cifar_ad99a_00000   TERMINATED     32      1   0.0123941                2       10           615.167    2.31109       0.1017 │
│ train_cifar_ad99a_00001   TERMINATED      4      1   0.0101213               16       10            89.6874   2.30332       0.0997 │
│ train_cifar_ad99a_00002   TERMINATED    256      4   0.000710914              2       10           605.089    1.38372       0.5492 │
│ train_cifar_ad99a_00003   TERMINATED     16      1   0.00261551               8       10           165.841    1.77842       0.262  │
│ train_cifar_ad99a_00005   PENDING         8     32   0.0264297                4                                                    │
│ train_cifar_ad99a_00006   PENDING       256    128   0.0848744               16                                                    │
│ train_cifar_ad99a_00007   PENDING       128    128   0.0010942                2                                                    │
│ train_cifar_ad99a_00008   PENDING         2      8   0.019285                16                                                    │
│ train_cifar_ad99a_00009   PENDING       128     32   0.00107603               4                                                    │
╰────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────╯
(func pid=11253) [1,  2000] loss: 2.272
(func pid=11253) [1,  4000] loss: 1.078
(func pid=11253) [1,  6000] loss: 0.683
(func pid=11253) [1,  8000] loss: 0.496
(func pid=11253) [1, 10000] loss: 0.389
(func pid=11253) Checkpoint successfully created at: Checkpoint(filesystem=local, path=/var/lib/ci-user/ray_results/train_cifar_2026-05-08_21-48-41/train_cifar_ad99a_00004_4_batch_size=4,l1=256,l2=1,lr=0.0008_2026-05-08_21-48-41/checkpoint_000000)
Trial status: 4 TERMINATED | 1 RUNNING | 5 PENDING
Current time: 2026-05-08 22:14:14. Total running time: 25min 33s
Logical resource usage: 2.0/16 CPUs, 1.0/1 GPUs (0.0/1.0 accelerator_type:A10G)
Current best trial: ad99a_00002 with loss=1.3837229575751349 and params={'l1': 256, 'l2': 4, 'lr': 0.0007109136804996362, 'batch_size': 2, 'device': 'cuda'}
╭────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────╮
│ Trial name                status         l1     l2            lr     batch_size     iter     total time (s)      loss     accuracy │
├────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────┤
│ train_cifar_ad99a_00004   RUNNING       256      1   0.000823347              4        1            34.0985   1.92921       0.1928 │
│ train_cifar_ad99a_00000   TERMINATED     32      1   0.0123941                2       10           615.167    2.31109       0.1017 │
│ train_cifar_ad99a_00001   TERMINATED      4      1   0.0101213               16       10            89.6874   2.30332       0.0997 │
│ train_cifar_ad99a_00002   TERMINATED    256      4   0.000710914              2       10           605.089    1.38372       0.5492 │
│ train_cifar_ad99a_00003   TERMINATED     16      1   0.00261551               8       10           165.841    1.77842       0.262  │
│ train_cifar_ad99a_00005   PENDING         8     32   0.0264297                4                                                    │
│ train_cifar_ad99a_00006   PENDING       256    128   0.0848744               16                                                    │
│ train_cifar_ad99a_00007   PENDING       128    128   0.0010942                2                                                    │
│ train_cifar_ad99a_00008   PENDING         2      8   0.019285                16                                                    │
│ train_cifar_ad99a_00009   PENDING       128     32   0.00107603               4                                                    │
╰────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────╯
(func pid=11253) [2,  2000] loss: 1.915
(func pid=11253) [2,  4000] loss: 0.953
(func pid=11253) [2,  6000] loss: 0.626
(func pid=11253) [2,  8000] loss: 0.466
(func pid=11253) [2, 10000] loss: 0.372
Trial status: 4 TERMINATED | 1 RUNNING | 5 PENDING
Current time: 2026-05-08 22:14:44. Total running time: 26min 3s
Logical resource usage: 2.0/16 CPUs, 1.0/1 GPUs (0.0/1.0 accelerator_type:A10G)
Current best trial: ad99a_00002 with loss=1.3837229575751349 and params={'l1': 256, 'l2': 4, 'lr': 0.0007109136804996362, 'batch_size': 2, 'device': 'cuda'}
╭────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────╮
│ Trial name                status         l1     l2            lr     batch_size     iter     total time (s)      loss     accuracy │
├────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────┤
│ train_cifar_ad99a_00004   RUNNING       256      1   0.000823347              4        1            34.0985   1.92921       0.1928 │
│ train_cifar_ad99a_00000   TERMINATED     32      1   0.0123941                2       10           615.167    2.31109       0.1017 │
│ train_cifar_ad99a_00001   TERMINATED      4      1   0.0101213               16       10            89.6874   2.30332       0.0997 │
│ train_cifar_ad99a_00002   TERMINATED    256      4   0.000710914              2       10           605.089    1.38372       0.5492 │
│ train_cifar_ad99a_00003   TERMINATED     16      1   0.00261551               8       10           165.841    1.77842       0.262  │
│ train_cifar_ad99a_00005   PENDING         8     32   0.0264297                4                                                    │
│ train_cifar_ad99a_00006   PENDING       256    128   0.0848744               16                                                    │
│ train_cifar_ad99a_00007   PENDING       128    128   0.0010942                2                                                    │
│ train_cifar_ad99a_00008   PENDING         2      8   0.019285                16                                                    │
│ train_cifar_ad99a_00009   PENDING       128     32   0.00107603               4                                                    │
╰────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────╯
(func pid=11253) Checkpoint successfully created at: Checkpoint(filesystem=local, path=/var/lib/ci-user/ray_results/train_cifar_2026-05-08_21-48-41/train_cifar_ad99a_00004_4_batch_size=4,l1=256,l2=1,lr=0.0008_2026-05-08_21-48-41/checkpoint_000001)
(func pid=11253) [3,  2000] loss: 1.838
(func pid=11253) [3,  4000] loss: 0.911
(func pid=11253) [3,  6000] loss: 0.609
(func pid=11253) [3,  8000] loss: 0.457
(func pid=11253) [3, 10000] loss: 0.363
Trial status: 4 TERMINATED | 1 RUNNING | 5 PENDING
Current time: 2026-05-08 22:15:14. Total running time: 26min 33s
Logical resource usage: 2.0/16 CPUs, 1.0/1 GPUs (0.0/1.0 accelerator_type:A10G)
Current best trial: ad99a_00002 with loss=1.3837229575751349 and params={'l1': 256, 'l2': 4, 'lr': 0.0007109136804996362, 'batch_size': 2, 'device': 'cuda'}
╭────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────╮
│ Trial name                status         l1     l2            lr     batch_size     iter     total time (s)      loss     accuracy │
├────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────┤
│ train_cifar_ad99a_00004   RUNNING       256      1   0.000823347              4        2            66.0751   1.84362       0.2373 │
│ train_cifar_ad99a_00000   TERMINATED     32      1   0.0123941                2       10           615.167    2.31109       0.1017 │
│ train_cifar_ad99a_00001   TERMINATED      4      1   0.0101213               16       10            89.6874   2.30332       0.0997 │
│ train_cifar_ad99a_00002   TERMINATED    256      4   0.000710914              2       10           605.089    1.38372       0.5492 │
│ train_cifar_ad99a_00003   TERMINATED     16      1   0.00261551               8       10           165.841    1.77842       0.262  │
│ train_cifar_ad99a_00005   PENDING         8     32   0.0264297                4                                                    │
│ train_cifar_ad99a_00006   PENDING       256    128   0.0848744               16                                                    │
│ train_cifar_ad99a_00007   PENDING       128    128   0.0010942                2                                                    │
│ train_cifar_ad99a_00008   PENDING         2      8   0.019285                16                                                    │
│ train_cifar_ad99a_00009   PENDING       128     32   0.00107603               4                                                    │
╰────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────╯
(func pid=11253) Checkpoint successfully created at: Checkpoint(filesystem=local, path=/var/lib/ci-user/ray_results/train_cifar_2026-05-08_21-48-41/train_cifar_ad99a_00004_4_batch_size=4,l1=256,l2=1,lr=0.0008_2026-05-08_21-48-41/checkpoint_000002)
(func pid=11253) [4,  2000] loss: 1.788
(func pid=11253) [4,  4000] loss: 0.893
(func pid=11253) [4,  6000] loss: 0.598
(func pid=11253) [4,  8000] loss: 0.447
Trial status: 4 TERMINATED | 1 RUNNING | 5 PENDING
Current time: 2026-05-08 22:15:44. Total running time: 27min 3s
Logical resource usage: 2.0/16 CPUs, 1.0/1 GPUs (0.0/1.0 accelerator_type:A10G)
Current best trial: ad99a_00002 with loss=1.3837229575751349 and params={'l1': 256, 'l2': 4, 'lr': 0.0007109136804996362, 'batch_size': 2, 'device': 'cuda'}
╭────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────╮
│ Trial name                status         l1     l2            lr     batch_size     iter     total time (s)      loss     accuracy │
├────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────┤
│ train_cifar_ad99a_00004   RUNNING       256      1   0.000823347              4        3            97.9666   1.808         0.2487 │
│ train_cifar_ad99a_00000   TERMINATED     32      1   0.0123941                2       10           615.167    2.31109       0.1017 │
│ train_cifar_ad99a_00001   TERMINATED      4      1   0.0101213               16       10            89.6874   2.30332       0.0997 │
│ train_cifar_ad99a_00002   TERMINATED    256      4   0.000710914              2       10           605.089    1.38372       0.5492 │
│ train_cifar_ad99a_00003   TERMINATED     16      1   0.00261551               8       10           165.841    1.77842       0.262  │
│ train_cifar_ad99a_00005   PENDING         8     32   0.0264297                4                                                    │
│ train_cifar_ad99a_00006   PENDING       256    128   0.0848744               16                                                    │
│ train_cifar_ad99a_00007   PENDING       128    128   0.0010942                2                                                    │
│ train_cifar_ad99a_00008   PENDING         2      8   0.019285                16                                                    │
│ train_cifar_ad99a_00009   PENDING       128     32   0.00107603               4                                                    │
╰────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────╯
(func pid=11253) [4, 10000] loss: 0.357
(func pid=11253) Checkpoint successfully created at: Checkpoint(filesystem=local, path=/var/lib/ci-user/ray_results/train_cifar_2026-05-08_21-48-41/train_cifar_ad99a_00004_4_batch_size=4,l1=256,l2=1,lr=0.0008_2026-05-08_21-48-41/checkpoint_000003)
(func pid=11253) [5,  2000] loss: 1.749
(func pid=11253) [5,  4000] loss: 0.878
(func pid=11253) [5,  6000] loss: 0.580
(func pid=11253) [5,  8000] loss: 0.436
Trial status: 4 TERMINATED | 1 RUNNING | 5 PENDING
Current time: 2026-05-08 22:16:14. Total running time: 27min 33s
Logical resource usage: 2.0/16 CPUs, 1.0/1 GPUs (0.0/1.0 accelerator_type:A10G)
Current best trial: ad99a_00002 with loss=1.3837229575751349 and params={'l1': 256, 'l2': 4, 'lr': 0.0007109136804996362, 'batch_size': 2, 'device': 'cuda'}
╭────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────╮
│ Trial name                status         l1     l2            lr     batch_size     iter     total time (s)      loss     accuracy │
├────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────┤
│ train_cifar_ad99a_00004   RUNNING       256      1   0.000823347              4        4           130.017    1.78843       0.2788 │
│ train_cifar_ad99a_00000   TERMINATED     32      1   0.0123941                2       10           615.167    2.31109       0.1017 │
│ train_cifar_ad99a_00001   TERMINATED      4      1   0.0101213               16       10            89.6874   2.30332       0.0997 │
│ train_cifar_ad99a_00002   TERMINATED    256      4   0.000710914              2       10           605.089    1.38372       0.5492 │
│ train_cifar_ad99a_00003   TERMINATED     16      1   0.00261551               8       10           165.841    1.77842       0.262  │
│ train_cifar_ad99a_00005   PENDING         8     32   0.0264297                4                                                    │
│ train_cifar_ad99a_00006   PENDING       256    128   0.0848744               16                                                    │
│ train_cifar_ad99a_00007   PENDING       128    128   0.0010942                2                                                    │
│ train_cifar_ad99a_00008   PENDING         2      8   0.019285                16                                                    │
│ train_cifar_ad99a_00009   PENDING       128     32   0.00107603               4                                                    │
╰────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────╯
(func pid=11253) [5, 10000] loss: 0.350
(func pid=11253) Checkpoint successfully created at: Checkpoint(filesystem=local, path=/var/lib/ci-user/ray_results/train_cifar_2026-05-08_21-48-41/train_cifar_ad99a_00004_4_batch_size=4,l1=256,l2=1,lr=0.0008_2026-05-08_21-48-41/checkpoint_000004)
(func pid=11253) [6,  2000] loss: 1.723
(func pid=11253) [6,  4000] loss: 0.859
(func pid=11253) [6,  6000] loss: 0.578
(func pid=11253) [6,  8000] loss: 0.433
Trial status: 4 TERMINATED | 1 RUNNING | 5 PENDING
Current time: 2026-05-08 22:16:45. Total running time: 28min 3s
Logical resource usage: 2.0/16 CPUs, 1.0/1 GPUs (0.0/1.0 accelerator_type:A10G)
Current best trial: ad99a_00002 with loss=1.3837229575751349 and params={'l1': 256, 'l2': 4, 'lr': 0.0007109136804996362, 'batch_size': 2, 'device': 'cuda'}
╭────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────╮
│ Trial name                status         l1     l2            lr     batch_size     iter     total time (s)      loss     accuracy │
├────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────┤
│ train_cifar_ad99a_00004   RUNNING       256      1   0.000823347              4        5           161.897    1.78536       0.2969 │
│ train_cifar_ad99a_00000   TERMINATED     32      1   0.0123941                2       10           615.167    2.31109       0.1017 │
│ train_cifar_ad99a_00001   TERMINATED      4      1   0.0101213               16       10            89.6874   2.30332       0.0997 │
│ train_cifar_ad99a_00002   TERMINATED    256      4   0.000710914              2       10           605.089    1.38372       0.5492 │
│ train_cifar_ad99a_00003   TERMINATED     16      1   0.00261551               8       10           165.841    1.77842       0.262  │
│ train_cifar_ad99a_00005   PENDING         8     32   0.0264297                4                                                    │
│ train_cifar_ad99a_00006   PENDING       256    128   0.0848744               16                                                    │
│ train_cifar_ad99a_00007   PENDING       128    128   0.0010942                2                                                    │
│ train_cifar_ad99a_00008   PENDING         2      8   0.019285                16                                                    │
│ train_cifar_ad99a_00009   PENDING       128     32   0.00107603               4                                                    │
╰────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────╯
(func pid=11253) [6, 10000] loss: 0.345
(func pid=11253) Checkpoint successfully created at: Checkpoint(filesystem=local, path=/var/lib/ci-user/ray_results/train_cifar_2026-05-08_21-48-41/train_cifar_ad99a_00004_4_batch_size=4,l1=256,l2=1,lr=0.0008_2026-05-08_21-48-41/checkpoint_000005)
(func pid=11253) [7,  2000] loss: 1.693
(func pid=11253) [7,  4000] loss: 0.846
(func pid=11253) [7,  6000] loss: 0.565
Trial status: 4 TERMINATED | 1 RUNNING | 5 PENDING
Current time: 2026-05-08 22:17:15. Total running time: 28min 33s
Logical resource usage: 2.0/16 CPUs, 1.0/1 GPUs (0.0/1.0 accelerator_type:A10G)
Current best trial: ad99a_00002 with loss=1.3837229575751349 and params={'l1': 256, 'l2': 4, 'lr': 0.0007109136804996362, 'batch_size': 2, 'device': 'cuda'}
╭────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────╮
│ Trial name                status         l1     l2            lr     batch_size     iter     total time (s)      loss     accuracy │
├────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────┤
│ train_cifar_ad99a_00004   RUNNING       256      1   0.000823347              4        6           193.615    1.77252       0.3095 │
│ train_cifar_ad99a_00000   TERMINATED     32      1   0.0123941                2       10           615.167    2.31109       0.1017 │
│ train_cifar_ad99a_00001   TERMINATED      4      1   0.0101213               16       10            89.6874   2.30332       0.0997 │
│ train_cifar_ad99a_00002   TERMINATED    256      4   0.000710914              2       10           605.089    1.38372       0.5492 │
│ train_cifar_ad99a_00003   TERMINATED     16      1   0.00261551               8       10           165.841    1.77842       0.262  │
│ train_cifar_ad99a_00005   PENDING         8     32   0.0264297                4                                                    │
│ train_cifar_ad99a_00006   PENDING       256    128   0.0848744               16                                                    │
│ train_cifar_ad99a_00007   PENDING       128    128   0.0010942                2                                                    │
│ train_cifar_ad99a_00008   PENDING         2      8   0.019285                16                                                    │
│ train_cifar_ad99a_00009   PENDING       128     32   0.00107603               4                                                    │
╰────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────╯
(func pid=11253) [7,  8000] loss: 0.423
(func pid=11253) [7, 10000] loss: 0.338
(func pid=11253) Checkpoint successfully created at: Checkpoint(filesystem=local, path=/var/lib/ci-user/ray_results/train_cifar_2026-05-08_21-48-41/train_cifar_ad99a_00004_4_batch_size=4,l1=256,l2=1,lr=0.0008_2026-05-08_21-48-41/checkpoint_000006)
(func pid=11253) [8,  2000] loss: 1.651
(func pid=11253) [8,  4000] loss: 0.833
(func pid=11253) [8,  6000] loss: 0.558
Trial status: 4 TERMINATED | 1 RUNNING | 5 PENDING
Current time: 2026-05-08 22:17:45. Total running time: 29min 3s
Logical resource usage: 2.0/16 CPUs, 1.0/1 GPUs (0.0/1.0 accelerator_type:A10G)
Current best trial: ad99a_00002 with loss=1.3837229575751349 and params={'l1': 256, 'l2': 4, 'lr': 0.0007109136804996362, 'batch_size': 2, 'device': 'cuda'}
╭────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────╮
│ Trial name                status         l1     l2            lr     batch_size     iter     total time (s)      loss     accuracy │
├────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────┤
│ train_cifar_ad99a_00004   RUNNING       256      1   0.000823347              4        7           225.55     1.75134       0.3088 │
│ train_cifar_ad99a_00000   TERMINATED     32      1   0.0123941                2       10           615.167    2.31109       0.1017 │
│ train_cifar_ad99a_00001   TERMINATED      4      1   0.0101213               16       10            89.6874   2.30332       0.0997 │
│ train_cifar_ad99a_00002   TERMINATED    256      4   0.000710914              2       10           605.089    1.38372       0.5492 │
│ train_cifar_ad99a_00003   TERMINATED     16      1   0.00261551               8       10           165.841    1.77842       0.262  │
│ train_cifar_ad99a_00005   PENDING         8     32   0.0264297                4                                                    │
│ train_cifar_ad99a_00006   PENDING       256    128   0.0848744               16                                                    │
│ train_cifar_ad99a_00007   PENDING       128    128   0.0010942                2                                                    │
│ train_cifar_ad99a_00008   PENDING         2      8   0.019285                16                                                    │
│ train_cifar_ad99a_00009   PENDING       128     32   0.00107603               4                                                    │
╰────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────╯
(func pid=11253) [8,  8000] loss: 0.420
(func pid=11253) [8, 10000] loss: 0.339
(func pid=11253) Checkpoint successfully created at: Checkpoint(filesystem=local, path=/var/lib/ci-user/ray_results/train_cifar_2026-05-08_21-48-41/train_cifar_ad99a_00004_4_batch_size=4,l1=256,l2=1,lr=0.0008_2026-05-08_21-48-41/checkpoint_000007)
(func pid=11253) [9,  2000] loss: 1.654
(func pid=11253) [9,  4000] loss: 0.825
(func pid=11253) [9,  6000] loss: 0.550
Trial status: 4 TERMINATED | 1 RUNNING | 5 PENDING
Current time: 2026-05-08 22:18:15. Total running time: 29min 33s
Logical resource usage: 2.0/16 CPUs, 1.0/1 GPUs (0.0/1.0 accelerator_type:A10G)
Current best trial: ad99a_00002 with loss=1.3837229575751349 and params={'l1': 256, 'l2': 4, 'lr': 0.0007109136804996362, 'batch_size': 2, 'device': 'cuda'}
╭────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────╮
│ Trial name                status         l1     l2            lr     batch_size     iter     total time (s)      loss     accuracy │
├────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────┤
│ train_cifar_ad99a_00004   RUNNING       256      1   0.000823347              4        8           257.992    1.78931       0.3201 │
│ train_cifar_ad99a_00000   TERMINATED     32      1   0.0123941                2       10           615.167    2.31109       0.1017 │
│ train_cifar_ad99a_00001   TERMINATED      4      1   0.0101213               16       10            89.6874   2.30332       0.0997 │
│ train_cifar_ad99a_00002   TERMINATED    256      4   0.000710914              2       10           605.089    1.38372       0.5492 │
│ train_cifar_ad99a_00003   TERMINATED     16      1   0.00261551               8       10           165.841    1.77842       0.262  │
│ train_cifar_ad99a_00005   PENDING         8     32   0.0264297                4                                                    │
│ train_cifar_ad99a_00006   PENDING       256    128   0.0848744               16                                                    │
│ train_cifar_ad99a_00007   PENDING       128    128   0.0010942                2                                                    │
│ train_cifar_ad99a_00008   PENDING         2      8   0.019285                16                                                    │
│ train_cifar_ad99a_00009   PENDING       128     32   0.00107603               4                                                    │
╰────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────╯
(func pid=11253) [9,  8000] loss: 0.416
(func pid=11253) [9, 10000] loss: 0.335
(func pid=11253) Checkpoint successfully created at: Checkpoint(filesystem=local, path=/var/lib/ci-user/ray_results/train_cifar_2026-05-08_21-48-41/train_cifar_ad99a_00004_4_batch_size=4,l1=256,l2=1,lr=0.0008_2026-05-08_21-48-41/checkpoint_000008)
(func pid=11253) [10,  2000] loss: 1.606
(func pid=11253) [10,  4000] loss: 0.799
Trial status: 4 TERMINATED | 1 RUNNING | 5 PENDING
Current time: 2026-05-08 22:18:45. Total running time: 30min 3s
Logical resource usage: 2.0/16 CPUs, 1.0/1 GPUs (0.0/1.0 accelerator_type:A10G)
Current best trial: ad99a_00002 with loss=1.3837229575751349 and params={'l1': 256, 'l2': 4, 'lr': 0.0007109136804996362, 'batch_size': 2, 'device': 'cuda'}
╭────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────╮
│ Trial name                status         l1     l2            lr     batch_size     iter     total time (s)      loss     accuracy │
├────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────┤
│ train_cifar_ad99a_00004   RUNNING       256      1   0.000823347              4        9           289.855    1.76786       0.322  │
│ train_cifar_ad99a_00000   TERMINATED     32      1   0.0123941                2       10           615.167    2.31109       0.1017 │
│ train_cifar_ad99a_00001   TERMINATED      4      1   0.0101213               16       10            89.6874   2.30332       0.0997 │
│ train_cifar_ad99a_00002   TERMINATED    256      4   0.000710914              2       10           605.089    1.38372       0.5492 │
│ train_cifar_ad99a_00003   TERMINATED     16      1   0.00261551               8       10           165.841    1.77842       0.262  │
│ train_cifar_ad99a_00005   PENDING         8     32   0.0264297                4                                                    │
│ train_cifar_ad99a_00006   PENDING       256    128   0.0848744               16                                                    │
│ train_cifar_ad99a_00007   PENDING       128    128   0.0010942                2                                                    │
│ train_cifar_ad99a_00008   PENDING         2      8   0.019285                16                                                    │
│ train_cifar_ad99a_00009   PENDING       128     32   0.00107603               4                                                    │
╰────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────╯
(func pid=11253) [10,  6000] loss: 0.544
(func pid=11253) [10,  8000] loss: 0.414
(func pid=11253) [10, 10000] loss: 0.327

Trial train_cifar_ad99a_00004 completed after 10 iterations at 2026-05-08 22:19:01. Total running time: 30min 19s
╭────────────────────────────────────────────────────────────╮
│ Trial train_cifar_ad99a_00004 result                       │
├────────────────────────────────────────────────────────────┤
│ checkpoint_dir_name                      checkpoint_000009 │
│ time_this_iter_s                                  31.32033 │
│ time_total_s                                     321.17495 │
│ training_iteration                                      10 │
│ accuracy                                            0.3073 │
│ loss                                               1.77184 │
╰────────────────────────────────────────────────────────────╯
(func pid=11253) Checkpoint successfully created at: Checkpoint(filesystem=local, path=/var/lib/ci-user/ray_results/train_cifar_2026-05-08_21-48-41/train_cifar_ad99a_00004_4_batch_size=4,l1=256,l2=1,lr=0.0008_2026-05-08_21-48-41/checkpoint_000009)

Trial train_cifar_ad99a_00005 started with configuration:
╭──────────────────────────────────────────────────╮
│ Trial train_cifar_ad99a_00005 config             │
├──────────────────────────────────────────────────┤
│ batch_size                                     4 │
│ device                                      cuda │
│ l1                                             8 │
│ l2                                            32 │
│ lr                                       0.02643 │
╰──────────────────────────────────────────────────╯
(func pid=12572) [1,  2000] loss: 2.317

Trial status: 5 TERMINATED | 1 RUNNING | 4 PENDING
Current time: 2026-05-08 22:19:15. Total running time: 30min 33s
Logical resource usage: 2.0/16 CPUs, 1.0/1 GPUs (0.0/1.0 accelerator_type:A10G)
Current best trial: ad99a_00002 with loss=1.3837229575751349 and params={'l1': 256, 'l2': 4, 'lr': 0.0007109136804996362, 'batch_size': 2, 'device': 'cuda'}
╭────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────╮
│ Trial name                status         l1     l2            lr     batch_size     iter     total time (s)      loss     accuracy │
├────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────┤
│ train_cifar_ad99a_00005   RUNNING         8     32   0.0264297                4                                                    │
│ train_cifar_ad99a_00000   TERMINATED     32      1   0.0123941                2       10           615.167    2.31109       0.1017 │
│ train_cifar_ad99a_00001   TERMINATED      4      1   0.0101213               16       10            89.6874   2.30332       0.0997 │
│ train_cifar_ad99a_00002   TERMINATED    256      4   0.000710914              2       10           605.089    1.38372       0.5492 │
│ train_cifar_ad99a_00003   TERMINATED     16      1   0.00261551               8       10           165.841    1.77842       0.262  │
│ train_cifar_ad99a_00004   TERMINATED    256      1   0.000823347              4       10           321.175    1.77184       0.3073 │
│ train_cifar_ad99a_00006   PENDING       256    128   0.0848744               16                                                    │
│ train_cifar_ad99a_00007   PENDING       128    128   0.0010942                2                                                    │
│ train_cifar_ad99a_00008   PENDING         2      8   0.019285                16                                                    │
│ train_cifar_ad99a_00009   PENDING       128     32   0.00107603               4                                                    │
╰────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────╯
(func pid=12572) [1,  4000] loss: 1.160
(func pid=12572) [1,  6000] loss: 0.772
(func pid=12572) [1,  8000] loss: 0.580
(func pid=12572) [1, 10000] loss: 0.463

Trial train_cifar_ad99a_00005 completed after 1 iterations at 2026-05-08 22:19:39. Total running time: 30min 57s
╭────────────────────────────────────────────────────────────╮
│ Trial train_cifar_ad99a_00005 result                       │
├────────────────────────────────────────────────────────────┤
│ checkpoint_dir_name                      checkpoint_000000 │
│ time_this_iter_s                                  33.05294 │
│ time_total_s                                      33.05294 │
│ training_iteration                                       1 │
│ accuracy                                            0.1003 │
│ loss                                               2.31767 │
╰────────────────────────────────────────────────────────────╯
(func pid=12572) Checkpoint successfully created at: Checkpoint(filesystem=local, path=/var/lib/ci-user/ray_results/train_cifar_2026-05-08_21-48-41/train_cifar_ad99a_00005_5_batch_size=4,l1=8,l2=32,lr=0.0264_2026-05-08_21-48-41/checkpoint_000000)

Trial train_cifar_ad99a_00006 started with configuration:
╭──────────────────────────────────────────────────╮
│ Trial train_cifar_ad99a_00006 config             │
├──────────────────────────────────────────────────┤
│ batch_size                                    16 │
│ device                                      cuda │
│ l1                                           256 │
│ l2                                           128 │
│ lr                                       0.08487 │
╰──────────────────────────────────────────────────╯

Trial status: 6 TERMINATED | 1 RUNNING | 3 PENDING
Current time: 2026-05-08 22:19:45. Total running time: 31min 3s
Logical resource usage: 2.0/16 CPUs, 1.0/1 GPUs (0.0/1.0 accelerator_type:A10G)
Current best trial: ad99a_00002 with loss=1.3837229575751349 and params={'l1': 256, 'l2': 4, 'lr': 0.0007109136804996362, 'batch_size': 2, 'device': 'cuda'}
╭────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────╮
│ Trial name                status         l1     l2            lr     batch_size     iter     total time (s)      loss     accuracy │
├────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────┤
│ train_cifar_ad99a_00006   RUNNING       256    128   0.0848744               16                                                    │
│ train_cifar_ad99a_00000   TERMINATED     32      1   0.0123941                2       10           615.167    2.31109       0.1017 │
│ train_cifar_ad99a_00001   TERMINATED      4      1   0.0101213               16       10            89.6874   2.30332       0.0997 │
│ train_cifar_ad99a_00002   TERMINATED    256      4   0.000710914              2       10           605.089    1.38372       0.5492 │
│ train_cifar_ad99a_00003   TERMINATED     16      1   0.00261551               8       10           165.841    1.77842       0.262  │
│ train_cifar_ad99a_00004   TERMINATED    256      1   0.000823347              4       10           321.175    1.77184       0.3073 │
│ train_cifar_ad99a_00005   TERMINATED      8     32   0.0264297                4        1            33.0529   2.31767       0.1003 │
│ train_cifar_ad99a_00007   PENDING       128    128   0.0010942                2                                                    │
│ train_cifar_ad99a_00008   PENDING         2      8   0.019285                16                                                    │
│ train_cifar_ad99a_00009   PENDING       128     32   0.00107603               4                                                    │
╰────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────╯
(func pid=12775) [1,  2000] loss: 2.267

Trial train_cifar_ad99a_00006 completed after 1 iterations at 2026-05-08 22:19:54. Total running time: 31min 13s
╭────────────────────────────────────────────────────────────╮
│ Trial train_cifar_ad99a_00006 result                       │
├────────────────────────────────────────────────────────────┤
│ checkpoint_dir_name                      checkpoint_000000 │
│ time_this_iter_s                                  10.59678 │
│ time_total_s                                      10.59678 │
│ training_iteration                                       1 │
│ accuracy                                            0.1804 │
│ loss                                               2.19539 │
╰────────────────────────────────────────────────────────────╯
(func pid=12775) Checkpoint successfully created at: Checkpoint(filesystem=local, path=/var/lib/ci-user/ray_results/train_cifar_2026-05-08_21-48-41/train_cifar_ad99a_00006_6_batch_size=16,l1=256,l2=128,lr=0.0849_2026-05-08_21-48-41/checkpoint_000000)

Trial train_cifar_ad99a_00007 started with configuration:
╭──────────────────────────────────────────────────╮
│ Trial train_cifar_ad99a_00007 config             │
├──────────────────────────────────────────────────┤
│ batch_size                                     2 │
│ device                                      cuda │
│ l1                                           128 │
│ l2                                           128 │
│ lr                                       0.00109 │
╰──────────────────────────────────────────────────╯
(func pid=12943) [1,  2000] loss: 2.086
(func pid=12943) [1,  4000] loss: 0.908

Trial status: 7 TERMINATED | 1 RUNNING | 2 PENDING
Current time: 2026-05-08 22:20:15. Total running time: 31min 34s
Logical resource usage: 2.0/16 CPUs, 1.0/1 GPUs (0.0/1.0 accelerator_type:A10G)
Current best trial: ad99a_00002 with loss=1.3837229575751349 and params={'l1': 256, 'l2': 4, 'lr': 0.0007109136804996362, 'batch_size': 2, 'device': 'cuda'}
╭────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────╮
│ Trial name                status         l1     l2            lr     batch_size     iter     total time (s)      loss     accuracy │
├────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────┤
│ train_cifar_ad99a_00007   RUNNING       128    128   0.0010942                2                                                    │
│ train_cifar_ad99a_00000   TERMINATED     32      1   0.0123941                2       10           615.167    2.31109       0.1017 │
│ train_cifar_ad99a_00001   TERMINATED      4      1   0.0101213               16       10            89.6874   2.30332       0.0997 │
│ train_cifar_ad99a_00002   TERMINATED    256      4   0.000710914              2       10           605.089    1.38372       0.5492 │
│ train_cifar_ad99a_00003   TERMINATED     16      1   0.00261551               8       10           165.841    1.77842       0.262  │
│ train_cifar_ad99a_00004   TERMINATED    256      1   0.000823347              4       10           321.175    1.77184       0.3073 │
│ train_cifar_ad99a_00005   TERMINATED      8     32   0.0264297                4        1            33.0529   2.31767       0.1003 │
│ train_cifar_ad99a_00006   TERMINATED    256    128   0.0848744               16        1            10.5968   2.19539       0.1804 │
│ train_cifar_ad99a_00008   PENDING         2      8   0.019285                16                                                    │
│ train_cifar_ad99a_00009   PENDING       128     32   0.00107603               4                                                    │
╰────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────╯
(func pid=12943) [1,  6000] loss: 0.570
(func pid=12943) [1,  8000] loss: 0.413
(func pid=12943) [1, 10000] loss: 0.322
(func pid=12943) [1, 12000] loss: 0.269
(func pid=12943) [1, 14000] loss: 0.226
(func pid=12943) [1, 16000] loss: 0.195
Trial status: 7 TERMINATED | 1 RUNNING | 2 PENDING
Current time: 2026-05-08 22:20:45. Total running time: 32min 4s
Logical resource usage: 2.0/16 CPUs, 1.0/1 GPUs (0.0/1.0 accelerator_type:A10G)
Current best trial: ad99a_00002 with loss=1.3837229575751349 and params={'l1': 256, 'l2': 4, 'lr': 0.0007109136804996362, 'batch_size': 2, 'device': 'cuda'}
╭────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────╮
│ Trial name                status         l1     l2            lr     batch_size     iter     total time (s)      loss     accuracy │
├────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────┤
│ train_cifar_ad99a_00007   RUNNING       128    128   0.0010942                2                                                    │
│ train_cifar_ad99a_00000   TERMINATED     32      1   0.0123941                2       10           615.167    2.31109       0.1017 │
│ train_cifar_ad99a_00001   TERMINATED      4      1   0.0101213               16       10            89.6874   2.30332       0.0997 │
│ train_cifar_ad99a_00002   TERMINATED    256      4   0.000710914              2       10           605.089    1.38372       0.5492 │
│ train_cifar_ad99a_00003   TERMINATED     16      1   0.00261551               8       10           165.841    1.77842       0.262  │
│ train_cifar_ad99a_00004   TERMINATED    256      1   0.000823347              4       10           321.175    1.77184       0.3073 │
│ train_cifar_ad99a_00005   TERMINATED      8     32   0.0264297                4        1            33.0529   2.31767       0.1003 │
│ train_cifar_ad99a_00006   TERMINATED    256    128   0.0848744               16        1            10.5968   2.19539       0.1804 │
│ train_cifar_ad99a_00008   PENDING         2      8   0.019285                16                                                    │
│ train_cifar_ad99a_00009   PENDING       128     32   0.00107603               4                                                    │
╰────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────╯
(func pid=12943) [1, 18000] loss: 0.172
(func pid=12943) [1, 20000] loss: 0.152
(func pid=12943) Checkpoint successfully created at: Checkpoint(filesystem=local, path=/var/lib/ci-user/ray_results/train_cifar_2026-05-08_21-48-41/train_cifar_ad99a_00007_7_batch_size=2,l1=128,l2=128,lr=0.0011_2026-05-08_21-48-41/checkpoint_000000)
(func pid=12943) [2,  2000] loss: 1.438
(func pid=12943) [2,  4000] loss: 0.754
Trial status: 7 TERMINATED | 1 RUNNING | 2 PENDING
Current time: 2026-05-08 22:21:15. Total running time: 32min 34s
Logical resource usage: 2.0/16 CPUs, 1.0/1 GPUs (0.0/1.0 accelerator_type:A10G)
Current best trial: ad99a_00002 with loss=1.3837229575751349 and params={'l1': 256, 'l2': 4, 'lr': 0.0007109136804996362, 'batch_size': 2, 'device': 'cuda'}
╭────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────╮
│ Trial name                status         l1     l2            lr     batch_size     iter     total time (s)      loss     accuracy │
├────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────┤
│ train_cifar_ad99a_00007   RUNNING       128    128   0.0010942                2        1            62.3062   1.5084        0.4415 │
│ train_cifar_ad99a_00000   TERMINATED     32      1   0.0123941                2       10           615.167    2.31109       0.1017 │
│ train_cifar_ad99a_00001   TERMINATED      4      1   0.0101213               16       10            89.6874   2.30332       0.0997 │
│ train_cifar_ad99a_00002   TERMINATED    256      4   0.000710914              2       10           605.089    1.38372       0.5492 │
│ train_cifar_ad99a_00003   TERMINATED     16      1   0.00261551               8       10           165.841    1.77842       0.262  │
│ train_cifar_ad99a_00004   TERMINATED    256      1   0.000823347              4       10           321.175    1.77184       0.3073 │
│ train_cifar_ad99a_00005   TERMINATED      8     32   0.0264297                4        1            33.0529   2.31767       0.1003 │
│ train_cifar_ad99a_00006   TERMINATED    256    128   0.0848744               16        1            10.5968   2.19539       0.1804 │
│ train_cifar_ad99a_00008   PENDING         2      8   0.019285                16                                                    │
│ train_cifar_ad99a_00009   PENDING       128     32   0.00107603               4                                                    │
╰────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────╯
(func pid=12943) [2,  6000] loss: 0.483
(func pid=12943) [2,  8000] loss: 0.362
(func pid=12943) [2, 10000] loss: 0.293
(func pid=12943) [2, 12000] loss: 0.245
(func pid=12943) [2, 14000] loss: 0.207
(func pid=12943) [2, 16000] loss: 0.181
Trial status: 7 TERMINATED | 1 RUNNING | 2 PENDING
Current time: 2026-05-08 22:21:45. Total running time: 33min 4s
Logical resource usage: 2.0/16 CPUs, 1.0/1 GPUs (0.0/1.0 accelerator_type:A10G)
Current best trial: ad99a_00002 with loss=1.3837229575751349 and params={'l1': 256, 'l2': 4, 'lr': 0.0007109136804996362, 'batch_size': 2, 'device': 'cuda'}
╭────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────╮
│ Trial name                status         l1     l2            lr     batch_size     iter     total time (s)      loss     accuracy │
├────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────┤
│ train_cifar_ad99a_00007   RUNNING       128    128   0.0010942                2        1            62.3062   1.5084        0.4415 │
│ train_cifar_ad99a_00000   TERMINATED     32      1   0.0123941                2       10           615.167    2.31109       0.1017 │
│ train_cifar_ad99a_00001   TERMINATED      4      1   0.0101213               16       10            89.6874   2.30332       0.0997 │
│ train_cifar_ad99a_00002   TERMINATED    256      4   0.000710914              2       10           605.089    1.38372       0.5492 │
│ train_cifar_ad99a_00003   TERMINATED     16      1   0.00261551               8       10           165.841    1.77842       0.262  │
│ train_cifar_ad99a_00004   TERMINATED    256      1   0.000823347              4       10           321.175    1.77184       0.3073 │
│ train_cifar_ad99a_00005   TERMINATED      8     32   0.0264297                4        1            33.0529   2.31767       0.1003 │
│ train_cifar_ad99a_00006   TERMINATED    256    128   0.0848744               16        1            10.5968   2.19539       0.1804 │
│ train_cifar_ad99a_00008   PENDING         2      8   0.019285                16                                                    │
│ train_cifar_ad99a_00009   PENDING       128     32   0.00107603               4                                                    │
╰────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────╯
(func pid=12943) [2, 18000] loss: 0.161
(func pid=12943) [2, 20000] loss: 0.141
(func pid=12943) Checkpoint successfully created at: Checkpoint(filesystem=local, path=/var/lib/ci-user/ray_results/train_cifar_2026-05-08_21-48-41/train_cifar_ad99a_00007_7_batch_size=2,l1=128,l2=128,lr=0.0011_2026-05-08_21-48-41/checkpoint_000001)
(func pid=12943) [3,  2000] loss: 1.355
(func pid=12943) [3,  4000] loss: 0.688
Trial status: 7 TERMINATED | 1 RUNNING | 2 PENDING
Current time: 2026-05-08 22:22:15. Total running time: 33min 34s
Logical resource usage: 2.0/16 CPUs, 1.0/1 GPUs (0.0/1.0 accelerator_type:A10G)
Current best trial: ad99a_00002 with loss=1.3837229575751349 and params={'l1': 256, 'l2': 4, 'lr': 0.0007109136804996362, 'batch_size': 2, 'device': 'cuda'}
╭────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────╮
│ Trial name                status         l1     l2            lr     batch_size     iter     total time (s)      loss     accuracy │
├────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────┤
│ train_cifar_ad99a_00007   RUNNING       128    128   0.0010942                2        2           122.735    1.43182       0.4891 │
│ train_cifar_ad99a_00000   TERMINATED     32      1   0.0123941                2       10           615.167    2.31109       0.1017 │
│ train_cifar_ad99a_00001   TERMINATED      4      1   0.0101213               16       10            89.6874   2.30332       0.0997 │
│ train_cifar_ad99a_00002   TERMINATED    256      4   0.000710914              2       10           605.089    1.38372       0.5492 │
│ train_cifar_ad99a_00003   TERMINATED     16      1   0.00261551               8       10           165.841    1.77842       0.262  │
│ train_cifar_ad99a_00004   TERMINATED    256      1   0.000823347              4       10           321.175    1.77184       0.3073 │
│ train_cifar_ad99a_00005   TERMINATED      8     32   0.0264297                4        1            33.0529   2.31767       0.1003 │
│ train_cifar_ad99a_00006   TERMINATED    256    128   0.0848744               16        1            10.5968   2.19539       0.1804 │
│ train_cifar_ad99a_00008   PENDING         2      8   0.019285                16                                                    │
│ train_cifar_ad99a_00009   PENDING       128     32   0.00107603               4                                                    │
╰────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────╯
(func pid=12943) [3,  6000] loss: 0.467
(func pid=12943) [3,  8000] loss: 0.340
(func pid=12943) [3, 10000] loss: 0.272
(func pid=12943) [3, 12000] loss: 0.229
(func pid=12943) [3, 14000] loss: 0.200
(func pid=12943) [3, 16000] loss: 0.173
Trial status: 7 TERMINATED | 1 RUNNING | 2 PENDING
Current time: 2026-05-08 22:22:45. Total running time: 34min 4s
Logical resource usage: 2.0/16 CPUs, 1.0/1 GPUs (0.0/1.0 accelerator_type:A10G)
Current best trial: ad99a_00002 with loss=1.3837229575751349 and params={'l1': 256, 'l2': 4, 'lr': 0.0007109136804996362, 'batch_size': 2, 'device': 'cuda'}
╭────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────╮
│ Trial name                status         l1     l2            lr     batch_size     iter     total time (s)      loss     accuracy │
├────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────┤
│ train_cifar_ad99a_00007   RUNNING       128    128   0.0010942                2        2           122.735    1.43182       0.4891 │
│ train_cifar_ad99a_00000   TERMINATED     32      1   0.0123941                2       10           615.167    2.31109       0.1017 │
│ train_cifar_ad99a_00001   TERMINATED      4      1   0.0101213               16       10            89.6874   2.30332       0.0997 │
│ train_cifar_ad99a_00002   TERMINATED    256      4   0.000710914              2       10           605.089    1.38372       0.5492 │
│ train_cifar_ad99a_00003   TERMINATED     16      1   0.00261551               8       10           165.841    1.77842       0.262  │
│ train_cifar_ad99a_00004   TERMINATED    256      1   0.000823347              4       10           321.175    1.77184       0.3073 │
│ train_cifar_ad99a_00005   TERMINATED      8     32   0.0264297                4        1            33.0529   2.31767       0.1003 │
│ train_cifar_ad99a_00006   TERMINATED    256    128   0.0848744               16        1            10.5968   2.19539       0.1804 │
│ train_cifar_ad99a_00008   PENDING         2      8   0.019285                16                                                    │
│ train_cifar_ad99a_00009   PENDING       128     32   0.00107603               4                                                    │
╰────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────╯
(func pid=12943) [3, 18000] loss: 0.153
(func pid=12943) [3, 20000] loss: 0.139
(func pid=12943) Checkpoint successfully created at: Checkpoint(filesystem=local, path=/var/lib/ci-user/ray_results/train_cifar_2026-05-08_21-48-41/train_cifar_ad99a_00007_7_batch_size=2,l1=128,l2=128,lr=0.0011_2026-05-08_21-48-41/checkpoint_000002)
(func pid=12943) [4,  2000] loss: 1.311
(func pid=12943) [4,  4000] loss: 0.642
Trial status: 7 TERMINATED | 1 RUNNING | 2 PENDING
Current time: 2026-05-08 22:23:15. Total running time: 34min 34s
Logical resource usage: 2.0/16 CPUs, 1.0/1 GPUs (0.0/1.0 accelerator_type:A10G)
Current best trial: ad99a_00007 with loss=1.3442824479543138 and params={'l1': 128, 'l2': 128, 'lr': 0.0010942045681162035, 'batch_size': 2, 'device': 'cuda'}
╭────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────╮
│ Trial name                status         l1     l2            lr     batch_size     iter     total time (s)      loss     accuracy │
├────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────┤
│ train_cifar_ad99a_00007   RUNNING       128    128   0.0010942                2        3           182.697    1.34428       0.5276 │
│ train_cifar_ad99a_00000   TERMINATED     32      1   0.0123941                2       10           615.167    2.31109       0.1017 │
│ train_cifar_ad99a_00001   TERMINATED      4      1   0.0101213               16       10            89.6874   2.30332       0.0997 │
│ train_cifar_ad99a_00002   TERMINATED    256      4   0.000710914              2       10           605.089    1.38372       0.5492 │
│ train_cifar_ad99a_00003   TERMINATED     16      1   0.00261551               8       10           165.841    1.77842       0.262  │
│ train_cifar_ad99a_00004   TERMINATED    256      1   0.000823347              4       10           321.175    1.77184       0.3073 │
│ train_cifar_ad99a_00005   TERMINATED      8     32   0.0264297                4        1            33.0529   2.31767       0.1003 │
│ train_cifar_ad99a_00006   TERMINATED    256    128   0.0848744               16        1            10.5968   2.19539       0.1804 │
│ train_cifar_ad99a_00008   PENDING         2      8   0.019285                16                                                    │
│ train_cifar_ad99a_00009   PENDING       128     32   0.00107603               4                                                    │
╰────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────╯
(func pid=12943) [4,  6000] loss: 0.437
(func pid=12943) [4,  8000] loss: 0.328
(func pid=12943) [4, 10000] loss: 0.260
(func pid=12943) [4, 12000] loss: 0.221
(func pid=12943) [4, 14000] loss: 0.192
(func pid=12943) [4, 16000] loss: 0.166
Trial status: 7 TERMINATED | 1 RUNNING | 2 PENDING
Current time: 2026-05-08 22:23:45. Total running time: 35min 4s
Logical resource usage: 2.0/16 CPUs, 1.0/1 GPUs (0.0/1.0 accelerator_type:A10G)
Current best trial: ad99a_00007 with loss=1.3442824479543138 and params={'l1': 128, 'l2': 128, 'lr': 0.0010942045681162035, 'batch_size': 2, 'device': 'cuda'}
╭────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────╮
│ Trial name                status         l1     l2            lr     batch_size     iter     total time (s)      loss     accuracy │
├────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────┤
│ train_cifar_ad99a_00007   RUNNING       128    128   0.0010942                2        3           182.697    1.34428       0.5276 │
│ train_cifar_ad99a_00000   TERMINATED     32      1   0.0123941                2       10           615.167    2.31109       0.1017 │
│ train_cifar_ad99a_00001   TERMINATED      4      1   0.0101213               16       10            89.6874   2.30332       0.0997 │
│ train_cifar_ad99a_00002   TERMINATED    256      4   0.000710914              2       10           605.089    1.38372       0.5492 │
│ train_cifar_ad99a_00003   TERMINATED     16      1   0.00261551               8       10           165.841    1.77842       0.262  │
│ train_cifar_ad99a_00004   TERMINATED    256      1   0.000823347              4       10           321.175    1.77184       0.3073 │
│ train_cifar_ad99a_00005   TERMINATED      8     32   0.0264297                4        1            33.0529   2.31767       0.1003 │
│ train_cifar_ad99a_00006   TERMINATED    256    128   0.0848744               16        1            10.5968   2.19539       0.1804 │
│ train_cifar_ad99a_00008   PENDING         2      8   0.019285                16                                                    │
│ train_cifar_ad99a_00009   PENDING       128     32   0.00107603               4                                                    │
╰────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────╯
(func pid=12943) [4, 18000] loss: 0.151
(func pid=12943) [4, 20000] loss: 0.134
(func pid=12943) Checkpoint successfully created at: Checkpoint(filesystem=local, path=/var/lib/ci-user/ray_results/train_cifar_2026-05-08_21-48-41/train_cifar_ad99a_00007_7_batch_size=2,l1=128,l2=128,lr=0.0011_2026-05-08_21-48-41/checkpoint_000003)
(func pid=12943) [5,  2000] loss: 1.276
(func pid=12943) [5,  4000] loss: 0.651
Trial status: 7 TERMINATED | 1 RUNNING | 2 PENDING
Current time: 2026-05-08 22:24:15. Total running time: 35min 34s
Logical resource usage: 2.0/16 CPUs, 1.0/1 GPUs (0.0/1.0 accelerator_type:A10G)
Current best trial: ad99a_00002 with loss=1.3837229575751349 and params={'l1': 256, 'l2': 4, 'lr': 0.0007109136804996362, 'batch_size': 2, 'device': 'cuda'}
╭────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────╮
│ Trial name                status         l1     l2            lr     batch_size     iter     total time (s)      loss     accuracy │
├────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────┤
│ train_cifar_ad99a_00007   RUNNING       128    128   0.0010942                2        4           243.339    1.45474       0.5054 │
│ train_cifar_ad99a_00000   TERMINATED     32      1   0.0123941                2       10           615.167    2.31109       0.1017 │
│ train_cifar_ad99a_00001   TERMINATED      4      1   0.0101213               16       10            89.6874   2.30332       0.0997 │
│ train_cifar_ad99a_00002   TERMINATED    256      4   0.000710914              2       10           605.089    1.38372       0.5492 │
│ train_cifar_ad99a_00003   TERMINATED     16      1   0.00261551               8       10           165.841    1.77842       0.262  │
│ train_cifar_ad99a_00004   TERMINATED    256      1   0.000823347              4       10           321.175    1.77184       0.3073 │
│ train_cifar_ad99a_00005   TERMINATED      8     32   0.0264297                4        1            33.0529   2.31767       0.1003 │
│ train_cifar_ad99a_00006   TERMINATED    256    128   0.0848744               16        1            10.5968   2.19539       0.1804 │
│ train_cifar_ad99a_00008   PENDING         2      8   0.019285                16                                                    │
│ train_cifar_ad99a_00009   PENDING       128     32   0.00107603               4                                                    │
╰────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────╯
(func pid=12943) [5,  6000] loss: 0.423
(func pid=12943) [5,  8000] loss: 0.329
(func pid=12943) [5, 10000] loss: 0.254
(func pid=12943) [5, 12000] loss: 0.214
(func pid=12943) [5, 14000] loss: 0.189
(func pid=12943) [5, 16000] loss: 0.160
Trial status: 7 TERMINATED | 1 RUNNING | 2 PENDING
Current time: 2026-05-08 22:24:46. Total running time: 36min 4s
Logical resource usage: 2.0/16 CPUs, 1.0/1 GPUs (0.0/1.0 accelerator_type:A10G)
Current best trial: ad99a_00002 with loss=1.3837229575751349 and params={'l1': 256, 'l2': 4, 'lr': 0.0007109136804996362, 'batch_size': 2, 'device': 'cuda'}
╭────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────╮
│ Trial name                status         l1     l2            lr     batch_size     iter     total time (s)      loss     accuracy │
├────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────┤
│ train_cifar_ad99a_00007   RUNNING       128    128   0.0010942                2        4           243.339    1.45474       0.5054 │
│ train_cifar_ad99a_00000   TERMINATED     32      1   0.0123941                2       10           615.167    2.31109       0.1017 │
│ train_cifar_ad99a_00001   TERMINATED      4      1   0.0101213               16       10            89.6874   2.30332       0.0997 │
│ train_cifar_ad99a_00002   TERMINATED    256      4   0.000710914              2       10           605.089    1.38372       0.5492 │
│ train_cifar_ad99a_00003   TERMINATED     16      1   0.00261551               8       10           165.841    1.77842       0.262  │
│ train_cifar_ad99a_00004   TERMINATED    256      1   0.000823347              4       10           321.175    1.77184       0.3073 │
│ train_cifar_ad99a_00005   TERMINATED      8     32   0.0264297                4        1            33.0529   2.31767       0.1003 │
│ train_cifar_ad99a_00006   TERMINATED    256    128   0.0848744               16        1            10.5968   2.19539       0.1804 │
│ train_cifar_ad99a_00008   PENDING         2      8   0.019285                16                                                    │
│ train_cifar_ad99a_00009   PENDING       128     32   0.00107603               4                                                    │
╰────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────╯
(func pid=12943) [5, 18000] loss: 0.142
(func pid=12943) [5, 20000] loss: 0.136
(func pid=12943) Checkpoint successfully created at: Checkpoint(filesystem=local, path=/var/lib/ci-user/ray_results/train_cifar_2026-05-08_21-48-41/train_cifar_ad99a_00007_7_batch_size=2,l1=128,l2=128,lr=0.0011_2026-05-08_21-48-41/checkpoint_000004)
(func pid=12943) [6,  2000] loss: 1.218
(func pid=12943) [6,  4000] loss: 0.619
Trial status: 7 TERMINATED | 1 RUNNING | 2 PENDING
Current time: 2026-05-08 22:25:16. Total running time: 36min 34s
Logical resource usage: 2.0/16 CPUs, 1.0/1 GPUs (0.0/1.0 accelerator_type:A10G)
Current best trial: ad99a_00007 with loss=1.356649359957641 and params={'l1': 128, 'l2': 128, 'lr': 0.0010942045681162035, 'batch_size': 2, 'device': 'cuda'}
╭────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────╮
│ Trial name                status         l1     l2            lr     batch_size     iter     total time (s)      loss     accuracy │
├────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────┤
│ train_cifar_ad99a_00007   RUNNING       128    128   0.0010942                2        5           303.774    1.35665       0.5222 │
│ train_cifar_ad99a_00000   TERMINATED     32      1   0.0123941                2       10           615.167    2.31109       0.1017 │
│ train_cifar_ad99a_00001   TERMINATED      4      1   0.0101213               16       10            89.6874   2.30332       0.0997 │
│ train_cifar_ad99a_00002   TERMINATED    256      4   0.000710914              2       10           605.089    1.38372       0.5492 │
│ train_cifar_ad99a_00003   TERMINATED     16      1   0.00261551               8       10           165.841    1.77842       0.262  │
│ train_cifar_ad99a_00004   TERMINATED    256      1   0.000823347              4       10           321.175    1.77184       0.3073 │
│ train_cifar_ad99a_00005   TERMINATED      8     32   0.0264297                4        1            33.0529   2.31767       0.1003 │
│ train_cifar_ad99a_00006   TERMINATED    256    128   0.0848744               16        1            10.5968   2.19539       0.1804 │
│ train_cifar_ad99a_00008   PENDING         2      8   0.019285                16                                                    │
│ train_cifar_ad99a_00009   PENDING       128     32   0.00107603               4                                                    │
╰────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────╯
(func pid=12943) [6,  6000] loss: 0.432
(func pid=12943) [6,  8000] loss: 0.317
(func pid=12943) [6, 10000] loss: 0.255
(func pid=12943) [6, 12000] loss: 0.212
(func pid=12943) [6, 14000] loss: 0.182
(func pid=12943) [6, 16000] loss: 0.163
Trial status: 7 TERMINATED | 1 RUNNING | 2 PENDING
Current time: 2026-05-08 22:25:46. Total running time: 37min 4s
Logical resource usage: 2.0/16 CPUs, 1.0/1 GPUs (0.0/1.0 accelerator_type:A10G)
Current best trial: ad99a_00007 with loss=1.356649359957641 and params={'l1': 128, 'l2': 128, 'lr': 0.0010942045681162035, 'batch_size': 2, 'device': 'cuda'}
╭────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────╮
│ Trial name                status         l1     l2            lr     batch_size     iter     total time (s)      loss     accuracy │
├────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────┤
│ train_cifar_ad99a_00007   RUNNING       128    128   0.0010942                2        5           303.774    1.35665       0.5222 │
│ train_cifar_ad99a_00000   TERMINATED     32      1   0.0123941                2       10           615.167    2.31109       0.1017 │
│ train_cifar_ad99a_00001   TERMINATED      4      1   0.0101213               16       10            89.6874   2.30332       0.0997 │
│ train_cifar_ad99a_00002   TERMINATED    256      4   0.000710914              2       10           605.089    1.38372       0.5492 │
│ train_cifar_ad99a_00003   TERMINATED     16      1   0.00261551               8       10           165.841    1.77842       0.262  │
│ train_cifar_ad99a_00004   TERMINATED    256      1   0.000823347              4       10           321.175    1.77184       0.3073 │
│ train_cifar_ad99a_00005   TERMINATED      8     32   0.0264297                4        1            33.0529   2.31767       0.1003 │
│ train_cifar_ad99a_00006   TERMINATED    256    128   0.0848744               16        1            10.5968   2.19539       0.1804 │
│ train_cifar_ad99a_00008   PENDING         2      8   0.019285                16                                                    │
│ train_cifar_ad99a_00009   PENDING       128     32   0.00107603               4                                                    │
╰────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────╯
(func pid=12943) [6, 18000] loss: 0.148
(func pid=12943) [6, 20000] loss: 0.132
(func pid=12943) Checkpoint successfully created at: Checkpoint(filesystem=local, path=/var/lib/ci-user/ray_results/train_cifar_2026-05-08_21-48-41/train_cifar_ad99a_00007_7_batch_size=2,l1=128,l2=128,lr=0.0011_2026-05-08_21-48-41/checkpoint_000005)
(func pid=12943) [7,  2000] loss: 1.213
(func pid=12943) [7,  4000] loss: 0.607
Trial status: 7 TERMINATED | 1 RUNNING | 2 PENDING
Current time: 2026-05-08 22:26:16. Total running time: 37min 34s
Logical resource usage: 2.0/16 CPUs, 1.0/1 GPUs (0.0/1.0 accelerator_type:A10G)
Current best trial: ad99a_00007 with loss=1.3532648556645261 and params={'l1': 128, 'l2': 128, 'lr': 0.0010942045681162035, 'batch_size': 2, 'device': 'cuda'}
╭────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────╮
│ Trial name                status         l1     l2            lr     batch_size     iter     total time (s)      loss     accuracy │
├────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────┤
│ train_cifar_ad99a_00007   RUNNING       128    128   0.0010942                2        6           364.05     1.35326       0.5364 │
│ train_cifar_ad99a_00000   TERMINATED     32      1   0.0123941                2       10           615.167    2.31109       0.1017 │
│ train_cifar_ad99a_00001   TERMINATED      4      1   0.0101213               16       10            89.6874   2.30332       0.0997 │
│ train_cifar_ad99a_00002   TERMINATED    256      4   0.000710914              2       10           605.089    1.38372       0.5492 │
│ train_cifar_ad99a_00003   TERMINATED     16      1   0.00261551               8       10           165.841    1.77842       0.262  │
│ train_cifar_ad99a_00004   TERMINATED    256      1   0.000823347              4       10           321.175    1.77184       0.3073 │
│ train_cifar_ad99a_00005   TERMINATED      8     32   0.0264297                4        1            33.0529   2.31767       0.1003 │
│ train_cifar_ad99a_00006   TERMINATED    256    128   0.0848744               16        1            10.5968   2.19539       0.1804 │
│ train_cifar_ad99a_00008   PENDING         2      8   0.019285                16                                                    │
│ train_cifar_ad99a_00009   PENDING       128     32   0.00107603               4                                                    │
╰────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────╯
(func pid=12943) [7,  6000] loss: 0.419
(func pid=12943) [7,  8000] loss: 0.307
(func pid=12943) [7, 10000] loss: 0.256
(func pid=12943) [7, 12000] loss: 0.208
(func pid=12943) [7, 14000] loss: 0.191
(func pid=12943) [7, 16000] loss: 0.163
Trial status: 7 TERMINATED | 1 RUNNING | 2 PENDING
Current time: 2026-05-08 22:26:46. Total running time: 38min 4s
Logical resource usage: 2.0/16 CPUs, 1.0/1 GPUs (0.0/1.0 accelerator_type:A10G)
Current best trial: ad99a_00007 with loss=1.3532648556645261 and params={'l1': 128, 'l2': 128, 'lr': 0.0010942045681162035, 'batch_size': 2, 'device': 'cuda'}
╭────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────╮
│ Trial name                status         l1     l2            lr     batch_size     iter     total time (s)      loss     accuracy │
├────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────┤
│ train_cifar_ad99a_00007   RUNNING       128    128   0.0010942                2        6           364.05     1.35326       0.5364 │
│ train_cifar_ad99a_00000   TERMINATED     32      1   0.0123941                2       10           615.167    2.31109       0.1017 │
│ train_cifar_ad99a_00001   TERMINATED      4      1   0.0101213               16       10            89.6874   2.30332       0.0997 │
│ train_cifar_ad99a_00002   TERMINATED    256      4   0.000710914              2       10           605.089    1.38372       0.5492 │
│ train_cifar_ad99a_00003   TERMINATED     16      1   0.00261551               8       10           165.841    1.77842       0.262  │
│ train_cifar_ad99a_00004   TERMINATED    256      1   0.000823347              4       10           321.175    1.77184       0.3073 │
│ train_cifar_ad99a_00005   TERMINATED      8     32   0.0264297                4        1            33.0529   2.31767       0.1003 │
│ train_cifar_ad99a_00006   TERMINATED    256    128   0.0848744               16        1            10.5968   2.19539       0.1804 │
│ train_cifar_ad99a_00008   PENDING         2      8   0.019285                16                                                    │
│ train_cifar_ad99a_00009   PENDING       128     32   0.00107603               4                                                    │
╰────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────╯
(func pid=12943) [7, 18000] loss: 0.147
(func pid=12943) [7, 20000] loss: 0.130
(func pid=12943) Checkpoint successfully created at: Checkpoint(filesystem=local, path=/var/lib/ci-user/ray_results/train_cifar_2026-05-08_21-48-41/train_cifar_ad99a_00007_7_batch_size=2,l1=128,l2=128,lr=0.0011_2026-05-08_21-48-41/checkpoint_000006)
(func pid=12943) [8,  2000] loss: 1.214
(func pid=12943) [8,  4000] loss: 0.604
Trial status: 7 TERMINATED | 1 RUNNING | 2 PENDING
Current time: 2026-05-08 22:27:16. Total running time: 38min 34s
Logical resource usage: 2.0/16 CPUs, 1.0/1 GPUs (0.0/1.0 accelerator_type:A10G)
Current best trial: ad99a_00002 with loss=1.3837229575751349 and params={'l1': 256, 'l2': 4, 'lr': 0.0007109136804996362, 'batch_size': 2, 'device': 'cuda'}
╭────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────╮
│ Trial name                status         l1     l2            lr     batch_size     iter     total time (s)      loss     accuracy │
├────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────┤
│ train_cifar_ad99a_00007   RUNNING       128    128   0.0010942                2        7           424.28     1.40256       0.525  │
│ train_cifar_ad99a_00000   TERMINATED     32      1   0.0123941                2       10           615.167    2.31109       0.1017 │
│ train_cifar_ad99a_00001   TERMINATED      4      1   0.0101213               16       10            89.6874   2.30332       0.0997 │
│ train_cifar_ad99a_00002   TERMINATED    256      4   0.000710914              2       10           605.089    1.38372       0.5492 │
│ train_cifar_ad99a_00003   TERMINATED     16      1   0.00261551               8       10           165.841    1.77842       0.262  │
│ train_cifar_ad99a_00004   TERMINATED    256      1   0.000823347              4       10           321.175    1.77184       0.3073 │
│ train_cifar_ad99a_00005   TERMINATED      8     32   0.0264297                4        1            33.0529   2.31767       0.1003 │
│ train_cifar_ad99a_00006   TERMINATED    256    128   0.0848744               16        1            10.5968   2.19539       0.1804 │
│ train_cifar_ad99a_00008   PENDING         2      8   0.019285                16                                                    │
│ train_cifar_ad99a_00009   PENDING       128     32   0.00107603               4                                                    │
╰────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────╯
(func pid=12943) [8,  6000] loss: 0.415
(func pid=12943) [8,  8000] loss: 0.311
(func pid=12943) [8, 10000] loss: 0.252
(func pid=12943) [8, 12000] loss: 0.223
(func pid=12943) [8, 14000] loss: 0.182
(func pid=12943) [8, 16000] loss: 0.161
Trial status: 7 TERMINATED | 1 RUNNING | 2 PENDING
Current time: 2026-05-08 22:27:46. Total running time: 39min 4s
Logical resource usage: 2.0/16 CPUs, 1.0/1 GPUs (0.0/1.0 accelerator_type:A10G)
Current best trial: ad99a_00002 with loss=1.3837229575751349 and params={'l1': 256, 'l2': 4, 'lr': 0.0007109136804996362, 'batch_size': 2, 'device': 'cuda'}
╭────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────╮
│ Trial name                status         l1     l2            lr     batch_size     iter     total time (s)      loss     accuracy │
├────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────┤
│ train_cifar_ad99a_00007   RUNNING       128    128   0.0010942                2        7           424.28     1.40256       0.525  │
│ train_cifar_ad99a_00000   TERMINATED     32      1   0.0123941                2       10           615.167    2.31109       0.1017 │
│ train_cifar_ad99a_00001   TERMINATED      4      1   0.0101213               16       10            89.6874   2.30332       0.0997 │
│ train_cifar_ad99a_00002   TERMINATED    256      4   0.000710914              2       10           605.089    1.38372       0.5492 │
│ train_cifar_ad99a_00003   TERMINATED     16      1   0.00261551               8       10           165.841    1.77842       0.262  │
│ train_cifar_ad99a_00004   TERMINATED    256      1   0.000823347              4       10           321.175    1.77184       0.3073 │
│ train_cifar_ad99a_00005   TERMINATED      8     32   0.0264297                4        1            33.0529   2.31767       0.1003 │
│ train_cifar_ad99a_00006   TERMINATED    256    128   0.0848744               16        1            10.5968   2.19539       0.1804 │
│ train_cifar_ad99a_00008   PENDING         2      8   0.019285                16                                                    │
│ train_cifar_ad99a_00009   PENDING       128     32   0.00107603               4                                                    │
╰────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────╯
(func pid=12943) [8, 18000] loss: 0.144
(func pid=12943) [8, 20000] loss: 0.131
(func pid=12943) Checkpoint successfully created at: Checkpoint(filesystem=local, path=/var/lib/ci-user/ray_results/train_cifar_2026-05-08_21-48-41/train_cifar_ad99a_00007_7_batch_size=2,l1=128,l2=128,lr=0.0011_2026-05-08_21-48-41/checkpoint_000007)
(func pid=12943) [9,  2000] loss: 1.203
(func pid=12943) [9,  4000] loss: 0.624
Trial status: 7 TERMINATED | 1 RUNNING | 2 PENDING
Current time: 2026-05-08 22:28:16. Total running time: 39min 35s
Logical resource usage: 2.0/16 CPUs, 1.0/1 GPUs (0.0/1.0 accelerator_type:A10G)
Current best trial: ad99a_00002 with loss=1.3837229575751349 and params={'l1': 256, 'l2': 4, 'lr': 0.0007109136804996362, 'batch_size': 2, 'device': 'cuda'}
╭────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────╮
│ Trial name                status         l1     l2            lr     batch_size     iter     total time (s)      loss     accuracy │
├────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────┤
│ train_cifar_ad99a_00007   RUNNING       128    128   0.0010942                2        8           484.606    1.49033       0.5264 │
│ train_cifar_ad99a_00000   TERMINATED     32      1   0.0123941                2       10           615.167    2.31109       0.1017 │
│ train_cifar_ad99a_00001   TERMINATED      4      1   0.0101213               16       10            89.6874   2.30332       0.0997 │
│ train_cifar_ad99a_00002   TERMINATED    256      4   0.000710914              2       10           605.089    1.38372       0.5492 │
│ train_cifar_ad99a_00003   TERMINATED     16      1   0.00261551               8       10           165.841    1.77842       0.262  │
│ train_cifar_ad99a_00004   TERMINATED    256      1   0.000823347              4       10           321.175    1.77184       0.3073 │
│ train_cifar_ad99a_00005   TERMINATED      8     32   0.0264297                4        1            33.0529   2.31767       0.1003 │
│ train_cifar_ad99a_00006   TERMINATED    256    128   0.0848744               16        1            10.5968   2.19539       0.1804 │
│ train_cifar_ad99a_00008   PENDING         2      8   0.019285                16                                                    │
│ train_cifar_ad99a_00009   PENDING       128     32   0.00107603               4                                                    │
╰────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────╯
(func pid=12943) [9,  6000] loss: 0.406
(func pid=12943) [9,  8000] loss: 0.311
(func pid=12943) [9, 10000] loss: 0.252
(func pid=12943) [9, 12000] loss: 0.210
(func pid=12943) [9, 14000] loss: 0.179
(func pid=12943) [9, 16000] loss: 0.162
Trial status: 7 TERMINATED | 1 RUNNING | 2 PENDING
Current time: 2026-05-08 22:28:46. Total running time: 40min 5s
Logical resource usage: 2.0/16 CPUs, 1.0/1 GPUs (0.0/1.0 accelerator_type:A10G)
Current best trial: ad99a_00002 with loss=1.3837229575751349 and params={'l1': 256, 'l2': 4, 'lr': 0.0007109136804996362, 'batch_size': 2, 'device': 'cuda'}
╭────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────╮
│ Trial name                status         l1     l2            lr     batch_size     iter     total time (s)      loss     accuracy │
├────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────┤
│ train_cifar_ad99a_00007   RUNNING       128    128   0.0010942                2        8           484.606    1.49033       0.5264 │
│ train_cifar_ad99a_00000   TERMINATED     32      1   0.0123941                2       10           615.167    2.31109       0.1017 │
│ train_cifar_ad99a_00001   TERMINATED      4      1   0.0101213               16       10            89.6874   2.30332       0.0997 │
│ train_cifar_ad99a_00002   TERMINATED    256      4   0.000710914              2       10           605.089    1.38372       0.5492 │
│ train_cifar_ad99a_00003   TERMINATED     16      1   0.00261551               8       10           165.841    1.77842       0.262  │
│ train_cifar_ad99a_00004   TERMINATED    256      1   0.000823347              4       10           321.175    1.77184       0.3073 │
│ train_cifar_ad99a_00005   TERMINATED      8     32   0.0264297                4        1            33.0529   2.31767       0.1003 │
│ train_cifar_ad99a_00006   TERMINATED    256    128   0.0848744               16        1            10.5968   2.19539       0.1804 │
│ train_cifar_ad99a_00008   PENDING         2      8   0.019285                16                                                    │
│ train_cifar_ad99a_00009   PENDING       128     32   0.00107603               4                                                    │
╰────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────╯
(func pid=12943) [9, 18000] loss: 0.142
(func pid=12943) [9, 20000] loss: 0.133
(func pid=12943) Checkpoint successfully created at: Checkpoint(filesystem=local, path=/var/lib/ci-user/ray_results/train_cifar_2026-05-08_21-48-41/train_cifar_ad99a_00007_7_batch_size=2,l1=128,l2=128,lr=0.0011_2026-05-08_21-48-41/checkpoint_000008)
(func pid=12943) [10,  2000] loss: 1.173
(func pid=12943) [10,  4000] loss: 0.636
Trial status: 7 TERMINATED | 1 RUNNING | 2 PENDING
Current time: 2026-05-08 22:29:16. Total running time: 40min 35s
Logical resource usage: 2.0/16 CPUs, 1.0/1 GPUs (0.0/1.0 accelerator_type:A10G)
Current best trial: ad99a_00002 with loss=1.3837229575751349 and params={'l1': 256, 'l2': 4, 'lr': 0.0007109136804996362, 'batch_size': 2, 'device': 'cuda'}
╭────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────╮
│ Trial name                status         l1     l2            lr     batch_size     iter     total time (s)      loss     accuracy │
├────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────┤
│ train_cifar_ad99a_00007   RUNNING       128    128   0.0010942                2        9           545.24     1.43734       0.5111 │
│ train_cifar_ad99a_00000   TERMINATED     32      1   0.0123941                2       10           615.167    2.31109       0.1017 │
│ train_cifar_ad99a_00001   TERMINATED      4      1   0.0101213               16       10            89.6874   2.30332       0.0997 │
│ train_cifar_ad99a_00002   TERMINATED    256      4   0.000710914              2       10           605.089    1.38372       0.5492 │
│ train_cifar_ad99a_00003   TERMINATED     16      1   0.00261551               8       10           165.841    1.77842       0.262  │
│ train_cifar_ad99a_00004   TERMINATED    256      1   0.000823347              4       10           321.175    1.77184       0.3073 │
│ train_cifar_ad99a_00005   TERMINATED      8     32   0.0264297                4        1            33.0529   2.31767       0.1003 │
│ train_cifar_ad99a_00006   TERMINATED    256    128   0.0848744               16        1            10.5968   2.19539       0.1804 │
│ train_cifar_ad99a_00008   PENDING         2      8   0.019285                16                                                    │
│ train_cifar_ad99a_00009   PENDING       128     32   0.00107603               4                                                    │
╰────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────╯
(func pid=12943) [10,  6000] loss: 0.416
(func pid=12943) [10,  8000] loss: 0.319
(func pid=12943) [10, 10000] loss: 0.247
(func pid=12943) [10, 12000] loss: 0.199
(func pid=12943) [10, 14000] loss: 0.184
(func pid=12943) [10, 16000] loss: 0.164
Trial status: 7 TERMINATED | 1 RUNNING | 2 PENDING
Current time: 2026-05-08 22:29:46. Total running time: 41min 5s
Logical resource usage: 2.0/16 CPUs, 1.0/1 GPUs (0.0/1.0 accelerator_type:A10G)
Current best trial: ad99a_00002 with loss=1.3837229575751349 and params={'l1': 256, 'l2': 4, 'lr': 0.0007109136804996362, 'batch_size': 2, 'device': 'cuda'}
╭────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────╮
│ Trial name                status         l1     l2            lr     batch_size     iter     total time (s)      loss     accuracy │
├────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────┤
│ train_cifar_ad99a_00007   RUNNING       128    128   0.0010942                2        9           545.24     1.43734       0.5111 │
│ train_cifar_ad99a_00000   TERMINATED     32      1   0.0123941                2       10           615.167    2.31109       0.1017 │
│ train_cifar_ad99a_00001   TERMINATED      4      1   0.0101213               16       10            89.6874   2.30332       0.0997 │
│ train_cifar_ad99a_00002   TERMINATED    256      4   0.000710914              2       10           605.089    1.38372       0.5492 │
│ train_cifar_ad99a_00003   TERMINATED     16      1   0.00261551               8       10           165.841    1.77842       0.262  │
│ train_cifar_ad99a_00004   TERMINATED    256      1   0.000823347              4       10           321.175    1.77184       0.3073 │
│ train_cifar_ad99a_00005   TERMINATED      8     32   0.0264297                4        1            33.0529   2.31767       0.1003 │
│ train_cifar_ad99a_00006   TERMINATED    256    128   0.0848744               16        1            10.5968   2.19539       0.1804 │
│ train_cifar_ad99a_00008   PENDING         2      8   0.019285                16                                                    │
│ train_cifar_ad99a_00009   PENDING       128     32   0.00107603               4                                                    │
╰────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────╯
(func pid=12943) [10, 18000] loss: 0.145
(func pid=12943) [10, 20000] loss: 0.127

Trial train_cifar_ad99a_00007 completed after 10 iterations at 2026-05-08 22:30:05. Total running time: 41min 23s
╭────────────────────────────────────────────────────────────╮
│ Trial train_cifar_ad99a_00007 result                       │
├────────────────────────────────────────────────────────────┤
│ checkpoint_dir_name                      checkpoint_000009 │
│ time_this_iter_s                                  60.81447 │
│ time_total_s                                     606.05492 │
│ training_iteration                                      10 │
│ accuracy                                            0.5369 │
│ loss                                               1.38719 │
╰────────────────────────────────────────────────────────────╯
(func pid=12943) Checkpoint successfully created at: Checkpoint(filesystem=local, path=/var/lib/ci-user/ray_results/train_cifar_2026-05-08_21-48-41/train_cifar_ad99a_00007_7_batch_size=2,l1=128,l2=128,lr=0.0011_2026-05-08_21-48-41/checkpoint_000009)

Trial train_cifar_ad99a_00008 started with configuration:
╭──────────────────────────────────────────────────╮
│ Trial train_cifar_ad99a_00008 config             │
├──────────────────────────────────────────────────┤
│ batch_size                                    16 │
│ device                                      cuda │
│ l1                                             2 │
│ l2                                             8 │
│ lr                                       0.01929 │
╰──────────────────────────────────────────────────╯

Trial status: 8 TERMINATED | 1 RUNNING | 1 PENDING
Current time: 2026-05-08 22:30:16. Total running time: 41min 35s
Logical resource usage: 2.0/16 CPUs, 1.0/1 GPUs (0.0/1.0 accelerator_type:A10G)
Current best trial: ad99a_00002 with loss=1.3837229575751349 and params={'l1': 256, 'l2': 4, 'lr': 0.0007109136804996362, 'batch_size': 2, 'device': 'cuda'}
╭────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────╮
│ Trial name                status         l1     l2            lr     batch_size     iter     total time (s)      loss     accuracy │
├────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────┤
│ train_cifar_ad99a_00008   RUNNING         2      8   0.019285                16                                                    │
│ train_cifar_ad99a_00000   TERMINATED     32      1   0.0123941                2       10           615.167    2.31109       0.1017 │
│ train_cifar_ad99a_00001   TERMINATED      4      1   0.0101213               16       10            89.6874   2.30332       0.0997 │
│ train_cifar_ad99a_00002   TERMINATED    256      4   0.000710914              2       10           605.089    1.38372       0.5492 │
│ train_cifar_ad99a_00003   TERMINATED     16      1   0.00261551               8       10           165.841    1.77842       0.262  │
│ train_cifar_ad99a_00004   TERMINATED    256      1   0.000823347              4       10           321.175    1.77184       0.3073 │
│ train_cifar_ad99a_00005   TERMINATED      8     32   0.0264297                4        1            33.0529   2.31767       0.1003 │
│ train_cifar_ad99a_00006   TERMINATED    256    128   0.0848744               16        1            10.5968   2.19539       0.1804 │
│ train_cifar_ad99a_00007   TERMINATED    128    128   0.0010942                2       10           606.055    1.38719       0.5369 │
│ train_cifar_ad99a_00009   PENDING       128     32   0.00107603               4                                                    │
╰────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────╯
(func pid=14809) [1,  2000] loss: 2.220

Trial train_cifar_ad99a_00008 completed after 1 iterations at 2026-05-08 22:30:21. Total running time: 41min 39s
╭────────────────────────────────────────────────────────────╮
│ Trial train_cifar_ad99a_00008 result                       │
├────────────────────────────────────────────────────────────┤
│ checkpoint_dir_name                      checkpoint_000000 │
│ time_this_iter_s                                    10.781 │
│ time_total_s                                        10.781 │
│ training_iteration                                       1 │
│ accuracy                                            0.1068 │
│ loss                                                2.3047 │
╰────────────────────────────────────────────────────────────╯
(func pid=14809) Checkpoint successfully created at: Checkpoint(filesystem=local, path=/var/lib/ci-user/ray_results/train_cifar_2026-05-08_21-48-41/train_cifar_ad99a_00008_8_batch_size=16,l1=2,l2=8,lr=0.0193_2026-05-08_21-48-41/checkpoint_000000)

Trial train_cifar_ad99a_00009 started with configuration:
╭──────────────────────────────────────────────────╮
│ Trial train_cifar_ad99a_00009 config             │
├──────────────────────────────────────────────────┤
│ batch_size                                     4 │
│ device                                      cuda │
│ l1                                           128 │
│ l2                                            32 │
│ lr                                       0.00108 │
╰──────────────────────────────────────────────────╯
(func pid=14962) [1,  2000] loss: 2.097
(func pid=14962) [1,  4000] loss: 0.869
(func pid=14962) [1,  6000] loss: 0.531

Trial status: 9 TERMINATED | 1 RUNNING
Current time: 2026-05-08 22:30:46. Total running time: 42min 5s
Logical resource usage: 2.0/16 CPUs, 1.0/1 GPUs (0.0/1.0 accelerator_type:A10G)
Current best trial: ad99a_00002 with loss=1.3837229575751349 and params={'l1': 256, 'l2': 4, 'lr': 0.0007109136804996362, 'batch_size': 2, 'device': 'cuda'}
╭────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────╮
│ Trial name                status         l1     l2            lr     batch_size     iter     total time (s)      loss     accuracy │
├────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────┤
│ train_cifar_ad99a_00009   RUNNING       128     32   0.00107603               4                                                    │
│ train_cifar_ad99a_00000   TERMINATED     32      1   0.0123941                2       10           615.167    2.31109       0.1017 │
│ train_cifar_ad99a_00001   TERMINATED      4      1   0.0101213               16       10            89.6874   2.30332       0.0997 │
│ train_cifar_ad99a_00002   TERMINATED    256      4   0.000710914              2       10           605.089    1.38372       0.5492 │
│ train_cifar_ad99a_00003   TERMINATED     16      1   0.00261551               8       10           165.841    1.77842       0.262  │
│ train_cifar_ad99a_00004   TERMINATED    256      1   0.000823347              4       10           321.175    1.77184       0.3073 │
│ train_cifar_ad99a_00005   TERMINATED      8     32   0.0264297                4        1            33.0529   2.31767       0.1003 │
│ train_cifar_ad99a_00006   TERMINATED    256    128   0.0848744               16        1            10.5968   2.19539       0.1804 │
│ train_cifar_ad99a_00007   TERMINATED    128    128   0.0010942                2       10           606.055    1.38719       0.5369 │
│ train_cifar_ad99a_00008   TERMINATED      2      8   0.019285                16        1            10.781    2.3047        0.1068 │
╰────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────╯
(func pid=14962) [1,  8000] loss: 0.382
(func pid=14962) [1, 10000] loss: 0.298
(func pid=14962) Checkpoint successfully created at: Checkpoint(filesystem=local, path=/var/lib/ci-user/ray_results/train_cifar_2026-05-08_21-48-41/train_cifar_ad99a_00009_9_batch_size=4,l1=128,l2=32,lr=0.0011_2026-05-08_21-48-41/checkpoint_000000)
(func pid=14962) [2,  2000] loss: 1.394
(func pid=14962) [2,  4000] loss: 0.692
(func pid=14962) [2,  6000] loss: 0.462
Trial status: 9 TERMINATED | 1 RUNNING
Current time: 2026-05-08 22:31:16. Total running time: 42min 35s
Logical resource usage: 2.0/16 CPUs, 1.0/1 GPUs (0.0/1.0 accelerator_type:A10G)
Current best trial: ad99a_00002 with loss=1.3837229575751349 and params={'l1': 256, 'l2': 4, 'lr': 0.0007109136804996362, 'batch_size': 2, 'device': 'cuda'}
╭────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────╮
│ Trial name                status         l1     l2            lr     batch_size     iter     total time (s)      loss     accuracy │
├────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────┤
│ train_cifar_ad99a_00009   RUNNING       128     32   0.00107603               4        1            32.532    1.45985       0.4632 │
│ train_cifar_ad99a_00000   TERMINATED     32      1   0.0123941                2       10           615.167    2.31109       0.1017 │
│ train_cifar_ad99a_00001   TERMINATED      4      1   0.0101213               16       10            89.6874   2.30332       0.0997 │
│ train_cifar_ad99a_00002   TERMINATED    256      4   0.000710914              2       10           605.089    1.38372       0.5492 │
│ train_cifar_ad99a_00003   TERMINATED     16      1   0.00261551               8       10           165.841    1.77842       0.262  │
│ train_cifar_ad99a_00004   TERMINATED    256      1   0.000823347              4       10           321.175    1.77184       0.3073 │
│ train_cifar_ad99a_00005   TERMINATED      8     32   0.0264297                4        1            33.0529   2.31767       0.1003 │
│ train_cifar_ad99a_00006   TERMINATED    256    128   0.0848744               16        1            10.5968   2.19539       0.1804 │
│ train_cifar_ad99a_00007   TERMINATED    128    128   0.0010942                2       10           606.055    1.38719       0.5369 │
│ train_cifar_ad99a_00008   TERMINATED      2      8   0.019285                16        1            10.781    2.3047        0.1068 │
╰────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────╯
(func pid=14962) [2,  8000] loss: 0.340
(func pid=14962) [2, 10000] loss: 0.264
(func pid=14962) Checkpoint successfully created at: Checkpoint(filesystem=local, path=/var/lib/ci-user/ray_results/train_cifar_2026-05-08_21-48-41/train_cifar_ad99a_00009_9_batch_size=4,l1=128,l2=32,lr=0.0011_2026-05-08_21-48-41/checkpoint_000001)
(func pid=14962) [3,  2000] loss: 1.255
(func pid=14962) [3,  4000] loss: 0.624
(func pid=14962) [3,  6000] loss: 0.415
Trial status: 9 TERMINATED | 1 RUNNING
Current time: 2026-05-08 22:31:46. Total running time: 43min 5s
Logical resource usage: 2.0/16 CPUs, 1.0/1 GPUs (0.0/1.0 accelerator_type:A10G)
Current best trial: ad99a_00009 with loss=1.3064815080732106 and params={'l1': 128, 'l2': 32, 'lr': 0.001076028094323717, 'batch_size': 4, 'device': 'cuda'}
╭────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────╮
│ Trial name                status         l1     l2            lr     batch_size     iter     total time (s)      loss     accuracy │
├────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────┤
│ train_cifar_ad99a_00009   RUNNING       128     32   0.00107603               4        2            63.6578   1.30648       0.5449 │
│ train_cifar_ad99a_00000   TERMINATED     32      1   0.0123941                2       10           615.167    2.31109       0.1017 │
│ train_cifar_ad99a_00001   TERMINATED      4      1   0.0101213               16       10            89.6874   2.30332       0.0997 │
│ train_cifar_ad99a_00002   TERMINATED    256      4   0.000710914              2       10           605.089    1.38372       0.5492 │
│ train_cifar_ad99a_00003   TERMINATED     16      1   0.00261551               8       10           165.841    1.77842       0.262  │
│ train_cifar_ad99a_00004   TERMINATED    256      1   0.000823347              4       10           321.175    1.77184       0.3073 │
│ train_cifar_ad99a_00005   TERMINATED      8     32   0.0264297                4        1            33.0529   2.31767       0.1003 │
│ train_cifar_ad99a_00006   TERMINATED    256    128   0.0848744               16        1            10.5968   2.19539       0.1804 │
│ train_cifar_ad99a_00007   TERMINATED    128    128   0.0010942                2       10           606.055    1.38719       0.5369 │
│ train_cifar_ad99a_00008   TERMINATED      2      8   0.019285                16        1            10.781    2.3047        0.1068 │
╰────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────╯
(func pid=14962) [3,  8000] loss: 0.315
(func pid=14962) [3, 10000] loss: 0.248
(func pid=14962) Checkpoint successfully created at: Checkpoint(filesystem=local, path=/var/lib/ci-user/ray_results/train_cifar_2026-05-08_21-48-41/train_cifar_ad99a_00009_9_batch_size=4,l1=128,l2=32,lr=0.0011_2026-05-08_21-48-41/checkpoint_000002)
(func pid=14962) [4,  2000] loss: 1.151
(func pid=14962) [4,  4000] loss: 0.586
(func pid=14962) [4,  6000] loss: 0.388
Trial status: 9 TERMINATED | 1 RUNNING
Current time: 2026-05-08 22:32:16. Total running time: 43min 35s
Logical resource usage: 2.0/16 CPUs, 1.0/1 GPUs (0.0/1.0 accelerator_type:A10G)
Current best trial: ad99a_00009 with loss=1.2515093033343554 and params={'l1': 128, 'l2': 32, 'lr': 0.001076028094323717, 'batch_size': 4, 'device': 'cuda'}
╭────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────╮
│ Trial name                status         l1     l2            lr     batch_size     iter     total time (s)      loss     accuracy │
├────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────┤
│ train_cifar_ad99a_00009   RUNNING       128     32   0.00107603               4        3            94.6531   1.25151       0.559  │
│ train_cifar_ad99a_00000   TERMINATED     32      1   0.0123941                2       10           615.167    2.31109       0.1017 │
│ train_cifar_ad99a_00001   TERMINATED      4      1   0.0101213               16       10            89.6874   2.30332       0.0997 │
│ train_cifar_ad99a_00002   TERMINATED    256      4   0.000710914              2       10           605.089    1.38372       0.5492 │
│ train_cifar_ad99a_00003   TERMINATED     16      1   0.00261551               8       10           165.841    1.77842       0.262  │
│ train_cifar_ad99a_00004   TERMINATED    256      1   0.000823347              4       10           321.175    1.77184       0.3073 │
│ train_cifar_ad99a_00005   TERMINATED      8     32   0.0264297                4        1            33.0529   2.31767       0.1003 │
│ train_cifar_ad99a_00006   TERMINATED    256    128   0.0848744               16        1            10.5968   2.19539       0.1804 │
│ train_cifar_ad99a_00007   TERMINATED    128    128   0.0010942                2       10           606.055    1.38719       0.5369 │
│ train_cifar_ad99a_00008   TERMINATED      2      8   0.019285                16        1            10.781    2.3047        0.1068 │
╰────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────╯
(func pid=14962) [4,  8000] loss: 0.296
(func pid=14962) [4, 10000] loss: 0.235
(func pid=14962) Checkpoint successfully created at: Checkpoint(filesystem=local, path=/var/lib/ci-user/ray_results/train_cifar_2026-05-08_21-48-41/train_cifar_ad99a_00009_9_batch_size=4,l1=128,l2=32,lr=0.0011_2026-05-08_21-48-41/checkpoint_000003)
(func pid=14962) [5,  2000] loss: 1.081
(func pid=14962) [5,  4000] loss: 0.556
Trial status: 9 TERMINATED | 1 RUNNING
Current time: 2026-05-08 22:32:46. Total running time: 44min 5s
Logical resource usage: 2.0/16 CPUs, 1.0/1 GPUs (0.0/1.0 accelerator_type:A10G)
Current best trial: ad99a_00009 with loss=1.276932258567214 and params={'l1': 128, 'l2': 32, 'lr': 0.001076028094323717, 'batch_size': 4, 'device': 'cuda'}
╭────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────╮
│ Trial name                status         l1     l2            lr     batch_size     iter     total time (s)      loss     accuracy │
├────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────┤
│ train_cifar_ad99a_00009   RUNNING       128     32   0.00107603               4        4           125.601    1.27693       0.5479 │
│ train_cifar_ad99a_00000   TERMINATED     32      1   0.0123941                2       10           615.167    2.31109       0.1017 │
│ train_cifar_ad99a_00001   TERMINATED      4      1   0.0101213               16       10            89.6874   2.30332       0.0997 │
│ train_cifar_ad99a_00002   TERMINATED    256      4   0.000710914              2       10           605.089    1.38372       0.5492 │
│ train_cifar_ad99a_00003   TERMINATED     16      1   0.00261551               8       10           165.841    1.77842       0.262  │
│ train_cifar_ad99a_00004   TERMINATED    256      1   0.000823347              4       10           321.175    1.77184       0.3073 │
│ train_cifar_ad99a_00005   TERMINATED      8     32   0.0264297                4        1            33.0529   2.31767       0.1003 │
│ train_cifar_ad99a_00006   TERMINATED    256    128   0.0848744               16        1            10.5968   2.19539       0.1804 │
│ train_cifar_ad99a_00007   TERMINATED    128    128   0.0010942                2       10           606.055    1.38719       0.5369 │
│ train_cifar_ad99a_00008   TERMINATED      2      8   0.019285                16        1            10.781    2.3047        0.1068 │
╰────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────╯
(func pid=14962) [5,  6000] loss: 0.371
(func pid=14962) [5,  8000] loss: 0.282
(func pid=14962) [5, 10000] loss: 0.222
(func pid=14962) Checkpoint successfully created at: Checkpoint(filesystem=local, path=/var/lib/ci-user/ray_results/train_cifar_2026-05-08_21-48-41/train_cifar_ad99a_00009_9_batch_size=4,l1=128,l2=32,lr=0.0011_2026-05-08_21-48-41/checkpoint_000004)
(func pid=14962) [6,  2000] loss: 1.021
(func pid=14962) [6,  4000] loss: 0.515
Trial status: 9 TERMINATED | 1 RUNNING
Current time: 2026-05-08 22:33:17. Total running time: 44min 35s
Logical resource usage: 2.0/16 CPUs, 1.0/1 GPUs (0.0/1.0 accelerator_type:A10G)
Current best trial: ad99a_00009 with loss=1.1884577242083847 and params={'l1': 128, 'l2': 32, 'lr': 0.001076028094323717, 'batch_size': 4, 'device': 'cuda'}
╭────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────╮
│ Trial name                status         l1     l2            lr     batch_size     iter     total time (s)      loss     accuracy │
├────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────┤
│ train_cifar_ad99a_00009   RUNNING       128     32   0.00107603               4        5           156.483    1.18846       0.5877 │
│ train_cifar_ad99a_00000   TERMINATED     32      1   0.0123941                2       10           615.167    2.31109       0.1017 │
│ train_cifar_ad99a_00001   TERMINATED      4      1   0.0101213               16       10            89.6874   2.30332       0.0997 │
│ train_cifar_ad99a_00002   TERMINATED    256      4   0.000710914              2       10           605.089    1.38372       0.5492 │
│ train_cifar_ad99a_00003   TERMINATED     16      1   0.00261551               8       10           165.841    1.77842       0.262  │
│ train_cifar_ad99a_00004   TERMINATED    256      1   0.000823347              4       10           321.175    1.77184       0.3073 │
│ train_cifar_ad99a_00005   TERMINATED      8     32   0.0264297                4        1            33.0529   2.31767       0.1003 │
│ train_cifar_ad99a_00006   TERMINATED    256    128   0.0848744               16        1            10.5968   2.19539       0.1804 │
│ train_cifar_ad99a_00007   TERMINATED    128    128   0.0010942                2       10           606.055    1.38719       0.5369 │
│ train_cifar_ad99a_00008   TERMINATED      2      8   0.019285                16        1            10.781    2.3047        0.1068 │
╰────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────╯
(func pid=14962) [6,  6000] loss: 0.353
(func pid=14962) [6,  8000] loss: 0.274
(func pid=14962) [6, 10000] loss: 0.219
(func pid=14962) Checkpoint successfully created at: Checkpoint(filesystem=local, path=/var/lib/ci-user/ray_results/train_cifar_2026-05-08_21-48-41/train_cifar_ad99a_00009_9_batch_size=4,l1=128,l2=32,lr=0.0011_2026-05-08_21-48-41/checkpoint_000005)
(func pid=14962) [7,  2000] loss: 0.985
(func pid=14962) [7,  4000] loss: 0.501
Trial status: 9 TERMINATED | 1 RUNNING
Current time: 2026-05-08 22:33:47. Total running time: 45min 5s
Logical resource usage: 2.0/16 CPUs, 1.0/1 GPUs (0.0/1.0 accelerator_type:A10G)
Current best trial: ad99a_00009 with loss=1.2195294665135443 and params={'l1': 128, 'l2': 32, 'lr': 0.001076028094323717, 'batch_size': 4, 'device': 'cuda'}
╭────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────╮
│ Trial name                status         l1     l2            lr     batch_size     iter     total time (s)      loss     accuracy │
├────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────┤
│ train_cifar_ad99a_00009   RUNNING       128     32   0.00107603               4        6           187.272    1.21953       0.5871 │
│ train_cifar_ad99a_00000   TERMINATED     32      1   0.0123941                2       10           615.167    2.31109       0.1017 │
│ train_cifar_ad99a_00001   TERMINATED      4      1   0.0101213               16       10            89.6874   2.30332       0.0997 │
│ train_cifar_ad99a_00002   TERMINATED    256      4   0.000710914              2       10           605.089    1.38372       0.5492 │
│ train_cifar_ad99a_00003   TERMINATED     16      1   0.00261551               8       10           165.841    1.77842       0.262  │
│ train_cifar_ad99a_00004   TERMINATED    256      1   0.000823347              4       10           321.175    1.77184       0.3073 │
│ train_cifar_ad99a_00005   TERMINATED      8     32   0.0264297                4        1            33.0529   2.31767       0.1003 │
│ train_cifar_ad99a_00006   TERMINATED    256    128   0.0848744               16        1            10.5968   2.19539       0.1804 │
│ train_cifar_ad99a_00007   TERMINATED    128    128   0.0010942                2       10           606.055    1.38719       0.5369 │
│ train_cifar_ad99a_00008   TERMINATED      2      8   0.019285                16        1            10.781    2.3047        0.1068 │
╰────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────╯
(func pid=14962) [7,  6000] loss: 0.341
(func pid=14962) [7,  8000] loss: 0.256
(func pid=14962) [7, 10000] loss: 0.208
(func pid=14962) Checkpoint successfully created at: Checkpoint(filesystem=local, path=/var/lib/ci-user/ray_results/train_cifar_2026-05-08_21-48-41/train_cifar_ad99a_00009_9_batch_size=4,l1=128,l2=32,lr=0.0011_2026-05-08_21-48-41/checkpoint_000006)
(func pid=14962) [8,  2000] loss: 0.924
(func pid=14962) [8,  4000] loss: 0.480
Trial status: 9 TERMINATED | 1 RUNNING
Current time: 2026-05-08 22:34:17. Total running time: 45min 35s
Logical resource usage: 2.0/16 CPUs, 1.0/1 GPUs (0.0/1.0 accelerator_type:A10G)
Current best trial: ad99a_00009 with loss=1.162755216448754 and params={'l1': 128, 'l2': 32, 'lr': 0.001076028094323717, 'batch_size': 4, 'device': 'cuda'}
╭────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────╮
│ Trial name                status         l1     l2            lr     batch_size     iter     total time (s)      loss     accuracy │
├────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────┤
│ train_cifar_ad99a_00009   RUNNING       128     32   0.00107603               4        7           217.892    1.16276       0.6105 │
│ train_cifar_ad99a_00000   TERMINATED     32      1   0.0123941                2       10           615.167    2.31109       0.1017 │
│ train_cifar_ad99a_00001   TERMINATED      4      1   0.0101213               16       10            89.6874   2.30332       0.0997 │
│ train_cifar_ad99a_00002   TERMINATED    256      4   0.000710914              2       10           605.089    1.38372       0.5492 │
│ train_cifar_ad99a_00003   TERMINATED     16      1   0.00261551               8       10           165.841    1.77842       0.262  │
│ train_cifar_ad99a_00004   TERMINATED    256      1   0.000823347              4       10           321.175    1.77184       0.3073 │
│ train_cifar_ad99a_00005   TERMINATED      8     32   0.0264297                4        1            33.0529   2.31767       0.1003 │
│ train_cifar_ad99a_00006   TERMINATED    256    128   0.0848744               16        1            10.5968   2.19539       0.1804 │
│ train_cifar_ad99a_00007   TERMINATED    128    128   0.0010942                2       10           606.055    1.38719       0.5369 │
│ train_cifar_ad99a_00008   TERMINATED      2      8   0.019285                16        1            10.781    2.3047        0.1068 │
╰────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────╯
(func pid=14962) [8,  6000] loss: 0.328
(func pid=14962) [8,  8000] loss: 0.251
(func pid=14962) [8, 10000] loss: 0.200
(func pid=14962) Checkpoint successfully created at: Checkpoint(filesystem=local, path=/var/lib/ci-user/ray_results/train_cifar_2026-05-08_21-48-41/train_cifar_ad99a_00009_9_batch_size=4,l1=128,l2=32,lr=0.0011_2026-05-08_21-48-41/checkpoint_000007)
(func pid=14962) [9,  2000] loss: 0.902
(func pid=14962) [9,  4000] loss: 0.474
Trial status: 9 TERMINATED | 1 RUNNING
Current time: 2026-05-08 22:34:47. Total running time: 46min 5s
Logical resource usage: 2.0/16 CPUs, 1.0/1 GPUs (0.0/1.0 accelerator_type:A10G)
Current best trial: ad99a_00009 with loss=1.1954964253924787 and params={'l1': 128, 'l2': 32, 'lr': 0.001076028094323717, 'batch_size': 4, 'device': 'cuda'}
╭────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────╮
│ Trial name                status         l1     l2            lr     batch_size     iter     total time (s)      loss     accuracy │
├────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────┤
│ train_cifar_ad99a_00009   RUNNING       128     32   0.00107603               4        8           248.793    1.1955        0.6022 │
│ train_cifar_ad99a_00000   TERMINATED     32      1   0.0123941                2       10           615.167    2.31109       0.1017 │
│ train_cifar_ad99a_00001   TERMINATED      4      1   0.0101213               16       10            89.6874   2.30332       0.0997 │
│ train_cifar_ad99a_00002   TERMINATED    256      4   0.000710914              2       10           605.089    1.38372       0.5492 │
│ train_cifar_ad99a_00003   TERMINATED     16      1   0.00261551               8       10           165.841    1.77842       0.262  │
│ train_cifar_ad99a_00004   TERMINATED    256      1   0.000823347              4       10           321.175    1.77184       0.3073 │
│ train_cifar_ad99a_00005   TERMINATED      8     32   0.0264297                4        1            33.0529   2.31767       0.1003 │
│ train_cifar_ad99a_00006   TERMINATED    256    128   0.0848744               16        1            10.5968   2.19539       0.1804 │
│ train_cifar_ad99a_00007   TERMINATED    128    128   0.0010942                2       10           606.055    1.38719       0.5369 │
│ train_cifar_ad99a_00008   TERMINATED      2      8   0.019285                16        1            10.781    2.3047        0.1068 │
╰────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────╯
(func pid=14962) [9,  6000] loss: 0.320
(func pid=14962) [9,  8000] loss: 0.238
(func pid=14962) [9, 10000] loss: 0.195
(func pid=14962) Checkpoint successfully created at: Checkpoint(filesystem=local, path=/var/lib/ci-user/ray_results/train_cifar_2026-05-08_21-48-41/train_cifar_ad99a_00009_9_batch_size=4,l1=128,l2=32,lr=0.0011_2026-05-08_21-48-41/checkpoint_000008)
(func pid=14962) [10,  2000] loss: 0.885
(func pid=14962) [10,  4000] loss: 0.451
Trial status: 9 TERMINATED | 1 RUNNING
Current time: 2026-05-08 22:35:17. Total running time: 46min 35s
Logical resource usage: 2.0/16 CPUs, 1.0/1 GPUs (0.0/1.0 accelerator_type:A10G)
Current best trial: ad99a_00009 with loss=1.2090966225519777 and params={'l1': 128, 'l2': 32, 'lr': 0.001076028094323717, 'batch_size': 4, 'device': 'cuda'}
╭────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────╮
│ Trial name                status         l1     l2            lr     batch_size     iter     total time (s)      loss     accuracy │
├────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────┤
│ train_cifar_ad99a_00009   RUNNING       128     32   0.00107603               4        9           279.822    1.2091        0.5995 │
│ train_cifar_ad99a_00000   TERMINATED     32      1   0.0123941                2       10           615.167    2.31109       0.1017 │
│ train_cifar_ad99a_00001   TERMINATED      4      1   0.0101213               16       10            89.6874   2.30332       0.0997 │
│ train_cifar_ad99a_00002   TERMINATED    256      4   0.000710914              2       10           605.089    1.38372       0.5492 │
│ train_cifar_ad99a_00003   TERMINATED     16      1   0.00261551               8       10           165.841    1.77842       0.262  │
│ train_cifar_ad99a_00004   TERMINATED    256      1   0.000823347              4       10           321.175    1.77184       0.3073 │
│ train_cifar_ad99a_00005   TERMINATED      8     32   0.0264297                4        1            33.0529   2.31767       0.1003 │
│ train_cifar_ad99a_00006   TERMINATED    256    128   0.0848744               16        1            10.5968   2.19539       0.1804 │
│ train_cifar_ad99a_00007   TERMINATED    128    128   0.0010942                2       10           606.055    1.38719       0.5369 │
│ train_cifar_ad99a_00008   TERMINATED      2      8   0.019285                16        1            10.781    2.3047        0.1068 │
╰────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────╯
(func pid=14962) [10,  6000] loss: 0.306
(func pid=14962) [10,  8000] loss: 0.235
(func pid=14962) [10, 10000] loss: 0.190

Trial train_cifar_ad99a_00009 completed after 10 iterations at 2026-05-08 22:35:36. Total running time: 46min 55s
╭────────────────────────────────────────────────────────────╮
│ Trial train_cifar_ad99a_00009 result                       │
├────────────────────────────────────────────────────────────┤
│ checkpoint_dir_name                      checkpoint_000009 │
│ time_this_iter_s                                  30.73978 │
│ time_total_s                                     310.56216 │
│ training_iteration                                      10 │
│ accuracy                                            0.5976 │
│ loss                                               1.23208 │
╰────────────────────────────────────────────────────────────╯
2026-05-08 22:35:36,689 INFO tune.py:1001 -- Wrote the latest version of all result files and experiment state to '/var/lib/ci-user/ray_results/train_cifar_2026-05-08_21-48-41' in 0.0105s.

Trial status: 10 TERMINATED
Current time: 2026-05-08 22:35:36. Total running time: 46min 55s
Logical resource usage: 2.0/16 CPUs, 1.0/1 GPUs (0.0/1.0 accelerator_type:A10G)
Current best trial: ad99a_00009 with loss=1.2320763322196902 and params={'l1': 128, 'l2': 32, 'lr': 0.001076028094323717, 'batch_size': 4, 'device': 'cuda'}
╭────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────╮
│ Trial name                status         l1     l2            lr     batch_size     iter     total time (s)      loss     accuracy │
├────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────┤
│ train_cifar_ad99a_00000   TERMINATED     32      1   0.0123941                2       10           615.167    2.31109       0.1017 │
│ train_cifar_ad99a_00001   TERMINATED      4      1   0.0101213               16       10            89.6874   2.30332       0.0997 │
│ train_cifar_ad99a_00002   TERMINATED    256      4   0.000710914              2       10           605.089    1.38372       0.5492 │
│ train_cifar_ad99a_00003   TERMINATED     16      1   0.00261551               8       10           165.841    1.77842       0.262  │
│ train_cifar_ad99a_00004   TERMINATED    256      1   0.000823347              4       10           321.175    1.77184       0.3073 │
│ train_cifar_ad99a_00005   TERMINATED      8     32   0.0264297                4        1            33.0529   2.31767       0.1003 │
│ train_cifar_ad99a_00006   TERMINATED    256    128   0.0848744               16        1            10.5968   2.19539       0.1804 │
│ train_cifar_ad99a_00007   TERMINATED    128    128   0.0010942                2       10           606.055    1.38719       0.5369 │
│ train_cifar_ad99a_00008   TERMINATED      2      8   0.019285                16        1            10.781    2.3047        0.1068 │
│ train_cifar_ad99a_00009   TERMINATED    128     32   0.00107603               4       10           310.562    1.23208       0.5976 │
╰────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────╯

Best trial config: {'l1': 128, 'l2': 32, 'lr': 0.001076028094323717, 'batch_size': 4, 'device': 'cuda'}
Best trial final validation loss: 1.2320763322196902
Best trial final validation accuracy: 0.5976
(func pid=14962) Checkpoint successfully created at: Checkpoint(filesystem=local, path=/var/lib/ci-user/ray_results/train_cifar_2026-05-08_21-48-41/train_cifar_ad99a_00009_9_batch_size=4,l1=128,l2=32,lr=0.0011_2026-05-08_21-48-41/checkpoint_000009)
Best trial test set accuracy: 0.5975

Results#

Your Ray Tune trial summary output looks something like this. The text table summarizes the validation performance of the trials and highlights the best hyperparameter configuration:

Number of trials: 10/10 (10 TERMINATED)
+-----+--------------+------+------+-------------+--------+---------+------------+
| ... |   batch_size |   l1 |   l2 |          lr |   iter |    loss |   accuracy |
|-----+--------------+------+------+-------------+--------+---------+------------|
| ... |            2 |    1 |  256 | 0.000668163 |      1 | 2.31479 |     0.0977 |
| ... |            4 |   64 |    8 | 0.0331514   |      1 | 2.31605 |     0.0983 |
| ... |            4 |    2 |    1 | 0.000150295 |      1 | 2.30755 |     0.1023 |
| ... |           16 |   32 |   32 | 0.0128248   |     10 | 1.66912 |     0.4391 |
| ... |            4 |    8 |  128 | 0.00464561  |      2 | 1.7316  |     0.3463 |
| ... |            8 |  256 |    8 | 0.00031556  |      1 | 2.19409 |     0.1736 |
| ... |            4 |   16 |  256 | 0.00574329  |      2 | 1.85679 |     0.3368 |
| ... |            8 |    2 |    2 | 0.00325652  |      1 | 2.30272 |     0.0984 |
| ... |            2 |    2 |    2 | 0.000342987 |      2 | 1.76044 |     0.292  |
| ... |            4 |   64 |   32 | 0.003734    |      8 | 1.53101 |     0.4761 |
+-----+--------------+------+------+-------------+--------+---------+------------+

Best trial config: {'l1': 64, 'l2': 32, 'lr': 0.0037339984519545164, 'batch_size': 4}
Best trial final validation loss: 1.5310075663924216
Best trial final validation accuracy: 0.4761
Best trial test set accuracy: 0.4737

Most trials stopped early to conserve resources. The best performing trial achieved a validation accuracy of approximately 47%, which the test set confirms.

Observability#

Monitoring is critical when running large-scale experiments. Ray provides a dashboard that lets you view the status of your trials, check cluster resource use, and inspect logs in real time.

For debugging, Ray also offers distributed debugging tools that let you attach a debugger to running trials across the cluster.

Conclusion#

In this tutorial, you learned how to tune the hyperparameters of a PyTorch model using Ray Tune. You saw how to integrate Ray Tune into your PyTorch training loop, define a search space for your hyperparameters, use an efficient scheduler like ASHAScheduler to terminate low-performing trials early, save checkpoints and report metrics to Ray Tune, and run the hyperparameter search and analyze the results.

Ray Tune makes it straightforward to scale your experiments from a single machine to a large cluster, helping you find the best model configuration efficiently.

Further reading#

Total running time of the script: (47 minutes 44.398 seconds)