Pytorch Lightning Distributed Accelerators using Ray

Overview

Distributed PyTorch Lightning Training on Ray

This library adds new PyTorch Lightning accelerators for distributed training using the Ray distributed computing framework.

These PyTorch Lightning Accelerators on Ray enable quick and easy parallel training while still leveraging all the benefits of PyTorch Lightning and using your desired training protocol, either PyTorch Distributed Data Parallel or Horovod.

Once you add your accelerator to the PyTorch Lightning Trainer, you can parallelize training to all the cores in your laptop, or across a massive multi-node, multi-GPU cluster with no additional code changes.

This library also comes with an integration with Ray Tune for distributed hyperparameter tuning experiments.

Installation

You can install the master branch of ray_lightning_accelerators like so:

pip install git+https://github.com/ray-project/ray_lightning_accelerators#ray_lightning

PyTorch Distributed Data Parallel Accelerator on Ray

The RayAccelerator provides Distributed Data Parallel training on a Ray cluster. PyTorch DDP is used as the distributed training protocol, and Ray is used to launch and manage the training worker processes.

Here is a simplified example:

import pytorch_lightning as ptl
from ray_lightning import RayAccelerator

# Create your PyTorch Lightning model here.
ptl_model = MNISTClassifier(...)
accelerator = RayAccelerator(num_workers=4, cpus_per_worker=1, use_gpu=True)

# If using GPUs, set the ``gpus`` arg to a value > 0.
# The actual number of GPUs is determined by ``num_workers``.
trainer = pl.Trainer(..., gpus=1, accelerator=accelerator)
trainer.fit(ptl_model)

Because Ray is used to launch processes, instead of the same script being called multiple times, you CAN use this accelerator even in cases when you cannot use the standard DDPAccelerator such as

  • Jupyter Notebooks, Google Colab, Kaggle
  • Calling fit or test multiple times in the same script

Horovod Accelerator on Ray

Or if you prefer to use Horovod as the distributed training protocol, use the HorovodRayAccelerator instead.

import pytorch_lightning as ptl
from ray.util.lightning_accelerators import HorovodRayAccelerator

# Create your PyTorch Lightning model here.
ptl_model = MNISTClassifier(...)

# 2 nodes, 4 workers per node, each using 1 CPU and 1 GPU.
accelerator = HorovodRayAccelerator(num_hosts=2, num_slots=4, use_gpu=True)

# If using GPUs, set the ``gpus`` arg to a value > 0.
# The actual number of GPUs is determined by ``num_slots``.
trainer = pl.Trainer(..., gpus=1, accelerator=accelerator)
trainer.fit(ptl_model)

Multi-node Distributed Training

Using the same examples above, you can run distributed training on a multi-node cluster with just 2 simple steps.

  1. Use Ray's cluster launcher to start a Ray cluster- ray up my_cluster_config.yaml.
  2. Execute your Python script on the Ray cluster- ray submit my_cluster_config.yaml train.py. This will rsync your training script to the head node, and execute it on the Ray cluster.

You no longer have to set environment variables or configurations and run your training script on every single node.

Hyperparameter Tuning with Ray Tune

ray_lightning also integrates with Ray Tune to provide distributed hyperparameter tuning for your distributed model training. You can run multiple PyTorch Lightning training runs in parallel, each with a different hyperparameter configuration, and each training run parallelized by itself. All you have to do is move your training code to a function, pass the function to tune.run, and make sure to add the appropriate callback (Either TuneReportCallback or TuneReportCheckpointCallback) to your PyTorch Lightning Trainer.

Example using ray_lightning with Tune:

def train_mnist(config):
    
    # Create your PTL model.
    model = MNISTClassifier(config)

    # Create the Tune Reporting Callback
    metrics = {"loss": "ptl/val_loss", "acc": "ptl/val_accuracy"}
    callbacks = [TuneReportCallback(metrics, on="validation_end")]
    
    trainer = pl.Trainer(
        max_epochs=4,
        callbacks=callbacks,
        accelerator=RayAccelerator(num_workers=4, use_gpu=False))
    trainer.fit(model)
    
config = {
    "layer_1": tune.choice([32, 64, 128]),
    "layer_2": tune.choice([64, 128, 256]),
    "lr": tune.loguniform(1e-4, 1e-1),
    "batch_size": tune.choice([32, 64, 128]),
}

# Make sure to specify how many actors each training run will create via the "extra_cpu" field.
analysis = tune.run(
        train_mnist,
        metric="loss",
        mode="min",
        config=config,
        num_samples=num_samples,
        resources_per_trial={
            "cpu": 1,
            "extra_cpu": 4
        },
        name="tune_mnist")
        
print("Best hyperparameters found were: ", analysis.best_config)

FAQ

RaySGD already has a Pytorch Lightning integration. What's the difference between this integration and that?

The key difference is which Trainer you'll be interacting with. In this library, you will still be using Pytorch Lightning's Trainer. You'll be able to leverage all the features of Pytorch Lightning, and Ray is used just as a backend to handle distributed training.

With RaySGD's integration, you'll be converting your LightningModule to be RaySGD compatible, and will be interacting with RaySGD's TorchTrainer. RaySGD's TorchTrainer is not as feature rich nor as easy to use as Pytorch Lightning's Trainer (no built in support for logging, early stopping, etc.). However, it does have built in support for fault-tolerant and elastic training. If these are hard requirements for you, then RaySGD's integration with PTL might be a better option.

I see that RayAccelerator is based off of Pytorch Lightning's DDPSpawnAccelerator. However, doesn't the PTL team discourage the use of spawn?

As discussed here, using a spawn approach instead of launch is not all that detrimental. The original factors for discouraging spawn were:

  1. not being able to use 'spawn' in a Jupyter or Colab notebook, and
  2. not being able to use multiple workers for data loading.

Neither of these should be an issue with the RayAccelerator due to Ray's serialization mechanisms. The only thing to keep in mind is that when using this accelerator, your model does have to be serializable/pickleable.

Comments
  • Unable to use both GPUs

    Unable to use both GPUs

    Hello, thanks for this amazing library for Lightning!

    I am trying to run Lighting and Ray Tune on a system with 2 GPU. As a start, I want to use both of GPU to train 1 trial at a time.

    However, when I use

    def train_model(config):
        ...
        trainer = pl.Trainer(
            gpus=2,
            accelerator="ddp",
            callbacks=[checkpoint_callback, tune_report_callback],
            plugins=[RayPlugin(num_workers=1, use_gpu=True)],
            precision=16,
        )
        trainer.fit(model, dm)
    
    if __name__ == "__main__":
    
        ray.init()
    
        config = {"batch_size": 256}
    
        analysis = tune.run(
            train_model,
            metric="loss",
            mode="min",
            config=config,
            num_samples=1,
            resources_per_trial={"gpu": 2},
            name="test",
        )
    

    I get the error about the actor or task not being able to be scheduled.

    == Status ==
    Memory usage on this node: 17.5/125.8 GiB
    Using FIFO scheduling algorithm.
    Resources requested: 1/32 CPUs, 2/2 GPUs, 0.0/70.46 GiB heap, 0.0/23.58 GiB objects (0/1.0 accelerator_type:RTX)
    Result logdir: /home/x/ray_results/test
    Number of trials: 1/1 (1 RUNNING)
    
    ...
    
    (pid=109446) GPU available: True, used: True
    (pid=109446) TPU available: None, using: 0 TPU cores
    (pid=109446) Using native 16bit precision.
    (pid=109446) LOCAL_RANK: 0 - CUDA_VISIBLE_DEVICES: [0,1]
    2021-03-30 01:12:21,546 WARNING worker.py:1107 -- The actor or task with ID ffffffffffffffff609630d00bec4e0790a0da3f01000000 cannot be scheduled right now. It requires {CPU: 1.000000}, {GPU: 1.000000} for placement, but this node only has remaining {31.000000/32.000000 CPU, 70.556641 GiB/70.556641 GiB memory, 0.000000/2.000000 GPU, 23.583984 GiB/23.583984 GiB object_store_memory, 1.000000/1.000000 node:192.168.1.159, 1.000000/1.000000 accelerator_type:RTX}
    . In total there are 0 pending tasks and 1 pending actors on this node. This is likely due to all cluster resources being claimed by actors. To resolve the issue, consider creating fewer actors or increase the resources available to this Ray cluster. You can ignore this message if this Ray cluster is expected to auto-scale.
    

    Firstly, I tried changing Trainer to use gpus=1 and num_workers=2, but we still get the same error

        trainer = pl.Trainer(
            gpus=1,
            accelerator="ddp",
            callbacks=[checkpoint_callback, tune_report_callback],
            plugins=[RayPlugin(num_workers=2, use_gpu=True)],
            precision=16,
        )
    
    == Status ==
    Memory usage on this node: 17.1/125.8 GiB
    Using FIFO scheduling algorithm.
    Resources requested: 1/32 CPUs, 2/2 GPUs, 0.0/70.07 GiB heap, 0.0/23.44 GiB objects (0/1.0 accelerator_type:RTX)
    Result logdir: /home/x/ray_results/test
    Number of trials: 1/1 (1 RUNNING)
    
    ...
    
    (pid=111559) GPU available: True, used: True
    (pid=111559) TPU available: None, using: 0 TPU cores
    (pid=111559) Using native 16bit precision.
    (pid=111559) LOCAL_RANK: 0 - CUDA_VISIBLE_DEVICES: [0,1]
    2021-03-30 01:14:17,641 WARNING worker.py:1107 -- The actor or task with ID ffffffffffffffff98dc17f283c4962e8f312ee401000000 cannot be scheduled right now. It requires {CPU: 1.000000}, {GPU: 1.000000} for placement, but this node only has remaining {31.000000/32.000000 CPU, 70.068359 GiB/70.068359 GiB memory, 0.000000/2.000000 GPU, 1.000000/1.000000 accelerator_type:RTX, 1.000000/1.000000 node:192.168.1.159, 23.437500 GiB/23.437500 GiB object_store_memory}
    . In total there are 0 pending tasks and 2 pending actors on this node. This is likely due to all cluster resources being claimed by actors. To resolve the issue, consider creating fewer actors or increase the resources available to this Ray cluster. You can ignore this message if this Ray cluster is expected to auto-scale.
    

    Next, I reduced num_workers=2 to num_workers=1, and the error remains

        trainer = pl.Trainer(
            val_check_interval=0.1,
            gpus=1,
            accelerator="ddp",
            callbacks=[checkpoint_callback, tune_report_callback],
            plugins=[RayPlugin(num_workers=1, use_gpu=True)],
            precision=16,
            progress_bar_refresh_rate=1000,  # refresh every 1000 iterations
        )
    
    == Status ==
    Memory usage on this node: 19.4/125.8 GiB
    Using FIFO scheduling algorithm.
    Resources requested: 1/32 CPUs, 2/2 GPUs, 0.0/68.7 GiB heap, 0.0/23.05 GiB objects (0/1.0 accelerator_type:RTX)
    Result logdir: /home/x/ray_results/test
    Number of trials: 1/1 (1 RUNNING)
    
    ...
    
    (pid=113817) GPU available: True, used: True
    (pid=113817) TPU available: None, using: 0 TPU cores
    (pid=113817) Using native 16bit precision.
    (pid=113817) LOCAL_RANK: 0 - CUDA_VISIBLE_DEVICES: [0,1]
    2021-03-30 01:16:23,959 WARNING worker.py:1107 -- The actor or task with ID ffffffffffffffffa8b7c70cbc343023efcc60f501000000 cannot be scheduled right now. It requires {CPU: 1.000000}, {GPU: 1.000000} for placement, but this node only has remaining {31.000000/32.000000 CPU, 68.701172 GiB/68.701172 GiB memory, 0.000000/2.000000 GPU, 23.046875 GiB/23.046875 GiB object_store_memory, 1.000000/1.000000 accelerator_type:RTX, 1.000000/1.000000 node:192.168.1.159}
    . In total there are 0 pending tasks and 1 pending actors on this node. This is likely due to all cluster resources being claimed by actors. To resolve the issue, consider creating fewer actors or increase the resources available to this Ray cluster. You can ignore this message if this Ray cluster is expected to auto-scale.
    

    Finally, I reduced resources_per_trial to resources_per_trial={"gpu": 1}, and it finally runs, but appears to be using only 1 GPU, not 2.

        trainer = pl.Trainer(
            gpus=1,
            accelerator="ddp",
            callbacks=[checkpoint_callback, tune_report_callback],
            plugins=[RayPlugin(num_workers=1, use_gpu=True)],
            precision=16,
        )
    
        analysis = tune.run(
            train_model,
            metric="loss",
            mode="min",
            config=config,
            num_samples=1,
            resources_per_trial={"gpu": 1},
            name="test",
        )
    
    == Status ==
    Memory usage on this node: 18.8/125.8 GiB
    Using FIFO scheduling algorithm.
    Resources requested: 1/32 CPUs, 1/2 GPUs, 0.0/69.24 GiB heap, 0.0/23.19 GiB objects (0/1.0 accelerator_type:RTX)
    Result logdir: /home/x/ray_results/test
    Number of trials: 1/1 (1 RUNNING)
    
    ...
    
    (pid=119147) GPU available: True, used: True
    (pid=119147) TPU available: None, using: 0 TPU cores
    (pid=119147) Using native 16bit precision.
    (pid=119147) LOCAL_RANK: 0 - CUDA_VISIBLE_DEVICES: [0]
    (pid=119167) initializing ddp: GLOBAL_RANK: 0, MEMBER: 1/1
    (pid=119167) 
    (pid=119167)   | Name  | Type          | Params
    (pid=119167) ----------------------------------------
    (pid=119167) 0 | model | TestPLModel00 | 3.7 K 
    (pid=119167) ----------------------------------------
    (pid=119167) 3.7 K     Trainable params
    (pid=119167) 0         Non-trainable params
    (pid=119167) 3.7 K     Total params
    (pid=119167) 0.015     Total estimated model params size (MB)
    

    I am using

    • pytorch 1.7.0
    • pytorch-lightning 1.2.5
    • ray 1.2.0
    • ray-lightning 0.0.1

    What should be the correct way to train a single trial using both GPU devices?

    opened by nyxynyx 36
  • NCCL peer access is not supported error

    NCCL peer access is not supported error

    just the training was about to start (nvidia-smi shows the GPU memory filling up), there's a new error

    RuntimeError: NCCL error in: /opt/conda/conda-bld/pytorch_1603729062494/work/torch/lib/c10d/ProcessGroupNCCL.cpp:784, unhandled cuda error, NCCL version 2.7.8
    

    Starting the Python training script using NCCL_IB_DISABLE=1 python tune.py does not help.

    Enabled debug messages using NCCL_DEBUG="INFO" NCCL_IB_DISABLE=1 python tune.py, and the following new messages appeared:

    (pid=909597) RuntimeError: NCCL error in: /opt/conda/conda-bld/pytorch_1603729062494/work/torch/lib/c10d/ProcessGroupNCCL.cpp:784, unhandled cuda error, NCCL version 2.7.8
    (pid=909608) z-pc:909608:910337 [0] NCCL INFO Channel 00/02 :    0   1
    (pid=909608) z-pc:909608:910337 [0] NCCL INFO Channel 01/02 :    0   1
    (pid=909608) z-pc:909608:910337 [0] NCCL INFO threadThresholds 8/8/64 | 16/8/64 | 8/8/64
    (pid=909608) z-pc:909608:910337 [0] NCCL INFO Trees [0] 1/-1/-1->0->-1|-1->0->1/-1/-1 [1] 1/-1/-1->0->-1|-1->0->1/-1/-1
    (pid=909608) z-pc:909608:910337 [0] NCCL INFO Channel 00 : 0[2d000] -> 1[2e000] via P2P/IPC
    (pid=909608) 
    (pid=909608) z-pc:909608:910337 [0] transport/p2p.cc:238 NCCL WARN failed to open CUDA IPC handle : 217 peer access is not supported between these two devices
    (pid=909608) z-pc:909608:910337 [0] NCCL INFO transport.cc:68 -> 1
    (pid=909608) z-pc:909608:910337 [0] NCCL INFO init.cc:766 -> 1
    (pid=909608) z-pc:909608:910337 [0] NCCL INFO init.cc:840 -> 1
    (pid=909608) z-pc:909608:910337 [0] NCCL INFO group.cc:73 -> 1 [Async thread]
    (pid=909611) z-pc:909611:910338 [0] NCCL INFO threadThresholds 8/8/64 | 16/8/64 | 8/8/64
    (pid=909611) z-pc:909611:910338 [0] NCCL INFO Trees [0] -1/-1/-1->1->0|0->1->-1/-1/-1 [1] -1/-1/-1->1->0|0->1->-1/-1/-1
    (pid=909611) z-pc:909611:910338 [0] NCCL INFO Channel 00 : 1[2e000] -> 0[2d000] via P2P/IPC
    (pid=909611) 
    (pid=909611) z-pc:909611:910338 [0] transport/p2p.cc:238 NCCL WARN failed to open CUDA IPC handle : 217 peer access is not supported between these two devices
    (pid=909611) z-pc:909611:910338 [0] NCCL INFO transport.cc:68 -> 1
    (pid=909611) z-pc:909611:910338 [0] NCCL INFO init.cc:766 -> 1
    (pid=909611) z-pc:909611:910338 [0] NCCL INFO init.cc:840 -> 1
    (pid=909611) z-pc:909611:910338 [0] NCCL INFO group.cc:73 -> 1 [Async thread]
    2021-03-30 18:20:32,548 ERROR trial_runner.py:616 -- Trial train_model_190f1_00000: Error processing event.
    $ nvidia-smi topo -m
    	GPU0	GPU1	CPU Affinity	NUMA Affinity
    GPU0	 X 	PHB	0-31		N/A
    GPU1	PHB	 X 	0-31		N/A
    

    Originally posted by @nyxynyx

    opened by amogkam 22
  • question: Do I need to use ray.init() before using the Ray Accelerator?

    question: Do I need to use ray.init() before using the Ray Accelerator?

    Hey, Thank you for creating a needed library.

    I am very new to using Ray, and I already had a project built around PL. I looked around on how to add Ray distributed training backend to my project, and I found this library that does not force me to not use the PL trainer.

    Now, I am trying to use the accelerator on my local machine, but I failed to do so. I think it's a really simple issue, because of my lack of knowledge.

    This the bit where I add the accelerator:

    if accelerator_use:
        ray.init()
        accelerator = RayAccelerator(num_workers=4, cpus_per_worker=1, use_gpu=True)
    else:
        accelerator = None
    

    I tried without using ray init and I got an error, and when I add ray init I get this:

    2021-03-06 14:54:45,217 WARNING worker.py:1107 -- The actor or task with ID ffffffffffffffff63964fa4841d4a2ecb45751801000000 cannot be scheduled right now. It requires {CPU: 1.000000}, {GPU: 1.000000} for placement, but this node only has remaining {7.000000/8.000000 CPU, 7.177734 GiB/7.177734 GiB memory, 0.000000/1.000000 GPU, 1.000000/1.000000 node:172.20.10.2, 2.441406 GiB/2.441406 GiB object_store_memory}
    . In total there are 0 pending tasks and 6 pending actors on this node. This is likely due to all cluster resources being claimed by actors. To resolve the issue, consider creating fewer actors or increase the resources available to this Ray cluster. You can ignore this message if this Ray cluster is expected to auto-scale.
    
    opened by MohammedAljahdali 12
  • Compatibility with PTL 1.6

    Compatibility with PTL 1.6

    Todos:

    • [x] Check if we need to_state_stream / load_state_stream P(0)
    • [x] Check multi node (P0)
    • [x] Check multi GPU/multi node (P0)
    • [x] Fix / change tests (P0)
    • [x] Check that recent PRs are included, e.g. https://github.com/ray-project/ray_lightning/pull/156 P(0.5-1)
    • [x] Check Ray client (P1)
    • [x] Check fractional GPUs (P2)
    • [x] DDP sharded (P2)
    opened by krfricke 11
  • ray ddp fails with 2 gpu workers

    ray ddp fails with 2 gpu workers

      File "/home/ubuntu/anaconda3/envs/automm-dev-pl-latest/lib/python3.8/site-packages/pytorch_lightning/trainer/trainer.py", line 1797, in __setup_profiler
        self.profiler.setup(stage=self.state.fn._setup_fn, local_rank=local_rank, log_dir=self.log_dir)
      File "/home/ubuntu/anaconda3/envs/automm-dev-pl-latest/lib/python3.8/site-packages/pytorch_lightning/trainer/trainer.py", line 2249, in log_dir
        dirpath = self.strategy.broadcast(dirpath)
      File "/home/ubuntu/anaconda3/envs/automm-dev-pl-latest/lib/python3.8/site-packages/pytorch_lightning/strategies/ddp_spawn.py", line 215, in broadcast
        torch.distributed.broadcast_object_list(obj, src, group=_group.WORLD)
      File "/home/ubuntu/anaconda3/envs/automm-dev-pl-latest/lib/python3.8/site-packages/torch/distributed/distributed_c10d.py", line 1817, in broadcast_object_list
        broadcast(object_sizes_tensor, src=src, group=group)
      File "/home/ubuntu/anaconda3/envs/automm-dev-pl-latest/lib/python3.8/site-packages/torch/distributed/distributed_c10d.py", line 1159, in broadcast
        work = default_pg.broadcast([tensor], opts)
    RuntimeError: NCCL error in: ../torch/csrc/distributed/c10d/ProcessGroupNCCL.cpp:891, internal error, NCCL version 21.0.3
    ncclInternalError: Internal check failed. This is either a bug in NCCL or due to memory corruptio
    

    use this branch: https://github.com/sxjscience/autogluon/tree/kaggle_california_house and install autogluon via bash full_install.sh Afterwards, try this script: https://gist.github.com/sxjscience/53bc799e37cc0680ca9e53c2fea75cd7 Internally, the ray strategy are constructed here: https://github.com/sxjscience/autogluon/blob/59f01b95381fba5651db17fd98fa84164ad168c2/multimodal/src/autogluon/multimodal/predictor.py#L1036-L1052 .

    opened by JiahaoYao 10
  • [Tune] Ray Tune + Ray Lightning too many tasks warning

    [Tune] Ray Tune + Ray Lightning too many tasks warning

    I noticed such warning log constantly being logged out while using Ray Tune + Ray Lightning. For example: Warning: More than 20000 tasks are pending submission to actor 386ebf690ec87ad0d825174701000000. To reduce memory usage, wait for these tasks to finish before sending more.

    Do I need to worry about it?

    opened by yinweisu 10
  • CUDA devices are not exposed when running in DDP mode with multiple GPUs

    CUDA devices are not exposed when running in DDP mode with multiple GPUs

    Hi all!

    First of all thanks for this great project!

    To my issue: When I tried your example for hyperparameter tuning I discovered that it only worked when using cpu. After some digging, I found out that the problem is related to get_tune_ddp_resources Since head_bundle only requests cpu it does not expose the required CUDA devices for the child bundles. Therefore, Lightning fails to run on GPU(s) since CUDA_VISIBLE_DEVICES is not set.

    As a workaround, I have added

    if use_gpu:
            os.environ["CUDA_VISIBLE_DEVICES"] = ",".join([str(i) for i in range(num_workers)])
    

    in __init__ of RayPlugin It gets the job done but seems incredibly hacky. Is there some other way around?

    Thanks in advance!

    opened by MarkusSpanring 10
  • Does not appear to be compatible with the current version of Lightning

    Does not appear to be compatible with the current version of Lightning

    I was excited to try this out but the code appears to not be working due to a missing import:

    Traceback (most recent call last):
      File "...", line 9, in <module>
        from ray_lightning import RayAccelerator
      File ".../lib/python3.8/site-packages/ray_lightning/__init__.py", line 1, in <module>
        from ray_lightning.ray_ddp import RayAccelerator
      File ".../lib/python3.8/site-packages/ray_lightning/ray_ddp.py", line 8, in <module>
        from pytorch_lightning.accelerators import DDPSpawnAccelerator
    ImportError: cannot import name 'DDPSpawnAccelerator' from 'pytorch_lightning.accelerators' (.../lib/python3.8/site-packages/pytorch_lightning/accelerators/__init__.py)
    
    opened by import-antigravity 10
  • [Windows] RuntimeError: Distributed package doesn't have NCCL built in

    [Windows] RuntimeError: Distributed package doesn't have NCCL built in

    Hey, I am having an issue when I run trainer.fit with the accelerator I get the following error:

    2021-03-08 13:45:49,085 INFO services.py:1172 -- View the Ray dashboard at http://127.0.0.1:8265
    GPU available: True, used: True
    TPU available: None, using: 0 TPU cores
    LOCAL_RANK: 0 - CUDA_VISIBLE_DEVICES: [0]
    Using native 16bit precision.
    Global seed set to 1234
    initializing ddp: GLOBAL_RANK: 0, MEMBER: 1/1
    Traceback (most recent call last):
      File "train.py", line 26, in cli_main
        train(None, cfg)
      File "train.py", line 102, in train
        trainer.fit(model, datamodule=dm)
      File "C:\Users\Mohammed\AppData\Local\conda\conda\envs\htts\lib\site-packages\pytorch_lightning\trainer\trainer.py", line 510, in fit
        results = self.accelerator_backend.train()
      File "C:\Users\Mohammed\AppData\Local\conda\conda\envs\htts\lib\site-packages\ray_lightning\ray_ddp.py", line 184, in train
        results = process_results(futures, queue)
      File "C:\Users\Mohammed\AppData\Local\conda\conda\envs\htts\lib\site-packages\ray_lightning\util.py", line 103, in process_results
        ray.get(ready)
      File "C:\Users\Mohammed\AppData\Roaming\Python\Python38\site-packages\ray\_private\client_mode_hook.py", line 47, in wrapper
        return func(*args, **kwargs)
      File "C:\Users\Mohammed\AppData\Roaming\Python\Python38\site-packages\ray\worker.py", line 1456, in get
        raise value.as_instanceof_cause()
    ray.exceptions.RayTaskError(RuntimeError): ray::RayExecutor.execute() (pid=27620, ip=192.168.8.100)
      File "python\ray\_raylet.pyx", line 480, in ray._raylet.execute_task
      File "python\ray\_raylet.pyx", line 432, in ray._raylet.execute_task.function_executor
      File "C:\Users\Mohammed\AppData\Roaming\Python\Python38\site-packages\ray\function_manager.py", line 556, in actor_method_executor
        return method(__ray_actor, *args, **kwargs)
      File "C:\Users\Mohammed\AppData\Local\conda\conda\envs\htts\lib\site-packages\ray_lightning\ray_ddp.py", line 31, in execute
        return fn(*args, **kwargs)
      File "C:\Users\Mohammed\AppData\Local\conda\conda\envs\htts\lib\site-packages\ray_lightning\ray_ddp.py", line 218, in train_remote
        super(RayAccelerator, self).ddp_train(
      File "C:\Users\Mohammed\AppData\Local\conda\conda\envs\htts\lib\site-packages\pytorch_lightning\accelerators\ddp_spawn_accelerator.py", line 127, in ddp_train
        self.init_ddp_connection(
      File "C:\Users\Mohammed\AppData\Local\conda\conda\envs\htts\lib\site-packages\ray_lightning\ray_ddp.py", line 232, in init_ddp_connection
        torch.distributed.init_process_group(
      File "C:\Users\Mohammed\AppData\Local\conda\conda\envs\htts\lib\site-packages\torch\distributed\distributed_c10d.py", line 503, in init_process_group
        _update_default_pg(_new_process_group_helper(
      File "C:\Users\Mohammed\AppData\Local\conda\conda\envs\htts\lib\site-packages\torch\distributed\distributed_c10d.py", line 597, in _new_process_group_helper
        raise RuntimeError("Distributed package doesn't have NCCL "
    RuntimeError: Distributed package doesn't have NCCL built in
    
    opened by MohammedAljahdali 9
  • Optimising with respect to the epoch that scored highest for a trial (instead of the last epoch)

    Optimising with respect to the epoch that scored highest for a trial (instead of the last epoch)

    Hi,

    I am using metric="acc" for ray.tune, with mode="max". However, I think that the score for the last epoch is being used as the "best" score for the trial.

    E.g., for trial train_mnist_c1a02_00000 from the following console output, the reported acc is 0.942736:

    +-------------------------+------------+-------+--------------+-----------+-----------+-------------+--------+------------------+----------+----------+
    | Trial name              | status     | loc   |   batch_size |   layer_1 |   layer_2 |          lr |   iter |   total time (s) |      loss |      acc |
    |-------------------------+------------+-------+--------------+-----------+-----------+-------------+--------+------------------+-----------+----------|
    | train_mnist_c1a02_00000 | TERMINATED |       |           64 |       128 |       256 | 0.000120742 |     16 |    166.36  | -0.938587 | 0.942736 |
    | train_mnist_c1a02_00001 | TERMINATED |       |          128 |       128 |        64 | 0.000120068 |     16 |    138.23  | -0.923084 | 0.929161 |
    | train_mnist_c1a02_00002 | TERMINATED |       |           64 |        32 |       256 | 0.000308457 |     16 |    168.73  | -0.942267 | 0.945811 |
    | train_mnist_c1a02_00003 | TERMINATED |       |           64 |        32 |       256 | 0.0927983   |     16 |    162.749 | -0.103807 | 0.103807 |
    +-------------------------+------------+-------+--------------+-----------+-----------+-------------+--------+------------------+-----------+----------+
    

    However, when looking at progress.csv in tune_mnist/train_mnist_c1a02_00000_0_batch_size=64,layer_1=128,layer_2=256,lr=0.00012074_2021-08-25_20-18-14, the highest acc is 0.9441488981246948 (where 0.942736 is the score for the last epoch):

    loss,acc,time_this_iter_s,done,timesteps_total,episodes_total,training_iteration,experiment_id,date,timestamp,time_total_s,pid,hostname,node_ip,time_since_restore,timesteps_since_restore,iterations_since_restore,trial_id
    -0.7927070260047913,0.8189826607704163,28.36010980606079,False,,,1,daa92b03869b42409a0f9a69b6d7918d,2021-08-25_20-18-49,1629886729,28.36010980606079,6438,b078,10.141.1.144,28.36010980606079,0,1,c1a02_00000
    -0.8159223198890686,0.829454779624939,9.054965734481812,False,,,2,daa92b03869b42409a0f9a69b6d7918d,2021-08-25_20-18-58,1629886738,37.4150755405426,6438,b078,10.141.1.144,37.4150755405426,0,2,c1a02_00000
    -0.8259921669960022,0.835106372833252,9.04287576675415,False,,,3,daa92b03869b42409a0f9a69b6d7918d,2021-08-25_20-19-07,1629886747,46.45795130729675,6438,b078,10.141.1.144,46.45795130729675,0,3,c1a02_00000
    -0.831539511680603,0.8390957117080688,8.813407182693481,False,,,4,daa92b03869b42409a0f9a69b6d7918d,2021-08-25_20-19-16,1629886756,55.271358489990234,6438,b078,10.141.1.144,55.271358489990234,0,4,c1a02_00000
    -0.8353389501571655,0.8413397073745728,9.657632112503052,False,,,5,daa92b03869b42409a0f9a69b6d7918d,2021-08-25_20-19-26,1629886766,64.92899060249329,6438,b078,10.141.1.144,64.92899060249329,0,5,c1a02_00000
    -0.896843671798706,0.9082446694374084,9.233526706695557,False,,,6,daa92b03869b42409a0f9a69b6d7918d,2021-08-25_20-19-35,1629886775,74.16251730918884,6438,b078,10.141.1.144,74.16251730918884,0,6,c1a02_00000
    -0.9138997197151184,0.9224567413330078,9.06982707977295,False,,,7,daa92b03869b42409a0f9a69b6d7918d,2021-08-25_20-19-44,1629886784,83.23234438896179,6438,b078,10.141.1.144,83.23234438896179,0,7,c1a02_00000
    -0.919984757900238,0.9273603558540344,9.12305760383606,False,,,8,daa92b03869b42409a0f9a69b6d7918d,2021-08-25_20-19-53,1629886793,92.35540199279785,6438,b078,10.141.1.144,92.35540199279785,0,8,c1a02_00000
    -0.9245654940605164,0.9311003684997559,9.59031629562378,False,,,9,daa92b03869b42409a0f9a69b6d7918d,2021-08-25_20-20-03,1629886803,101.94571828842163,6438,b078,10.141.1.144,101.94571828842163,0,9,c1a02_00000
    -0.9273290038108826,0.9330950379371643,9.259892702102661,False,,,10,daa92b03869b42409a0f9a69b6d7918d,2021-08-25_20-20-12,1629886812,111.20561099052429,6438,b078,10.141.1.144,111.20561099052429,0,10,c1a02_00000
    -0.9301624298095703,0.935339093208313,9.178678750991821,False,,,11,daa92b03869b42409a0f9a69b6d7918d,2021-08-25_20-20-21,1629886821,120.38428974151611,6438,b078,10.141.1.144,120.38428974151611,0,11,c1a02_00000
    -0.9327710270881653,0.9389959573745728,9.190800189971924,False,,,12,daa92b03869b42409a0f9a69b6d7918d,2021-08-25_20-20-30,1629886830,129.57508993148804,6438,b078,10.141.1.144,129.57508993148804,0,12,c1a02_00000
    -0.9332801103591919,0.9382479786872864,9.453409433364868,False,,,13,daa92b03869b42409a0f9a69b6d7918d,2021-08-25_20-20-40,1629886840,139.0284993648529,6438,b078,10.141.1.144,139.0284993648529,0,13,c1a02_00000
    -0.9350691437721252,0.9409075379371643,8.920868873596191,False,,,14,daa92b03869b42409a0f9a69b6d7918d,2021-08-25_20-20-49,1629886849,147.9493682384491,6438,b078,10.141.1.144,147.9493682384491,0,14,c1a02_00000
    -0.9383015632629395,0.9441488981246948,9.268950700759888,False,,,15,daa92b03869b42409a0f9a69b6d7918d,2021-08-25_20-20-58,1629886858,157.21831893920898,6438,b078,10.141.1.144,157.21831893920898,0,15,c1a02_00000
    -0.9385870099067688,0.942736029624939,9.141263961791992,False,,,16,daa92b03869b42409a0f9a69b6d7918d,2021-08-25_20-21-07,1629886867,166.35958290100098,6438,b078,10.141.1.144,166.35958290100098,0,16,c1a02_00000
    
    

    I see that this was an issue here: https://github.com/ray-project/ray/issues/5174, but it isn't clear what the fix was.

    Thanks.

    Source:

    """Simple example using RayAccelerator and Ray Tune"""
    
    from pl_bolts.datamodules.mnist_datamodule import MNISTDataModule
    from ray import tune
    from ray_lightning.tests.utils import LightningMNISTClassifier
    from ray_lightning.tune import TuneReportCallback, get_tune_ddp_resources
    from ray_lightning import RayPlugin
    import os
    import pytorch_lightning as pl
    import ray
    
    DATA_DIR = "/datasets/work/hb-mlaifsp-mm/source/Datasets/mnist"
    NUM_WORKERS = 1
    NUM_SAMPLES = 4
    MAX_EPOCHS = 16
    USE_GPU = True
    
    def train_mnist(config):
    
    
        model = LightningMNISTClassifier(config, DATA_DIR)
    
        metrics = {"loss": "ptl/val_loss", "acc": "ptl/val_accuracy"}
        callbacks = [TuneReportCallback(metrics, on="validation_end")]
    
        trainer = pl.Trainer(
            max_epochs=MAX_EPOCHS,
            callbacks=callbacks,
            progress_bar_refresh_rate=0,
            plugins=[RayPlugin(num_workers=NUM_WORKERS, use_gpu=USE_GPU)],
        )
    
        dm = MNISTDataModule(data_dir=DATA_DIR, num_workers=NUM_WORKERS, batch_size=config["batch_size"])
        trainer.fit(model, dm)
    
    def tune_mnist():
        config = {
            "layer_1": tune.choice([32, 64, 128]),
            "layer_2": tune.choice([64, 128, 256]),
            "lr": tune.loguniform(1e-4, 1e-1),
            "batch_size": tune.choice([32, 64, 128]),
        }
    
        ray.init()
        analysis = tune.run(
            train_mnist,
            metric="acc",
            mode="max",
            local_dir=os.getcwd(),
            config=config,
            num_samples=NUM_SAMPLES,
            resources_per_trial=get_tune_ddp_resources(num_workers=NUM_WORKERS, use_gpu=USE_GPU),
            name="tune_mnist",
        )
        print("Best hyperparameters found were: ", analysis.best_config)
    
    if __name__ == "__main__":
        tune_mnist()
    
    opened by anicolson 8
  • Cloudpickle Dataset deserialization error

    Cloudpickle Dataset deserialization error

    Hi,

    When I try to run the code with RayPlugin in my tests, I get the following error:

    (pid=2127144) 2021-07-14 06:16:52,345   ERROR serialization.py:250 -- No module named 'test_runner'
    (pid=2127144) Traceback (most recent call last):
    (pid=2127144)   File "/home/rizhiy/miniconda3/envs/ntf/lib/python3.8/site-packages/ray/serialization.py", line 248, in deserialize_objects
    (pid=2127144)     obj = self._deserialize_object(data, metadata, object_ref)
    (pid=2127144)   File "/home/rizhiy/miniconda3/envs/ntf/lib/python3.8/site-packages/ray/serialization.py", line 190, in _deserialize_object
    (pid=2127144)     return self._deserialize_msgpack_data(data, metadata_fields)
    (pid=2127144)   File "/home/rizhiy/miniconda3/envs/ntf/lib/python3.8/site-packages/ray/serialization.py", line 168, in _deserialize_msgpack_data
    (pid=2127144)     python_objects = self._deserialize_pickle5_data(pickle5_data)
    (pid=2127144)   File "/home/rizhiy/miniconda3/envs/ntf/lib/python3.8/site-packages/ray/serialization.py", line 158, in _deserialize_pickle5_data
    (pid=2127144)     obj = pickle.loads(in_band)
    (pid=2127144) ModuleNotFoundError: No module named 'test_runner'
    

    test_runner is the name of my testing script, but I'm not sure why you would need to serialize anything inside it. It just loads data and calls Runner, which is my wrapper around pl.Trainer.

    How should I properly launch the training?

    opened by Rizhiy 8
  • Bump pytorch-lightning from 1.6.4 to 1.8.6

    Bump pytorch-lightning from 1.6.4 to 1.8.6

    Bumps pytorch-lightning from 1.6.4 to 1.8.6.

    Release notes

    Sourced from pytorch-lightning's releases.

    Weekly patch release

    App

    Added

    • Added partial support for fastapi Request annotation in configure_api handlers (#16047)
    • Added a nicer UI with URL and examples for the autoscaler component (#16063)
    • Enabled users to have more control over scaling out/in intervals (#16093)
    • Added more datatypes to the serving component (#16018)
    • Added work.delete method to delete the work (#16103)
    • Added display_name property to LightningWork for the cloud (#16095)
    • Added ColdStartProxy to the AutoScaler (#16094)
    • Added status endpoint, enable ready (#16075)
    • Implemented ready for components (#16129)

    Changed

    • The default start_method for creating Work processes locally on macOS is now 'spawn' (previously 'fork') (#16089)
    • The utility lightning.app.utilities.cloud.is_running_in_cloud now returns True during the loading of the app locally when running with --cloud (#16045)
    • Updated Multinode Warning (#16091)
    • Updated app testing (#16000)
    • Changed overwrite to True (#16009)
    • Simplified messaging in cloud dispatch (#16160)
    • Added annotations endpoint (#16159)

    Fixed

    • Fixed PythonServer messaging "Your app has started" (#15989)
    • Fixed auto-batching to enable batching for requests coming even after the batch interval but is in the queue (#16110)
    • Fixed a bug where AutoScaler would fail with min_replica=0 (#16092
    • Fixed a non-thread safe deepcopy in the scheduler (#16114)
    • Fixed HTTP Queue sleeping for 1 sec by default if no delta was found (#16114)
    • Fixed the endpoint info tab not showing up in the AutoScaler UI (#16128)
    • Fixed an issue where an exception would be raised in the logs when using a recent version of streamlit (#16139)
    • Fixed e2e tests (#16146)

    Full Changelog: https://github.com/Lightning-AI/lightning/compare/1.8.5.post0...1.8.6

    Minor patch release

    App

    • Fixed install/upgrade - removing single quote (#16079)
    • Fixed bug where components that are re-instantiated several times failed to initialize if they were modifying self.lightningignore (#16080)
    • Fixed a bug where apps that had previously been deleted could not be run again from the CLI (#16082)

    Pytorch

    ... (truncated)

    Commits

    Dependabot compatibility score

    Dependabot will resolve any conflicts with this PR as long as you don't alter it yourself. You can also trigger a rebase manually by commenting @dependabot rebase.


    Dependabot commands and options

    You can trigger Dependabot actions by commenting on this PR:

    • @dependabot rebase will rebase this PR
    • @dependabot recreate will recreate this PR, overwriting any edits that have been made to it
    • @dependabot merge will merge this PR after your CI passes on it
    • @dependabot squash and merge will squash and merge this PR after your CI passes on it
    • @dependabot cancel merge will cancel a previously requested merge and block automerging
    • @dependabot reopen will reopen this PR if it is closed
    • @dependabot close will close this PR and stop Dependabot recreating it. You can achieve the same result by closing it manually
    • @dependabot ignore this major version will close this PR and stop Dependabot creating any more for this major version (unless you reopen the PR or upgrade to it yourself)
    • @dependabot ignore this minor version will close this PR and stop Dependabot creating any more for this minor version (unless you reopen the PR or upgrade to it yourself)
    • @dependabot ignore this dependency will close this PR and stop Dependabot creating any more for this dependency (unless you reopen the PR or upgrade to it yourself)
    dependencies 
    opened by dependabot[bot] 0
  • Make `_GPUAcceleratorl.get_parallel_devices` fit `GPUAccelerator` API

    Make `_GPUAcceleratorl.get_parallel_devices` fit `GPUAccelerator` API

    Closes https://github.com/ray-project/ray_lightning/issues/235

    The PTL Accelerator API is expected to return a List, and not None. This PR updates our _GPUAccelerator abstraction to fit this API.

    opened by amogkam 0
  • Support string based GPU ids

    Support string based GPU ids

    GPU device ids can be specified with an integer index, but may also be specified as strings.

    This PR ensures that both cases are supported by root_device. The code is taken from what is being done in Ray Train: https://sourcegraph.com/github.com/ray-project/ray/-/blob/python/ray/train/torch/train_loop_utils.py?L470-498

    Closes https://github.com/ray-project/ray_lightning/issues/236

    opened by amogkam 0
  • Update protobuf requirement from <=3.20.1 to <4.21.13

    Update protobuf requirement from <=3.20.1 to <4.21.13

    Updates the requirements on protobuf to permit the latest version.

    Commits

    Dependabot will resolve any conflicts with this PR as long as you don't alter it yourself. You can also trigger a rebase manually by commenting @dependabot rebase.


    Dependabot commands and options

    You can trigger Dependabot actions by commenting on this PR:

    • @dependabot rebase will rebase this PR
    • @dependabot recreate will recreate this PR, overwriting any edits that have been made to it
    • @dependabot merge will merge this PR after your CI passes on it
    • @dependabot squash and merge will squash and merge this PR after your CI passes on it
    • @dependabot cancel merge will cancel a previously requested merge and block automerging
    • @dependabot reopen will reopen this PR if it is closed
    • @dependabot close will close this PR and stop Dependabot recreating it. You can achieve the same result by closing it manually
    • @dependabot ignore this major version will close this PR and stop Dependabot creating any more for this major version (unless you reopen the PR or upgrade to it yourself)
    • @dependabot ignore this minor version will close this PR and stop Dependabot creating any more for this minor version (unless you reopen the PR or upgrade to it yourself)
    • @dependabot ignore this dependency will close this PR and stop Dependabot creating any more for this dependency (unless you reopen the PR or upgrade to it yourself)
    dependencies 
    opened by dependabot[bot] 0
  • Multi-GPU training fails with `ValueError` on systems with UUID GPU IDs

    Multi-GPU training fails with `ValueError` on systems with UUID GPU IDs

    I'm currently trying to use ray_lightning to distribute model training over the resources in my ray cluster, like so:

    ngpu = int(ray.cluster_resources().get("GPU", 0))
    use_gpu = ngpu > 0
    num_workers = ngpu
    ncpu = 8
    strategy = RayStrategy(num_workers,ncpu,use_gpu, find_unused_parameters=False)
    # define dataloaders
    # define callbacks
    trainer = PlTrainer(
        logger=False,
        max_epochs=50,
        callbacks=callbacks,
        gpus=1,
        enable_model_summary=False,
        enable_checkpointing=False,
        strategy=strategy,
    )
    trainer.fit(lit_model, train_dataloader, val_dataloader)
    

    However, this code results in a ValueError:

      File "/home/gridsan/dgraff/molpal/molpal/models/mpnmodels.py", line 207, in train
        trainer.fit(lit_model, train_dataloader, val_dataloader)
      File "/home/gridsan/dgraff/.conda/envs/molpal/lib/python3.8/site-packages/pytorch_lightning/trainer/trainer.py", line 770, in fit
        self._call_and_handle_interrupt(
      File "/home/gridsan/dgraff/.conda/envs/molpal/lib/python3.8/site-packages/pytorch_lightning/trainer/trainer.py", line 721, in _call_and_handle_interrupt
        return self.strategy.launcher.launch(trainer_fn, *args, trainer=self, **kwargs)
      File "/home/gridsan/dgraff/.conda/envs/molpal/lib/python3.8/site-packages/ray_lightning/launchers/ray_launcher.py", line 58, in launch
        ray_output = self.run_function_on_workers(
      File "/home/gridsan/dgraff/.conda/envs/molpal/lib/python3.8/site-packages/ray_lightning/launchers/ray_launcher.py", line 249, in run_function_on_workers
        results = process_results(self._futures, self.tune_queue)
      File "/home/gridsan/dgraff/.conda/envs/molpal/lib/python3.8/site-packages/ray_lightning/util.py", line 64, in process_results
        ray.get(ready)
      File "/home/gridsan/dgraff/.conda/envs/molpal/lib/python3.8/site-packages/ray/_private/client_mode_hook.py", line 105, in wrapper
        return func(*args, **kwargs)
      File "/home/gridsan/dgraff/.conda/envs/molpal/lib/python3.8/site-packages/ray/_private/worker.py", line 2289, in get
        raise value.as_instanceof_cause()
    ray.exceptions.RayTaskError(ValueError): ray::RayExecutor.execute() (pid=49053, ip=172.31.130.105, repr=<ray_lightning.launchers.utils.RayExecutor object at 0x7f392469a6d0>)
      File "/home/gridsan/dgraff/.conda/envs/molpal/lib/python3.8/site-packages/ray_lightning/launchers/utils.py", line 52, in execute
        return fn(*args, **kwargs)
      File "/home/gridsan/dgraff/.conda/envs/molpal/lib/python3.8/site-packages/ray_lightning/launchers/ray_launcher.py", line 295, in _wrapping_function
        self._strategy._worker_setup(process_idx=global_rank)
      File "/home/gridsan/dgraff/.conda/envs/molpal/lib/python3.8/site-packages/ray_lightning/ray_ddp.py", line 170, in _worker_setup
        self._process_group_backend = self._get_process_group_backend()
      File "/home/gridsan/dgraff/.conda/envs/molpal/lib/python3.8/site-packages/pytorch_lightning/strategies/ddp_spawn.py", line 166, in _get_process_group_backend
        or get_default_process_group_backend_for_device(self.root_device)
      File "/home/gridsan/dgraff/.conda/envs/molpal/lib/python3.8/site-packages/ray_lightning/ray_ddp.py", line 295, in root_device
        cuda_visible_list = [
      File "/home/gridsan/dgraff/.conda/envs/molpal/lib/python3.8/site-packages/ray_lightning/ray_ddp.py", line 296, in <listcomp>
        int(dev) for dev in cuda_visible_str.split(",")
    ValueError: invalid literal for int() with base 10: 'GPU-dade4b6e-8461-eee0-e8bb-4f7e570856f4'
    

    It seems like the internal code relies on an ordinal GPU device naming scheme. I.e.,

    $ echo $CUDA_VISIBLE_DEVICES
    0,1
    

    which seems reasonable, given that what I typically encounter on most systems. But on my system, the GPU device naming looks something like this:

    $ echo $CUDA_VISIBLE_DEVICES
    GPU-23c5e712-9b16-e21a-df00-7dab564ade42,GPU-cdaae969-b14c-6b80-2fa2-de8e9efe87a1
    

    So it seems like there are two options:

    1. I could ask my sys-admins to rename the GPUs on the cluster to the more "standard" ordinal scheme. They'll probably tell me "No." and reference the CUDA_VISIBLE_DEVICES specification where it states that device names of the form GPU-<UUID> is the second option in addition to integer indices
    2. This block of code is altered ray_lightning/ray_ddp.py#L292:
    gpu_id = ray.get_gpu_ids()[0]  # NOTE: this value is cast to `int(...)` in the main branch. The could would break _here_ in the current code but breaks later v0.3
    cuda_visible_str = os.environ.get("CUDA_VISIBLE_DEVICES", "")
    if cuda_visible_str and cuda_visible_str != "NoDevFiles":
        cuda_visible_list = [
            int(dev) for dev in cuda_visible_str.split(",")
        ]
        device_id = cuda_visible_list.index(gpu_id)
        return torch.device("cuda", device_id)
    

    I think the block should be changed to:

    gpu_id = ray.get_gpu_ids()[0]
    cuda_visible_str = os.environ.get("CUDA_VISIBLE_DEVICES", "")
    if cuda_visible_str and cuda_visible_str != "NoDevFiles":
        cuda_visible_list = list(cuda_visible_str.split(","))
        device_id = cuda_visible_list.index(gpu_id)
        return torch.device("cuda", device_id)
    

    Thanks for the great work so far!

    opened by davidegraff 1
  • TypeError in a SLURM environment due to internal API break

    TypeError in a SLURM environment due to internal API break

    Using the master branch of ray-lightning with pytorch-lightning v1.6 in a SLURM environment leads to following exception:

    ray.exceptions.RayTaskError(TypeError): ray::ImplicitFunc.train() (pid=117539, ip=10.181.76.37, repr=train)
      File ".../lib/python3.9/site-packages/ray/tune/trainable/trainable.py", line 367, in train
        raise skipped from exception_cause(skipped)
      File ".../lib/python3.9/site-packages/ray/tune/trainable/function_trainable.py", line 335, in entrypoint
        return self._trainable_func(
      File ".../lib/python3.9/site-packages/ray/tune/trainable/function_trainable.py", line 652, in _trainable_func
        output = fn()
      File ".../random_search.py", line 122, in train
        trainer = Trainer(
      File ".../lib/python3.9/site-packages/pytorch_lightning/utilities/argparse.py", line 339, in insert_env_defaults
        return fn(self, **kwargs)
      File ".../lib/python3.9/site-packages/pytorch_lightning/trainer/trainer.py", line 485, in __init__
        self._accelerator_connector = AcceleratorConnector(
      File ".../lib/python3.9/site-packages/pytorch_lightning/trainer/connectors/accelerator_connector.py", line 204, in __init__
        self.cluster_environment: ClusterEnvironment = self._choose_and_init_cluster_environment()
      File ".../lib/python3.9/site-packages/pytorch_lightning/trainer/connectors/accelerator_connector.py", line 549, in _choose_and_init_cluster_environment
        if self._is_slurm_managing_tasks():
      File ".../lib/python3.9/site-packages/pytorch_lightning/trainer/connectors/accelerator_connector.py", line 562, in _is_slurm_managing_tasks
        total_requested_devices = len(self._parallel_devices) * self._num_nodes_flag
    TypeError: object of type 'NoneType' has no len()
    

    The _GPUAccelerator.get_parallel_devices method breaks the internal Pytorch Lightning API by returning None in some cases, is this intentional? Returning an empty List instead of None fixes my issue, but I don't know if None is required in other ray-lightning use cases.

    I would be more than happy to provide a PR if you think the fix is fine.

    Thank you for this very convenient package and keep up the fantastic work!

    opened by dcfidalgo 2
Releases(v0.3.0)
  • v0.3.0(Aug 23, 2022)

    What's Changed

    • Bump version for development by @amogkam in https://github.com/ray-project/ray_lightning/pull/122
    • Update README to render on Ray docs by @amogkam in https://github.com/ray-project/ray_lightning/pull/135
    • Fix bash code block in Readme by @Yard1 in https://github.com/ray-project/ray_lightning/pull/136
    • Fix for fractional GPU by @amogkam in https://github.com/ray-project/ray_lightning/pull/125
    • Update broken PTL link by @amogkam in https://github.com/ray-project/ray_lightning/pull/137
    • Fix hanging trainer.test() by @amogkam in https://github.com/ray-project/ray_lightning/pull/142
    • Fix ray_ddp_sharded_example by @chongxiaoc in https://github.com/ray-project/ray_lightning/pull/153
    • Pop kwargs to support LightningCLI by @amogkam in https://github.com/ray-project/ray_lightning/pull/154
    • ray_ddp: support logged_metrics as part of remote worker return value by @chongxiaoc in https://github.com/ray-project/ray_lightning/pull/156
    • Support PyTorch Lightning 1.6 by @JiahaoYao in https://github.com/ray-project/ray_lightning/pull/163
    • Fix docs formatting by @JiahaoYao in https://github.com/ray-project/ray_lightning/pull/188
    • fix issue #189 by @JiahaoYao in https://github.com/ray-project/ray_lightning/pull/190
    • [Ray lightning 1.6] update the change according to the comment in #163 by @JiahaoYao in https://github.com/ray-project/ray_lightning/pull/195

    New Contributors

    • @Yard1 made their first contribution in https://github.com/ray-project/ray_lightning/pull/136
    • @chongxiaoc made their first contribution in https://github.com/ray-project/ray_lightning/pull/153
    • @JiahaoYao made their first contribution in https://github.com/ray-project/ray_lightning/pull/163

    Full Changelog: https://github.com/ray-project/ray_lightning/compare/0.2.0...v0.3.0

    Source code(tar.gz)
    Source code(zip)
    ray_lightning-0.3.0-py3-none-any.whl(49.62 KB)
  • 0.2.0(Feb 2, 2022)

    • Support for PyTorch Lightning v1.5 (#115, #121)!
    • Update HorovodRayPlugin API to match the new Horovod on Ray API. num_hosts and num_slots args have been deprecated in favor of a generic num_workers arg (#71).
    • get_tune_ddp_resouces has been renamed to get_tune_resources and can now be used for both RayPlugin and HorovodRayPlugin (#71).
    • Rename the cpus_per_worker arg in get_tune_resources utility to num_cpus_per_worker to match the arg name in RayPlugin (#96).
    • Annotate the APIs as beta (#88).
    Source code(tar.gz)
    Source code(zip)
  • 0.1.1(Aug 20, 2021)

  • 0.1.0(Aug 12, 2021)

QAHOI: Query-Based Anchors for Human-Object Interaction Detection (paper)

QAHOI QAHOI: Query-Based Anchors for Human-Object Interaction Detection (paper) Requirements PyTorch = 1.5.1 torchvision = 0.6.1 pip install -r requ

38 Dec 29, 2022
A full-fledged version of Pix2Seq

Stable-Pix2Seq A full-fledged version of Pix2Seq What it is. This is a full-fledged version of Pix2Seq. Compared with unofficial-pix2seq, stable-pix2s

peng gao 205 Dec 27, 2022
Neural Point-Based Graphics

Neural Point-Based Graphics Project   Video   Paper Neural Point-Based Graphics Kara-Ali Aliev1 Artem Sevastopolsky1,2 Maria Kolos1,2 Dmitry Ulyanov3

Ali Aliev 252 Dec 13, 2022
A particular navigation route using satellite feed and can help in toll operations & traffic managemen

How about adding some info that can quanitfy the stress on a particular navigation route using satellite feed and can help in toll operations & traffic management The current analysis is on the satel

Ashish Pandey 1 Feb 14, 2022
Python Implementation of algorithms in Graph Mining, e.g., Recommendation, Collaborative Filtering, Community Detection, Spectral Clustering, Modularity Maximization, co-authorship networks.

Graph Mining Author: Jiayi Chen Time: April 2021 Implemented Algorithms: Network: Scrabing Data, Network Construbtion and Network Measurement (e.g., P

Jiayi Chen 3 Mar 03, 2022
Expressive Body Capture: 3D Hands, Face, and Body from a Single Image

Expressive Body Capture: 3D Hands, Face, and Body from a Single Image [Project Page] [Paper] [Supp. Mat.] Table of Contents License Description Fittin

Vassilis Choutas 1.3k Jan 07, 2023
Single-Stage Instance Shadow Detection with Bidirectional Relation Learning (CVPR 2021 Oral)

Single-Stage Instance Shadow Detection with Bidirectional Relation Learning (CVPR 2021 Oral) Tianyu Wang*, Xiaowei Hu*, Chi-Wing Fu, and Pheng-Ann Hen

Steve Wong 51 Oct 20, 2022
A library for researching neural networks compression and acceleration methods.

A library for researching neural networks compression and acceleration methods.

Intel Labs 100 Dec 29, 2022
Provably Rare Gem Miner.

Provably Rare Gem Miner just another random project by yoyoismee.eth useful link main site market contract useful thing you should know read contract

34 Nov 22, 2022
Rede Neural Convolucional feita durante o processo seletivo do Laboratório de Inteligência Artificial da FACOM (UFMS)

Primeira_Rede_Neural_Convolucional Rede Neural Convolucional feita durante o processo seletivo do Laboratório de Inteligência Artificial da FACOM (UFM

Roney_Felipe 1 Jan 13, 2022
tensorflow implementation of 'YOLO : Real-Time Object Detection'

YOLO_tensorflow (Version 0.3, Last updated :2017.02.21) 1.Introduction This is tensorflow implementation of the YOLO:Real-Time Object Detection It can

Jinyoung Choi 1.7k Nov 21, 2022
DeepFashion2 is a comprehensive fashion dataset.

DeepFashion2 Dataset DeepFashion2 is a comprehensive fashion dataset. It contains 491K diverse images of 13 popular clothing categories from both comm

switchnorm 1.8k Jan 07, 2023
PURE: End-to-End Relation Extraction

PURE: End-to-End Relation Extraction This repository contains (PyTorch) code and pre-trained models for PURE (the Princeton University Relation Extrac

Princeton Natural Language Processing 657 Jan 09, 2023
Load What You Need: Smaller Multilingual Transformers for Pytorch and TensorFlow 2.0.

Smaller Multilingual Transformers This repository shares smaller versions of multilingual transformers that keep the same representations offered by t

Geotrend 79 Dec 28, 2022
Classifying cat and dog images using Kaggle dataset

PyTorch Image Classification Classifies an image as containing either a dog or a cat (using Kaggle's public dataset), but could easily be extended to

Robert Coleman 74 Nov 22, 2022
Deeplab-resnet-101 in Pytorch with Jaccard loss

Deeplab-resnet-101 Pytorch with Lovász hinge loss Train deeplab-resnet-101 with binary Jaccard loss surrogate, the Lovász hinge, as described in http:

Maxim Berman 95 Apr 15, 2022
A GPU-optional modular synthesizer in pytorch, 16200x faster than realtime, for audio ML researchers.

torchsynth The fastest synth in the universe. Introduction torchsynth is based upon traditional modular synthesis written in pytorch. It is GPU-option

torchsynth 229 Jan 02, 2023
Unofficial PyTorch implementation of Fastformer based on paper "Fastformer: Additive Attention Can Be All You Need"."

Fastformer-PyTorch Unofficial PyTorch implementation of Fastformer based on paper Fastformer: Additive Attention Can Be All You Need. Usage : import t

Hong-Jia Chen 126 Dec 06, 2022
MicroNet: Improving Image Recognition with Extremely Low FLOPs (ICCV 2021)

MicroNet: Improving Image Recognition with Extremely Low FLOPs (ICCV 2021) A pytorch implementation of MicroNet. If you use this code in your research

Yunsheng Li 293 Dec 28, 2022
[CVPR'20] TTSR: Learning Texture Transformer Network for Image Super-Resolution

TTSR Official PyTorch implementation of the paper Learning Texture Transformer Network for Image Super-Resolution accepted in CVPR 2020. Contents Intr

Multimedia Research 689 Dec 28, 2022