Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Use current device in _as_gpu #3586

Merged
merged 1 commit into from
Dec 21, 2021
Merged

Use current device in _as_gpu #3586

merged 1 commit into from
Dec 21, 2021

Conversation

klecki
Copy link
Contributor

@klecki klecki commented Dec 20, 2021

Signed-off-by: Krzysztof Lecki klecki@nvidia.com

Description

  • Bug fix (non-breaking change which fixes an issue)
  • New feature (non-breaking change which adds functionality)
  • Breaking change (fix or feature that would cause existing functionality to not work as expected)
  • Refactoring (Redesign of existing code that doesn't affect functionality)
  • Other (e.g. Documentation, Tests, Configuration)

What happened in this PR

In the internal Tensor[List]CPU._as_gpu used in debug mode
we used stream tied with the device of the newly created Tensor[List]GPU,
which is not initialized to any of current GPU.
Use the current GPU device for the returned Tensor[List]GPU and
do the copy & wait on the stream associated with that device.

Additional information

  • Affected modules and functionalities:
    Python/Backend/Debug Mode.

  • Key points relevant for the review:

Checklist

Tests

  • Existing tests apply
  • New tests added
    • Python tests
    • GTests
    • Benchmark
    • Other
  • N/A

Documentation

  • Existing documentation applies
  • Documentation updated
    • Docstring
    • Doxygen
    • RST
    • Jupyter
    • Other
  • N/A

DALI team only

Requirements

  • Implements new requirements
  • Affects existing requirements
  • N/A

REQ IDs: N/A

JIRA TASK: N/A

Signed-off-by: Krzysztof Lecki <klecki@nvidia.com>
@klecki
Copy link
Contributor Author

klecki commented Dec 20, 2021

!build

@dali-automaton
Copy link
Collaborator

CI MESSAGE: [3628219]: BUILD STARTED

@JanuszL JanuszL self-assigned this Dec 20, 2021
@dali-automaton
Copy link
Collaborator

CI MESSAGE: [3628219]: BUILD PASSED

@klecki klecki merged commit 396ff86 into NVIDIA:main Dec 21, 2021
cyyever pushed a commit to cyyever/DALI that referenced this pull request Jan 23, 2022
In the internal `Tensor[List]CPU._as_gpu` used in debug mode
we used stream tied with the device of the newly created `Tensor[List]GPU`,
which is not initialized to any of current GPU.
Use the current GPU device for the returned `Tensor[List]GPU` and 
do the copy & wait on the stream associated with that device.

Signed-off-by: Krzysztof Lecki <klecki@nvidia.com>
cyyever pushed a commit to cyyever/DALI that referenced this pull request Jan 23, 2022
In the internal `Tensor[List]CPU._as_gpu` used in debug mode
we used stream tied with the device of the newly created `Tensor[List]GPU`,
which is not initialized to any of current GPU.
Use the current GPU device for the returned `Tensor[List]GPU` and 
do the copy & wait on the stream associated with that device.

Signed-off-by: Krzysztof Lecki <klecki@nvidia.com>
cyyever pushed a commit to cyyever/DALI that referenced this pull request Jan 23, 2022
In the internal `Tensor[List]CPU._as_gpu` used in debug mode
we used stream tied with the device of the newly created `Tensor[List]GPU`,
which is not initialized to any of current GPU.
Use the current GPU device for the returned `Tensor[List]GPU` and 
do the copy & wait on the stream associated with that device.

Signed-off-by: Krzysztof Lecki <klecki@nvidia.com>
cyyever pushed a commit to cyyever/DALI that referenced this pull request Jan 23, 2022
In the internal `Tensor[List]CPU._as_gpu` used in debug mode
we used stream tied with the device of the newly created `Tensor[List]GPU`,
which is not initialized to any of current GPU.
Use the current GPU device for the returned `Tensor[List]GPU` and 
do the copy & wait on the stream associated with that device.

Signed-off-by: Krzysztof Lecki <klecki@nvidia.com>
cyyever pushed a commit to cyyever/DALI that referenced this pull request Jan 23, 2022
In the internal `Tensor[List]CPU._as_gpu` used in debug mode
we used stream tied with the device of the newly created `Tensor[List]GPU`,
which is not initialized to any of current GPU.
Use the current GPU device for the returned `Tensor[List]GPU` and 
do the copy & wait on the stream associated with that device.

Signed-off-by: Krzysztof Lecki <klecki@nvidia.com>
cyyever pushed a commit to cyyever/DALI that referenced this pull request Jan 23, 2022
In the internal `Tensor[List]CPU._as_gpu` used in debug mode
we used stream tied with the device of the newly created `Tensor[List]GPU`,
which is not initialized to any of current GPU.
Use the current GPU device for the returned `Tensor[List]GPU` and 
do the copy & wait on the stream associated with that device.

Signed-off-by: Krzysztof Lecki <klecki@nvidia.com>
cyyever pushed a commit to cyyever/DALI that referenced this pull request Jan 23, 2022
In the internal `Tensor[List]CPU._as_gpu` used in debug mode
we used stream tied with the device of the newly created `Tensor[List]GPU`,
which is not initialized to any of current GPU.
Use the current GPU device for the returned `Tensor[List]GPU` and 
do the copy & wait on the stream associated with that device.

Signed-off-by: Krzysztof Lecki <klecki@nvidia.com>
cyyever pushed a commit to cyyever/DALI that referenced this pull request Jan 23, 2022
In the internal `Tensor[List]CPU._as_gpu` used in debug mode
we used stream tied with the device of the newly created `Tensor[List]GPU`,
which is not initialized to any of current GPU.
Use the current GPU device for the returned `Tensor[List]GPU` and 
do the copy & wait on the stream associated with that device.

Signed-off-by: Krzysztof Lecki <klecki@nvidia.com>
cyyever pushed a commit to cyyever/DALI that referenced this pull request Jan 23, 2022
In the internal `Tensor[List]CPU._as_gpu` used in debug mode
we used stream tied with the device of the newly created `Tensor[List]GPU`,
which is not initialized to any of current GPU.
Use the current GPU device for the returned `Tensor[List]GPU` and 
do the copy & wait on the stream associated with that device.

Signed-off-by: Krzysztof Lecki <klecki@nvidia.com>
cyyever pushed a commit to cyyever/DALI that referenced this pull request Jan 23, 2022
In the internal `Tensor[List]CPU._as_gpu` used in debug mode
we used stream tied with the device of the newly created `Tensor[List]GPU`,
which is not initialized to any of current GPU.
Use the current GPU device for the returned `Tensor[List]GPU` and 
do the copy & wait on the stream associated with that device.

Signed-off-by: Krzysztof Lecki <klecki@nvidia.com>
cyyever pushed a commit to cyyever/DALI that referenced this pull request Jan 23, 2022
In the internal `Tensor[List]CPU._as_gpu` used in debug mode
we used stream tied with the device of the newly created `Tensor[List]GPU`,
which is not initialized to any of current GPU.
Use the current GPU device for the returned `Tensor[List]GPU` and 
do the copy & wait on the stream associated with that device.

Signed-off-by: Krzysztof Lecki <klecki@nvidia.com>
cyyever pushed a commit to cyyever/DALI that referenced this pull request Jan 23, 2022
In the internal `Tensor[List]CPU._as_gpu` used in debug mode
we used stream tied with the device of the newly created `Tensor[List]GPU`,
which is not initialized to any of current GPU.
Use the current GPU device for the returned `Tensor[List]GPU` and 
do the copy & wait on the stream associated with that device.

Signed-off-by: Krzysztof Lecki <klecki@nvidia.com>
cyyever pushed a commit to cyyever/DALI that referenced this pull request Jan 23, 2022
In the internal `Tensor[List]CPU._as_gpu` used in debug mode
we used stream tied with the device of the newly created `Tensor[List]GPU`,
which is not initialized to any of current GPU.
Use the current GPU device for the returned `Tensor[List]GPU` and 
do the copy & wait on the stream associated with that device.

Signed-off-by: Krzysztof Lecki <klecki@nvidia.com>
cyyever pushed a commit to cyyever/DALI that referenced this pull request Feb 21, 2022
In the internal `Tensor[List]CPU._as_gpu` used in debug mode
we used stream tied with the device of the newly created `Tensor[List]GPU`,
which is not initialized to any of current GPU.
Use the current GPU device for the returned `Tensor[List]GPU` and 
do the copy & wait on the stream associated with that device.

Signed-off-by: Krzysztof Lecki <klecki@nvidia.com>
cyyever pushed a commit to cyyever/DALI that referenced this pull request May 13, 2022
In the internal `Tensor[List]CPU._as_gpu` used in debug mode
we used stream tied with the device of the newly created `Tensor[List]GPU`,
which is not initialized to any of current GPU.
Use the current GPU device for the returned `Tensor[List]GPU` and 
do the copy & wait on the stream associated with that device.

Signed-off-by: Krzysztof Lecki <klecki@nvidia.com>
cyyever pushed a commit to cyyever/DALI that referenced this pull request Jun 7, 2022
In the internal `Tensor[List]CPU._as_gpu` used in debug mode
we used stream tied with the device of the newly created `Tensor[List]GPU`,
which is not initialized to any of current GPU.
Use the current GPU device for the returned `Tensor[List]GPU` and 
do the copy & wait on the stream associated with that device.

Signed-off-by: Krzysztof Lecki <klecki@nvidia.com>
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

4 participants