Skip to content
This repository has been archived by the owner on Jan 3, 2023. It is now read-only.

Change GPU check #19

Closed
Kaixhin opened this issue May 8, 2015 · 4 comments
Closed

Change GPU check #19

Kaixhin opened this issue May 8, 2015 · 4 comments

Comments

@Kaixhin
Copy link
Contributor

Kaixhin commented May 8, 2015

I've been making Docker images for deep learning libraries, with CPU-only and CUDA-enabled versions. I've successfully made the base version of neon , but my CUDA version doesn't get built properly. The issue in the Makefile is that nvidia-smi is used to check that a CUDA-enabled GPU is available, and whilst this isn't available in Docker build environments it has been possible to build libraries with the installed SDK. If a GPU isn't needed for the build, is it possible to change this test to check for something that is e.g. nvcc --version?

scttl added a commit that referenced this issue May 11, 2015
Change GPU check to CUDA SDK check. Closes issue #19
@scttl
Copy link
Contributor

scttl commented May 11, 2015

Yes this seems like a reasonable alternative check. Thanks for the commit and producing Docker images!

@scttl scttl closed this as completed May 11, 2015
@Kaixhin
Copy link
Contributor Author

Kaixhin commented May 13, 2015

Just following up since I've found time to run some models. The CUDA version fails on Kepler GPUs (as expected) but works beautifully on Maxwell GPUs!

@scttl
Copy link
Contributor

scttl commented May 13, 2015

Great to hear!

As a suggestion you may want to rename the docker image to make it clear that it is Maxwell only since you're using the nervanagpu backend.

In fact it would probably good to have another docker image based off our cudanet GPU backend since this will work on Kepler cards too.

@Kaixhin
Copy link
Contributor Author

Kaixhin commented May 14, 2015

Good point - I've now replaced the CUDA version with nervanagpu-neon and cudanet-neon.

As far as I can tell it isn't necessary to install cuda-convnet2 beforehand as the neon installer does it. One weird issue I've had on doing local builds is that when attempting to do a multithreaded make on the cudanet repo it works fine, but it consistently fails when run on the neon repo with the cudanet flag. I'm not going to raise an issue but if you wish to investigate further the most pertinent log I saw was

read jobs pipe: Resource temporarily unavailable. Stop

which looks like somewhere along the line pip isn't handling memory well.

Sign up for free to subscribe to this conversation on GitHub. Already have an account? Sign in.
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants