Questions tagged [nvidia]

23 questions
13
votes
1 answer

How to make my Neural Netwok run on GPU instead of CPU

I have installed Anaconda3 and have installed latest versions of Keras and Tensorflow. Running this command : from tensorflow.python.client import device_lib print(device_lib.list_local_devices()) I find the Notebook is running in CPU: [name:…
Deni Avinash
  • 133
  • 1
  • 1
  • 5
10
votes
2 answers

interpret results of nvidia-smi

Every 1.0s: nvidia-smi Tue Feb 20 12:49:34 2018 Tue Feb 20 12:49:34…
kRazzy R
  • 203
  • 2
  • 7
3
votes
2 answers

What does images per second mean when benchmarking Deep Learning GPU?

I've been reviewing performance of several NVIDIA GPU's and I see that typically results are presented in terms of "images per second" that can be processed. Experiments are typically being performed on classical network architectures such as Alex…
Javierfdr
  • 1,490
  • 12
  • 14
2
votes
1 answer

Is Nvidia Jetson product family also suitable for machine learing model training?

I recently came accross these products (Nvidia Jetson) and they are all tagged as "edge", so i think they are designed only for machine learning inference and not model training. They are quite interesting for their low power consumpion and price…
Skary
  • 125
  • 5
2
votes
1 answer

What are "belief maps" and "affinity maps"?

When evaluating Nvidia-Deep Object Pose Estimation, I came across these terms, belief maps and affinity maps. I haven't been able to find a satisfying answer online on what these terms mean.
rajput
  • 31
  • 3
2
votes
1 answer

How can I monitor the usage of NVLink connections?

If I'm running a Keras model on some Nvidia GPUs which are connected via NVLink, how can I monitor the usage of the NVLink connections? I want to be sure the NVLink connections are being used and see at what rate.
user3731622
  • 121
  • 1
  • 3
2
votes
0 answers

Is attention cache useful during transformer pretraining?

I am looking at the MegatronLM implementation, and the only thing that is cached are the results of xK and xV computation: https://github.com/NVIDIA/Megatron-LM/blob/b44dca25727c294a7f825e74a3c4a53744cc8404/megatron/model/transformer.py#L339 Which…
LOST
  • 131
  • 1
1
vote
1 answer

DIGITS Docker container not picking up GPU

I am running DIGITS Docker container but for some reason it fails to recognize host's GPU: it does not report any GPUs (where I expect 1 to be reported) so in the upper right corner of the DIGITS home page there is no indication of any GPUs and also…
1
vote
1 answer

Not able to connect to GPU on Google Colab

I'm trying to use tensorflow with a GPU on Google Colab. I followed the steps listed at https://www.tensorflow.org/install/gpu I confirmed that gpu is visible and CUDA is installed with the commands - !nvcc --version !nvidia-smi This works as…
anirudh
  • 143
  • 1
  • 1
  • 5
1
vote
0 answers

Why doesn't this CNN model need fetures for reducing overfitting?

I found this CNN model by Nvidia end-to-end-deeplearning and with training this model, I'm wondering why this model doesn't need to have dropout layers to reduce overfitting. Neither, this doesn't have activation function. I know we can tune the…
Yuki.U
  • 83
  • 5
1
vote
1 answer

Two different GPUs for Keras (Python)?

One question guys, someone knows if it should be ok to get one more GPU of type Nvidia Geforce GTX 1070 (gaming version), given that now I have GTX 1070 Titanium? They don't have another Titanium card available here, so I have to get a different…
user68541
1
vote
1 answer

How to setup my Ubuntu PC for these Deep Learning frameworks?

I need to use both NVIDIA Digits and tensorflow Object Detection API for different deep learning purposes. I am a bit lost with the compatibility issues, especially for what concerns CUDA. So which version should I choose for DIGITS, Tensorflow-gpu,…
firion
  • 199
  • 4
1
vote
1 answer

Does it make sense to parallelize machine learning algorithms as part of PhD research?

I'm developing machine learning algorithms to aid in the diagnosis and prognosis of various cancers for my PhD. My lab is an Nvidia teaching center (CUDA). My supervisor thinks that I need to also optimize ML by parallelizing it in CUDA. However,…
Sterls
  • 160
  • 1
  • 7
1
vote
0 answers

Huge variance for RandomForestRegressor models

The experiment is the following: train a RFR with a 15k train rows get predictions on 8k test rows, save predictions as y_hat0 remove 1 random row from the training set and retrain the RFR save prediction for the newly trained model as…
Oleg
  • 11
  • 3
1
vote
0 answers

Is there a reason not to wirk with AMP (automatic mixed precision)?

According to: Introducing native PyTorch automatic mixed precision for faster training on NVIDIA GPUs It's better to use AMP (with 16 Floating Point) due to: Shorter training time. Lower memory requirements, enabling larger batch sizes, larger…
user3668129
  • 363
  • 2
  • 11
1
2