Tag: GeForce GTX

Install cp2k with cuda in a geforce gtx 1650 card

Hello,  I am trying to install cp2k with cuda enable in a  geforce gtx 1650 card, but in the gpu-ver –gpu-ver                 Selects the GPU architecture for which to compile. Available                          …

Continue Reading Install cp2k with cuda in a geforce gtx 1650 card

Reinstalling GROMACS with CUDA GPU – User discussions

GROMACS version: 2023.2GROMACS modification: Yes/NoHere post your question $ nvcc –versionnvcc: NVIDIA (R) Cuda compiler driverCopyright (c) 2005-2021 NVIDIA CorporationBuilt on Thu_Nov_18_09:45:30_PST_2021Cuda compilation tools, release 11.5, V11.5.119Build cuda_11.5.r11.5/compiler.30672275_0 $ nvidia-smiTue Oct 3 13:42:40 2023±————————————————————————————–+| NVIDIA-SMI 535.103 Driver Version: 537.13 CUDA Version: 12.2 ||—————————————–±———————±———————+| GPU Name Persistence-M | Bus-Id Disp.A |…

Continue Reading Reinstalling GROMACS with CUDA GPU – User discussions

CUDA error: CUBLAS_STATUS_EXECUTION_FAILED when using roberta – nlp

Hi everyone, I’m training a model that uses roberta to encoder sentence. I got an error that is:RuntimeError: CUDA error: CUBLAS_STATUS_EXECUTION_FAILED when calling cublasGemmStridedBatchedExFix( handle, opa, opb, m, n, k, (void*)(&falpha), a, CUDA_R_16F, lda, stridea, b, CUDA_R_16F, ldb, strideb, (void*)(&fbeta), c, CUDA_R_16F, ldc, stridec, num_batches, CUDA_R_32F, CUBLAS_GEMM_DEFAULT_TENSOR_OP) My problem is…

Continue Reading CUDA error: CUBLAS_STATUS_EXECUTION_FAILED when using roberta – nlp

Minimum and Mean Euclidean distance between two tensors of different shape

I am quite new to Pytorch and currently running into issues with Memory Overflow. Task:I have two 2D tensors of respective shapes A: [1000, 14] & B: [100000, 14]. I have to find the distance of each row of tensor-A from all rows from tensor-B. Later using the calculated distance…

Continue Reading Minimum and Mean Euclidean distance between two tensors of different shape

Step-by-Step Guide to Setup Pytorch for Your GPU on Windows 10/11

In this competitive world of technology, Machine Learning and Artificial Intelligence technologies have emerged as a breakthrough for developing advanced AI applications like image recognition, natural language processing, speech translation, and more. However, developing such AI-powered applications would require massive amounts of computational power far beyond the capabilities of CPUs…

Continue Reading Step-by-Step Guide to Setup Pytorch for Your GPU on Windows 10/11

python – Old pytorch version on new gpu

I am trying to setup a project from github. It is based on python 3.8 and pytorch 1.6. The gpu I am using is NVIDIA GeForce RTX 3080 Ti GPU with Driver Version: 535.54.03 and CUDA Version: 12.2 according to nvidia-smi. (Using torch 13.1 torch.version.cuda shows 11.7. I could run…

Continue Reading python – Old pytorch version on new gpu

KeyError: ‘No action for destination key “trainer.devices” to set its default.’ – Trainer

I have been creating a wrapper for CLIP using pytorch-lightning in order to finetune CLIP on Image Remote Sesing Captioning. I have been using LightningCLI to handle the argument parsing for the model, the trainer etc. I am finetuning CLIP using RSICD dataset. When I start the training, something strange…

Continue Reading KeyError: ‘No action for destination key “trainer.devices” to set its default.’ – Trainer

Nvcc fatal : Unsupported gpu architecture ‘compute_89’ – CUDA Setup and Installation

After installing cuda in my laptop it apparently is still missing the support for the specific GPU, on ubuntu 22.04. I have followed the instructions in CUDA Toolkit 12.1 Update 1 Downloads | NVIDIA Developer When running some of the examples I get a message like this: cuda-samples/Samples/6_Performance/LargeKernelParameter$ make/usr/bin/nvcc -ccbin…

Continue Reading Nvcc fatal : Unsupported gpu architecture ‘compute_89’ – CUDA Setup and Installation

Torch.compile, Triton cuda capability – compile

Hello everyone,Since i am working with PINN, the new capabilities of pytorch2.0 is very interesting.On making the transition to pytorch 2.0, the torch.compile seemed to work fine but on calling the model i get the following error, “torch._dynamo.exc.BackendCompilerFailed: debug_wrapper raised RuntimeError: Found NVIDIA GeForce GTX 1080 Ti which is too…

Continue Reading Torch.compile, Triton cuda capability – compile

python – Pytorch cannot open libgomp-a34b3233.so.1

I have been using PyTorch for year now and out of nowhere, couple days ago I got this error when importing torch: >>> import torch Traceback (most recent call last): File “<stdin>”, line 1, in <module> File “/home/jechos/.local/lib/python3.8/site-packages/torch/__init__.py”, line 201, in <module> _load_global_deps() File “/home/jechos/.local/lib/python3.8/site-packages/torch/__init__.py”, line 154, in _load_global_deps ctypes.CDLL(lib_path,…

Continue Reading python – Pytorch cannot open libgomp-a34b3233.so.1

E5-2666v4-x10drl-ct-2023-06-07 Lammps Performance – OpenBenchmarking.org

Intel Xeon E5-2666 v4 testing with a Supermicro X10DRL-CT v0123456789 (3.2 BIOS) and NVIDIA GeForce GTX 1080 Ti 11GB on Ubuntu 20.04 via the Phoronix Test Suite. Compare your own system(s) to this result file with the Phoronix Test Suite by running the command: phoronix-test-suite benchmark 2306074-NE-E52666V4X88 Intel Xeon E5-2666…

Continue Reading E5-2666v4-x10drl-ct-2023-06-07 Lammps Performance – OpenBenchmarking.org

QLoRa, Reprompting, Retrieval Augmented Generation, Triton PyTorch Backend, Google Dialogflow ES in Twilio, PointNet Data Preparation

👋 Hey, “Causal inference is the key to unlocking the true value of data in making informed decisions.” – Judea Pearl, Computer Scientist  Inferences of causality allow us to understand cause-and-effect relationships, uncover hidden connections, predict outcomes, and design effective interventions, thereby guiding us in informed decision-making. Without it, we would…

Continue Reading QLoRa, Reprompting, Retrieval Augmented Generation, Triton PyTorch Backend, Google Dialogflow ES in Twilio, PointNet Data Preparation

How to run detectron2, as python embedded code in C++, on GPU?

I want to run detectron2 prediction as a Python Code embedded in a WebRTC C++ code. I am using webrtc-streamer as a C++ server in which I want to embed detectron2 Python code to call a prediction function. Everything works fine when I run the executable of webrtc-streamer and call…

Continue Reading How to run detectron2, as python embedded code in C++, on GPU?

Why is CUDA unavailable in anaconda environment with pytorch even though its installed successfully, and works in python?

I Installed CUDA on my system (windows 10), then also installed pytorch. I then created an Anaconda environment and installed the same into it as well. Now my gpu is recognized and is set to available in torch when running through python. and the necessary cuda installation is there within…

Continue Reading Why is CUDA unavailable in anaconda environment with pytorch even though its installed successfully, and works in python?

Cuda illegal memory access(kokkos) when using multiple GPUs – LAMMPS Development

Dear all, I have encountered cuda illegal memory access(lib kokkos) when using multiple GPUs. The system is mixture of 2 beads, 3 beads, and 100 beads chains with harmonic bond and angle potential style. The atom style is set as angle and the pair style is lj/expand. The exact same…

Continue Reading Cuda illegal memory access(kokkos) when using multiple GPUs – LAMMPS Development

machine learning – pytorch cannot built running processes on Tesla V100-PCIE-32GB (type: (7, 0))

On remote server, PyTorch can access the GPU that can place variables and models on GPUs, but the GPU shows no running processes found on the GPU. Thu Apr 20 09:21:59 2023 +—————————————————————————–+ | NVIDIA-SMI 511.65 Driver Version: 452.39 CUDA Version: 11.0 | |——————————-+———————-+———————-+ | GPU Name TCC/WDDM | Bus-Id…

Continue Reading machine learning – pytorch cannot built running processes on Tesla V100-PCIE-32GB (type: (7, 0))

Why cast and unpack take so long – LAMMPS General Discussion

dxu17 March 13, 2023, 12:39am 1 I have run a lammps simjulation.The outupt is very strange. Device Time Info (average): Data Transfer: 126.2653 s.Neighbor copy: 0.3334 s.Neighbor build: 6.2128 s.Force calc: 38.1705 s.Device Overhead: 0.0872 s.Average split: 1.0000.Lanes / atom: 4.Vector width: 32.Max Mem / Proc: 9171.50 MB.CPU Neighbor: 7.9423…

Continue Reading Why cast and unpack take so long – LAMMPS General Discussion

python – PyTorch working in Miniconda but “not compiled with CUDA enabled” in PyCharm

I have had quite the journey trying to get PyCharm to use my GPU (NVIDIA GeForce GTX 1080 ti) when running code from this github: github.com/gordicaleksa/pytorch-neural-style-transfer After a whole lot of back and forth setting up CUDA, cuDNN etc., I have finally got PyTorch working (pretty sure) in my Miniconda…

Continue Reading python – PyTorch working in Miniconda but “not compiled with CUDA enabled” in PyCharm