diff --git a/check_gpu.py b/check_gpu.py deleted file mode 100644 index f8daa9e..0000000 --- a/check_gpu.py +++ /dev/null @@ -1,18 +0,0 @@ -import torch - -if torch.cuda.is_available(): - print(f"CUDA is available. GPU: {torch.cuda.get_device_name(0)}") - print(f"PyTorch CUDA version: {torch.version.cuda}") -else: - print("CUDA is not available. PyTorch is running on CPU.") - -# Check torchvision -try: - import torchvision - - print(f"Torchvision version: {torchvision.__version__}") - # You might need specific torchvision ops to fully check GPU integration, - # but successful import and version check is a good start. - print("Torchvision imported successfully.") -except ImportError: - print("Torchvision could not be imported.")