suisuyy commited on
Commit
00f1499
1 Parent(s): 09dab11
Files changed (2) hide show
  1. __pycache__/app.cpython-310.pyc +0 -0
  2. app.py +12 -11
__pycache__/app.cpython-310.pyc CHANGED
Binary files a/__pycache__/app.cpython-310.pyc and b/__pycache__/app.cpython-310.pyc differ
 
app.py CHANGED
@@ -74,18 +74,19 @@ def update_gpu_status():
74
 
75
  except Exception as e:
76
  print(f"Error getting GPU stats: {e}")
77
-
78
- # def update_gpu_status():
79
- # if torch.cuda.is_available():
80
- # gpu_info = torch.cuda.get_device_name(0)
81
- # gpu_memory = torch.cuda.mem_get_info(0)
82
- # total_memory = gpu_memory[1] / (1024 * 1024)
83
- # used_memory = (gpu_memory[1] - gpu_memory[0]) / (1024 * 1024)
 
84
 
85
- # gpu_status = f"GPU: {gpu_info}\nTotal Memory: {total_memory:.2f} MB\nUsed Memory: {used_memory:.2f} MB"
86
- # else:
87
- # gpu_status = "No GPU available"
88
- # return gpu_status
89
 
90
  def update_cpu_status():
91
  import datetime
 
74
 
75
  except Exception as e:
76
  print(f"Error getting GPU stats: {e}")
77
+ return torch_update_gpu_status()
78
+
79
+ def torch_update_gpu_status():
80
+ if torch.cuda.is_available():
81
+ gpu_info = torch.cuda.get_device_name(0)
82
+ gpu_memory = torch.cuda.mem_get_info(0)
83
+ total_memory = gpu_memory[1] / (1024 * 1024)
84
+ used_memory = (gpu_memory[1] - gpu_memory[0]) / (1024 * 1024)
85
 
86
+ gpu_status = f"GPU: {gpu_info}\nTotal Memory: {total_memory:.2f} MB\nUsed Memory: {used_memory:.2f} MB"
87
+ else:
88
+ gpu_status = "No GPU available"
89
+ return gpu_status
90
 
91
  def update_cpu_status():
92
  import datetime