Update app.py
Browse files
app.py
CHANGED
|
@@ -5,16 +5,16 @@ from transformers import AutoModelForCausalLM, AutoTokenizer
|
|
| 5 |
|
| 6 |
os.environ["HF_HOME"] = "/data/hf_cache"
|
| 7 |
|
| 8 |
-
tokenizer = AutoTokenizer.from_pretrained("
|
| 9 |
-
model = AutoModelForCausalLM.from_pretrained("
|
| 10 |
|
| 11 |
def get_stats():
|
| 12 |
process = psutil.Process(os.getpid())
|
| 13 |
ram = process.memory_info().rss / 1024 ** 3
|
| 14 |
disk_tmp = psutil.disk_usage('/tmp').used / 1024 ** 3
|
| 15 |
-
|
| 16 |
cpu = psutil.cpu_percent(interval=1)
|
| 17 |
-
return f"RAM: {ram:.2f} GB | /tmp: {disk_tmp:.2f} GB |
|
| 18 |
|
| 19 |
def chat(message, history):
|
| 20 |
messages = []
|
|
|
|
| 5 |
|
| 6 |
os.environ["HF_HOME"] = "/data/hf_cache"
|
| 7 |
|
| 8 |
+
tokenizer = AutoTokenizer.from_pretrained("/data/model")
|
| 9 |
+
model = AutoModelForCausalLM.from_pretrained("/data/model", device_map="auto", offload_folder="/tmp/offload")
|
| 10 |
|
| 11 |
def get_stats():
|
| 12 |
process = psutil.Process(os.getpid())
|
| 13 |
ram = process.memory_info().rss / 1024 ** 3
|
| 14 |
disk_tmp = psutil.disk_usage('/tmp').used / 1024 ** 3
|
| 15 |
+
disk_data = psutil.disk_usage('/data').used / 1024 ** 3
|
| 16 |
cpu = psutil.cpu_percent(interval=1)
|
| 17 |
+
return f"RAM: {ram:.2f} GB | /tmp: {disk_tmp:.2f} GB | /data: {disk_data:.2f} GB | CPU: {cpu:.1f}%"
|
| 18 |
|
| 19 |
def chat(message, history):
|
| 20 |
messages = []
|