zetavg commited on
Commit
85fb243
1 Parent(s): f090da6
llama_lora/globals.py CHANGED
@@ -163,6 +163,7 @@ def load_gpu_info():
163
  preserve_loaded_models_count = math.floor(
164
  (available_cpu_ram * 0.8) / total_memory) - 1
165
  if preserve_loaded_models_count > 1:
 
166
  print(
167
  f"Will keep {preserve_loaded_models_count} offloaded models in CPU RAM.")
168
  Global.loaded_models = ModelLRUCache(preserve_loaded_models_count)
 
163
  preserve_loaded_models_count = math.floor(
164
  (available_cpu_ram * 0.8) / total_memory) - 1
165
  if preserve_loaded_models_count > 1:
166
+ ModelLRUCache = dynamic_import('.utils.model_lru_cache').ModelLRUCache
167
  print(
168
  f"Will keep {preserve_loaded_models_count} offloaded models in CPU RAM.")
169
  Global.loaded_models = ModelLRUCache(preserve_loaded_models_count)
llama_lora/utils/eta_predictor.py CHANGED
@@ -31,7 +31,7 @@ class ETAPredictor:
31
  self._cleanup_old_data()
32
 
33
  # Only predict if we have enough data
34
- if len(self.data) < 2 or self.data[-1][1] - self.data[0][1] < 5:
35
  return None
36
 
37
  first_step, first_time = self.data[0]
 
31
  self._cleanup_old_data()
32
 
33
  # Only predict if we have enough data
34
+ if len(self.data) < 2 or self.data[-1][1] - self.data[0][1] < 1:
35
  return None
36
 
37
  first_step, first_time = self.data[0]