Gen-HVAC commited on
Commit
50abebb
·
verified ·
1 Parent(s): 52440df

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +8 -9
README.md CHANGED
@@ -142,17 +142,16 @@ actions lead to what kind of consequenses.
142
 
143
  5) LLM deployment phase
144
 
145
- One of the main high lights of our System is LLM and digital human in the loop which is possible because of flexible RTG conditioning posible because of transformer architecture.
146
- Different RTG can lead to different sequences which is completely different approach where we give control to LLM and because of that we never face sudden uneven temperature settings.
147
-
148
- For LLM utilisation please download and setup OLLAMA
149
-
150
- For this experiments we used several opensource Quantized versions of both resonsing and non reasoning LLMS liek DEEPseek V1 and r1 lamma, etc.
151
- you have to create the llm client and server setup which is provided in the llm folder.
152
 
153
- Our system use LLM to capture the humnistic value of the human behaviour so you can change or tweak the prompt as per your behaviour or create your own personal digital persona.
 
154
 
155
- After this you have to move on to the inference Server side to tie up all these together
 
156
 
157
  7) Inference
158
 
 
142
 
143
  5) LLM deployment phase
144
 
145
+ After this you have to move on to the inference Server side to tie up all these together
146
+ Gen-HVAC supports an optional LLM + Digital Human-in-the-Loop (DHIL) layer that modulates preference/RTG targets and high-level
147
+ constraints while the controller produces smooth setpoint sequences. For local LLM hosting, install Ollama, pull a quantized model
148
+ , and launch the service.
 
 
 
149
 
150
+ On Linux/macOS you can install Ollama via curl -fsSL https://ollama.com/install.sh | sh, start the daemon with ollama serve (leave it running), and pull recommended models using ollama pull deepseek-r1:7b (lightweight reasoning), ollama pull llama3.1:8b (strong general instruction-following), ollama pull qwen2.5:7b (efficient general model), or ollama pull mistral:instruct (fast instruct model). If you want a slightly heavier but still practical model, ollama pull deepseek-r1:14b or ollama pull qwen2.5:14b.
151
+ In our testing we choose Deepseek R1.
152
 
153
+ Once pulled, sanity-check locally with ollama run deepseek-r1:7b, then in another terminal point your Gen-HVAC LLM client to the default endpoint and run your integration from the llm/ folder (e.g., python -m llm.server --host 0.0.0.0 --port 8000 and python -m llm.client --base_url http://localhost:xxxx --model deepseek-r1:7b.
154
+ After the LLM endpoint is up, you can proceed to the inference server step to bind the persona/prompt layer to RTG conditioning and the control loop in one end to end pipeline.
155
 
156
  7) Inference
157