Update README.md
Browse files
README.md
CHANGED
|
@@ -142,17 +142,16 @@ actions lead to what kind of consequenses.
|
|
| 142 |
|
| 143 |
5) LLM deployment phase
|
| 144 |
|
| 145 |
-
|
| 146 |
-
|
| 147 |
-
|
| 148 |
-
|
| 149 |
-
|
| 150 |
-
For this experiments we used several opensource Quantized versions of both resonsing and non reasoning LLMS liek DEEPseek V1 and r1 lamma, etc.
|
| 151 |
-
you have to create the llm client and server setup which is provided in the llm folder.
|
| 152 |
|
| 153 |
-
|
|
|
|
| 154 |
|
| 155 |
-
|
|
|
|
| 156 |
|
| 157 |
7) Inference
|
| 158 |
|
|
|
|
| 142 |
|
| 143 |
5) LLM deployment phase
|
| 144 |
|
| 145 |
+
After this you have to move on to the inference Server side to tie up all these together
|
| 146 |
+
Gen-HVAC supports an optional LLM + Digital Human-in-the-Loop (DHIL) layer that modulates preference/RTG targets and high-level
|
| 147 |
+
constraints while the controller produces smooth setpoint sequences. For local LLM hosting, install Ollama, pull a quantized model
|
| 148 |
+
, and launch the service.
|
|
|
|
|
|
|
|
|
|
| 149 |
|
| 150 |
+
On Linux/macOS you can install Ollama via curl -fsSL https://ollama.com/install.sh | sh, start the daemon with ollama serve (leave it running), and pull recommended models using ollama pull deepseek-r1:7b (lightweight reasoning), ollama pull llama3.1:8b (strong general instruction-following), ollama pull qwen2.5:7b (efficient general model), or ollama pull mistral:instruct (fast instruct model). If you want a slightly heavier but still practical model, ollama pull deepseek-r1:14b or ollama pull qwen2.5:14b.
|
| 151 |
+
In our testing we choose Deepseek R1.
|
| 152 |
|
| 153 |
+
Once pulled, sanity-check locally with ollama run deepseek-r1:7b, then in another terminal point your Gen-HVAC LLM client to the default endpoint and run your integration from the llm/ folder (e.g., python -m llm.server --host 0.0.0.0 --port 8000 and python -m llm.client --base_url http://localhost:xxxx --model deepseek-r1:7b.
|
| 154 |
+
After the LLM endpoint is up, you can proceed to the inference server step to bind the persona/prompt layer to RTG conditioning and the control loop in one end to end pipeline.
|
| 155 |
|
| 156 |
7) Inference
|
| 157 |
|