Update README.md
Browse files
README.md
CHANGED
|
@@ -152,5 +152,13 @@ Once pulled, sanity-check locally with ollama run deepseek-r1:7b, then in anothe
|
|
| 152 |
After the LLM endpoint is up, you can proceed to the inference server step to bind the persona/prompt layer to RTG conditioning and the control loop in one end to end pipeline.
|
| 153 |
|
| 154 |
7) Inference
|
| 155 |
-
|
| 156 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 152 |
After the LLM endpoint is up, you can proceed to the inference server step to bind the persona/prompt layer to RTG conditioning and the control loop in one end to end pipeline.
|
| 153 |
|
| 154 |
7) Inference
|
| 155 |
+
During inference, we deploy Gen-HVAC as a stateless HTTP microservice
|
| 156 |
+
|
| 157 |
+
that loads the trained Decision Transformer checkpoint and normalization statistics at startup, maintains a short autoregressive context window internally,
|
| 158 |
+
and returns multi-zone heating/cooling setpoints per control step.
|
| 159 |
+
In our experiments, EnergyPlus/Sinergym executes inside the Docker container, while the inference service runs on the host/server (CPU/GPU),
|
| 160 |
+
so the simulator can stream observation vectors to POST /predict (payload: {step, obs, info}) and receive an action vector in the response, with POST /reset
|
| 161 |
+
used to clear policy history at episode boundaries.
|
| 162 |
+
|
| 163 |
+
When enabled, the DHIL module queries a local Ollama endpoint and updates the comfort RTG target at a low frequency (e.g., every 4 steps).
|
| 164 |
+
|