Update README.md
Browse files
README.md
CHANGED
@@ -15,11 +15,11 @@ tags:
|
|
15 |
|
16 |
## Introduction
|
17 |
|
18 |
-
***Combinatorial Optimization Problems(COPs)*** has long been an active field of research. Generally speaking, there exists two main
|
19 |
-
On the other hand, *heuristic algorithms* can compute solutions efficiently, but are not able to
|
20 |
|
21 |
-
In the realistic business scenarios, COPs are usually large-scale(>=1000 nodes), which have very strict requirements for the execution time and performance of solutions. To better solve these problems, we
|
22 |
-
|
23 |
|
24 |
## 🏆Award
|
25 |
|
@@ -34,11 +34,11 @@ tags:
|
|
34 |
|
35 |
* **HIGH-PERFORMANCE**
|
36 |
|
37 |
-
🤠GreedRL have improved the DRL environment(Env) simulation speed by **CUDA and C++ implementations**. At the same time, we have also implemented some **Operators** to replace the native operators of PyTorch, like *Masked Matrix Multiplication* and *Masked Additive Attention*, to achive the ultimate computing performance.
|
38 |
|
39 |
* **USER-FRIENDLY**
|
40 |
|
41 |
-
🤠GreedRL have **warped commonly used modules**, such as Neural Network(NN) components, RL training
|
42 |
|
43 |
## Architecture
|
44 |
![](./images/GREEDRL-Framwork_en.png)
|
@@ -46,8 +46,7 @@ tags:
|
|
46 |
## COPs Modeling examples
|
47 |
|
48 |
|
49 |
-
###
|
50 |
-
Capacitated Vehicle Routing Problem (CVRP)
|
51 |
<details>
|
52 |
<summary>CVRP</summary>
|
53 |
|
@@ -564,14 +563,14 @@ class Objective:
|
|
564 |
|
565 |
We are delighted to release 🤠GreedRL Community Edition, as well as pretrained models, which are specialized to CVRP with problem size ranging from 100 to 5000 nodes.
|
566 |
|
567 |
-
The model is trained using a deep reinforcement learning(DRL) algorithm known as REINFORCE. The model consists of two main components, an Encoder and a Decoder. The encoder produces embedding of all input nodes. The decoder then generates a solution sequence autoregressively. Feasibility of the solution is ensured by a *mask* procedure that prevents the model from selecting nodes that would result in a violation of constraints, e.g. exceeding the vehicle capacity.
|
568 |
|
569 |
|
570 |
## Intended uses & limitations
|
571 |
|
572 |
-
You can use these default models for solving the Capacitated VRP(CVRP) with deep reinforcement learning(DRL).
|
573 |
|
574 |
-
These
|
575 |
|
576 |
## How to use
|
577 |
|
|
|
15 |
|
16 |
## Introduction
|
17 |
|
18 |
+
***Combinatorial Optimization Problems (COPs)*** has long been an active field of research. Generally speaking, there exists two main approaches for solving COPs, each of them having pros and cons. On one hand, the *exact algorithms* can find the optimal solution, but they may be prohibitive for solving large instances because of the exponential increase of the execution time.
|
19 |
+
On the other hand, *heuristic algorithms* can compute solutions efficiently, but are not able to guarantee the optimality of solutions.
|
20 |
|
21 |
+
In the realistic business scenarios, COPs are usually large-scale (>=1000 nodes), which have very strict requirements for the execution time and performance of solutions. To better solve these problems, we
|
22 |
+
propose a generic and complete solver, named **🤠GreedRL**, based on **Deep Reinforcement Learning (DRL)**, which achieves improved speed and performance of solutions than *heuristic algorithms* .
|
23 |
|
24 |
## 🏆Award
|
25 |
|
|
|
34 |
|
35 |
* **HIGH-PERFORMANCE**
|
36 |
|
37 |
+
🤠GreedRL have improved the DRL environment (Env) simulation speed by **CUDA and C++ implementations**. At the same time, we have also implemented some **Operators** to replace the native operators of PyTorch, like *Masked Matrix Multiplication* and *Masked Additive Attention*, to achive the ultimate computing performance.
|
38 |
|
39 |
* **USER-FRIENDLY**
|
40 |
|
41 |
+
🤠GreedRL have **warped commonly used modules**, such as Neural Network (NN) components, RL training algorithms and COPs constraints implementations, which makes it easy to use.
|
42 |
|
43 |
## Architecture
|
44 |
![](./images/GREEDRL-Framwork_en.png)
|
|
|
46 |
## COPs Modeling examples
|
47 |
|
48 |
|
49 |
+
### Capacitated Vehicle Routing Problem (CVRP)
|
|
|
50 |
<details>
|
51 |
<summary>CVRP</summary>
|
52 |
|
|
|
563 |
|
564 |
We are delighted to release 🤠GreedRL Community Edition, as well as pretrained models, which are specialized to CVRP with problem size ranging from 100 to 5000 nodes.
|
565 |
|
566 |
+
The model is trained using a deep reinforcement learning (DRL) algorithm known as REINFORCE. The model consists of two main components, an Encoder and a Decoder. The encoder produces embedding of all input nodes. The decoder then generates a solution sequence autoregressively. Feasibility of the solution is ensured by a *mask* procedure that prevents the model from selecting nodes that would result in a violation of constraints, e.g. exceeding the vehicle capacity.
|
567 |
|
568 |
|
569 |
## Intended uses & limitations
|
570 |
|
571 |
+
You can use these default models for solving the Capacitated VRP (CVRP) with deep reinforcement learning(DRL).
|
572 |
|
573 |
+
These models are limited by the training dataset, which may not generalize well for all use cases in different domains.
|
574 |
|
575 |
## How to use
|
576 |
|