File size: 2,785 Bytes
be5ae25
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
---
license: apache-2.0
language:
- en
pipeline_tag: text2text-generation
tags:
- bpmn
- Business Process
---
# T5-Small Finetuned for Purchase Order Workflow Business Processes

## Model Description
This is a fine-tuned version of the T5-Small model, designed specifically for the extraction of BPMN (Business Process Model and Notation) diagrams from textual descriptions related to Purchase Order Workflow Business Processes. This AI-driven approach leverages advanced language modeling techniques to transform natural language descriptions into BPMN models, facilitating the modernization and automation of business processes in this specific area. **This model serves as a proof of concept and is not yet ready for real-life applications.**

## Key Features
- **Language Model Base**: T5-Small, known for its efficiency and efficacy in understanding and generating text.
- **Specialization**: Fine-tuned specifically for BPMN generation in Purchase Order Workflows, improving accuracy and relevancy in business process modeling.
- **Dataset**: Trained on the "MaD: A Dataset for Interview-based BPM in Business Process Management" dataset, cited from the research article available at [IEEE Xplore](https://ieeexplore.ieee.org/document/10191898).

## Applications
- **Business Process Management**: Automates the generation of BPMN diagrams, which are crucial for documenting and improving Purchase Order Workflow Business Processes.
- **AI Research and Development**: Provides a research basis for further exploration into the integration of NLP and business process management.
- **Educational Tool**: Assists in teaching the concepts of BPMN and AI's role in business process automation, particularly in the context of Purchase Order Workflows.

## Configuration
- **Pre-trained Model**: Google's T5-Small
- **Training Environment**: Utilized a dataset from "MaD: A Dataset for Interview-based BPM in Business Process Management" for training and validation.
- **Hardware Used**: 
  - **CPU**: Apple M1 MAX with 10 cores (8 performance + 2 efficiency), 3.20 GHz
  - **GPU**: Integrated, 32 cores
  - **RAM**: 64 GB LPDDR5
  - **Storage**: 2 TB SSD
  - **OS**: macOS 12.7.1 (Monterey)
- **Training Script**: [Finetuning T5-Small BPMN](https://github.com/ofachati/Finetuning-T5small-BPMN)

## Installation and Requirements
The model can be accessed and installed via the Hugging Face model hub. Requirements for using this model include Python 3.6 or newer and access to a machine with adequate computational capabilities to run inference with the T5 architecture.

## Contributors
- **Pascal Poizat**: [Hugging Face Profile](https://huggingface.co/pascalpoizat) - Training hardware
- **Omar El Fachati**: [Hugging Face Profile](https://huggingface.co/fachati) - Training script