pascalpoizat's picture
correct: ordering of contributors is alpha
64da9ab verified
---
license: apache-2.0
language:
- en
pipeline_tag: text2text-generation
tags:
- bpmn
- Business Process
---
# T5-Small Finetuned for Purchase Order Workflow Business Processes
## Model Description
This is a fine-tuned version of the T5-Small model, designed specifically for the extraction of BPMN (Business Process Model and Notation) diagrams from textual descriptions related to Purchase Order Workflow Business Processes. This AI-driven approach leverages advanced language modeling techniques to transform natural language descriptions into BPMN models, facilitating the modernization and automation of business processes in this specific area. **This model serves as a proof of concept and is not yet ready for real-life applications.**
## Key Features
- **Language Model Base**: T5-Small, known for its efficiency and efficacy in understanding and generating text.
- **Specialization**: Fine-tuned specifically for BPMN generation in Purchase Order Workflows, improving accuracy and relevancy in business process modeling.
- **Dataset**: Trained on the "MaD: A Dataset for Interview-based BPM in Business Process Management" dataset, cited from the research article available at [IEEE Xplore](https://ieeexplore.ieee.org/document/10191898).
## Applications
- **Business Process Management**: Automates the generation of BPMN diagrams, which are crucial for documenting and improving Purchase Order Workflow Business Processes.
- **AI Research and Development**: Provides a research basis for further exploration into the integration of NLP and business process management.
- **Educational Tool**: Assists in teaching the concepts of BPMN and AI's role in business process automation, particularly in the context of Purchase Order Workflows.
## Configuration
- **Pre-trained Model**: Google's T5-Small
- **Training Environment**: Utilized a dataset from "MaD: A Dataset for Interview-based BPM in Business Process Management" for training and validation.
- **Hardware Used**:
- **CPU**: Apple M1 MAX with 10 cores (8 performance + 2 efficiency), 3.20 GHz
- **GPU**: Integrated, 32 cores
- **RAM**: 64 GB LPDDR5
- **Storage**: 2 TB SSD
- **OS**: macOS 12.7.1 (Monterey)
- **Training Script**: [Finetuning T5-Small BPMN](https://github.com/ofachati/Finetuning-T5small-BPMN)
## Installation and Requirements
The model can be accessed and installed via the Hugging Face model hub. Requirements for using this model include Python 3.6 or newer and access to a machine with adequate computational capabilities to run inference with the T5 architecture.
## Contributors
- **Omar El Fachati**: [Hugging Face Profile](https://huggingface.co/fachati) - Training script
- **Pascal Poizat**: [Hugging Face Profile](https://huggingface.co/pascalpoizat) - Training hardware