--- license: apache-2.0 language: - en pipeline_tag: text2text-generation tags: - bpmn - Business Process --- # T5-Small Finetuned for Purchase Order Workflow Business Processes ## Model Description This is a fine-tuned version of the T5-Small model, designed specifically for the extraction of BPMN (Business Process Model and Notation) diagrams from textual descriptions related to Purchase Order Workflow Business Processes. This AI-driven approach leverages advanced language modeling techniques to transform natural language descriptions into BPMN models, facilitating the modernization and automation of business processes in this specific area. **This model serves as a proof of concept and is not yet ready for real-life applications.** ## Key Features - **Language Model Base**: T5-Small, known for its efficiency and efficacy in understanding and generating text. - **Specialization**: Fine-tuned specifically for BPMN generation in Purchase Order Workflows, improving accuracy and relevancy in business process modeling. - **Dataset**: Trained on the "MaD: A Dataset for Interview-based BPM in Business Process Management" dataset, cited from the research article available at [IEEE Xplore](https://ieeexplore.ieee.org/document/10191898). ## Applications - **Business Process Management**: Automates the generation of BPMN diagrams, which are crucial for documenting and improving Purchase Order Workflow Business Processes. - **AI Research and Development**: Provides a research basis for further exploration into the integration of NLP and business process management. - **Educational Tool**: Assists in teaching the concepts of BPMN and AI's role in business process automation, particularly in the context of Purchase Order Workflows. ## Configuration - **Pre-trained Model**: Google's T5-Small - **Training Environment**: Utilized a dataset from "MaD: A Dataset for Interview-based BPM in Business Process Management" for training and validation. - **Hardware Used**: - **CPU**: Apple M1 MAX with 10 cores (8 performance + 2 efficiency), 3.20 GHz - **GPU**: Integrated, 32 cores - **RAM**: 64 GB LPDDR5 - **Storage**: 2 TB SSD - **OS**: macOS 12.7.1 (Monterey) - **Training Script**: [Finetuning T5-Small BPMN](https://github.com/ofachati/Finetuning-T5small-BPMN) ## Installation and Requirements The model can be accessed and installed via the Hugging Face model hub. Requirements for using this model include Python 3.6 or newer and access to a machine with adequate computational capabilities to run inference with the T5 architecture. ## Contributors - **Pascal Poizat**: [Hugging Face Profile](https://huggingface.co/pascalpoizat) - Training hardware - **Omar El Fachati**: [Hugging Face Profile](https://huggingface.co/fachati) - Training script