AI & ML interests

None defined yet.

Foundation Model Stack

Foundation Model Stack (fms) is a collection of components developed out of IBM Research used for development, inference, training, and tuning of foundation models leveraging PyTorch native components.

Optimizations

In FMS, we aim to bring the latest optimizations for pre-training/inference/fine-tuning to all of our models. A few of these optimizations include, but are not limited to:

  • fully compilable models with no graph breaks
  • full tensor-parallel support for all applicable modules developed in fms
  • training scripts leveraging FSDP
  • state of the art light-weight speculators for improving inference performance

Usage

FMS is currently being deployed in Text Generation Inference Server

Repositories

  • foundation-model-stack: Main repository for which all fms models are based
  • fms-extras: New features staged to be integrated with foundation-model-stack
  • fms-fsdp: Pre-Training Examples using FSDP wrapped foundation models
  • fms-hf-tuning: Basic Tuning scripts for fms models leveraging SFTTrainer

datasets

None public yet