YAML Metadata Warning: empty or missing yaml metadata in repo card (https://huggingface.co/docs/hub/model-cards#model-card-metadata)

Quantization made by Richard Erkhov.

Github

Discord

Request more models

StructLM-7B - GGUF

Name Quant method Size
StructLM-7B.Q2_K.gguf Q2_K 2.36GB
StructLM-7B.IQ3_XS.gguf IQ3_XS 2.6GB
StructLM-7B.IQ3_S.gguf IQ3_S 2.75GB
StructLM-7B.Q3_K_S.gguf Q3_K_S 2.75GB
StructLM-7B.IQ3_M.gguf IQ3_M 2.9GB
StructLM-7B.Q3_K.gguf Q3_K 3.07GB
StructLM-7B.Q3_K_M.gguf Q3_K_M 3.07GB
StructLM-7B.Q3_K_L.gguf Q3_K_L 3.35GB
StructLM-7B.IQ4_XS.gguf IQ4_XS 3.4GB
StructLM-7B.Q4_0.gguf Q4_0 3.56GB
StructLM-7B.IQ4_NL.gguf IQ4_NL 3.58GB
StructLM-7B.Q4_K_S.gguf Q4_K_S 3.59GB
StructLM-7B.Q4_K.gguf Q4_K 3.8GB
StructLM-7B.Q4_K_M.gguf Q4_K_M 3.8GB
StructLM-7B.Q4_1.gguf Q4_1 3.95GB
StructLM-7B.Q5_0.gguf Q5_0 4.33GB
StructLM-7B.Q5_K_S.gguf Q5_K_S 4.33GB
StructLM-7B.Q5_K.gguf Q5_K 4.45GB
StructLM-7B.Q5_K_M.gguf Q5_K_M 4.45GB
StructLM-7B.Q5_1.gguf Q5_1 4.72GB
StructLM-7B.Q6_K.gguf Q6_K 5.15GB
StructLM-7B.Q8_0.gguf Q8_0 6.67GB

Original model description:

license: mit datasets: - TIGER-Lab/SKGInstruct language: - en

πŸ—οΈ StructLM: Towards Building Generalist Models for Structured Knowledge Grounding

This checkpoing seems to have some issue, please use https://huggingface.co/TIGER-Lab/StructLM-7B-Mistral instead.

Project Page: https://tiger-ai-lab.github.io/StructLM/

Paper: https://arxiv.org/pdf/2402.16671.pdf

Code: https://github.com/TIGER-AI-Lab/StructLM

Alt text

Introduction

StructLM, is a series of open-source large language models (LLMs) finetuned for structured knowledge grounding (SKG) tasks. We release 3 models:

7B | StructLM-7B

13B | StructLM-13B

34B | StructLM-34B

Training Data

These models are trained on πŸ€— SKGInstruct Dataset, an instruction-tuning dataset containing mixture of 19 SKG tasks combined with πŸ€— SlimOrca. Check out the dataset card for more details.

Training Procedure

The models are fine-tuned with CodeLlama-Instruct-hf models as base models. Each model is trained for 3 epochs, and the best checkpoint is selected.

Evaluation

Here are a subset of model evaluation results:

Held in

Model ToTTo GrailQA CompWebQ MMQA Feverous Spider TabFact Dart
StructLM-7B 49.4 80.4 78.3 85.2 84.4 72.4 80.8 62.2
StructLM-13B 49.3 79.2 80.4 86.0 85.0 74.1 84.7 61.4
StructLM-34B 50.2 82.2 81.9 88.1 85.7 74.6 86.6 61.8

Held out

Model BIRD InfoTabs FinQA SQA
StructLM-7B 22.3 55.3 27.3 49.7
StructLM-13B 22.8 58.1 25.6 36.1
StructLM-34B 24.7 61.8 36.2 44.2

Usage

You can use the models through Huggingface's Transformers library. Check our Github repo for the evaluation code: https://github.com/TIGER-AI-Lab/StructLM

Prompt Format

For this 7B model, the prompt format (different from 13B, 34B) is

[INST] <<SYS>>
You are an AI assistant that specializes in analyzing and reasoning over structured information. You will be given a task, optionally with some structured knowledge input. Your answer must strictly adhere to the output format, if specified.
<</SYS>>
{instruction} [/INST]

To see concrete examples of this linearization, you can directly reference the πŸ€— SKGInstruct Dataset (coming soon). We will provide code for linearizing this data shortly.

A few examples:

Tabular data

col : day | kilometers row 1 : tuesday | 0 row 2 : wednesday | 0 row 3 : thursday | 4 row 4 : friday | 0 row 5 : saturday | 0

Knowledge triples (dart)

Hawaii Five-O : notes : Episode: The Flight of the Jewels | [TABLECONTEXT] : [title] : Jeff Daniels | [TABLECONTEXT] : title : Hawaii Five-O

Knowledge graph schema (grailqa)

top antiquark: m.094nrqp | physics.particle_antiparticle.self_antiparticle physics.particle_family physics.particle.antiparticle physics.particle_family.subclasses physics.subatomic_particle_generation physics.particle_family.particles physics.particle common.image.appears_in_topic_gallery physics.subatomic_particle_generation.particles physics.particle.family physics.particle_family.parent_class physics.particle_antiparticle physics.particle_antiparticle.particle physics.particle.generation

Example input

[INST] <<SYS>>
You are an AI assistant that specializes in analyzing and reasoning over structured information. You will be given a task, optionally with some structured knowledge input. Your answer must strictly adhere to the output format, if specified.
<</SYS>>

Use the information in the following table to solve the problem, choose between the choices if they are provided. table:

col : day | kilometers row 1 : tuesday | 0 row 2 : wednesday | 0 row 3 : thursday | 4 row 4 : friday | 0 row 5 : saturday | 0


question:

Allie kept track of how many kilometers she walked during the past 5 days. What is the range of the numbers? [/INST]

Intended Uses

These models are trained for research purposes. They are designed to be proficient in interpreting linearized structured input. Downstream uses can potentially include various applications requiring the interpretation of structured data.

Limitations

While we've tried to build an SKG-specialized model capable of generalizing, we have shown that this is a challenging domain, and it may lack performance characteristics that allow it to be directly used in chat or other applications.

Citation

If you use the models, data, or code from this project, please cite the original paper:

@misc{zhuang2024structlm,
      title={StructLM: Towards Building Generalist Models for Structured Knowledge Grounding}, 
      author={Alex Zhuang and Ge Zhang and Tianyu Zheng and Xinrun Du and Junjie Wang and Weiming Ren and Stephen W. Huang and Jie Fu and Xiang Yue and Wenhu Chen},
      year={2024},
      eprint={2402.16671},
      archivePrefix={arXiv},
      primaryClass={cs.CL}
}
Downloads last month
37
GGUF
Model size
6.74B params
Architecture
llama

2-bit

3-bit

4-bit

5-bit

6-bit

8-bit

Inference API
Unable to determine this model's library. Check the docs .