ContaLLM

ContaLLM-Food-Beverage-8B-Instruct

ContaLLM-Food-Beverage-8B-Instruct is a large Chinese vertical marketing model for the food and beverage industry. You can customize and generate marketing texts according to users' specific marketing needs, product selection, product selection knowledge base, keywords, main recommended selling points, main recommended scenes, hashtags, article types, etc. Use the LLM's capabilities and training on existing high-quality marketing materials to help companies generate diversified, high-quality marketing content and improve marketing conversion rates.

Model description

  • Model type: A model trained on a mix of publicly available, synthetic and human-annotated datasets.
  • Language(s) (NLP): Primarily Chinese
  • Industry: Food And Beverage Industry Marketing
  • License: Llama 3.1 Community License Agreement
  • Finetuned from model: meta-llama/Llama-3.1-8B-Instruct

Model Stage

Industry Version Llama 3.1 8B
Food And Beverage bf16 ContaAI/ContaLLM-Food-Beverage-8B-Instruct
Food And Beverage 8bit ContaAI/ContaLLM-Food-Beverage-8B-Instruct-8bit
Food And Beverage 4bit ContaAI/ContaLLM-Food-Beverage-8B-Instruct-4bit

Using the model

Loading with HuggingFace

To load the model with HuggingFace, use the following snippet:

from transformers import AutoModelForCausalLM

model = AutoModelForCausalLM.from_pretrained("ContaAI/ContaLLM-Food-Beverage-8B-Instruct")

System Prompt

This model is a Chinese marketing model for food and beverage industry, so we use this system prompt by default:

system_prompt = '่ฏทๆ นๆฎ็”จๆˆทๆไพ›็š„่ฅ้”€้œ€ๆฑ‚ใ€้€‰ๅ“ๅŠๅ…ถไป–ไฟกๆฏๅ†™ไธ€็ฏ‡้ฃŸๅ“้ฅฎๆ–™่กŒไธš็š„่ฅ้”€ๆŽจๆ–‡ใ€‚'

User Prompt

Users can enter the required marketing needs according to their own needs, non-required including keywords, topics, label marketing nodes, people, related materials, content length, which content length has three specifications, respectively, shorter, medium, longer. The details are as follows:

Parameter name Required Meaning and optional range
่ฅ้”€้œ€ๆฑ‚ required Fill in your marketing requirements, cannot be blank
้€‰ๅ“ required Fill in your product selection, cannot be blank
้€‰ๅ“็Ÿฅ่ฏ†ๅบ“ required Fill in the relevant information/materials about your product, cannot be blank
ๅ…ณ้”ฎ่ฏ optional Fill in your marketing keywords, or remove this row from the prompt
ๆ ‡็ญพ optional Fill in the hashtag, or remove this row from the prompt
ไธปๆŽจๅ–็‚น optional Fill in the main recommended selling points, or remove this row from the prompt
ไธปๆŽจๅœบๆ™ฏ optional Fill in the main recommended scenes, or remove this row from the prompt
ๆ–‡็ซ ็ฑปๅž‹ optional Fill in the article type, or remove this row from the prompt

Example:

user_prompt = """่ฅ้”€้œ€ๆฑ‚๏ผšๅคๆ—ฅๆธ…ๅ‡‰๏ผŒๆ—ฅๆ–™้ฃŽๅ‘ณไฝ“้ชŒ
้€‰ๅ“๏ผšๆธ…ๆ–ฐๆŸ ๆชฌๅฏฟๅธๅท
้€‰ๅ“็Ÿฅ่ฏ†ๅบ“๏ผš1ใ€้€‰็”จๆ–ฐ้ฒœ็š„ไธ‰ๆ–‡้ฑผๅ’Œ็‰›ๆฒนๆžœ๏ผŒๆญ้…ๆธ…็ˆฝๆŸ ๆชฌๆฑ๏ผŒๅฃๆ„Ÿๅฑ‚ๆฌกไธฐๅฏŒใ€‚2ใ€ไฝŽ่„‚ๅฅๅบท๏ผŒ้€‚ๅˆๅฅ่บซไบบๅฃซใ€‚3ใ€ๆฏไปฝไป…ๅซ200ๅคงๅก๏ผŒ่ฝปๆพไบซๅ—็พŽๅ‘ณใ€‚
ๅ…ณ้”ฎ่ฏ๏ผšๆ—ฅๆ–™ใ€ๅฏฟๅธใ€ๅฅๅบท้ฅฎ้ฃŸใ€ๅคๆ—ฅ็พŽ้ฃŸ
ไธปๆŽจๅ–็‚น๏ผšๆธ…ๆ–ฐๅฅๅบท
ไธปๆŽจๅœบๆ™ฏ๏ผšๅคๆ—ฅ่šไผš
ๆ ‡็ญพ๏ผš#ๆ—ฅๆ–™# #ๅฏฟๅธ# #ๅฅๅบท็พŽ้ฃŸ
ๆ–‡็ซ ็ฑปๅž‹๏ผš็พŽ้ฃŸๆŽจ่"""

Use example (with template)

import torch
from transformers import AutoModelForCausalLM, AutoTokenizer
model_name = "ContaAI/ContaLLM-Food-Beverage-8B-Instruct"
model = AutoModelForCausalLM.from_pretrained(model_name, device_map="auto")
tokenizer = AutoTokenizer.from_pretrained(model_name)

system_prompt = '่ฏทๆ นๆฎ็”จๆˆทๆไพ›็š„่ฅ้”€้œ€ๆฑ‚ใ€้€‰ๅ“ๅŠๅ…ถไป–ไฟกๆฏๅ†™ไธ€็ฏ‡้ฃŸๅ“้ฅฎๆ–™่กŒไธš็š„่ฅ้”€ๆŽจๆ–‡ใ€‚'

user_prompt = """่ฅ้”€้œ€ๆฑ‚๏ผšๅคๆ—ฅๆธ…ๅ‡‰๏ผŒๆ—ฅๆ–™้ฃŽๅ‘ณไฝ“้ชŒ
้€‰ๅ“๏ผšๆธ…ๆ–ฐๆŸ ๆชฌๅฏฟๅธๅท
้€‰ๅ“็Ÿฅ่ฏ†ๅบ“๏ผš1ใ€้€‰็”จๆ–ฐ้ฒœ็š„ไธ‰ๆ–‡้ฑผๅ’Œ็‰›ๆฒนๆžœ๏ผŒๆญ้…ๆธ…็ˆฝๆŸ ๆชฌๆฑ๏ผŒๅฃๆ„Ÿๅฑ‚ๆฌกไธฐๅฏŒใ€‚2ใ€ไฝŽ่„‚ๅฅๅบท๏ผŒ้€‚ๅˆๅฅ่บซไบบๅฃซใ€‚3ใ€ๆฏไปฝไป…ๅซ200ๅคงๅก๏ผŒ่ฝปๆพไบซๅ—็พŽๅ‘ณใ€‚
ๅ…ณ้”ฎ่ฏ๏ผšๆ—ฅๆ–™ใ€ๅฏฟๅธใ€ๅฅๅบท้ฅฎ้ฃŸใ€ๅคๆ—ฅ็พŽ้ฃŸ
ไธปๆŽจๅ–็‚น๏ผšๆธ…ๆ–ฐๅฅๅบท
ไธปๆŽจๅœบๆ™ฏ๏ผšๅคๆ—ฅ่šไผš
ๆ ‡็ญพ๏ผš#ๆ—ฅๆ–™# #ๅฏฟๅธ# #ๅฅๅบท็พŽ้ฃŸ
ๆ–‡็ซ ็ฑปๅž‹๏ผš็พŽ้ฃŸๆŽจ่"""

prompt_template = '''<|begin_of_text|><|start_header_id|>system<|end_header_id|>
{}<|eot_id|><|start_header_id|>user<|end_header_id|>
{}<|eot_id|><|start_header_id|>assistant<|end_header_id|>'''

prompt = prompt_template.format(system_prompt, user_prompt)

tokenized_message = tokenizer(
  prompt,
  max_length=2048,
  return_tensors="pt",
  add_special_tokens=False
)

response_token_ids= model.generate(
  **tokenized_message,
  max_new_tokens=1024,
  do_sample=True,
  top_p=1.0,
  temperature=0.5,
  min_length=None,
  use_cache=True,
  top_k=50,
  repetition_penalty=1.2,
  length_penalty=1,
)

generated_tokens = response_token_ids[0, tokenized_message['input_ids'].shape[-1]:]
generated_text = tokenizer.decode(generated_tokens, skip_special_tokens=True)
print(generated_text)

Bias, Risks, and Limitations

The ContaLLM models implemented safety techniques during data generation and training, but they are not deployed automatically with in-the-loop filtering of responses like ChatGPT during inference, so the model can produce problematic outputs (especially when prompted to do so). It is also unknown what the size and composition of the corpus was used to train the base Llama 3.1 models, however it is likely to have included a mix of Web data and technical sources like books and code. The use of the models is at your own risk. You may need to monitor the outputs of the model and take appropriate actions such as content filtering if necessary.

License and use

All Llama 3.1 ContaAI models are released under Meta's Llama 3.1 Community License Agreement.

Downloads last month
7
Safetensors
Model size
8.03B params
Tensor type
BF16
ยท
Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support

Model tree for ContaAI/ContaLLM-Food-Beverage-8B-Instruct

Finetuned
(1179)
this model
Quantizations
1 model