--- library_name: transformers.js --- https://huggingface.co/Qwen/Qwen1.5-0.5B-Chat with ONNX weights to be compatible with Transformers.js. ## Usage (Transformers.js) If you haven't already, you can install the [Transformers.js](https://huggingface.co/docs/transformers.js) JavaScript library from [NPM](https://www.npmjs.com/package/@xenova/transformers) using: ```bash npm i @xenova/transformers ``` **Example:** Text generation with `Xenova/Qwen1.5-0.5B-Chat`. ```js import { pipeline } from '@xenova/transformers'; // Create text-generation pipeline const generator = await pipeline('text-generation', 'Xenova/Qwen1.5-0.5B-Chat'); // Define the prompt and list of messages const prompt = 'Give me a short introduction to large language model.' const messages = [ { role: 'system', content: 'You are a helpful assistant.' }, { role: 'user', content: prompt } ] // Apply chat template const text = generator.tokenizer.apply_chat_template(messages, { tokenize: false, add_generation_prompt: true, }); // Generate text const output = await generator(text, { max_new_tokens: 128, do_sample: false, return_full_text: false, }); console.log(output[0].generated_text); // 'A large language model is a type of artificial intelligence system that can generate text based on the input provided by users, such as books, articles, or websites. It uses advanced algorithms and techniques to learn from vast amounts of data and improve its performance over time through machine learning and natural language processing (NLP). Large language models have become increasingly popular in recent years due to their ability to handle complex tasks such as generating human-like text quickly and accurately. They have also been used in various fields such as customer service chatbots, virtual assistants, and search engines for information retrieval purposes.' ``` --- Note: Having a separate repo for ONNX weights is intended to be a temporary solution until WebML gains more traction. If you would like to make your models web-ready, we recommend converting to ONNX using [🤗 Optimum](https://huggingface.co/docs/optimum/index) and structuring your repo like this one (with ONNX weights located in a subfolder named `onnx`).