from ctransformers import AutoModelForCausalLM class TestModel: def test_generate(self, lib): llm = AutoModelForCausalLM.from_pretrained("marella/gpt-2-ggml", lib=lib) response = llm("AI is going to", seed=5, max_new_tokens=3) assert response == " be a big" token = llm.sample() logits = llm.logits value = logits[token] logits[token] -= 1 assert logits[token] == llm.logits[token] == value - 1 llm.logits[token] *= 2 assert logits[token] == llm.logits[token] == (value - 1) * 2 assert llm.eos_token_id == 50256 assert llm.vocab_size == 50257 assert llm.context_length == 1024