jeffrey-fong commited on
Commit
5df5630
1 Parent(s): 2b6b788
Files changed (1) hide show
  1. modeling_functionary.py +3 -2
modeling_functionary.py CHANGED
@@ -48,7 +48,9 @@ class FunctionaryForCausalLM(LlamaForCausalLM):
48
  negative_prompt_attention_mask: Optional[torch.Tensor] = None,
49
  **kwargs,
50
  ) -> Union[GenerateOutput, torch.LongTensor]:
51
-
 
 
52
  results = self.generate(
53
  inputs=inputs,
54
  generation_config=generation_config,
@@ -63,7 +65,6 @@ class FunctionaryForCausalLM(LlamaForCausalLM):
63
  **kwargs,
64
  )
65
 
66
- tokenizer = kwargs.pop("tokenizer", None) # Pull this out first, we use it to parse raw output
67
  input_ids = kwargs.pop("input_ids")
68
  function_call_token = "<|reserved_special_token_249|>"
69
 
 
48
  negative_prompt_attention_mask: Optional[torch.Tensor] = None,
49
  **kwargs,
50
  ) -> Union[GenerateOutput, torch.LongTensor]:
51
+
52
+ tokenizer = kwargs.pop("tokenizer", None) # Pull this out first, we use it to parse raw output
53
+
54
  results = self.generate(
55
  inputs=inputs,
56
  generation_config=generation_config,
 
65
  **kwargs,
66
  )
67
 
 
68
  input_ids = kwargs.pop("input_ids")
69
  function_call_token = "<|reserved_special_token_249|>"
70