PeterPinetree commited on
Commit
95e113e
·
1 Parent(s): 8aa8163

Update index.html

Browse files
Files changed (1) hide show
  1. index.html +27 -10
index.html CHANGED
@@ -354,7 +354,8 @@
354
  await model.generate(enc.input_ids, {
355
  attention_mask: enc.attention_mask,
356
  max_new_tokens: 1,
357
- do_sample: false
 
358
  });
359
  } catch (e) {
360
  console.warn("Warm-up skipped:", e);
@@ -392,15 +393,31 @@
392
  const padId = tokenizer?.pad_token_id ?? eosId;
393
 
394
  const t0 = performance.now();
395
- const gen = await model.generate(enc.input_ids, {
396
- attention_mask: enc.attention_mask,
397
- max_new_tokens: 1,
398
- do_sample: false,
399
- output_scores: true,
400
- return_dict_in_generate: true,
401
- eos_token_id: eosId,
402
- pad_token_id: padId
403
- });
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
404
  const dt = (performance.now() - t0) | 0;
405
 
406
  // --- from here, keep your existing logits -> softmax -> Top-K code ---
 
354
  await model.generate(enc.input_ids, {
355
  attention_mask: enc.attention_mask,
356
  max_new_tokens: 1,
357
+ do_sample: false,
358
+ use_cache: false // avoid cached branch on first pass
359
  });
360
  } catch (e) {
361
  console.warn("Warm-up skipped:", e);
 
393
  const padId = tokenizer?.pad_token_id ?? eosId;
394
 
395
  const t0 = performance.now();
396
+ let gen;
397
+ try {
398
+ // try normal path first
399
+ gen = await model.generate(enc.input_ids, {
400
+ attention_mask: enc.attention_mask,
401
+ max_new_tokens: 1,
402
+ do_sample: false,
403
+ output_scores: true,
404
+ return_dict_in_generate: true,
405
+ eos_token_id: eosId,
406
+ pad_token_id: padId
407
+ });
408
+ } catch (e) {
409
+ console.warn("generate() failed; retrying with use_cache:false", e);
410
+ gen = await model.generate(enc.input_ids, {
411
+ attention_mask: enc.attention_mask,
412
+ max_new_tokens: 1,
413
+ do_sample: false,
414
+ output_scores: true,
415
+ return_dict_in_generate: true,
416
+ eos_token_id: eosId,
417
+ pad_token_id: padId,
418
+ use_cache: false // <-- key change
419
+ });
420
+ }
421
  const dt = (performance.now() - t0) | 0;
422
 
423
  // --- from here, keep your existing logits -> softmax -> Top-K code ---