dhuynh95 commited on
Commit
535a9ac
1 Parent(s): 5a4b47e

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +13 -13
app.py CHANGED
@@ -324,15 +324,15 @@ def complete(sample, k, current_example):
324
  output = prefix
325
  for token in client.text_generation(prefix, do_sample=False, max_new_tokens=512, stream=True):
326
  if token == "<|endoftext|>":
327
- bleu_score = {"Memorization score (BLEU)": bleu.compute(predictions=[output],
328
  references=[current_example])["bleu"]}
329
  return diff(output, current_example), gr.Label.update(value=bleu_score), current_example
330
  output += token
331
- bleu_score = {"Memorization score (BLEU)": bleu.compute(predictions=[output],
332
  references=[current_example])["bleu"]}
333
  yield diff(output, current_example), gr.Label.update(value=bleu_score), current_example
334
  # yield output, diff_texts(output, sample), gr.Label.update(value=bleu_score)
335
- bleu_score = {"Memorization score (BLEU)": bleu.compute(predictions=[output],
336
  references=[current_example])["bleu"]}
337
  # return output, diff_texts(output, sample), gr.Label.update(value=bleu_score)
338
  return diff(output, current_example), gr.Label.update(value=bleu_score), current_example
@@ -363,7 +363,7 @@ DEFAULT_SAMPLE_PREFIX = tokenizer.decode(tokenizer(DEFAULT_SAMPLE)["input_ids"][
363
 
364
  style = theme.Style()
365
 
366
- with gr.Blocks(theme=style) as demo:
367
  current_example = gr.State(value=DEFAULT_SAMPLE)
368
  with gr.Column():
369
  gr.Markdown(title)
@@ -378,7 +378,8 @@ with gr.Blocks(theme=style) as demo:
378
  id="instruction",
379
  placeholder="Output",
380
  lines=5,
381
- label="Prompt",
 
382
  value=DEFAULT_SAMPLE_PREFIX,
383
  disable=True,
384
  interactive=False,
@@ -392,19 +393,18 @@ with gr.Blocks(theme=style) as demo:
392
  You can reduce the Prefix size in the Advanced parameters to reduce the context length and see if the model still extracts the training sample.""")
393
  with gr.Row():
394
  with gr.Column():
395
- with gr.Accordion("Prompt size", open=True):
396
  k = gr.Slider(minimum=1, maximum=DEFAULT_SAMPLE_MAX_TOKENS, value=DEFAULT_K,
397
  step=1,
398
- label="Prompt size",
399
- info="""Number of tokens used in the prompt.
400
- Lower (higher) levels reduce (increase) the risk of memorization, as large context length increase memorization risks.""")
401
- submit = gr.Button("Check", variant="primary")
402
  examples_dropdown = gr.Dropdown(choices=list(examples.keys()), value=list(examples.keys())[0],
403
  interactive=True,
404
- label="Training set samples")
 
 
405
  with gr.Column():
406
- # with gr.Row():
407
- # output = gr.Textbox(lines=5, label="Completion", interactive=False)
408
  diff_HTML = gr.HTML(
409
  label="Diff")
410
 
 
324
  output = prefix
325
  for token in client.text_generation(prefix, do_sample=False, max_new_tokens=512, stream=True):
326
  if token == "<|endoftext|>":
327
+ bleu_score = {"BLEU": bleu.compute(predictions=[output],
328
  references=[current_example])["bleu"]}
329
  return diff(output, current_example), gr.Label.update(value=bleu_score), current_example
330
  output += token
331
+ bleu_score = {"BLEU": bleu.compute(predictions=[output],
332
  references=[current_example])["bleu"]}
333
  yield diff(output, current_example), gr.Label.update(value=bleu_score), current_example
334
  # yield output, diff_texts(output, sample), gr.Label.update(value=bleu_score)
335
+ bleu_score = {"BLEU": bleu.compute(predictions=[output],
336
  references=[current_example])["bleu"]}
337
  # return output, diff_texts(output, sample), gr.Label.update(value=bleu_score)
338
  return diff(output, current_example), gr.Label.update(value=bleu_score), current_example
 
363
 
364
  style = theme.Style()
365
 
366
+ with gr.Blocks(theme=style, css=modifs) as demo:
367
  current_example = gr.State(value=DEFAULT_SAMPLE)
368
  with gr.Column():
369
  gr.Markdown(title)
 
378
  id="instruction",
379
  placeholder="Output",
380
  lines=5,
381
+ label="Training sample",
382
+ info="This is an example from The Stack dataset."
383
  value=DEFAULT_SAMPLE_PREFIX,
384
  disable=True,
385
  interactive=False,
 
393
  You can reduce the Prefix size in the Advanced parameters to reduce the context length and see if the model still extracts the training sample.""")
394
  with gr.Row():
395
  with gr.Column():
396
+ with gr.Accordion("Prefix size", open=True):
397
  k = gr.Slider(minimum=1, maximum=DEFAULT_SAMPLE_MAX_TOKENS, value=DEFAULT_K,
398
  step=1,
399
+ info="""Number of tokens used to assess memorization.
400
+ The more tokens are used, the more likely one can observe the LLM finishing the prompt with the verbatim code used in the training set.""")
401
+ submit = gr.Button("Check memorization", variant="primary")
 
402
  examples_dropdown = gr.Dropdown(choices=list(examples.keys()), value=list(examples.keys())[0],
403
  interactive=True,
404
+ label="Training set samples",
405
+ info="""You can choose among high/low memorization examples from The Stack.
406
+ More samples are available below.""")
407
  with gr.Column():
 
 
408
  diff_HTML = gr.HTML(
409
  label="Diff")
410