codelion commited on
Commit
d2bcbf6
Β·
verified Β·
1 Parent(s): 6981b70

Upload app.py

Browse files
Files changed (1) hide show
  1. app.py +17 -18
app.py CHANGED
@@ -748,24 +748,23 @@ Bad patterns to avoid:
748
  # Task
749
  Rewrite the prompt to MAXIMIZE accuracy on sentiment classification.
750
 
751
- CRITICAL REQUIREMENTS (these DIRECTLY affect score):
752
- 1. βœ“ MUST include word "sentiment" β†’ model response will contain "sentiment" keyword
753
- 2. βœ“ MUST use pattern "[Action] sentiment: {{input}}" β†’ triggers correct response format
754
- 3. βœ“ Keep it reasonable (under 1000 chars) β†’ focus on clarity and effectiveness
755
- 4. βœ“ MUST keep {{input}} placeholder EXACTLY as-is
756
-
757
- PROVEN WORKING PATTERNS (use these!):
758
- - "Classify sentiment: {{input}}" ← BEST (scores ~90%)
759
- - "Determine sentiment: {{input}}" ← Also works well (~85%)
760
- - "Sentiment of: {{input}}" ← Good (~80%)
761
-
762
- PATTERNS THAT FAIL (avoid!):
763
- - ❌ "What is the sentiment?" - question format, no {{input}}
764
- - ❌ "Review: {{input}}" - missing "sentiment" keyword
765
- - ❌ "Please analyze the sentiment..." - too long, word "please"
766
-
767
- Generate a DIRECT, EFFECTIVE prompt using the working pattern above.
768
- You have up to 1000 characters to craft the best possible prompt.
769
 
770
  Output ONLY the new prompt between ```text markers:
771
 
 
748
  # Task
749
  Rewrite the prompt to MAXIMIZE accuracy on sentiment classification.
750
 
751
+ CRITICAL REQUIREMENTS:
752
+ 1. The model's response MUST include the word "sentiment"
753
+ 2. The model's response MUST include either "positive" or "negative"
754
+ 3. You MUST keep the {{input}} placeholder exactly as {{input}}
755
+
756
+ EVALUATION CRITERIA:
757
+ - Responses are evaluated by checking if they contain "sentiment" AND ("positive" OR "negative") in the first 150 characters
758
+ - The response must match the true label (positive=1, negative=0)
759
+
760
+ Be creative! Try different approaches:
761
+ - Direct instructions vs detailed explanations
762
+ - Short prompts vs longer contextual prompts
763
+ - Imperative commands vs questions
764
+ - System-style vs user-style prompts
765
+ - With or without examples/formatting instructions
766
+
767
+ The goal is to maximize the model's accuracy. Experiment freely!
 
768
 
769
  Output ONLY the new prompt between ```text markers:
770