The Google Prompt Engineering Whitepaper is excellent, so I created a set of knowledge cards with ChatGPT, 😄.
🛠️ Best Practices for Effective Prompting
| Principle |
Key Idea |
Example / Tip |
| Provide Examples |
Use one-shot or few-shot examples to show the model what good output looks like. |
✅ Include 3-5 varied examples in classification prompts. |
| Design with Simplicity |
Clear, concise, and structured prompts work better than vague or verbose ones. |
❌ "What should we do in NY?" -> ✅ "List 3 family attractions in Manhattan." |
| Be Specific About Output |
Explicitly define output length, format, tone, or constraints. |
"Write a 3-paragraph summary in JSON format." |
| Instructions > Constraints |
Tell the model what to do, not what not to do. |
✅ "List top consoles and their makers." vs ❌ "Don't mention video game names." |
| Control Token Length |
Use model config or prompt phrasing to limit response length. |
"Explain in 1 sentence" or set token limit. |
| Use Variables |
Template prompts for reuse by inserting dynamic values. |
Tell me a fact about {city} |
| Experiment with Input Style |
Try different formats: questions, statements, instructions. |
🔄 Compare: "What is X?", "Explain X.", "Write a blog about X." |
| Shuffle Classes (Few-Shot) |
Mix up response class order to avoid overfitting to prompt pattern. |
✅ Randomize class label order in few-shot tasks. |
| Adapt to Model Updates |
LLMs evolve; regularly test and adjust prompts. |
🔄 Re-tune for new Gemini / GPT / Claude versions. |
| Experiment with Output Format |
For structured tasks, ask for output in JSON/XML to reduce ambiguity. |
"Return response as valid JSON." |
| Document Prompt Iterations |
Keep track of changes and tests for each prompt. |
📝 Use a table or versioning system. |
🎯 Core Prompting Techniques
| Technique |
Description |
Example Summary |
| Zero-Shot |
Ask the model directly without any example. |
🧠 "Classify this review as positive/neutral/negative." |
| One-Shot |
Provide one example to show expected format/output. |
🖋️ Input + Example -> New input |
| Few-Shot |
Provide multiple examples to show a pattern. |
🎓 Use 3-5 varied examples. Helps with parsing, classification, etc. |
| System Prompting |
Set high-level task goals and output instructions. |
🛠️ "Return the answer as JSON. Only use uppercase for labels." |
| Role Prompting |
Assign a persona or identity to the model. |
🎭 "Act as a travel guide. I'm in Tokyo." |
| Contextual Prompting |
Provide relevant background info to guide output. |
📜 "You're writing for a retro games blog." |
| Step-Back Prompting |
Ask a general question first, then solve the specific one. |
🔄 Extract relevant themes -> Use as context -> Ask final question |
| Chain of Thought (CoT) |
Ask the model to think step-by-step. Improves reasoning. |
🤔 "Let's think step by step." |
| Self-Consistency |
Generate multiple CoTs and pick the most common answer. |
🗳️ Run same CoT prompt multiple times, use majority vote |
| Tree of Thoughts (ToT) |
Explore multiple reasoning paths in parallel for more complex problems. |
🌳 LLM explores different paths like a decision tree |
| ReAct (Reason & Act) |
Mix reasoning + action. Model decides, acts (e.g. via tool/API), observes, and iterates. |
🤖 Thought -> Action -> Observation -> Thought |
| Automatic Prompting |
Use LLM to generate prompt variants automatically, then evaluate best ones. |
💡 "Generate 10 ways to say 'Order a small Metallica t-shirt.'" |
⚙️ LLM Output Configuration Essentials
| Config Option |
What It Does |
Best Use Cases |
| Max Token Length |
Limits response size by number of tokens. |
📦 Prevent runaway generations, control cost/speed. |
| Temperature |
Controls randomness of token selection (0 = deterministic). |
🎯 0 for precise answers (e.g., math/code), 0.7+ for creativity. |
| Top-K Sampling |
Picks next token from top K probable tokens. |
🎨 Higher K = more diverse output. K=1 = greedy decoding. |
| Top-P Sampling |
Picks from smallest set of tokens with cumulative probability ≥ P. |
💡 Top-P ~0.9-0.95 gives quality + diversity. |
🔁 How These Settings Interact
| If You Set... |
Then... |
temperature = 0 |
Top-K/Top-P are ignored. Most probable token is always chosen. |
top-k = 1 |
Like greedy decoding. Temperature/Top-P become irrelevant. |
top-p = 0 |
Only most probable token considered. |
high temperature (e.g. >1) |
Makes Top-K/Top-P dominant. Token sampling becomes more random. |
✅ Starting Config Cheat Sheet
| Goal |
Temp |
Top-P |
Top-K |
Notes |
| 🧠 Precise Answer |
0 |
Any |
Any |
For logic/math problems, deterministic output |
| 🛠️ Semi-Creative |
0.2 |
0.95 |
30 |
Balanced, informative output |
| 🎨 Highly Creative |
0.9 |
0.99 |
40 |
For stories, ideas, writing |
Top comments (0)