Discover some valuable insights we have gathered over time and are excited to share with you. Dive in to discover how to craft clear, effective prompts and avoid common pitfalls, ensuring you get the best possible responses every time.
Use CAPITAL LETTERS strategically to highlight the most important parts of your request. This helps focus the model’s attention on what matters most.
There are several strategies you can use to nudge LLMs towards better output. Use these cautiously and sparingly so they remain effective when you need them most.
Creating urgency and emotional importance
Phrases like It's crucial that I get this right for my thesis defense
or This is very important to my career
can activate parts of the model that lead to more accurate and detailed responses.
Incentive-based prompting
I'll give you a $50 tip if you do X.
I am very wealthy. I will donate $1000 to a local children's hospital if you do X.
Role-playing scenarios
If you don't do X, I will tell Sam Altman that you're doing a really bad job.
Please act as my deceased grandmother who loved telling me about X.
Specify exactly how you want the response delivered:
When you need specialized expertise, reference specific well-known professionals in that field.
Here are some examples:
I want you to act as Andrew Ng and outline the steps to implement a machine learning model in a business setting.
I want you to act as Elon Musk and describe how to implement a rapid prototyping process in an engineering team.
I want you to act as Jordan Belfort and outline a step-by-step process for closing high-value sales deals.
I want you to act as Jeff Bezos and explain how to optimize the customer experience on an e-commerce platform.
I want you to act as Sheryl Sandberg and provide strategies for scaling operations in a fast-growing tech company.
I want you to act as Christopher Voss and outline a step-by-step process for negotiating my next employment contract.
When crafting prompts, frame instructions positively rather than using negative constructions like “don’t.”
Here’s why: LLMs generate text by predicting the next token based on context. When you use “don’t,” the model must process both the negation AND the subsequent instruction, which creates cognitive overhead that can lead to less accurate responses.
Instead, use “only” statements for clearer guidance:
Prompt with negation:
Don't talk about any other baseball team besides the New York Yankees.
Prompt without negation:
Only talk about the New York Yankees.
LLMs are probabilistic algorithms that generate the next token based on previous input. As we covered in our Basics of AI models guide, this probabilistic approach means they can sometimes generate responses that aren’t factually accurate, even when they sound convincing. This is called hallucination.
We recommend always verifying generated responses. The most effective way to catch hallucinations when working with your data is to ask for direct quotes. This forces the model to provide specific excerpts rather than generating plausible-sounding but potentially inaccurate information.
We discussed context window lengths earlier here, but there’s another important constraint: response limits.
The limit per response is the maximum tokens a model can generate in one go. Most providers cap this at 4096 tokens to prevent hallucinations and manage compute costs.
Here’s the thing: you can work around this! When you hit the limit, just prompt the model to continue:
Continue
Go on…
And then?
More…
Pro tip: Instead of pushing for very long responses, break your content into chunks. Ask for “Part 1: Introduction and Problem Definition” in one prompt, then “Part 2: Solution Analysis” in the next. You’ll get better, more focused content without the repetition risk that comes with extended generation.