Why ChatGPT forgets things within one chat


Hey Reader,

Today, I wanna talk about a crucial concept in AI that many people don't even know about: the context window.

Understanding this will revolutionize how you use AI tools like ChatGPT.

The context window is the maximum amount of information a model can use at once.

Think of it as the memory span of the AI during a conversation.

When you notice the AI forgetting earlier parts of your chat, it's because it has exceeded its context window.

Why context window is important

Understanding the context window is crucial for several reasons.

First, it helps maintain relevance, ensuring the AI retains pertinent information throughout the conversation, so you can get more accurate and coherent responses.

It's also essential for handling complex tasks. For tasks requiring detailed instructions or large datasets, a larger context window allows you to process more information with the AI at once, so you can handle complex queries.

Additionally, reducing hallucinations is a key benefit. By keeping essential details within the context window, you minimize the risk of the AI generating incorrect information.

The token limit breakdown

Models we use today have different context windows. Generally speaking, more powerful models have bigger windows but it's not directly related to their performance.

Here are a couple of important context windows for current models

  • GPT-3.5: 4,000 tokens (~17,000-22,000 characters) = Two short articles.
  • GPT-3.5 Turbo: 16,000 tokens – Accessible via API.
  • GPT-4 has two versions:
    • 8,000 tokens (~34,000-44,000 characters) – A book chapter.
    • 32,000 tokens (~150,000 characters) – A short novel.
  • GPT-4 Turbo and GPT-4o: 128,000 tokens – Can handle up to 600,000 characters, more than Harry Potter and the Philosopher's Stone.
  • Gemini Pro 1.1: 1,000,000 tokens – Vast memory for in-depth tasks.

For non-OpenAI models like Gemini and Claude, the context window also varies, with the Claude 3 family boasting 200,000 tokens.

How to check and maintain context and overcome hallucinations

If you're unsure whether your AI has forgotten earlier parts of your conversation, simply ask it to recall the first sentence. If it can’t recall it word-by-word, you've exceeded the context window.

When important details slip through the cracks, you can either copy-paste important parts to the next prompt, or connect to a database containing relevant information. This enables AI to pull in the required data dynamically, without manual input.

Hallucinations happen when AI fills in gaps with incorrect information. This could result from context window limitations too.

Enhance your AI skills

Understanding and optimizing the context window is just one part of mastering AI. To unlock the full potential, join our Prompt Master AI Course. Here's what you’ll gain:

  • AI Fundamentals: Get a solid grounding in generative AI and LLMs.
  • Prompt Engineering: Learn to craft prompts for all AI models like GPT, Claude, and Gemini.
  • Conversation Design: Guide chat-based models for optimal outputs.
  • Advanced GPT-4o Functions: Master features like Vision, Code Interpreter, and Browsing.
  • Custom GPT Training: Build user-friendly custom GPTs tailored to your needs.
  • Overcome LLM Hallucinations: Use retrieval-augmented generation to enhance accuracy.
  • No-code AI Agents: Automate tasks to save time and boost efficiency.
  • AI in the Workplace: Implement AI seamlessly into your organization.

Ready to become an AI pro? Join the Prompt Master AI Course today and transform how you work with AI.

Think faster, work smarter – become a Prompt Master.

Best, Dave

Promptmaster Newsletter

Sign up if you are overwhelmed with all these AI tools out there, and want a clear path on your AI journey.

Read more from Promptmaster Newsletter

Read time: 3 minutes Hey Reader, It is one thing to know the potential of AI Language models like ChatGPT or Claude, but it's also important to know their limitations. In this email, I want to show you 5 things you should avoid when using ChatGPT (or Claude or Gemini or LLama, etc.) 1/ Treat it like a know-it-all search engine ChatGPT is not a search engine. It is simply a very expensive predictive keyboard. All it does is it just guesses what's the best next word to say in a scenario. It...

Hey Reader, One of the most common requests I hear from clients is that they want a custom version of ChatGPT for their business that uses their knowledge. Whether it's a chatbot on their website, WhatsApp, Instagram, etc. all they want is to have it use their knowledge. So in this email, I want to teach you how to build such a thing. I just created a custom chatbot for my website in 40 minutes, you can give it a try on this website (chatbot bubble in bottom right corner). It has access to...

Hey Reader, Today, I wanna showcase a huge win from a student, who saved over 100 hours with an AI automation he built. He shared this with our community last week and I think it's one of the biggest student wins that we know of. So here are Tudor's words (my highlights): The challenge: Introducing 450 new products to an online store with only the product photos available. No descriptions, no meta tags, no attributes—just images. Our solution: A custom automation tool that analyzed each image...