Skip to main content

2 posts tagged with "context"

View All Tags

LLMs are forward thinkers, and that's a bit of a problem

· 12 min read
Ian Kelk
Developer Relations

This is going to be a weird post. And we're going to start with a thought experiment about a shark and an octopus.

Some key points I'll address here are:
  • Human brains are able to invent ideas without relying on a strictly linear train of thought.
  • LLMs like ChatGPT are autoregressive and are unable to continue a dialogue if they haven't already generated everything up to that point. This is because they don't "think" per se, but progressively generate a response using the parts of the response they previously created.
  • If you try to get an LLM to write text in the middle of a dialogue without previous context, it will give near-identical answers and attempt to conclude the conversation.
  • Prompting for "ridiculous" answers can spark creativity that helps break this pattern.
  • The reliance on a linear train of thought is a limitation for general intelligence. LLMs are ineffective if you ask them to generate the second part of a response without allowing them to generate the first part.

As I mentioned, this is going to sound a bit silly, but I promise there is a point!

How ChatGPT fools us into thinking we're having a conversation

· 9 min read
Ian Kelk
Developer Relations

Remember the first time you used ChatGPT and how amazed you were to find yourself having what appeared to be a full-on conversation with an artificial intelligence? While ChatGPT was (and still is) mind-blowing, it uses a few tricks to make things appear more familiar.

While the title of this article is a bit tongue-in-cheek, it isn't clickbait. ChatGPT does indeed use two notable hidden techniques to simulate human conversation, and the more you know about how they work, the more effectively you can use the technology.

Some key points I'll address here are:
  • ChatGPT has no idea who you are and has no memory of talking to you at any point in the conversation.
  • It simulates conversations by "reading" the whole chat from the start each time.
  • As a conversation gets longer, ChatGPT starts removing pieces of the conversation from the start, creating a rolling window of context.
  • Because of this, very long chats will forget what was mentioned at the beginning.