



Instead of the classic ChatGPT’s fixed personality, API users can customize user experiences to make it talk like a pirate or tutor in the Socratic method. OpenAI says users can now prescribe their AI’s style and task by giving it directions via a system message. The company also asserts that GPT-4 is more collaborative for creative and technical writing tasks, noting it is reliable, creative, and able to handle much more nuanced instructions than GPT-3.5. OpenAI says GPT-4 has a wider general knowledge across multiple domains and can pass the bar exam in the 90th percentile, compared to GPT-3.5’s 10th percentile score. It Understands Nuance, Can Beat You at Trivia, and Ace the Bar Exam The old version of ChatGPT could only remember content as far back as 8,000 words, or four to five pages. It also has a longer working memory of about 64,000 words, or about 50 pages, so it can remember and refer back to things from earlier in a conversation. GPT-4 can now handle 25,000 words of text, allowing for use cases like long-form content creation and long document search and analysis. GPT-4 is powering a “virtual volunteer” mode on the app that OpenAI claims can often generate the same level of detail and assistance as a human volunteer. To prepare the image input capability for wider availability, the company says it has partnered with Be My Eyes, a mobile app that connects blind and low-vision people with a roster of over 6 million sighted volunteers who can help with accessibility using live video calls. One caveat: GPT-4’s image modality is still in research mode and is not yet available to the public, even to those who subscribe to ChatGPT+. When asked why the image is funny, GPT-4 correctly identifies the objects in the picture and replies with “The humor in this image comes from the absurdity of plugging a large, outdated VGA connector into a small, modern smartphone charging port.” It detected the humor in a meme where an iPhone is plugged into a charger with a VGA cable instead of a Lightning cable. GPT-4 is now a multi-modal system that can accept images as inputs to do tasks like generating captions, classifying images, and analyzing the context of the images, including humor. This example pulled from GPT-4’s research paper shows the model can understand the contents and context of an image.
