The new GPT-4 from OpenAI can comprehend inputs in both text and images

In advance of Thursday’s Microsoft Future of Work event and just after Google announced Workspace AI on Tuesday, OpenAI has published GPT-4, the most recent version of their generative pre-trained transformer system. The new and enhanced GPT-4 will be able to generate text on input photos as well, unlike the current generation GPT-3.5, which powers OpenAI’s immensely successful ChatGPT conversational bot, which can only read and answer with text. The OpenAI team said on Tuesday that while it “exhibits human-level performance on various professional and academic benchmarks,” it “is less capable than humans in many real-world scenarios.”

According to reports, OpenAI, which collaborates with Microsoft to improve GPT’s capabilities, has just renewed their partnership and spent the last six months optimising and improving the system’s performance in response to user comments from the recent ChatGPT hullabaloo. According to the firm, GPT-4 performed “around the top 10 percent of test takers” on simulated exams (including the Uniform Bar, LSAT, GRE, and other AP tests) whereas GPT-3.5 performed “in the bottom 10 percent of test takers.” Furthermore, in numerous benchmark tests, the new GPT has outscored existing cutting-edge large language models (LLMs). In addition, the business asserts that the new system has outperformed its predecessor in “factuality, steerability, and refusing to go outside of guardrails.”