TECH

How Gemini and OpenAI Updates Affect Apple's Artificial Intelligence Strategy

Apple still has a lot of catching up to do if it wants to compete with Google and OpenAI

Google and OpenAI have announced significant updates to their artificial intelligence models and capabilities, providing additional competition to Apple ahead of WWDC.

On Monday, OpenAI announced its innovative GPT-4o artificial intelligence model and an all-new app for Mac, and on Tuesday, Google announced major improvements to its Gemini software. Both companies have demonstrated many great features, which has resulted in making the market even more competitive.

Although Apple appears to be lagging far behind in its AI endeavors, partnering with Google or OpenAI could be an easy way to offer generative AI features to its users. At the very least, rumors suggest that Apple is ready to go down this route.

OpenAI Updates

OpenAI recently introduced GPT-4o, a new multimodal version of the company's GPT Artificial Intelligence Model that contains advanced capabilities for processing various types of input data.

Unlike its predecessors, GPT-4o will be able to use a single neural network to process audio, images and text, offering significant model improvements as a result. Speed ​​and language processing improvements were also touted during the product announcement.

GPT-4o OpenAI will be able to understand and convey emotions. During a recent company event, team members demonstrated this by asking a model to analyze facial expressions and identify the specific emotions the user was expressing.

ChatGPT OpenAI is now officially available on macOS

With an enhanced voice mode feature that provides audio output as speech, GPT-4o can adjust the tone of its voice , making it more robotic or more natural depending on the user's request.

The company also launched a new ChatGPT desktop application available for macOS and introduced a new API for developers. GPT-4o will be available to users through a gradual rollout process,

Google Gemini updates

Google unveiled a host of improvements to its Gemini model at its I/O Developer Conference on Tuesday. The new and improved Google Gemini will be able to understand more complex user input and images by taking into account the context behind them.

Google Gemini is a generative AI tool

AI software will have new context-aware capabilities, meaning it will be able to see everything on screen, be it a PDF file, a video, or a series of text messages. Gemini will be able to collect information and generate output, but only on select Android devices.

For example, with the new Search Circle option, users will be able to select individual objects in an image and instantly get Google search results about that object.

Another feature available exclusively on Android will give users the ability to analyze YouTube videos and PDF files using Gemini Advanced. With a paid service, users will be able to ask specific questions and receive answers taken from the content of the specified video or PDF file.

Google's updated Gemini will be able to summarize long conversations and isolate key information from documents, images and videos, all of which will be very useful for end users. Apple implements similar features in its own products.

What we know about Apple's AI strategy

Apple is noticeably behind the competition when it comes to AI offerings, but that could all change very soon with the announcement of iOS 18 in early June.

For over a year, Apple has been working on its own Large Language Model (LLM) known as Ajax. With its generative AI software, the company aims to offer new features similar to those Google and OpenAI announced in early May.

As part of its recent AI push, Apple is expected to introduce several AI-powered features in its new operating systems. Document and web page analysis, text summarization, image captioning, and response generation are all in development.

The company aims to introduce generative AI technology into its existing portfolio of core system applications. As a result, apps like Notes, Safari, Messages, Mail, Siri, and Spotlight Search are expected to get AI-enabled improvements in one way or another.

Apple's Ajax LLM will improve Safari, Spotlight, and Messages

In terms of actual functionality, however, there are limits to what Apple has been able to achieve. The on-device artificial intelligence model being tested is only capable of basic text analysis and generating basic responses on the device.

More advanced features will likely require cloud processing, which is why Apple is reportedly seeking a licensing deal with OpenAI. This will allow Apple to offer various AI-related improvements that its own models cannot provide on devices.

A separate rumor claims that Apple wants to create an “AI App Store” through which users could purchase AI-themed apps and products from other companies. In theory, this will give users the ability to use paid versions of products such as Gemini Advanced.

We'll soon have a better understanding of Apple's AI efforts as the company is expected to unveil its new generative AI features at its annual Worldwide Developers Conference on June 10.

Follow AppleInsider on Google News.

Leave a Reply