Apple, in collaboration with researchers from the University of California, has released its own AI-based image editing model that and anybody can try.
The model has been dubbed "MGIE," which stands for MLLM-Guided Image Editing and allows users to edit images based on natural language instructions. It leverages multimodal large language models (MLLMs), which combine various types of information, such as text, photos, and videos, to understand and generate human-like language.
While studies on MLLMs have shown promising results in understanding and generating images, they haven't been implemented widely yet.
MGIE was presented in a paper at the International Conference on Learning Representations 2024. The paper showcases how MGIE can improve automatic metrics and human evaluation while also maintaining competitive inference efficiency.
MGIE utilizes MLLMs to derive expressive instructions from user input. Venture Beat offers an example: a user could tell MGIE to "make the sky more blue," and MGIE would interpret this as "increase the saturation of the sky region by 20%."
Then, once the instructions are derived from the natural-language input, it generates an edited version of the image. It can be used to create a wide range of edits, ranging from color adjustments to generating or removing parts of an image.
MGIE is open-source and available on GitHub for anyone to try. The GitHub page allows users to snag the code, data, and pre-trained models.
The speed of image generation will vary significantly based on hardware performance. Also, models such as this are incredibly RAM hungry.
Apple has been working behind the scenes on its AI-assisted features for some time now. In January, code in the iOS 17.4 beta suggested that Apple is working on a new AI-powered version of Siri.
Apple stock has taken a beating as of late, in part because analysts have loudly proclaimed that the company is behind Meta, Google, and Microsoft in generative AI implementation. It's not clear why this wasn't a problem when it wasn't first to a mobile phone, a tablet, a smartwatch, or a VR headset, but is with generative AI.
Perhaps to dissuade those analysts, Apple CEO Tim Cook recently said that Apple is spending "a tremendous amount of time and effort" on AI features that will be announced in the coming months. He also noted that Apple has been working internally on generative AI.
36 Comments
Ultimately, it's just a huge waste of electricity similar to crypto.
Anyone thinking that if you click on the link provided to try editing your photos will be disappointed. Unless you are a coder or computer scientist it’s just gobbledegook.
It's fairly obvious that Apple is rushing out at least a half-baked something in order to claim "we do AI too, just like those other guys," and for the silliest of reasons: Prop up a stock price.
Apple the corporation doesn't get a cent more if the stock goes up, or a penny less if it goes down. But Cook and the rest of the upper management team certainly does, with much of their pay coming in the form of company stock rather than cash. Thus we have these kinds of kneejerk things announced by them now instead of the slow and steady development without broadcasting the internal baby steps that we typically expect from Apple.
It's silly when Samsung/Google/random tech company, responds to some new Apple service or product with "Hey, look what we're working on, it's just as good!" and equally as silly when Apple turns around and does the same thing.