
Immediately, most generative picture fashions principally fall into two essential classes: diffusion fashions, like Secure Diffusion, or autoregressive fashions, like OpenAI’s GPT-4o. However Apple simply launched two papers that present how there could be room for a 3rd, forgotten method: Normalizing Flows. And with a splash of Transformers on high, they could be extra succesful than beforehand thought.
First issues first: What are Normalizing Flows?
Normalizing Flows (NFs) are a kind of AI mannequin that works by studying mathematically remodel real-world knowledge (like pictures) into structured noise, after which reverse that course of to generate new samples.
The large benefit is that they’ll calculate the precise probability of every picture they generate, a property that diffusion fashions can’t do. This makes flows particularly interesting for duties the place understanding the likelihood of an end result actually issues.
However there’s a purpose most individuals haven’t heard a lot about them these days: Early flow-based fashions produced pictures that seemed blurry or lacked the element and variety supplied by diffusion and transformer-based programs.
Examine #1: TarFlow
Within the paper “Normalizing Flows are Succesful Generative Fashions”, Apple introduces a brand new mannequin referred to as TarFlow, quick for Transformer AutoRegressive Circulate.
At its core, TarFlow replaces the outdated, handcrafted layers utilized in earlier move fashions with Transformer blocks. Mainly, it splits pictures into small patches, and generates them in blocks, with every block predicted primarily based on all those that got here earlier than. That’s what’s referred to as autoregressive, which is similar underlying methodology that OpenAI at the moment makes use of for picture technology.

The important thing distinction is that whereas OpenAI generates discrete tokens, treating pictures like lengthy sequences of text-like symbols, Apple’s TarFlow generates pixel values instantly, with out tokenizing the picture first. It’s a small, however important distinction as a result of it lets Apple keep away from the standard loss and rigidity that usually include compressing pictures into a hard and fast vocabulary of tokens.
Nonetheless, there have been limitations, particularly when it got here to scaling as much as bigger, high-res pictures. And that’s the place the second research is available in.
Examine #2: STARFlow
Within the paper “STARFlow: Scaling Latent Normalizing Flows for Excessive-resolution Picture Synthesis”, Apple builds instantly on TarFlow and presents STARFlow (Scalable Transformer AutoRegressive Circulate), with key upgrades.
The most important change: STARFlow now not generates pictures instantly in pixel house. As a substitute, it principally works on a compressed model of the picture, after which arms issues off to a decoder that upsamples all the pieces again to full decision on the ultimate step.

This shift to what’s referred to as latent house means STARFlow doesn’t must predict tens of millions of pixels instantly. It might probably give attention to the broader picture construction first, leaving positive texture element to the decoder.
Apple additionally reworked how the mannequin handles textual content prompts. As a substitute of constructing a separate textual content encoder, STARFlow can plug in current language fashions (like Google’s small language mannequin Gemma, which in concept may run on-device) to deal with language understanding when the person prompts the mannequin to create the picture. This retains the picture technology aspect of the mannequin centered on refining visible particulars.
How STARFlow compares with OpenAI’s 4o picture generator
Whereas Apple is rethinking flows, OpenAI has additionally just lately moved past diffusion with its GPT-4o mannequin. However their method is basically completely different.
GPT-4o treats pictures as sequences of discrete tokens, very similar to phrases in a sentence. While you ask ChatGPT to generate a picture, the mannequin predicts one picture token at a time, constructing the image piece by piece. This provides OpenAI monumental flexibility: the identical mannequin can generate textual content, pictures, and audio inside a single, unified token stream.
The tradeoff? Token-by-token technology might be sluggish, particularly for big or high-resolution pictures. And it’s extraordinarily computationally costly. However since GPT-4o runs completely within the cloud, OpenAI isn’t as constrained by latency or energy use.
In brief: each Apple and OpenAI are shifting past diffusion, however whereas OpenAI is constructing for its knowledge facilities, Apple is clearly constructing for our pockets.
FTC: We use earnings incomes auto affiliate hyperlinks. Extra.