{"id":433,"date":"2024-02-23T13:00:00","date_gmt":"2024-02-23T14:00:00","guid":{"rendered":"https:\/\/reshebniki-online.com\/?p=433"},"modified":"2024-02-29T15:39:47","modified_gmt":"2024-02-29T15:39:47","slug":"what-two-years-of-ai-development-can-tell-us-about-sora","status":"publish","type":"post","link":"https:\/\/reshebniki-online.com\/index.php\/2024\/02\/23\/what-two-years-of-ai-development-can-tell-us-about-sora\/","title":{"rendered":"What two years of AI development can tell us about Sora"},"content":{"rendered":"
\n
\n \"An
A screenshot of a video generated by Sora, OpenAI\u2019s generative video model. | Sora\/OpenAI CEO Sam Altman<\/a><\/figcaption><\/figure>\n

If you want to know the future of OpenAI\u2019s latest tool, take a look at Midjourney and DALL-E 2.<\/p>\n

Remember when AI<\/a> art generators became widely available in 2022 and suddenly the internet was full of uncanny pictures that were very cool but didn\u2019t look quite right on close inspection? Get ready for that to happen again \u2014 but this time for video.<\/p>\n

Last week, OpenAI released Sora, a generative AI model that produces videos based on a simple prompt. It\u2019s not available to the public yet, but CEO Sam Altman showed off its capabilities by taking requests on X, formerly known as Twitter<\/a>. Users replied with short prompts: \u201ca monkey playing chess in a park<\/a>,\u201d or \u201ca bicycle race on ocean with different animals as athletes<\/a>.\u201d It\u2019s uncanny, mesmerizing, weird, beautiful \u2014 and prompting the usual cycle of commentary.<\/p>\n

Some people are making strong claims about Sora\u2019s negative effects<\/a>, expecting a \u201cwave of disinformation<\/a>\u201d \u2014 but while I (and experts) think future powerful AI systems pose really serious risks<\/a>, claims that a specific model will bring the disinformation wave upon us have not held up so far. <\/p>\n

Others are pointing at Sora\u2019s many flaws as representing fundamental limitations<\/a> of the technology \u2014 which was a mistake when people did it with image generator models and which, I suspect, will be a mistake again. As my colleague A.W. Ohlheiser pointed out<\/a>, \u201cjust as DALL-E and ChatGPT improved over time, so could Sora.\u201d<\/p>\n

The predictions, both bullish and bearish, may yet pan out \u2014 but the conversation around Sora and generative AI would be more productive if people on all sides took into greater account all the ways in which we\u2019ve been proven wrong these last couple of years.<\/p>\n

What DALL-E 2 and Midjourney can teach us about Sora<\/h3>\n

Two years ago, OpenAI announced DALL-E 2<\/a>, a model that could produce still images from a text prompt. The high-resolution fantastical images it produced were quickly all over social media, as were the takes<\/a> on what to think of it: Real art? Fake art? A threat to artists? A tool for artists? A disinformation machine? Two years later, it\u2019s worth a bit of a retrospective if we want our takes on Sora to age better. <\/p>\n

DALL-E 2\u2019s release was only a few months ahead of Midjourney<\/a> and Stable Diffusion<\/a>, two popular competitors. They each had their strengths and weaknesses. DALL-E 2 did more photorealistic pictures and adhered a little better to prompts; Midjourney was \u201cartsier.\u201d Collectively, they made AI art available at the click of a button to millions. <\/p>\n

Much of the societal impact of generative AI then didn\u2019t come directly from DALL-E 2, but from the wave of image models it led. Likewise, we might expect that the important question about Sora isn\u2019t just what Sora can do, but what its imitators and competitors will be able to do.<\/p>\n

Many people thought that DALL-E and its competitors heralded a flood of deepfake propaganda and scams<\/a> that\u2019d threaten our democracy. While we may well see an effect like that some day, those calls now seem to have been premature. The effect of deepfakes on our democracy \u201calways seems just around the corner,\u201d analyst Peter Carlyon wrote in December<\/a>, noting that most propaganda continues to be of a more boring kind \u2014 for example, taking remarks out of context, or images of one conflict shared and mislabeled as being from another. <\/p>\n

Presumably at some point this will change, but there should be some humility about claims that Sora will be that change. It doesn\u2019t take deepfakes to lie to people, and they remain an expensive way to do it. <\/strong>(AI generations are relatively cheap, but if you\u2019re going for something specific and convincing, that\u2019s much pricier. A tsunami of deepfakes implies a scale that spammers mostly can\u2019t afford at the moment.)<\/p>\n

But the place where it seems most crucial to me to remember the last two years of AI history is when I read criticisms of Sora\u2019s images for being clumsy, stilted, inhuman, or obviously flawed. It\u2019s true, they are. Sora \u201cdoes not accurately model the physics of many basic interactions,\u201d OpenAI\u2019s research release acknowledges, adding<\/a> that it has trouble with cause and effect, mixing up left and right, and following a trajectory. <\/p>\n