Meta Puts Two New Generative AI Projects On Display That Assist Users With Video Creation Through Text Prompts

Tech giant Meta will soon give its Instagram and Facebook users the chance to produce videos with ease, using AI-based text prompts.

The company recently unveiled a sneak-peak preview version of what users can expect from the firm’s latest range of generative AI campaigns. This includes giving people the chance to enable customization of pictures in-stream, allowing them access to a wide range of integral applications.

Both of the latest offerings by Meta are linked to its popular Emu project that has to do with AI research. The latter is designed to explore some interesting ways by which generative AI technology may be used via prompts to give rise to such interesting and powerful visual projects.


The first one in question is dubbed Emu Video which enables users to generate short clips, all depending on what text prompt is added. This was better described with the help of examples where the EMU Video can produce videos of high quality, all depending on the simple degree of text incorporated as well as the types of still pictures being included at the forefront.

Meta went on to mention that it’s a unified architecture designed to generate videos, giving out responses to so many different types of inputs. That can be picture-based, text-based, or a combo of both realms. They further added how it’s a simplified process that is divided into two. The first one produces pictures supported by text prompts and it produces video depending on both the picture generated and the text prompt added.

You can consider that option to be a split technique for video generation that gives the company the chance to train its video production models in a great manner.

Hence, if you need to, you would be able to produce video clips depending on product pictures with text prompts that open up a whole array of creative opportunities for brands to play with.

Emu Video can produce videos that are four seconds in duration and run at a whopping 16 frames each second. They will be super impressive, better than what Meta had imagined with its previous talked-about endeavor where it was supporting a text-to-video production process that was unveiled last year.

Meta agrees that this new model for creating videos is better preferred than other projects with similar themes, adding how 96% of the majority felt the quality was better while 85% had full faith in the whole text prompt producing accurate results.

In the end, such models can add that degree of animation to images provided by users depending on text prompts which outperforms by a huge margin when compared to other options.

From the looks of it, the tool is very impressive and might actually have plenty of uses as well. It all depends on whether it can outperform in real-world applications or not because for now, it does seem tempting and promising, making it a huge deal for the company’s range of tools based on Generative AI technology.

Before we forget, a special mention must be given to the tiny watermark toward the clip’s bottom left-hand corner. This is the firm’s latest tag for AI-generated material. It really is working hard toward producing digitalized watermarks across content produced via AI, better known as synthetic material.

And we agree that they can be altered, but with videos, it’s going to be a tougher approach than what was seen in the past.

Meanwhile, the second feature being talked about is dubbed Emu Edit, and that gives users the chance to give rise to more personalized edits in the pictures. Interestingly, it is all based on conversational prompts. Therefore, users will not be required to highlight which part of the picture needs an edit. You can just ask it to make the edit and they’ll know immediately which image part needs the edit.

This is guaranteed to be a huge assistance when it comes to making edits to AI pictures and producing a more specific variation, depending on what the user really needs.

As you can probably guess by now, both projects have endless possibilities and potential. Moreover, we can surely see how so many brands and creators can now make better use of generative AI technology in a plethora of ways.

Read next: Galileo Claims GPT-4 Excels in Multitasking with Minimal Hallucinations Among LLMs
Previous Post Next Post