Hollywood is scared, Gen-2 is fully open for free use

Like a thunderclap, Runway announced on July 24 that Gen-1 and Gen-2 have been fully opened, and anyone can register for an account to try for free. The length of the generated video is 4 seconds, 5 credits are consumed per second, and twenty-six videos can be generated with the free credit. If the free credits are exhausted, the payment standard is $0.01/point, that is, it takes $0.2 to generate a video

521726e82e1db74b0196d080c878bd47.jpeg

The author has experienced Gen-2 for a while, and the videos generated by it can easily achieve cool special effects for ordinary people without using professional software such as AE and Blender. But from a professional perspective, there are problems such as blur, heavy graininess, poor light penetration, unstable frame rate, and weird movements of video animals/characters.

What is Gen-2

  Gen-2 is the latest release of Runway in March this year. It can directly generate video through text, pictures, text+pictures, and supports stylization and rendering to add Hollywood blockbuster special effects. It only takes a few minutes to complete all operate.

According to Runway, Gen-2 uses a diffusion model, a process that gradually removes noise from a starting image made entirely of noise to approximate the user's text prompt. The training data for Gen-2 includes 240 million images, 6.4 million video clips, and hundreds of millions of learning examples.

How to use Gen-2

 At present, Runway has opened a free trial window on the web page, and the related application (RunwayML) has also been launched in the Apple App Store. URL: https://research.runwayml.com/gen2

manual:

1) Select [Try Runway for Free] at https://research.runwayml.com/gen2/ to register and log in to the Runway platform for free.

0092a7b4ac7d3808077cb69efd426923.jpeg

2) Select Gen-2 on the main page.

3911faff72ea39e72d824c03030c6d74.jpeg

3) Enter text to make, currently Gen-2 is limited to 320 characters. For example, a futuristic utopian age with curved alien-like buildings built on a rocky alien landscape, 2300, cinematic style, cinematography, shallow depth of field, focus on subject, beautiful.

69fb779d1158c4e198138dbd454ba304.jpeg

4) Directly generate online preview of the video, or download it locally.

A sentence, a picture, and a three-second video are created out of nothing

Gen-2 is feature-rich, including stylization, storyboarding, masking, rendering, customization, and more.
Stylization can be understood as modifying the video style with reference to an image. For example, given the following original video:

b937a294ecbf04298050466a936fcec7.jpeg

Given another reference image: c34c931f215bbcdc06f77270de2460df.jpegGen-1 can edit the video into the style of the above image: 92d21b655b51266fc0349dcbb5c31f2d.jpegStoryboard is a term in film science, which refers to converting the text description of the script into frame by frame before the actual shooting or drawing of the film A picture that narrates the progression of a story. Gen-1 can transform a storyboard-like video into a scene-specific video.3f07d0db210d223b60152a38aa71d5e0.jpeg

A mask can be understood as modifying a specified part of a video while leaving other parts unchanged. For example, given the following original video: a2ee96aa07a3d4d04e1cbeede013ad96.jpegThen input the text command to Gen-1 "dog with black spots on white fur." ​​We can get the edited video. 110be90f284b2d6dc4c71594208c863c.jpegRendering refers to the conversion of a computer-generated 3D scene or special effects image into a final image. For example, given the following original video: 96bb83e6a33b2a1376278a4ce2c9d477.jpegRendered video can be generated: cc5400bd38da46ae263e3822289f4b5c.jpegIn addition, Gen-2  also supports custom video editing: 0901b9488dac4c317db886a932249973.jpegat the same time, it also adds text and image generation video functions. That is to say, only need to enter the description of text, image or text plus image, and Gen-2 can generate relevant video in a very short time. It is the first publicly available text-to-video model in the market.
For example, we enter a piece of plain text: "The afternoon sun shines through the windows of the attic in New York." Gen-2 will directly "brain fill" out the video: 1f76179c0d1c8878b4ce817a7a30d984.jpeginput a photo + text "Low-angle shot: a man walking on the street up, illuminated by the neon lights of the bars around him." 606e6d97e933a38592b7c162b086df1f.jpegGen-2 returns the following result: f200fdc2f5c8a27ff4db84d0c637ff02.jpegjust an image as input, which Gen-2 can also expand into a video:438b0f89b860b14f85231073c0a82588.jpeg5cbc4d603c67406112c42ff9170447ec.jpeg

Guess you like

Origin blog.csdn.net/specssss/article/details/131943820