Skip to content
Limits and Next

Limits and Next

Updated: 2026-05

1. About This Page

This page summarizes the current limitations of video-generating AI, including Runway, as of May 2026, and predicts what the landscape will look like in six months and a year. Since this is a field where educational materials quickly become outdated, this page serves as a guide to help you understand the current state of affairs and know where to go from here on your own.

2. Limitations as of May 2026

2.1 Video Length

  • 1 shot: 5–10 seconds, up to about 20 seconds
  • Even with top-tier commercial AI, it’s difficult to produce a single shot lasting over 1 minute
  • Long single takes” are not realistic with AI-generated content

Editing multiple shots to increase the total runtime is the practical solution. This is where Runway Editor really shines.

2.2 Consistency

  • Character Faces: They can be maintained for a few shots, but start to look off after more than 10 shots.
  • Environmental Consistency: It’s difficult to recreate the same room or street corner in different shots.
  • Depiction of the Passage of Time: If you cover morning, noon, and evening in a single work, it breaks the sense of unity.

→ While it can be improved using Reference, Act-Two, and Aleph, it does not offer the same level of stability as live-action or CG.

2.3 Physical Representation

  • Fluids (water, smoke): Performs quite well, but breaks down during intense movement
  • Finger movements: Breaks down when all five fingers move independently
  • Physical contact: Low accuracy for handshakes, kicking a ball, passing objects, etc.
  • Crowds: Faces break down when five or more people move independently

Use composition to work around your weak points. Change the scenario so that your hands aren’t shown or you don’t shake hands.

2.4 Characters

  • Text on signs, books, and screens often appears as unrecognizable symbols
  • Japanese text is more prone to distortion than English
  • If you want to display text, you need to composite it later

2.5 Precise Synchronization of Audio and Video

  • While the lip-syncing is good, whether the rhythm matches the overall performance depends on Act-Two’s skills
  • The choreography that synchronizes movements with key moments in the BGM is done manually (adjusted during editing)

2.6 Limits of Control

  • Camera movement has become more precise in Director Mode, but it lacks pixel-level precision
  • Area-specific operations, such as “I want only the right side of the screen to be red,” are partially possible using Aleph or Motion Brush

3. Limitations of Plans and Costs

  • Free 125 cr (one-time offer): Enough for a trial, but not enough for production
  • Standard: $12–15 for 625 cr/month—a good fit for personal prototyping
  • Pro or higher is required for full-scale production, starting at $28/month
  • As of May 2026, no official student discount is available

→ In university and vocational school classes, sharing plans within groups is the practical solution.

4. Comparison of the Highest Commercial Frequency Bands (as of May 2026)

Model Strengths Weaknesses
Runway Gen-4.5 + Aleph Integrated editing, Director Mode, Act-Two Weaker physics than Sora 2
Sora 2 Physics, consistency in long-form content Weak editing workflow, paid subscription model
Veo 3.1 Realism, color rendering Limited customization
Kling 3.0 Pro Low cost, high quality UI is primarily in Chinese; English version has limited features
Hailuo 02 Responsiveness, generation speed Moderate consistency

Rather than “doing everything with a single model,” the mainstream approach in 2026 will be to use multiple models in combination. Runway is also beginning to serve as an “aggregator,” capable of calling third-party models (such as Veo and Kling) from within its platform.

5. Things That Are Likely to Change in Six Months or a Year

Forecast (as of May 2026):

5.1 Resolution and Duration

  • Shots lasting 30 seconds or longer will become the norm
  • Native 4K and 8K production

5.2 Control Precision

  • “Pixel-level region control” has become widespread
  • Real-time preview (adjust parameters while rendering)

5.3 Real-Time Generation

  • A 5-second clip is generated in 5 seconds
  • AI processing during live streams

5.4 The Rise of Open Source

  • Further advancements in the Wan and HunyuanVideo families
  • Quality approaching that of top-tier commercial software is now achievable even on personal computers

5.5 Legal Regulations

  • Regulations regarding copyright, image rights, and deepfakes are advancing in various countries
  • Proof of model origin may be required for commercial use

These are predictions, not certainties. You’ll need to keep track of developments on your own even after class ends.

6. If you plan to continue outside of school

If you want to continue exploring AI video generation after class:

  • Runway Academy — Extensive official tutorials, all available for free
  • Runway Changelog — Update history for new features
  • Artificial Analysis — Model comparison benchmarks; check monthly
  • Reddit r/aivideo — Community and the latest case studies
  • Numerous AI channels on YouTube (channels with different names will likely emerge in six months)

There is vastly more information available in English. English search skills are essential in the age of AI.

7. Take a look at open-source video AI as well

If you only use commercial tools, you’ll be left with nothing when the service is discontinued.

It’s also a good idea to try out open-source video AI tools like Comfy Cloud:

  • Wan 2.2 / 2.5 — A leading open-source video generation tool
  • HunyuanVideo — Developed by Tencent, open-source
  • LTX-Video 2.3 — Real-time performance is its key strength

Lightweight versions of these are available that can run on personal computers. They can be used for free for research and experimental purposes.

8. Relationship with Media Art

The overall theme of this course is “Media Art Expression.” Runway is one of the tools used for this:

  • Used for project proposals for media art works (video prototyping)
  • Creating simulation videos for installation art
  • Creating background videos for live performances

Runway doesn’t create a media art piece on its own. The venue, the installation, and the audience’s experience—all of these are part of the work. Runway is a tool for quickly bringing those elements to life.

9. What Comes Next After This Course

During class:

  • Practical Exercises: Mastering Runway
  • Hands-On Sessions: Creating short films in groups (approximately 5 sessions)
  • Final Presentation: Screening

Here’s how to use Runway if you plan to work as a media artist:

  • Attach a short prototype to the proposal to make it more persuasive
  • Create concept art for scenes that cannot be filmed live-action (e.g., futuristic cities, sci-fi settings)
  • Position yourself as a writer experimenting with “works created using AI”

Whether or not to call oneself an “AI writer” is a personal choice. The stance of simply using it as a tool is also a valid option.

10. In Conclusion

This is a field where technology advances rapidly, and the landscape can change in just six months. Fundamentals that remain relevant even as tools evolve are what hold the most value. The fundamentals we want you to master in this course are:

  • The ability to articulate concepts (whether through prompts or briefs)
  • The ability to think in terms of shots (camera position, movement, and transitions)
  • The judgment to identify and avoid limitations
  • The mindset to maintain your intent without being swayed by the tools

The runway is a training ground, not the end goal.

From here on out, it’s a journey for each of us to discover how to incorporate AI into our own creative work.

11. Navigation

  • Runway Overview — What Is Runway?
  • Video Prototyping Mindset — The Approach to Prototyping
  • Shot Planning — Shot Design