Limits and Next
Updated: 2026-05
1. About This Page
This page summarizes the current limitations of video-generating AI, including Runway, as of May 2026, and predicts what the landscape will look like in six months and a year. Since this is a field where educational materials quickly become outdated, this page serves as a guide to help you understand the current state of affairs and know where to go from here on your own.
2. Limitations as of May 2026
2.1 Video Length
- 1 shot: 5–10 seconds, up to about 20 seconds
- Even with top-tier commercial AI, it’s difficult to produce a single shot lasting over 1 minute
- “Long single takes” are not realistic with AI-generated content
→ Editing multiple shots to increase the total runtime is the practical solution. This is where Runway Editor really shines.
2.2 Consistency
- Character Faces: They can be maintained for a few shots, but start to look off after more than 10 shots.
- Environmental Consistency: It’s difficult to recreate the same room or street corner in different shots.
- Depiction of the Passage of Time: If you cover morning, noon, and evening in a single work, it breaks the sense of unity.
→ While it can be improved using Reference, Act-Two, and Aleph, it does not offer the same level of stability as live-action or CG.
2.3 Physical Representation
- Fluids (water, smoke): Performs quite well, but breaks down during intense movement
- Finger movements: Breaks down when all five fingers move independently
- Physical contact: Low accuracy for handshakes, kicking a ball, passing objects, etc.
- Crowds: Faces break down when five or more people move independently
→ Use composition to work around your weak points. Change the scenario so that your hands aren’t shown or you don’t shake hands.
2.4 Characters
- Text on signs, books, and screens often appears as unrecognizable symbols
- Japanese text is more prone to distortion than English
- If you want to display text, you need to composite it later
2.5 Precise Synchronization of Audio and Video
- While the lip-syncing is good, whether the rhythm matches the overall performance depends on Act-Two’s skills
- The choreography that synchronizes movements with key moments in the BGM is done manually (adjusted during editing)
2.6 Limits of Control
- Camera movement has become more precise in Director Mode, but it lacks pixel-level precision
- Area-specific operations, such as “I want only the right side of the screen to be red,” are partially possible using Aleph or Motion Brush
3. Limitations of Plans and Costs
- Free 125 cr (one-time offer): Enough for a trial, but not enough for production
- Standard: $12–15 for 625 cr/month—a good fit for personal prototyping
- Pro or higher is required for full-scale production, starting at $28/month
- As of May 2026, no official student discount is available
→ In university and vocational school classes, sharing plans within groups is the practical solution.
4. Comparison of the Highest Commercial Frequency Bands (as of May 2026)
| Model | Strengths | Weaknesses |
|---|---|---|
| Runway Gen-4.5 + Aleph | Integrated editing, Director Mode, Act-Two | Weaker physics than Sora 2 |
| Sora 2 | Physics, consistency in long-form content | Weak editing workflow, paid subscription model |
| Veo 3.1 | Realism, color rendering | Limited customization |
| Kling 3.0 Pro | Low cost, high quality | UI is primarily in Chinese; English version has limited features |
| Hailuo 02 | Responsiveness, generation speed | Moderate consistency |
Rather than “doing everything with a single model,” the mainstream approach in 2026 will be to use multiple models in combination. Runway is also beginning to serve as an “aggregator,” capable of calling third-party models (such as Veo and Kling) from within its platform.
5. Things That Are Likely to Change in Six Months or a Year
Forecast (as of May 2026):
5.1 Resolution and Duration
- Shots lasting 30 seconds or longer will become the norm
- Native 4K and 8K production
5.2 Control Precision
- “Pixel-level region control” has become widespread
- Real-time preview (adjust parameters while rendering)
5.3 Real-Time Generation
- A 5-second clip is generated in 5 seconds
- AI processing during live streams
5.4 The Rise of Open Source
- Further advancements in the Wan and HunyuanVideo families
- Quality approaching that of top-tier commercial software is now achievable even on personal computers
5.5 Legal Regulations
- Regulations regarding copyright, image rights, and deepfakes are advancing in various countries
- Proof of model origin may be required for commercial use
These are predictions, not certainties. You’ll need to keep track of developments on your own even after class ends.
6. If you plan to continue outside of school
If you want to continue exploring AI video generation after class:
- Runway Academy — Extensive official tutorials, all available for free
- Runway Changelog — Update history for new features
- Artificial Analysis — Model comparison benchmarks; check monthly
- Reddit r/aivideo — Community and the latest case studies
- Numerous AI channels on YouTube (channels with different names will likely emerge in six months)
There is vastly more information available in English. English search skills are essential in the age of AI.
7. Take a look at open-source video AI as well
If you only use commercial tools, you’ll be left with nothing when the service is discontinued.
It’s also a good idea to try out open-source video AI tools like Comfy Cloud:
- Wan 2.2 / 2.5 — A leading open-source video generation tool
- HunyuanVideo — Developed by Tencent, open-source
- LTX-Video 2.3 — Real-time performance is its key strength
Lightweight versions of these are available that can run on personal computers. They can be used for free for research and experimental purposes.
8. Relationship with Media Art
The overall theme of this course is “Media Art Expression.” Runway is one of the tools used for this:
- Used for project proposals for media art works (video prototyping)
- Creating simulation videos for installation art
- Creating background videos for live performances
Runway doesn’t create a media art piece on its own. The venue, the installation, and the audience’s experience—all of these are part of the work. Runway is a tool for quickly bringing those elements to life.
9. What Comes Next After This Course
During class:
- Practical Exercises: Mastering Runway
- Hands-On Sessions: Creating short films in groups (approximately 5 sessions)
- Final Presentation: Screening
Here’s how to use Runway if you plan to work as a media artist:
- Attach a short prototype to the proposal to make it more persuasive
- Create concept art for scenes that cannot be filmed live-action (e.g., futuristic cities, sci-fi settings)
- Position yourself as a writer experimenting with “works created using AI”
Whether or not to call oneself an “AI writer” is a personal choice. The stance of simply using it as a tool is also a valid option.
10. In Conclusion
This is a field where technology advances rapidly, and the landscape can change in just six months. Fundamentals that remain relevant even as tools evolve are what hold the most value. The fundamentals we want you to master in this course are:
- The ability to articulate concepts (whether through prompts or briefs)
- The ability to think in terms of shots (camera position, movement, and transitions)
- The judgment to identify and avoid limitations
- The mindset to maintain your intent without being swayed by the tools
The runway is a training ground, not the end goal.
From here on out, it’s a journey for each of us to discover how to incorporate AI into our own creative work.
11. Navigation
- Runway Overview — What Is Runway?
- Video Prototyping Mindset — The Approach to Prototyping
- Shot Planning — Shot Design
