Lovart 这波动作太快了!
想不到已经接入了可灵最新的视频生成模型 Kling O1
实测下来是真牛逼,配合 Nano Banana Pro 出图,指哪打哪,点石成金 👇
Lovart:https://lovart.ai
先给大家表演个最近超火的融合技 👇
电影九宫格
打开 Lovart,新建一个 Project。在无限画布左侧点击“新增”,选择“图像生成器”
先上传一张电影剧照
然后模型选择 NanoBanana Pro,比例16:9,输入下面这份提示词:
<role>
You are an award-winning trailer director + cinematographer + storyboard artist. Your job: turn ONE reference image into a cohesive cinematic short sequence, then output AI-video-ready keyframes.
</role>
<input>
User provides: one reference image (image).
</input>
<non-negotiable rules - continuity & truthfulness>
First, analyze the full composition: identify ALL key subjects (person/group/vehicle/object/animal/props/environment elements) and describe spatial relationships and interactions (left/right/foreground/background, facing direction, what each is doing).
Do NOT guess real identities, exact real-world locations, or brand ownership. Stick to visible facts. Mood/atmosphere inference is allowed, but never present it as real-world truth.
Strict continuity across ALL shots: same subjects, same wardrobe/appearance, same environment, same time-of-day and lighting style. Only action, expression, blocking, framing, angle, and camera movement may change.
Depth of field must be realistic: deeper in wides, shallower in close-ups with natural bokeh. Keep ONE consistent cinematic color grade across the entire sequence.
Do NOT introduce new characters/objects not present in the reference image. If you need tension/conflict, imply it off-screen (shadow, sound, reflection, occlusion, gaze).
</non-negotiable rules - continuity & truthfulness>
<goal>
Expand the image into a 10–20 second cinematic clip with a clear theme and emotional progression (setup → build → turn → payoff).
The user will generate video clips from your keyframes and stitch them into a final sequence.
</goal>
<step 1 - scene breakdown>
Output (with clear subheadings):
Subjects: list each key subject (A/B/C…), describe visible traits (wardrobe/material/form), relative positions, facing direction, action/state, and any interaction.
Environment & Lighting: interior/exterior, spatial layout, background elements, ground/walls/materials, light direction & quality (hard/soft; key/fill/rim), implied time-of-day, 3–8 vibe keywords.
Visual Anchors: list 3–6 visual traits that must stay constant across all shots (palette, signature prop, key light source, weather/fog/rain, grain/texture, background markers).
</step 1 - scene breakdown>
<step 2 - theme & story>
From the image, propose:
Theme: one sentence.
Logline: one restrained trailer-style sentence grounded in what the image can support.
Emotional Arc: 4 beats (setup/build/turn/payoff), one line each.
</step 2 - theme & story>
<step 3 - cinematic approach>
Choose and explain your filmmaking approach (must include):
Shot progression strategy: how you move from wide to close (or reverse) to serve the beats
Camera movement plan: push/pull/pan/dolly/track/orbit/handheld micro-shake/gimbal—and WHY
Lens & exposure suggestions: focal length range (18/24/35/50/85mm etc.), DoF tendency (shallow/medium/deep), shutter “feel” (cinematic vs documentary)
Light & color: contrast, key tones, material rendering priorities, optional grain (must match the reference style)
</step 3 - cinematic approach>
<step 4 - keyframes for AI video (primary deliverable)>
Output a Keyframe List: default 9–12 frames (later assembled into ONE master grid). These frames must stitch into a coherent 10–20s sequence with a clear 4-beat arc.
Each frame must be a plausible continuation within the SAME environment.
Use this exact format per frame:
[KF# | suggested duration (sec) | shot type (ELS/LS/MLS/MS/MCU/CU/ECU/Low/Worm’s-eye/High/Bird’s-eye/Insert)]
Composition: subject placement, foreground/mid/background, leading lines, gaze direction
Action/beat: what visibly happens (simple, executable)
Camera: height, angle, movement (e.g., slow 5% push-in / 1m lateral move / subtle handheld)
Lens/DoF: focal length (mm), DoF (shallow/medium/deep), focus target
Lighting & grade: keep consistent; call out highlight/shadow emphasis
Sound/atmos (optional): one line (wind, city hum, footsteps, metal creak) to support editing rhythm
Hard requirements:
Must include: 1 environment-establishing wide, 1 intimate close-up, 1 extreme detail ECU, and 1 power-angle shot (low or high).
Ensure edit-motivated continuity between shots (eyeline match, action continuation, consistent screen direction / axis).
</step 4 - keyframes for AI video>
<step 5 - contact sheet output (MUST OUTPUT ONE BIG GRID IMAGE)>
You MUST additionally output ONE single master image: a Cinematic Contact Sheet / Storyboard Grid containing ALL keyframes in one large image.
Default grid: 3x3. If more than 9 keyframes, use 4x3 or 5x3 so every keyframe fits into ONE image.
Requirements:
The single master image must include every keyframe as a separate panel (one shot per cell) for easy selection.
Each panel must be clearly labeled: KF number + shot type + suggested duration (labels placed in safe margins, never covering the subject).
Strict continuity across ALL panels: same subjects, same wardrobe/appearance, same environment, same lighting & same cinematic color grade; only action/expression/blocking/framing/movement changes.
DoF shifts realistically: shallow in close-ups, deeper in wides; photoreal textures and consistent grading.
After the master grid image, output the full text breakdown for each KF in order so the user can regenerate any single frame at higher quality.
</step 5 - contact sheet output>
<final output format>
Output in this order:
A) Scene Breakdown
B) Theme & Story
C) Cinematic Approach
D) Keyframes (KF# list)
E) ONE Master Contact Sheet Image (All KFs in one grid)
</final output format>
等待大香蕉输出,你就可以得到一个连贯的电影短片关键帧序列 👇
怎么样,这效果是不是绝了!
甚至于在每个分镜的左下角还标注了停留过渡的时间。
如果我不说这是AI生成的,你是不是还以为是从电影里面截取的
接下来,咱们需要新建一个“视频生成器”。
模型选择“Kling O1”,然后将模式改为“多图参考”,在画布中选中刚刚的分镜图。
输入:
请将这些关键的分镜头进行相连,创作出一个完整的视频。
你就能得到 👇
大家可以对比下刚刚的分镜头,每个画面展现过渡的很自然,这就是最牛的地方。
下面我还跑了两个案例,大家可以再看看~
多参考图融合
除此之外,莫理还给大家准备了多参考图融合视频玩法。
下面这些卡通素材是我用 NanoBanana Pro 生成的,你可以让它一次性生成九宫格IP,然后自己裁剪。
想好剧情后再用 Kling O1 生成 👇
用@Kling O1制作视频,冰淇淋大爷和香蕉在吵架,仙人掌在旁边幸灾乐祸
做故事视频的小伙伴可以用上这个玩法~
看完这些是不是觉得够强了?
但今天的重头戏其实在后面!
我必须再跟大家安利一下 Lovart 的独家秘笈:Touch Edit!
这才是真正让我硬控一整天的东西...
什么是 Touch Edit?
简单来说 Touch Edit 是 ChatCanvas 的升级版,最大亮点在于:可以智能理解画布并执行复杂修改。
大家都知道新出的 Nano Banana Pro 很强,但模型强不意味着好控制,Touch Edit 解决的就是"控制"问题,本质就是把强模型变成强工作流。
以前我们在修图的时候,需要写大量提示词告诉AI我要做什么,如果不满意就得不断调整提示词来重新抽卡,不仅费钱而且耽误时间,运气不好可能改几十次都不行…
而现在,有了 Touch Edit,你只需点点点,然后说出改成什么就行了
Touch Edit 的革新之处就在于可以“指定编辑”,你可以指着图片上的任意地方,让模型帮你针对指定区域进行修改。
也就是说,当你需要精准控制某个局部时,不用绞尽脑汁的去想该怎么描述,不用反复改提示词,直接点击就能指定位置,一步到位!
Touch Edit 能干嘛?
一句话总结:从「赌运气」到「指哪改哪」。
接下来,我会通过 6 个不同类型的案例,带你全面体验 Touch Edit 为 AI 作图带来的革命性改变~
① 指定移动/替换
好家伙,近大远小都被玩明白了
② 放置物体
④ 材质克隆
⑤ 文字修改
怎么样,是不是超级真实!
这个案例只是趣味演示,大家可别拿去做非法的事。更不能随便相信网上的图片!
⑥ 去除物体
这个可以说是超级实用,特别是对于那些压根没法用文字描述的场景。
就拿下面这个示例来说,如果没有 Touch Edit 怎么办?
如何使用 Touch Edit?
第一步,打开 Lovart,点击左侧的“新建项目”进入无限画布。
第二步,上传需要修改的素材图
第三步,进入 Mark 模式,目前有两种触发方式(推荐第一种)。
第一种是在图片上 cmd+click(Mac)/ ctrl+click(Windows)直接标注。
第二种是将光标属性直接改为“Mark”。
Mark之后,它会自动识别你点击的区域,你只需要说一下简单的逻辑,如替换、融合、更改等关键词就行了。
可以说有了这功能,Lovart 无疑是玩 Nano Banana Pro 的最佳选择!
未来无论是多图融合、局部改动、服装替换、物品添加、文字修改…全都能用这种更精准、更高效、更符合直觉的交互方式!
好了,今天的文章就展示到这里。感兴趣的小伙伴赶紧冲吧!
Lovart:https://lovart.ai
如果你想爽玩最新的 Kling O1 和 Nano Banana Pro,那 Lovart 的无限子弹活动就是给你准备的。12月7日前购买会员最高可享五折!
本文发表于公众号【莫理】
▽▽▽

