3 Models Available
A concept artist generates a clean interior reference on Nano Banana 2, then turns it into a navigable scene where the camera can orbit and capture matched-angle stills. On Martini's canvas, drop the reference image into a world node (or chain Flux as an alt-look reference), capture 5-10 angles, and feed each as a starting frame into Sora 2 video nodes for shots that all share the same world. Note: Martini does not export navigable worlds as glTF or USD — captured stills are the deliverable. Pick a model below to walk through the image-to-world workflow.
Generate the canonical reference image for an Image-to-3D-World workflow on Martini using Nano Banana 2 — the cleaner the source, the navigable the resulting scene. The output of the world node is a navigable canvas-internal scene preview you can orbit and screenshot, not a portable .obj, .fbx, .glb, or USD mesh file. Concept artists use this to lock a location once on Nano Banana 2, pass the locked still into the World Labs or Image-to-3D-World node, and capture matched-angle stills that feed downstream Sora 2 or Kling 3 nodes for shots that all share the same world.
Black Forest Labs
Generate the source reference image for an Image-to-3D-World workflow on Martini using FLUX.2 — its prompt-fidelity rendering produces clean, literal scene compositions that the world node can reconstruct cleanly. The world node's output is a navigable canvas-internal scene preview you can orbit and screenshot, not a portable .obj, .fbx, .glb, or USD mesh file. Concept artists use FLUX.2 when they need an alt-look reference (different palette, different lighting, different style) than what Nano Banana 2 produces — same workflow, different aesthetic.
OpenAI
Use Sora 2 as the downstream camera-move engine for an Image-to-3D-World workflow on Martini — the captured stills from the navigable world feed directly into Sora 2 video nodes for matched-angle motion shots. The world node's output is a canvas-internal navigable scene preview, not a portable .obj, .fbx, .glb, or USD mesh. Sora 2 takes the captured stills as starting frames and produces video clips that all share the same locked location, with cinematographic camera moves that respect the spatial structure of the source world.