the core interaction is dialogue with latent space via tangible controls. initial scene generation, then direct dive into hybrid 3d/nodal interface. selecting an entity highlights its node. geometry: draw splines for ridges, force brushes for terrain deformation. textures: connect procedural nodes, image samplers, ai style transfer nodes via graph; parameters are dials, color pickers. physics properties are visualized as force vectors or zones in 3d, adjusted with gizmos. dynamic events are timeline sequences; event nodes have conditional inputs, probabilistic outputs like sliders for 'meteor_impact_radius'. feedback is tight; tweak a parameter, see local effect instantly. ai refines details contextually, offers suggestions. select a dinosaur's 'locomotion' node, tweak stride length, or feed new reference like 'walk like a heron'. it is always rendering, always live. scripting layer per node offers raw parameter access for custom behaviors. parameter changes propagate.