View publication

Large language models (LLMs) have the potential to impact a wide range of creative domains, as exemplified in popular text-to-image generators like DALL·E and Midjourney. However, the application of LLMs to motion-based visual design has not yet been explored and presents novel challenges such as how users might effectively describe motion in natural language. Further, many existing generative design tools lack support for iterative refinement of designs beyond prompt engineering. In this paper, we present Keyframer, a design tool that leverages the code generation capabilities of LLMs to support design exploration for animations. Informed by interviews with professional motion designers, animators, and engineers, we designed Keyframer to support both ideation and refinement stages of animation design processes by enabling users to explore design variants throughout their process. We evaluated Keyframer with 13 users with a range of animation and programming experience, examining their prompting strategies and how they considered incorporating design variants into their process. We share a series of design principles for applying LLM to motion design prototyping tools and their potential implication for visual design tools more broadly.

Related readings and updates.

ToolSandbox: A Stateful, Conversational, Interactive Evaluation Benchmark for LLM Tool Use Capabilities

Recent large language models (LLMs) advancements sparked a growing research interest in tool assisted LLMs solving real-world challenges, which calls for comprehensive evaluation of tool-use capabilities. While previous works focused on either evaluating over stateless web services (RESTful API), based on a single turn user prompt, or an off-policy dialog trajectory, ToolSandbox includes stateful tool execution, implicit state dependencies…
See paper details

On the Role of Lip Articulation in Visual Speech Perception

*= Equal Contribution Generating realistic lip motion from audio to simulate speech production is critical for driving natural character animation. Previous research has shown that traditional metrics used to optimize and assess models for generating lip motion from speech are not a good indicator of subjective opinion of animation quality. Devising metrics that align with subjective opinion first requires understanding what impacts human…
See paper details