Papers
Topics
Authors
Recent
Assistant
AI Research Assistant
Well-researched responses based on relevant abstracts and paper content.
Custom Instructions Pro
Preferences or requirements that you'd like Emergent Mind to consider when generating responses.
Gemini 2.5 Flash
Gemini 2.5 Flash 134 tok/s
Gemini 2.5 Pro 41 tok/s Pro
GPT-5 Medium 26 tok/s Pro
GPT-5 High 35 tok/s Pro
GPT-4o 99 tok/s Pro
Kimi K2 192 tok/s Pro
GPT OSS 120B 440 tok/s Pro
Claude Sonnet 4.5 37 tok/s Pro
2000 character limit reached

Prompt Middleware: Mapping Prompts for Large Language Models to UI Affordances (2307.01142v1)

Published 3 Jul 2023 in cs.HC

Abstract: To help users do complex work, researchers have developed techniques to integrate AI and human intelligence into user interfaces (UIs). With the recent introduction of LLMs, which can generate text in response to a natural language prompt, there are new opportunities to consider how to integrate LLMs into UIs. We present Prompt Middleware, a framework for generating prompts for LLMs based on UI affordances. These include prompts that are predefined by experts (static prompts), generated from templates with fill-in options in the UI (template-based prompts), or created from scratch (free-form prompts). We demonstrate this framework with FeedbackBuffet, a writing assistant that automatically generates feedback based on a user's text input. Inspired by prior research showing how templates can help non-experts perform more like experts, FeedbackBuffet leverages template-based prompt middleware to enable feedback seekers to specify the types of feedback they want to receive as options in a UI. These options are composed using a template to form a feedback request prompt to GPT-3. We conclude with a discussion about how Prompt Middleware can help developers integrate LLMs into UIs.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (28)
  1. The value of feedback in improving collaborative writing assignments in an online learning environment. Studies in Higher Education 37, 4 (2012), 387–400.
  2. Soylent: a word processor with a crowd inside. In Proceedings of the 23nd annual ACM symposium on User interface software and technology. 313–322.
  3. Language models are few-shot learners. Advances in Neural Information Processing Systems 33 (2020), 1877–1901.
  4. Wizard of Oz studies—why and how. Knowledge-based systems 6, 4 (1993), 258–266.
  5. The Idea Machine: LLM-Based Expansion, Rewriting, Combination, and Suggestion of Ideas. In Proceedings of the 14th Conference on Creativity and Cognition (Venice, Italy) (C&C ’22). Association for Computing Machinery, New York, NY, USA, 623–627. https://doi.org/10.1145/3527927.3535197
  6. Fluid Transformers and Creative Analogies: Exploring Large Language Models’ Capacity for Augmenting Cross-Domain Analogical Creativity. In Proceedings of the 15th Conference on Creativity and Cognition (Virtual Event, USA) (C&C ’23). Association for Computing Machinery, New York, NY, USA, 489–505. https://doi.org/10.1145/3591196.3593516
  7. Dennis M Docheff. 1990. The feedback sandwich. Journal of Physical Education, Recreation & Dance 61, 9 (1990), 17–18.
  8. Atul Gawande. 2009. The Checklist Manifesto: How to Get Things Right. Metropolitan Books.
  9. CausalMapper: Challenging Designers to Think in Systems with Causal Maps and Large Language Model. In Proceedings of the 15th Conference on Creativity and Cognition (Virtual Event, USA) (C&C ’23). Association for Computing Machinery, New York, NY, USA, 325–329. https://doi.org/10.1145/3591196.3596818
  10. Introassist: A tool to support writing introductory help requests. In Proceedings of the 2018 CHI Conference on Human Factors in Computing Systems. 1–13.
  11. PromptMaker: Prompt-based Prototyping with Large Language Models. In CHI Conference on Human Factors in Computing Systems Extended Abstracts. 1–8.
  12. Motif: Supporting Novice Creativity through Expert Patterns. In Proceedings of the 33rd Annual ACM Conference on Human Factors in Computing Systems (Seoul, Republic of Korea) (CHI ’15). Association for Computing Machinery, New York, NY, USA, 1211–1220. https://doi.org/10.1145/2702123.2702507
  13. CrowdWeaver: visually managing complex crowd work. In Proceedings of the ACM 2012 Conference on Computer Supported Cooperative Work. 1033–1036.
  14. Critique Style Guide: Improving Crowdsourced Design Feedback with a Natural Language Model. In Proceedings of the 2017 CHI Conference on Human Factors in Computing Systems (Denver, Colorado, USA) (CHI ’17). Association for Computing Machinery, New York, NY, USA, 4627–4639. https://doi.org/10.1145/3025453.3025883
  15. Collaboratively crowdsourcing workflows with turkomatic. In Proceedings of the acm 2012 conference on computer supported cooperative work. 1003–1012.
  16. Ask Me or Tell Me? Enhancing the Effectiveness of Crowdsourced Design Feedback. In Proceedings of the 2021 CHI Conference on Human Factors in Computing Systems (Yokohama, Japan) (CHI ’21). Association for Computing Machinery, New York, NY, USA, Article 564, 12 pages. https://doi.org/10.1145/3411764.3445507
  17. Framing Creative Work: Helping Novices Frame Better Problems through Interactive Scaffolding. In Creativity and Cognition (Virtual Event, Italy) (C&C ’21). Association for Computing Machinery, New York, NY, USA, Article 30, 10 pages. https://doi.org/10.1145/3450741.3465261
  18. Combining Freeform Curation with Structured Templates. In Creativity and Cognition (Gathertown) (C&C ’19). ACM, New York, NY, USA, 11 pages. https://doi.org/10.1145/3591196.3593337
  19. Prototyping an Intelligent Agent through Wizard of Oz. In Proceedings of the INTERACT ’93 and CHI ’93 Conference on Human Factors in Computing Systems (Amsterdam, The Netherlands) (CHI ’93). Association for Computing Machinery, New York, NY, USA, 277–284. https://doi.org/10.1145/169059.169215
  20. Co-Writing Screenplays and Theatre Scripts with Language Models: Evaluation by Industry Professionals. In Proceedings of the 2023 CHI Conference on Human Factors in Computing Systems. 1–34.
  21. Docent: transforming personal intuitions to scientific hypotheses through content learning and process training. In Proceedings of the Fifth Annual ACM Conference on Learning at Scale. 1–10.
  22. Laria Reynolds and Kyle McDonell. 2021. Prompt Programming for Large Language Models: Beyond the Few-Shot Paradigm. In Extended Abstracts of the 2021 CHI Conference on Human Factors in Computing Systems (Yokohama, Japan) (CHI EA ’21). Association for Computing Machinery, New York, NY, USA, Article 314, 7 pages. https://doi.org/10.1145/3411763.3451760
  23. What is ”Intelligent” in Intelligent User Interfaces? A Meta-Analysis of 25 Years of IUI. In Proceedings of the 25th International Conference on Intelligent User Interfaces (Cagliari, Italy) (IUI ’20). Association for Computing Machinery, New York, NY, USA, 477–487. https://doi.org/10.1145/3377325.3377500
  24. AI Chains: Transparent and Controllable Human-AI Interaction by Chaining Large Language Model Prompts. In Proceedings of the 2022 CHI Conference on Human Factors in Computing Systems (New Orleans, LA, USA) (CHI ’22). Association for Computing Machinery, New York, NY, USA, Article 385, 22 pages. https://doi.org/10.1145/3491102.3517582
  25. Re-Examining Whether, Why, and How Human-AI Interaction Is Uniquely Difficult to Design. Association for Computing Machinery, New York, NY, USA, 1–13. https://doi.org/10.1145/3313831.3376301
  26. Robert K Yin et al. 2018. Case study research and applications: Design and methods. Los Angeles, UK: Sage (2018).
  27. Wordcraft: story writing with large language models. In 27th International Conference on Intelligent User Interfaces. 841–852.
  28. Almost an Expert: The Effects of Rubrics and Expertise on Perceived Value of Crowdsourced Design Critiques. In Proceedings of the 19th ACM Conference on Computer-Supported Cooperative Work & Social Computing (San Francisco, California, USA) (CSCW ’16). Association for Computing Machinery, New York, NY, USA, 1005–1017. https://doi.org/10.1145/2818048.2819953
Citations (13)

Summary

  • The paper introduces Prompt Middleware, a framework that maps LLM prompts to UI affordances using static, template-based, and free-form methods.
  • It details the FeedbackBuffet implementation, where template-based prompts combine expert options to generate effective GPT-3 feedback.
  • The study highlights improvements in user interaction and sets the stage for future research on middleware efficacy in integrating AI with UIs.

Prompt Middleware: Mapping Prompts for LLMs to UI Affordances

Introduction

The introduction of LLMs such as GPT-3 has opened new avenues for integrating AI within User Interfaces (UIs). The paper "Prompt Middleware: Mapping Prompts for LLMs to UI Affordances" presents a framework to efficiently leverage LLMs through three distinct methodologies: static prompts, template-based prompts, and free-form prompts. Each methodology offers varying degrees of control over prompt generation, facilitating diverse applications across different user expertise levels.

Dynamic Integration of LLMs in UIs

Prompt Middleware serves as a bridge between UI affordances and LLM interfacing, functioning to simplify prompt creation while infusing domain-specific knowledge into the prompts. Static prompts, being predefined and optimized by experts, facilitate seamless integration by abstracting the complexities in prompt formulation for end-users. Figure 1

Figure 1: FeedbackBuffet enables users to insert writing samples (1) and select from a set of predefined options for the type of feedback they want to receive (2). Using a template, these options are combined to form a prompt (3) which is sent to GPT-3 using OpenAI's public API (4). GPT-3 then generates feedback, which is displayed in a text box for the user to review (5).

Implementation of FeedbackBuffet

FeedbackBuffet exemplifies the effectiveness of template-based prompts. Implemented as a ReactJS application, it provides users the ability to generate high-quality prompts through an interface that merges best practices in feedback design with interactive functionalities. This system allows users to select options which are then combined into prompts for GPT-3, offering flexibility without necessitating deep AI knowledge. Figure 2

Figure 2: Examples of feedback that GPT-3 can provide by combining different options within our feedback template. These results can be compared across two example contexts—an email and a short statement of purpose.

Evaluation and Future Directions

The implementation of FeedbackBuffet highlights potential in simplifying LLM utilization within UIs through middleware frameworks. The integration transcends mere prompt generation to include best practices and expert-derived templates that significantly streamline the user experience. Upcoming research can focus on evaluating user interaction with these middleware types to quantify utility, control trade-offs, and further refine UI integration.

Conclusion

The paper outlines a significant advance in LLM interfacing through Prompt Middleware—a framework ensuring effective LLM utilization within user-friendly interfaces. By abstracting prompt complexities and encapsulating domain expertise, this approach promises enhanced LLM accessibility for diverse user bases, signifying a vital step towards bridging the gap between AI capabilities and user needs.

Dice Question Streamline Icon: https://streamlinehq.com

Open Problems

We haven't generated a list of open problems mentioned in this paper yet.

List To Do Tasks Checklist Streamline Icon: https://streamlinehq.com

Collections

Sign up for free to add this paper to one or more collections.