Jihye Lee
2025
Taxonomy and Analysis of Sensitive User Queries in Generative AI Search System
Hwiyeol Jo
|
Taiwoo Park
|
Hyunwoo Lee
|
Nayoung Choi
|
Changbong Kim
|
Ohjoon Kwon
|
Donghyeon Jeon
|
Eui-Hyeon Lee
|
Kyoungho Shin
|
Sun Suk Lim
|
Kyungmi Kim
|
Jihye Lee
|
Sun Kim
Findings of the Association for Computational Linguistics: NAACL 2025
Although there has been a growing interest among industries in integrating generative LLMs into their services, limited experience and scarcity of resources act as a barrier in launching and servicing large-scale LLM-based services. In this paper, we share our experiences in developing and operating generative AI models within a national-scale search engine, with a specific focus on the sensitiveness of user queries. We propose a taxonomy for sensitive search queries, outline our approaches, and present a comprehensive analysis report on sensitive queries from actual users. We believe that our experiences in launching generative AI search systems can contribute to reducing the barrier in building generative LLM-based services.
2023
Meta-Learning of Prompt Generation for Lightweight Prompt Engineering on Language-Model-as-a-Service
Hyeonmin Ha
|
Jihye Lee
|
Wookje Han
|
Byung-Gon Chun
Findings of the Association for Computational Linguistics: EMNLP 2023
Recently, many companies have been providing the capabilities of large language models as services. These Language-Model-as-a-Service (LMaaS) offerings support a variety of user tasks through in-context learning from prompts, which include instructions and demonstrations of the task. However, for users, manually crafting prompts or running automatic prompt tuning methods themselves can be demanding. Despite these challenges, LMaaS providers do not offer automatic prompt engineering methods as part of their services. One of the major obstacles to deploying them on an LMaaS is the heavy computational costs associated with automatic prompt engineering methods. These methods are typically designed to iterate through tens of thousands of examples, which impose unaffordable overheads for LMaaS providers. In this paper, we introduce MetaL-Prompt, a novel lightweight automatic prompt generation method for LMaaS. MetaL-Prompt meta-trains a prompt generation model (PGM) to enable robust learning by the language model from the contexts created by the generated prompts (i.e., in-context learning). Thanks to our meta-learning approach, a PGM can generate prompts for unseen tasks without requiring additional training for those specific tasks. Furthermore, the PGM can generate prompts with a single forward pass, significantly reducing computational costs compared to previous methods. We evaluate MetaL-Prompt on a range of unseen tasks and find that it improves performance by up to 19.4% in terms of mean F1 score on QA datasets compared to the state-of-the-art baseline P-tuning, with limited computational cost.
Search
Fix data
Co-authors
- Nayoung Choi 1
- Byung-Gon Chun 1
- Hyeonmin Ha 1
- Wookje Han 1
- Donghyeon Jeon 1
- show all...