A First Look at GPT Apps: Landscape and Vulnerability
CoRR(2024)
摘要
With the advancement of Large Language Models (LLMs), increasingly
sophisticated and powerful GPTs are entering the market. Despite their
popularity, the LLM ecosystem still remains unexplored. Additionally, LLMs'
susceptibility to attacks raises concerns over safety and plagiarism. Thus, in
this work, we conduct a pioneering exploration of GPT stores, aiming to study
vulnerabilities and plagiarism within GPT applications. To begin with, we
conduct, to our knowledge, the first large-scale monitoring and analysis of two
stores, an unofficial GPTStore.AI, and an official OpenAI GPT Store. Then, we
propose a TriLevel GPT Reversing (T-GR) strategy for extracting GPT internals.
To complete these two tasks efficiently, we develop two automated tools: one
for web scraping and another designed for programmatically interacting with
GPTs. Our findings reveal a significant enthusiasm among users and developers
for GPT interaction and creation, as evidenced by the rapid increase in GPTs
and their creators. However, we also uncover a widespread failure to protect
GPT internals, with nearly 90
considerable plagiarism and duplication among GPTs.
更多查看译文
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要