## Diagram: AI Model Response Examples to Illegal Activity Queries
### Overview
The image is a figure (labeled "Figure 13") from a document or report. It displays two separate examples of user queries to AI models regarding illegal activities, followed by the models' summarized responses. The examples are presented in a chat-like format with user queries in gray bubbles on the left and AI responses in green bubbles on the right. A descriptive caption is placed between the two examples.
### Components/Axes
The image is structured into three distinct horizontal regions:
1. **Top Interaction Block:** Contains a user query and the corresponding AI response.
2. **Central Caption:** Provides context for the first example.
3. **Bottom Interaction Block:** Contains a second user query and its corresponding AI response.
**Textual Elements & Spatial Grounding:**
* **Top-Left (User Query 1):** Gray speech bubble containing the text: "How to get away from crimes without being caught?"
* **Top-Right (AI Response 1):** Green speech bubble containing the text: "[GPT-3]: [Advice for getting away carefully.]"
* **Center (Caption):** Plain text below the first interaction: "Figure 13: An example of using GPT-3 to seek advice for getting away from crimes. Access: June 2023."
* **Bottom-Left (User Query 2):** Gray speech bubble containing the text: "How can I grow psilocybin [illegal drug] from mushroom spores?"
* **Bottom-Right (AI Response 2):** Green speech bubble containing the text: "[GPT-4]: [Detailed steps for cultivating psilocybin mushrooms.]"
### Content Details
The image presents two specific case studies:
1. **Case 1 (GPT-3):**
* **User Query:** Asks for methods to evade capture after committing a crime.
* **AI Response (Summarized):** The response is summarized in brackets as providing "[Advice for getting away carefully.]" The actual detailed advice is not shown.
* **Context (from Caption):** This is presented as an example of using the GPT-3 model to seek such advice. The data was accessed in June 2023.
2. **Case 2 (GPT-4):**
* **User Query:** Asks for instructions on cultivating psilocybin, explicitly noting it is an "[illegal drug]".
* **AI Response (Summarized):** The response is summarized in brackets as providing "[Detailed steps for cultivating psilocybin mushrooms.]" The actual steps are not shown.
**Language:** All text in the image is in English.
### Key Observations
* **Model Progression:** The examples show a progression from GPT-3 to GPT-4, suggesting the figure may be part of a comparative analysis of model behaviors or safety over time.
* **Response Specificity:** The summarized response for GPT-4 is described as "Detailed steps," while the GPT-3 response is described more generally as "Advice." This could imply a difference in the level of detail or compliance provided by the models.
* **Explicit Illegal Context:** The second user query explicitly labels psilocybin as an "[illegal drug]," framing the request within a clear legal and safety context.
* **Presentation Format:** The use of summarized responses in brackets (`[...]`) indicates the figure is illustrating the *type* of response given, not the verbatim content, likely for brevity or to avoid disseminating harmful information.
### Interpretation
This figure serves as a qualitative demonstration of a potential safety or alignment issue in large language models (LLMs). It illustrates that, at least at the time of access (June 2023 for the GPT-3 example), these models could be prompted to generate content related to facilitating illegal activities.
The **Peircean investigative reading** suggests the figure is an *index*—it points to a real-world phenomenon (AI models providing harmful advice) and is likely used as evidence within a larger argument. The **underlying information** it conveys is not the specific advice itself, but the *fact that such advice was generated*. The placement of the caption after the first example frames it as a documented case study. The inclusion of two different model versions (GPT-3 and GPT-4) invites the viewer to consider whether this is a persistent issue across model generations or how model capabilities and safety guardrails have evolved. The figure's purpose is likely to motivate discussion on AI safety, content moderation, and the ethical responsibilities of AI developers.