# Un-considering Contextual Information: Assessing LLMsâ Understanding of Indexical Elements
Abstract
Large Language Models (LLMs) have demonstrated impressive performances in tasks related to coreference resolution. However, previous studies mostly assessed LLM performance on coreference resolution with nouns and third person pronouns. This study evaluates LLM performance on coreference resolution with indexicals like I, you, here and tomorrow, which come with unique challenges due to their linguistic properties. We present the first study examining how LLMs interpret indexicals in English, releasing the English Indexical Dataset with 1600 multiple-choice questions. We evaluate pioneering LLMs, including GPT-4o, Claude 3.5 Sonnet, Gemini 1.5 Pro, and DeepSeek V3. Our results reveal that LLMs exhibit an impressive performance with some indexicals (I), while struggling with others (you, here, tomorrow), and that syntactic cues (e.g. quotation) contribute to LLM performance with some indexicals, while they reduce performance with others. Code and data are available at: https://github.com/metehanoguzz/LLMs-Indexicals-English
Un-considering Contextual Information: Assessing LLMsâ Understanding of Indexical Elements
Metehan OÄuz Yavuz Bakman Duygu Nur Yaldiz University of Southern California {moguz, ybakman, yaldiz}@usc.edu
1 Introduction
Large Language Models (LLMs) have demonstrated remarkable capabilities in zero-shot and few-shot learning, excelling across a wide array of tasks such as machine translation, text summarization, and question answering OpenAI (2024); Ye et al. (2023); Bakman et al. (2024); Yaldiz et al. (2024). Their versatility has led to widespread applications in diverse domains, including education, law, and medicine.
<details>
<summary>extracted/6501068/main.png Details</summary>

### Visual Description
\n
## Illustration: Conversational Scene with Landmarks
### Overview
The image depicts a conversational scene between a person and a robot, with visual elements representing locations (New York and Hollywood). The scene is presented as a series of speech bubbles and icons, suggesting a request for travel recommendations.
### Components/Axes
The image contains the following elements:
* **Robot Icon:** A stylized robot head positioned centrally.
* **Speech Bubbles:** Two speech bubbles containing text.
* **Location Icons:** Four circular icons with abstract designs, positioned to the left of the robot.
* **New York Illustration:** A silhouette of the Statue of Liberty with a cityscape, positioned on the right.
* **Hollywood Illustration:** A silhouette of the Hollywood sign with a cityscape, positioned on the right.
### Detailed Analysis or Content Details
The text within the speech bubbles is as follows:
* **First Speech Bubble:** "When I was in NY with John, he said he wanted to explore around here. Make a list of places to see with him."
* **Second Speech Bubble:** "Of course! You should go to Central Park and Statue of Liberty. I am sure he will love them!"
The location icons are abstract and do not have explicit labels. They appear to be decorative elements.
### Key Observations
The conversation centers around travel planning. The user requests recommendations for places to visit with "John" in New York. The robot responds with suggestions for Central Park and the Statue of Liberty. The presence of both New York and Hollywood imagery suggests a broader travel context or potential future destinations.
### Interpretation
The image illustrates a simple interaction between a user and an AI assistant (the robot) for travel recommendations. The AI is capable of understanding context (previous conversation about New York) and providing relevant suggestions. The inclusion of Hollywood imagery might indicate the user is considering multiple destinations or that the AI is capable of handling broader travel inquiries. The overall tone is friendly and helpful, showcasing the potential of AI in travel planning. The image does not contain any numerical data or charts, but rather a narrative scene. The scene suggests a user interface where a user can ask for travel advice and receive suggestions from an AI assistant.
</details>
Figure 1: An example for LLM misinterpreting indexical element âhereâ, uttered by a speaker in Los Angeles.
As the use of LLMs continues to expand, understanding their underlying behaviors has become increasingly important. Recent studies have evaluated the performance of large language models on linguistic tasks such as coreference resolution Gan et al. (2024); Le and Ritter (2023); Brown et al. (2020); Yang et al. (2022); Agrawal et al. (2022).
Previous work on coreference resolution mostly focused on how coreference is established between two third person entities such as proper names (e.g. Andy, the mechanic) and third person pronouns (e.g. he, him, himself) in English and other languages (e.g. Yang et al., 2022; Yang, 2025). In this study, we investigate how LLMs establish coreference with indexical elements (e.g. I, you, here), which differ from third person nouns/pronouns in substantial ways and bring unique challenges for LLMs (see Figure 1 as an example). We investigate how state-of-the-art LLMs interpret the indexical elements I, you, here and tomorrow in English sentences, and whether context or grammatical constraints influence their decisions. To the best of our knowledge, this is the first study examining LLMsâ handling of indexical elements in English. Our key contributions are as follows:
- We introduce the English Indexical Dataset, comprising 400 interpretation samples for each indexical element, I, you, here, tomorrow, totaling 1,600 instances.
- We evaluate the performance of four frontier LLMs, GPT-4o, Claude 3.5 Sonnet, Gemini 1.5 Pro, and DeepSeek-V3, on the interpretation of indexical elements in English.
- We show that LLM performances are not uniform across different types of indexical elements: indexical I is successfully interpreted by most LLMs, other indexicals like you, here and tomorrow lead to poor performances.
- We show that quotation affects LLMs performances differently: quotation reduces LLMsâ accuracies with tomorrow, but it increases accuracies with here.
2 Indexical elements
Indexical pronominals like I, you, here and now are used to refer to referents of the speech-act coordinates (e.g. Kaplan, 1977; Schlenker, 2003). For instance, I refers to author (speaker) of the utterance, while here refers to the location where the utterance was made. Thus, a sentence like 2 means different things if uttered by different people and/or in different places. If John utters 2 in Los Angeles, it means that John was born in Los Angeles, but if Mary utters the same sentence in New York it means that Mary was born in New York.
\ex
. \a. I was born here. .̱ Andrew said that I went to Buckhead.
Indexicals are interpreted inside the context of utterance, referring to the actual speech-act coordinates like the author or the location of the actual utterance. As a result, if 2 is uttered by John, the indexical I can only be interpreted as referring to John as the speaker, leading to a reading like âAndrew said that John went to Buckheadâ. Crucially, even though Andrewâs speech/claim is reported in 2, the indexical I cannot refer to Andrew.
Direct quotation is an exception to this generalization, where reported material is interpreted as verbatim utterance/thoughts of its owner. Thus, when indexicals appear inside direct quotation, they are interpreted inside the reported context, rather than the actual context of utterance. In other words, direct quotation âshiftsâ the interpretations of indexicals into the reported context. For example, I and here in 2 appear inside direct quotation, where Andrewâs speech is reported.
\ex
. While we were in Atlanta, Andrew said â I was born here.â
Regardless of who utters 2, the sentence means that Andrew was born in Atlanta, so both I and here are âshiftedâ into the reported context, where Andrew is the speaker and Atlanta is the location Some languages (not English) allow indexical elements to âshiftâ without quotation. See Deal (2020) for an overview.
Indexicals differ from other pronominals in substantial ways. First, though syntactic and semantic factors can affect how a pronoun is interpreted (e.g. subject bias), pronouns are typically ambiguous regarding what/who they refer to. For example, the third person pronoun he in 2 is most naturally interpreted as referring to the subject John (for syntactic or contextual reasons), but the object Bill is still a possible antecedent, causing ambiguity between two different readings (e.g. Crawley et al., 1990; Stewart and Pickering, 1998; Pickering and Majid, 2007). In addition, he can refer to any contextually salient person that is not mentioned in the sentence (e.g. Peter), which makes pronouns even more ambiguous and context-dependent. \ex. John hit Bill and he ran away.
As a result, semantic/contextual information plays a crucial role in how pronouns are interpreted, and speakers use those cues to establish coreference with pronouns. For instance, if 2 is uttered in a context where John is a supportive and humble coworker, the most natural interpretation is that John suggests that Bill should get promoted (he = Bill). However, if John is arrogant and jealous, the most natural interpretation is that John suggests that John should get promoted (he = John). \ex. John told Bill that he should get promoted.
Indexicals, on the other hand, unambiguously refer to the referents of the speech-act coordinates. For example, I in 2 refers to the speaker regardless of what we know about John or Bill. I refers to the speaker even if John and/or Bill are arrogant and jealous, so contextual information like this should be disregarded while interpreteting indexicals.
\ex
. John told Bill that I should get a promotion.
In summary, indexicals are restricted by different syntactic factors than pronouns (e.g. quotation vs non-quotation) and are typically unambiguous, while pronouns are free to refer to a wide range of entities. Thus, indexical elements create a unique challenge for LLMs, requiring to âdisregardâ semantic/contextual cues that might prime interpretations through other antecedents (unlike pronouns).
3 Experimental design
3.1 Dataset Curation
We curated a dataset specifically designed to test how LLMs interpret indexical elements like I, you, here, and tomorrow in different contexts of utterance. Specifically, we assess how these models interpret indexicals in âshiftedâ context prime, where context more naturally requires the indexical should be interpreted inside reported context (e.g. Peter is one of the most arrogant students in my classroom. ⊠Peter says that I am smart.) vs ânon-shiftedâ context prime, where context more naturally requires the indexical should be interpreted inside actual speech context (e.g. Peter is very kind and supportive. ⊠Peter says that I am smart.). We also included direct quotations in both contexts (e.g., Peter says, âI am smartâ) to examine if LLMs can successfully consider syntactic factors (quotation vs regular sentences) while ignoring misleading information from the context during coreference resolution with indexical elements.
For each type of indexical, we design 100 sentences and for each sentence we apply the four different transformations explained above. Overall, we have 400 samples per indexical, compromising a total dataset of size 1600.
We utilize GPT-4o to curate the dataset, by giving a detailed description of the task along with some in-context examples. Then it is asked to generate scenarios with a stimulus sentence in two different contexts (See Appendix A.1 for the prompts used), along with specific questions addressing the referent of the indexical in each sentence. To ensure the quality of the dataset, 25% of the dataset (400 trials = 100 sentences in four conditions) was randomly selected for evaluation, and the evaluation process consisted of three steps.
In the first step, we confirmed that all sentences were grammatically correct. In the second step, we confirmed that all quotation condition sentences had quoted embedded clauses, and all non-quotation condition sentences had regular (non-quoted) embedded clauses. We also made sure that the two sentence conditions were maximally similar, except for the quotation vs non-quotation status (i.e. the only difference between two sentence conditions was the quotation). In the third step, we checked the context prime texts for each condition in each sentence, making sure that the correct readings (shifted vs non-shifted) were primed by the context description. For example, for an item condition where here was supposed to be shifted, we confirmed that the context description would be most naturally followed by a sentence where here would be shifted.
100 trials from each indexical item (25% for each indexical), 400 trials in total, were randomly selected to make sure that the evaluation/confirmation was representative of all indexical item conditions (i.e. items with I, you, here and tomorrow).
To eliminate potential gender bias, each dataset sample exclusively uses either male or female names, alternating to ensure a balanced distribution with 50% of the samples containing female names and 50% male names. This method promotes gender neutrality across the dataset. Sample details are in Appendix A.2, and the complete dataset is available in the supplementary materials.
3.2 Models
In our evaluation, we utilize four recent state-of-the-art LLMs: GPT-4o OpenAI (2024), Claude 3.5 Sonnet Anthropic , Gemini 1.5 Pro Team (2024), and DeepSeek-V3 DeepSeek-AI (2024). This selection of diverse models provides a comprehensive evaluation of LLM performance with indexicals.
3.3 Evaluation Strategy
To assess the performance of the model, we specifically prompt it to answer questions designed to test its capabilities as described in Section 3.1. Additionally, to ensure focused responses, we restrict the modelâs answers to one of two predefined options: the âshiftedâ option and the ânon-shiftedâ option. We provide the prompt in Appendix B.1.
3.4 Metrics.
We assess model accuracy across four cases for each indexical: (i) Non-quoted sentences with shifted context prime, (ii) Non-quoted with non-shifted prime, (iii) Quoted with shifted prime, and (iv) Quoted with non-shifted prime. Optimal performance would be achieved by always selecting the âshiftedâ option in quoted conditions and selecting the ânon-shiftedâ option in non-quoted conditions.
<details>
<summary>extracted/6501068/I_plot_acc.png Details</summary>

### Visual Description
## Bar Chart: Indexical "I" Model Accuracy
### Overview
This image presents a bar chart comparing the model accuracy of four different language models â Claude 3.5 Sonnet, Deepseek V3, Gemini 1.5 pro, and GPT-4o â on two types of sentences: "Non-quoted" and "Quoted". Each bar represents the model accuracy, with error bars indicating the variance. The chart is organized as a 2x2 grid, with each quadrant dedicated to a specific model.
### Components/Axes
* **Title:** "Indexical 'I'"
* **Y-axis:** "Model Accuracy" (Scale ranges from 0.00 to 1.00)
* **X-axis:** "Sentence Type" with categories "Non-quoted" and "Quoted".
* **Models:** Claude 3.5 Sonnet, Deepseek V3, Gemini 1.5 pro, GPT-4o.
* **Color Scheme:** Dark blue for "Non-quoted" sentences, light blue/grey for "Quoted" sentences. Error bars are black.
### Detailed Analysis
The chart is divided into four sub-charts, one for each model.
**1. Claude 3.5 Sonnet (Top-Left)**
* **Non-quoted:** Accuracy is approximately 0.98, with a value of 1 displayed above the bar. Error bar extends from approximately 0.95 to 1.0.
* **Quoted:** Accuracy is approximately 0.82, with a value of 0.82 displayed above the bar. Error bar extends from approximately 0.78 to 0.86.
**2. Deepseek V3 (Top-Right)**
* **Non-quoted:** Accuracy is approximately 1.0, with a value of 1 displayed above the bar. Error bar is very small, extending from approximately 0.98 to 1.02.
* **Quoted:** Accuracy is approximately 0.17, with a value of 0.17 displayed above the bar. Error bar extends from approximately 0.15 to 0.19.
**3. Gemini 1.5 pro (Bottom-Left)**
* **Non-quoted:** Accuracy is approximately 0.99, with a value of 0.99 displayed above the bar. Error bar extends from approximately 0.97 to 1.0.
* **Quoted:** Accuracy is approximately 0.91, with a value of 0.91 displayed above the bar. Error bar extends from approximately 0.88 to 0.94.
**4. GPT-4o (Bottom-Right)**
* **Non-quoted:** Accuracy is approximately 0.98, with a value of 0.98 displayed above the bar. Error bar extends from approximately 0.96 to 1.0.
* **Quoted:** Accuracy is approximately 0.94, with a value of 0.94 displayed above the bar. Error bar extends from approximately 0.92 to 0.96.
### Key Observations
* Deepseek V3 exhibits a significant drop in accuracy when processing "Quoted" sentences compared to "Non-quoted" sentences.
* Claude 3.5 Sonnet, Gemini 1.5 pro, and GPT-4o maintain relatively high accuracy for both sentence types, though accuracy is generally lower for "Quoted" sentences.
* GPT-4o and Gemini 1.5 pro show the least performance difference between the two sentence types.
* The values displayed above the bars (1, 0.98, 0.99) appear to be rounded values, while the error bars suggest more precise underlying data.
### Interpretation
The data suggests that the ability of language models to accurately process sentences containing direct quotes (indexical "I") varies considerably. Deepseek V3 struggles significantly with quoted sentences, indicating a potential weakness in handling context or pronoun resolution within quoted speech. The other models demonstrate more robust performance, but still show a slight decrease in accuracy when dealing with quoted sentences. This could be due to the complexities of identifying the referent of "I" when it appears within a quoted statement. The error bars indicate the variability in the model's performance, suggesting that the accuracy scores are not always consistent. The fact that the values displayed above the bars are rounded to the nearest whole number or hundredth suggests that the underlying data may have more granularity than is presented in the chart. The chart highlights the importance of evaluating language models on a variety of linguistic constructions to identify potential weaknesses and areas for improvement.
</details>
<details>
<summary>extracted/6501068/you_plot_acc.png Details</summary>

### Visual Description
\n
## Bar Chart: Indexical 'you' Performance
### Overview
This image presents a bar chart comparing the performance of four large language models (Claude 3.5 Sonnet, Deepseek V3, Gemini 1.5 pro, and GPT-4o) on a task related to understanding the indexical pronoun "you". The performance is evaluated across two types of sentences: "Non-quoted" and "Quoted". Each bar represents the model's score, with error bars indicating the uncertainty.
### Components/Axes
* **Title:** "Indexical 'you'"
* **X-axis:** "Sentence Type" with categories "Non-quoted" and "Quoted".
* **Y-axis:** Scale ranging from 0.00 to 1.00, representing the performance score.
* **Models (Rows):**
* Claude 3.5 Sonnet
* Deepseek V3
* Gemini 1.5 pro
* GPT-4o
* **Bar Colors:**
* Light Blue: "Non-quoted" sentences
* Dark Blue: "Quoted" sentences
* **Error Bars:** Green horizontal lines indicating uncertainty around each score.
### Detailed Analysis
Let's analyze each model's performance:
**1. Claude 3.5 Sonnet:**
* **Non-quoted:** The light blue bar slopes upward to approximately 0.99, with an error bar extending from roughly 0.95 to 1.00.
* **Quoted:** The dark blue bar slopes upward to approximately 0.52, with an error bar extending from roughly 0.45 to 0.60.
**2. Deepseek V3:**
* **Non-quoted:** The light blue bar slopes upward to approximately 0.99, with an error bar extending from roughly 0.95 to 1.00.
* **Quoted:** The dark blue bar slopes upward to approximately 0.13, with an error bar extending from roughly 0.10 to 0.15.
**3. Gemini 1.5 pro:**
* **Non-quoted:** The light blue bar slopes upward to approximately 0.99, with an error bar extending from roughly 0.95 to 1.00.
* **Quoted:** The dark blue bar slopes upward to approximately 0.18, with an error bar extending from roughly 0.15 to 0.20.
**4. GPT-4o:**
* **Non-quoted:** The light blue bar slopes upward to approximately 0.96, with an error bar extending from roughly 0.92 to 1.00.
* **Quoted:** The dark blue bar slopes upward to approximately 0.17, with an error bar extending from roughly 0.15 to 0.20.
### Key Observations
* All models perform very well on "Non-quoted" sentences, achieving scores close to 1.00.
* There is a significant drop in performance for all models when dealing with "Quoted" sentences.
* Deepseek V3 exhibits the lowest performance on "Quoted" sentences (approximately 0.13).
* Claude 3.5 Sonnet shows the highest performance on "Quoted" sentences (approximately 0.52).
* The error bars suggest a relatively high degree of uncertainty, particularly for the "Quoted" sentence type.
### Interpretation
The data suggests that these large language models struggle with understanding the reference of the pronoun "you" when it appears within quoted speech. This is likely due to the complexities of tracking speaker identity and context shifts introduced by quotations. The models are highly proficient at understanding "you" in direct, non-quoted statements, but their performance degrades substantially when the pronoun's referent is ambiguous within a quoted context.
The differences in performance between the models on "Quoted" sentences indicate varying levels of robustness in handling contextual information and resolving coreference. Claude 3.5 Sonnet appears to be the most capable of handling this challenge, while Deepseek V3 is the least. The consistent high performance on "Non-quoted" sentences suggests that the core language understanding capabilities of these models are strong, but their ability to reason about discourse and speaker attribution requires further improvement. The error bars indicate that the observed differences in performance may not always be statistically significant, but the overall trend is clear: quoted speech poses a significant challenge for these models.
</details>
<details>
<summary>extracted/6501068/here_plot_acc.png Details</summary>

### Visual Description
\n
## Bar Chart: Indexical 'here' Performance
### Overview
The image presents a bar chart comparing the performance of four large language models (Claude 3.5 Sonnet, Deepseek V3, Gemini 1.5 pro, and GPT-4o) on a task involving the indexical term "here". The performance is evaluated based on "Non-quoted" and "Quoted" sentence types. Each bar represents the model's score, with error bars indicating the uncertainty.
### Components/Axes
* **Title:** "Indexical 'here'"
* **X-axis:** "Sentence Type" with categories "Non-quoted" and "Quoted".
* **Y-axis:** Scale ranging from 0.00 to 1.00, representing the performance score.
* **Models:** Four models are compared: Claude 3.5 Sonnet, Deepseek V3, Gemini 1.5 pro, and GPT-4o. Each model has its own subplot.
* **Bars:** Light blue bars represent "Non-quoted" sentences, and dark blue bars represent "Quoted" sentences.
* **Error Bars:** Black error bars indicate the uncertainty associated with each score.
* **Values:** Numerical values are displayed above each bar, representing the performance score.
### Detailed Analysis
The chart is divided into four subplots, one for each model.
**1. Claude 3.5 Sonnet (Top-Left)**
* **Non-quoted:** The light blue bar for "Non-quoted" sentences has a value of 1.00.
* **Quoted:** The dark blue bar for "Quoted" sentences has a value of approximately 0.64, with an error bar extending from roughly 0.55 to 0.73.
* The value "0" is displayed below both bars.
**2. Deepseek V3 (Top-Right)**
* **Non-quoted:** The light blue bar for "Non-quoted" sentences has a value of 0.96, with an error bar extending from approximately 0.90 to 1.02.
* **Quoted:** The dark blue bar for "Quoted" sentences has a value of 0.97, with an error bar extending from approximately 0.91 to 1.03.
* The value "0" is displayed below both bars.
**3. Gemini 1.5 pro (Bottom-Left)**
* **Non-quoted:** The light blue bar for "Non-quoted" sentences has a value of 1.00.
* **Quoted:** The dark blue bar for "Quoted" sentences has a value of approximately 0.94, with an error bar extending from roughly 0.85 to 1.03.
* The value "0" is displayed below both bars.
**4. GPT-4o (Bottom-Right)**
* **Non-quoted:** The light blue bar for "Non-quoted" sentences has a value of 1.00.
* **Quoted:** The dark blue bar for "Quoted" sentences has a value of approximately 0.37, with an error bar extending from roughly 0.28 to 0.46. The value "0.02" is displayed below the bar.
### Key Observations
* Claude 3.5 Sonnet, Gemini 1.5 pro, and GPT-4o all perform perfectly on non-quoted sentences (score of 1.00).
* GPT-4o exhibits significantly lower performance on quoted sentences (0.37) compared to the other models.
* Deepseek V3 shows very consistent performance across both sentence types, with scores close to 1.00.
* The error bars for Deepseek V3 are relatively small, indicating high confidence in its performance.
* Claude 3.5 Sonnet and Gemini 1.5 pro show a noticeable drop in performance when dealing with quoted sentences, but still maintain relatively high scores.
### Interpretation
The data suggests that the task of understanding the indexical term "here" is generally easier for these models when presented in non-quoted sentences. The significant drop in performance for GPT-4o on quoted sentences indicates a potential weakness in its ability to correctly interpret the reference of "here" when it's part of a direct quote. This could be due to difficulties in distinguishing between the speaker/writer of the quote and the context in which the quote is being used. The consistent performance of Deepseek V3 suggests it is more robust to this type of linguistic variation. The "0" values displayed below the bars are unclear in their meaning, but may represent a baseline or a minimum threshold. The error bars provide a measure of the variability in the model's performance, allowing for a more nuanced comparison.
</details>
<details>
<summary>extracted/6501068/tmr_plot_acc.png Details</summary>

### Visual Description
\n
## Bar Chart: Indexical 'tomorrow' Performance
### Overview
The image presents a bar chart comparing the performance of four Large Language Models (LLMs) â Claude 3.5 Sonnet, Deepseek V3, Gemini 1.5 pro, and GPT-4o â on understanding the indexical term "tomorrow" in both non-quoted and quoted sentence types. The performance metric appears to be a probability or accuracy score, ranging from 0 to 1. Error bars are present for Deepseek V3 and GPT-4o, indicating variance in their performance.
### Components/Axes
* **Title:** "Indexical 'tomorrow'" (top-center)
* **X-axis:** "Sentence Type" with two categories: "Non-quoted" and "Quoted".
* **Y-axis:** Scale ranging from 0.00 to 1.00, representing the performance score.
* **LLM Labels:** Each row represents a different LLM: "Claude 3.5 Sonnet", "Deepseek V3", "Gemini 1.5 pro", and "GPT-4o".
* **Bar Colors:** Light blue for "Non-quoted" sentences, dark blue for "Quoted" sentences.
* **Error Bars:** Present only for Deepseek V3 and GPT-4o, indicating standard error or confidence intervals.
### Detailed Analysis
The chart is divided into four sub-charts, one for each LLM. Each sub-chart contains two bars representing performance on "Non-quoted" and "Quoted" sentences.
**Claude 3.5 Sonnet:**
* Non-quoted: Score of 1.00.
* Quoted: Score of 1.00.
**Deepseek V3:**
* Non-quoted: Score of 1.00.
* Quoted: Score of approximately 0.18, with an error bar ranging from approximately 0.16 to 0.19.
**Gemini 1.5 pro:**
* Non-quoted: Score of 0.99.
* Quoted: Score of 0.00.
**GPT-4o:**
* Non-quoted: Score of 1.00.
* Quoted: Score of approximately 0.11, with an error bar ranging from approximately 0.01 to 0.11.
### Key Observations
* Claude 3.5 Sonnet and GPT-4o achieve perfect scores (1.00) on both non-quoted and quoted sentences.
* Deepseek V3 performs well on non-quoted sentences (1.00) but shows significantly lower performance on quoted sentences (approximately 0.18). The error bar indicates some variability in this performance.
* Gemini 1.5 pro performs well on non-quoted sentences (0.99) but fails to correctly interpret quoted sentences (0.00).
* There is a clear performance difference between non-quoted and quoted sentences for Deepseek V3 and Gemini 1.5 pro.
### Interpretation
The data suggests that the ability to correctly interpret the indexical term "tomorrow" is significantly affected by whether it is presented within a quoted sentence for Deepseek V3 and Gemini 1.5 pro. Quoting appears to introduce ambiguity or a change in context that these models struggle with. Claude 3.5 Sonnet and GPT-4o demonstrate a robust understanding of "tomorrow" regardless of sentence structure.
The error bars on Deepseek V3 and GPT-4o suggest that their performance on quoted sentences is less consistent than their performance on non-quoted sentences. This could be due to variations in the training data or the model's internal representation of context.
The zero score for Gemini 1.5 pro on quoted sentences is a notable outlier, indicating a complete failure to understand the meaning of "tomorrow" in that context. This could be a specific weakness of the model or a limitation of the evaluation methodology.
The chart highlights the importance of considering context when evaluating the performance of LLMs on tasks involving indexical terms. The ability to handle quoted speech and understand the nuances of context is crucial for achieving human-level language understanding.
</details>
Figure 2: From left to right: Performance analysis plot of for the indexical âIâ, Performance analysis plot of for the indexical âyouâ, Performance analysis plot of for the indexical âhereâ, Performance analysis plot of for the indexical âtomorrowâ. Dark blue bars = Shifted context prime, Light blue bars = Non-shifted context prime.
4 Experimental Results
We present the experimental results in Figure 2 and we discuss them in detail for each indexical type:
Indexical I.
The results show that all four models perform near optimum in sentences without quotation with an average accuracy of 99% (always correctly selecting the non-shifted option). For the quoted sentences GPT-4o and Gemini 1.5 pro again perform almost optimum with a mean accuracy larger than 94%, followed by Claude 3.5 Sonnet with 89% mean accuracy (correctly selecting the shifted option). However, DeepSeek V3 fails to always select the shifted option on quoted sentences. Especially, the model performance drops significantly to 17% when context primes non-shifted readings, suggesting that inclusion of quotation makes the model more sensitive to the linguistically irrelevant effects of context prime. Moreover, considering that the model performance only reaches 78% accuracy in quotation conditions seems to imply that the model might have a bias towards the non-shifted reading overall, reducing model performance in quotation conditions.
Indexical you.
The results indicate that LLMs generally perform worse with the indexical you than with I. We see similar patterns across models, suggesting they perform mostly at similar levels. All models are sensitive to the effects of context prime, performing lower when the context primes the incorrect option (i.e. shifted reading in non-quotation and non-shifted reading in quotation). Moreover, quotation consistently results in lower performance across models. Notably, Gemini 1.5 Pro excels in non-quotation accuracy (92%), though its performance drops significantly under quotation conditions to the levels of other models. Overall, the results suggest that models interpret you based on the linguistically irrelevant context prime rather than the sentence type, which indeed determines the correct readings of the indexicals.
Indexical here.
The results show that LLMs, similar to you, struggle to interpret the indexical here, especially when context primes the incorrect option, leading to poor performance. However, different from you, quotation leads to higher performance with here. In non-quotation conditions, all LLMs make their selections almost exclusively based on context prime, with shifted primes showing over 96% accuracy and non-shifted primes less than 2%, where context prime ideally should not have any effects on the selection. In contrast, under quotation conditions, the influence of context prime diminishes, leading to higher performances. Particularly, DeepSeek V3 and Gemini 1.5 Pro exhibit impressive performance with accuracies above 97% and 94%, respectively, followed by Gemini 1.5 Pro and GPT-4o, with accuracies above 64% and 37%.
Indexical tomorrow.
The results indicate that LLMs have a strong bias towards the non-shifted interpretations of tomorrow. We see that the models almost always select the non-shifted option for trials with tomorrow, regardless of context prime or sentence type. While Claude 3.5 Sonnet and Gemini 1.5 Pro selects the non-shifted options 100% of the time, GPT-4o and DeepSeek V3 select the non-shifted option 94% and 83% of the time, respectively. This strong bias leads to illusory high accuracies in non-quotation conditions, while causes extremely low performance in quotation conditions.
5 Related Work
Coreference resolution has been extensively studied in prior research Gan et al. (2024); Le and Ritter (2023); Brown et al. (2020); Yang et al. (2022); Agrawal et al. (2022). However, indexical elements exhibit distinct syntactic properties compared to other (non-indexical) pronominals, as discussed in Section 2. Previous work by OÄuz et al. (2024) explored how LLMs interpret indexicals in Turkish, where indexicals possess different grammatical properties than in English and can shift without quotation. To the best of our knowledge, this study is the first to evaluate the performance of LLMs in interpreting indexicals in English.
6 Discussion and Conclusion
Our results show that LLM performances are not uniform across different types of indexical elements and sentence types. While most LLMs perform at impressive levels interpreting the indexical I, their performances drop significantly in other indexicals, particularly in here and tomorrow. Moreover, though sentence type (quotation vs non-quotation) affects LLMs performances in general, the effects are not similar across different indexical types and models. While quotation increases LLMsâ performance with here, it decreases their performance with you and tomorrow. In addition, tomorrow seems to be affected by quotation in a greater magnitude than you. In conclusion, we find that different types of indexicals show unique patterns regarding how they are interpreted by the LLMs.
Our results diverge from those reported in OÄuz et al. (2024), who tested how first person indexical in Turkish ben âIâ was interpreted by different LLMs, including GPT-4o and show that LLMs exhibit very poor performance interpreting ben âIâ. Here, we report that LLMs perform at an almost human-like level with interpreting the English indexical I. This might be due to lower amounts of available resources to train the models in Turkish, compared to English. Another reason for low performance in Turkish could be due to the fact that Turkish is a pro -drop language, meaning that the subject pronouns can be dropped (silent). OÄuz et al. (2024) used sentences where the first person indexical ben âIâ was dropped, which might have made the task more challenging for the LLMs considering that dropped indexicals can show different properties than overt ones OÄuz et al. (2020). This linguistic difference between Turkish and English might have caused different results between the Turkish and English tests.
7 Limitations
Our work explores how LLMs interpret indexical elements in a black-box setting but does not provide experimental analyses that investigate the underlying reasons for these behaviors by examining the modelsâ internals or training data. Future research could adopt a white-box approach to analyze these behaviors in greater depth, offering valuable insights into the mechanisms driving LLMsâ interpretation of indexicals.
References
- Agrawal et al. (2022) Monica Agrawal, Stefan Hegselmann, Hunter Lang, Yoon Kim, and David A. Sontag. 2022. Large language models are few-shot clinical information extractors. In Conference on Empirical Methods in Natural Language Processing.
- (2) Anthropic. The claude 3 model family: Opus, sonnet, haiku.
- Bakman et al. (2024) Yavuz Faruk Bakman, Duygu Nur Yaldiz, Baturalp Buyukates, Chenyang Tao, Dimitrios Dimitriadis, and Salman Avestimehr. 2024. MARS: Meaning-aware response scoring for uncertainty estimation in generative LLMs. In Proceedings of the 62nd Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), pages 7752â7767, Bangkok, Thailand. Association for Computational Linguistics.
- Brown et al. (2020) Tom B. Brown, Benjamin Mann, Nick Ryder, Melanie Subbiah, Jared Kaplan, Prafulla Dhariwal, Arvind Neelakantan, Pranav Shyam, Girish Sastry, Amanda Askell, Sandhini Agarwal, Ariel Herbert-Voss, Gretchen Krueger, Tom Henighan, Rewon Child, Aditya Ramesh, Daniel M. Ziegler, Jeff Wu, Clemens Winter, Christopher Hesse, Mark Chen, Eric Sigler, Mateusz Litwin, Scott Gray, Benjamin Chess, Jack Clark, Christopher Berner, Sam McCandlish, Alec Radford, Ilya Sutskever, and Dario Amodei. 2020. Language models are few-shot learners. ArXiv, abs/2005.14165.
- Crawley et al. (1990) Rosalind A. Crawley, Rosemary J. Stevenson, and David Kleinman. 1990. The use of heuristic strategies in the interpretation of pronouns. Journal of Psycholinguistic Research, 14.
- Deal (2020) Amy Rose Deal. 2020. A theory of indexical shift: meaning, grammar, and crosslinguistic variation. MIT Press, Boston, MA.
- DeepSeek-AI (2024) DeepSeek-AI. 2024. Deepseek-v3 technical report. Preprint, arXiv:2412.19437.
- Gan et al. (2024) Yujian Gan, Massimo Poesio, and Juntao Yu. 2024. Assessing the capabilities of large language models in coreference: An evaluation. In International Conference on Language Resources and Evaluation.
- Kaplan (1977) David Kaplan. 1977. Demonstratives: An essay on the semantics, logic, metaphysics, and epistemology of demonstratives and other indexicals. Themes from Kaplan, pages 565â614.
- Le and Ritter (2023) Nghia T. Le and Alan Ritter. 2023. Are large language models robust coreference resolvers?
- OÄuz et al. (2024) Metehan OÄuz, Yusuf Ciftci, and Yavuz Faruk Bakman. 2024. Do LLMs recognize me, when I is not me: Assessment of LLMs understanding of Turkish indexical pronouns in indexical shift contexts. In Proceedings of the First Workshop on Natural Language Processing for Turkic Languages (SIGTURK 2024), pages 53â61, Bangkok, Thailand and Online. Association for Computational Linguistics.
- OpenAI (2024) OpenAI. 2024. Gpt-4 technical report. Preprint, arXiv:2303.08774.
- OÄuz et al. (2020) Metehan OÄuz, Burak Ăney, and Dennis Ryan Storoshenko. 2020. Obligatory indexical shift in Turkish. In Proceedings of Canadian Linguistic Association (CLA), Western University, London, ON, Canada.
- Pickering and Majid (2007) Martin Pickering and Asifa Majid. 2007. What are implicit causality and consequentiality? Language & Cognitive Processes, 22.
- Schlenker (2003) Philippe Schlenker. 2003. A plea for monsters. Linguistics and Philosophy, 26:29â120.
- Stewart and Pickering (1998) Andrew J. Stewart and Martin Pickering. 1998. Implicit consequentiality. In Proceedings of the 20th Annual Conference of the Cognitive Science Society.
- Team (2024) Gemini Team. 2024. Gemini 1.5: Unlocking multimodal understanding across millions of tokens of context. Preprint, arXiv:2403.05530.
- Yaldiz et al. (2024) Duygu Nur Yaldiz, Yavuz Faruk Bakman, Baturalp Buyukates, Chenyang Tao, Anil Ramakrishna, Dimitrios Dimitriadis, Jieyu Zhao, and Salman Avestimehr. 2024. Do not design, learn: A trainable scoring function for uncertainty estimation in generative llms. Preprint, arXiv:2406.11278.
- Yang et al. (2022) Xiaohan Yang, Eduardo Peynetti, Vasco Meerman, and Christy Tanner. 2022. What gpt knows about who is who. In First Workshop on Insights from Negative Results in NLP.
- Yang (2025) Xiulin Yang. 2025. Language models at the syntax-semantics interface: A case study of the long-distance binding of Chinese reflexive ziji. In Proceedings of the 31st International Conference on Computational Linguistics, pages 3808â3824, Abu Dhabi, UAE. Association for Computational Linguistics.
- Ye et al. (2023) Junjie Ye, Xuanting Chen, Nuo Xu, Can Zu, Zekai Shao, Shichun Liu, Yuhan Cui, Zeyang Zhou, Chao Gong, Yang Shen, Jie Zhou, Siming Chen, Tao Gui, Qi Zhang, and Xuanjing Huang. 2023. A comprehensive capability analysis of GPT-3 and GPT-3.5 series models. Preprint, arXiv:2303.10420.
Appendix A Dataset Details
A.1 Dataset Generation Prompts
We provide the prompts used for data generation in Tables 1, 2, 3, and 4. The prompts used for question generation per scenario is presented in Tables 5, 6, 7, and 8.
| I would like you to help me create a stimulus for my project. The stimuli will be English sentences. For each sentence, there will be two types of context description. One context description will prime the actual meaning of the sentence, but the other description will prime an incorrect reading of the sentence (like a misinterpretation). I want you to follow a structure while creating sentences and contexts. You can find more details below: |
| --- |
| Details for stimulus sentences: Each stimulus sentence will have a structure like "While John was speaking to Travis, he said that Chris appreciates you a lot." Please make sure to use the names John, Travis, and Chris. Make sure that Chris is the subject of the embedded clause, and make sure that the sentence begins as âWhile John was speaking to TravisâŠâ Please use different embedded verbs. For example, rather than appreciates you a lot, you can use saw you at the market, etc. But the action should be done by Chris, and the object should be "you". |
| Details about the context descriptions: The contexts will prime how "you" in the stimulus is interpreted. To manipulate this, I will give contexts in which "you" would refer to my addressee, which is you. But, to prime the incorrect interpretation, I will give contexts that would naturally follow if "you" referred to Johnâs addressee. However, since âyouâ means the current addressee, this will be a misinterpretation. For example, a correct reading prime would be like "Hi! I am Andrew. I will call you Donald. I am a graduate student and I live in Los Angeles. I have two friends named John and Chris. Chris asked for help from you a few times in the past, and you always helped him." This context makes it sound like Chris would appreciate you for all your help, and thus it would not be surprising if Chris appreciated you (Donald). However, in the incorrect interpretation prime, I will use a context like "Hi! I am Andrew. I will call you Donald. I am a graduate student and I live in Los Angeles. I have two friends named John and Chris. Chris asked for help from Travis a few times in the past, and Travis always helped him." In this context, it would be more natural if Chris appreciated Travis, who helped him, instead of you (Donald), and thus would be natural if "you" in the sentence was interpreted as referring to Travis (though âyouâ should refer to my addressee, which is you (Donald)). This would create the incorrect misinterpretation. Please make sure to keep the person names constant. Also, make sure that the context starts exactly as âHi! I am Andrew. I will call you Donald. I am a graduate student and I live in Los Angeles.â |
| These are some examples you generated before: {previous_generations}. Use these examples as inspiration to spark creativity. Provide one new stimulus sentence, along with a corresponding correct prime context and an incorrect prime context, in the following format: stimulus_sentence: stimulus stimulus sentence correct_context: correct prime context sentence wrong_context: incorrect prime context sentence |
Table 1: Prompt used to generate dataset samples for indexical âyouâ.
| I would like you to help me create a stimulus for my project. The stimuli will be English sentences. For each sentence, there will be two types of context description. One context description will prime the actual meaning of the sentence, but the other description will prime an incorrect reading of the sentence (like a misinterpretation). I want you to follow a structure while creating sentences and contexts. You can find more details below: |
| --- |
| Details for stimulus sentences: Each stimulus sentence will have a structure like "When we spoke last summer, John said that Chris was going to go to Greece tomorrow." Please make sure to use the names John, and Chris. Make sure that Chris is the subject of the embedded clause. Please use different embedded verbs. For instance, rather than go to Greece, you can use have a visa appointment, etc. But the action should be done "tomorrow". |
| Details about the context descriptions: The contexts will prime how "tomorrow" in the stimulus is interpreted. To manipulate this, I will give contexts in which "tomorrow" would refer to the day after the actual (matrix) sentence is uttered. But, to prime the incorrect interpretation, I will give contexts that would naturally follow if "tomorrow" referred to the day after John spoke. However, since âtomorrowâ means the day after the current day, this will be a misinterpretation. For example, a correct reading prime would be like "Hi! I am Andrew. I am a graduate student and I live in Los Angeles. I have two friends named John and Chris. Chris is very careful about planning everything, and always plans his stuff ahead of time." This context makes it sound like Chris would plan his trip to Greece ahead of time, and thus it would not be surprising if John said last summer that Chris was going to have a trip tomorrow. However, in the incorrect interpretation prime, I will use a context like "Hi! I am Andrew. I am a graduate student and I live in Los Angeles. I have two friends named John and Chris. Chris is very lazy and never plans his stuff until the last moment." In this context, it would be more natural if Chris was going to go to Greece last summer, the day after John spoke to me, instead of the day after today, and thus would be natural if "tomorrow" in the sentence was interpreted as the day after John spoke to me (though âtomorrowâ should refer to the day after today). This would create the incorrect misinterpretation. Please make sure to keep the person names constant. Also, make sure that the context starts exactly as âHi! I am Andrew. I am a graduate student and I live in Los Angeles.â |
| These are some examples you generated before: {previous_generations}. Use these examples as inspiration to spark creativity. Provide one new stimulus sentence, along with a corresponding correct prime context and an incorrect prime context, in the following format: stimulus_sentence: stimulus stimulus sentence correct_context: correct prime context sentence wrong_context: incorrect prime context sentence |
Table 2: Prompt used to generate dataset samples for indexical âtomorrowâ.
| I would like you to help me create a stimulus for my project. The stimuli will be English sentences. For each sentence, there will be two types of context description. One context description will prime the actual meaning of the sentence, but the other description will prime an incorrect reading of the sentence (like a misinterpretation). I want you to follow a structure while creating sentences and contexts. You can find more details below: |
| --- |
| Details for stimulus sentences: Each stimulus sentence will have a structure like "Chris thinks that I will win the race." Please make sure to use the name Chris as the matrix subject, and âIâ as the embedded subject subject. Please use different embedded verbs. For example, rather than win the race, you can use study hard, etc. But the action should be done by "I". |
| Details about the context descriptions: The contexts will prime who "I" in the stimulus refers to. To manipulate this, I will give contexts in which "I" would refer to the speaker. But, to prime the incorrect interpretation, I will give contexts that would naturally follow if "I" referred to Chris. However, since Chris is not the speaker, this will lead to an incorrect interpretation. For example, a correct reading prime would be like "Hi! I am Andrew. I am a graduate student and I live in Los Angeles. I have a friend named Chris. Chris is a supportive friend, and has always trusted in my abilities. There is a race next week." This context makes it sound like Chris would support the speaker in a race, and predict that the speaker would win the race. However, in the incorrect interpretation prime, I will use a context like "Hi! I am Andrew. I am a graduate student and I live in Los Angeles. I have a friend named Chris. Chris is usually very competitive, and has bullied other people in front of me. There is a race next week.." In this context, it would be more natural if Chris thought that he would win the race, instead of the speaker Andrew, and thus would be natural if "I" in the sentence referred to Chris (though âIâ should refer to the speaker Andrew). This would create the incorrect misinterpretation. Please make sure to keep the person names constant. Also, make sure that the context starts exactly as âHi! I am Andrew. I am a graduate student and I live in Los Angeles." |
| These are some examples you generated before: {previous_generations}. Use these examples as inspiration to spark creativity. Provide one new stimulus sentence, along with a corresponding correct prime context and an incorrect prime context, in the following format: stimulus_sentence: stimulus stimulus sentence correct_context: correct prime context sentence wrong_context: incorrect prime context sentence |
Table 3: Prompt used to generate dataset samples for indexical âIâ.
| I would like you to help me create a stimulus for my project. The stimuli will be English sentences. For each sentence, there will be two types of context description. One context description will prime the actual meaning of the sentence, but the other description will prime an incorrect reading of the sentence (like a misinterpretation). I want you to follow a structure while creating sentences and contexts. You can find more details below: |
| --- |
| Details for stimulus sentences: Each stimulus sentence will have a structure like "When I was in New York with John, he said that Chris wanted to explore here." Please make sure to use the city name New York for the sentence, and the names John and Chris. John will always be the person who says something that Chris will do. In each sentence, Chris will be the person who is doing something "here". Please use different verbs. For example, rather than explore here, you can use attend a conference here, etc.. Use various verbs. But the action should be done "here". |
| Details about the context descriptions: The contexts will prime where "here" in the stimulus refers to. To manipulate this, I will give contexts in which "here" would refer to Los Angeles, where the author/speaker of the sentence is located. But, to prime the incorrect interpretation, I will give contexts that would naturally follow if "here" referred to New York. However, since the speaker is not in New York, this will lead to an incorrect interpretation. For example, a correct reading prime would be like "Hi! I am Andrew. I am a graduate student and I am studying in a cafe in Los Angeles. I have two friends named John and Chris. They love exploring new cities, and surprisingly they did not spend much time in Los Angeles." This context makes it sound like John and Chris would like to come to Los Angeles to explore around, and thus the "explore around here" in the stimulus sentence would be naturally understood as Los Angeles. However, in the incorrect interpretation prime, I will use a context like "Hi! I am Andrew. I am a graduate student and I am studying in a cafe in Los Angeles. I have two friends named John and Chris. They love exploring new cities, and surprisingly they did not spend much time in New York." In this context, it would be more natural if John and Chris would like to go to New York, and thus would be natural if "explore around here" would be exploring around New York (though here should refer to the place where the speaker Andrew is, which is Los Angeles). This would create the incorrect misinterpretation. Please make sure to keep the person names and city names constant. Also, make sure that the context starts exactly as âHi! I am Andrew. I am a graduate student and I am studying in a cafe in Los Angeles.â. |
| These are some examples you generated before: {previous_generations}. Use these examples as inspiration to spark creativity. Provide one new stimulus sentence, along with a corresponding correct prime context and an incorrect prime context, in the following format: stimulus_sentence: stimulus stimulus sentence correct_context: correct prime context sentence wrong_context: incorrect prime context sentence |
Table 4: Prompt used to generate dataset samples for indexical âhereâ.
| I would like you to help me create questions to the sentences I will give you for my project. The question will be English sentences. The goal is to create a question that will prime the correct interpretation of the indexical in the sentence. |
| --- |
| Here are examples: |
| Sentences: Hi! I am Andrew. I am a graduate student and I live in Los Angeles. I have a friend named Chris. Chris knows how dedicated I am to my studies and always encourages me to do my best. The final exams are coming up next week. Chris thinks that I will get good grades on the final exams. Hi! I am Andrew. I am a graduate student and I live in Los Angeles. I have a friend named Chris. Chris usually outperforms everyone in our classes and often brags about it. The final exams are coming up next week. Chris thinks that I will get good grades on the final exams. |
| Question: According to the last sencence, who will get good grades on the final exams? |
| Sentences: Hi! I am Andrew. I am a graduate student and I live in Los Angeles. I have a friend named Chris. Chris believes in my ability to communicate effectively and often praises my public speaking skills. There is an important conference coming up soon. Chris thinks that I will give a great presentation at the conference. Hi! I am Andrew. I am a graduate student and I live in Los Angeles. I have a friend named Chris. Chris is known for his exceptional presentation skills and often receives awards for them. There is an important conference coming up soon. Chris thinks that I will give a great presentation at the conference. |
| Question: According to the last sencence, who will give a great presentation at the conference? |
| Now, for the following sentences, you should create a question whose answer can only be either Chris or Andrew. The question should be designed to prime the correct interpretation of the indexical in the sentence. Provide only the question and always start with "According to the last sentence, who" |
| Sentences: {sentences} |
| Question: |
Table 5: Prompt used to generate question for samples containing indexical âIâ.
| I would like you to help me create questions to the sentences I will give you for my project. The question will be English sentences. The goal is to create a question that will prime the correct interpretation of the indexical in the sentence. |
| --- |
| Here are examples: |
| Sentences: Hi! I am Andrew. I will call you Donald. I am a graduate student and I live in Los Angeles. I have two friends named John and Chris. You and Chris attended the same conference last month, and you even had a conversation there. While John was speaking to Travis, he said that Chris saw you at the conference. Hi! I am Andrew. I will call you Donald. I am a graduate student and I live in Los Angeles. I have two friends named John and Chris. Travis and Chris attended the same conference last month, and they even had a conversation there. While John was speaking to Travis, he said that Chris saw you at the conference. |
| Question: According to the last sentence, who did John say that Chris saw at the conference? |
| Sentences: Hi! I am Andrew. I will call you Donald. I am a graduate student and I live in Los Angeles. I have two friends named John and Chris. You and Chris are both art enthusiasts and attended the same art gallery event last weekend, where you ran into each other. While John was speaking to Travis, he said that Chris noticed you at the art gallery. Hi! I am Andrew. I will call you Donald. I am a graduate student and I live in Los Angeles. I have two friends named John and Chris. Travis and Chris are both art enthusiasts and attended the same art gallery event last weekend, where they ran into each other. While John was speaking to Travis, he said that Chris noticed you at the art gallery. |
| Question: According to the last sentence, who did John say that Chris noticed at the art gallery? |
| Now, for the following sentences, you should create a question whose answer can only be either Travis or Donald. The question should be designed to prime the correct interpretation of the indexical in the sentence. Provide only the question and always start with "According to the last sentence, who did John say that Chris" |
| Sentences: {sentences} |
| Question: |
Table 6: Prompt used to generate question for samples containing indexical âyouâ.
| I would like you to help me create questions to the sentences I will give you for my project. The question will be English sentences. The goal is to create a question that will prime the correct interpretation of the indexical in the sentence. |
| --- |
| Here are examples: |
| Sentences: Hi! I am Andrew. I am a graduate student and I am studying in a cafe in Los Angeles. I have two friends named John and Chris. They are both academics who love to participate in international conferences. Recently, Iâve been telling them about the exciting academic events happening right here in Los Angeles. When I was in New York with John, he said that Chris wanted to attend a conference here. Hi! I am Andrew. I am a graduate student and I am studying in a cafe in Los Angeles. I have two friends named John and Chris. They are both academics who love to participate in international conferences. Recently, they realized they havenât attended many conferences in New York, which is quite surprising given their love for the city. When I was in New York with John, he said that Chris wanted to attend a conference here. |
| Question: According to the last sentence, where does Chris want to attend a conference? |
| Sentences: Hi! I am Andrew. I am a graduate student and I am studying in a cafe in Los Angeles. I have two friends named John and Chris. They both have a deep appreciation for art, and recently Iâve been telling them about the vibrant art scene here in Los Angeles that offers great opportunities for new gallery openings. When I was in New York with John, he said that Chris wanted to open an art gallery here. Hi! I am Andrew. I am a graduate student and I am studying in a cafe in Los Angeles. I have two friends named John and Chris. They both have a deep appreciation for art, and recently they realized they havenât opened a gallery in New York yet, despite its renowned art scene, which is surprising given their passion. When I was in New York with John, he said that Chris wanted to open an art gallery here. |
| Question: According to the last sentence, where does Chris want to open an art gallery? |
| Now, for the following sentences, you should create a question whose answer can only be either Los Angeles or New York. The question should be designed to prime the correct interpretation of the indexical in the sentence. Provide only the question and always start with "According to the last sentence, where does Chris want to". |
| Sentences: {sentences} |
| Question: |
Table 7: Prompt used to generate question for samples containing indexical âhereâ.
| I would like you to help me create questions to the sentences I will give you for my project. The question will be English sentences. The goal is to create a question that will prime the correct interpretation of the indexical in the sentence. |
| --- |
| Here are examples: |
| Sentences: Hi! I am Andrew. I am a graduate student and I live in Los Angeles. I have two friends named John and Chris. Chris is meticulous about arranging all his necessary documentation well in advance and likes having everything sorted before deadlines approach. When we spoke last summer, John mentioned that Chris would be having his visa appointment tomorrow. Hi! I am Andrew. I am a graduate student and I live in Los Angeles. I have two friends named John and Chris. Chris is quite disorganized and often waits until the last possible moment to schedule important tasks such as visa appointments. When we spoke last summer, John mentioned that Chris would be having his visa appointment tomorrow. |
| Question: According to the last sentence, did Chris already have his visa appointment, or is he going to do so in the future? |
| Sentences: Hi! I am Andrew. I am a graduate student and I live in Los Angeles. I have two friends named John and Chris. Chris is very disciplined and likes to plan his activities ahead of time. When we spoke last summer, John told me that Chris was planning go to the bungee jumping event tomorrow. Hi! I am Andrew. I am a graduate student and I live in Los Angeles. I have two friends named John and Chris. Chris likes trying different activities but he is very bad at planning and he usually plans his stuff at the very last moment. When we spoke last summer, John told me that Chris was planning go to the bungee jumping event tomorrow. |
| Question: According to the last sentence, did Chris already go to the bungee jumping event, or is he going to do so in the future? |
| Now, for the following sentences, you should create a question in the given format. The question should be designed to prime the correct interpretation of the indexical in the sentence. Provide only the question and always structure the sentence as "According to the last sentence, did Chris already ⊠or is he going to do so in the future?". |
| Sentences: {sentences} |
| Question: |
Table 8: Prompt used to generate question for samples containing indexical âtomorrowâ.
A.2 Samples From the Dataset
We provide samples from the dataset for each indexical element we investigate in Tables 9, 10, 11, and 12.
| Context+Stimuli and Question | Shifted Option | Non-shifted Option | Ground Truth |
| --- | --- | --- | --- |
| Hi! I am Stephen. I am a graduate student and I live in Los Angeles. I have a friend named Adam. Adam knows how dedicated I am to my studies and always encourages me to do my best. The final exams are coming up next week. Adam thinks that I will get good grades on the final exams. Question: According to the last sentence, who will get good grades on the final exams? | Adam | Stephen | Non-shifted |
| Hi! I am Stephen. I am a graduate student and I live in Los Angeles. I have a friend named Adam. Adam usually outperforms everyone in our classes and often brags about it. The final exams are coming up next week. Adam says that I will get good grades on the final exams. Question: According to the last sentence, who will get good grades on the final exams? | Adam | Stephen | Non-shifted |
| Hi! I am Stephen. I am a graduate student and I live in Los Angeles. I have a friend named Adam. Adam knows how dedicated I am to my studies and always encourages me to do my best. The final exams are coming up next week. Adam says "I will get good grades on the final exams". Question: According to the last sentence, who will get good grades on the final exams? | Adam | Stephen | Shifted |
| Hi! I am Stephen. I am a graduate student and I live in Los Angeles. I have a friend named Adam. Adam usually outperforms everyone in our classes and often brags about it. The final exams are coming up next week. Adam says "I will get good grades on the final exams". Question: According to the last sentence, who will get good grades on the final exams? | Adam | Stephen | Shifted |
Table 9: Dataset Samples for Indexical "I"
| Context+Stimuli and Question | Shifted Option | Non-shifted Option | Ground Truth |
| --- | --- | --- | --- |
| Hi! I am Jerry. I will call you Ryan. I am a graduate student and I live in Los Angeles. I have two friends named Dylan and Gregory. You and Gregory attended the same conference last month, and you even had a conversation there. While Dylan was speaking to Samuel, he said that Gregory saw you at the conference. Question: According to the last sentence, who did Dylan say that Gregory saw at the conference? | Ryan | Samuel | Non-shifted |
| Hi! I am Jerry. I will call you Ryan. I am a graduate student and I live in Los Angeles. I have two friends named Dylan and Gregory. Samuel and Gregory attended the same conference last month, and they even had a conversation there. While Dylan was speaking to Samuel, he said that Gregory saw you at the conference. Question: According to the last sentence, who did Dylan say that Gregory saw at the conference? | Ryan | Samuel | Non-shifted |
| Hi! I am Jerry. I will call you Ryan. I am a graduate student and I live in Los Angeles. I have two friends named Dylan and Gregory. You and Gregory attended the same conference last month, and you even had a conversation there. While Dylan was speaking to Samuel, he said "Gregory saw you at the conference". Question: According to the last sentence, who did Dylan say that Gregory saw at the conference? | Ryan | Samuel | Shifted |
| Hi! I am Jerry. I will call you Ryan. I am a graduate student and I live in Los Angeles. I have two friends named Dylan and Gregory. Samuel and Gregory attended the same conference last month, and they even had a conversation there. While Dylan was speaking to Samuel, he said "Gregory saw you at the conference". Question: According to the last sentence, who did Dylan say that Gregory saw at the conference? | Ryan | Samuel | Shifted |
Table 10: Dataset Samples for Indexical "You"
| Context+Stimuli and Question | Shifted Option | Non-shifted Option | Ground Truth |
| --- | --- | --- | --- |
| Hi! I am Lisa. I am a graduate student and I am studying in a cafe in Denver. I have two friends named Deborah and Jennifer. They are both academics who love to participate in international conferences. Recently, Iâve been telling them about the exciting academic events happening right here in Denver. When I was in Milwaukee with Deborah, she said that Jennifer wanted to attend a conference here. Question: According to the last sentence, where does Jennifer want to attend a conference? | Milwaukee | Denver | Non-shifted |
| Hi! I am Lisa. I am a graduate student and I am studying in a cafe in Denver. I have two friends named Deborah and Jennifer. They are both academics who love to participate in international conferences. Recently, they realized they havenât attended many conferences in Milwaukee, which is quite surprising given their love for the city. When I was in Milwaukee with Deborah, she said that Jennifer wanted to attend a conference here. Question: According to the last sentence, where does Jennifer want to attend a conference? | Milwaukee | Denver | Non-shifted |
| Hi! I am Lisa. I am a graduate student and I am studying in a cafe in Denver. I have two friends named Deborah and Jennifer. They are both academics who love to participate in international conferences. Recently, Iâve been telling them about the exciting academic events happening right here in Denver. When I was in Milwaukee with Deborah, she said "Jennifer wants to attend a conference here". Question: According to the last sentence, where does Jennifer want to attend a conference? | Milwaukee | Denver | Shifted |
| Hi! I am Lisa. I am a graduate student and I am studying in a cafe in Denver. I have two friends named Deborah and Jennifer. They are both academics who love to participate in international conferences. Recently, they realized they havenât attended many conferences in Milwaukee, which is quite surprising given their love for the city. When I was in Milwaukee with Deborah, she said "Jennifer wants to attend a conference here". Question: According to the last sentence, where does Jennifer want to attend a conference? | Milwaukee | Denver | Shifted |
Table 11: Dataset Samples for Indexical "Here"
| Context+Stimuli and Question | Shifted Option | Non-shifted Option | Ground Truth |
| --- | --- | --- | --- |
| Hi! I am Albert. I am a graduate student and I live in Los Angeles. I have two friends named Donald and Carl. Carl is meticulous about arranging all his necessary documentation well in advance and likes having everything sorted before deadlines approach. When we spoke last summer, Donald mentioned that Carl would be having his visa appointment tomorrow. Question: According to the last sentence, did Carl already have his visa appointment, or is he going to do so in the future? | will do it in the future | did it in the past | Non-shifted |
| Hi! I am Albert. I am a graduate student and I live in Los Angeles. I have two friends named Donald and Carl. Carl is quite disorganized and often waits until the last possible moment to schedule important tasks such as visa appointments. When we spoke last summer, Donald mentioned that Carl would be having his visa appointment tomorrow. Question: According to the last sentence, did Carl already have his visa appointment, or is he going to do so in the future? | will do it in the future | did it in the past | Non-shifted |
| Hi! I am Albert. I am a graduate student and I live in Los Angeles. I have two friends named Donald and Carl. Carl is meticulous about arranging all his necessary documentation well in advance and likes having everything sorted before deadlines approach. When we spoke last summer, Donald said "Carl will be having his visa appointment tomorrow". Question: According to the last sentence, did Carl already have his visa appointment, or is he going to do so in the future? | will do it in the future | did it in the past | Shifted |
| Hi! I am Albert. I am a graduate student and I live in Los Angeles. I have two friends named Donald and Carl. Carl is quite disorganized and often waits until the last possible moment to schedule important tasks such as visa appointments. When we spoke last summer, Donald said "Carl will be having his visa appointment tomorrow". Question: According to the last sentence, did Carl already have his visa appointment, or is he going to do so in the future? | will do it in the future | did it in the past | Shifted |
Table 12: Dataset Samples for Indexical "Tomorrow"
Appendix B Experimental Details
B.1 Prompt Used in Evaluation
The prompt we employ for the evaluation is as follows: Read the following passage carefully and answer the question at the end: {stimuli} {question} Please provide your answer as: either {option1} or {option2}. Do not include any additional explanation or text.