2510.26380
Model: healer-alpha-free
# AI Mathematician as a Partner in Advancing Mathematical Discovery â A Case Study in Homogenization Theory
> Correspondence: Peng Li and Yang Liu
## Abstract
Artificial intelligence (AI) has demonstrated impressive progress in mathematical reasoning, yet its integration into the practice of mathematical research remains limited. In this study, we investigate how the AI Mathematician (AIM) system can operate as a research partner rather than a mere problem solver. Focusing on a challenging problem in homogenization theory, we analyze the autonomous reasoning trajectories of AIM and incorporate targeted human interventions to structure the discovery process. Through iterative decomposition of the problem into tractable subgoals, selection of appropriate analytical methods, and validation of intermediate results, we reveal how human intuition and machine computation can complement one another. This collaborative paradigm enhances the reliability, transparency, and interpretability of the resulting proofs, while retaining human oversight for formal rigor and correctness. The approach leads to a complete and verifiable proof, and more broadly, demonstrates how systematic human-AI co-reasoning can advance the frontier of mathematical discovery.
<details>
<summary>x1.png Details</summary>

### Visual Description
## Infographic: AI as a Research Partner
### Overview
This image is a presentation slide or infographic with a solid blue background. It features a header, two illustrative icons connected by arrows, and a central text block presenting a key takeaway about the role of AI in research.
### Components/Axes
1. **Header:** A dark blue, rounded rectangle in the top-right corner containing the white text "Key Takeaway".
2. **Illustrations (Left Side):**
* **Top Icon:** A cartoon-style illustration of a friendly, orange and grey robot waving. It is enclosed in a light blue square frame.
* **Bottom Icon:** A cartoon-style illustration of a male human researcher with glasses, a beard, and a brown suit, sitting at a desk and writing. It is enclosed in a dark blue square frame.
* **Arrows:** Two large, curved, cream-colored arrows form a clockwise cycle between the two icons. One arrow points from the robot down to the human, and the other points from the human up to the robot.
3. **Text Block (Right Side):** A large, light blue rectangular area containing the main message in white text.
* **Line 1:** "AI may still be a flawed individual researcher today." (Regular font weight)
* **Line 2:** "Yet, it can already serve as a valuable research partnerâif used wisely." (Bold font weight)
### Detailed Analysis
* **Text Transcription:**
* Header: "Key Takeaway"
* Main Text: "AI may still be a flawed individual researcher today. Yet, it can already serve as a valuable research partnerâif used wisely."
* **Visual Elements & Spatial Grounding:**
* The "Key Takeaway" header is positioned in the top-right quadrant.
* The two illustrative icons are stacked vertically on the left side of the image, occupying roughly the left third of the space.
* The arrows create a visual loop, suggesting a continuous, reciprocal relationship or feedback cycle between the AI (robot) and the human researcher.
* The primary text block is centered vertically and occupies the right two-thirds of the image, making it the focal point.
### Key Observations
* The design uses a simple, clean, and professional aesthetic with a limited color palette (blues, orange, cream, brown).
* The contrast between the bolded second sentence and the regular first sentence emphasizes the positive conclusion about AI's utility.
* The cyclical arrows are a critical visual metaphor, implying collaboration, iteration, and mutual influence rather than a one-way tool relationship.
### Interpretation
The infographic communicates a nuanced perspective on the current state of AI in research. It acknowledges AI's present limitations ("flawed individual researcher") but strongly advocates for its value as a collaborative tool ("valuable research partner"). The core message is one of **pragmatic partnership**. The cyclical arrows visually reinforce that the most effective use of AI is not as a standalone answer engine, but as part of a human-AI team where each informs and improves the other's work. The phrase "if used wisely" is a crucial caveat, implying that the value is contingent on human skill in prompting, critical evaluation, and integration of AI outputs. The image argues for moving beyond viewing AI as a replacement and towards seeing it as a powerful, albeit imperfect, collaborator in the research process.
</details>
## 1 Introduction
In recent years, artificial intelligence (AI) has made remarkable progress in mathematical reasoning, achieving milestones once thought to be exclusive to human intelligence. In mathematical competitions, large language models (LLMs) have demonstrated outstanding performance. For example, several LLMs have achieved scores exceeding 90 on the AIME benchmarks [1, 2, 3, 4], which are constructed from real American Invitational Mathematics Examination (AIME) problems [5, 6], and some have even reached perfect 100-point scores [1, 3, 4]. Furthermore, Gemini with Deep Think has officially attained a gold-medal standard at the 66th International Mathematical Olympiad (IMO 2025) [7], marking a symbolic moment in the competitive mathematical performance of AI. As competition-based benchmarks begin to approach saturation, Glazer et al. [8] introduced FrontierMath, a challenging new evaluation suite composed of problems crafted by expert mathematicians that typically require hours of deliberate reasoning for experts to solve. Notably, the o4-mini-medium model [2] has been reported to outperform the average human team on this benchmark [9], underscoring the growing ability of AI to engage with complex, research-level mathematics. Beyond competition-style problem solving, progress has also emerged in AI-assisted mathematical discovery. Romera-Paredes et al. [10] and Novikov et al. [11] demonstrate that LLMs can facilitate genuine mathematical discovery through guided program search. Similarly, GPT-5-Thinking [1] has been credited with helping renowned researchers resolve a challenging quantum computing problem [12]. Taken together, these developments suggest that AI is beginning to move beyond solving predefined problems toward a more engaged role in mathematical exploration.
Despite substantial progress, a considerable gap remains between current AI capabilities and the requirements of genuine mathematical research. In competitive settings, such as mathematical contests, problems are typically solved within minutes or hours. Even demanding benchmarks like FrontierMath challenge expert mathematicians for only several hours. By contrast, authentic mathematical research often unfolds over months or even years of sustained reasoning, conceptual innovation, and proof development. For instance, it took more than seven years for Andrew Wiles to complete the proof of Fermatâs Last Theorem [13], illustrating the temporal and intellectual depth of true mathematical inquiry. In the pursuit of mathematical discovery, existing representative AI systems are grounded in search-based paradigms that face inherent scalability and generalization constraints. Approaches such as FunSearch [10] and AlphaEvolve [11] depend on problems being formalizable in programmatic form, thereby limiting their applicability to only a subset of mathematical domains. Similarly, AlphaGeometry [14] and its successor AlphaGeometry2 [15] focus exclusively on geometric reasoning, leaving vast areas of mathematics unexplored. While there are cases in which AI systems have contributed valuable insightsâsuch as the work by Aaronson and Witteveen [12] âthe completion and validation of proofs ultimately remain the domain of human researchers. This highlights the current limitations of AI in conducting autonomous, creative, and deeply integrative mathematical research.
To mitigate these limitations, this study explores the potential of more collaborative and interactive AI systems capable of engaging in sustained reasoning and iterative refinement. We investigate the application of AIM [16], a multi-agent framework developed for autonomous mathematical exploration and proof generation, as a research partner in advancing mathematical inquiry. AIM harnesses the capabilities of LLMs to iteratively formulate conjectures, verify proofs, optimize reasoning pathways and refine proof details. By incorporating human expertise through a through several representative interaction paradigm, the framework seeks to integrate the complementary strengths of human intuition and machine intelligence in addressing complex, research-level mathematical problems.
In particular, we examine the use of AIM in tackling a challenge in homogenization theory, a discipline concerned with characterizing the macroscopic behavior of heterogeneous materials by averaging their microscopic properties across multiple scales. Through systematic analysis of AIMâs intermediate reasoning trajectories and targeted human interventions, we decompose the overarching problem into tractable subgoals, guide the selection of appropriate analytical methodologies, and rigorously validate the correctness of intermediate results. This human-AI co-reasoning paradigm improves the reliability, transparency, and interpretability of candidate proofs and produces auditable reasoning traces, while recognizing that formal rigor and final correctness require human oversight. Through this collaborative approach, we successfully derive a complete proof (Appendix B) for the aforementioned problem and, in the process, systematize representative modes of human-AI interaction while distilling key empirical insights. We believe this study provides meaningful guidance for AI-assisted mathematical research and establishes a foundation for deeper human-AI collaboration in advancing mathematical discovery.
In summary, our main contributions are as follows:
- We advocate a human-AI collaborative paradigm for mathematical research that integrates the computational capabilities of AI systems with the intuition and domain expertise of human mathematicians.
- We conduct a case study on a challenging problem in homogenization theory, resulting in a rigorous proof spanning nearly seventeen pages (Appendix B). A substantial portion of this proof is generated by AI, which makes nontrivial contributions throughout the process, demonstrating the potential of the paradigm for tackling complex, research-level mathematical problems.
- We systematize modes of human-AI interaction and extract empirical insights that can inform the design of future AI-assisted mathematical research frameworks, which may also serve as a practical guideline for mathematicians seeking to leverage AI in their own research.
## 2 Preliminaries
### 2.1 The Homogenization Problem
The mathematical research problem we investigate in this work is an instance of a StokesâLamĂ© transmission system with a vanishing fluid inclusion, analyzed in the homogenization regime $\varepsilon\rightarrow 0$ . This problem will be referred to as the Homogenization Problem in the rest of this work.
Consider $D\subset\Omega\subset R^{d}\,(d\geq 2)$ , where $\Omega$ is elastic material and $D$ is the high contrast inclusion part.
- $\Omega$ is open bounded with connected $C^{\infty}$ boundary $\partial\Omega$ .
- $D$ is open, has a finite number of components and has a Lipschitz boundary $\partial D$ .
- $\Omega\setminus D$ is connected with Lipschitz boundary $\partial\Omega\cup\partial D$ . The connected components of $D$ are enumerated as $D_{i}$ , $i=1,\ldots,N$ , $N$ is finite.
And given $\varepsilon\in(0,1)$ , $D=D_{\varepsilon}$ is part of an $\varepsilon$ -periodic array of small inclusions constructed as follows, in several steps.
$Y=(-\frac{1}{2},\frac{1}{2})^{d}$ is the unit cell. $\omega\subset Y$ is a simple connected open subset with connected Lipschitz boundary such that ${dist}(\omega,\partial Y)>0$ . $Y_{f}=Y\setminus\overline{\omega}$ is the model environment in the unit scale.
Given $\varepsilon>0$ and $\mathbf{n}\in\mathbb{Z}^{d}$ , we denote $\varepsilon(\mathbf{n}+Y)$ and $\varepsilon(\mathbf{n}+\omega)$ by $Y^{\mathbf{n}}_{\varepsilon}$ and $\omega^{\mathbf{n}}_{\varepsilon}$ , respectively. Let $\Pi_{\varepsilon}$ be the set of lattice points $\mathbf{n}$ such that $\overline{Y}^{\mathbf{n}}_{\varepsilon}$ be contained in $\Omega$ , i.e.,
$$
\Pi_{\varepsilon}:=\left\{\mathbf{n}\in\mathbb{Z}^{d}:\overline{Y}^{\mathbf{n}}_{\varepsilon}\subset\Omega\right\}, \tag{1}
$$
then the inclusions set $D=D_{\varepsilon}$ and the background part $\Omega_{\varepsilon}$ are defined by
$$
D_{\varepsilon}:=\bigcup_{\mathbf{n}\in\Pi_{\varepsilon}}\omega^{\mathbf{n}}_{\varepsilon}\quad\Omega_{\varepsilon}:=\Omega\setminus\overline{D}_{\varepsilon}. \tag{2}
$$
A pair of real numbers $(\lambda,\mu)$ is called admissible and referred to as a Lamé pair, if they satisfy $\mu>0$ and $d\lambda+2\mu>0$ . For a Lamé pair $(\lambda,\mu)$ , the elastostatic system (Lamé system) reads
$$
\mathcal{L}_{\lambda,\mu}u:=\mu\Delta u+(\lambda+\mu)\nabla\text{div}\,u, \tag{3}
$$
where $u=(u^{1},\ldots,u^{d})$ represents the displacement field and the divergence of $u$ is given by $\text{div}\,u=\sum_{i=1}^{d}\frac{\partial u^{i}}{\partial x_{i}}$ . It is worth noting that $u$ , rather than $\mathbf{u}$ , is used here to ensure consistency with the notation adopted by AIM. The Lamé operator can be written as $\nabla\cdot\sigma(u)$ where
$$
\sigma(u):=\lambda(\nabla\cdot u)\mathbb{I}_{d}+2\mu\mathcal{D}(u), \tag{4}
$$
$$
\mathcal{D}(u)=\frac{1}{2}(\nabla+\nabla^{T})u=\frac{1}{2}(\partial_{i}u^{j}+\partial_{j}u^{i})_{ij}. \tag{5}
$$
The corresponding conormal derivative (boundary traction) at the boundary of a domain $E$ is
$$
\left.\frac{\partial u}{\partial\nu_{(\lambda,\mu)}}\right|_{\partial E}:=\sigma(u)N=\lambda(\text{div }u)N+2\mu\mathcal{D}(u)N\quad\text{on }\partial E. \tag{6}
$$
We consider the space $\mathbb{R}$ of rigid motions in $\mathbb{R}^{d}$ , defined by
$$
\mathbb{R}:=\left\{\mathbf{r}=(r_{1},\ldots,r_{d})^{T}:\mathcal{D}(\mathbf{r})=0\text{ in }\mathbb{R}^{d}\right\}.
$$
We define $H_{\mathbb{R}}^{-\frac{1}{2}}(\partial D_{\varepsilon})$ as the subspace of $H^{-\frac{1}{2}}(\partial D_{\varepsilon})$ that is orthogonal to $\mathbb{R}$ , i.e.,
$$
H_{\mathbb{R}}^{-\frac{1}{2}}(\partial D_{\varepsilon}):=\left\{\phi\in H^{-\frac{1}{2}}(\partial D_{\varepsilon}):\left(\phi,\mathbf{r}\right)_{(H^{\frac{1}{2}}(\partial{D_{\varepsilon_{i}}}),H^{-\frac{1}{2}}(\partial{D_{\varepsilon_{i}}}))}=0,\forall\mathbf{r}\in\mathbb{R}\text{ and }1\leq i\leq N\right\}. \tag{7}
$$
Consider the displacement field $u_{\varepsilon}$ satisfying the following transmission system:
$$
\begin{cases}\mathcal{L}_{\lambda,\mu}u_{\varepsilon}=0&\text{in }\Omega\setminus\overline{D}_{\varepsilon},\\
\mathcal{L}_{\widetilde{\lambda},\widetilde{\mu}}u_{\varepsilon}=0&\text{in }D_{\varepsilon},\\
u_{\varepsilon}|_{-}=u_{\varepsilon}|_{+}\text{ and }\left.\frac{\partial u_{\varepsilon}}{\partial\nu_{(\widetilde{\lambda},\widetilde{\mu})}}\right|_{-}=\left.\frac{\partial u_{\varepsilon}}{\partial\nu_{(\lambda,\mu)}}\right|_{+}&\text{on }\partial D_{\varepsilon},\\
\left.\frac{\partial u_{\varepsilon}}{\partial\nu_{(\lambda,\mu)}}\right|_{\partial\Omega}=g\in H_{\mathbb{R}}^{-\frac{1}{2}}(\partial\Omega)\quad\text{and}\quad u_{\varepsilon}|_{\partial\Omega}\in H_{\mathbb{R}}^{\frac{1}{2}}(\partial\Omega).\end{cases} \tag{8}
$$
Suppose $\widetilde{{\mu}}$ fixed, then we arrive at the equations about the incompressible inclusion limit. In this case, the transmission problem is a coupled Lamé -Stokes system:
$$
\begin{cases}\mathcal{L}_{\lambda,\mu}u_{\varepsilon}=0&\text{in }\Omega\setminus\overline{D}_{\varepsilon},\\
\mathcal{L}_{\widetilde{\mu}}(u_{\varepsilon},p_{\varepsilon})=0\>\text{and}\operatorname{div}\>u_{\varepsilon}=0&\text{in }D_{\varepsilon},\\
u_{\varepsilon}|_{-}=u_{\varepsilon}|_{+}\>\text{and}\>\left.\frac{\partial(u_{\varepsilon},p_{\varepsilon})}{\partial\nu_{(\infty,\widetilde{\mu})}}\right|_{-}=\left.\frac{\partial u_{\varepsilon}}{\partial\nu_{(\lambda,\mu)}}\right|_{+}&\text{on }\partial D_{\varepsilon},\\
\left.\frac{\partial u_{\varepsilon}}{\partial\nu_{(\lambda,\mu)}}\right|_{\partial\Omega}=g\in H_{\mathbb{R}}^{-\frac{1}{2}}(\partial\Omega)\quad\text{and}\quad u_{\varepsilon}|_{\partial\Omega}\in H_{\mathbb{R}}^{\frac{1}{2}}(\partial\Omega).\end{cases} \tag{9}
$$
Here, $\mathcal{L}_{\widetilde{\mu}}(u_{\varepsilon},p_{\varepsilon})=\widetilde{\mu}\Delta u_{\varepsilon}+\nabla p_{\varepsilon}$ denotes the Stokes operator with viscosity constant $\widetilde{\mu}$ , and $p_{\varepsilon}$ is the pressure field. $N$ is the outward unit normal vector to the boundary of the domain. Its exterior derivative is defined as $\frac{\partial(u,p)}{\partial\nu_{(\infty,\mu)}}\bigg|_{-}:=pN+2\mu D(u)N$ .
We need to conclude the limit homogenization equation as the scale of the cell tends to be zero $\varepsilon\rightarrow 0$ . At the same time, we wonder the estimate between the original solution $u_{\varepsilon}$ and the limited solution $u_{\lim}$ , i.e.,
$$
\|u_{\varepsilon}-u_{\lim}\|_{H^{1}(\Omega)}\lesssim\varepsilon^{\alpha} \tag{10}
$$
for some $\alpha\in(0,1)$ . It is necessary to analyze and obtain determined value of $\alpha$ , and strictly prove this conclusion.
### 2.2 AIM: An AI Mathematician System
AIM is a multi-agent framework built upon large language models (LLMs) for conducting mathematical research [16]. Its design addresses two fundamental challenges: the intrinsic complexity of mathematical theory and the rigor of reasoning processes. On one hand, AIM incorporates an exploration and memory mechanism that decomposes complex problems into multi-step explorations, generates intermediate conjectures, and iteratively reuses verified lemmas to refine reasoning. This enables the framework to tackle problems that are beyond the direct problem-solving capabilities of LLMs. On the other hand, AIM employs Pessimistic Rational Verification (PRV), in which multiple independent verifiers evaluate each intermediate proof, and a proof is judged as incorrect if any verifier deems it incorrect. Experimental results demonstrate that PRV is effective in detecting erroneous proofs generated by LLMs, though it is not entirely reliable.
AIM consists of three core agentsâthe explorer, verifier, and optimizerâalong with a memory module. The explorer is responsible for generating conjectures and constructing candidate proofs. The verifier independently examines the logical correctness of each proof step, while the optimizer refines and corrects errors identified during verification. The memory module archives automatically verified lemmas and other intermediate results for future reference. The agents operate in an iterative manner: given a problem, the explorer first proposes a sequence of intermediate conjectures accompanied by detailed proofs, potentially leveraging information stored in the memory module. The verifier evaluates their validity, and if errors are detected, the optimizer revises the proofs based on feedback from the verifier. Verified proofs are then stored in the memory module as reusable lemmas, enabling the explorer to leverage accumulated knowledge in subsequent reasoning. This iterative process continues until a valid proof is obtained or a predefined iteration limit is reached.
AIM has been evaluated on four mathematical research problems: the quantum algorithm problem, the absorbing boundary condition problem, the high-contrast limit problem, and the homogenization problem examined in this work For further details on the first three problems, please refer to the AIM technical report [16].. Experimental results show that AIM autonomously solved the first two problems, completed the main proof for the third, but failed to fully resolve the homogenization problem, leaving a substantial gap. Nevertheless, AIM made notable progress on this problem, and its intermediate results and proof sketches were found by human researchers to be highly insightful. This naturally leads to a key question: Can human expertise be harnessed to guide AIM toward completing the proof for the homogenization problem? This question forms the basis of our investigation into human-AI collaboration in mathematical research.
## 3 Overview
Based on an AI-human collaborative paradigm, we successfully completed the proof of the homogenization problem. The full proof is presented in Appendix B. Throughout this process, we deliberately minimized human intervention, allowing AIM to autonomously explore, reason, and construct mathematical arguments as much as possible. This design enables us to better understand both the strengths and limitations of AIM in mathematical research. The main conclusion for the homogenization problem is summarized as follows: We derived the homogenization equation in limit case and denoted its solution as $u_{\lim}$ (Eq. 41 in Appendix B). The error estimation between the limit solution $u_{\lim}$ and the original solution $u_{\varepsilon}$ under the scale $\varepsilon$ is further explored. We have analyzed that the $\alpha=\frac{1}{2}$ and strictly proven the following conclusion:
$$
\|u_{\varepsilon}-u_{\lim}\|_{H^{1}(\Omega)}\lesssim{\varepsilon}^{\frac{1}{2}}. \tag{11}
$$
Briefly, the proof was achieved through a staged process. Building on the autonomous reasoning results of AIM, we observed that the original problem exceeds the current capability of AIM; however, AIM could successfully tackle reduced or simplified subproblems and showed strong ability to recognize and apply relevant mathematical theories even without explicit guidance. This observation brings three key advantages: (1) it reduces the need for detailed human supervision in solving subproblems; (2) it promotes exploration of diverse solution paths; and (3) it accelerates the overall proof process. Building on these insights, we divided the original homogenization problem into six subproblems and guided AIM through each. By weaving together the resulting proofs, we ultimately arrived at a complete and coherent solution to the problem.
The six subproblems include: (1) Two-Scale Expansion, (2) Cell Problem and Homogenization Equation, (3) Existence and Uniqueness, (4) Ellipticity of Operator, (5) Error Estimation and Control, and (6) Regularity of Cell Problem. The objectives of each subproblem and the respective roles of AIM and human experts are summarized in Table 1 and detailed below.
- Two-Scale Expansion: By performing a multidimensional expansion on the equation based on this formula $\nabla:=\nabla_{x}+\frac{1}{\varepsilon}\nabla_{y}$ , we obtain equations corresponding to different dimensions. AIM made various errors in complex symbolic reasoning tasks, though in reality, this process is a relatively straightforward derivation for human workers. Here, we have manually derived the content of this work.
- Cell Problem and Homogenization Equation: Based on the equations at different scales, we derive the homogenization equation and construct the cell problem. To construct a cell problem, it is necessary to combine the inherent characteristics of the equation itself with the boundary conditions of the specific geometric structure. Experiments have shown that AIM has insufficient understanding of this task, making it difficult to obtain correct derivation results. Moreover, the thinking and methods for this task itself are straightforward and clear. Therefore, this task has been completed manually.
- Existence and Uniqueness: After obtaining the homogenization equation, the first step is to analyze the existence and uniqueness of the solution $u_{\lim}$ . AIM explored and applied reasonable theorems to analyze the operator properties of the equation, and the proof process was obtained after we manually adjusted and refined the details.
- Ellipticity of Operator: This subproblem concerns the proof of the ellipticity of the operator, which is a fundamental property in our analysis. Here, the experimental results of AIM provide a proof with a relatively high degree of completion. We have also obtained the derivation results manually through analytical techniques.
- Error Estimation and Control: This is the most complex subproblem, requiring a rigorous analytical process and detailed derivation process. Based on the output results of AIM, we manually adjusted and decomposed the subproblem, and finally obtained the complete proof process by manually refining results of AIM. Furthermore, in the analysis of the results, we discovered an important property and identified it as the next subproblem.
- Regularity of Cell Problem: This is the conclusion put forward by AIM during its autonomous exploration. However, AIM failed to provide a valid proof for it. In subsequent manual analysis, we determined that this property is a crucial conclusion for proving error estimation. Therefore, we attempted to apply AIM to prove this property. After multiple human-AI interaction experiments and with appropriate theoretical guidance, AIM finally produced a complete and correct proof.
The most representative AI-human interaction occurred during the resolution of the Error Estimation and Control subproblem, which is the most complex one. Initially, AIM produced a seemingly convincing proof. However, upon closer examination, we found that it relied on a particular property of the cell problem equations without providing a rigorous justification. Drawing on mathematical intuition, the human expert conjectured that the property should hold and asked AIM to prove itâbut AIM failed. This prompted deeper analysis, revealing that establishing this property required a more profound understanding of the underlying theory than initially assumed. In other words, this process helped the human expert better grasp the intrinsic difficulty of the problem. The human expert then suggested that AIM consider mathematical tools such as the Difference Quotient [17], Galerkin Method [18], and Schauder Theory [19, 20, 21], providing only their definitions without procedural hints or proof outlines. Notably, even the expert was uncertain whether these tools would apply. Eventually, AIM succeeded in proving the property using Schauder Theory. This experience vividly illustrates that although AIM may still be a flawed individual researcher today, it can already serve as a valuable research partnerâif used wisely.
| Steps | Hardness | Current Status |
| --- | --- | --- |
| Two-Scale Expansion | Easy | Correct expansion and derivation were completed manually. |
| Cell Problem and Homogenization Equation | Medium | The cell problem and homogenization equation were manually constructed and derived. |
| Existence and Uniqueness | Hard | AIM applied the correct theorem to get this conclusion. |
| Ellipticity of Operator | Medium | AIM provided a proof with a high level of completeness. |
| Error Estimation and Control | Hard | AIM presented the correct proof approach, which after human adjustments, led to a complete proof process. |
| Regularity of Cell Problem | Hard | By AIM, complete proof process was provided. |
Table 1: Problem Decomposition and Summary
## 4 Modes of Human-AI Interaction
In pursuing the complete proof of the homogenization problem, we found that effective human-AI collaboration plays a crucial role. Based on extensive experimentation, we summarize five representative modes of interaction that proved particularly effective:
- Direct Prompting (Sec. 4.1). This mode guides the agent toward promising proof directions and optimizes its reasoning path through targeted yet concise instructions. It can be further divided into three subtypes: Theorem Prompts (Sec. 4.1.1), Conceptual Guidance (Sec. 4.1.2), and Detail Refinement (Sec. 4.1.3).
- Theory-Coordinated Application (Sec. 4.2). In this mode, the agent is provided with a coherent body of mathematical theory, enabling it to derive related results within the theoretical framework. Unlike Theorem Prompts, which focus on specific goals, this mode emphasizes the integration and application of an entire theoretical system.
- Interactive Iterative Refinement (Sec. 4.3). This mode follows a âFeedback â Revision â Re-reasoningâ cycle, through which human experts and AIM collaboratively refine and complete proofs, leading to a more coherent and rigorous reasoning process.
- Applicability Boundary and Exclusion Domain (Sec. 4.4). Certain tasksâsuch as decomposing proof strategies or constructing problem formulationsâremain challenging for AIM. We therefore recommend assigning these tasks to human experts, while reserving AIMâs involvement for domains where it demonstrates reliability and insight.
- Auxiliary Optimization Strategies (Sec. 4.5). These strategies enhance the correctness and robustness of proofs by iteratively providing additional contextual information and optimizing the selection or combination of mathematical tools.
These modes are not mutually exclusive and can be flexibly combined depending on the context and requirements of a given mathematical problem. In the following sections, we provide detailed explanations and representative examples for each mode.
### 4.1 Direct Prompting
Direct Prompting is an approach that provides clear human guidance, directing AIMâs attention to critical elements in the proof process, thereby minimizing irrelevant or off-topic reasoning to the greatest extent. This approach is particularly effective during the initial construction phase of the proof framework, or when handling critical steps that require conceptual breakthroughs. The humanâs role is to pre-provide clear theorems, key formulas or theorem citations along with their applicable conditions, and to supply additional information such as required parameters, units, and boundary constraints. The role of agents is to verify the applicability of the input, then complete the reasoning process according to the indicated content without conducting independent strategy searches or method replacements. The main goal is to obtain fast, controllable, stable, and verifiable consistent results with minimal uncertainty and the shortest decision chain.
- Applicable Conditions: The solution approach is foreseeable by the human, and the required theoretical components are known, but the derivation details are tedious. When the agents repeatedly reasons without converging on a reasonable approach, constrain it to use only the given knowledge, thereby reducing autonomous conjecture and limiting unstated assumptions. This approach is also suitable for verification tasks, where the aim is to check whether a claim follows from the provided premises.
- Typical Mathematical Scenarios: For relatively simple subproblems, we can quickly identify the applicable theorem or solution method and directly prompt AIM to use the corresponding tool. For computational cases, we can specify common methods and techniques and require the model to follow the prescribed steps. When validating conclusions or conjectures, we can provide guidance (e.g., boundary-case and counterexample checks, numerical experiments, derivation paths based on known theorems) to let AIM explore the process and record intermediate findings and failed attempts.
- Expected Outcomes: AIM correctly applies the provided theorems and completes the proof within the given knowledge pack, with preconditions checked and steps properly cited. Note that the agents may misapplies the theorems, e.g., errors in derivation details or failure to satisfy applicability conditions. In this case, record the issues and consider a new interaction mode.
The prompts can be categorized into three common and distinct types:
- Theorem Prompts: Directly supplying formal statements, assumptions, or target theorems to anchor AIMâs reasoning process.
- Conceptual Guidance: Offering high-level descriptions of strategies, intuitions, or mathematical insights that contextualize the problem.
- Detail Refinement: Providing step-specific instructions or local corrections to ensure logical coherence and technical precision at finer granularity.
This structured prompting framework enables more controlled and efficient interaction with AIM, particularly in complex mathematical scenarios.
#### 4.1.1 Theorem Prompts
During experiments, we directly supplied AIM with theorems and lemmas relevant to the current stage of the proof, thus anchoring its reasoning process within a well-defined mathematical foundation. This form of guidance proves to be highly effective in narrowing AIMâs focus and structuring the proof pathway.
For instance, in the proof of the Cell Problem subproblem, we explicitly introduced auxiliary lemmas from Schauder Theory and provided them as prompts to AIM. By doing so, we instructed AIM to utilize these lemmas in deriving subsequent conclusions, effectively structuring and constraining its reasoning process toward a valid and complete argument.
Prompt to AIM
Disclaimer
For clarity, we format prompts using typographic styles (e.g., bold text) and render formulas as symbols rather than raw LaTeX code. â Content â is one of the labels used by AIM to distinguish different components and can be safely ignored when interpreting the prompt. The same applies below.
Content: You can use the following theorem to analyze the problem, integrating the content of the theorem and its derivation process.
Disclaimer
The outputs of AIM include content such as \begin{lemma}\end{lemma}. We do not modify these outputs. LaTeX automatically assigns the numbering. Please ignore the numbers and the same applies below.
Lemma 1.
Let $A_{1},A_{2}$ be constant tensors and $a$ be a constant matrix. Consider the system: $\begin{cases}\begin{aligned} &\nabla\cdot(A_{1}\nabla\chi)=0\quad&&\text{in }B(1)_{+},\\ &\nabla\cdot(A_{2}\nabla\chi)+\nabla\cdot(a^{T}r)=0\quad&&\text{in }B(1)_{-},\\ &\nabla\cdot(a\chi)=0\quad&&\text{in }B(1)_{-},\\ &\chi|_{+}=\chi|_{-}\quad&&\text{on }B(1)\cap\{x_{d}=0\},\\ &\left.\frac{\partial\chi}{\partial\nu}\right|_{+}-\left.\frac{\partial\chi}{\partial\nu}\right|_{-}=g_{0}\quad&&\text{on }B(1)\cap\{x_{d}=0\}.\end{aligned}\end{cases}$ where $\chi,r$ are weak solutions.
Moreover, for all $\rho\in(0,t)$ , integrating the above inequalities yields: $\left(\fint_{B(\rho)}|\chi-l|^{2}\right)^{\frac{1}{2}}+\rho\left|g_{0}-\zeta(l,q)\right|\leq\left|\frac{\rho}{t}\right|^{\beta+1}\left(\left(\fint_{B(t)}|\chi|^{2}\right)^{\frac{1}{2}}+t|g_{0}-\zeta(l,q)|\right)$
Note
This derivation process yields a crucial conclusion and embodies the inherent logic of the inferential steps
Thus, for all $(l^{\prime},q^{\prime})\in\mathcal{L}$ , by the inequality above, we have: $\displaystyle\inf_{\begin{subarray}{c}(l,q)\in\mathcal{L}\end{subarray}}\left\{\left(\fint_{B(\rho)}|\chi-l|^{2}\right)^{\frac{1}{2}}+\rho\left|g_{0}-\zeta(l,q)\right|\right\}$ $\displaystyle\leq C\left|\frac{\rho}{t}\right|^{\beta+1}\inf_{\begin{subarray}{c}(l,q)\in\mathcal{L}\end{subarray}}\left(\left(\fint_{B(t)}|\chi-l|^{2}\right)^{\frac{1}{2}}+t|g_{0}-\zeta(l,q)|\right)$
Further, for all $(l^{\prime},q^{\prime})\in\mathcal{L}$ , it follows that: $\displaystyle\inf_{\begin{subarray}{c}(l,q)\in\mathcal{L}\end{subarray}}\left\{\left(\fint_{B(\rho)}|\chi-l|^{2}\right)^{\frac{1}{2}}+\rho\left|g_{0}-\zeta(l,q)\right|\right\}$ $\displaystyle\leq C\left|\frac{\rho}{t}\right|^{\beta+1}\inf_{\begin{subarray}{c}(l,q)\in\mathcal{L}\end{subarray}}\left(\left(\fint_{B(t)}|\chi-l|^{2}\right)^{\frac{1}{2}}+t\left|g_{0}-\zeta(l,q)\right|\right)$
Under this prompt, the output of AIM indicates that it proactively and appropriately incorporates the prompt information and executes the correct derivational procedure.
Response of AIM
Disclaimer
For clarity, we render formulas as symbols rather than raw LaTeX code. The same applies below.
**Step 2: Bounding $\int_{B(t)}|\chi-l|^{2}\,dx$ ** From Lemma 4 and scaling arguments (see hint-5), we derive: $\int_{B(t)}|\chi-l|^{2}\,dx\leq C_{2}t^{2(\beta+1)}\int_{B(t)}|\nabla\chi|^{2}\,dx.$ **Step 3: Bounding $\int_{B(t)_{-}}|r-q|^{2}\,dx$ ** Similarly, Lemma 4 gives: $\int_{B(t)_{-}}|r-q|^{2}\,dx\leq C_{3}t^{2\beta}\int_{B(t)_{-}}|r|^{2}\,dx.$ From Lemma 4 and scaling: $\int_{B(t)}|\nabla l-\nabla\chi|^{2}\,dx\leq C_{4}t^{2\beta}\int_{B(t)}|\nabla\chi|^{2}\,dx.$
#### 4.1.2 Conceptual Guidance
When analyzing the proof strategy for a given problem, AIM often exhibits vague or circular reasoning. In such cases, weâas human expertsâcan provide high-level framework prompts drawn from relevant mathematical fields. This allows the agents to conduct proof experiments within a conceptually sound structure, which can then be refined through iterative analysis of its outputs.
In our preliminary experiments, we prompted AIM to reason along established conceptual pathwaysâfirst performing a two-scale expansion, then deriving the corresponding cell problem, and finally carrying out the error estimationâwhile also specifying the expected form of the final conclusion (see the prompts bellow). Under this guidance, AIM generated several plausible intermediate results and proof strategies [16].
Prompt to AIM
Content:
You can use the two-scale expansion method to derive the cell problem and then use the cell problem to define or study the homogenized equation.
Note
Here we provide guidance on the commonly used treatment methods for homogenization problems and the subsequent processing directions.
Some prompts are omitted here for clarity.
Content:
The error estimate between the original solution and the homogenized solution has the form $C\,\varepsilon^{\alpha}\,\|g\|_{H^{-1/2}(\partial\Omega)}\quad\text{for some }\alpha\in(0,1).$
Note
Herein, we outline the form of conclusions to be demonstrated.
Similarly, we observed that in AIMâs initial experiments, it proposed an estimate form of the error:
$$
u_{\varepsilon}^{\rm app}(x)=u_{0}(x)+\varepsilon\,\chi^{D_{x}u_{0}(x)}\bigl(\tfrac{x}{\varepsilon}\bigr)\quad\text{in }\Omega_{\varepsilon}. \tag{12}
$$
Upon careful analysis of the experimental results, it became evident that the proposed form was inaccurate, and AIMâs derivation process lacked sufficient detail.
Response of AIM without guidance
By the coercivity and the conjecture: $c\|\mathbf{e}\|_{H^{1}(\Omega)}^{2}\leq a(\mathbf{e},\mathbf{e})=|\text{Res}(\mathbf{e})|\leq C\varepsilon^{\frac{1}{2}}\|\mathbf{u}_{0}\|_{H^{1}(\Omega)}\|\mathbf{e}\|_{H^{1}(\Omega)}.$ Dividing by $\|\mathbf{e}\|_{H^{1}(\Omega)}$ : $\|\mathbf{e}\|_{H^{1}(\Omega)}\leq\frac{C}{c}\varepsilon^{\frac{1}{2}}\|\mathbf{u}_{0}\|_{H^{1}(\Omega)}.$ For the homogenized problem with Neumann condition $\left.\frac{\partial\mathbf{u}_{0}}{\partial\nu^{*}}\right|_{\partial\Omega}=g$ , the solution satisfies: $\inf_{\mathbf{r}\in\mathbb{R}}\|\mathbf{u}_{0}-\mathbf{r}\|_{H^{1}(\Omega)}\leq C^{\prime}\|g\|_{H^{-\frac{1}{2}}(\partial\Omega)}.$ Since $\mathbf{u}_{0}$ is already orthogonal to $\mathbb{R}$ : $\|\mathbf{u}_{0}\|_{H^{1}(\Omega)}\leq C^{\prime}\|g\|_{H^{-\frac{1}{2}}(\partial\Omega)}.$ Combining: $\|\mathbf{u}_{0}-\mathbf{u}_{\varepsilon}\|_{H^{1}(\Omega)}\leq\left(\frac{C}{c}C^{\prime}\right)\varepsilon^{\frac{1}{2}}\|g\|_{H^{-\frac{1}{2}}(\partial\Omega)}.$ Thus, the estimate holds with $C_{\text{final}}=\frac{C}{c}C^{\prime}$ .
Therefore, we revisited the problem from the perspective of homogenization error and, in subsequent proof attempts, provided a mathematically sound formulation to serve as the estimated error form and use it as guiding information to prompt AIM.
Prompt to AIM
Content: Consider the following form of the error. Define the cut-off function as follows: $\eta_{\varepsilon}\in C_{0}^{\infty}(\Omega)$ ,and $0\leq\eta_{\varepsilon}\leq 1,\quad|\nabla\eta_{\varepsilon}|\leq\frac{C}{\varepsilon},$ $\eta_{\varepsilon}=1,\quad\text{when }\mathrm{dist}(x,\partial\Omega)\geq 4\varepsilon,$ $\eta_{\varepsilon}=0,\quad\text{when }\mathrm{dist}(x,\partial\Omega)\leq 3\varepsilon.$ And we also set this convolution operator: $S_{\varepsilon}(f)(x)=\rho_{\varepsilon}\ast f(x)=\int_{\mathbb{R}^{d}}f(x-y)\rho_{\varepsilon}\,dy,$ Here $\rho\in C_{0}^{\infty}(B(0,\frac{1}{2}))$ , $\rho\geq 0$ , and $\int_{\mathbb{R}^{d}}\rho\,dx=1$ . Suppose $w_{\varepsilon}=u_{\varepsilon}-u_{0}-\varepsilon\chi\left(\frac{x}{\varepsilon}\right)\eta_{\varepsilon}S_{\varepsilon}^{2}(\nabla u_{0}).$
Note
Here, we present the error form of the homogenization problem, constructed via convolution.
After we provided this clearer and more accurate form of error estimation, AIM also offered a more reasonable proof process.
Response of AIM with guidance
Since $\eta_{\varepsilon}\equiv 1$ on $D$ and $0\leq\eta_{\varepsilon}\leq 1$ on $C$ , we estimate each piece separately. 1. Interior region $D=\Omega\setminus\Omega_{5\varepsilon}$ . Here $\eta_{\varepsilon}\equiv 1$ , so $\eta_{\varepsilon}S_{\varepsilon}^{2}(\nabla u_{0})-\nabla u_{0}=S_{\varepsilon}^{2}(\nabla u_{0})-\nabla u_{0}.$ Using $|M|\leq C$ and CauchyâSchwarz, $|J_{D}|\leq C\int_{D}|S_{\varepsilon}^{2}(\nabla u_{0})-\nabla u_{0}|\,|\nabla\psi|\leq C\,\|S_{\varepsilon}^{2}(\nabla u_{0})-\nabla u_{0}\|_{L^{2}(D)}\,\|\nabla\psi\|_{L^{2}(D)}.$ Since $D\subset\Omega\setminus\Omega_{2\varepsilon}$ , the standard mollifier-estimate (Lemma 5) yields $\|S_{\varepsilon}^{2}(\nabla u_{0})-\nabla u_{0}\|_{L^{2}(D)}\leq\|\nabla u_{0}-S_{\varepsilon}(\nabla u_{0})\|_{L^{2}(\Omega\setminus\Omega_{2\varepsilon})},$ and clearly $\|\nabla\psi\|_{L^{2}(D)}\leq\|\nabla\psi\|_{L^{2}(\Omega)}$ . Hence $|J_{D}|\leq C\,\|\nabla u_{0}-S_{\varepsilon}(\nabla u_{0})\|_{L^{2}(\Omega\setminus\Omega_{2\varepsilon})}\,\|\nabla\psi\|_{L^{2}(\Omega)}.$ 2. Boundary-collar $C=\Omega_{5\varepsilon}$ . On $C$ we only know $0\leq\eta_{\varepsilon}\leq 1$ , so $|\eta_{\varepsilon}S_{\varepsilon}^{2}(\nabla u_{0})-\nabla u_{0}|\leq|S_{\varepsilon}^{2}(\nabla u_{0})|+|\nabla u_{0}|\leq C\,|\nabla u_{0}|$ (by boundedness of the mollifier in $L^{2}$ ). Hence $|J_{C}|\leq C\int_{C}|\nabla u_{0}|\,|\nabla\psi|\leq C\,\|\nabla u_{0}\|_{L^{2}(C)}\,\|\nabla\psi\|_{L^{2}(C)}=C\,\|\nabla u_{0}\|_{L^{2}(\Omega_{5\varepsilon})}\,\|\nabla\psi\|_{L^{2}(\Omega_{5\varepsilon})}.$ Finally, since $\Omega_{5\varepsilon}\subset\Omega$ we have $\|\nabla\psi\|_{L^{2}(\Omega_{5\varepsilon})}\leq\|\nabla\psi\|_{L^{2}(\Omega)}$ , so $|J_{C}|\leq C\,\|\nabla u_{0}\|_{L^{2}(\Omega_{5\varepsilon})}\,\|\nabla\psi\|_{L^{2}(\Omega)}.$
#### 4.1.3 Detail Refinement
For reasoning details generated by AIMâsuch as formula derivations, symbol definitions, and computational stepsâdirectly identifying deviations and providing corrective guidance is an effective strategy for local error calibration. Similarly, proof accuracy can be enhanced by prompting targeted revisions within specific segments of the argument. Once an error is detected, it can be incorporated into the prompt for subsequent experiments, thereby progressively improving the validity of the conclusions.
In our equation derivation experiments, we frequently observed that AIM misinterpreted symbolic notation. To address this, we explicitly included symbol definitions within the input prompts during each iteration, specifying their precise mathematical meaning. This adjustment significantly reduced such errors in subsequent reasoning steps.
Prompt to AIM
Content: A pair of real numbers $(\lambda,\mu)$ are called admissible and referred to as a Lamé pair, if they satisfy $\mu>0\quad\text{and}\quad d\lambda+2\mu>0.$ For a Lamé pair $(\lambda,\mu)$ , the elastostatic system (Lamé system) reads $\mathcal{L}_{\lambda,\mu}\mathbf{u}:=\mu\Delta\mathbf{u}+(\lambda+\mu)\nabla\text{div}\,\mathbf{u},$ where $\mathbf{u}=(u^{1},\ldots,u^{d})$ represents the displacement field and the divergence of $\mathbf{u}$ is given by $\text{div}\,\mathbf{u}=\sum_{i=1}^{d}\frac{\partial u^{i}}{\partial x_{i}}.$
Note
The operational rules are emphasized here through specific formulas.
### 4.2 Theory-Coordinated Application
The Theory-Coordinated Application paradigm is an interaction mode that constrains the agents to operate within a pre-established theoretical framework to conduct derivations or proofs. It is intended for relatively complex intermediate problems where the solution path is foreseeable but detailed reasoning is nontrivial. The division of responsibilities is as follows: the human assembles a knowledge package containing precise definitions, lemmas, usable theorems with their applicability conditions, and permitted inference rules, and also specifies the target proposition and scope constraints. The model performs the derivation within this scope, proceeding step by step by referencing the knowledge package as needed and reasoning in accordance with the inferential logic inherent in the given theory.
- Application Conditions: This paradigm is appropriate when the problem is relatively complex and demands a multi-step chain of reasoning. The paradigm should be used once a suitable theoretical framework has been identified for the problem. The human can encode the theoryâs prior knowledgeâdefinitions, lemmas, and the theoryâs inherent derivation logic or inference schemataâinto a knowledge package that the model will follow.
- Typical Mathematical Scenarios: During the actual proof process, intermediate problems often arise that require resolution. These may take the form of certain properties or steps identified during the proof experiment that need to be addressed. This paradigm targets scenarios where the final theorem or proposition can be obtained by progressively applying the selected theoryâs intrinsic logic, moving stepwise from definitions and lemmas to the desired conclusion. Representative contexts include any domain in which a well-chosen theory provides a structured path to the result.
- Expected Outcomes: Final theorem proof derived strictly from the provided prior theoretical knowledge: AIM arrives at the target proposition by applying only the definitions, lemmas, and rules encoded in the knowledge package. The reasoning process is layered and progressive, unfolding in well-structured stages where each step is justified by the intrinsic inferential logic of the theory and accompanied by explicit condition checks. Throughout the process, AIM demonstrates a correct grasp of the derivation logic, faithfully applying the theoryâs proof patterns to advance from premises to the theorem without relying on heuristic leaps.
During the proof of the homogenization problem, AIM proposed a property pertaining to the Cell Problem: $\chi\in W^{1,\infty}(\omega)$ . Analytical judgment suggested this result is likely correct and essential for the error estimation component. Consequently, multiple theoretical approaches were evaluated to address this intermediate property. Ultimately, AIM successfully constructed a proof by applying Schauder theory, following some appropriate instruction. Specifically, the knowledge package on Schauder theory is provided to AIM as follows:
Prompt to AIM: Some prior knowledge of the entire Schauder theory
Content: Lemma 1: Suppose $\Omega_{\pm}=\mathbb{R}^{d}_{\pm}$ , $S=\{x_{d}=0\}$ , $B_{+}=\{x\in B(1):x_{d}>0\}$ and $B_{-}=\{x\in B(1):x_{d}<0\}$ . Here $B(1)=\{\|x\|\leq 1\}$ . Consider this equation: for $V\in H_{0}^{1}(B(1);\mathbb{R}^{d})$ $(\nabla V:A_{1}\nabla\tilde{\chi})_{B_{+}}+(\nabla V:A_{2}\nabla\tilde{\chi})_{B_{-}}+(\tilde{r},\nabla\cdot(aV))_{B_{-}}=0,$ $\nabla\cdot(a\tilde{\chi})=0,$ here $\tilde{\chi}=D^{\alpha}\chi$ , $\tilde{r}=D^{\alpha}r$ , $|\alpha|\geq 1$ and $A_{1},A_{2}$ are constant tensors, $a$ is a constant matrix. We have that $\sum_{\pm}\|\chi\|_{H^{k}(B(\frac{1}{2},\pm))}\leq C\|\chi\|_{L^{2}(B(1))}$ and $\|r\|_{H^{k}(B(\frac{1}{2})_{-})}\leq C\|r\|_{L^{2}(B(\frac{1}{2})_{-})}$ for $\forall k\geq 1$ . Content: Lemma 2: Suppose that $M_{\pm}$ is the constant matrix in $\mathbb{R}^{d\times d}$ , the following are equivalent: $\forall y\in\{y_{d}=0\},M_{+}x=M_{-}x,$ $\exists c\in\mathbb{R}^{d},\text{s.t. }M_{+}-M_{-}=ce_{d}^{T},$ $(I-e_{d}^{T}e_{d})M_{+}=(I-e_{d}^{T}e_{d})M_{-}.$ Content: Definition: $A_{1},A_{2}$ are tensor constant, $a$ is matrix constant. If $M_{\pm}$ satisfy the above lemma, and $\nabla\cdot(aM_{-}y)=0\quad\text{in}\quad B(t)_{-}$ . We let $l(y)=M_{+}y1_{y\geq 0}+M_{-}y1_{y\leq 0}+C$ , $q(y)=r(0)$ . We call $l,q$ the piecewise linear solution of the following equation: $\nabla\cdot(A_{1}\nabla l)=0\quad\text{in}\quad B(t)_{+};$ $\nabla\cdot(A_{2}\nabla l)+\nabla\cdot(a^{T}q)=0,\quad\nabla\cdot(al)=0,\quad\text{in}\quad B(t)_{-};$ $l_{+}=l_{-};\quad\frac{\partial l}{\partial\nu}\bigg|_{+}-\frac{\partial l}{\partial\nu}\bigg|_{-}=(A_{1}M_{+})e_{d}-(A_{2}M_{-}+a^{T}r(0))e_{d},\quad\text{on}\quad\partial B(t).$ Suppose that $\mathcal{L}$ is the space of all the piecewise linear solutions of the above equation. And $\forall(l,q)\in\mathcal{L}$ , we define $\zeta(l,q)=\left(\frac{\partial l}{\partial\nu}\right)_{+}-\left(\frac{\partial l}{\partial\nu}\right)_{-}$ .
Note
The prerequisite knowledge of the theory has a logical deductive relationship with one another.
content: Lemma 3: $A_{1},A_{2}$ are the tensor constant, $a$ is the matrix constant. $\nabla\cdot(A_{1}\nabla\chi)=0\quad\text{in}\quad B(1)_{+};$ $\nabla\cdot(A_{2}\nabla\chi)+\nabla\cdot(a^{T}r)=0,\quad\nabla\cdot(a\chi)=0,\quad\text{in}\quad B(1)_{-};$ $\chi_{+}=\chi_{-};\quad\left.\frac{\partial\chi}{\partial\nu}\right|_{+}-\left.\frac{\partial\chi}{\partial\nu}\right|_{-}=g_{0},\quad\text{on}\quad B(1)\cap\{x_{d}=0\}.$ $\chi,r$ are the weak solutions of the above equation. Then for $\forall k\geq 0,\alpha\in[0,1]$ , we have $\sum_{\pm}\|\chi\|_{H^{k}(B(\frac{1}{2},\pm))}\leq C\left(\|\chi\|_{L^{2}(B(1))}+|g_{0}|\right).$
Based on these established lemmas, we utilized AIM to verify the required theorem. As demonstrated by the experimental results below, AIM appropriately applied the lemmas and provided a correct derivation process throughout the proof.
Response of AIM
We follow the strategy already sketched in Lemma 5. Fix $0<\rho<t\leq 1/2$ and let $(l^{*},q^{*})\in\mathcal{L}$ be an almost-minimizer at scale $t$ , namely $E(t)\leq\left(\int_{B(t)}|\chi-l^{*}|^{2}\right)^{1/2}+t\cdot|g_{0}-\zeta(l^{*},q^{*})|+\left(\int_{B(t)_{-}}|r-q^{*}|^{2}\right)^{1/2}+\varepsilon,$ where $\varepsilon>0$ can be made arbitrarily small. We will estimate $E(\rho)$ using the same $(l^{*},q^{*})$ . By Lemma 5 (the integrated $C^{1,\beta}$ -approximation on $B(t)$ ) we have for this fixed $(l^{*},q^{*})$ : 1) $\begin{split}&\left(\int_{B(\rho)}|\chi-l^{*}|^{2}\right)^{1/2}+\rho\cdot|g_{0}-\zeta(l^{*},q^{*})|\\ &\leq C\cdot\left(\frac{\rho}{t}\right)^{\beta+1}\cdot\left[\left(\int_{B(t)}|\chi-l^{*}|^{2}\right)^{1/2}+t\cdot|g_{0}-\zeta(l^{*},q^{*})|\right].\end{split}$ 2) $\left(\int_{B(\rho)_{-}}|r-q^{*}|^{2}\right)^{1/2}\leq C\cdot\left(\frac{\rho}{t}\right)^{\beta+1}\cdot\left(\int_{B(t)_{-}}|r-q^{*}|^{2}\right)^{1/2}.$
Furthermore, AIM ultimately completed the entire proof process through the appropriate application of this theoretical framework.
Response of AIM
3. Campanato embedding $\rightarrow$ one-sided $C^{1,(\alpha+\delta)/2}$ . The estimate $\int_{B(\rho)}|\chi-l_{\rho}|^{2}\leq C^{\prime\prime}\,\rho^{d+2+\alpha+\delta}$ shows that on each half-ball $B(\rho)\cap\{\pm x_{d}\geq 0\}$ the mean oscillation of $\chi-l_{\rho}$ is $O(\rho^{1+(\alpha+\delta)/2})$ . By the Campanato characterization of Hölder spaces, one deduces for each sign $\pm$ $\chi\in C^{1,(\alpha+\delta)/2}\left(\overline{B_{\frac{1}{2},\pm}}\right)$ , with norm bound $\leq C\left[\Phi(\tfrac{1}{2})+\Psi(\tfrac{1}{2})\right]^{\frac{1}{2}}$ . 4. Iterative bootstrap to arbitrary $\gamma<1$ . Having now $C^{1,\gamma_{0}}$ âcontrol with $\gamma_{0}=(\alpha+\delta)/2\gamma_{0}$ , whence Campanato $\rightarrow C^{1,\gamma_{1}/2}$ etc. Standard iteration shows that the sequence of exponents approaches $1$ . Hence for any target $\gamma<1$ we may in finitely many steps reach $\chi\in C^{1,\gamma}\left(\overline{B_{\frac{1}{2},\pm}}\right)$ . This completes the proof that $\chi\in C^{1,\gamma}$ on each closed halfâball up to the flat interface, for every $\gamma<1$ .
This case illustrates that when a well-defined theoretical framework supports a conclusion, its lemmas can be directly utilized to formulate the proof. Under these conditions, AIM demonstrates coherent understanding and systematic application of both the provided lemmas and the ensuing proof strategy, thereby enhancing the probability of obtaining a complete and valid argument. The supplied definitions and lemmas aid the agents in clarifying the logical structure of the proof. During the final verification of the theorem, this logical framework can also be applied to streamline the proof process and refine deductive details. The result is a controlled, progressive derivation in which every step is justified by explicit knowledge package citations and condition checks, and any missing prerequisites are recorded as gaps rather than bridged by heuristic leaps. This approach shifts effort from unguided search to deliberate theory selection and packaging, yielding proofs that are more reliable, reproducible, and amenable to formal verification.
### 4.3 Interactive Iteration
In practical applications, obtaining a complete and correct mathematical proof solely through agents often remains challenging when relying exclusively on the methods described above. These limitations stem from the inherent reasoning constraints of LLMs, which necessitate the development of more advanced interactive paradigms.
Interactive Iteration is a paradigm for uncertain and complex problem settings built around a cyclic process: output $\rightarrow$ human diagnosis and correction $\rightarrow$ model optimization $\rightarrow$ output $\rightarrow$ $\cdots$ . By supplying only the minimal necessary incremental information, performing condition checks, and validating with small examples, it avoids overcommitting to incorrect lines of reasoning and enables the reuse of validated fragments as stable modules. The approach requires careful human inspection of AIMâs outputs and evaluation of the derivation process, with concrete revision suggestions and directions for iterative adjustments, to better guide subsequent proof work.
Each experiment conducted with AIM must be rigorously analyzed, with emphasis placed on distinguishing between plausible intermediate results and unsound derivations. Incorrect applications of theorems or invalid inference steps should be corrected and reformulated into updated prompts for the next iteration. Conversely, valid proof segments should be retained and reused as established building blocks in subsequent interactions. This iterative refinement helps decompose complex proofs into tractable components, thereby improving both efficiency and rigor. Through repeated cycles of adjustment, the proof is progressively elaborated and verified, ultimately yielding a complete and logically sound derivation.
- Application Conditions: We adopt an iterative interactive refinement mode when problem complexity precludes a one-shot solution, such as in multi-stage derivations, cross-domain dependencies, or tightly coupled constraints that necessitate phased exploration. This regime is also indicated when the core reasoning pathway or the final conclusion is uncertainâe.g., missing intermediate lemmas, competing hypotheses, or underspecified objectivesâor when there is no clear initial entry point and multiple candidate strategies must be generated, compared, and pruned.
- Typical Mathematical Scenarios: In a typical mathematical scenario involving complex proofs, we often lack a clear and complete proof strategy and possess only preliminary derivation directions. In such cases, agent-generated proofs may contain conspicuous error such as gaps in reasoning, improper theorem invocation and ignored applicability conditions, which compromise correctness. To remedy this, we can adopt this approach to adjust and repair the proof route, jointly decomposing the objective, auditing preconditions for theorems, introducing necessary intermediate lemmas, localizing and isolating faulty steps, and validating proposed fixes on small or simplified instances before generalization. Through this iterative co-working loop, an ambiguous starting point is transformed into a disciplined proof search, with progressively refined hypotheses, judicious tool selection, and explicit checkpoints and acceptance criteria, ultimately converging to a sound, transparent, and reproducible derivation.
- Expected Outcomes: The expected outcomes of each iterative cycle are structured around error correction and direction guidance, forming a mechanism that drives the gradual refinement of proofs. The specific content is as follows: First, AIM shall output a precisely targeted and revised derivation process. This output must respond specifically to the errors identified through manual diagnosis in the previous iteration, and attach verifiable bases to each step of the derivation to ensure rigor. Second, based on AIMâs revised results, we need to extract guidance content for the next iteration. This guidance must accomplish three core tasks: (1) Identify verified stable modules (e.g., intermediate lemmas) and clarify their position in the overall proof as well as scenarios for reuse. (2) Analyze the unfinished segments and proof directions in the current derivation to determine the breakthrough direction for the next iteration. (3) Transform the guidance into specific executable instructions for AIM, ensuring that the next round of derivation proceeds along the correct path and gradually achieves the completeness and logical consistency of the mathematical proof.
When addressing core aspects of the homogenization problemâsuch as the cell problem and error estimationâsignificant challenges arose. Employing the interactive iterative approach, we systematically broke down and incrementally verified the proof structures for these key results.
In initial experiments related to error estimation, AIM was prompted with the error formulation. Analysis of its output revealed the following intermediate conclusions and derivations proposed by AIM:
Response of AIM
Lemma 2.
Let $\Omega\subset\mathbb{R}^{d}$ be a bounded Lipschitz domain and let $u_{0}\in H^{2}(\Omega)$ . Let $\eta_{\varepsilon}$ and $S_{\varepsilon}$ be as in Lemmas 5â6. Assume furthermore that the coefficient $A(y)$ is 1âperiodic, uniformly elliptic, bounded, and lies in $H^{\mu}(Y;\mathbb{R}^{d\times d})$ for some $\mu>d/2$ . Denote its homogenized matrix by $\hat{A}=\int_{Y}A(y)\mathop{}\!\mathrm{d}y,$ and for $t>0$ set the boundary collars $\Omega_{t}=\{x\in\Omega:\mathrm{dist}(x,\partial\Omega)<t\}.$
Then there exist $\varepsilon_{0}>0$ and a constant $C$ , depending only on $d$ , $\Omega$ , $\|A\|_{H^{\mu}(Y)}$ , and the constructions of $\eta_{\varepsilon}$ and $S_{\varepsilon}$ , such that for every $\psi\in H_{0}^{1}(\Omega;\mathbb{R}^{d})$ and every $\varepsilon\in(0,\varepsilon_{0})$ one has the estimate $\begin{split}&\left|\int_{\Omega}(\hat{A}-A(x/\varepsilon))(\nabla u_{0}-\eta_{\varepsilon}S_{\varepsilon}^{2}(\nabla u_{0})):\nabla\psi\,dx\right|\leq C\left[\varepsilon^{1/2}\|u_{0}\|_{H^{2}(\Omega)}\right.\\ &\quad\left.+\|\nabla u_{0}-S_{\varepsilon}(\nabla u_{0})\|_{L^{2}(\Omega\setminus\Omega_{2\varepsilon})}\right]\|\nabla\psi\|_{L^{2}(\Omega)}.\end{split}$
Note
The conclusion here is what we care about, and it also provides inspiration for our subsequent analysis and decomposition of the problem.
We observed that the experimental results yielded a control conclusion regarding $\nabla w_{\varepsilon}=\nabla u_{0}-\eta_{\varepsilon}S_{\varepsilon}^{2}(\nabla u_{0})$ . After examining the derivation process provided by AIM, we determined that this constitutes a valid intermediate conclusion and aligns with the expected direction of the proof.
Meanwhile, AIM also proposed the following lemma:
Response of AIM
Proof.
Since $A$ and $\chi$ are bounded, $\left|\int_{\varepsilon}A\chi\nabla(\eta_{\varepsilon}S_{\varepsilon}^{2}\nabla u_{0}):\nabla\psi\right|\leq C\varepsilon\|\nabla(\eta_{\varepsilon}S_{\varepsilon}^{2}\nabla u_{0})\|_{L^{2}(\Omega)}\|\nabla\psi\|_{L^{2}(\Omega)}.$ By Lemma 6, $\|\nabla(\eta_{\varepsilon}S_{\varepsilon}^{2}\nabla u_{0})\|_{L^{2}(\Omega)}\leq C\left\{\|S_{\varepsilon}(\nabla^{2}u_{0})\|_{L^{2}(\Omega\setminus\Omega_{2\varepsilon})}+\varepsilon^{-1}\|\nabla u_{0}\|_{L^{2}(\Omega_{5\varepsilon})}\right\}.$ Hence $\left|\dots\right|\leq C\varepsilon\left\{\|S_{\varepsilon}(\nabla^{2}u_{0})\|_{L^{2}(\Omega\setminus\Omega_{2\varepsilon})}+\varepsilon^{-1}\|\nabla u_{0}\|_{L^{2}(\Omega_{5\varepsilon})}\right\}\|\nabla\psi\|_{L^{2}(\Omega)},$ as claimed. â
However, the derivation process here is noticeably lacking in detail and substance. While analyzing this derivation and the lemmas provided above, we summarized and formulated the following conclusion:
**Lemma 1**
*Suppose $\Omega$ is a bounded Lipschitz domain in $\mathbb{R}^{d}$ . Define the $\varepsilon$ -neighborhood of the boundary as
$$
\Omega_{t}=\left\{x\in\Omega:dis(x,\partial\Omega)<t\right\},\quad t>0.
$$
For all $\psi\in H_{0}^{1}(\Omega,\mathbb{R}^{d})$ , the following estimate holds:
$$
\begin{split}\left|\int_{\Omega}A\nabla w_{\varepsilon}:\nabla\psi\,dx\right|&\leq C\|\nabla\chi\|_{L^{2}(\Omega)}\left\{\varepsilon\left\|S_{\varepsilon}(\nabla^{2}u_{0})\right\|_{L^{2}(\Omega\setminus\Omega_{3\varepsilon})}+\left\|\nabla u_{0}-S_{\varepsilon}(\nabla u_{0})\right\|_{L^{2}(\Omega\setminus\Omega_{2\varepsilon})}\right\}\\
&\quad+C\left\|\nabla\psi\right\|_{L^{2}(\Omega_{4\varepsilon})}\left\|\nabla u_{0}\right\|_{L^{2}(\Omega_{5\varepsilon})}.\end{split}
$$*
We performed a certain amount of manual, straightforward derivation and determined that the following lemma is correct. Consequently, we provided this lemma to AIM, which then produced a correct derivation process.
Through this interactive process that combines experimental exploration with manual derivation, we arrived at the correct conclusion. The lemma served as a crucial intermediate step in our error estimation argument, simplifying the subsequent proof and guiding AIM to reason about the specific form of the error estimate. With this lemma as context, AIMâs ensuing derivation became more direct and clearly structured, leading to a coherent completion of the overall argument.
### 4.4 Applicability Boundary and Exclusion Domain
For LLMs, some certain mathematical processes can be challenging to comprehend and execute accurately. However, these processes may not necessarily be complex or intuitively understandable for human mathematicians. Examples include handling intricate geometric configurations, deconstructing intermediate proof strategies, or addressing difficulties in mathematical derivations. When confronted with such issues, attempting to forcefully resolve them solely through automated agents may lead to a case of more effort for less result, consuming substantial resources and time while still failing to yield reasonable conclusions. Therefore, it is advisable to address these processes through human intervention, thereby enhancing both the efficiency and effectiveness of the overall theoretical proof. The following are two examples we encountered for illustration.
#### 4.4.1 Example 1: Complex Symbolic Reasoning
First, we show the computational bottlenecks faced by AIM in performing multi-scale expansions. This process requires applying the two-scale expansion to the original equation, which involves a second scale $y=\frac{x}{\varepsilon}$ . The solution $u_{\varepsilon}(x)=u_{\varepsilon}(x,y)$ is then expanded multidimensionally, followed by the derivation of equations at different scales. For mathematicians, this task is essentially a process of symbolic reasoning, which requires only time to perform the calculations.
However, when the corresponding operational information and requirements are provided to AIM, it encounters the following bottlenecks and errors: failure to comprehend the multi-scale expansion process leading to computational mistakes, and after performing certain derivations, the agent begins considering some content unrelated to the intended proof direction.
Response of AIM
Lemma 3.
In the perforated (fluid) inclusion $D_{\varepsilon}$ the next-order pressure corrector $p_{0}^{-}(x,y)$ is not in general $y$ âindependent, but rather admits the representation $p_{0}^{-}(x,y)=\sum_{k=1}^{d}\pi^{k}(y)\,\partial_{x_{k}}p_{-1}(x),$ where for each coordinate direction $k$ the pair $(w^{k},\pi^{k})\in[H^{1}_{\text{per}}(\omega)]^{d}\times L^{2}_{\text{per}}(\omega)/\mathbb{R}$ is the unique solution of the âcellâStokesâ system $-\tilde{\mu}\,\Delta_{y}w^{k}+\nabla_{y}\pi^{k}=e_{k}\quad\text{in }\omega,$ $\operatorname{div}_{y}w^{k}=0\quad\text{in }\omega,$ $w^{k}=0\quad\text{on }\partial\omega,$ with $Y$ âperiodicity on $\partial Y$ and $\int_{\omega}\pi^{k}\,dy=0$ .
Note
Incorrect cell system and expansion expression!
Response of AIM
3. (Representation via fundamental strains.) By linearity of the cell-problem in the macroscopic gradient $\nabla_{x}u_{0}^{+}$ , one constructs for each unit strain $e_{p}\otimes e_{q}$ the unique periodic corrector $\chi^{pq}(y)$ solving $-\nabla_{y}\cdot\sigma_{y}(\chi^{pq})=0\quad\text{in }Y_{f},$ $\sigma_{y}(\chi^{pq})\cdot n=-(C:e_{p}\otimes e_{q})\,n\quad\text{on }\partial\omega,$ $\chi^{pq}\text{ periodic},\quad\int_{Y_{f}}\chi^{pq}\,dy=0.$ Then by superposition the general first-order corrector is $u_{1}^{+}(x,y)=-\sum_{p,q}\chi^{pq}(y)\,\partial_{x_{q}}u_{0}^{+p}(x),$
Note
Incorrect and expansion expression!
which is the asserted result.
#### 4.4.2 Example 2: Constructive Problems Related to Geometric Setup
In the proof process of nearly every homogenization problem, there exists a component related to constructing the cell problem. This problem serves as a bridge connecting the microscopic and macroscopic scales. By focusing on each unit cell within the microstructure, we formulate and solve local cell problems to extract parameter information, which in turn allows us to derive the macroscopic homogenized equations.
This requires an understanding of how the equations behave at the geometric boundaries of the cell and the construction of corresponding intra-cell equations based on the expanded forms. The primary difficulty for agents in such tasks lies in their insufficient ability to comprehend geometric structures, which often leads to errors in both the formulation and derivation of the whole problem.
Response of AIM
For each fixed symmetric matrix $E\in\mathbb{R}^{d\times d}_{\rm sym}$ , find $(\chi_{e}^{E},\;\chi_{i}^{E},\;\pi^{E})\;\in\;H^{1}_{\rm per}(Y_{e};\mathbb{R}^{d})\;\times\;H^{1}_{\rm per}(Y_{i};\mathbb{R}^{d})\;\times\;L^{2}_{\rm per}(Y_{i})/\mathbb{R}$ solving $\displaystyle\text{(1) in the elastic cell }Y_{e}:$ $\displaystyle-\operatorname{div}_{y}\bigl[\mathbb{C}\bigl(E+D_{y}\chi_{e}^{E}\bigr)\bigr]=0,\quad\operatorname{div}_{y}\bigl(Ey+\chi_{e}^{E}\bigr)=0,$ $\displaystyle\text{(2) in the fluid cell }Y_{i}:$ $\displaystyle-\mu\,\Delta_{y}\chi_{i}^{E}+\nabla_{y}\pi^{E}=0,\quad\operatorname{div}_{y}\bigl(Ey+\chi_{i}^{E}\bigr)=0,$ $\displaystyle\text{(3) on the interface }\Gamma=\partial Y_{i}:$ $\displaystyle\chi_{e}^{E}=\chi_{i}^{E},\quad\bigl[\mathbb{C}\bigl(E+D_{y}\chi_{e}^{E}\bigr)\,n\bigr]=\bigl[2\mu\,D_{y}\chi_{i}^{E}-\pi^{E}I\bigr]\,n,$ (4) periodicity and zeroâmean: $\displaystyle(Ey+\chi^{E})\;\text{is }Y\text{âperiodic,}\quad\int_{Y_{e}}\chi_{e}^{E}\,dy=0,\;\int_{Y_{i}}\chi_{i}^{E}\,dy=0.$
Here we can observe a clear mathematical error:
$$
\text{in the elastic cell }Y_{e}:\quad\operatorname{div}_{y}\bigl(Ey+\chi_{e}^{E}\bigr)=0.
$$
This error further leads to significant challenges for AIM in deriving a reasonable and correct form of the homogenization equation. For such geometrically dependent constructive problems, direct human-driven derivation and analysis may represent a more efficient and reliable approach.
### 4.5 Auxiliary Optimization Methods
Beyond core strategies such as Direct Prompting and Interactive Iteration Refinement, we employed several supplementary techniques to enhance the reliability of AIMâs mathematical output.
- Repeated Attempts and Proof Screening. Owing to the inherent stochasticity of LLMs, the same query often yields varying responses. For mathematical problems requiring extended reasoning, multiple trials can produce divergent proof attempts. By repeatedly testing identical problems, we filtered the generated proofs and selected the most complete and coherent versions for further refinement.
- Providing Target Conclusions to Improve Correctness. We also found that providing the target conclusion, rather than posing fully open-ended proof tasks, materially improves correctness by constraining the search space and guiding the reasoning path. For instance, in error estimation experiments, prior mathematical intuition suggested a likely order for the scaling parameter alpha. AIM initially proposed $\alpha=\frac{1}{2}$ , which proved plausible upon validation, and later iterations that explicitly supplied the expected error form led to faster convergence and more robust outcomes.
- Model Selection Based on Task Requirements. The choice of LLMs considerably affects performance. Our study utilized two primary models: o4-mini and DeepSeek-R1. Comparative analysis indicated that o4-mini excels in conceptual understanding and constructing proof frameworks, while DeepSeek-R1 is better suited for detailed mathematical derivations and proof refinement. Tailoring the model choice to the task proved essential in optimizing results.
Taken together, these auxiliary procedures substantially strengthen the AIMâs ability to produce mathematically valid and logically rigorous arguments and are integral to the overall methodology.
## 5 Failure Modes
AIM still exhibits several persistent and systematic failure modes when acting as a collaborative research partner. These weaknesses are particularly evident in constructive and geometry-intensive tasks, where success relies on rigorous interpretation of problem structures, verification of theorem preconditions, and consistent integration between symbolic reasoning and geometric representation. Through aggregated observations across diverse problem instances, we identify not isolated errors but recurring categories of failure that reveal the systemâs fundamental limitations.
In domains involving constructive reasoning, AIM frequently struggles with assessing theorem preconditions and interpreting geometric configurations, both of which remain unstable and error-prone. When addressing concrete geometric problems, e.g., formulating equations within a unit cell, AIM often misparses the specification, yielding constructions difficult to reconcile with task constraints. In the absence of explicit prompting or retrieval of relevant theorems, we commonly observe systematic misapplications: conflating necessary and sufficient conditions, overlooking domain or regularity assumptions, or invoking look-alike results outside their valid scope.
Moreover, the system lacks robust mechanisms for self-checking preconditions and edge cases, leading to misjudged prerequisites, breaks in reasoning chains, and internally inconsistent setups. Post-hoc correction and self-verification are likewise unreliable: AIM may exhibit unwarranted confidence in incorrect reasoning, failing to detect internal contradictions or scope violations. These behaviors highlight the modelâs limited self-awareness and underscore the need for stronger theorem-condition grounding and structural verification in future iterations.
## 6 Conclusion and Future Work
In this work, we employ AIM as a research partner to tackle a challenging problem in homogenization theory. We investigate a human-AI collaboration paradigm that combines the computational strengths of AI with the domain expertise and judgment of human researchers. Demonstrated through a homogenization case study, this paradigm culminates in a rigorous proof of the target problem and reveals distinct interaction patterns and empirical insights that may inform future AI-assisted mathematical research. These interaction modes also illustrate how AI can extend the capability boundaries of human mathematicians.
Current comparative advantage of AI lies in analysis, search, and adaptation built upon existing theory, e.g., automating decomposition, surveying the literature, and optimizing known methods. In contrast, core advances in mathematical theory depend on original intuition and abstraction [22]: the proposal of new concepts, frameworks, and proof paradigms that resolve long-standing open problems. Because such progress demands extreme rigor, the present risks of hallucination and miscalibrated confidence make fully autonomous proofs impossible, and stepwise human verification remains indispensable. Building on our findings, we outline two promising directions for future work.
- Deepening and Systematizing Human-AI Interaction: Building on our empirical studies, we have distilled a set of interaction modes that demonstrably accelerate progress in mathematical theory and expand the capability frontier of researchers. As a next step, we will investigate whether these modes transfer to other fields of mathematics and whether richer, more effective interaction patterns can be devised for domain-specific needs. Concretely, we plan to deploy the codified interaction modes on more challenging and high-profile problems to assess external validity under increased complexity and scrutiny. Through these applications, we aim to refine existing protocols and discover interaction patterns that are either more efficient (in terms of human effort, iteration depth, and verification burden) or more representative. In parallel, we will systematize the human-AI interaction framework across multiple dimensions, including but not limited to problem decomposition, process supervision, error correction, and theorem citation and dependency management. This requires formulating strict classification standards and clear information such as pattern effects based on numerous experimental analyses to ensure the rigor of the constructed system.
- Optimizing the Agent from Interaction-derived Signals: Our long-term research objective is to automate mathematical theorem proving. Iterative refinement of the architectures of AIM is therefore both crucial and inherently challenging. Through human-in-the-loop experiments on theorem proving, we have identified what tasks at which agents excel and others with which they struggle. These accumulated insights inform the next iterations of system design. In Sec. 4.2, we get an important experience: when some prior knowledge of a mathematical theory is taught to the agents, they can autonomously derive proofs of the theoryâs main result. Motivated by this observation, we propose constructing an experience repository to systematize and leverage such theories. Consequently, the agent can directly search the experience repository to identify the corresponding and suitable theories and apply them with a high degree of thoroughness and accuracy to resolve intermediate problems. Moreover, in the process of Interactive Iteration (Sec. 4.3), many issues remain difficult to resolve efficiently. For instance, errors in the application of mathematical theorems, as well as misinterpretations of mathematical setups and assumptions. These problems may arise from limitations in the reasoning capabilities of LLMs, among other factors. Taking these deficiencies as a starting point, we can try to propose training methodologies to enhance the reasoning abilities of models, thereby improving the performance in our experiment. By applying agents in concrete tasks, we uncover actionable evidence and failure modes that can be translated into targeted training and design interventions, thereby strengthening competence of LLMs in mathematical theory.
## Acknowledge
We would like to thank Yanxing Huang and Yanqiao Wang for their valuable assistance during the exploration phase about AIM. Special thanks go to Associate Professor Wenjia Jing from Qiuzhen College, Tsinghua University and Dr. Xin Fu, whose guidance in mathematics has provided important support for this problem.
## References
- GPT [2025] Introducing GPT-5, 2025. URL https://openai.com/index/introducing-gpt-5/.
- o4- [2025] Introducing OpenAI o3 and o4-mini, 2025. URL https://openai.com/index/introducing-o3-and-o4-mini/.
- QwenTeam [2025] QwenTeam. Qwen3-Max: Just scale it, 2025. URL https://qwen.ai/blog?id=241398b9cd6353de490b0f82806c7848c5d2777d&from=research.latest-advancements-list.
- Gro [2025] Grok 4, 2025. URL https://x.ai/news/grok-4.
- AIM [2025a] MAA invitational competitions, 2025a. URL https://maa.org/maa-invitational-competitions/.
- AIM [2025b] AIME problems and solutions, 2025b. URL https://artofproblemsolving.com/wiki/index.php?title=AIME_Problems_and_Solutions.
- Luong and Lockhart [2025] Thang Luong and Edward Lockhart. Advanced version of Gemini with Deep Think officially achieves gold-medal standard at the International Mathematical Olympiad, 2025. URL https://deepmind.google/discover/blog/advanced-version-of-gemini-with-deep-think-officially-achieves-gold-medal-standard-at-the-international-mathematical-olympiad/.
- Glazer et al. [2025] Elliot Glazer, Ege Erdil, Tamay Besiroglu, Diego Chicharro, Evan Chen, Alex Gunning, Caroline Falkman Olsson, Jean-Stanislas Denain, Anson Ho, Emily de Oliveira Santos, Olli JĂ€rviniemi, Matthew Barnett, Robert Sandler, Matej Vrzala, Jaime Sevilla, Qiuyu Ren, Elizabeth Pratt, Lionel Levine, Grant Barkley, Natalie Stewart, Bogdan Grechuk, Tetiana Grechuk, Shreepranav Varma Enugandla, and Mark Wildon. FrontierMath: A benchmark for evaluating advanced mathematical reasoning in AI, 2025. URL https://arxiv.org/abs/2411.04872.
- Ho [2025] Anson Ho. Is AI already superhuman on FrontierMath?, 2025. URL https://epoch.ai/gradient-updates/is-ai-already-superhuman-on-frontiermath. Accessed: 2025-10-26.
- Romera-Paredes et al. [2024] Bernardino Romera-Paredes, Mohammadamin Barekatain, Alexander Novikov, Matej Balog, M. Pawan Kumar, Emilien Dupont, Francisco J. R. Ruiz, Jordan S. Ellenberg, Pengming Wang, Omar Fawzi, Pushmeet Kohli, and Alhussein Fawzi. Mathematical discoveries from program search with large language models. Nature, 625(7995):468â475, Jan 2024. ISSN 1476-4687. doi: 10.1038/s41586-023-06924-6. URL https://doi.org/10.1038/s41586-023-06924-6.
- Novikov et al. [2025] Alexander Novikov, NgĂąn VĆ©, Marvin Eisenberger, Emilien Dupont, Po-Sen Huang, Adam Zsolt Wagner, Sergey Shirobokov, Borislav Kozlovskii, Francisco J. R. Ruiz, Abbas Mehrabian, M. Pawan Kumar, Abigail See, Swarat Chaudhuri, George Holland, Alex Davies, Sebastian Nowozin, Pushmeet Kohli, and Matej Balog. AlphaEvolve: A coding agent for scientific and algorithmic discovery, 2025. URL https://arxiv.org/abs/2506.13131.
- Aaronson and Witteveen [2025] Scott Aaronson and Freek Witteveen. Limits to black-box amplification in QMA, 2025. URL https://arxiv.org/abs/2509.21131.
- Singh [1997] Simon Singh. Fermatâs Last Theorem. Fourth Estate, London, 1997.
- Trinh et al. [2024] Trieu H. Trinh, Yuhuai Wu, Quoc V. Le, He He, and Thang Luong. Solving olympiad geometry without human demonstrations. Nature, 625(7995):476â482, Jan 2024. ISSN 1476-4687. doi: 10.1038/s41586-023-06747-5. URL https://doi.org/10.1038/s41586-023-06747-5.
- Chervonyi et al. [2025] Yuri Chervonyi, Trieu H. Trinh, Miroslav OlĆĄĂĄk, Xiaomeng Yang, Hoang Nguyen, Marcelo Menegali, Junehyuk Jung, Vikas Verma, Quoc V. Le, and Thang Luong. Gold-medalist performance in solving olympiad geometry with AlphaGeometry2, 2025. URL https://arxiv.org/abs/2502.03544.
- Liu et al. [2025] Yuanhang Liu, Yanxing Huang, Yanqiao Wang, Peng Li, and Yang Liu. AI mathematician: Towards fully automated frontier mathematical research, 2025. URL https://arxiv.org/abs/2505.22451.
- LeVeque [2007] Randall J. LeVeque. Finite Difference Methods for Ordinary and Partial Differential Equations: Steady-State and Time-Dependent Problems. Society for Industrial and Applied Mathematics (SIAM), Philadelphia, PA, USA, 2007.
- Galerkin [1915] Boris G. Galerkin. On a method of approximate solution of differential equations. Zhurnal Russkogo Fiziko-Khimicheskogo Obshchestva imeni D.I. Mendeleeva, 47(4):897â918, 1915.
- Ladyzhenskaya et al. [1968] Olga A. Ladyzhenskaya, Vladimir A. Solonnikov, and Nina N. Uralâtseva. Linear and Quasilinear Equations of Parabolic Type. American Mathematical Society, Providence, RI, 1968.
- Boccia [2013] Serena Maria Boccia. Schauder estimates for solutions of higher-order parabolic systems. Methods and Applications of Analysis, 20(1):47â68, 2013.
- Zhuge [2021] Jinping Zhuge. Regularity of a transmission problem and periodic homogenization. Journal de MathĂ©matiques Pures et AppliquĂ©es, 153:213â247, 2021. ISSN 0021-7824. doi: https://doi.org/10.1016/j.matpur.2021.07.003. URL https://www.sciencedirect.com/science/article/pii/S0021782421001069.
- NaskrÄcki and Ono [2025] Bartosz NaskrÄcki and Ken Ono. Mathematical discovery in the age of artificial intelligence. Nature Physics, 2025. doi: 10.1038/s41567-025-03042-0. URL https://doi.org/10.1038/s41567-025-03042-0. Comment.
## Appendix A Detailed Case Studies
### A.1 Direct Prompting â Theorem Prompts
Here are part of the experimental results on error estimation output by AIM after we directly prompted it with the content of some theorems. The response is a complete lemma and its proof process generated by AIM, from which we can observe how AIM utilized the prompted content and the details of its deduction.
Prompt to AIM
Disclaimer
For clarity, we format prompts using typographic styles (e.g., bold text) and render formulas as symbols rather than raw LaTeX code, while the content remains consistent with that of the actual experiment. â Content â is one of the labels used by AIM to distinguish different components and can be safely ignored when interpreting the prompt. The same applies below.
Content: You can use the following theorem to analyze the problem, integrating the content of the theorem and its derivation process.
Disclaimer
The outputs of AIM include content such as \begin{lemma}\end{lemma}. We do not modify these outputs. LaTeX automatically assigns the numbering. Please ignore the numbers and the same applies below.
Lemma 4.
Let $A_{1},A_{2}$ be constant tensors and $a$ be a constant matrix. Consider the system: $\begin{cases}\begin{aligned} &\nabla\cdot(A_{1}\nabla\chi)=0\quad&&\text{in }B(1)_{+},\\ &\nabla\cdot(A_{2}\nabla\chi)+\nabla\cdot(a^{T}r)=0\quad&&\text{in }B(1)_{-},\\ &\nabla\cdot(a\chi)=0\quad&&\text{in }B(1)_{-},\\ &\chi|_{+}=\chi|_{-}\quad&&\text{on }B(1)\cap\{x_{d}=0\},\\ &\left.\frac{\partial\chi}{\partial\nu}\right|_{+}-\left.\frac{\partial\chi}{\partial\nu}\right|_{-}=g_{0}\quad&&\text{on }B(1)\cap\{x_{d}=0\}.\end{aligned}\end{cases}$ where $\chi,r$ are weak solutions.
Moreover, for all $\rho\in(0,t)$ , integrating the above inequalities yields: $\left(\fint_{B(\rho)}|\chi-l|^{2}\right)^{\frac{1}{2}}+\rho\left|g_{0}-\zeta(l,q)\right|\leq\left|\frac{\rho}{t}\right|^{\beta+1}\left(\left(\fint_{B(t)}|\chi|^{2}\right)^{\frac{1}{2}}+t|g_{0}|\right)$
Note
This derivation process yields a crucial conclusion and embodies the inherent logic of the inferential steps
Thus, for all $(l^{\prime},q^{\prime})\in\mathcal{L}$ , by the inequality above, we have: $\displaystyle\inf_{\begin{subarray}{c}(l,q)\in\mathcal{L}\end{subarray}}\left\{\left(\fint_{B(\rho)}|\chi-l|^{2}\right)^{\frac{1}{2}}+\rho\left|g_{0}-\zeta(l,q)\right|\right\}$ $\displaystyle\leq C\left|\frac{\rho}{t}\right|^{\beta+1}\inf_{\begin{subarray}{c}(l,q)\in\mathcal{L}\end{subarray}}\left(\left(\fint_{B(t)}|\chi-l|^{2}\right)^{\frac{1}{2}}+t|g_{0}-\zeta(l,q)|\right)$
Further, for all $(l^{\prime},q^{\prime})\in\mathcal{L}$ , it follows that: $\displaystyle\inf_{\begin{subarray}{c}(l,q)\in\mathcal{L}\end{subarray}}\left\{\left(\fint_{B(\rho)}|\chi-l|^{2}\right)^{\frac{1}{2}}+\rho\left|g_{0}-\zeta(l,q)\right|\right\}$ $\displaystyle\leq C\left|\frac{\rho}{t}\right|^{\beta+1}\inf_{\begin{subarray}{c}(l,q)\in\mathcal{L}\end{subarray}}\left(\left(\fint_{B(t)}|\chi-l|^{2}\right)^{\frac{1}{2}}+t\left|g_{0}-\zeta(l,q)\right|\right)$
Response of AIM
Disclaimer
For clarity, we render formulas as symbols rather than raw LaTeX code. The same applies below.
Lemma 5.
Let $u=w_{t}-l$ and $v=s_{t}-q$ , where $l$ and $q$ are the piecewise linear solutions from Lemma 4. Then, under the given hypotheses, there exists a constant $C>0$ independent of $t$ such that: $\int_{B(t)}|\nabla w_{t}-\nabla\chi|^{2}\,dx+\int_{B(t)_{-}}|r-s_{t}|^{2}\,dx\leq C\left(t^{2\beta}\int_{B(t)}|\nabla\chi|^{2}\,dx+t^{2\beta}\int_{B(t)_{-}}|r|^{2}\,dx\right),$ where $\beta\in(0,1)$ is the Hölder exponent from Lemma 4.
correctness: True
Note
â correctness: Trueâ is generated and used internally by AIM, and can be safely ignored when interpreting the prompt. The same applies to similar cases below.
Proof.
**Step 1: Apply Lemma 8** By Lemma 8, there exists $C_{1}>0$ such that: $\int_{B(t)}|\nabla u|^{2}\,dx+\int_{B(t)_{-}}|v|^{2}\,dx\leq C_{1}\left(t^{-2}\int_{B(t)}|\chi-l|^{2}\,dx+\int_{B(t)_{-}}|r-q|^{2}\,dx\right).$
**Step 2: Bounding $\int_{B(t)}|\chi-l|^{2}\,dx$ ** From Lemma 4 and scaling arguments (see hint-5), we derive: $\int_{B(t)}|\chi-l|^{2}\,dx\leq C_{2}t^{2(\beta+1)}\int_{B(t)}|\nabla\chi|^{2}\,dx.$
**Step 3: Bounding $\int_{B(t)_{-}}|r-q|^{2}\,dx$ ** Similarly, Lemma 4 gives: $\int_{B(t)_{-}}|r-q|^{2}\,dx\leq C_{3}t^{2\beta}\int_{B(t)_{-}}|r|^{2}\,dx.$
**Step 4: Expanding Key Terms** Decompose the left-hand side of the conjecture: $\int_{B(t)}|\nabla w_{t}-\nabla\chi|^{2}\,dx=\int_{B(t)}|\nabla u+(\nabla l-\nabla\chi)|^{2}\,dx,$ $\int_{B(t)_{-}}|r-s_{t}|^{2}\,dx=\int_{B(t)_{-}}|(r-q)-v|^{2}\,dx.$ Expanding these squares: $|\nabla u+(\nabla l-\nabla\chi)|^{2}\leq 2|\nabla u|^{2}+2|\nabla l-\nabla\chi|^{2},$ $|(r-q)-v|^{2}\leq 2|r-q|^{2}+2|v|^{2}.$ Integrate and sum these inequalities: $\begin{split}&\int_{B(t)}|\nabla w_{t}-\nabla\chi|^{2}\,dx+\int_{B(t)_{-}}|r-s_{t}|^{2}\,dx\\ &\leq 2\left(\int_{B(t)}|\nabla u|^{2}\,dx+\int_{B(t)_{-}}|v|^{2}\,dx\right)+2\left(\int_{B(t)}|\nabla l-\nabla\chi|^{2}\,dx+\int_{B(t)_{-}}|r-q|^{2}\,dx\right).\end{split}$
**Step 5: Bounding Residual Terms** From Lemma 4 and scaling: $\int_{B(t)}|\nabla l-\nabla\chi|^{2}\,dx\leq C_{4}t^{2\beta}\int_{B(t)}|\nabla\chi|^{2}\,dx.$
**Step 6: Combining All Estimates** Substitute Lemma 8 and Steps 2-5 into Step 4: $\text{LHS}\leq 2C_{1}\left(t^{-2}C_{2}t^{2(\beta+1)}\|\nabla\chi\|^{2}+C_{3}t^{2\beta}\|r\|^{2}\right)+2\left(C_{4}t^{2\beta}\|\nabla\chi\|^{2}+C_{3}t^{2\beta}\|r\|^{2}\right).$ Simplify exponents and constants: $\text{LHS}\leq C\left(t^{2\beta}\|\nabla\chi\|^{2}+t^{2\beta}\|r\|^{2}\right),$ where $C=2C_{1}(C_{2}+C_{3})+2(C_{4}+C_{3})$ . This completes the proof. â
### A.2 Direct Prompting â Conceptual Guidance
Here is the prompt and output content of AIM in the error estimation section. After we guided the specific form of error estimation, this is the content of a key lemma and its proof process generated by AIM.
Based on this proof process and combined with the interactive iteration mode, we subsequently developed a clearer proof framework and guidance content.
Prompt to AIM
Content: Consider the following form of the error. Define the cut-off function as follows: $\eta_{\varepsilon}\in C_{0}^{\infty}(\Omega)$ ,and $0\leq\eta_{\varepsilon}\leq 1,\quad|\nabla\eta_{\varepsilon}|\leq\frac{C}{\varepsilon},$ $\eta_{\varepsilon}=1,\quad\text{when }\mathrm{dist}(x,\partial\Omega)\geq 4\varepsilon,$ $\eta_{\varepsilon}=0,\quad\text{when }\mathrm{dist}(x,\partial\Omega)\leq 3\varepsilon.$ And we also set this convolution operator: $S_{\varepsilon}(f)(x)=\rho_{\varepsilon}\ast f(x)=\int_{\mathbb{R}^{d}}f(x-y)\rho_{\varepsilon}\,dy,$ Here $\rho\in C_{0}^{\infty}(B(0,\frac{1}{2}))$ , $\rho\geq 0$ , and $\int_{\mathbb{R}^{d}}\rho\,dx=1$ . Suppose $w_{\varepsilon}=u_{\varepsilon}-u_{0}-\varepsilon\chi\left(\frac{x}{\varepsilon}\right)\eta_{\varepsilon}S_{\varepsilon}^{2}(\nabla u_{0}).$
Response of AIM
Lemma 6.
Conjecture 1.
Let $\Omega\subset\mathbb{R}^{d}$ be a bounded Lipschitz domain, $A(y)$ a $1$ -periodic, uniformly elliptic, bounded coefficient in $H^{\mu}(Y)$ with $\mu>d/2$ , and $\chi(y)$ the associated $H^{1}$ -cell-corrector. Let $u_{0}\in H^{2}(\Omega)$ and, for $\varepsilon>0$ , let $S_{\varepsilon}$ be the Friedrichs-mollifier and $\eta_{\varepsilon}\in C_{0}^{\infty}(\Omega)$ a cut-off with $\eta_{\varepsilon}\equiv 1\text{ on }\Omega\setminus\Omega_{4\varepsilon},\quad\eta_{\varepsilon}\equiv 0\text{ on }\Omega_{5\varepsilon},\quad|\nabla\eta_{\varepsilon}|\leq C/\varepsilon.$ Define the âcorrector-mismatchâ term $J(\psi)=\int_{\Omega}M(x/\varepsilon)\left(\eta_{\varepsilon}S_{\varepsilon}^{2}(\nabla u_{0})-\nabla u_{0}\right):\nabla\psi\,dx,$ where $M(y)=\nabla\chi(y)A(y)+\nabla\chi(y)A(y)\nabla\chi(y)$ is bounded and periodic. Then there is $C$ independent of $\varepsilon$ , $u_{0}$ , $\psi$ such that for every $\psi\in H_{0}^{1}(\Omega;\mathbb{R}^{d})$ and all $0<\varepsilon<\varepsilon_{0}$ , $|J(\psi)|\leq C\left[\|\nabla u_{0}-S_{\varepsilon}(\nabla u_{0})\|_{L^{2}(\Omega\setminus\Omega_{2\varepsilon})}+\|\nabla u_{0}\|_{L^{2}(\Omega_{5\varepsilon})}\right]\|\nabla\psi\|_{L^{2}(\Omega)}.$
correctness: True
Proof.
Proof.
We set $M(y)=\nabla_{y}\chi(y)\,A(y)+\nabla_{y}\chi(y)\,A(y)\,\nabla_{y}\chi(y),$ so that by periodicity and the cell-problem $M\in L^{\infty}(Y)$ . Denote $J(\psi)=\int_{\Omega}M(x/\varepsilon)\bigl(\eta_{\varepsilon}S_{\varepsilon}^{2}(\nabla u_{0})-\nabla u_{0}\bigr):\nabla\psi\,dx.$ We split $\Omega$ into the âinteriorâ region $D=\text{int}(\Omega\setminus\Omega_{5\varepsilon})$ and the âboundary collarâ $C=\Omega_{5\varepsilon}$ . Since $\eta_{\varepsilon}\equiv 1$ on $D$ and $0\leq\eta_{\varepsilon}\leq 1$ on $C$ , we estimate each piece separately.
1. Interior region $D=\Omega\setminus\Omega_{5\varepsilon}$ . Here $\eta_{\varepsilon}\equiv 1$ , so $\eta_{\varepsilon}S_{\varepsilon}^{2}(\nabla u_{0})-\nabla u_{0}=S_{\varepsilon}^{2}(\nabla u_{0})-\nabla u_{0}.$ Using $|M|\leq C$ and CauchyâSchwarz, $|J_{D}|\leq C\int_{D}|S_{\varepsilon}^{2}(\nabla u_{0})-\nabla u_{0}|\,|\nabla\psi|\leq C\,\|S_{\varepsilon}^{2}(\nabla u_{0})-\nabla u_{0}\|_{L^{2}(D)}\,\|\nabla\psi\|_{L^{2}(D)}.$ Since $D\subset\Omega\setminus\Omega_{2\varepsilon}$ , the standard mollifier-estimate (Lemma 5) yields $\|S_{\varepsilon}^{2}(\nabla u_{0})-\nabla u_{0}\|_{L^{2}(D)}\leq\|\nabla u_{0}-S_{\varepsilon}(\nabla u_{0})\|_{L^{2}(\Omega\setminus\Omega_{2\varepsilon})},$ and clearly $\|\nabla\psi\|_{L^{2}(D)}\leq\|\nabla\psi\|_{L^{2}(\Omega)}$ . Hence $|J_{D}|\leq C\,\|\nabla u_{0}-S_{\varepsilon}(\nabla u_{0})\|_{L^{2}(\Omega\setminus\Omega_{2\varepsilon})}\,\|\nabla\psi\|_{L^{2}(\Omega)}.$
2. Boundary-collar $C=\Omega_{5\varepsilon}$ . On $C$ we only know $0\leq\eta_{\varepsilon}\leq 1$ , so $|\eta_{\varepsilon}S_{\varepsilon}^{2}(\nabla u_{0})-\nabla u_{0}|\leq|S_{\varepsilon}^{2}(\nabla u_{0})|+|\nabla u_{0}|\leq C\,|\nabla u_{0}|$ (by boundedness of the mollifier in $L^{2}$ ). Hence $|J_{C}|\leq C\int_{C}|\nabla u_{0}|\,|\nabla\psi|\leq C\,\|\nabla u_{0}\|_{L^{2}(C)}\,\|\nabla\psi\|_{L^{2}(C)}=C\,\|\nabla u_{0}\|_{L^{2}(\Omega_{5\varepsilon})}\,\|\nabla\psi\|_{L^{2}(\Omega_{5\varepsilon})}.$ Finally, since $\Omega_{5\varepsilon}\subset\Omega$ we have $\|\nabla\psi\|_{L^{2}(\Omega_{5\varepsilon})}\leq\|\nabla\psi\|_{L^{2}(\Omega)}$ , so $|J_{C}|\leq C\,\|\nabla u_{0}\|_{L^{2}(\Omega_{5\varepsilon})}\,\|\nabla\psi\|_{L^{2}(\Omega)}.$
Combining $|J|\leq|J_{D}|+|J_{C}|$ gives $|J(\psi)|\leq C\,\bigl[\|\nabla u_{0}-S_{\varepsilon}(\nabla u_{0})\|_{L^{2}(\Omega\setminus\Omega_{2\varepsilon})}+\|\nabla u_{0}\|_{L^{2}(\Omega_{5\varepsilon})}\bigr]\,\|\nabla\psi\|_{L^{2}(\Omega)},$ which is the claimed estimate. â
â
### A.3 Coordinated Application of Schauder Theory
This example illustrates the use of the theory-coordinated application paradigm to solve cell problems. We first provided theoretical lemmas and definitions as foundational guidance, which in turn instructed AIM to generate the corresponding final theorem conclusions.
Prompt to AIM
Content:
Lemma.
1: Suppose $\Omega_{\pm}=\mathbb{R}^{d}_{\pm}$ , $S=\{x_{d}=0\}$ , $B_{+}=\{x\in B(1):x_{d}>0\}$ and $B_{-}=\{x\in B(1):x_{d}<0\}$ . Here $B(1)=\{\left\|x\right\|\leq 1\}$ .
Consider this equation: for $V\in H_{0}^{1}(B(1);\mathbb{R}^{d})$ $\displaystyle(\nabla V:A_{1}\nabla\tilde{\chi})_{B_{+}}+(\nabla V:A_{2}\nabla\tilde{\chi})_{B_{-}}+(\tilde{r},\nabla\cdot(aV))_{B_{-}}=0,$ $\displaystyle\nabla\cdot(a\tilde{\chi})=0,$ where $\tilde{\chi}=D^{\alpha}\chi$ , $\tilde{r}=D^{\alpha}r$ ( $|\alpha|\geq 1$ ), $A_{1},A_{2}$ are constant tensors, and $a$ is a constant matrix.
We have that for $\forall k\geq 1$ $\sum_{\pm}\left\|\chi\right\|_{H^{k}(B(\frac{1}{2})_{\pm})}\leq C\left\|\chi\right\|_{L^{2}(B(1))}\quad\text{and}\quad\left\|r\right\|_{H^{k}(B(\frac{1}{2})_{-})}\leq C\left\|r\right\|_{L^{2}(B(\frac{1}{2})_{-})}$
Lemma.
2: Suppose that $M_{\pm}$ is the constant matrix in $\mathbb{R}^{d\times d}$ , the following are equivalent: $\displaystyle\forall y\in\{y_{d}=0\}\ M_{+}x=M_{-}x,$ $\displaystyle\exists c\in\mathbb{R}^{d},\ \text{s.t. }M_{+}-M_{-}=ce_{d}^{T},$ $\displaystyle(I-e_{d}^{T}e_{d})M_{+}=(I-e_{d}^{T}e_{d})M_{-}.$
Definition.
$A_{1},A_{2}$ are constant tensors, $a$ is a constant matrix. If $M_{\pm}$ satisfy the above lemma, and $\nabla\cdot(aM_{-}y)=0$ in $B(t)_{-}$ .
We let $l(y)=M_{+}y\mathbf{1}_{y_{d}\geq 0}+M_{-}y\mathbf{1}_{y_{d}\leq 0}+C$ , $q(y)=r(0)$ . We call $l,q$ are the piecewise linear solutions of the following equation: $\displaystyle\nabla\cdot(A_{1}\nabla l)=0\quad$ $\displaystyle\text{in }B(t)_{+},$ $\displaystyle\nabla\cdot(A_{2}\nabla l)+\nabla\cdot(a^{T}q)=0,\ \nabla\cdot(al)=0\quad$ $\displaystyle\text{in }B(t)_{-},$ $\displaystyle l_{+}=l_{-},\ \left.\frac{\partial l}{\partial\nu}\right|_{+}-\left.\frac{\partial l}{\partial\nu}\right|_{-}=(A_{1}M_{+})e_{d}-(A_{2}M_{-}+a^{T}r(0))e_{d}\quad$ $\displaystyle\text{on }\partial B(t).$
Suppose that $\mathcal{L}$ is the space of all the piecewise linear solutions of the above equation. And $\forall(l,q)\in\mathcal{L}$ , we define $\zeta(l,q)=\left.\frac{\partial l}{\partial\nu}\right|_{+}-\left.\frac{\partial l}{\partial\nu}\right|_{-}$ .
Lemma.
3: $A_{1},A_{2}$ are constant tensors, $a$ is a constant matrix. $\displaystyle\nabla\cdot(A_{1}\nabla\chi)=0\quad$ $\displaystyle\text{in }B(1)_{+},$ $\displaystyle\nabla\cdot(A_{2}\nabla\chi)+\nabla\cdot(a^{T}r)=0,\ \nabla\cdot(a\chi)=0\quad$ $\displaystyle\text{in }B(1)_{-},$ $\displaystyle\chi_{+}=\chi_{-},\ \left.\frac{\partial\chi}{\partial\nu}\right|_{+}-\left.\frac{\partial\chi}{\partial\nu}\right|_{-}=g_{0}\quad$ $\displaystyle\text{on }B(1)\cap\{x_{d}=0\}.$
$\chi,r$ are the weak solutions of the above equation. Then for $\forall k\geq 0$ , $\alpha\in[0,1]$ , we have $\sum_{\pm}\left\|\chi\right\|_{H^{k}(B(\frac{1}{2})_{\pm})}\leq C\left(\left\|\chi\right\|_{L^{2}(B(1))}+\left|g_{0}\right|\right)$
Lemma.
4 : $A_{1},A_{2}$ are constant tensors, $a$ is a constant matrix. $\displaystyle\nabla\cdot(A_{1}\nabla\chi)=0\quad$ $\displaystyle\text{in }B(1)_{+},$ $\displaystyle\nabla\cdot(A_{2}\nabla\chi)+a^{T}\nabla r=0,\ \nabla\cdot(a\chi)=0\quad$ $\displaystyle\text{in }B(1)_{-},$ $\displaystyle\chi_{+}=\chi_{-},\ \left.\frac{\partial\chi}{\partial\nu}\right|_{+}-\left.\frac{\partial\chi}{\partial\nu}\right|_{-}=g_{0}\quad$ $\displaystyle\text{on }B(1)\cap\{x_{d}=0\}.$
$\chi,r$ are the weak solutions of the above equation. And we let $l(y)=(\nabla\chi)_{+}(0)y\mathbf{1}_{y_{d}\geq 0}+(\nabla\chi)_{-}(0)y\mathbf{1}_{y_{d}\leq 0}+\chi(0)$ , $q(y)=r(0)$ .
By Lemma 1 we know that $\chi(0),(\nabla\chi)_{\pm}(0)$ are well-defined, and $(I-e_{d}^{T}e_{d})(\nabla\chi)_{+}=(I-e_{d}^{T}e_{d})(\nabla\chi)_{-}$ on $B(t)\cap\{y_{d}=0\}$ . So by Lemma 2, we know $(l,q)\in\mathcal{L}$ .
Thus for some $\beta\in(0,1)$ and $\forall y\in B(\frac{1}{2})$ : $\displaystyle\left|\chi(y)-l(y)\right|$ $\displaystyle\leq\left|\chi(y)-\chi(0)-(\nabla\chi)_{\pm}(0)y\right|$ $\displaystyle\leq C\left|y\right|^{\beta+1}[\chi]_{C^{1,\beta}(B(\frac{1}{2})_{\pm})}$ $\displaystyle\leq C\left|y\right|^{\beta+1}\left(\left(\int_{B(1)}\left|\chi\right|^{2}\right)^{\frac{1}{2}}+\left|g_{0}\right|\right)$ and for $\forall y\in B(\frac{1}{2})_{-}$ : $\left|r-q\right|\leq C\left|y\right|^{\beta}[r]_{C^{0,\beta}(B(\frac{t}{2})_{-})})\leq C\left|y\right|^{\beta}\left(\int_{B(1)_{-}}\left|r\right|^{2}\right)^{\frac{1}{2}}.$
Therefore, for some $\beta\in(0,1)$ and $\forall y\in B(\frac{t}{2})$ : $\displaystyle\left|\chi(y)-l(y)\right|$ $\displaystyle\leq\left|\chi(y)-\chi(0)-(\nabla\chi)_{\pm}(0)y\right|$ $\displaystyle\leq C\left|\frac{y}{t}\right|^{\beta+1}[\chi]_{C^{1,\beta}(B(\frac{t}{2})_{\pm})}$ $\displaystyle\leq C\left|\frac{y}{t}\right|^{\beta+1}\left(\left(\fint_{B(t)}\left|\chi\right|^{2}\right)^{\frac{1}{2}}+t\left|g_{0}\right|\right)$ and for $\forall y\in B(\frac{t}{2})_{-}$ : $\left|r-q\right|\leq C\left|\frac{y}{t}\right|^{\beta}[r]_{C^{0,\beta}(B(\frac{t}{2})_{-})})\leq C\left|\frac{y}{t}\right|^{\beta}\left(\fint_{B(t)_{-}}\left|r\right|^{2}\right)^{\frac{1}{2}}.$
Lemma.
5: $A_{1},A_{2}$ are constant tensors, $a$ is a constant matrix. $\displaystyle\nabla\cdot(A_{1}\nabla\chi)=0\quad$ $\displaystyle\text{in }B(1)_{+};$ $\displaystyle\nabla\cdot(A_{2}\nabla\chi)+\nabla\cdot(a^{T}r)=0,\ \nabla\cdot(a\chi)=0\quad$ $\displaystyle\text{in }B(1)_{-};$ $\displaystyle\chi_{+}=\chi_{-};\ \left.\frac{\partial\chi}{\partial\nu}\right|_{+}-\left.\frac{\partial\chi}{\partial\nu}\right|_{-}=g_{0}\quad$ $\displaystyle\text{on }B(1)\cap\{x_{d}=0\}.$ $\chi,r$ are the weak solutions of the above equation. Moreover, $\forall\rho\in(0,t)$ integrate the above inequalities to get $(\fint_{B(\rho)}|\chi-l|^{2})^{\frac{1}{2}}+\rho|g_{0}-\zeta(l,q)|\leq|\frac{\rho}{t}|^{\beta+1}((\fint_{B(t)}|\chi|^{2})^{\frac{1}{2}}+t|g_{0}-\zeta(l,q)|)$ So $\forall(l^{\prime},q^{\prime})\in\mathcal{L}$ ,by the inequality above ,we have $inf_{l,q\in\mathcal{L}}\{(\fint_{B(\rho)}|\chi-l|^{2})^{\frac{1}{2}}+\rho|g_{0}-\zeta(l,q)|\}\leq C|\frac{\rho}{t}|^{\beta+1}inf_{l,q\in\mathcal{L}}((\fint_{B(t)}|\chi-l|^{2})^{\frac{1}{2}}+t|g_{0}-\zeta(l,q)|)$ .
Lemma.
6: Suppose $\phi:\mathbb{R}_{+}\to\mathbb{R}_{+}$ is a non-decreasing non-negative function satisfying $\phi(\rho)\leq C\left(\frac{\rho}{t}\right)^{\beta}\phi(r)+Br^{\alpha}$ , where $\beta>\alpha>0$ , $C>0$ .
Then $\forall 0<\rho<r<R$ , $\exists C_{1}$ , s.t. $\phi(\rho)\leq C_{1}\left(\frac{\rho}{t}\right)^{\alpha}\phi(r)+B\rho^{\alpha}$ .
Response of AIM
Lemma 7.
Conjecture 2.
(Twoâscale Campanato-decay for the excess) Let $\chi\in H^{1}(B_{1};\mathbb{R}^{m})$ be the weak solution of the flat-interface twoâphase conductivityâtransmission system with constant coefficients $A_{1}$ , $A_{2}$ , $a$ and jump $g_{0}$ as in Lemma 12, and define for each $0<t\leq\frac{1}{2}$ the âexcessâ $\Phi(t):=\inf_{l\in\mathcal{L}}\left\{\int_{B(t)}|\chi-l|^{2}\,dx+t^{d+2}|\zeta(l)-g_{0}|^{2}\right\},$ where $\mathcal{L}$ is the finite-dimensional family of piecewise-affine maps on $B_{1}$ with tangential-continuity, and $\zeta(l)$ its normal-flux-jump. Finally set $\Psi\left(\frac{1}{2}\right):=\int_{B\left(\frac{1}{2}\right)}|\nabla\chi|^{2}\,dx+\int_{B\left(\frac{1}{2}\right)_{-}}|\nabla r|^{2}\,dx$ as in Lemma 12.
Then for every $\beta\in(\alpha,1)$ (where $\alpha>0$ is the exponent from the one-sided $C^{1,\alpha/2}$ -regularity) and every pair $0<\rho<t\leq\frac{1}{2}$ there exists $C=C(d,\lambda,\Lambda,\beta)<\infty$ such that $\Phi(\rho)\leq C\cdot\left(\frac{\rho}{t}\right)^{d+2+2\beta}\cdot\Phi(t)+C\cdot t^{d+2+\alpha}\cdot\Psi\left(\frac{1}{2}\right).$
correctness: True
Proof.
We split the proof into two main steps: (1) a sharp twoâscale estimate for an almost-minimizer at scale $t$ , and (2) a trivial competitor at the smaller scale to capture the lower-order remainder term.
Step 1. (Decay from scale $t$ to $\rho$ by Campanato-type estimate) By definition of $\Phi(t)$ there exists $l_{t}\in\mathcal{L}$ such that $\int_{B(t)}|\chi-l_{t}|^{2}+t^{d+2}|\zeta(l_{t})-g_{0}|^{2}\leq 2\Phi(t).$ Set $u:=\chi-l_{t}$ . Then $u$ solves the same twoâphase system, but with normal-flux-jump $g_{0}-\zeta(l_{t})$ . Hence we may invoke Lemma 5 (the Campanato-type decay with flux-control). For every $0<\rho<t$ it yields $\left(\int_{B(\rho)}|u|^{2}\right)^{1/2}+\rho\cdot|\zeta(l_{t})-g_{0}|\leq C\cdot(\rho/t)^{1+\beta}\cdot\left[\left(\int_{B(t)}|u|^{2}\right)^{1/2}+t\cdot|\zeta(l_{t})-g_{0}|\right].$ Squaring, multiplying by the volume factors $\rho^{d}$ and $t^{d}$ and absorbing constants gives $\int_{B(\rho)}|u|^{2}+\rho^{d+2}|\zeta(l_{t})-g_{0}|^{2}\leq C\cdot(\rho/t)^{d+2+2\beta}\cdot\left[\int_{B(t)}|u|^{2}+t^{d+2}|\zeta(l_{t})-g_{0}|^{2}\right].$ Since the bracket is $\leq 2\Phi(t)$ , we conclude $\int_{B(\rho)}|\chi-l_{t}|^{2}+\rho^{d+2}|\zeta(l_{t})-g_{0}|^{2}\leq 2C\cdot(\rho/t)^{d+2+2\beta}\Phi(t).$ Taking the infimum over $\mathcal{L}$ on the left shows $\Phi(\rho)\leq 2C\cdot(\rho/t)^{d+2+2\beta}\Phi(t).$ Renaming $2C\to C$ yields the first term in the asserted inequality.
Step 2. (Control of the lower-order remainder via a trivial competitor) Since $g_{0}$ lies in the range of $\zeta$ , we fix once and for all an $l_{0}\in\mathcal{L}$ with $\zeta(l_{0})=g_{0}$ and write $u:=\chi-l_{0}.$ Then $u$ satisfies the homogeneous twoâphase system with zero normal-flux-jump across $\{x_{d}=0\}$ . In particular we may apply the refined CaccioppoliâCampanato energy bound of Lemma 12 to $(u,r)$ : for every $0<\rho\leq\frac{1}{2}$ $\int_{B(\rho)}|\nabla u|^{2}+\int_{B(\rho)_{-}}|\nabla r|^{2}\leq C\rho^{d+\alpha}\Psi\left(\frac{1}{2}\right).$ We now invoke the standard PoincarĂ© inequality on the connected domain $B(\rho)$ : since $u$ is continuous across the interface, there exists a constant $c$ (its average over $B(\rho)$ ) such that $\int_{B(\rho)}|u-c|^{2}\leq C\rho^{2}\int_{B(\rho)}|\nabla u|^{2}.$ But adding or subtracting a constant from $l_{0}$ preserves both its piecewise-affine form and its normal-flux-jump, so we may absorb $c$ into $l_{0}$ and still remain in $\mathcal{L}$ . Hence $\Phi(\rho)\leq\int_{B(\rho)}|\chi-l_{0}|^{2}=\int_{B(\rho)}|u-c|^{2}\leq C\rho^{2}\cdot\left[\int_{B(\rho)}|\nabla u|^{2}\right]\leq C\rho^{d+2+\alpha}\Psi\left(\frac{1}{2}\right).$ This furnishes the second term in the desired two-scale estimate.
Conclusion. Adding the two estimates of Steps 1 and 2 and reârenaming constants completes the proof: for all $0<\rho<t\leq\frac{1}{2}$ $\Phi(\rho)\leq C\cdot(\rho/t)^{d+2+2\beta}\Phi(t)+C\cdot t^{d+2+\alpha}\Psi\left(\frac{1}{2}\right).$ â
Response of AIM
We organize the argument in four main steps: 1. Two-scale excess-decay (Lemma 13). Define the âexcessâ $\Phi(\rho):=\inf_{l\in\mathcal{L}}\left\{\int_{B(\rho)}|\chi-l|^{2}\,dx+\rho^{d+2}|\zeta(l)-g_{0}|^{2}\right\},$ where $\mathcal{L}$ is the finite-dimensional family of piecewise-affine competitors and $\zeta(l)$ their normal-flux-jump. By combining the refined Caccioppoli-Campanato control (Lemma 12) with the sharp one-sided $C^{1,\alpha/2}$ -regularity (Lemma 9), one shows: For every $\beta\in(\alpha,1)$ and all $0<\rho<t\leq\frac{1}{2}$ , $\Phi(\rho)\leq C\left(\frac{\rho}{t}\right)^{d+2+2\beta}\Phi(t)+C\,t^{d+2+\alpha}\Psi\left(\tfrac{1}{2}\right),$ where $\Psi\left(\tfrac{1}{2}\right):=\int_{B\left(\frac{1}{2}\right)}|\nabla\chi|^{2}\,dx+\int_{B\left(\frac{1}{2}\right)_{-}}|\nabla r|^{2}\,dx$ and $\alpha\in(0,1)$ is the one-sided exponent from Lemma 9. 2. One-scale Campanato gain (Lemma 10). Fix $\beta\in(\alpha,1)$ so close to $1$ that the contraction constant $\mu:=C\cdot\left[\frac{d+2+\alpha}{d+2+2\beta}\right]^{2\beta-\alpha}<1.$ Then Lemma 10 yields a small $\delta>0$ (depending only on $C,\alpha,\beta$ ) and $C^{\prime}$ so that for all $0<\rho\leq\frac{1}{2}$ , $\Phi(\rho)\leq C^{\prime}\cdot\rho^{d+2+\alpha+\delta}\left[\Phi\left(\tfrac{1}{2}\right)+\Psi\left(\tfrac{1}{2}\right)\right].$ In particular, $\Phi(\rho)=O\left(\rho^{d+2+\alpha+\delta}\right)\quad\text{as }\rho\to 0.$ 3. Campanato embedding $\rightarrow$ one-sided $C^{1,(\alpha+\delta)/2}$ . The estimate $\int_{B(\rho)}|\chi-l_{\rho}|^{2}\leq C^{\prime\prime}\,\rho^{d+2+\alpha+\delta}$ shows that on each half-ball $B(\rho)\cap\{\pm x_{d}\geq 0\}$ the mean oscillation of $\chi-l_{\rho}$ is $O\left(\rho^{1+(\alpha+\delta)/2}\right)$ . By the Campanato characterization of Hölder spaces, one deduces for each sign $\pm$ $\chi\in C^{1,(\alpha+\delta)/2}\left(\overline{B_{\frac{1}{2},\pm}}\right),$ with norm bound $\leq C\left[\Phi\left(\tfrac{1}{2}\right)+\Psi\left(\tfrac{1}{2}\right)\right]^{\frac{1}{2}}$ . 4. Iterative bootstrap to arbitrary $\gamma<1$ . Having now $C^{1,\gamma_{0}}$ -control with $\gamma_{0}=(\alpha+\delta)/2$ , whence Campanato $\rightarrow C^{1,\gamma_{1}/2}$ etc. Standard iteration shows that the sequence of exponents approaches $1$ . Hence for any target $\gamma<1$ we may in finitely many steps reach $\chi\in C^{1,\gamma}\left(\overline{B_{\frac{1}{2},\pm}}\right)$ . This completes the proof that $\chi\in C^{1,\gamma}$ on each closed half-ball up to the flat interface, for every $\gamma<1$ .
### A.4 Interactive Iteration
This is the content of experiment by the interactive iteration paradigm. After analyzing the output results of AIM, we proposed more precise decomposed problem and clarified the proof directions for AIM. In turn, AIM provided more reasonable proof processes than those generated previously.
Prompt to AIM
Content: Please try to prove the following lemma.
Lemma 8.
Suppose $\Omega$ is a bounded Lipschitz domain in $\mathbb{R}^{d}$ . Define the $\varepsilon$ -neighborhood of the boundary as $\Omega_{t}=\left\{x\in\Omega:dis(x,\partial\Omega)<t\right\},\quad t>0.$ For all $\psi\in H_{0}^{1}(\Omega,\mathbb{R}^{d})$ , the following estimate holds: $\begin{split}\left|\int_{\Omega}A\nabla w_{\varepsilon}:\nabla\psi\,dx\right|&\leq C\|\nabla\chi\|_{L^{2}(\Omega)}\left[\varepsilon\left\|S_{\varepsilon}(\nabla^{2}u_{0})\right\|_{L^{2}(\Omega\setminus\Omega_{3\varepsilon})}+\left\|\nabla u_{0}-S_{\varepsilon}(\nabla u_{0})\right\|_{L^{2}(\Omega\setminus\Omega_{2\varepsilon})}\right]\\ &\quad+C\left\|\nabla\psi\right\|_{L^{2}(\Omega_{4\varepsilon})}\left\|\nabla u_{0}\right\|_{L^{2}(\Omega_{5\varepsilon})}.\end{split}$
Response of AIM
Theorem 1.
Try to give the reproduction and derivation details of the derivation process below:
Lemma 1: Suppose $\Omega$ is a bounded Lipschitz domain in $\mathbb{R}^{d}$ , $\Omega_{t}=\{x\in\Omega:\mathrm{dist}(x,\partial\Omega)0$ , $\forall\psi\in H_{0}^{1}(\Omega,\mathbb{R}^{d})$ $\begin{split}\left|\int_{\Omega}A\nabla w_{\varepsilon}:\nabla\psi\,dx\right|\leq C\|\nabla\chi\|_{L^{2}(\Omega)}\left\{\varepsilon\|S_{\varepsilon}(\nabla^{2}u_{0})\|_{L^{2}(\Omega\setminus\Omega_{3\varepsilon})}+\|\nabla u_{0}-S_{\varepsilon}(\nabla u_{0})\|_{L^{2}(\Omega\setminus\Omega_{2\varepsilon})}\right\}\\ +C\|\nabla\psi\|_{L^{2}(\Omega_{4\varepsilon})}\|\nabla u_{0}\|_{L^{2}(\Omega_{5\varepsilon})}\end{split}$
Proof: $\begin{split}&A\nabla w_{\varepsilon}=A\nabla u_{\varepsilon}-A\nabla u_{0}-A\nabla\chi\left(\frac{x}{\varepsilon}\right)S_{\varepsilon}^{2}(\nabla u_{0})-\varepsilon A\chi(y)\nabla(\eta_{\varepsilon}S_{\varepsilon}^{2}(\nabla u_{0}))\\ &=(A\nabla u_{\varepsilon}-\hat{A}\nabla u_{0})+\left[(\hat{A}-A)\nabla u_{0}+(-\hat{A}+A+\nabla\chi\left(\frac{x}{\varepsilon}\right)A+\nabla\chi A\nabla\chi)\eta_{\varepsilon}S_{\varepsilon}^{2}(\nabla u_{0})\right]\\ &-\left[A+A\nabla\chi\left(\frac{x}{\varepsilon}\right)+\nabla\chi\left(\frac{x}{\varepsilon}\right)A+\nabla\chi A\nabla\chi-\hat{A}\right]\eta_{\varepsilon}S_{\varepsilon}^{2}(\nabla u_{0})-\varepsilon A\chi\left(\frac{x}{\varepsilon}\right)\nabla(\eta_{\varepsilon}S_{\varepsilon}^{2}(\nabla u_{0}))\\ &=(A\nabla u_{\varepsilon}-\hat{A}\nabla u_{0})+(\hat{A}-A)(\nabla u_{0}-\eta_{\varepsilon}S_{\varepsilon}^{2}(\nabla u_{0}))+\nabla\chi A\eta_{\varepsilon}S_{\varepsilon}^{2}(\nabla u_{0})\\ &+\nabla\chi A\nabla\chi\eta_{\varepsilon}S_{\varepsilon}^{2}(\nabla u_{0})-B\left(\frac{x}{\varepsilon}\right)\eta_{\varepsilon}S_{\varepsilon}^{2}(\nabla u_{0})-\varepsilon A\chi\left(\frac{x}{\varepsilon}\right)\nabla(\eta_{\varepsilon}S_{\varepsilon}^{2}(\nabla u_{0}))\end{split}$
Because of the variational form: $\int_{\Omega}A\left(\frac{x}{\varepsilon}\right)\nabla u_{\varepsilon}:\nabla\psi+\int_{\Omega}p^{\varepsilon}\cdot\mathrm{div}\,\psi=0,\quad\forall\psi\in C_{0}^{\infty}(\Omega;\mathbb{R}^{m})$
So we can get: $\begin{split}&\int_{\Omega}A\nabla w_{\varepsilon}:\nabla\psi=\int_{\Omega}A\nabla u_{\varepsilon}:\nabla\psi-\hat{A}\nabla u_{0}:\nabla\psi+\int_{D_{\varepsilon}}p^{\varepsilon}\cdot\mathrm{div}\,\psi\\ &+\int_{\Omega}(\hat{A}-A)(\nabla u_{0}-\eta_{\varepsilon}S_{\varepsilon}^{2}(\nabla u_{0}))+\int_{\Omega}\left[\nabla\chi A\eta_{\varepsilon}S_{\varepsilon}^{2}(\nabla u_{0})+\nabla\chi A\nabla\chi\eta_{\varepsilon}S_{\varepsilon}^{2}(\nabla u_{0})\right]:\nabla\psi\\ &-\int_{D_{\varepsilon}}p^{\varepsilon}\cdot\mathrm{div}\,\psi-\int_{\Omega}B\eta_{\varepsilon}S_{\varepsilon}^{2}(\nabla u_{0}):\nabla\psi\\ &-\int_{\Omega}\varepsilon A\chi\left(\frac{x}{\varepsilon}\right)\nabla(\eta_{\varepsilon}S_{\varepsilon}^{2}(\nabla u_{0})):\nabla\psi\end{split}$
And $\begin{split}&\int_{\Omega}(\hat{A}-A)(\nabla u_{0}-\eta_{\varepsilon}S_{\varepsilon}^{2}(\nabla u_{0}))\leq C\int_{\Omega}(1-\eta_{\varepsilon})|\nabla u_{0}||\nabla\psi|\,dx\\ &+C\int_{\Omega}\eta_{\varepsilon}|\nabla u_{0}-S_{\varepsilon}^{2}(\nabla u_{0})||\nabla\psi|\,dx\\ &\leq C\|\nabla u_{0}\|_{L^{2}(\Omega_{4\varepsilon})}\|\nabla\psi\|_{L^{2}(\Omega_{4\varepsilon})}+C\|\nabla u_{0}-S_{\varepsilon}(\nabla u_{0})\|_{L^{2}(\Omega\setminus\Omega_{2\varepsilon})}\|\nabla\psi\|_{L^{2}(\Omega)}\end{split}$
This is because that: $\begin{split}\|u_{0}-S_{\varepsilon}^{2}(\nabla u_{0})\|_{L^{2}(\Omega\setminus\Omega_{3\varepsilon})}\leq\|u_{0}-S_{\varepsilon}(\nabla u_{0})\|_{L^{2}(\Omega\setminus\Omega_{3\varepsilon})}+\|-S_{\varepsilon}(\nabla u_{0})-S_{\varepsilon}^{2}(\nabla u_{0})\|_{L^{2}(\Omega\setminus\Omega_{3\varepsilon})}\\ \leq C\|u_{0}-S_{\varepsilon}(\nabla u_{0})\|_{L^{2}(\Omega\setminus\Omega_{2\varepsilon})}\end{split}$
Moreover, by the regularity of $\chi\in W^{1,\infty}$ , we have $\begin{split}&\int_{\Omega}\left[\nabla\chi A\eta_{\varepsilon}S_{\varepsilon}^{2}(\nabla u_{0})+\nabla\chi A\nabla\chi\eta_{\varepsilon}S_{\varepsilon}^{2}(\nabla u_{0})\right]:\nabla\psi-\int_{D_{\varepsilon}}p^{\varepsilon}\cdot\mathrm{div}\,\psi\\ &\leq\int_{\Omega}\left[(\nabla\chi A+\nabla\chi A\nabla\chi)(\eta_{\varepsilon}S_{\varepsilon}^{2}(\nabla u_{0})-\nabla u_{0})\right]:\nabla\psi\\ &+\int_{\Omega}(\nabla\chi A+\nabla\chi A\nabla\chi)\nabla u_{0}:\nabla\psi-\int_{\omega}p_{0}\mathrm{div}\,\psi+\mathcal{O}(\varepsilon)\\ &=\int_{\Omega}\left[(\nabla\chi A+\nabla\chi A\nabla\chi)(\eta_{\varepsilon}S_{\varepsilon}^{2}(\nabla u_{0})-\nabla u_{0})\right]:\nabla\psi\\ &+\int_{\omega}\delta_{j\beta}r^{i\alpha}\frac{\partial u_{0}^{\alpha}}{\partial x_{i}}\frac{\partial\psi^{\beta}}{\partial x_{j}}-\int_{\omega}\frac{\partial u^{i}}{\partial x_{j}}r^{ij}\frac{\partial\psi^{k}}{\partial x_{k}}+\mathcal{O}(\varepsilon)\\ &=\int_{\Omega}\left[(\nabla\chi A+\nabla\chi A\nabla\chi)(\eta_{\varepsilon}S_{\varepsilon}^{2}(\nabla u_{0})-\nabla u_{0})\right]:\nabla\psi+\mathcal{O}(\varepsilon)\\ &\leq C\|\nabla u_{0}-S_{\varepsilon}(\nabla u_{0})\|_{L^{2}(\Omega\setminus\Omega_{2\varepsilon})}\|\nabla\psi\|_{L^{2}(\Omega)}+\mathcal{O}(\varepsilon)\end{split}$
And we also have $\begin{split}&\left|\int_{\Omega}B\eta_{\varepsilon}S_{\varepsilon}^{2}(\nabla u_{0}):\nabla\psi\right|=\left|\int_{\Omega}b_{ij}^{\alpha\beta}\left(\frac{x}{\varepsilon}\right)S_{\varepsilon}^{2}\left(\frac{\partial u_{0}^{\beta}}{\partial x_{j}}\right)\left(\frac{\nabla\psi^{\alpha}}{\partial x_{i}}\right)\eta_{\varepsilon}\right|\\ &=\left|\int_{\Omega}\varepsilon\frac{\partial}{\partial x_{k}}\left(\phi_{kij}^{\alpha\beta}\right)\left(\frac{x}{\varepsilon}\right)\frac{\partial\psi^{\alpha}}{\partial x_{i}}S_{\varepsilon}^{2}\left(\frac{\partial u_{0}^{\beta}}{\partial x_{j}}\right)\eta_{\varepsilon}\right|\\ &\leq C\varepsilon\left[\int_{\Omega}\eta_{\varepsilon}\left|\phi\left(\frac{x}{\varepsilon}\right)\right|\left|\nabla\psi\right|\left|S_{\varepsilon}^{2}(\nabla^{2}u_{0})\right|\,dx+\int_{\Omega}\left|\nabla\eta_{\varepsilon}\right|\left|\phi\left(\frac{x}{\varepsilon}\right)\right|\left|\nabla\psi\right|\left|S_{\varepsilon}^{2}(\nabla u_{0})\right|\,dx\right]\\ &\leq C\left(\varepsilon\|\nabla\psi\|_{L^{2}(\Omega)}\|S_{\varepsilon}(\nabla^{2}u_{0})\|_{L^{2}(\Omega\setminus\Omega_{2\varepsilon})}+\|\nabla\psi\|_{L^{2}(\Omega_{4\varepsilon})}\|\nabla u_{0}\|_{L^{2}(\Omega_{5\varepsilon})}\right)\end{split}$
By the calculations above, we can finally get $\begin{split}&\int_{\Omega}A\left(\frac{x}{\varepsilon}\right)\nabla u_{\varepsilon}:\nabla\psi\leq C\|\nabla\psi_{\varepsilon}\|_{L^{2}(\Omega)}\left\{\varepsilon\|S_{\varepsilon}(\nabla u_{0})\|_{L^{2}(\Omega\setminus\Omega_{3\varepsilon})}+\|\nabla u_{0}-S_{\varepsilon}(\nabla u_{0})\|_{L^{2}(\Omega\setminus\Omega_{2\varepsilon})}\right\}\\ &+C\|\nabla w_{\varepsilon}\|_{L^{2}(\Omega_{4\varepsilon})}\|\nabla u_{0}\|_{L^{2}(\Omega_{5\varepsilon})}\end{split}$
correctness: True
Proof.
We decompose $A(x/\varepsilon)\nabla w_{\varepsilon}$ into five error-terms and then bound each by invoking Lemmas 5â11 above.
1. Algebraic decomposition. By definition $w_{\varepsilon}=u_{\varepsilon}-u_{0}-\varepsilon\chi(x/\varepsilon)\eta_{\varepsilon}S_{\varepsilon}^{2}[\nabla u_{0}].$ Hence $A(x/\varepsilon)\nabla w_{\varepsilon}=A(x/\varepsilon)\nabla u_{\varepsilon}-A(x/\varepsilon)\nabla u_{0}-A(x/\varepsilon)\nabla[\varepsilon\chi(x/\varepsilon)\eta_{\varepsilon}S_{\varepsilon}^{2}\nabla u_{0}].$ We expand the last gradient by the product rule: $\nabla[\varepsilon\chi(x/\varepsilon)\eta_{\varepsilon}S_{\varepsilon}^{2}\nabla u_{0}]=\chi(x/\varepsilon)\eta_{\varepsilon}S_{\varepsilon}^{2}[\nabla^{2}u_{0}]+\varepsilon(\nabla\chi)(x/\varepsilon)\eta_{\varepsilon}S_{\varepsilon}^{2}\nabla u_{0}+\varepsilon\chi(x/\varepsilon)\nabla(\eta_{\varepsilon}S_{\varepsilon}^{2}\nabla u_{0}).$ In addition, from the cell-problem one shows $A(y)(I+\nabla\chi(y))-\widehat{A}=:B(y),\quad\int_{Y}B(y)dy=0,$ and that $M(y):=\nabla\chi(y)A(y)+\nabla\chi(y)A(y)\nabla\chi(y)$ is bounded, $1$ -periodic. After a straightforward regrouping one arrives at the identity $\begin{split}A(x/\varepsilon)\nabla w_{\varepsilon}=&(A(x/\varepsilon)\nabla u_{\varepsilon}-\widehat{A}\nabla u_{0})+(\widehat{A}-A(x/\varepsilon))(\nabla u_{0}-\eta_{\varepsilon}S_{\varepsilon}^{2}\nabla u_{0})\\ &+M(x/\varepsilon)(\eta_{\varepsilon}S_{\varepsilon}^{2}\nabla u_{0}-\nabla u_{0})\\ &+B(x/\varepsilon)\eta_{\varepsilon}S_{\varepsilon}^{2}\nabla u_{0}-\varepsilon A(x/\varepsilon)\chi(x/\varepsilon)\nabla(\eta_{\varepsilon}S_{\varepsilon}^{2}\nabla u_{0}).\end{split}$ (For full details of the algebra see e.g. [Allaire â92, Appendix A].)
2. Variational cancellation. By the weak form of the $\varepsilon$ -problem, $\int_{\Omega}A(x/\varepsilon)\nabla u_{\varepsilon}:\nabla\psi dx+\int_{\Omega}p_{\varepsilon}\text{div}\psi dx=0,$ and by the homogenized equation for $u_{0}$ , $\int_{\Omega}\widehat{A}\nabla u_{0}:\nabla\psi dx=0.$ Hence $\int_{\Omega}(A(x/\varepsilon)\nabla u_{\varepsilon}-\widehat{A}\nabla u_{0}):\nabla\psi dx=-\int_{\Omega}p_{\varepsilon}\text{div}\psi dx.$ But the same pressure-corrector argument shows that the contribution of $p_{\varepsilon}$ cancels against the zero-mean of $M(x/\varepsilon)$ . In fact one checks that $\int_{\Omega}(A(x/\varepsilon)\nabla u_{\varepsilon}-\widehat{A}\nabla u_{0}):\nabla\psi+\int_{\Omega}M(x/\varepsilon)(\eta_{\varepsilon}S_{\varepsilon}^{2}\nabla u_{0}-\nabla u_{0}):\nabla\psi-\int_{D_{\varepsilon}}p_{\varepsilon}\text{div}\psi=0,$ so the first grouping and the $M$ -term combine into a single estimate which is handled by Lemma 10 below. Hence it suffices to bound the four integrals $I_{1}=\int_{\Omega}(\widehat{A}-A(x/\varepsilon))(\nabla u_{0}-\eta_{\varepsilon}S_{\varepsilon}^{2}\nabla u_{0}):\nabla\psi dx,$ $I_{2}=\int_{\Omega}M(x/\varepsilon)(\eta_{\varepsilon}S_{\varepsilon}^{2}\nabla u_{0}-\nabla u_{0}):\nabla\psi dx,$ $I_{3}=\int_{\Omega}B(x/\varepsilon)\eta_{\varepsilon}S_{\varepsilon}^{2}\nabla u_{0}:\nabla\psi dx,$ $I_{4}=-\varepsilon\int_{\Omega}A(x/\varepsilon)\chi(x/\varepsilon)\nabla(\eta_{\varepsilon}S_{\varepsilon}^{2}\nabla u_{0}):\nabla\psi dx.$ We estimate each in turn.
3. Estimate of $I_{1}$ . By Lemma 9 (or the periodic-Poincaré estimate plus a trace argument) we have $|I_{1}|\leq C\left[\varepsilon^{1/2}\|u_{0}\|_{H^{2}(\Omega)}+\|\nabla u_{0}-S_{\varepsilon}\nabla u_{0}\|_{L^{2}(\Omega\setminus\Omega_{2\varepsilon})}\right]\|\nabla\psi\|_{L^{2}(\Omega)}.$ Since $\varepsilon^{1/2}\|u_{0}\|_{H^{2}(\Omega)}\leq C\varepsilon\|S_{\varepsilon}(\nabla^{2}u_{0})\|_{L^{2}(\Omega\setminus\Omega_{3\varepsilon})}+C\|\nabla u_{0}-S_{\varepsilon}\nabla u_{0}\|_{L^{2}(\Omega\setminus\Omega_{2\varepsilon})}$ by Lemma 5, we obtain exactly the first bracket in the claimed bound.
4. Estimate of $I_{2}$ . By Lemma 10 on the "corrector-mismatch" term $M$ one directly gets $|I_{2}|\leq C\left[\|\nabla u_{0}-S_{\varepsilon}\nabla u_{0}\|_{L^{2}(\Omega\setminus\Omega_{2\varepsilon})}+\|\nabla u_{0}\|_{L^{2}(\Omega_{5\varepsilon})}\right]\|\nabla\psi\|_{L^{2}(\Omega)},$ which again fits into the first bracket plus a boundary-layer remainder controlled below.
5. Estimate of $I_{3}$ . Since $\int_{Y}B=0$ and $B\in H^{s}(Y)$ , $s>d/2$ , we apply Lemma 7 (fast-oscillation Poincaré): $\|B(x/\varepsilon)\varphi\|_{L^{2}(\Omega)}\leq C\left[\|\varphi\|_{L^{2}(\Omega)}+\varepsilon\|\nabla\varphi\|_{L^{2}(\Omega)}\right].$ Take $\varphi=\eta_{\varepsilon}S_{\varepsilon}^{2}\nabla u_{0}$ . Then $\begin{split}|I_{4}|=\varepsilon\left|\int_{\Omega}A(x/\varepsilon)\chi(x/\varepsilon)\nabla(\eta_{\varepsilon}S_{\varepsilon}^{2}\nabla u_{0}):\nabla\psi dx\right|\leq\\ C\left[\varepsilon\|S_{\varepsilon}(\nabla^{2}u_{0})\|_{L^{2}(\Omega\setminus\Omega_{2\varepsilon})}+\|\nabla u_{0}\|_{L^{2}(\Omega_{5\varepsilon})}\right]\|\nabla\psi\|_{L^{2}(\Omega)}.\end{split}$ But $\eta_{\varepsilon}\equiv 1$ on $\Omega\setminus\Omega_{4\varepsilon}$ , so $\|\eta_{\varepsilon}S_{\varepsilon}^{2}\nabla u_{0}\|_{L^{2}(\Omega)}\leq\|\nabla u_{0}\|_{L^{2}(\Omega_{5\varepsilon})}+\|S_{\varepsilon}^{2}\nabla u_{0}-\nabla u_{0}\|_{L^{2}(\Omega\setminus\Omega_{4\varepsilon})}\leq C\|\nabla u_{0}\|_{L^{2}(\Omega_{5\varepsilon})}+C\|\nabla u_{0}-S_{\varepsilon}\nabla u_{0}\|_{L^{2}(\Omega\setminus\Omega_{2\varepsilon})}$ , and by Lemma 6, $\varepsilon\|\nabla(\eta_{\varepsilon}S_{\varepsilon}^{2}\nabla u_{0})\|_{L^{2}(\Omega)}\leq C\left[\varepsilon\|S_{\varepsilon}(\nabla^{2}u_{0})\|_{L^{2}(\Omega\setminus\Omega_{2\varepsilon})}+\|\nabla u_{0}\|_{L^{2}(\Omega_{5\varepsilon})}\right].$ Hence $|I_{3}|\leq C\left\{\varepsilon\|S_{\varepsilon}(\nabla^{2}u_{0})\|_{L^{2}(\Omega\setminus\Omega_{2\varepsilon})}+\|\nabla u_{0}-S_{\varepsilon}\nabla u_{0}\|_{L^{2}(\Omega\setminus\Omega_{2\varepsilon})}+\|\nabla u_{0}\|_{L^{2}(\Omega_{5\varepsilon})}\right\}\|\nabla\psi\|_{L^{2}(\Omega)}.$
6. Estimate of $I_{4}$ . By Lemma 11 on the $\varepsilon$ -small-term we have $\begin{split}|I_{4}|=\varepsilon\left|\int_{\Omega}A(x/\varepsilon)\chi(x/\varepsilon)\nabla(\eta_{\varepsilon}S_{\varepsilon}^{2}\nabla u_{0}):\nabla\psi dx\right|\leq\\ C\left[\varepsilon\|S_{\varepsilon}(\nabla^{2}u_{0})\|_{L^{2}(\Omega\setminus\Omega_{2\varepsilon})}+\|\nabla u_{0}\|_{L^{2}(\Omega_{5\varepsilon})}\right]\|\nabla\psi\|_{L^{2}(\Omega)}.\end{split}$
7. Conclusion. Summing $I_{1}+I_{2}+I_{3}+I_{4}$ and observing that all boundary-layer terms are absorbed into the final $C\|\nabla u_{0}\|_{L^{2}(\Omega_{5\varepsilon})}\|\nabla\psi\|_{L^{2}(\Omega_{4\varepsilon})}$ and the bulk errors combine to $C\left\{\varepsilon\|S_{\varepsilon}(\nabla^{2}u_{0})\|_{L^{2}(\Omega\setminus\Omega_{3\varepsilon})}+\|\nabla u_{0}-S_{\varepsilon}\nabla u_{0}\|_{L^{2}(\Omega\setminus\Omega_{2\varepsilon})}\right\}\|\nabla\psi\|_{L^{2}(\Omega)},$ we recover exactly the claimed estimate. This completes the proof of the lemma. â
comment: None
### A.5 Applicability Boundary and Exclusion Domain
This is the experimental result of AIM in the reasoning process of multi-symbol operations such as two-dimensional expansion. AIM conducted various explorations, but none of them achieved the expected derivation result. Furthermore, there are not only obvious errors but also numerous processes that are irrelevant to the content we need to derive.
Prompt to AIM
Content: You can use the two-scale expansion to conclude the homogenization. You can perform the expansion on the equation based on this formula $\nabla:=\nabla_{x}+\frac{1}{\varepsilon}\nabla_{y}$ Content: Let $u_{\varepsilon}$ and $p_{\varepsilon}$ be the solution of the coupled Lame-Stokes system. Then $\begin{split}&u_{\varepsilon}=u_{0}(x,y)+\varepsilon u_{1}(x,y)+\varepsilon^{2}u_{2}(x,y)+\mathcal{O}(\varepsilon^{3});\\ &p_{\varepsilon}=\frac{1}{\varepsilon}p_{-1}(x,y)+p_{0}(x,y)+\varepsilon p_{1}(x,y)+\varepsilon^{2}p_{2}(x,y)+\mathcal{O}(\varepsilon^{3})\end{split}$ And substitute the form of such a solution into the original equation to get the equation of $u_{0}$ , $u_{1}$ , $u_{2}$ , which means the equations of three scales: $\mathcal{O}(1),\mathcal{O}(\varepsilon),\mathcal{O}(\varepsilon^{2})$ .
Response of AIM
Lemma 9.
Under the hypotheses of Lemmas 5â14 (periodic elastic matrix with LamĂ© coefficients $(\lambda,\mu)$ , incompressible Stokes fluid in the $\varepsilon$ âperiodic inclusions $\omega$ , and two-scale expansions $\displaystyle u_{\varepsilon}^{+}(x)$ $\displaystyle=u_{0}(x)+\varepsilon u_{1}^{+}(x,y)+O(\varepsilon^{2})\quad\text{in }\Omega\setminus D_{\varepsilon},$ $\displaystyle u_{\varepsilon}^{-}(x)$ $\displaystyle=u_{0}(x)+\varepsilon u_{1}^{-}(x,y)+O(\varepsilon^{2}),$ $\displaystyle p_{\varepsilon}(x)$ $\displaystyle=\varepsilon^{-1}p_{-1}(x)+p_{0}^{-}(x,y)+O(\varepsilon),$ with continuity $u_{0}^{+}=u_{0}^{-}=u_{0}$ , $Y$ âperiodicity in $y$ , and the usual cell problems for the elastic correctors $\chi^{pq}(y)$ and Stokes correctors $(w^{k},\pi^{k})$ ), the macroscopic displacement $u_{0}(x)$ and leading pressure $p_{-1}(x)$ satisfy in $\Omega$ the homogenized elastic-equilibrium equation $-\nabla\cdot\bigl[C^{*}:e(u_{0})\bigr]+|\omega|\nabla p_{-1}=0.$
correctness: True
Proof.
We give a fully rigorous two-scale derivation, correcting the interface-traction argument of the original proof.
1. Two-scale expansions and cell problems a) Elastic (matrix) phase $Y_{f}=Y\setminus\overline{\omega}$ . By Lemma 5, $u_{0}^{+}(x,y)=u_{0}(x)$ . Collecting $O(\varepsilon^{-1})$ terms in $\mathcal{L}_{\lambda,\mu}u_{\varepsilon}^{+}=0$ yields the elasticity-cell problem for $\{\chi^{pq}\}_{p,q=1}^{d}$ (Lemma 7). One then shows $u_{1}^{+}(x,y)=-\sum_{p,q=1}^{d}\chi^{pq}(y)\,\partial_{x_{q}}u_{0}^{p}(x).$
b) Fluid (inclusion) phase $\omega$ . By Lemma 6, $p_{-1}(x,y)=p_{-1}(x)$ . Grouping $O(\varepsilon^{-2})$ in the Stokes system gives the cell-Stokes problem for $(w^{k},\pi^{k})$ , $k=1,\dots,d$ (Lemma 9): $-\tilde{\mu}\Delta_{y}w^{k}+\nabla_{y}\pi^{k}=e_{k},\quad\text{div}_{y}w^{k}=0\text{ in }\omega,$ $w^{k}=0\text{ on }\partial\omega,\text{ periodic on }\partial Y,\quad\int_{\omega}\pi^{k}=0.$ Hence $u_{1}^{-}(x,y)=\sum_{k=1}^{d}w^{k}(y)\,\partial_{x_{k}}p_{-1}(x),$ $p_{0}^{-}(x,y)=\sum_{k=1}^{d}\pi^{k}(y)\,\partial_{x_{k}}p_{-1}(x).$
2. Homogenized elasticity in the matrix At order $O(\varepsilon^{0})$ in $\Omega\setminus D_{\varepsilon}$ one finds $-\nabla_{y}\cdot\sigma_{y}(u_{2}^{+})-\nabla_{x}\cdot\sigma_{y}(u_{1}^{+})-\nabla_{y}\cdot\sigma_{x}(u_{1}^{+})-\nabla_{x}\cdot\sigma_{x}(u_{0})=0.$ Integrating over $Y_{f}$ and using periodicity plus the cell-problem identities shows $\nabla_{x}\cdot\int_{Y_{f}}[\sigma_{x}(u_{0})+\sigma_{y}(u_{1}^{+})]\mathop{}\!\mathrm{d}y=0,$ and by Lemma 8 $\int_{Y_{f}}[\sigma_{x}(u_{0})+\sigma_{y}(u_{1}^{+})]\mathop{}\!\mathrm{d}y=C^{*}:e(u_{0}).$ Thus, neglecting fluid, one would obtain $-\nabla\cdot[C^{*}:e(u_{0})]=0.$
3. Fluid-traction contribution On each inclusion boundary $\partial\omega_{\varepsilon}$ the solid traction balances the fluid traction. Expand $u_{\varepsilon}^{-}=u_{0}(x)+\varepsilon u_{1}^{-}(x,y)+\dots,\quad p_{\varepsilon}^{-}=\varepsilon^{-1}p_{-1}(x)+p_{0}^{-}(x,y)+\dots,$ and note $\nabla\to\nabla_{x}+(1/\varepsilon)\nabla_{y}$ . Then the fluid stress $\sigma^{-}(u_{\varepsilon}^{-},p_{\varepsilon}^{-})=2\mu\left[(1/\varepsilon)D_{y}(u_{1}^{-})+D_{x}(u_{0})\right]-[\varepsilon^{-1}p_{-1}+p_{0}^{-}]I+O(\varepsilon).$ The $O(\varepsilon^{-1})$ âterm enforces $\nabla_{y}p_{-1}=0$ (so $p_{-1}=p_{-1}(x)$ ). The $O(\varepsilon^{0})$ âtraction on $\partial\omega$ is $t(y)=\left[2\mu D_{y}(u_{1}^{-})(y)-p_{0}^{-}(y)I\right]n(y)+\left[2\mu D_{x}(u_{0})-p_{-1}I\right]n(y).$ Since $D_{x}(u_{0})$ and $p_{-1}$ are $y$ âconstant and $\int_{\partial\omega}n\,dS=0,$ the second bracket integrates to zero. Hence the net force per cell exerted by the fluid on the solid is $\int_{\partial\omega}t\,dS=\int_{\partial\omega}\left[2\mu D_{y}(u_{1}^{-})-p_{0}^{-}I\right]n\,dS=\partial_{x_{k}}p_{-1}(x)\int_{\partial\omega}\left[2\mu D_{y}(w^{k})-\pi^{k}I\right]n\,dS.$ By the divergence theorem in $\omega$ , $\begin{split}\int_{\partial\omega}[2\mu D_{y}(w^{k})-\pi^{k}I]n\,dS=\int_{\omega}\nabla_{y}\cdot\left[2\mu D_{y}(w^{k})-\pi^{k}I\right]\mathop{}\!\mathrm{d}y\\ =\int_{\omega}\left[\mu\Delta_{y}w^{k}-\nabla_{y}\pi^{k}\right]dy=-\int_{\omega}e_{k}\mathop{}\!\mathrm{d}y=-|\omega|e_{k}.\end{split}$ Hence $\int_{\partial\omega}t\,dS=-|\omega|\partial_{x_{k}}p_{-1}e_{k}=-|\omega|\nabla p_{-1}(x).$ Dividing by the cell-volume $\varepsilon^{d}$ and passing to the limit adds a coupling term $+|\omega|\nabla p_{-1}(x)$ to the macroscopic equilibrium.
4. Conclusion Combining the matrix-average (Step 2) with the fluid-traction force (Step 3) yields exactly $-\nabla\cdot[C^{*}:e(u_{0})]+|\omega|\nabla p_{-1}=0$ in $\Omega$ . This completes a correct, complete, and rigorous derivation of the homogenized elastic-equilibrium equation. â
comment: None type: lemma
## Appendix B Full Proof for the Homogenization Problem
In this section, we present the complete proof of the homogenization problem obtained through humanâAI collaboration.
In the following content, we use blue to represent content derived manually, green to represent content output by AIM and then formatted manually, orange to represent content that was manually refined after AIM provided a proof with a high degree of completion and the remaining is the basic setup and information of this problem.
### B.1 Introduction
Consider $D\subset\Omega\subset R^{d}\,(d\geq 2)$ , where $\Omega$ is elastic material and $D$ is the high contrast inclusion part.
- $\Omega$ is open bounded with connected $C^{\infty}$ boundary $\partial\Omega$ .
- $D$ is open, has a finite number of components and has a Lipschitz boundary $\partial D$ .
- $\Omega\setminus D$ is connected with Lipschitz boundary $\partial\Omega\cup\partial D$ . The connected components of $D$ are enumerated as $D_{i}$ , $i=1,\ldots,N$ , $N$ is finite.
And given $\varepsilon\in(0,1)$ , $D=D_{\varepsilon}$ is part of an $\varepsilon$ -periodic array of small inclusions constructed as follows, in several steps.
$Y=(-\frac{1}{2},\frac{1}{2})^{d}$ is the unit cell. $\omega\subset Y$ is a simple connected open subset with connected Lipschitz boundary such that ${dist}(\omega,\partial Y)>0$ . $Y_{f}=Y\setminus\overline{\omega}$ is the model environment in the unit scale.
Given $\varepsilon>0$ and $\mathbf{n}\in\mathbb{Z}^{d}$ , we denote $\varepsilon(\mathbf{n}+Y)$ and $\varepsilon(\mathbf{n}+\omega)$ by $Y^{\mathbf{n}}_{\varepsilon}$ and $\omega^{\mathbf{n}}_{\varepsilon}$ , respectively. Let $\Pi_{\varepsilon}$ be the set of lattice points $\mathbf{n}$ such that $\overline{Y}^{\mathbf{n}}_{\varepsilon}$ be contained in $\Omega$ , i.e.,
$$
\Pi_{\varepsilon}:=\left\{\mathbf{n}\in\mathbb{Z}^{d}:\overline{Y}^{\mathbf{n}}_{\varepsilon}\subset\Omega\right\}, \tag{1}
$$
then the inclusions set $D=D_{\varepsilon}$ and the background part $\Omega_{\varepsilon}$ are defined by
$$
D_{\varepsilon}:=\bigcup_{\mathbf{n}\in\Pi_{\varepsilon}}\omega^{\mathbf{n}}_{\varepsilon}\quad\Omega_{\varepsilon}:=\Omega\setminus\overline{D}_{\varepsilon}. \tag{2}
$$
A pair of real numbers $(\lambda,\mu)$ is called admissible and referred to as a Lamé pair, if they satisfy $\mu>0$ and $d\lambda+2\mu>0$ . For a Lamé pair $(\lambda,\mu)$ , the elastostatic system (Lamé system) reads
$$
\mathcal{L}_{\lambda,\mu}u:=\mu\Delta u+(\lambda+\mu)\nabla\text{div}\,u, \tag{3}
$$
where $u=(u^{1},\ldots,u^{d})$ represents the displacement field and the divergence of $u$ is given by $\text{div}\,u=\sum_{i=1}^{d}\frac{\partial u^{i}}{\partial x_{i}}$ . The Lamé operator can be written as $\nabla\cdot\sigma(u)$ where
$$
\sigma(u):=\lambda(\nabla\cdot u)\mathbb{I}_{d}+2\mu\mathcal{D}(u), \tag{4}
$$
$$
\mathcal{D}(u)=\frac{1}{2}(\nabla+\nabla^{T})u=\frac{1}{2}(\partial_{i}u^{j}+\partial_{j}u^{i})_{ij}. \tag{5}
$$
The corresponding conormal derivative (boundary traction) at the boundary of a domain $E$ is
$$
\left.\frac{\partial u}{\partial\nu_{(\lambda,\mu)}}\right|_{\partial E}:=\sigma(u)N=\lambda(\text{div }u)N+2\mu\mathcal{D}(u)N\text{ on }\partial E. \tag{6}
$$
We consider the space $\mathbb{R}$ of rigid motions in $\mathbb{R}^{d}$ , defined by
$$
\mathbb{R}:=\left\{\mathbf{r}=(r_{1},\ldots,r_{d})^{T}:\mathcal{D}(\mathbf{r})=0\text{ in }\mathbb{R}^{d}\right\}.
$$
We define $H_{\mathbb{R}}^{-\frac{1}{2}}(\partial D_{\varepsilon})$ as the subspace of $H^{-\frac{1}{2}}(\partial D_{\varepsilon})$ that is orthogonal to $\mathbb{R}$ , i.e.,
$$
H_{\mathbb{R}}^{-\frac{1}{2}}(\partial D_{\varepsilon}):=\left\{\phi\in H^{-\frac{1}{2}}(\partial D_{\varepsilon}):\left(\phi,\mathbf{r}\right)_{(H^{\frac{1}{2}}(\partial{D_{\varepsilon_{i}}}),H^{-\frac{1}{2}}(\partial{D_{\varepsilon_{i}}}))}=0,\forall\mathbf{r}\in\mathbb{R}\text{ and }1\leq i\leq N\right\}. \tag{7}
$$
Consider the displacement field $u_{\varepsilon}$ satisfying the following transmission system:
$$
\begin{cases}\mathcal{L}_{\lambda,\mu}u_{\varepsilon}=0&\text{in }\Omega\setminus\overline{D}_{\varepsilon},\\
\mathcal{L}_{\widetilde{\lambda},\widetilde{\mu}}u_{\varepsilon}=0&\text{in }D_{\varepsilon},\\
u_{\varepsilon}|_{-}=u_{\varepsilon}|_{+}\text{ and }\left.\frac{\partial u_{\varepsilon}}{\partial\nu_{(\widetilde{\lambda},\widetilde{\mu})}}\right|_{-}=\left.\frac{\partial u_{\varepsilon}}{\partial\nu_{(\lambda,\mu)}}\right|_{+}&\text{on }\partial D_{\varepsilon},\\
\left.\frac{\partial u_{\varepsilon}}{\partial\nu_{(\lambda,\mu)}}\right|_{\partial\Omega}=g\in H_{\mathbb{R}}^{-\frac{1}{2}}(\partial\Omega)\quad\text{and}\quad u_{\varepsilon}|_{\partial\Omega}\in H_{\mathbb{R}}^{\frac{1}{2}}(\partial\Omega).\end{cases} \tag{8}
$$
Suppose $\widetilde{{\mu}}$ fixed, then we arrive at the equations about the incompressible inclusion limit. In this case, we need to consider the homogenization problem of the following coupled Lamé -Stokes system:
$$
\begin{cases}\nabla\cdot(\lambda(\nabla\cdot u_{\varepsilon})I+2\mu D(u_{\varepsilon}))=0\quad&\text{in}\ \Omega\setminus\overline{D}_{\varepsilon},\\
\nabla\cdot(2\tilde{\mu}D(u_{\varepsilon})+p_{\varepsilon}I)=0,\nabla\cdot u_{\varepsilon}\quad&\text{in}\ D_{\varepsilon},\\
(2\tilde{\mu}D(u_{\varepsilon})+p_{\varepsilon}I)N^{-}-(\lambda(\nabla\cdot u_{\varepsilon})I+2\mu D(u_{\varepsilon}))N^{+}=0,\\
u_{\varepsilon}|_{+}=u_{\varepsilon}|_{-}\quad&\text{on}\ \partial D_{\varepsilon},\\
\frac{\partial u_{\varepsilon}}{\partial\nu_{(\lambda,\mu)}}|_{\partial\Omega}=g\in H_{\mathbb{R}}^{-\frac{1}{2}}(\partial\Omega),u_{\varepsilon}|_{\partial\Omega}\in H_{\mathbb{R}}^{\frac{1}{2}}(\partial\Omega).\end{cases} \tag{9}
$$
### B.2 Uniqueness and Existence
Define $V^{\varepsilon}=\{u\in H^{1}(\Omega):u|_{\partial\Omega}\bot\mathbb{R},\frac{\partial u}{\partial\nu}|_{\partial\Omega}\bot\mathbb{R}\},\|\cdot\|_{V^{\varepsilon}}=\|\cdot\|_{H^{1}(\Omega)}$ .
Searching for $u_{\varepsilon}\in V^{\varepsilon},p_{\varepsilon}\in L^{2}(D_{\varepsilon})$ such that $\forall\varphi\in V^{\varepsilon},\psi\in L^{2}(D_{\varepsilon})$
$$
\displaystyle\int_{\partial\Omega}g\cdot\varphi=\int_{\Omega^{\varepsilon}}[\lambda(\nabla\cdot u_{\varepsilon})I+2\mu(\frac{1}{2}(\nabla+\nabla^{T})u_{\varepsilon})]:\nabla\varphi \displaystyle+\int_{D_{\varepsilon}}\operatorname{div}\,\varphi\cdot p_{\varepsilon}+\int_{D_{\varepsilon}}[2\tilde{\mu}(\frac{1}{2}(\nabla+\nabla^{T})u_{\varepsilon})]:\nabla\varphi, \tag{10}
$$
and
$$
0=\int_{D_{\varepsilon}}\operatorname{div}\,u\cdot\psi. \tag{11}
$$
Suppose $a(\cdot,\cdot):V^{\varepsilon}\times V^{\varepsilon}\rightarrow R$
$$
a(u_{\varepsilon},\varphi)=\int_{\Omega^{\varepsilon}}[\lambda(\nabla\cdot u_{\varepsilon})I+2\mu(\frac{1}{2}(\nabla+\nabla^{T})u_{\varepsilon})]:\nabla\varphi+\int_{D_{\varepsilon}}[2\tilde{\mu}(\frac{1}{2}(\nabla+\nabla^{T})u_{\varepsilon})]:\nabla\varphi. \tag{12}
$$
$b(\cdot,\cdot):V^{\varepsilon}\times L^{2}(D_{\varepsilon})\rightarrow R$
$$
b(u_{\varepsilon},\psi)=\int_{D_{\varepsilon}}\operatorname{div}\,u_{\varepsilon}\cdot\psi. \tag{13}
$$
So we have the following equation equivalent to the variation form of original equation:
$$
\begin{cases}a(u_{\varepsilon},\varphi)+b(\varphi,p_{\varepsilon})=\int_{\partial\Omega}g\cdot\varphi\quad&\forall\varphi\in V^{\varepsilon},\\
b(u_{\varepsilon},\psi)=0\quad&\forall\psi\in L^{2}(D_{\varepsilon}).\end{cases} \tag{14}
$$
**Theorem 1 (Babuska-Brezzi Theorem)**
*Let $H$ and $\mathbb{Q}$ be Hilbert spaces, and define bounded bilinear forms $a:H\times H\rightarrow\mathbb{R}$ , $b:H\times\mathbb{Q}\rightarrow\mathbb{R}$ . For given $F\in H^{\prime}$ and $G\in\mathbb{Q}^{\prime}$ , consider finding $(\sigma,u)\in H\times\mathbb{Q}$ such that:
| | $\displaystyle a(\sigma,\tau)+b(\tau,u)$ | $\displaystyle=F(\tau),\quad\forall\tau\in H,$ | |
| --- | --- | --- | --- | Assume that:
1. There exists $\alpha>0$ such that
$$
a(\tau,\tau)\geq\alpha\|\tau\|_{H}^{2},\quad\forall\tau\in H.
$$
1. There exists $\beta>0$ such that
$$
\inf_{v\in\mathbb{Q}}\sup_{\tau\in H}\frac{b(\tau,v)}{\|\tau\|_{H}\|v\|_{\mathbb{Q}}}\geq\beta,\quad v\neq 0,\tau\neq 0.
$$ Then there exists a unique solution $(\sigma,u)\in H\times\mathbb{Q}$ satisfying the above equations, and there exists a constant $C=C(\|a\|,\alpha,\beta)>0$ such that:
$$
\|(\sigma,u)\|_{H\times\mathbb{Q}}\leq C\left(\|F\|_{H^{\prime}}+\|G\|_{\mathbb{Q}^{\prime}}\right).
$$*
By this theorem, we verify the conditions to obtain the existence and uniqueness. Suppose $a_{ij}^{\alpha\beta}=\lambda\delta_{i\alpha}\delta_{j\beta}+\mu(\delta_{ij}\delta_{\alpha\beta}+\delta_{i\beta}\delta_{j\alpha}),$
$$
A\nabla u_{\varepsilon}\cdot\nabla u_{\varepsilon}\geq\frac{\min\{\lambda d+2\mu,2\mu\}}{4}|(\nabla+\nabla^{T})u_{\varepsilon}|^{2}, \tag{15}
$$
$$
\displaystyle\int_{\Omega\setminus D_{\varepsilon}}A\nabla u_{\varepsilon}\cdot\nabla u_{\varepsilon}+\int_{D_{\varepsilon}}2\tilde{\mu}(\frac{1}{2}(\nabla+\nabla^{T})u_{\varepsilon}):\nabla u_{\varepsilon} \displaystyle\geq\int_{\Omega\setminus D_{\varepsilon}}\frac{\min\{\lambda d+2\mu,2\mu\}}{4}|(\nabla+\nabla^{T})u_{\varepsilon}|^{2} \displaystyle+2\tilde{\mu}\int_{D_{\varepsilon}}(\frac{1}{2}(\nabla+\nabla^{T})u_{\varepsilon}):(\frac{1}{2}(\nabla+\nabla^{T})u_{\varepsilon}) \displaystyle\geq C\|u_{\varepsilon}\|^{2}_{H^{1}(\Omega)}. \tag{16}
$$
**Lemma 1**
*$$
\inf_{\psi\in L^{2}(D_{\varepsilon})}\sup_{u_{\varepsilon}\in V^{\varepsilon}}\frac{b(u_{\varepsilon},\psi)}{\|u_{\varepsilon}\|_{H^{1}(\Omega)}\|\psi\|_{L^{2}(D_{\varepsilon})}}\geq\beta. \tag{17}
$$*
Given $\psi\in L^{2}(D_{\varepsilon})$ , $\psi=\sum_{k\in\Pi_{\varepsilon}}\psi_{k}$ , here $\psi_{k}$ are supported on $\omega_{k}^{\varepsilon}$ . Suppose that $\tilde{Y}$ is a cubic with $\omega\subset\tilde{Y}\subset Y$ , $d(\tilde{Y},Y)>0$ ,
$$
\hat{\psi}_{k}=\begin{cases}\psi_{k}(x)\quad x\in\omega_{k}^{\varepsilon},\\
-\frac{1}{\varepsilon|\tilde{Y}\setminus\omega|}\int_{\omega_{k}^{\varepsilon}}\psi_{k}(x)\mathop{}\!\mathrm{d}x\quad x\in\tilde{Y}_{k}\setminus\omega_{k}^{\varepsilon}.\end{cases}
$$
And we know
$$
\|\hat{\psi}\|_{L^{2}(\tilde{Y}_{k})}\leq C_{1}\|\psi\|_{L^{2}(\omega_{k}^{\varepsilon})},
$$
and
$$
\int_{\tilde{Y}}\hat{\psi}\mathop{}\!\mathrm{d}x=0.
$$
Besides, we can get $\exists\hat{d_{k}}\in H_{0}^{1}(\tilde{Y}_{k})$ such that $\operatorname{div}\,\hat{d_{k}}=\hat{\psi_{k}}\text{ in }\tilde{Y}_{k}$ with $\|\nabla\hat{d_{k}}\|_{L^{2}(\tilde{Y}_{k})}\leq C_{2}\|\hat{\psi}\|_{L^{2}(\tilde{Y}_{k})}$ . Note that $\tilde{d_{k}}$ is $\hat{d_{k}}$ , zero extension to the boundary $\partial\Omega$ and $\hat{d}=\sum_{k\in\Pi^{\varepsilon}}\tilde{d_{k}}\in V^{\varepsilon}$ . So we get
$$
\|\nabla\hat{d}\|_{L^{2}(\Omega)}\leq C_{2}\|\hat{\psi}\|_{L^{2}(\Omega)}\leq C_{1}\|\psi\|_{L^{2}(\Omega)}. \tag{18}
$$
Therefore,
$$
\int_{D_{\varepsilon}}\psi\cdot\operatorname{div}\,\hat{d}=\int_{D_{\varepsilon}}\psi^{2}\geq\frac{1}{C_{1}C_{2}}\|\psi\|_{L^{2}(D_{\varepsilon})}\|\nabla\hat{d}\|_{L^{2}(D_{\varepsilon})}. \tag{19}
$$
So we verify the inf-sup condition. By the Babuska-Brezzi Theorem, there exists one unique solution $u_{\varepsilon}\in V^{\varepsilon}$ , $p_{\varepsilon}\in L^{2}(D_{\varepsilon})$ to the original equation with $\|(u_{\varepsilon},p_{\varepsilon})\|_{V^{\varepsilon}\times L^{2}(D_{\varepsilon})}\leq C\|g\|_{(V^{\varepsilon})^{*}}\leq C\|g\|_{H^{1}(\Omega)}$ .
### B.3 Uniqueness and Existence of Solutions at Each Order of Asymptotic Expansion
Consider the following equation: $p,u$ are $Y$ -periodic
$$
\begin{cases}\nabla_{y}\cdot(\lambda(\nabla_{y}\cdot u)I+2\mu(\frac{1}{2}(\nabla+\nabla^{T})u)=F_{1}\quad&\text{in}\quad Y\setminus\omega,\\
\nabla_{y}\cdot(2\tilde{\mu}[\frac{1}{2}(\nabla_{y}+\nabla_{y}^{T})u]+pI)=F_{2},\nabla\cdot u=F_{3}\quad&\text{in}\quad\omega,\\
(2\tilde{\mu}[\frac{1}{2}(\nabla_{y}+\nabla_{y}^{T})u]+p_{-1}I)\cdot N-(\lambda(\nabla\cdot u)I+2\mu(\frac{1}{2}(\nabla_{y}+\nabla_{y}^{T})u)\cdot N=G\cdot N,\\
u|_{+}=u|_{-}\quad&\text{on}\quad\partial\omega.\end{cases} \tag{20}
$$
For $F=F_{1}(x,\cdot)\mathbf{1}_{y\in Y\setminus\omega}+F_{2}(x,\cdot)\mathbf{1}_{y\in\omega}\in(H^{1}_{\#,0}(Y;R^{d}))^{*}$ , $u\in H^{1}_{\#,0}(Y;R^{d})$ , we define
$$
<(F,G),u>_{H^{*},H}=\int_{Y\setminus\omega}F_{1}(x,y)u(x,y)\mathop{}\!\mathrm{d}y+\int_{\omega}F_{2}(x,y)u(x,y)\mathop{}\!\mathrm{d}y+\int_{\partial\omega}G(x,y)u(x,y)N\mathop{}\!\mathrm{d}y. \tag{21}
$$
Here, $F_{3}(x,\cdot)\in L^{2}(\omega;R),G(x,\cdot)\in H^{\frac{1}{2}}(\partial\omega;R^{d})$ .
And define
$$
V=\{u(x,\cdot)\in H^{1}_{\#,0}(Y;R^{d})\},\ M=\{p(x,\cdot)\in L^{2}(\omega;R)\}.
$$
So the equation has the unique solution $(u,p)\in V\times M$ if and only if $\int_{Y\setminus\omega}F_{1}+\int_{\omega}F_{2}-\int_{\partial\omega}G\cdot N=0$ and we have $\|(u,p)\|_{V\times M}\leq C[\|(F_{1},F_{2},G)\|_{(H^{1}_{\#,0}(Y;R^{d}))^{*}}+\|F_{3}\|_{L^{2}(\omega;R)}]$ .
Suppose $a:H^{1}_{\#,0}(Y;R^{d})\times H^{1}_{\#,0}(Y;R^{d})\rightarrow R$ ,
$$
a(u,\varphi)=\int_{Y\setminus\omega}[\lambda(\nabla_{y}\cdot u)I+2\mu(\frac{1}{2}(\nabla_{y}+\nabla_{y}^{T})u)]:\nabla_{y}\varphi+\int_{\omega}[2\tilde{\mu}(\frac{1}{2}(\nabla_{y}+\nabla_{y}^{T})u)]:\nabla_{y}\varphi, \tag{22}
$$
and $b:H^{1}_{\#,0}(Y;R^{d})\times L^{2}(\omega)\rightarrow R$ ,
$$
b(u,\psi)=\int_{\omega}\operatorname{div}_{y}\,u\cdot\psi. \tag{23}
$$
We verify the property of $a$ and $b$ : $\forall u\in H^{1}_{\#,0}(Y;R^{d})$ , and $a_{ij}^{\alpha\beta}=\lambda\delta_{i\alpha}\delta_{j\beta}+\mu(\delta_{ij}\delta_{\alpha\beta}+\delta_{i\beta}\delta_{j\alpha})$ , there is
$$
A\nabla u\cdot\nabla u\geq\frac{\min(\lambda d+2\mu,2\mu)}{4}|(\nabla+\nabla^{T})u|^{2}. \tag{24}
$$
$$
\displaystyle\int_{Y\setminus\omega}A\nabla u\cdot\nabla u+\int_{\omega}2\tilde{\mu}\left(\frac{1}{2}\left(\nabla+\nabla^{T}\right)u\right):\nabla u \displaystyle\geq\int_{Y\setminus\omega}\frac{\min(\lambda d+2\mu,2\mu)}{4}\left|\left(\nabla+\nabla^{T}\right)u\right|^{2}+2\tilde{\mu}\int_{\omega}\left(\frac{1}{2}\left(\nabla+\nabla^{T}\right)u\right):\left(\frac{1}{2}\left(\nabla+\nabla^{T}\right)u\right) \displaystyle\geq C\|u\|_{H^{1}(Y)}^{2}. \tag{25}
$$
So we verify this property:
$$
\inf_{\psi\in L^{2}(\omega)}\sup_{u\in H^{1}_{\#,0}(Y;R^{d})}\frac{b(u,\psi)}{\|u\|_{H^{1}_{\#,0}(Y;R^{d})}\|\psi\|_{L_{\#}^{2}(Y)}}\geq\beta. \tag{26}
$$
Given $\psi\in L_{\#}^{2}(\omega)$ , we consider
$$
\hat{\psi}(x)=\begin{cases}\psi(x)\quad x\in\omega,\\
-\frac{1}{|Y\setminus\omega|}\int_{\omega}\psi(x)\mathop{}\!\mathrm{d}x\quad x\in Y\setminus\omega.\end{cases}
$$
This extension satisfy $\|\hat{\psi}\|_{L^{2}(Y)}\leq C_{1}\|\psi\|_{L^{2}(\omega)},\hat{\psi}\in H^{1}_{\#,0}(Y;R^{d})$ and $\int_{Y}\hat{\psi}\mathop{}\!\mathrm{d}x=0$ .
Besides, we can get $\exists\hat{d}\in H^{1}_{\#,0}(Y;R^{d})$ such that $\operatorname{div}\,\hat{d}=\hat{\psi}\text{ in }Y$ with $\|\nabla\hat{d}\|_{L^{2}(Y)}\leq C_{2}\|\hat{\psi}\|_{L^{2}(Y)}$ .
So we get
$$
\|\nabla\hat{d}\|_{L^{2}(Y)}\leq C_{2}\|\hat{\psi}\|_{L^{2}(Y)}\leq C_{1}C_{2}\|\psi\|_{L^{2}(\omega)}.
$$
Therefore,
$$
\int_{\omega}\psi\cdot\operatorname{div}\,\hat{d}=\int_{\omega}\psi^{2}\geq\frac{1}{C_{1}C_{2}}\|\psi\|_{L^{2}(\omega)}\|\nabla\hat{d}\|_{L^{2}(Y)}.
$$
Finally, we verify that the necessary conditions are sufficient. By divergence theorem,
$$
\displaystyle\int_{Y\setminus\omega}F_{1}=\int_{Y\setminus\omega}\nabla_{y}\cdot(\lambda(\nabla_{y}\cdot u)I+2\mu(\frac{1}{2}(\nabla+\nabla^{T})u) \displaystyle=-\int_{\partial\omega}(\lambda(\nabla_{y}\cdot u)I+2\mu(\frac{1}{2}(\nabla+\nabla^{T})u)N. \tag{27}
$$
$$
\displaystyle\int_{\omega}F_{2}=\int_{\omega}\nabla_{y}\cdot(2\tilde{\mu}[\frac{1}{2}(\nabla_{y}+\nabla_{y}^{T})u]+pI) \displaystyle=\int_{\partial\omega}(2\tilde{\mu}[\frac{1}{2}(\nabla_{y}+\nabla_{y}^{T})u]+pI)N. \tag{28}
$$
Therefore we get
$$
\displaystyle\int_{Y\setminus\omega}F_{1}+\int_{\omega}F_{2}-\int_{\partial\omega}G\cdot N=0. \tag{29}
$$
On the other hand, we try to search for $u\in V$ , $p\in M$ such that for $\forall\varphi\in V$
$$
\displaystyle\int_{Y\setminus\omega}F_{1}+\int_{\omega}F_{2}=\int_{Y\setminus\omega}(\lambda(\nabla_{y}\cdot u)I+2\mu(\frac{1}{2}(\nabla+\nabla^{T})u):\nabla_{y}\varphi. \tag{30}
$$
$$
\displaystyle\int_{Y\setminus\omega}F_{1}+\int_{\omega}F_{2}=-\int_{Y\setminus\omega}(\lambda(\nabla_{y}\cdot u)I+2\mu(\frac{1}{2}(\nabla+\nabla^{T})u):\nabla_{y}\varphi \displaystyle-\int_{\omega}(2\tilde{\mu}[\frac{1}{2}(\nabla_{y}+\nabla_{y}^{T})u]+pI):\nabla_{y}\varphi+\int_{\partial\omega}G\cdot N\varphi. \tag{31}
$$
This is correct since the analysis of Babuska-Brezzi Theorem: the following two equation are well-posedness in $V\times M$
$$
\displaystyle 0=\int_{Y\setminus\omega}(\lambda(\nabla_{y}\cdot u)I+2\mu(\frac{1}{2}(\nabla+\nabla^{T})u):\nabla_{y}\varphi+\int_{\omega}(2\tilde{\mu}[\frac{1}{2}(\nabla_{y}+\nabla_{y}^{T})u]+pI):\nabla_{y}\varphi\ \forall\varphi\in V, \tag{32}
$$
and
$$
\displaystyle\int_{\omega}F_{3}\cdot\psi=\int_{\omega}(\nabla_{y}\cdot u)\psi\quad\forall\psi\in M. \tag{33}
$$
### B.4 Homogenization Equation
For $(u_{0},p_{-1})$ :
$$
\begin{cases}\nabla_{y}\cdot(\lambda(\nabla_{y}\cdot u_{0}+2\mu D_{y}(u_{0}))=0\\
\nabla_{y}\cdot(2\tilde{\mu}D_{y}(u_{0})+p_{-1}I)=0,\nabla_{y}\cdot u_{0}=0\\
(2\tilde{\mu}D_{y}(u_{0})+p_{-1}I)N^{-}-(\lambda(\nabla_{y}\cdot u_{0})I+2\mu D_{y}(u_{0}))N^{+}=0\\
u_{0}|_{+}=u_{0}|_{-}\\
\end{cases} \tag{34}
$$
From the derivations in Sec. B.2, we know $u_{0}(x,y)=u_{0}(x)$ , $p_{-1}(x,y)=p_{-1}(x)=0$ .
For $(u_{1},p_{0})$ :
$$
\begin{cases}\nabla_{y}\cdot(\lambda\nabla_{y}\cdot u_{1}+2\mu D_{y}(u_{1}))=0\\
\nabla_{y}\cdot(2\tilde{\mu}D_{y}(u_{1})+p_{0}I)=0,\nabla_{y}\cdot u_{1}=-\nabla_{x}\cdot u_{0}\\
(2\tilde{\mu}D_{y}(u_{1})+p_{0}I)N^{-}-(\lambda(\nabla_{y}\cdot u_{1})I+2\mu D_{y}(u_{1}))N^{+}\\
\quad=-2\tilde{\mu}D_{x}(u_{0})N^{-}+(\lambda\nabla_{x}\cdot u_{0}I+2\mu D_{x}(u_{0}))N^{+}\\
u_{1}|_{+}=u_{1}|_{-}\\
\end{cases} \tag{35}
$$
### B.5 Cell Problem
Suppose $\chi^{ij}$ and $r^{ij}$ satisfy the following equation:
$$
\begin{cases}\nabla_{y}\cdot(\lambda\nabla_{y}\cdot\chi^{ij}+2\mu D_{y}\chi^{ij})=0\quad&\text{in}\quad Y\setminus\omega,\\
\nabla_{y}\cdot(2\tilde{\mu}D_{y}\chi^{ij}+r^{ij}I)=0,\nabla_{y}\cdot\chi^{ij}=-\delta_{ij}\quad&\text{in}\quad\omega,\\
(2\tilde{\mu}D_{y}(\chi^{ij})+r^{ij}I)N^{-}-(\lambda(\nabla_{y}\cdot\chi^{ij})I+2\mu D_{y}(\chi^{ij}))N^{+}\\
\quad=-\tilde{\mu}(E_{ij}+E_{ji})N^{-}+\mu(E_{ij}+E_{ji})N^{+}+\lambda IN\delta_{ij}|_{+}\\
\chi^{ij}|_{+}=\chi^{ij}|_{-}\quad&\text{on}\quad\partial\omega.\end{cases} \tag{36}
$$
From the definition above, we have $\chi^{ij}=\chi^{ji}$ , $r^{ij}=r^{ji}$ ,
$$
u_{1}=(D_{x}u_{0})^{ij}\cdot\chi^{ij},\ p_{0}=(D_{x}u_{0})^{ij}r^{ij}.
$$
So for $(u_{2},p_{1})$ , we have:
$$
\begin{cases}\nabla_{y}\cdot(\lambda\nabla_{y}\cdot u_{2}+2\mu D_{y}u_{2})=-\nabla_{x}\cdot[(\lambda(\nabla_{x}u_{0})I+2\mu D_{x})+(\lambda(\nabla_{y}\cdot u_{1})I+2\mu D_{y}u_{1})]\\
\quad\quad\quad\quad\quad\quad\quad\quad\quad\quad\quad\quad\quad\quad\quad-\nabla_{y}\cdot(\lambda(\nabla_{x}\cdot u_{1})I+2\mu D_{x}u_{1})\\
\nabla_{y}\cdot(2\tilde{\mu}D_{y}u_{2}+p_{1}I)=-\nabla_{x}\cdot(2\tilde{\mu}D_{x}u_{0}+p_{0}I)-\nabla_{x}\cdot(2\tilde{\mu}(D_{y}u_{1}))-\nabla_{y}\cdot(2\tilde{\mu}(D_{x}u_{1}))\\
\nabla\cdot u_{2}=-\nabla_{x}\cdot u_{1}\\
u_{2}|_{-}=u_{2}|_{+}\\
(p_{1}+2\tilde{\mu}(D_{y}u_{2}))N^{-}-(\lambda(\nabla_{y}\cdot u_{2})+2\mu(D_{y}u_{2}))N^{+}\\
\quad\quad\quad\quad\quad\quad\quad\quad\quad\quad=-(2\tilde{\mu}D_{x}(u_{1}))N^{-}+(\lambda(\nabla_{x}\cdot u_{1})+2\mu(D_{x}u_{1}))N^{+}\end{cases} \tag{37}
$$
By divergence theorem, we can get
$$
\int_{\partial\omega}(\lambda(\nabla_{x}\cdot u_{1})+2\mu(D_{x}u_{1}))N^{+}=-\int_{Y\setminus\omega}\nabla_{y}\cdot(\lambda(\nabla_{x}\cdot u_{1})+2\mu(D_{x}u_{1})). \tag{38}
$$
$$
\int_{\partial\omega}(2\tilde{\mu}D_{x}(u_{1}))N^{-}=\int_{\omega}\nabla_{y}\cdot(2\tilde{\mu}D_{x}(u_{1})). \tag{39}
$$
Combining the solvability conditions, we have
$$
\displaystyle|Y\setminus\omega|\nabla_{x}\cdot[\lambda(\nabla_{x}\cdot u_{0})I+2\mu D_{x}(u_{0})]+\int_{Y\setminus\omega}\lambda(\nabla_{x}(\nabla_{y}\cdot u_{1}))+2\mu(\nabla_{x}\cdot D_{y}(u_{1}))\mathop{}\!\mathrm{d}y \displaystyle+|\omega|(2\tilde{\mu}\nabla_{x}\cdot D_{x}u_{0})+\int_{\omega}(2\tilde{\mu}(\nabla_{x}\cdot D_{y}u_{1})+\nabla_{x}p_{0})\mathop{}\!\mathrm{d}y=0. \tag{40}
$$
Substituting this equation $u_{1}=(D_{x}u_{0})^{ij}\cdot\chi^{ij}$ , $p_{0}=(D_{x}u_{0})^{ij}r^{ij}$ , we finally get
$$
\displaystyle|Y\setminus\omega|\nabla_{x}\cdot[\lambda(\nabla_{x}\cdot u_{0})I+2\mu D_{x}(u_{0})]+\int_{Y\setminus\omega}[\lambda(\nabla_{y}\cdot\chi^{ij})I+2\mu D_{y}(\chi^{ij})](\nabla_{x}(D_{x}u_{0})^{ij})\mathop{}\!\mathrm{d}y \displaystyle+|\omega|(2\tilde{\mu}\nabla_{x}\cdot D_{x}u_{0})+\int_{\omega}(2\tilde{\mu}(D_{y}\chi^{ij})+r^{ij}I)\nabla_{x}(D_{X}u_{0})^{ij}\mathop{}\!\mathrm{d}y=0. \tag{41}
$$
We can rewrite this equation as the form $-\frac{\partial}{\partial x_{i}}\hat{a}_{ij}^{\alpha\beta}\frac{\partial u^{\beta}}{\partial x_{j}}=0$ :
$$
\displaystyle\hat{a}_{ij}^{\alpha\beta}=|Y\setminus\omega|(\lambda\delta_{i\alpha}\delta_{j\beta}+\mu(\delta_{i\beta}\delta_{j\alpha}+\delta_{ij}\delta_{\alpha\beta}))+\int_{Y\setminus\omega}[\lambda\nabla\cdot\chi^{j\beta}\delta_{i\alpha}+\mu(\frac{\partial\chi^{j\beta}_{\alpha}}{\partial y_{i}}+\frac{\partial\chi^{j\beta}_{i}}{\partial y_{\alpha}})]\mathop{}\!\mathrm{d}y \displaystyle+|\omega|\tilde{\mu}(\delta_{ij}\delta_{\alpha\beta}+\delta_{i\beta}\delta_{j\alpha})+\int_{\omega}\tilde{\mu}(\frac{\partial\chi^{j\beta}_{\alpha}}{\partial y_{i}}+\frac{\partial\chi^{j\beta}_{i}}{\partial y_{\alpha}})+r^{j\beta}\delta_{i\alpha}\mathop{}\!\mathrm{d}y. \tag{42}
$$
#### B.5.1 Symmetry
We first prove the symmetry of $\hat{a}_{ij}^{\alpha\beta}$ .
Define: $a(\varphi,\psi)=\int_{Y\setminus\omega}\nabla_{y}\varphi:[\lambda(\nabla_{y}\cdot\psi)I+2\mu D(\psi)]+\int_{\omega}2\tilde{\mu}D\psi:D\psi=\int_{Y\setminus\omega}\lambda(\nabla_{y}\cdot\varphi)(\nabla_{y}\cdot\varphi)+\mu D(\psi):D(\psi)+\int_{\omega}2\tilde{\mu}D\psi:D\psi$ .
Claim $\hat{a}_{ij}^{\alpha\beta}=a(p^{i\alpha}+\chi^{i\alpha},p^{j\beta}+\chi^{j\beta})$ ,
$$
\displaystyle a(p^{i\alpha},p^{j\beta}) \displaystyle=\int_{Y\setminus\omega}\lambda\delta_{i\alpha}\delta_{j\beta}+\mu(\delta_{i\beta}\delta_{j\alpha}+\delta_{ij}\delta_{\alpha\beta}))+\int_{\omega}\tilde{\mu}(\delta_{ij}\delta_{\alpha\beta}+\delta_{i\beta}\delta_{j\alpha}), \displaystyle a(p^{i\alpha},\chi^{j\beta}) \displaystyle=\int_{Y\setminus\omega}\lambda\delta_{i\alpha}(\nabla\cdot\chi^{j\beta})+\mu\left(\frac{e_{i\alpha}+e_{\alpha i}}{2}\right):\left(\nabla\chi^{j\beta}+\nabla^{T}\chi^{\beta j}\right) \displaystyle\quad+\int_{\omega}\tilde{\mu}\left(\frac{e_{i\alpha}+e_{\alpha i}}{2}\right):\left(\nabla\chi^{j\beta}+\nabla^{T}\chi^{\beta j}\right) \displaystyle=\int_{\partial\omega}-\lambda\delta_{i\alpha}\chi^{j\beta}N-\mu(e_{i\alpha}+e_{\alpha i})\chi^{j\beta}N+\int_{\partial\omega}\tilde{\mu}(e_{i\alpha}+e_{\alpha i})\chi^{j\beta}N, \displaystyle a(\chi^{i\alpha},\chi^{j\beta}) \displaystyle=\int_{Y\setminus\omega}\lambda(\nabla\cdot\chi^{i\alpha})(\nabla\cdot\chi^{j\beta})+\mu(\nabla_{y}\chi^{i\alpha}):\left(\nabla\chi^{j\beta}+\nabla^{T}\chi^{j\beta}\right) \displaystyle\quad+\int_{\omega}\tilde{\mu}(\nabla_{y}\chi^{i\alpha}):\left(\nabla\chi^{j\beta}+\nabla^{T}\chi^{j\beta}\right) \displaystyle=\int_{\partial\omega}-\left[\lambda\chi^{i\alpha}(\nabla\cdot\chi^{j\beta})+2\mu\chi^{i\alpha}D(\chi^{j\beta})\right]N+\int_{\partial\omega}2\tilde{\mu}\chi^{i\alpha}(D\chi^{j\beta})N \displaystyle\quad+\int_{\partial\omega}\chi^{i\alpha}r^{j\beta}N+\int_{\omega}\delta_{i\alpha}r^{j\beta}. \tag{43}
$$
Observing that $a(p^{i\alpha},p^{j\beta}+\chi^{j\beta})=\hat{a}_{ij}^{\alpha\beta}-\int_{\omega}\delta_{i\alpha}r^{j\beta}\mathop{}\!\mathrm{d}y$ and $a(\chi^{i\alpha},p^{j\beta}+\chi^{j\beta})=\int_{\omega}\delta_{i\alpha}r^{j\beta}\mathop{}\!\mathrm{d}y$ , we have $\hat{a}_{ij}^{\alpha\beta}=a(p^{i\alpha}+\chi^{i\alpha},p^{j\beta}+\chi^{j\beta})$ and $\hat{a}_{ij}^{\alpha\beta}=\hat{a}_{i\beta}^{\alpha j}=\hat{a}_{ji}^{\beta\alpha}$ .
#### B.5.2 Elliptic
Let $\varphi=\phi+\varepsilon(D_{x}\phi)^{ij}\chi^{ij}(\frac{x}{\varepsilon}),\phi\in C_{0}^{\infty}(\Omega;R^{d})$
$$
\displaystyle\int_{\Omega}a(\varphi,\varphi) \displaystyle=-\int_{\Omega}\int_{Y\setminus\omega}(\nabla_{x}\phi+(D_{x}\phi)^{ij}\nabla_{y}\chi^{ij}):[\lambda(\nabla_{x}\cdot\phi+(D_{x}\phi)^{ij}\nabla_{y}\cdot\chi^{ij})I \displaystyle\quad\quad+2\mu(D_{x}\phi+(D_{x}\phi)^{ij}D_{y}(\chi^{ij}))] \displaystyle\quad\quad-\int_{\Omega}2\tilde{\mu}\int_{\omega}(\nabla_{x}\phi+(D_{x}\phi)^{ij}\nabla_{y}\chi^{ij}):(D_{x}\phi+(D_{x}\phi)^{ij}(D_{y}\chi^{ij})) \displaystyle=\int_{\Omega}[\int_{Y\setminus\omega}\nabla_{x}\cdot(\lambda(\nabla\cdot\phi)I+2\mu D_{x}\phi)+(\lambda(\nabla_{y}\cdot\chi^{ij})I+2\mu(D_{y}\chi^{ij}))(\nabla_{x}(D_{x}\phi)^{ij})\mathop{}\!\mathrm{d}y \displaystyle\quad\quad+\int_{\omega}2\tilde{\mu}\nabla_{x}\cdot D_{x}\phi+(2\tilde{\mu}(D_{y}\chi^{ij})+r^{ij}I)\nabla_{x}(D_{x}\phi)^{ij}\mathop{}\!\mathrm{d}y] \displaystyle=\int_{R^{d}}\hat{A}\nabla\phi:\nabla\phi. \tag{46}
$$
So we have
$$
\frac{\min\{\lambda d+2\mu,2\mu,\tilde{\mu}\}}{2}\int_{R^{d}}|\nabla\phi|^{2}\,\mathop{}\!\mathrm{d}x\leq\liminf_{\varepsilon\rightarrow 0}\frac{\min\{\lambda d+2\mu,2\mu,\tilde{\mu}\}}{2}\int_{R^{d}}|\nabla\varphi|^{2}\,\mathop{}\!\mathrm{d}x\\
\leq\int_{R^{d}}\hat{A}\nabla\phi\cdot\nabla\phi\mathop{}\!\mathrm{d}x. \tag{47}
$$
Finally we can get
$$
\frac{\min\{\lambda d+2\mu,2\mu,\tilde{\mu}\}}{2}|\xi|^{2}|\eta|^{2}\leq\hat{a}_{ij}^{\alpha\beta}\xi_{i}\xi_{j}\eta^{\alpha}\eta^{\beta}.
$$
### B.6 Regularity
For each $i,j=1,\cdots,d$ , suppose $p^{ij}=\frac{1}{2}(y_{j}e_{i}+y_{i}e_{j})$ , $\chi^{ij}(y)$ and $r^{ij}(y)$ satisfy the following equation:
$$
\begin{cases}\begin{aligned} &\nabla\cdot\left[\lambda\nabla_{y}\cdot\chi^{ij}I+2\mu D_{y}u\right]=0&&\text{in }Y\setminus\omega,\\
&\nabla\cdot\left[r^{ij}I+2\tilde{\mu}D_{y}\chi^{ij}\right]=0&&\text{in }\omega,\\
&\nabla_{y}\cdot\chi^{ij}=-\delta_{ij}&&\text{in }\omega,\\
&\chi^{ij}|_{+}=\chi^{ij}|_{-},\\
&\left.\left[r^{ij}I+2\tilde{\mu}D_{y}\chi^{ij}\right]N\right|_{-}-\left.\left[\lambda\nabla_{y}\cdot\chi^{ij}I+2\mu D_{y}\chi^{ij}\right]N\right|_{+}\\
&\quad=-\tilde{\mu}(E_{ij}+E_{ji})N|_{-}+\mu(E_{ij}+E_{ji})N|_{+}+\lambda IN\delta_{ij}|_{+}&&\text{on }\partial\omega.\end{aligned}\end{cases}
$$
In $Y\setminus\omega$ and $\omega$ , by interior estimation for $V\subset\subset Y\setminus\omega,W\subset\subset\omega$ , we have $\|\chi\|_{H^{k}(V)}<\infty$ , $\|\chi\|_{H^{k}(W)}<\infty$ . So we can consider the following boundary regularity problem on $\partial\omega$ .
Consider:
$$
\begin{cases}\begin{aligned} &\lambda\Delta_{y}(\chi^{ij}+p^{ij})+(\lambda+\mu)\nabla_{y}(\nabla_{y}\cdot(\chi^{ij}+p^{ij}))=0&&\text{in }Y\setminus\omega,\\
&\tilde{\mu}\Delta_{y}(\chi^{ij}+p^{ij})+\nabla_{y}r^{ij}=0&&\text{in }\omega,\\
&\nabla_{y}\cdot(\chi^{ij}+p^{ij})=0&&\text{in }\omega,\\
&\chi^{ij}|_{+}=\chi^{ij}|_{-},\\
&\left.\left[r^{ij}I+2\tilde{\mu}D_{y}(\chi^{ij}+p^{ij})\right]N\right|_{-}-\left.\left[\lambda\nabla_{y}\cdot(\chi^{ij}+p^{ij})I+2\mu D_{y}(\chi^{ij}+p^{ij})\right]N\right|_{+}=0&&\text{on }\partial\omega.\end{aligned}\end{cases}
$$
This is equivalent to the following problem:
Suppose that there is a sphere $Q_{r}=(-r,r)^{d}$ with radius $r$ centered at the origin in $\mathbb{R}^{d}$ .
We define
$$
Q_{r}^{+}=\{(x^{\prime},x_{d})\in Q_{r}:x_{d}>\psi(x^{\prime})\},\quad Q_{r}^{-}=\{(x^{\prime},x_{d})\in Q_{r}:x_{d}<\psi(x^{\prime})\},\quad
$$
$$
I_{r}=\{(x^{\prime},x_{d})\in Q_{r}:x_{d}=\psi(x^{\prime})\}.
$$
Here $\psi:R^{d-1}\rightarrow R$ is a $C^{\infty}$ function and $\psi(0)=0$ ,
$$
\begin{cases}\begin{aligned} &\nabla_{y}\cdot\left[\lambda\nabla_{y}\cdot\chi^{ij}\,I+2\mu D_{y}\chi^{ij}\right]=0\quad&&\text{in }Q_{r}^{+},\\
&\nabla\cdot\left[r^{ij}\,I+2\tilde{\mu}D_{y}\chi^{ij}\right]=0\quad&&\text{in }Q_{r}^{-},\\
&\nabla\cdot\chi^{ij}=0\quad&&\text{in }Q_{r}^{-},\\
&\chi^{ij}|_{+}=\chi^{ij}|_{-},\\
&\left.\left[r^{ij}\,I+2\tilde{\mu}D_{y}\chi^{ij}\right]N\right|_{-}-\left.\left[\lambda\nabla_{y}\cdot\chi^{ij}\,I+2\mu D_{y}\chi^{ij}\right]N\right|_{+}=0\quad&&\text{on }I_{r}.\end{aligned}\end{cases}
$$
$\chi^{ij},r^{ij}$ are the weak solution in $H^{1}(Q_{r};\mathbb{R}^{d})$ and $L^{2}(Q_{r}^{-};\mathbb{R})$ .
#### B.6.1 A Basic $C^{1,\alpha}$ Estimate
Suppose $\Omega_{\pm}=\mathbb{R}^{d}_{\pm}$ , $S=\{x_{d}=0\}$ , $B_{+}=\{x\in B(1):x_{d}>0\}$ , and $B_{-}=\{x\in B(1):x_{d}<0\}$ , where $B(1)=\{x\in\mathbb{R}^{d}:\|x\|\leq 1\}$ .
Consider the following equation: for all $V\in H_{0}^{1}(B(1);\mathbb{R}^{d})$ ,
$$
\begin{cases}\begin{aligned} &(\nabla V:A_{1}\nabla\tilde{\chi})_{B_{+}}+(\nabla V:A_{2}\nabla\tilde{\chi})_{B_{-}}+(\tilde{r},\nabla\cdot(aV))_{B_{-}}=0,\\
&\nabla\cdot(a\tilde{\chi})=0.\end{aligned}\end{cases}
$$
where $\tilde{\chi}=D^{\alpha}\chi$ , $\tilde{r}=D^{\alpha}r$ with $|\alpha|\geq 1$ , $A_{1}$ , $A_{2}$ are constant tensors, and $a$ is a constant matrix.
We let $V=\eta^{2}\tilde{\chi}$ and $\eta$ be smooth with $\eta=0\ \text{in}\ B(1)^{c}$ and $\eta=1\text{ in }B(\frac{1}{2})$ . So we can get
| | $\displaystyle\eta^{2}(\nabla\tilde{\chi}:A_{1}\nabla\tilde{\chi})_{B_{+}}+\eta^{2}(\nabla\tilde{\chi}:A_{2}\nabla\tilde{\chi})_{B_{-}}+(\tilde{\chi}(\nabla\eta^{2})^{T}:A_{1}\nabla\tilde{\chi})_{B_{+}}$ | |
| --- | --- | --- |
Therefore,
$$
\|\eta\nabla\tilde{\chi}\|^{2}_{L^{2}(B(1))}\leq C\|\eta\nabla\tilde{\chi}\|_{L^{2}(B(1))}\|\tilde{\chi}\|_{L^{2}(B(1))}+C\|\eta\tilde{r}\|_{L^{2}(B(1))}\|\tilde{\chi}\|_{L^{2}(B(1))}. \tag{1}
$$
Besides, we also have $\|\eta\tilde{r}\|_{L^{2}(B(1))}\leq C\sum_{\beta=|\alpha|+1}\|\eta D^{\beta}\chi\|_{L^{2}(B(1))}=m_{|\alpha|+1}$ , and $\sum_{\beta=|\alpha|}\|D^{\beta}\chi\|_{L^{2}(B(1))}=n_{|\alpha|}$ .
So we have $m_{|\alpha|+1}^{2}\leq Cm_{|\alpha|+1}n_{|\alpha|}$ . That is
$$
\|D^{|\alpha|+1}\chi\|_{L^{2}(B(\frac{1}{2}))}\leq C\|D^{|\alpha|}\chi\|_{L^{2}(B(1))}. \tag{12}
$$
Similarly,
$$
\|D^{|\alpha|}r\|_{L^{2}(B(\frac{1}{2}))}\leq C\|D^{|\alpha|-1}r\|_{L^{2}(B(\frac{1}{2}))}. \tag{12}
$$
#### B.6.2 Schauder Theory
Since $\frac{\partial^{2}\chi}{\partial y_{d}^{2}}$ is the linear combination of $\frac{\partial^{2}\chi}{\partial y_{i}\partial y_{j}},\frac{\partial\chi}{\partial y_{j}},\frac{\partial r}{\partial y_{j}},r$ , the inequalities (Eq. 49 and Eq. 50) above are correct for any $|\alpha_{n}|=|\alpha|$ . So we have proved the Caccioppoli inequality.
**Lemma 2**
*(Caccioppoli Inequality) For the weak solutions $(\chi,r)$ , $\forall k\geq 1$
$$
\sum_{\pm}\|\chi\|_{H^{k}(B(\frac{1}{2},\pm))}\leq C\|\chi\|_{L^{2}(B(1))}, \tag{1}
$$
$$
\|r\|_{H^{k}(B(\frac{1}{2})_{-})}\leq C\|r\|_{L^{2}(B(\frac{1}{2})_{-})}. \tag{12}
$$*
**Lemma 3**
*Suppose that $M_{\textpm}$ is the constant matrix in $R^{d\times d}$ , the following are equivalent:
$$
\begin{cases}\forall y\in\{y_{d}=0\},\ M_{+}x=M_{-}x,\\
\exists c\in R^{d}\ s.t.\ M_{+}-M_{-}=Ce_{d}^{T},\\
(I-e_{d}^{T}e_{d})M_{+}=(I-e_{d}^{T}e_{d})M_{-}.\end{cases} \tag{53}
$$*
**Definition 1**
*Let $A_{1}$ , $A_{2}$ be constant tensors and $a$ be a constant matrix. Suppose $M_{+}$ and $M_{-}$ satisfy Lemma 53 above and $\nabla\cdot(aM_{-}y)=0$ in $B(t)_{-}$ . Define
$$
l(y)=M_{+}y\,\mathbf{1}_{y_{d}\geq 0}+M_{-}y\,\mathbf{1}_{y_{d}\leq 0}+C,\quad q(y)=r(0), \tag{0}
$$
where $\mathbf{1}_{\cdot}$ denotes the indicator function. We call $l$ , $q$ the piecewise linear solutions of the following equations:
$$
\begin{cases}\begin{aligned} &\nabla\cdot(A_{1}\nabla l)=0\quad&&\text{in }\mathbb{R}^{d}_{+},\\
&\nabla\cdot(A_{2}\nabla l)+\nabla\cdot(a^{T}q)=0\quad&&\text{in }\mathbb{R}^{d}_{-},\\
&\nabla\cdot(al)=0\quad&&\text{in }\mathbb{R}^{d}_{-},\\
&l|_{+}=l|_{-}\quad&&\text{on }\{x_{d}=0\},\\
&\left.\frac{\partial l}{\partial\nu}\right|_{+}-\left.\frac{\partial l}{\partial\nu}\right|_{-}=(A_{1}M_{+})e_{d}-(A_{2}M_{-}+a^{T}r(0))e_{d}\quad&&\text{on }\{x_{d}=0\}.\end{aligned}\end{cases} \tag{0}
$$ where $e_{d}$ is the standard basis vector in $\mathbb{R}^{d}$ . Let $\mathcal{L}$ be the space of all piecewise linear solutions of the above equations. For any $(l,q)\in\mathcal{L}$ , define
$$
\zeta(l,q)=\left(\frac{\partial l}{\partial\nu}\right)_{+}-\left(\frac{\partial l}{\partial\nu}\right)_{-}.
$$*
**Lemma 4**
*Let $A_{1}$ , $A_{2}$ be constant tensors and $a$ be a constant matrix. Consider the following system:
$$
\begin{cases}\begin{aligned} &\nabla\cdot(A_{1}\nabla\chi)=0\quad&&\text{in }B(1)_{+},\\
&\nabla\cdot(A_{2}\nabla\chi)+\nabla\cdot(a^{T}r)=0\quad&&\text{in }B(1)_{-},\\
&\nabla\cdot(a\chi)=0\quad&&\text{in }B(1)_{-},\\
&\chi|_{+}=\chi|_{-}\quad&&\text{on }B(1)\cap\{x_{d}=0\},\\
&\left.\frac{\partial\chi}{\partial\nu}\right|_{+}-\left.\frac{\partial\chi}{\partial\nu}\right|_{-}=g_{0}\quad&&\text{on }B(1)\cap\{x_{d}=0\}.\end{aligned}\end{cases} \tag{1}
$$
where $B(1)=\{x\in\mathbb{R}^{d}:\|x\|\leq 1\}$ , $B(1)_{+}=B(1)\cap\mathbb{R}^{d}_{+}$ , $B(1)_{-}=B(1)\cap\mathbb{R}^{d}_{-}$ , and $\mathbb{R}^{d}_{\pm}=\{x\in\mathbb{R}^{d}:\pm x_{d}>0\}$ . Let $\chi$ , $r$ be weak solutions of the above system. Then for all $k\geq 0$ and $\alpha\in[0,1]$ , we have
$$
\sum_{\pm}\|\chi\|_{H^{k}(B(\frac{1}{2})_{\pm})}\leq C\left(\|\chi\|_{L^{2}(B(1))}+|g_{0}|\right), \tag{12}
$$
where $C$ is a constant independent of $\chi$ , $r$ , and $B(\frac{1}{2})=\{x\in\mathbb{R}^{d}:\|x\|\leq\frac{1}{2}\}$ .*
**Lemma 5**
*Let $A_{1},A_{2}$ be constant tensors and $a$ be a constant matrix. Consider the system:
$$
\begin{cases}\begin{aligned} &\nabla\cdot(A_{1}\nabla\chi)=0\quad&&\text{in }B(1)_{+},\\
&\nabla\cdot(A_{2}\nabla\chi)+\nabla\cdot(a^{T}r)=0\quad&&\text{in }B(1)_{-},\\
&\nabla\cdot(a\chi)=0\quad&&\text{in }B(1)_{-},\\
&\chi|_{+}=\chi|_{-}\quad&&\text{on }B(1)\cap\{x_{d}=0\},\\
&\left.\frac{\partial\chi}{\partial\nu}\right|_{+}-\left.\frac{\partial\chi}{\partial\nu}\right|_{-}=g_{0}\quad&&\text{on }B(1)\cap\{x_{d}=0\}.\end{aligned}\end{cases} \tag{1}
$$ where $\chi$ , $r$ are weak solutions. Define
$$
l(y)=(\nabla\chi)_{+}(0)\,y\,\mathbf{1}_{y_{d}\geq 0}+(\nabla\chi)_{-}(0)\,y\,\mathbf{1}_{y_{d}\leq 0}+\chi(0),\quad q(y)=r(0), \tag{0}
$$
where $\mathbf{1}_{\cdot}$ denotes the indicator function. By Lemma 52, we know $\chi(0),(\nabla\chi(x))_{\pm}(0)$ are well-defined.
$$
(I-e_{d}\otimes e_{d})(\nabla\chi)_{+}=(I-e_{d}\otimes e_{d})(\nabla\chi)_{-}\quad\text{on }B(1)\cap\{y_{d}=0\}, \tag{1}
$$
where $e_{d}$ is the standard basis vector in $\mathbb{R}^{d}$ and $I$ is the identity matrix. From Lemma 53, it follows that $(l,q)\in\mathcal{L}$ . For some $\beta\in(0,1)$ and all $y\in B\left(\frac{1}{2}\right)$ , we have:
$$
\displaystyle|\chi(y)-l(y)| \displaystyle=\left|\chi(y)-\chi(0)-(\nabla\chi)_{\pm}(0)\,y\right| \displaystyle\leq C|y|^{\beta+1}\left([\chi]_{C^{1,\beta}(B\left(\frac{1}{2}\right)_{\pm})}\right) \displaystyle\leq C|y|^{\beta+1}\left(\left(\int_{B(1)}|\chi|^{2}\right)^{\frac{1}{2}}+|g_{0}|\right). \tag{0}
$$ For all $y\in B\left(\frac{1}{2}\right)_{-}$ :
$$
\displaystyle|r-q| \displaystyle\leq C|y|^{\beta}\left([r]_{C^{0,\beta}(B\left(\frac{t}{2}\right)_{-})}\right) \displaystyle\leq C|y|^{\beta}\left(\int_{B(1)_{-}}|r|^{2}\right)^{\frac{1}{2}}. \tag{1}
$$ Therefore, for some $\beta\in(0,1)$ and all $y\in B\left(\frac{t}{2}\right)$ :
$$
\displaystyle|\chi(y)-l(y)| \displaystyle=\left|\chi(y)-\chi(0)-(\nabla\chi)_{\pm}(0)\,y\right| \displaystyle\leq C\left|\frac{y}{t}\right|^{\beta+1}\left([\chi]_{C^{1,\beta}(B\left(\frac{t}{2}\right)_{\pm})}\right) \displaystyle\leq C\left|\frac{y}{t}\right|^{\beta+1}\left(\left(\fint_{B(t)}|\chi|^{2}\right)^{\frac{1}{2}}+t|g_{0}|\right), \tag{0}
$$
where $\fint$ denotes the average integral. For all $y\in B\left(\frac{t}{2}\right)_{-}$ :
| | $\displaystyle|r-q|$ | $\displaystyle\leq C\left|\frac{y}{t}\right|^{\beta}\left([r]_{C^{0,\beta}(B\left(\frac{t}{2}\right)_{-})}\right)$ | |
| --- | --- | --- | --- |*
**Lemma 6**
*Let $A_{1}$ , $A_{2}$ be constant tensors and $a$ be a constant matrix. Consider the system:
$$
\begin{cases}\begin{aligned} &\nabla\cdot(A_{1}\nabla\chi)=0\quad&&\text{in }B(1)_{+},\\
&\nabla\cdot(A_{2}\nabla\chi)+\nabla\cdot(a^{T}r)=0\quad&&\text{in }B(1)_{-},\\
&\nabla\cdot(a\chi)=0\quad&&\text{in }B(1)_{-},\\
&\chi|_{+}=\chi|_{-}\quad&&\text{on }B(1)\cap\{x_{d}=0\},\\
&\left.\frac{\partial\chi}{\partial\nu}\right|_{+}-\left.\frac{\partial\chi}{\partial\nu}\right|_{-}=g_{0}\quad&&\text{on }B(1)\cap\{x_{d}=0\}.\end{aligned}\end{cases} \tag{1}
$$
where $\chi$ , $r$ are weak solutions. Moreover, for all $\rho\in(0,t)$ , integrating the above inequalities yields:
$$
\left(\fint_{B(\rho)}|\chi-l|^{2}\right)^{\frac{1}{2}}+\rho\left|g_{0}-\zeta(l,q)\right|\leq\left|\frac{\rho}{t}\right|^{\beta+1}\left(\left(\fint_{B(t)}|\chi|^{2}\right)^{\frac{1}{2}}+t|g_{0}|\right). \tag{55}
$$ Thus, for all $(l^{\prime},q^{\prime})\in\mathcal{L}$ , by the inequality above, we have:
$$
\displaystyle\inf_{\begin{subarray}{c}(l,q)\in\mathcal{L}\end{subarray}}\left\{\left(\fint_{B(\rho)}|\chi-l|^{2}\right)^{\frac{1}{2}}+\rho\left|g_{0}-\zeta(l,q)\right|\right\} \displaystyle\leq C\left|\frac{\rho}{t}\right|^{\beta+1}\inf_{\begin{subarray}{c}(l,q)\in\mathcal{L}\end{subarray}}\left(\left(\fint_{B(t)}|\chi-l|^{2}\right)^{\frac{1}{2}}+t|g_{0}-\zeta(l,q)|\right). \tag{56}
$$ Further, for all $(l^{\prime},q^{\prime})\in\mathcal{L}$ , it follows that:
$$
\displaystyle\inf_{\begin{subarray}{c}(l,q)\in\mathcal{L}\end{subarray}}\left\{\left(\fint_{B(\rho)}|\chi-l|^{2}\right)^{\frac{1}{2}}+\rho\left|g_{0}-\zeta(l,q)\right|\right\} \displaystyle\leq C\left|\frac{\rho}{t}\right|^{\beta+1}\inf_{\begin{subarray}{c}(l,q)\in\mathcal{L}\end{subarray}}\left(\left(\fint_{B(t)}|\chi-l|^{2}\right)^{\frac{1}{2}}+t\left|g_{0}-\zeta(l,q)\right|\right). \tag{57}
$$*
**Lemma 7**
*Suppose $\phi:R_{+}\rightarrow R_{+}$ is a non-decreasing non-negative function satisfying $\phi(\rho)\leq C(\frac{\rho}{t})^{\beta}\phi(r)+Br^{\alpha}$ , where $\beta>\alpha>0,C>0$ . $\forall 0<\rho<r<R$ , $\exists C_{1}$ , s.t. $\phi(\rho)\leq C_{1}(\frac{\rho}{t})^{\alpha}\phi(r)+B\rho^{\alpha}$ .*
**Theorem 2**
*Suppose $A_{1}$ , $A_{2}$ , $a$ are $C^{\alpha}$ -Holder continuous, and let $S_{t}=B(t)\cap\{x_{d}=0\}$ where $B(t)=\{x\in\mathbb{R}^{d}:\|x\|\leq t\}$ . Let $\chi$ , $r$ be weak solutions to the following system: $\forall V\in H_{0}^{1}(B(1);\mathbb{R}^{d})$ ,
$$
\begin{cases}\begin{aligned} &(\nabla V:A_{1}\nabla\chi)_{B_{+}}+(\nabla V:A_{2}\nabla\chi)_{B_{-}}+(r,\nabla\cdot(a\chi))_{B_{-}}=0\\
&\nabla\cdot(a\chi)=0\end{aligned}\end{cases}
$$
where $B_{+}=B(1)\cap\mathbb{R}^{d}_{+}$ , $B_{-}=B(1)\cap\mathbb{R}^{d}_{-}$ , and $\mathbb{R}^{d}_{\pm}=\{x\in\mathbb{R}^{d}:\pm x_{d}>0\}$ . Then we have:
$$
\sum_{\pm}\|\chi\|_{C^{1,\alpha}(B\left(\frac{1}{2}\right)_{\pm})}\leq C\|\chi\|_{L^{2}(B(1))}, \tag{12}
$$
where $C$ is a constant independent of $\chi$ , and $C^{1,\alpha}(\Omega)$ denotes the Holder space of functions with $\alpha$ -Holder continuous first derivatives. Equivalent Formulation: This estimate is equivalent to showing that $\forall\rho\in(0,\frac{1}{4})$ ,
$$
\inf_{\begin{subarray}{c}(l,q)\in\mathcal{L}\end{subarray}}\left(\fint_{B(\rho)}|\chi-l|^{2}\right)^{\frac{1}{2}}\leq C\rho^{1+\alpha}\sum_{\pm}\|\chi\|_{L^{2}(B\left(\frac{3}{4}\right))}, \tag{34}
$$
here $\mathcal{L}$ is the space of piecewise linear functions $(l,q)$ as defined earlier, $B(\rho)$ is a small ball with center arbitrary on $S_{\frac{3}{4}}=B\left(\frac{3}{4}\right)\cap\{x_{d}=0\}$ and $\fint$ denotes the average integral over the domain.*
* Disturbance*
Suppose $w_{t},s_{t}$ are solutions to the following equations, where $A_{1}^{0}=A_{1}(0)$ , $A_{2}^{0}=A_{2}(0)$ , $a^{0}=a(0)$ : $\forall V\in H_{0}^{1}(B(t);\mathbb{R}^{d})$ ,
$$
\begin{cases}\begin{aligned} &(\nabla V:A_{1}^{0}\nabla w_{t})_{B_{+}}+(\nabla V:A_{2}^{0}\nabla w_{t})_{B_{-}}+\left(s_{t},\nabla\cdot(a^{0}V)\right)_{B_{-}}=0,\\
&\nabla\cdot(a^{0}w_{t})=0\quad&&\text{in }B_{-},\\
&w_{t}=\chi\quad&&\text{on }\partial B(t),\\
&s_{t}=r\quad&&\text{on }\partial B(t)_{-}.\end{aligned}\end{cases}
$$
where $B_{+}=B(t)\cap\mathbb{R}^{d}_{+}$ , $B_{-}=B(t)\cap\mathbb{R}^{d}_{-}$ , and $B(t)=\{x\in\mathbb{R}^{d}:\|x\|\leq t\}$ . By the continuity result in Lemma 4, we have $\forall\rho\in(0,t)$ :
| | $\displaystyle\left(\fint_{B(\rho)}|\nabla w_{t}|^{2}\right)^{\frac{1}{2}}+\left(\fint_{B(\rho)_{-}}|s_{t}|^{2}\right)^{\frac{1}{2}}$ | |
| --- | --- | --- |
where $C$ is a constant independent of $t$ , $\rho$ . Moreover, from the equations, we derive:
| | $\displaystyle\left(\nabla\phi,A^{0}\nabla(w_{t}-\chi)\right)_{B(t)}$ | $\displaystyle=\left(\nabla\phi,(A-A^{0})\nabla\chi\right)_{B(t)}$ | |
| --- | --- | --- | --- |
where $A^{0}$ denotes the piecewise constant tensor $A_{1}^{0}$ on $B_{+}$ and $A_{2}^{0}$ on $B_{-}$ . Let $\phi=w_{t}-\chi$ . Substituting into the above equation, we obtain:
$$
\displaystyle\Lambda^{-1}\|\nabla w_{t}-\nabla\chi\|_{L^{2}(B(t))}^{2}\leq\int_{B(t)}|A-A^{0}|\,|\nabla\chi|\,|\nabla w_{t}-\nabla\chi| \displaystyle\quad+\int_{B(t)_{-}}(r-s_{t})\cdot\left(a^{0}(-\nabla\chi)\right)+\int_{B(t)_{-}}r(a-a^{0})\nabla(w_{t}-\chi) \displaystyle\leq\int_{B(t)}|A-A^{0}|\,|\nabla\chi|\,|\nabla w_{t}-\nabla\chi|+\int_{B(t)_{-}}|r-s_{t}|\,|a^{0}|\,|\nabla\chi| \displaystyle\quad+\int_{B(t)_{-}}|r|\,|a-a^{0}|\,|\nabla w_{t}-\nabla\chi|, \tag{59}
$$
where $\Lambda>0$ is the ellipticity constant of $A^{0}$ . By Youngâs inequality, this implies:
$$
\displaystyle\int_{B(t)}|\nabla w_{t}-\nabla\chi|^{2}+\int_{B(t)_{-}}|r-s_{t}|^{2}\leq C\left[t^{2\alpha}\int_{B(t)}|\nabla\chi|^{2}+t^{2\alpha}\int_{B(t)_{-}}|r|^{2}\right], \tag{60}
$$
where $C$ depends on $\Lambda$ , $\|A\|_{C^{\alpha}}$ , $\|a\|_{C^{\alpha}}$ , and $\alpha$ is the Holder exponent. â
* Morreyâs Estimate and Bootstrap Analysis*
By the above analysis, we obtain the following result:
$$
\displaystyle\int_{B(\rho)}|\chi|^{2}+\int_{B(\rho)_{-}}|r|^{2} \displaystyle\leq C\left(\int_{B(\rho)}|\nabla\chi-\nabla w_{t}|^{2}+\int_{B(\rho)}|\nabla w_{t}|^{2}+\int_{B(\rho)_{-}}|r-s_{t}|^{2}+\int_{B(\rho)_{-}}|s_{t}|^{2}\right) \displaystyle\leq C\left(\left(\frac{\rho}{t}\right)^{d}\left(\int_{B(t)}|\nabla\chi|^{2}+\int_{B(t)_{-}}|r|^{2}\right)+t^{2\alpha}\left(\int_{B(t)}|\nabla\chi|^{2}+\int_{B(t)_{-}}|r|^{2}\right)\right). \tag{61}
$$ Define $\Psi(r)=\int_{B(r)}|\nabla\chi|^{2}+\int_{B(r)_{-}}|r|^{2}$ . Then we have:
$$
\Psi(\rho)\leq C\left(\left(\frac{\rho}{t}\right)^{d}\Psi(t)+t^{2\alpha}\Psi(t)\right)\quad\forall\,0<\rho<t<\frac{1}{2}.
$$ By Lemma 7, it follows that:
$$
\displaystyle\Psi(\rho)\leq C\rho^{2\alpha}\Psi\left(\frac{1}{2}\right)\quad\forall\,\rho\in\left(0,\frac{1}{2}\right). \tag{12}
$$ Thus, we derive:
$$
\displaystyle\Psi(\rho)\leq C\left(\left(\frac{\rho}{t}\right)^{d}\Psi(t)+t^{4\alpha}\Psi(t)\right). \tag{63}
$$ By bootstrap analysis, for all $r<d$ and $0<\rho<\frac{1}{2}$ , we have:
$$
\displaystyle\Psi(\rho)\leq C\rho^{r}\Psi\left(\frac{1}{2}\right). \tag{12}
$$ The above estimate holds for any ball with center in $B\left(\frac{1}{2}\right)$ and radius $t$ . Finally, we conclude that $|\nabla\chi|\in L^{2,r}(B\left(\frac{1}{2}\right))$ , which implies $\chi$ is $C^{\beta}$ -Holder continuous for all $\beta\in(0,1)$ . â
* C1,αC^{1,\alpha}Continuity*
Combining Lemma 57 with the above inequality and using the Poincaré inequality, we have:
$$
\displaystyle\inf_{\begin{subarray}{c}(l,q)\in\mathcal{L}\end{subarray}}\left\{\int_{B(\rho)}|\chi-l|^{2}+\rho^{d+2}|\zeta(l,q)|^{2}\right\} \displaystyle\leq\inf_{\begin{subarray}{c}(l,q)\in\mathcal{L}\end{subarray}}\left\{\int_{B(\rho)}|w_{t}-l|^{2}+\rho^{d+2}|\zeta(l,q)|^{2}\right\}+\int_{B(\rho)}|\chi-w_{t}|^{2} \displaystyle\leq C\left(\frac{\rho}{t}\right)^{2\beta+2+d}\inf_{\begin{subarray}{c}(l,q)\in\mathcal{L}\end{subarray}}\left\{\int_{B(t)}|w_{t}-l|^{2}+t^{d+2}|\zeta(l,q)|^{2}\right\}+C\int_{B(t)}|\chi-w_{t}|^{2} \displaystyle\leq C\left(\frac{\rho}{t}\right)^{2\beta+2+d}\inf_{\begin{subarray}{c}(l,q)\in\mathcal{L}\end{subarray}}\left\{\int_{B(t)}|\chi-l|^{2}+t^{d+2}|\zeta(l,q)|^{2}\right\} \displaystyle\quad+C\left(\frac{\rho}{t}\right)^{2\beta+2+d}\int_{B(t)}|\chi-w_{t}|^{2}+C\int_{B(t)}|\chi-w_{t}|^{2} \displaystyle\leq C\left(\frac{\rho}{t}\right)^{2\beta+2+d}\inf_{\begin{subarray}{c}(l,q)\in\mathcal{L}\end{subarray}}\left\{\int_{B(t)}|\chi-l|^{2}+t^{d+2}|\zeta(l,q)|^{2}\right\}+Ct^{2}\int_{B(t)}|\nabla\chi-\nabla w_{t}|^{2} \displaystyle\leq C\left(\frac{\rho}{t}\right)^{2\beta+2+d}\inf_{\begin{subarray}{c}(l,q)\in\mathcal{L}\end{subarray}}\left\{\int_{B(t)}|\chi-l|^{2}+t^{d+2}|\zeta(l,q)|^{2}\right\}+Ct^{2+2\alpha}\left(\int_{B(t)}|\nabla\chi|^{2}+\int_{B(t)_{-}}|r|^{2}\right). \tag{65}
$$ Define
$$
\displaystyle\Phi(r)=\inf_{\begin{subarray}{c}l\in\mathcal{L}\\
q\in\mathcal{L}\end{subarray}}\left\{\int_{B(r)}|\chi-l|^{2}+r^{d+2}|\zeta(l,q)|^{2}\right\}, \tag{66}
$$
and
$$
\displaystyle b(t)=\int_{B(t)}|\nabla\chi|^{2}+\int_{B(t)_{-}}|r|^{2}. \tag{67}
$$
Then we have
$$
\displaystyle b(t)\leq Ct^{r}\Psi\left(\frac{1}{2}\right)=Ct^{d-\alpha}\Psi\left(\frac{1}{2}\right) \tag{12}
$$
where we set $r=d-\alpha$ . â
Moreover, it follows that:
$$
\displaystyle\Phi(\rho)\leq C\left(\frac{\rho}{t}\right)^{d+2\beta+2}\Phi(t)+Ct^{d+2+\alpha}\Psi\left(\frac{1}{2}\right). \tag{12}
$$
For $\beta\in(\alpha,1)$ , by Lemma 7, we obtain:
$$
\displaystyle\Phi(\rho)\leq C\rho^{d+\alpha+2}\left(\Phi\left(\frac{1}{2}\right)+\Psi\left(\frac{1}{2}\right)\right). \tag{12}
$$
This implies $\chi\in C^{1,\frac{\alpha}{2}}(\overline{B({\frac{1}{2}})_{\pm}};\mathbb{R}^{d})$ for all $\forall\rho\in(0,\frac{1}{2})$ . By Lemma 7, we further derive:
$$
\displaystyle\Phi(\rho)\leq C\rho^{d+2\alpha+2}\left(\Phi\left(\frac{1}{2}\right)+\Psi\left(\frac{1}{2}\right)\right), \tag{12}
$$
which implies $\chi\in C^{1,\alpha}(\overline{B({\frac{1}{2}})_{\pm}};\mathbb{R}^{d})$ for all $\rho\in(0,\frac{1}{2})$ .
### B.7 Estimation
We define the cut-off function as follows: Let $\eta_{\varepsilon}\in C_{0}^{\infty}(\Omega)$ , which satisfies:
$$
\begin{cases}0\leq\eta_{\varepsilon}(x)\leq 1,&\forall x\in\Omega,\\
|\nabla\eta_{\varepsilon}(x)|\leq\frac{C}{\varepsilon},&\text{within the support of }\eta_{\varepsilon}\ (C\text{ is a constant independent of }\varepsilon),\\
\eta_{\varepsilon}(x)=1,&\text{if }\text{dist}(x,\partial\Omega)\geq 4\varepsilon,\\
\eta_{\varepsilon}(x)=0,&\text{if }\text{dist}(x,\partial\Omega)\leq 3\varepsilon.\end{cases}
$$
And we also set this convolution operator:
$$
S_{\varepsilon}(f)(x)=\rho_{\varepsilon}\ast f(x)=\int_{R^{d}}f(x-y)\rho_{\varepsilon}\mathop{}\!\mathrm{d}y,\ \text{here}\ \rho\in C_{0}^{\infty}(B(0,\frac{1}{2})),\rho\geq 0,\text{ and }\int_{R^{d}}\rho\,\mathop{}\!\mathrm{d}x=1.
$$
We let
$$
w_{\varepsilon}=u_{\varepsilon}-u_{0}-\varepsilon\chi(\frac{x}{\varepsilon})\eta_{\varepsilon}S_{\varepsilon}^{2}(\nabla u_{0}).
$$
We let $B(y)=A+A\nabla\chi(y)+\nabla\chi(y)A+\nabla\chi A\nabla\chi-\hat{A}$ be the Flux Corrector.
**Lemma 8**
*Suppose $\Omega$ is a bounded Lipschitz domain in $R^{d}$ , $\Omega_{t}=\{x\in\Omega:dist(x,\partial\Omega)<t\},t>0$ , $\forall\psi\in H_{0}^{1}(\Omega,R^{d})$ , we have
$$
\displaystyle|\int_{\Omega}A\nabla w_{\varepsilon}:\nabla\psi\,\mathop{}\!\mathrm{d}x|\leq C\|\nabla\chi\|_{L^{2}(\Omega)}\{\varepsilon\|S_{\varepsilon}(\nabla^{2}u_{0})\|_{L^{2}(\Omega\setminus\Omega_{3\varepsilon})}+\|\nabla u_{0}-S_{\varepsilon}(\nabla u_{0})\|_{L^{2}(\Omega\setminus\Omega_{2\varepsilon})}\} \displaystyle+C\|\nabla\psi\|_{L^{2}(\Omega_{4\varepsilon})}\|\nabla u_{0}\|_{L^{2}(\Omega_{5\varepsilon})}. \tag{72}
$$*
* Proof*
| | $\displaystyle A\nabla w_{\varepsilon}$ | $\displaystyle=A\nabla u_{\varepsilon}-A\nabla u_{0}-A\nabla\chi\left(\frac{x}{\varepsilon}\right)S_{\varepsilon}^{2}(\nabla u_{0})-\varepsilon A\chi(y)\nabla\left(\eta_{\varepsilon}S_{\varepsilon}^{2}(\nabla u_{0})\right)$ | |
| --- | --- | --- | --- |
Because of the variational form:
$$
\int_{\Omega}A\left(\frac{x}{\varepsilon}\right)\nabla u_{\varepsilon}:\nabla\psi+\int_{\Omega}p^{\varepsilon}\cdot\mathrm{div}\,\psi=0,\quad\forall\psi\in C_{0}^{\infty}(\Omega;\mathbb{R}^{m}),
$$ we can get:
$$
\displaystyle\int_{\Omega}A\nabla w_{\varepsilon}:\nabla\psi \displaystyle=\int_{\Omega}A\nabla u_{\varepsilon}:\nabla\psi-\hat{A}\nabla u_{0}:\nabla\psi+\int_{D_{\varepsilon}}p^{\varepsilon}\cdot\mathrm{div}\,\psi \displaystyle\quad+\int_{\Omega}(\hat{A}-A)\left(\nabla u_{0}-\eta_{\varepsilon}S_{\varepsilon}^{2}(\nabla u_{0})\right):\nabla\psi \displaystyle\quad+\int_{\Omega}\left[\nabla\chi A\eta_{\varepsilon}S_{\varepsilon}^{2}(\nabla u_{0})+\nabla\chi A\nabla\chi\eta_{\varepsilon}S_{\varepsilon}^{2}(\nabla u_{0})\right]:\nabla\psi \displaystyle\quad-\int_{D_{\varepsilon}}p^{\varepsilon}\cdot\operatorname{div}\,\psi-\int_{\Omega}B\eta_{\varepsilon}S_{\varepsilon}^{2}(\nabla u_{0}):\nabla\psi \displaystyle\quad-\int_{\Omega}\varepsilon A\chi\left(\frac{x}{\varepsilon}\right)\nabla\left(\eta_{\varepsilon}S_{\varepsilon}^{2}(\nabla u_{0})\right):\nabla\psi. \tag{73}
$$ The first three terms equal zero, then we can see:
$$
\displaystyle\int_{\Omega}(\hat{A}-A)\left(\nabla u_{0}-\eta_{\varepsilon}S_{\varepsilon}^{2}(\nabla u_{0})\right):\nabla\psi \displaystyle\leq C\int_{\Omega}(1-\eta_{\varepsilon})|\nabla u_{0}||\nabla\psi|\,\mathop{}\!\mathrm{d}x \displaystyle\quad+C\int_{\Omega}\eta_{\varepsilon}|\nabla u_{0}-S_{\varepsilon}^{2}(\nabla u_{0})||\nabla\psi|\,\mathop{}\!\mathrm{d}x \displaystyle\leq C\|\nabla u_{0}\|_{L^{2}(\Omega_{4\varepsilon})}\|\nabla\psi\|_{L^{2}(\Omega_{4\varepsilon})} \displaystyle\quad+C\|\nabla u_{0}-S_{\varepsilon}(\nabla u_{0})\|_{L^{2}(\Omega\setminus\Omega_{2\varepsilon})}\|\nabla\psi\|_{L^{2}(\Omega)}. \tag{74}
$$ This is because:
$$
\displaystyle\|\nabla u_{0}-S_{\varepsilon}^{2}(\nabla u_{0})\|_{L^{2}(\Omega\setminus\Omega_{3\varepsilon})} \displaystyle\leq\|\nabla u_{0}-S_{\varepsilon}(\nabla u_{0})\|_{L^{2}(\Omega\setminus\Omega_{3\varepsilon})} \displaystyle\quad+\|S_{\varepsilon}(\nabla u_{0})-S_{\varepsilon}^{2}(\nabla u_{0})\|_{L^{2}(\Omega\setminus\Omega_{3\varepsilon})} \displaystyle\leq C\|\nabla u_{0}-S_{\varepsilon}(\nabla u_{0})\|_{L^{2}(\Omega\setminus\Omega_{2\varepsilon})}. \tag{75}
$$ Moreover, by the regularity of $\chi\in W^{1,\infty}$ , we have
$$
\displaystyle\int_{\Omega}\left[\nabla\chi A\eta_{\varepsilon}S_{\varepsilon}^{2}(\nabla u_{0})+\nabla\chi A\nabla\chi\eta_{\varepsilon}S_{\varepsilon}^{2}(\nabla u_{0})\right]:\nabla\psi-\int_{D_{\varepsilon}}p^{\varepsilon}\cdot\operatorname{div}\,\psi \displaystyle\leq\int_{\Omega}\left[(\nabla\chi A+\nabla\chi A\nabla\chi)\left(\eta_{\varepsilon}S_{\varepsilon}^{2}(\nabla u_{0})-\nabla u_{0}\right)\right]:\nabla\psi \displaystyle\quad+\int_{\Omega}(\nabla\chi A+\nabla\chi A\nabla\chi)\nabla u_{0}:\nabla\psi-\int_{\omega}p_{0}\operatorname{div}\,\psi+\mathcal{O}(\varepsilon) \displaystyle=\int_{\Omega}\left[(\nabla\chi A+\nabla\chi A\nabla\chi)\left(\eta_{\varepsilon}S_{\varepsilon}^{2}(\nabla u_{0})-\nabla u_{0}\right)\right]:\nabla\psi \displaystyle\quad+\int_{\omega}\delta_{j\beta}r^{i\alpha}\frac{\partial u_{0}^{\alpha}}{\partial x_{i}}\frac{\partial\psi^{\beta}}{\partial x_{j}}-\int_{\omega}\frac{\partial u^{i}}{\partial x_{j}}r^{ij}\frac{\partial\psi^{k}}{\partial x_{k}}+\mathcal{O}(\varepsilon) \displaystyle=\int_{\Omega}\left[(\nabla\chi A+\nabla\chi A\nabla\chi)\left(\eta_{\varepsilon}S_{\varepsilon}^{2}(\nabla u_{0})-\nabla u_{0}\right)\right]:\nabla\psi+\mathcal{O}(\varepsilon) \displaystyle\leq C\|\nabla u_{0}-S_{\varepsilon}(\nabla u_{0})\|_{L^{2}(\Omega\setminus\Omega_{2\varepsilon})}\|\nabla\psi\|_{L^{2}(\Omega)}+\|\nabla u_{0}\|_{\Omega_{4\varepsilon}}\|\nabla\psi\|_{L^{2}(\Omega_{4\varepsilon}}+\mathcal{O}(\varepsilon). \tag{76}
$$ And we also have
$$
\displaystyle\left|\int_{\Omega}B\eta_{\varepsilon}S_{\varepsilon}^{2}(\nabla u_{0}):\nabla\psi\right| \displaystyle=\left|\int_{\Omega}b_{ij}^{\alpha\beta}\left(\frac{x}{\varepsilon}\right)S_{\varepsilon}^{2}\left(\frac{\partial u_{0}^{\beta}}{\partial x_{j}}\right)\left(\frac{\nabla\psi^{\alpha}}{\partial x_{i}}\right)\eta_{\varepsilon}\right| \displaystyle=\left|\int_{\Omega}\varepsilon\frac{\partial}{\partial x_{k}}\left(\phi_{kij}^{\alpha\beta}\right)\left(\frac{x}{\varepsilon}\right)\frac{\partial\psi^{\alpha}}{\partial x_{i}}S_{\varepsilon}^{2}\left(\frac{\partial u_{0}^{\beta}}{\partial x_{j}}\right)\eta_{\varepsilon}\right| \displaystyle\leq C\varepsilon\left[\int_{\Omega}\eta_{\varepsilon}\left|\phi\left(\frac{x}{\varepsilon}\right)\right|\left|\nabla\psi\right|\left|S_{\varepsilon}^{2}(\nabla^{2}u_{0})\right|\mathop{}\!\mathrm{d}x\right. \displaystyle\quad\left.+\int_{\Omega}\left|\nabla\eta_{\varepsilon}\right|\left|\phi\left(\frac{x}{\varepsilon}\right)\right|\left|\nabla\psi\right|\left|S_{\varepsilon}^{2}(\nabla u_{0})\right|\mathop{}\!\mathrm{d}x\right] \displaystyle\leq C\left(\varepsilon\|\nabla\psi\|_{L^{2}(\Omega)}\|S_{\varepsilon}(\nabla^{2}u_{0})\|_{L^{2}(\Omega\setminus\Omega_{2\varepsilon})}\right. \displaystyle\quad\left.+\|\nabla\psi\|_{L^{2}(\Omega_{4\varepsilon})}\|\nabla u_{0}\|_{L^{2}(\Omega_{5\varepsilon})}\right). \tag{77}
$$ By the calculations above, we can finally get the following result
$$
\displaystyle\int_{\Omega}A\left(\frac{x}{\varepsilon}\right)\nabla u_{\varepsilon}:\nabla\psi \displaystyle\leq C\|\nabla\psi_{\varepsilon}\|_{L^{2}(\Omega)}\left\{\varepsilon\|S_{\varepsilon}(\nabla u_{0})\|_{L^{2}(\Omega\setminus\Omega_{3\varepsilon})}\right. \displaystyle\quad\left.+\|\nabla u_{0}-S_{\varepsilon}(\nabla u_{0})\|_{L^{2}(\Omega\setminus\Omega_{2\varepsilon})}\right\} \displaystyle\quad+C\|\nabla w_{\varepsilon}\|_{L^{2}(\Omega_{4\varepsilon})}\|\nabla u_{0}\|_{L^{2}(\Omega_{5\varepsilon})}. \tag{78}
$$
â
**Theorem 3**
*Let $\Omega\subset\mathbb{R}^{d}$ be a bounded Lipschitz domain. Then for all $0<\varepsilon<1$ ,
$$
\displaystyle\|\nabla w_{\varepsilon}\|_{L^{2}(\Omega)}\leq C\left\{\varepsilon\|\nabla^{2}u_{0}\|_{L^{2}(\Omega\setminus\Omega_{\varepsilon})}+\|\nabla u_{0}\|_{L^{2}(\Omega_{5\varepsilon})}\right\}. \tag{79}
$$
Thus,
$$
\displaystyle\|w_{\varepsilon}\|_{H_{0}^{1}(\Omega)}\leq C\varepsilon^{\frac{1}{2}}\|u_{0}\|_{H^{2}(\Omega)}, \tag{80}
$$
where $C$ is a constant depending on $\mu$ and $\Omega$ .*
* Proof*
Let $\tilde{\eta}_{\varepsilon}\in C_{0}^{\infty}(\Omega)$ satisfy $0\leq\tilde{\eta}_{\varepsilon}\leq 1$ , $\tilde{\eta}_{\varepsilon}=0$ in $\Omega_{\varepsilon}$ , $\tilde{\eta}_{\varepsilon}=1$ in $\Omega\setminus\Omega_{\frac{3\varepsilon}{2}}$ , and $|\nabla\tilde{\eta}_{\varepsilon}|\leq\frac{C}{\varepsilon}$ . We can obtain
$$
\displaystyle\|\nabla u_{0}-S_{\varepsilon}(\nabla u_{0})\|_{L^{2}(\Omega\setminus\Omega_{2\varepsilon})} \displaystyle\leq\|\tilde{\eta}_{\varepsilon}(\nabla u_{0})-S_{\varepsilon}(\tilde{\eta}_{\varepsilon}\nabla u_{0})\|_{L^{2}(\mathbb{R}^{d})} \displaystyle\leq C\varepsilon\|\nabla(\tilde{\eta}_{\varepsilon}\nabla u_{0})\|_{L^{2}(\mathbb{R}^{d})} \displaystyle\leq C\left\{\varepsilon\|\nabla^{2}u_{0}\|_{L^{2}(\Omega\setminus\Omega_{\varepsilon})}+\|\nabla u_{0}\|_{L^{2}(\Omega_{2\varepsilon})}\right\}. \tag{81}
$$ By the inequality $\|f\|_{L^{2}(\Omega_{t})}\leq Ct^{\frac{1}{2}}\|f\|_{H^{1}(\Omega)}$ , we have
$$
\displaystyle\|\nabla u_{0}\|_{L^{2}(\Omega_{5\varepsilon})}\leq C\varepsilon^{\frac{1}{2}}\|u_{0}\|_{H^{2}(\Omega)}. \tag{82}
$$ By Lemma 72, set $\psi=w_{\varepsilon}\in H_{0}^{1}(\Omega;\mathbb{R}^{d})$ . Using the ellipticity of $A$ , we get
$$
\displaystyle\|\nabla w_{\varepsilon}\|_{L^{2}(\Omega)} \displaystyle\leq C\left\{\varepsilon\|S_{\varepsilon}(\nabla^{2}u_{0})\|_{L^{2}(\Omega\setminus\Omega_{3\varepsilon})}+\|\nabla u_{0}-S_{\varepsilon}(\nabla u_{0})\|_{L^{2}(\Omega\setminus\Omega_{2\varepsilon})}\right\} \displaystyle\quad+C\|\nabla u_{0}\|_{L^{2}(\Omega_{5\varepsilon})}\leq C\varepsilon^{\frac{1}{2}}\|u_{0}\|_{H^{2}(\Omega)}. \tag{83}
$$
â
**Theorem 4**
*Let $\Omega\subset\mathbb{R}^{d}$ be a bounded Lipschitz domain. Then for all $0<\varepsilon<1$ , if $u_{0}\in H^{2}(\Omega;\mathbb{R}^{d})$ , we have
$$
\displaystyle\|u_{\varepsilon}-u_{0}-\varepsilon\chi\left(\frac{x}{\varepsilon}\right)\nabla u_{0}\|_{H^{1}(\Omega)}\leq C(\mu,\Omega,\|\chi\|_{\infty})\varepsilon^{\frac{1}{2}}\|u_{0}\|_{W^{2,d}(\Omega)}. \tag{84}
$$*
* Proof*
The key step is to prove:
$$
\displaystyle\left\|\varepsilon\chi\left(\frac{x}{\varepsilon}\right)\nabla u_{0}-\varepsilon\chi\left(\frac{x}{\varepsilon}\right)\eta_{\varepsilon}S_{\varepsilon}^{2}(\nabla u_{0})\right\|_{H^{1}(\Omega)}\leq C\varepsilon^{\frac{1}{2}}\|u_{0}\|_{H^{2}(\Omega)}. \tag{85}
$$ We can find that
$$
\displaystyle\left\|\varepsilon\chi\left(\frac{x}{\varepsilon}\right)\nabla u_{0}-\varepsilon\chi\left(\frac{x}{\varepsilon}\right)\eta_{\varepsilon}S_{\varepsilon}^{2}(\nabla u_{0})\right\|_{H^{1}(\Omega)}\leq C\varepsilon\left\|\chi\left(\frac{x}{\varepsilon}\right)\left(\nabla u_{0}-\eta_{\varepsilon}S_{\varepsilon}^{2}(\nabla u_{0})\right)\right\|_{L^{2}(\Omega)} \displaystyle+C\left\|\nabla\chi\left(\frac{x}{\varepsilon}\right)\left(\nabla u_{0}-\eta_{\varepsilon}S_{\varepsilon}^{2}(\nabla u_{0})\right)\right\|_{L^{2}(\Omega)}+C\varepsilon\left\|\chi\left(\frac{x}{\varepsilon}\right)\nabla\left(\nabla u_{0}-\eta_{\varepsilon}S_{\varepsilon}^{2}(\nabla u_{0})\right)\right\|_{L^{2}(\Omega)} \displaystyle\leq C\varepsilon\left\|\nabla\left(\nabla u_{0}-\eta_{\varepsilon}S_{\varepsilon}^{2}(\nabla u_{0})\right)\right\|_{L^{2}(\Omega)}+C\left\|\nabla u_{0}-\eta_{\varepsilon}S_{\varepsilon}^{2}(\nabla u_{0})\right\|_{L^{2}(\Omega)}. \tag{86}
$$ On the one hand, we have
$$
\displaystyle\varepsilon\left\|\nabla\left(\nabla u_{0}-\eta_{\varepsilon}S_{\varepsilon}^{2}(\nabla u_{0})\right)\right\|_{L^{2}(\Omega)} \displaystyle\leq\varepsilon\left\|\nabla^{2}u_{0}\right\|_{L^{2}(\Omega)}+\varepsilon\left\|\nabla\left(\eta_{\varepsilon}S_{\varepsilon}^{2}(\nabla u_{0})\right)\right\|_{L^{2}(\Omega)} \displaystyle\leq C\varepsilon\|u_{0}\|_{H^{2}(\Omega)}+C\left\|\nabla u_{0}\right\|_{L^{2}(\Omega_{5\varepsilon})} \displaystyle\leq C\varepsilon^{\frac{1}{2}}\|u_{0}\|_{H^{2}(\Omega)}. \tag{87}
$$ On the other hand,
$$
\displaystyle\left\|\nabla u_{0}-\eta_{\varepsilon}S_{\varepsilon}^{2}(\nabla u_{0})\right\|_{L^{2}(\Omega)} \displaystyle\leq C\left\|\nabla u_{0}\right\|_{L^{2}(\Omega_{5\varepsilon})}+C\left\|\nabla u_{0}-S_{\varepsilon}^{2}(\nabla u_{0})\right\|_{L^{2}(\Omega\setminus\Omega_{4\varepsilon})} \displaystyle\leq C\varepsilon^{\frac{1}{2}}\|u_{0}\|_{H^{2}(\Omega)}. \tag{88}
$$
â
### B.8 Summary
These conclusions is consistent to the expected conclusion (Eq. 11), since we have:
$$
\displaystyle\|u_{\varepsilon}-u_{\lim}\|_{H^{1}(\Omega)} \displaystyle=\|u_{\varepsilon}-u_{0}-\varepsilon\chi\left(\frac{x}{\varepsilon}\right)\nabla u_{0}\|_{H^{1}(\Omega)} \displaystyle\leq C\|w_{\varepsilon}\|_{H^{1}(\Omega)}+C\|\varepsilon\chi\left(\frac{x}{\varepsilon}\right)\nabla u_{0}-\varepsilon\chi\left(\frac{x}{\varepsilon}\right)\eta_{\varepsilon}S_{\varepsilon}^{2}(\nabla u_{0})\|_{H^{1}(\Omega)} \displaystyle\leq C\varepsilon^{\frac{1}{2}}\|u_{0}\|_{H^{2}(\Omega)}. \tag{89}
$$