# Deep Sound Field Reconstruction in Real Rooms: Introducing the ISOBEL Sound Field Dataset
**Authors**: Miklas Strøm Kristoffersen, Martin Bo Møller, Pablo MartÃnez-Nuevo, Jan Østergaard
## Deep Sound Field Reconstruction in Real Rooms: Introducing the ISOBEL Sound Field Dataset
Miklas Strøm Kristoffersen, 1, 2 Martin Bo Møller, 1 Pablo Mart´ ınez-Nuevo, 1 and Jan Østergaard 2
1 Research Department, Bang & Olufsen a/s, Struer, Denmark
2 AI and Sound Section, Department of Electronic Systems, Aalborg University, Aalborg, Denmark
Knowledge of loudspeaker responses are useful in a number of applications, where a sound system is located inside a room that alters the listening experience depending on position within the room. Acquisition of sound fields for sound sources located in reverberant rooms can be achieved through labor intensive measurements of impulse response functions covering the room, or alternatively by means of reconstruction methods which can potentially require significantly fewer measurements. This paper extends evaluations of sound field reconstruction at low frequencies by introducing a dataset with measurements from four real rooms. The ISOBEL Sound Field dataset is publicly available, and aims to bridge the gap between synthetic and real-world sound fields in rectangular rooms. Moreover, the paper advances on a recent deep learning-based method for sound field reconstruction using a very low number of microphones, and proposes an approach for modeling both magnitude and phase response in a U-Net-like neural network architecture. The complex-valued sound field reconstruction demonstrates that the estimated room transfer functions are of high enough accuracy to allow for personalized sound zones with contrast ratios comparable to ideal room transfer functions using 15 microphones below 150 Hz.
The following article has been submitted to the Journal of the Acoustical Society of America. After it is published, it will be found at http://asa.scitation.org/journal/jas.
## I. INTRODUCTION
The response of a sound system in a room primarily varies with the room itself, the position of the loudspeakers, and the listening position. In order to deliver the intended sound system behavior to listeners, it is necessary to know about and compensate for this effect. Applications include among others room equalization (Cecchi et al. , 2018; Karjalainen et al. , 2001; Radlovic et al. , 2000), virtual reality sound field navigation (Tylka and Choueiri, 2015), source localization (Nowakowski et al. , 2017), and spatial sound field reproduction over predefined or dynamic regions of space also referred to as sound zones (Betlehem et al. , 2015; Møller and Østergaard, 2020). An approach to achieve this, is to measure the loudspeaker response at the desired listening locations and adjust the sound system accordingly. However, the task of measuring impulse responses on a sufficiently fine-grained grid in an entire room, quickly poses as a time-consuming and extensive manual labor that is not desirable. Instead, methods have been developed for the purpose of estimating impulse responses in a room based on a limited number of actual measurements. These methods are also referred to as sound field reconstruction and virtual microphones. The task of reconstructing room impulse responses in positions that have not been measured directly, is an active research field which has been explored in several studies (Ajdler et al. , 2006; Antonello et al. , 2017; Fernandez-Grande, 2019; Mignot et al. , 2014; Verburg and Fernandez-Grande, 2018; Vu and Lissek, 2020).
Machine learning, and in particular deep learning, is currently receiving widespread attention across scien- tific domains, and as an example within room acoustics, it has been used to estimate acoustical parameters of rooms (Genovese et al. , 2019; Yu and Kleijn, 2021). In recent work, deep learning-based methods were introduced to sound field reconstruction in reverberant rectangular rooms (Llu´ ıs et al. , 2020). This data-driven approach is able to learn sound field magnitude characteristics from large scale volumes of simulated data without prior information of room characteristics, such as room dimensions and reverberation time. The method is computationally efficient, and works with irregularly and arbitrarily distributed microphones for which there is no requirement of knowing absolute locations in the Euclidean space, in contrast to previous solutions. Furthermore, the reconstruction proves to work with a very low number of microphones, making real-world implementation feasible. To assess the issue of real-world sound field reconstruction, the method is evaluated using measurements in a single room (Llu´ ıs et al. , 2020). However, it is still unknown how much knowledge is transferred from the simulated to the real environment, as well as how well the model generalizes to different real rooms. This is a general problem in deep learning applications that rely on labor intensive data collections, which is our motivation for publishing an open access dataset of real-world sound fields in a diverse set of rooms.
This paper studies sound field reconstruction at low frequencies in rectangular rooms with a low number of microphones. The main contributions are:
- This paper introduces a sound field dataset, which is publicly available for development and evaluation of sound field reconstruction methods in four real rooms. It is our hope that the ISOBEL Sound Field
dataset will help the community in benchmarking and comparing state-of-the-art results.
- We assess the real-world performance of deep learning-based sound field magnitude reconstruction trained on simulated sound fields. For this purpose, we consider low frequencies, since lowfrequency room modes can significantly alter listening experience.Furthermore, we are interested in using a very low number of microphones.
- Moreover, we extend the deep learning-based sound field reconstruction to cover complex-valued inputs, i.e. both the magnitude and the phase of a sound field. Evaluation is performed in both simulated and real rooms, where a performance gap is observed. We argue why complex sound field reconstruction may have more difficulties in transferring useful knowledge from synthetic to real data.
- Lastly, we demonstrate the application of complexvalued sound field reconstruction within the field of sound zone control. Specifically, it is shown that sound fields reconstructed from as little as five microphones pose as valuable inputs to acoustic contrast control.
The paper is organized as follows: Section II introduces the concept of sound field reconstruction. Details of measurements from real rooms are presented in Section III. In Section IV, we focus on the problem of reconstructing the magnitude of sound fields, while Section V extends the model to complex-valued sound fields. Finally, Section VI investigates the application of sound zones through sound field reconstruction.
## II. SOUND FIELD RECONSTRUCTION
Our approach towards the sound field reconstruction problem is based on the observation that acoustic pressure in a room can be described using a three-dimensional regular grid of points defining a three-dimensional discrete function. The approach specifically for the purpose of magnitude reconstruction was introduced in (Llu´ ıs et al. , 2020). First, let R = [0 , l x ] × [0 , l y ] × [0 , l z ] denote a rectangular room, where l x , l y , l z > 0 are the length, width, and height of the room, respectively. Given such room, we define the grid as a discrete set of coordinates D o . However, for the sake of simplicity, we reduce the three-dimensional problem to a two-dimensional reconstruction on horizontal planes. The two-dimensional grid with a constant height z o is defined as
<!-- formula-not-decoded -->
for z o ∈ [0 , l z ], i = 0 , . . . , I -1, j = 0 , . . . , J -1, and integers I, J ≥ 2. Note, though, that the dataset collected for this study, which we will introduce in Section III, does in fact contain multiple horizontal planes at different heights. We keep the investigations of three-dimensional reconstruction for future work, and frame the core challenge of this paper as estimation of sound pressure in two-dimensional horizontal planes.
The function that we seek to reconstruct on this grid is the Fourier transform of the sound field in a frequency band that covers the low frequencies. The complexvalued frequency-domain sound field calculated using the Fourier transform is given by
<!-- formula-not-decoded -->
where ω ∈ R is a given excitation frequency, and p ( r , t ) denotes the spatio-temporal sound field with r ∈ R . We refer to the real and imaginary parts of the sound field using s Re ( r , ω ) and s Im ( r , ω ), respectively. Note that s is defined as the magnitude of the Fourier transform in (Llu´ ıs et al. , 2020). Instead, for magnitude reconstruction, we introduce the magnitude of the sound field
<!-- formula-not-decoded -->
for ω ∈ R and r ∈ R .
The procedure for reconstructing s ( r , ω ) on D o takes its starting point from actual observations of the sound field in select positions of the grid. We refer to the collected set of these available sample points as S o , which we further define to be a subset of the full grid. That is, S o ⊆ D o . The cardinality |S o | of the set S o is the number of available sample points, which we will also refer to as the number of microphones n mic in later experiments. We define the samples available to the reconstruction algorithm as
<!-- formula-not-decoded -->
An important aspect of these definitions is that the grid is unitless and positions can be defined in relative terms. That is, when sampling a point in the grid, only the relative position within the grid, and hence the room, needs to be known. This allows us to relax the data collection compared to alternative methods that require absolute locations. Another important element to consider is that the sampling pattern of S o can form any arrangement within D o as long as 1 ≤ |S o | ≤ |D o | . As an example, this means that sampled points can be irregularly distributed spatially in a room.
Situations may arise where the sound field resolution, as defined by l x , I , l y , and J , is too coarse. As an example, consider rooms that are either very long, wide, or in general large. Another example includes applications where fine-grained variations within a sound field are of importance. To compensate for this effect, we allow the reconstruction to base its output on another grid than D o . Such domain will typically be an upsampling of the original grid, but similarly it can be defined with other transformations, e.g. downsampling. Specifically, we define the grid as
<!-- formula-not-decoded -->
where i = 0 , . . . , IL -1, j = 0 , . . . , J P -1, and L, P must be chosen such that IL, JP ∈ Z + . Note that a value larger than one for either L or P results in an upsampling in the respective dimension.
The task of the sound field reconstruction is then to estimate the sound field on the grid D L,P o based on the sampled points S o . In particular, the objective of the reconstruction algorithm is to learn parameters w given
<!-- formula-not-decoded -->
where g w is an estimator and Ω = { ω k } K k =1 is the set of frequencies at which the sound field will be reconstructed. The remainder of the paper describes the procedure for learning parameters w using deep learning-based methods.
## A. Evaluation Metrics
The successfulness of the estimator is quantitatively judged using normalized mean square error (NMSE) at each frequency point in { ω k } K k =1
<!-- formula-not-decoded -->
The NMSE provides an average error over all positions in the grid between reconstructed and original sound fields for a single room at a single frequency. We also introduce an average NMSE, which is the NMSE performance averaged over all frequencies of interest as well as over all realizations from M trials, e.g. multiple rooms
MNMSE =
<!-- formula-not-decoded -->
This measure serves as an overall indication of the accuracy of a model, whereas the NMSE k allows a deeper insight of model behaviors at different frequencies. Note that the M trials are specific to each experiment and will be described accordingly.
## III. THE ISOBEL SOUND FIELD DATASET
A major contribution of this paper is the ISOBEL Sound Field dataset, which is released as open access alongside the manuscript. 1 The intended purpose is to use the measurements from real rooms for evaluation of sound field reconstruction in a diverse set of rooms. Note that the room-wide measurements of room impulse responses have several other use-cases that will not be further investigated in this paper, but we encourage the use outside sound field reconstruction as well. This section details the dataset and the measurement procedure.
The dataset consists of measurements from four different rooms as specified in Table I and depicted in Fig. 1. The data collection is an extension to the real room measured in (Llu´ ıs et al. , 2020), which is included in the ISOBEL Sound Field dataset as Room B for simple access to all measured rooms. The rooms are located at Aalborg University, Aalborg, Denmark, and Bang & Olufsen a/s, Struer, Denmark. The rooms have significantly different acoustic properties and also vary in size. Two types of measurements are conducted in each room: 1) Reverberation time; 2) Sound field. However, only the sound field measurements are released as part of the dataset.
The reverberation times are measured in conformity with ISO 3382-2 (ISO 3382-2:2008, 2008) and calculated based on resulting impulse responses using backwards integration and least-squares best fit evaluation of the decay curves. 2 The reverberation times reported in the table are the arithmetic averages of 1/3 octave T 20 estimates in the frequency range 50-316 Hz.
The sound field measurements are performed on a 32 by 32 grid with sample points distributed uniformly along the length and width of each room. That is, a total of 1024 positions are measured in each room if possible, but in some cases it is not feasible to measure all positions due to e.g. obstacles. 3 The horizontal grids are measured at four different heights: 1, 1.3, 1.6, and 1.9 meters above the floor. 4 This is achieved using the microphone rig depicted in Fig. 1. Two 10 inch loudspeakers are used to acquire sound fields from two different source positions in each room. Both loudspeakers are placed on the floor, one in a corner and one in an arbitrary position. The sound sources are kept in the same position, while the microphones are moved around the room to record impulse responses. For each microphone position in the grid, the two sources play logarithmic sine sweeps in the frequency range 0.1-24,000 Hz followed by a quiet tail, (Farina, 2000). We use a sampling frequency of 48,000 Hz. The equipment includes among others four G.R.A.S. 40AZ prepolarized free-field microphones connected to four G.R.A.S. 26CC CCP standard preamplifiers and an RME Fireface UFX+ sound card. The four microphones are level calibrated at 1,000 Hz using a Br¨ uel & Kjær sound calibrator type 4231 prior to the measurements.
TABLE I. Room characteristics in the ISOBEL Sound Field dataset. The reverberation times are the arithmetic averages of 1/3 octave T 20 estimates in the frequency range 50-316Hz.
| Room | Dim. [m] | Size [m 2 /m 3 ] | T 20 [s] |
|------------|---------------------|--------------------|------------|
| Room B | 4.16 x 6.46 x 2.30 | 27/ 62 | 0.39 |
| VR Lab | 6.98 x 8.12 x 3.03 | 57/172 | 0.37 |
| List. Room | 4.14 x 7.80 x 2.78 | 32/ 90 | 0.8 |
| Prod. Room | 9.13 x 12.03 x 2.60 | 110/286 | 0.77 |
FIG. 1. Left: Rig with four microphones. Rooms from top left to bottom right: Room B, VR Lab, Listening Room, and Product Room.
<details>
<summary>Image 1 Details</summary>

### Visual Description
## Collage of Experimental Room Setups with Spatial Markers
### Overview
The image is a collage of five photographs depicting different experimental room configurations, each annotated with spatial axes (x, y, z) and measurement markers. The leftmost image shows a vertical calibration pole with height markers, while the remaining four images display room layouts with labeled axes and equipment placements.
### Components/Axes
1. **Left Image (Calibration Pole)**:
- Vertical pole with red height markers at **1.0m**, **1.3m**, **1.6m**, and **1.9m**.
- Positioned on a blue carpeted floor adjacent to a gray wall.
2. **Top-Center Image (Room A)**:
- Axes labeled **x** (horizontal), **y** (vertical), and **z** (depth).
- Features: Wooden flooring, black acoustic curtains, mounted flat-screen TV, and a tripod-mounted camera.
3. **Top-Right Image (Room B)**:
- Axes labeled **x** (horizontal), **y** (vertical), and **z** (depth).
- Features: Light-colored wooden flooring, white walls, large windows with sheer curtains, and a projector on a gray mat.
4. **Bottom-Left Image (Room C)**:
- Axes labeled **x** (horizontal), **y** (vertical), and **z** (depth).
- Features: White walls with acoustic foam panels, recessed ceiling lights, and a tripod-mounted camera.
5. **Bottom-Right Image (Room D)**:
- Axes labeled **x** (horizontal), **y** (vertical), and **z** (depth).
- Features: Light wooden flooring, white walls with recessed lighting, and two windows with partially closed blinds.
### Detailed Analysis
- **Calibration Pole**: The pole in the left image serves as a reference for vertical measurements, with markers spaced at **0.3m intervals** (1.0m to 1.9m). The pole’s base is anchored to a tripod, suggesting use in height calibration for equipment.
- **Room A**: The TV is mounted at the far end of the room, aligned with the **z-axis** (depth). The tripod is positioned near the **y-axis** (vertical), likely for camera stabilization.
- **Room B**: The projector is centered on the gray mat, with the **x-axis** extending toward the window. The **z-axis** (depth) aligns with the wall-mounted projector.
- **Room C**: The tripod is placed near the **y-axis** (vertical), with acoustic foam panels on the walls to reduce sound reflection. The **z-axis** (depth) extends toward the far wall.
- **Room D**: The **x-axis** spans the room’s width, while the **z-axis** (depth) aligns with the windows. The recessed lighting suggests controlled illumination for experiments.
### Key Observations
1. **Consistent Axis Labeling**: All rooms use the same coordinate system (x, y, z), indicating standardized spatial referencing across experiments.
2. **Equipment Placement**: Tripods and cameras are consistently positioned near the **y-axis** (vertical), suggesting focus on vertical measurements or stabilization.
3. **Acoustic Treatments**: Rooms A and C include sound-dampening materials (curtains, foam panels), implying audio-sensitive experiments.
4. **Lighting Variations**: Rooms B and D have natural light sources (windows), while Room C uses artificial lighting, indicating controlled lighting conditions.
### Interpretation
The collage illustrates a controlled experimental environment designed for spatial and sensory studies. The consistent use of axes and calibration markers suggests a focus on **3D spatial mapping** or **human movement tracking**. The presence of acoustic treatments and lighting controls implies experiments requiring **minimized environmental interference**, such as robotics testing, VR calibration, or behavioral studies. The calibration pole’s precise height markers further support applications requiring **vertical alignment accuracy**, such as drone testing or optical sensor calibration.
</details>
## IV. SOUND FIELD MAGNITUDE RECONSTRUCTION
In the previous sections we have introduced the problem of reconstructing sound fields on two-dimensional grids in rectangular rooms, as well as introduced a realworld dataset specifically for evaluation of estimators solving such problem. In recent work, (Llu´ ıs et al. , 2020) showed that the problem fits within the context of deep learning-based methods for image reconstruction. Specifically, the tasks of inpainting, (Bertalmio et al. , 2000; Liu et al. , 2018), and super-resolution, (Dong et al. , 2016; Ledig et al. , 2017), which can be paralleled to the tasks of filling in the grid points that are not measured in the sound fields D L,P o \S o , as well as upsampling the grid resolution to achieve fine-grained variations in sound fields. One realization is that these methods are designed to work with real-valued images. To accommodate this, (Llu´ ıs et al. , 2020) propose to reconstruct only the magnitude of the sound field, i.e. | s ( r , ω ) | , using a U-Net-like architecture, (Ronneberger et al. , 2015).
To this end, the sampled grids are defined as tensors together with masks specifying which positions are measured (Llu´ ıs et al. , 2020). As an example, {| s ( r , ω k ) |} r ∈D L,P o ,k can be constructed as a tensor of the form S mag ∈ R IL × JP × K . The network is trained using a large number of simulated realizations of rooms, as will be described in the following section. For the experiments, we are interested in assessing the ability of the model to generalize to a wide range of real rooms.
## A. Simulation of Sound Fields for Training Data
Green's function can be used to approximate sound fields in rectangular rooms that are lightly damped, (Ja- cobsen and Juhl, 2013). The function provides a solution as an infinite summation of room modes in the three dimensions of a room, x , y , and z . It is defined as follows
<!-- formula-not-decoded -->
where ∑ N = ∑ ∞ n x =0 ∑ ∞ n y =0 ∑ ∞ n z =0 , for compactness, denotes summation across modal orders in the three dimensions of the room, and similarly the triplet of integers ( n x , n y , n z ) are represented by N . Furthermore, V denotes the volume of the room, ω 2 N represents angular resonance frequency of a mode associated with a specific N , the shape of the mode is denoted ψ N ( · ), τ N is the time constant of the mode, and c is the speed of sound. Assuming rigid boundaries, the shape is determined using the expression (Jacobsen and Juhl, 2013)
<!-- formula-not-decoded -->
Here, Λ N = √ x y z are constants used for normalization with 0 = 1, 1 = 2 = · · · = 2. Using Sabine's equation, the absorption coefficient is calculated and used to determine time constants of each mode.This is done by assuming that surfaces of a room have uniform distribution of absorption.
In the following experiments, two sets of training data are used. The first dataset is introduced in (Llu´ ıs et al. , 2020) and consists of 5,000 rectangular rooms. The room dimensions are sampled randomly in accordance with the recommendations for listening rooms in ITU-R BS.1116-3 (ITU-R BS.1116-3, 2015). The dataset uses a
FIG. 2. NMSE in dB of U-Net-based magnitude reconstruction in the four measured rooms with n mic = 15 using the original pretrained model presented in (Llu´ ıs et al. , 2020).
<details>
<summary>Image 2 Details</summary>

### Visual Description
## Line Graph: NMSE (dB) vs Frequency Across Different Rooms
### Overview
The image is a line graph comparing the Normalized Mean Squared Error (NMSE) in decibels (dB) across four different environments (Room B, VR Lab, List. Room, Prod. Room) as a function of frequency (Hz). The graph spans frequencies from 30 Hz to 300 Hz, with NMSE values ranging from -25 dB to 0 dB. Four distinct lines represent the data for each room, with color-coded legends for clarity.
---
### Components/Axes
- **X-axis (Frequency)**: Labeled "Frequency" with ticks at 30, 40, 50, ..., 300 Hz. The scale is linear.
- **Y-axis (NMSE)**: Labeled "NMSE (dB)" with ticks at -25, -20, -15, ..., 0 dB. The scale is linear.
- **Legend**: Located in the bottom-right corner, mapping colors to rooms:
- **Blue**: Room B
- **Green**: VR Lab
- **Black**: List. Room
- **Red**: Prod. Room
---
### Detailed Analysis
1. **Room B (Blue Line)**:
- **Trend**: Starts at ~-15 dB at 30 Hz, dips sharply to ~-18 dB at 40 Hz, then fluctuates with minor peaks and troughs. Ends near -5 dB at 300 Hz.
- **Key Points**:
- Sharp dip at 40 Hz (~-18 dB).
- Slight recovery to ~-10 dB at 50 Hz.
- Gradual increase to ~-5 dB at 300 Hz.
2. **VR Lab (Green Line)**:
- **Trend**: Relatively stable with minor fluctuations. Peaks at ~-3 dB at 100 Hz and ~-2 dB at 200 Hz. Ends near -4 dB at 300 Hz.
- **Key Points**:
- Smooth curve with no extreme dips.
- Slight upward trend toward higher frequencies.
3. **List. Room (Black Line)**:
- **Trend**: Starts at ~-10 dB at 30 Hz, drops sharply to ~-18 dB at 40 Hz, then fluctuates with peaks and troughs. Ends near -5 dB at 300 Hz.
- **Key Points**:
- Sharp dip at 40 Hz (~-18 dB).
- Recovery to ~-10 dB at 50 Hz.
- Oscillatory behavior with peaks at ~-5 dB (e.g., 70 Hz, 150 Hz).
4. **Prod. Room (Red Line)**:
- **Trend**: Most stable line with minimal fluctuations. Peaks at ~-2 dB at 100 Hz and ~-3 dB at 200 Hz. Ends near -4 dB at 300 Hz.
- **Key Points**:
- Consistent performance with no extreme dips.
- Slight upward trend toward higher frequencies.
---
### Key Observations
- **Sharp Dips**: Room B and List. Room exhibit significant drops at 40 Hz (~-18 dB), suggesting potential issues or anomalies at this frequency.
- **Stability**: Prod. Room shows the most consistent performance, with the least variation across frequencies.
- **Peaks**: VR Lab and Prod. Room have minor peaks at 100 Hz and 200 Hz, indicating possible resonance or measurement artifacts.
- **Convergence**: All lines converge near -5 dB at 300 Hz, suggesting similar performance at high frequencies.
---
### Interpretation
The data highlights differences in NMSE performance across environments. The **Prod. Room** demonstrates the most stable and reliable performance, with minimal fluctuations and no extreme dips. In contrast, **Room B** and **List. Room** show significant variability, particularly at 40 Hz, which may indicate environmental or measurement challenges. The **VR Lab** exhibits moderate stability but with slight upward trends at higher frequencies. These trends could reflect differences in room acoustics, equipment calibration, or experimental conditions. The sharp dips at 40 Hz in Room B and List. Room warrant further investigation to identify root causes, such as structural resonances or sensor limitations. Overall, the graph underscores the importance of environmental factors in NMSE measurements and highlights the Prod. Room as the optimal setting for consistent performance.
</details>
FIG. 3. NMSE in dB of U-Net-based magnitude reconstruction in the four measured rooms with n mic = 15 using the model presented in (Llu´ ıs et al. , 2020) trained using the extended dataset.
<details>
<summary>Image 3 Details</summary>

### Visual Description
## Line Chart: NMSE (dB) vs Frequency Across Different Rooms
### Overview
The chart displays the Normalized Mean Squared Error (NMSE) in decibels (dB) across a frequency range of 30–300 Hz for four distinct environments: Room B, VR Lab, List. Room, and Prod. Room. All lines exhibit fluctuating trends, with convergence toward -5 dB at higher frequencies. The VR Lab line shows the most pronounced variability, including a sharp dip below -15 dB near 40 Hz.
### Components/Axes
- **X-axis (Frequency)**: Labeled "Frequency" with markers at 30, 40, 50, ..., 300 Hz.
- **Y-axis (NMSE)**: Labeled "NMSE (dB)" with markers from 0 to -25 dB.
- **Legend**: Located in the bottom-right corner, associating colors with rooms:
- Blue: Room B
- Green: VR Lab
- Black: List. Room
- Red: Prod. Room
### Detailed Analysis
1. **Room B (Blue)**:
- Starts at ~-15 dB at 30 Hz, rises sharply to ~-5 dB by 50 Hz, then stabilizes with minor fluctuations.
- Ends near -5 dB at 300 Hz.
2. **VR Lab (Green)**:
- Begins at ~-10 dB, dips to ~-18 dB at 40 Hz, then oscillates upward to ~-5 dB by 100 Hz.
- Shows the highest variability, with peaks reaching ~-5 dB and troughs near -18 dB.
3. **List. Room (Black)**:
- Starts at ~-10 dB, fluctuates between -10 dB and -15 dB until 70 Hz, then stabilizes near -5 dB.
- Ends at ~-5 dB with minor oscillations.
4. **Prod. Room (Red)**:
- Begins at ~-10 dB, rises to ~-5 dB by 60 Hz, then fluctuates between -5 dB and -10 dB.
- Ends near -5 dB with consistent oscillations.
### Key Observations
- **Convergence at High Frequencies**: All lines approach -5 dB by 200–300 Hz, suggesting similar performance in higher-frequency ranges.
- **VR Lab Anomaly**: The sharp dip below -15 dB at 40 Hz is unique to VR Lab, indicating a potential outlier or environmental factor.
- **Initial Variability**: Room B and VR Lab start with the lowest NMSE values, while List. Room and Prod. Room begin higher but converge later.
### Interpretation
The data suggests that NMSE values (lower = better accuracy) are frequency-dependent and environment-specific. The convergence at higher frequencies implies that measurement accuracy stabilizes across rooms in this range. The VR Lab’s anomalous dip at 40 Hz may reflect unique acoustic properties or calibration issues in that environment. Room B’s rapid improvement from 30–50 Hz contrasts with its stability at higher frequencies, possibly indicating targeted adjustments in that range. The consistency of List. Room and Prod. Room at higher frequencies suggests they may share similar acoustic characteristics or calibration standards.
</details>
constant reverberation time T 60 of 0.6 s and only includes room modes in the x and y dimensions, i.e. n z = 0.
The second dataset consists of 20,000 rectangular rooms. Room dimensions are uniformly sampled with V ∼ U (50 , 300)m 3 , l x ∼ U (3 . 5 , 10)m, l z ∼ U (1 . 5 , 3 . 5)m, and l y = V/l x l z . Compared to the first dataset, the room dimensions span a larger range and allow us to represent e.g. the Product Room, which is not included in the original training data. The dataset uses reverberation times T 60 sampled from U (0 . 2 , 1 . 0)s and includes room modes in all three x -, y -, and z -dimensions.
For both datasets, a grid D L,P o is defined with I = J = 8 and L = P = 4, which effectively divides a sound field into 32x32 uniformly-spaced microphone positions. Using this grid, the magnitude of the sound field is reconstructed at 1/12 octave center-frequencies resolution in the range [30, 300] Hz. Simulations are specified to include all room modes with a resonance frequency below 400 Hz, which means that there is a total of K = 40 frequency slices.
## B. Experiments on the ISOBEL Sound Field Dataset
The U-Net-like architecture has shown promising results on simulated data and on measurements from a single real room (Llu´ ıs et al. , 2020). In the following experiments, we expose the model to the ISOBEL Sound Field dataset. We include results from the original model, as well as a model built around a similar architecture but using the extended training data with a larger range of room dimensions and reverberation characteristics. We investigate the performance of the model trained with the two different simulated datasets in the four rooms included in the real-world dataset. Special attention is paid to the number of available samples, i.e. the number of microphones n mic . We are mainly interested in settings with a very low number of microphones. In particular, we show results for 5, 15, and 25 microphones in the rooms with a total of 32 × 32 = 1024 available positions. In each room, a total of 40 different and randomly sampled realizations of microphone positions S o are used for each value of n mic . We report the average performance across the 40 realizations, and use the source located in one of the corners of each room.
Fig. 2 and Fig. 3 show NMSE k results for 15 microphones of model trained with the original and the extended datasets, respectively. It is clear that the model trained with the original dataset does not generalize well to all the rooms. This behavior is expected, since the training data are not designed to represent rooms that fall outside the recommendations for listening room dimensions. On the contrary, the extended training data are motivated in encompassing a wider selection of rooms, which also shows in the results for e.g. the Product Room. One important observation in this regard is that performance does not decrease in rooms that are already represented in the simulated data when more diverse simulated rooms are included, which can e.g. be seen from the performance in Room B. This result indicates that the capacity of the model is sufficient for generalizing to a wide range of diverse rooms and room
TABLE II. MNMSE in dB with M = 40 different and randomly sampled realizations of S o for each room in the ISOBEL SF dataset. A lower score is better.
| | | n mic | n mic | n mic |
|------------|-------|---------|---------|---------|
| Room | Model | 5 | 15 | 25 |
| Room B | Orig. | -6.33 | -8.71 | -9.62 |
| | Ext. | -6.27 | -8.84 | -10.25 |
| VR Lab | Orig. | -4.01 | -5.08 | -5.63 |
| | Ext. | -4.12 | -6.78 | -8.05 |
| List. Room | Orig. | -4.38 | -6.92 | -7.94 |
| | Ext. | -5 | -7.61 | -8.44 |
| Prod. Room | Orig. | -3.89 | -4.91 | -5.55 |
| | Ext. | -5.18 | -6.67 | -7.73 |
FIG. 4. Architecture of the U-Net-like convolutional neural network proposed for complex sound field reconstruction. S is the tensor with real and imaginary sound fields concatenated along the frequency-dimension, M is the mask tensor, and ˆ S is the reconstructed sound field tensor.
<details>
<summary>Image 4 Details</summary>

### Visual Description
## Diagram: U-Net Architecture for Image Segmentation
### Overview
The diagram illustrates a U-Net neural network architecture, commonly used for biomedical image segmentation. It shows the flow of data through encoder-decoder blocks with skip connections, highlighting dimensional changes and operations at each layer.
### Components/Axes
- **Input Block (SM)**: 80x80x32 (spatial dimensions x channels)
- **Encoder Path**:
- 128x128x16 → 256x256x8 → 512x512x4 → 1024x1024x2
- Operations: PConv 5x5 (purple), PConv 3x3 (green)
- **Decoder Path**:
- 1536x1536x2 → 512x512x4 → 768x768x8 → 1024x1024x16 → 128x128x32 → 80x80x32
- Operations: Upsample 2x2 (red), Skip/concat (gray)
- **Output Block (Ŝ)**: 80x80x32 (same as input)
- **Legend**:
- Purple: PConv 5x5
- Green: PConv 3x3
- Red: Upsample 2x2
- Gray: Skip/concat
### Detailed Analysis
- **Encoder**:
- Spatial resolution increases (80→128→256→512→1024) while channel depth decreases (32→16→8→4→2)
- Uses PConv 5x5 and PConv 3x3 for feature extraction
- **Decoder**:
- Spatial resolution decreases (1536→512→768→1024→128→80) while channel depth increases (2→4→8→16→32)
- Upsampling via 2x2 convolution (red arrows)
- **Skip Connections**:
- Gray arrows connect encoder and decoder layers at matching spatial resolutions (e.g., 128x128x16 ↔ 128x128x16)
- Preserve spatial information lost during downsampling
### Key Observations
1. **Symmetry**: Encoder and decoder paths mirror each other in spatial resolution progression
2. **Channel Depth**: Encoder reduces channels for feature abstraction; decoder increases channels for reconstruction
3. **Skip Connections**: Critical for maintaining positional accuracy in segmentation outputs
4. **Dimensional Consistency**: Input/output dimensions match (80x80x32), ensuring spatial alignment
### Interpretation
This architecture demonstrates a classic U-Net design optimized for medical image segmentation:
- **Encoder-Decoder Balance**: Progressive downsampling captures global context, while upsampling with skip connections preserves local details
- **Pixel Convolutions**: Efficient spatial operations tailored for image data
- **Skip Connections**: Enable direct feature reuse between encoder/decoder, critical for accurate segmentation
- **Channel Progression**: Balances feature complexity (encoder) with reconstruction capacity (decoder)
The architecture's design prioritizes spatial fidelity through skip connections while maintaining computational efficiency via pixel-wise convolutions. The symmetric structure ensures the network can learn both coarse and fine segmentation details.
</details>
acoustic characteristics, given that the model is provided with ample training samples.
Table II details MNMSE results, which are the NMSE results averaged across frequencies K = 40 and S o realizations M = 40. The MNMSE results for n mic = 15 are the condensed results shown for the NMSE k in Figs. 2 and 3. The scores in the table reiterate the observations from the figures, performance is improved with the extended training data for some rooms in particular, while performance is maintained in the other rooms. Interestingly, there seems to be a tendency of more pronounced improvements with a larger number of microphones. We attribute this effect to similar observations within classical methods that as the number of microphones increase, relative improvement for reconstruction is higher at low frequencies as opposed to the highfrequency range, (Ajdler et al. , 2006; Llu´ ıs et al. , 2020).
In summary, the deep learning-based model is confirmed to possess the ability to generalize to a diverse set of real rooms for sound field magnitude reconstruction. Based solely on training with simulated data, these promising results motivate further investigations, e.g. of reconstructing the complex-valued sound fields.
## V. COMPLEX SOUND FIELD RECONSTRUCTION
We propose to extend the U-Net-based model to work with complex-valued room transfer functions (RTFs). Reconstruction of both magnitude and phase of sound fields enable new opportunities, such as the application of sound zones. A topic, which we investigate in Section VI.
The proposed model is based on the model designed to work with the magnitude of sound fields. Note that deep learning-based models that work directly on complex-valued inputs have been introduced, e.g. within Transformers (Kim et al. , 2020; Yang et al. , 2020), but in this paper we instead choose to process the sound fields such that the U-Net-based model receives real-valued inputs. Specifically, we present the model to real and imaginary parts of sound fields separately. That is, where the magnitude-based model receive as input {| s ( r , ω k ) |} r ∈D L,P o ,k in the tensor form S mag ∈ R IL × JP × K , the complex-based model instead receives a concatenation of the real and imaginary sound fields. Specifically, using the real sound field { s Re ( r , ω k ) } r ∈D L,P o ,k with the tensor form S Re ∈ R IL × JP × K , and similarly the imaginary sound field tensor S Im ∈ R IL × JP × K , we define the concatenated input:
<!-- formula-not-decoded -->
where S ∈ R IL × JP × 2 K is the resulting tensor with real and imaginary sound fields concatenated along the frequency-dimension. Note that the complex-valued sound field is easily recovered from this tensor form. In addition, we define a mask tensor M ∈ R IL × JP × 2 K computed from S o and D L,P o .
We follow the pre- and postprocessing steps as described in (Llu´ ıs et al. , 2020), which entails completion, scaling, upsampling, mask generation, and rescaling based on linear regression. These steps are, however, adjusted such that they operate on a tensor that has doubled in size from K to 2 K in the third dimension. Furthermore, we have observed significant improvements by changing the min-max scaling of the input to a max scaling that takes into account both real and imaginary parts for each frequency slice. Specifically:
<!-- formula-not-decoded -->
<!-- formula-not-decoded -->
for each ω k . Note that this alters the scaling operation from working in the range [0,1] to working in [-1,1]. The
motivation in doing so, is that values can be negative, in contrast to the real values from the magnitude. By using max scaling we ensure that zero will not shift between realizations.
The architecture of the proposed neural network, as illustrated in Fig. 4, is based on a U-Net (Ronneberger et al. , 2015). We employ partial convolutions (PConv) as proposed for image inpainting in (Liu et al. , 2018). In the encoding part of the U-Net, we use a stride of two in the partial convolutions in order to halve the feature maps, while doubling the number of kernels in each layer. The decoder part acts opposite with upsampling feature maps and reducing the number of kernels to reach an output tensor ˆ S with matching dimensions to the input tensor S . We use ReLU as activation function in the encoding part, and leaky ReLU with a slope coefficient of -0.2 in the decoder. We initialize the weights using the uniform Xavier method (Glorot and Bengio, 2010), initialize the biases as zero, and use the Adam optimizer (Kingma and Ba, 2014) with early stopping when performance on a validation set stops increasing. Due to the increased input and output sizes, we double the number of kernels in all layers compared to the U-Net for magnitude reconstruction. We also do not use a 1x1 convolution with sigmoid activation in the last layer, since the range of our output is not constrained to [0,1] but instead [-1,1]. We have not experienced any decreases in performance from not including this layer.
## A. Experiments
In this section, we assess the complex-valued sound field reconstruction. The simulated extended dataset introduced in Section IV A is used to train the model. It is important to note that NMSE scores are not directly comparable between magnitude and complex reconstruction, for which reason it is not possible to scrutinize differences between the two types of models. That is, the results presented in the following experiments will stand on their own, and only indicative parallels can be drawn to the results from magnitude reconstruction.
First, we test how the model performs on the simulated data associated with the training data, but held out specifically for evaluation. This test set consists of 190 simulated rooms, the validation set contains approximately 1,000 rooms, and the training set holds the remaining rooms from the 20,000 available rooms. In each room, three different realizations of S o are used for each value of n mic . Results in terms of NMSE are shown in Fig. 5. Some tendencies are similar to those observed for magnitude reconstruction, such as improvements in performance with an increasing number of available microphones. At the same time, as frequency increases, performance degrades.
Next, we evaluate the complex reconstruction model on the ISOBEL Sound Field dataset. The approach is similar to the experiment in Section IV B, except the use of the complex-valued sound fields instead of the magnitude. As can be seen from the results in Fig. 6, per-
FIG. 5. NMSE in dB for complex reconstruction of simulated sound fields in the test set with 190 different rooms and three realizations of S o in each room ( M = 570 for each value of n mic ). The solid lines indicate average NMSE k shown with 95% confidence intervals. Colors indicate different values of n mic in the range [5, 55].
<details>
<summary>Image 5 Details</summary>

### Visual Description
## Line Chart: NMSE vs Frequency
### Overview
The image is a line chart depicting the relationship between Normalized Mean Squared Error (NMSE) in decibels (dB) and frequency in Hertz (Hz). Six distinct data series are plotted, each corresponding to a specific value (5, 15, 25, 35, 45, 55), with lines colored uniquely for differentiation. All lines exhibit a consistent upward trend as frequency increases, indicating a positive correlation between NMSE and frequency.
### Components/Axes
- **Y-Axis (Left)**: Labeled "NMSE (dB)" with a scale ranging from 0 to -25 dB in 5 dB increments.
- **X-Axis (Bottom)**: Labeled "Frequency" with a scale from 30 Hz to 300 Hz in 10 Hz increments.
- **Legend**: Positioned on the right side of the chart, mapping colors to values:
- Blue: 5
- Orange: 15
- Green: 25
- Red: 35
- Purple: 45
- Brown: 55
- **Lines**: Six colored lines (blue, orange, green, red, purple, brown) represent the six data series. Each line starts at a lower NMSE value at 30 Hz and increases (becomes less negative) as frequency rises to 300 Hz.
### Detailed Analysis
- **Line Trends**:
- **Blue (5)**: Starts at ~-10 dB at 30 Hz, ends near 0 dB at 300 Hz. Slope is the least steep.
- **Orange (15)**: Starts at ~-15 dB at 30 Hz, ends near -5 dB at 300 Hz. Slope is steeper than blue.
- **Green (25)**: Starts at ~-18 dB at 30 Hz, ends near -8 dB at 300 Hz. Slope is steeper than orange.
- **Red (35)**: Starts at ~-20 dB at 30 Hz, ends near -10 dB at 300 Hz. Slope is steeper than green.
- **Purple (45)**: Starts at ~-22 dB at 30 Hz, ends near -12 dB at 300 Hz. Slope is steeper than red.
- **Brown (55)**: Starts at ~-24 dB at 30 Hz, ends near -14 dB at 300 Hz. Slope is the steepest.
- **Spacing**: Lines are evenly spaced vertically at 30 Hz, with gaps decreasing slightly as frequency increases. At 300 Hz, the lines converge closer but remain distinct.
### Key Observations
1. **Positive Correlation**: All lines show a consistent upward trend, confirming that NMSE increases with frequency.
2. **Value-Specific Behavior**: Higher values (e.g., 55) start with lower NMSE at 30 Hz but exhibit steeper slopes, closing the gap with lower values (e.g., 5) at higher frequencies.
3. **Parallelism**: Lines maintain parallel trajectories, suggesting a linear relationship between frequency and NMSE for each value.
4. **Convergence**: At 300 Hz, the lines are closer together, indicating diminishing differences in NMSE across values at high frequencies.
### Interpretation
The data suggests that NMSE is inversely related to frequency across all values, with higher values (e.g., 55) demonstrating greater sensitivity to frequency changes. The parallel lines imply a uniform scaling of NMSE with frequency, independent of the value. This could indicate a systematic relationship between measurement accuracy (NMSE) and frequency in a technical system, where higher frequencies inherently introduce more error. The convergence at 300 Hz may reflect a saturation effect or a limit to the system's frequency-dependent performance degradation. The legend's color coding ensures clear differentiation of values, critical for analyzing value-specific trends.
</details>
FIG. 6. Average NMSE k in dB of complex reconstruction in the four measured rooms with n mic = 15.
<details>
<summary>Image 6 Details</summary>

### Visual Description
## Line Chart: NMSE (dB) vs Frequency
### Overview
The image is a line chart comparing the Normalized Mean Square Error (NMSE) in decibels (dB) across a frequency range of 30 Hz to 300 Hz. Four data series are plotted, each representing a different environment or condition: Room B (blue), VR Lab (green), List. Room (black), and Prod. Room (red). The chart shows how NMSE varies with frequency for each series.
### Components/Axes
- **X-axis (Frequency)**: Labeled "Frequency" with markers at 30, 40, 50, 60, 70, 80, 90, 100, 200, and 300 Hz. The scale is linear.
- **Y-axis (NMSE (dB))**: Labeled "NMSE (dB)" with markers at -25, -20, -15, -10, -5, and 0 dB. The scale is linear.
- **Legend**: Located in the bottom-right corner, with four entries:
- **Blue**: Room B
- **Green**: VR Lab
- **Black**: List. Room
- **Red**: Prod. Room
### Detailed Analysis
1. **Room B (Blue Line)**:
- Starts at approximately **-15 dB** at 30 Hz.
- Rises sharply to **-5 dB** around 50 Hz.
- Stabilizes near **-5 dB** from 60 Hz to 300 Hz.
- **Key Trend**: Significant initial increase, then plateau.
2. **VR Lab (Green Line)**:
- Begins at **-5 dB** at 30 Hz.
- Dips to **-10 dB** around 40 Hz.
- Rises to **-5 dB** by 60 Hz.
- Fluctuates slightly but remains near **-5 dB** from 70 Hz to 300 Hz.
- **Key Trend**: Initial dip followed by stabilization.
3. **List. Room (Black Line)**:
- Starts at **-5 dB** at 30 Hz.
- Fluctuates between **-5 dB and 0 dB** across the frequency range.
- Peaks near **0 dB** around 100 Hz.
- **Key Trend**: Moderate variability with a peak at mid-frequencies.
4. **Prod. Room (Red Line)**:
- Begins at **-5 dB** at 30 Hz.
- Rises sharply to **0 dB** around 60 Hz.
- Stabilizes near **0 dB** from 70 Hz to 300 Hz.
- **Key Trend**: Sharp increase at mid-frequencies, then plateau.
### Key Observations
- **Room B** exhibits the most pronounced initial increase in NMSE, suggesting higher error at lower frequencies.
- **VR Lab** shows a notable dip at 40 Hz, indicating a potential anomaly or improved performance at that frequency.
- **List. Room** has the least variability, maintaining a relatively stable NMSE between -5 dB and 0 dB.
- **Prod. Room** demonstrates a sharp rise in NMSE at 60 Hz, which could indicate a critical issue or calibration problem at that frequency.
### Interpretation
The data suggests that **Room B** and **Prod. Room** experience significant NMSE increases at specific frequencies, potentially pointing to environmental or measurement inconsistencies. **VR Lab** and **List. Room** show more stable performance, with **VR Lab** having a unique dip at 40 Hz that may require further investigation. The sharp rise in **Prod. Room** at 60 Hz could indicate a critical failure or calibration error, warranting immediate attention. The chart highlights the importance of frequency-specific analysis in evaluating NMSE performance across different environments.
</details>
formances in the real rooms are not comparable to those from simulated data. Moreover, although it is not possible to compare directly, performance seems worse than what is achieved with the magnitude-based reconstruction in the same rooms, see Fig. 3. That is, the complex reconstruction model is not transferring useful knowledge as successfully from the simulations-based training to the real world. Given that the network is able to reconstruct the simulated sound fields, it appears that the complex simulation model is a worse match for the real rooms than the magnitude simulation model. The outcome is that the framework is able to reconstruct sound fields which are close to fields included in the training data, it is indicated that the complex simulations are a poor match for the real rooms. Two apparent differences are the identical boundary conditions at all surfaces and perfectly rectangular geometry assumed in the simulations, but which are not true in the real rooms. To provide insights into how the network behaves relative to rooms which does not match the training data set we now present the following simulations.
FIG. 7. NMSE in dB for complex reconstruction of simulated sound fields in rooms with no or small variations in the room dimensions. Rows: Training data. Columns: Test data. Four random realizations of S o are used in each of the 11 test rooms ( M = 44). The solid lines indicate average NMSE k shown with 95% confidence intervals. Colors indicate different n mic values, i.e., n mic = 5 (blue), n mic = 15 (orange), n mic = 25 (green), n mic = 35 (red), n mic = 45 (purple), and n mic = 55 (brown).
<details>
<summary>Image 7 Details</summary>

### Visual Description
## Line Graphs: Simulated Room Amplitude vs Frequency with Parameter Variations
### Overview
The image contains three sets of line graphs comparing amplitude (y-axis) against frequency (x-axis) across three simulated room configurations. Each set includes:
1. Baseline "Simulated List Room"
2. "Simulated List Room" with spatial uncertainty l_x + U(-0.25,0.25)m
3. "Simulated List Room" with expanded uncertainty l_x + U(-1.0,1.0)m
### Components/Axes
- **X-axis**: Frequency (0-300 Hz, linear scale)
- **Y-axis**: Amplitude (-25 to +25 dB, linear scale)
- **Legend**: Located in top-left corner, color-coded for 25 rooms (Room 1-25)
- **Graph Layout**: Three rows of three graphs each, with:
- Top row: Baseline simulations
- Middle row: ±0.25m uncertainty
- Bottom row: ±1.0m uncertainty
### Detailed Analysis
1. **Baseline Simulations (Top Row)**
- Room 1 (blue): Smooth sinusoidal pattern with minor fluctuations
- Room 2 (orange): Similar pattern with slightly increased variability
- Room 3 (purple): Shows initial peak at ~50Hz, then gradual decay
- Room 4-25: Gradual amplitude increase across frequency range
2. **±0.25m Uncertainty (Middle Row)**
- All rooms show increased amplitude variability
- Room 1: Amplitude range expands from ±5dB to ±12dB
- Room 3: New peak at ~150Hz with ±18dB swing
- Room 25: Amplitude reaches ±22dB at 250Hz
3. **±1.0m Uncertainty (Bottom Row)**
- Extreme amplitude fluctuations across all rooms
- Room 1: Amplitude swings from -20dB to +20dB
- Room 3: Multiple peaks at 50Hz, 150Hz, and 250Hz
- Room 25: Amplitude exceeds ±25dB at multiple frequencies
### Key Observations
- Amplitude variability increases with uncertainty range
- Expanded uncertainty (±1.0m) causes:
- 300% increase in amplitude range (vs baseline)
- New resonant frequencies appearing
- Loss of coherent patterns in higher rooms
- Shaded areas (confidence intervals) widen proportionally with uncertainty
- Room 3 consistently shows most pronounced fluctuations
### Interpretation
The data demonstrates that spatial uncertainty directly impacts acoustic behavior:
1. **Mechanism**: Larger l_x ranges (U(-1.0,1.0)m) create more destructive/constructive interference patterns
2. **Practical Implication**: Room 3's behavior suggests potential for significant acoustic anomalies in uncertain environments
3. **Design Consideration**: The ±0.25m uncertainty already exceeds typical room tolerance standards (usually ±0.1m)
4. **Risk Assessment**: Room 25's extreme fluctuations indicate potential for catastrophic failure in high-precision applications
*Note: All values are approximate due to image resolution limitations. Legend colors were cross-verified with line placements in all three graph sets.*
</details>
## B. Discussion of Experiments
Several optimizations and fine-tuning approaches have been investigated for the complex reconstruction in real rooms without achieving notable improvements. Instead, we take another approach, and show what happens to the model, when it is exposed to data that are not represented in the training data. To this end, we are interested in assessing the performance of room specialized models. That is, if room dimensions and reverberation time are known, how well will a model trained specifically for that room perform. For this, we introduce new datasets each with 824 realizations for training, 165 for validation, and 11 for testing. Each simulated realization has a randomly positioned source. In total, three such datasets are generated according to the procedure described in Section IV A. The first dataset assumes that room characteristics are known perfectly, we use the parameters of the Listening Room. The second and third datasets introduce uncertainty in the room dimensions. In particular, we alter the length and width of rooms, while keeping the aspect ratio ( l x /l y ) of the room constant. We accomplish this by uniformly sampling an error, which is added to the length of a room, and correct the width to achieve the original aspect ratio. The two datasets sample errors in the range [-0.25, 0.25] m and [-1, 1] m, respectively. The results for the three models evaluated on each of the test sets are shown in Fig. 7. The first column shows how the three models perform on the dataset with no added uncertainties. Even with small variations of the 0.25 m scale, performance rapidly degrades with increasing frequency. On the diagonal, training data match test data, and once again high frequencies see a significant performance decrease with increasing uncertainty. In general, the models do not perform well on datasets with more variation than what is included in their own training data, which can be seen in the three upper right figures.
Further experiments showed that the three models do not generalize to the real-world measurements of the Listening Room. This result indicates that the simplifications imposed during the simulations of rooms causes the simulated sound fields to not represent the exact real rooms we intend it to. That is, a model trained with simulated data generated using exact parameters of a real room will not be able to reconstruct the sound field accurately in the real room. As suggested by our results, neither will a model trained with ± 1 m uncertainty. This calls for inclusion of diverse room parameters when training a model with simulated data if the intended purpose is to use the reconstruction in real rooms.
We showed in Section IV how magnitude reconstruction recovered performance in some of the real rooms by using an extended training dataset with more diverse simulated rooms. The same effect is not observed for complex reconstruction. We believe two factors are the main reasons: 1) the boundary conditions in the simulations assume nearly rigid walls and do not include e.g. phase shifts of real wall reflections; 2) the simulations assume perfectly rectangular rooms with a uniform dis-
tribution of absorption. Thus, we hypothesize that the model does not see representative data during training, analogous to not having the correct room dimensions represented in the training data.
## VI. THE SOUND ZONES APPLICATION
One potential application for the sound field reconstruction presented in this paper, is in the process of setting up sound zones. Sound zones generally refers to the scenario where multiple loudspeakers are used to reproduce individual audio signals to individual people within a room (Betlehem et al. , 2015). To control the sound field at the location of the listeners in the room, it is necessary to know the RTFs between each loudspeaker and locations sampling the listening regions. If the desired locations of the sound zones change over time, it becomes labor intensive to measure all the RTFs in situ. As an alternative, a small set of RTFs could be measured and used to extrapolate the RTFs at the positions of interest.
## 1. Setup
For this example, we will explore the scenario where sound is reproduced in one zone (the bright zone) and suppressed in another zone (the dark zone). 5
The question posed in a sound zones scenario, is how the output of the available loudspeakers should be adjusted to achieve the desired scenario. A simple formulation of this problem in the frequency domain is typically denoted acoustic contrast control and relies on maximizing the ratio of mean square pressure in the bright zone relative to the dark zone (Choi and Kim, 2002). This ratio is termed as the acoustic contrast and can be expressed as
<!-- formula-not-decoded -->
where H B ( ω ) ∈ C M × L is a matrix of RTFs from L loudspeakers to M microphone positions in the bright zone and H D ( ω ) ∈ C M × L are the RTFs from the loudspeakers to points in the dark zone. The adjustment of the loudspeaker responses q ( ω ) ∈ C L can be determined as the eigenvector of ( H H D ( ω ) H D ( ω )+ λ D I ) -1 H H B ( ω ) H B ( ω ) which corresponds to the maximal eigenvalue (Elliott et al. , 2012), where · H denotes the Hermitian transpose. In this investigation, the regularization parameter is chosen as
<!-- formula-not-decoded -->
This choice is made to scale the regularization relative to the maximal singular value of H H D ( ω ) H D ( ω ), thereby, controlling the condition number of the inverted matrix.
## 2. Sparse Reconstruction method
An alternative method for estimating the RTFs at positions of interest can be obtained by a sparse reconstruction problem inspired by (Fernandez-Grande, 2019).
Here, the sound pressure observed at the physical microphone locations are modeled as a combination of impinging plane waves
<!-- formula-not-decoded -->
where s ( · , · ) is defined in (2), φ n ( r m ) = e j k T n r m is the candidate plane wave, propagating with wave number k n ∈ R 3 , to observation point r m ∈ R 3 , and b n ( ω ) ∈ C is the complex weight of the n th candidate plane wave. The candidate plane waves can be obtained by sampling the wave number domain in a cubic grid. Note that the eigenfunctions of the room used in Green's function can be expanded into a number of plane waves whose propagation directions in the wave number domain equals the characteristic frequency of the eigenfunction ( ‖ k n ‖ 2 2 = ( ω/c ) 2 ). This fact was used in (Fernandez-Grande, 2019) to regularize the sparse reconstruction problem as
<!-- formula-not-decoded -->
where λ ∈ R + and L ( ω ) ∈ R N × N is a diagonal matrix, where the diagonal elements express the distance between the characteristic frequency associated with the n th candidate plane wave and the angular excitation frequency ω as |‖ k N ‖ 2 2 -( ω/c ) 2 | .
Note that the sparse reconstruction model is not directly comparable to the proposed sound field reconstruction. This is due to the sparse reconstruction relying on knowledge of the absolute locations of the microphone observations. The proposed algorithm, on the other hand, only requires the relative microphone locations on a unitless observation grid.
## 3. Experiments
For the experiments, we use the simulated Listening Room from the previous section, with eight loudspeakers placed at the corners of the floor and halfway between the corners. We have two predefined zones in the middle of the room, which are bright and dark zone respectively. We now, sample random positions in the 32 by 32 x,ygrid 1 m above the floor and use those observations to estimate the RTFs within the zones.
We compare the sparse reconstruction method to the deep learning-based model trained in the previous section. Specifically, the room specialized models are used.
The resulting performance is evaluated in terms of the acoustic contrast over 50 random microphone samplings for each number of microphones. In Fig. 8 the results are based on evaluations using the true RTFs when the loudspeaker weights are determined using either the true RTFs, estimated RTFs based on the model trained with simulated room with no added uncertainties, or estimates based on the sparse reconstruction. It
FIG. 8. Contrast results for the dataset with no added uncertainty to the simulated Listening Room (50 different observation masks). (blue): Perfectly known TFs. (black): Deep learning model. (red): Sparse reconstruction. (dashed): ± 1 standard deviation.
<details>
<summary>Image 8 Details</summary>

### Visual Description
## Line Graph: Frequency vs. Contrast Analysis
### Overview
The image contains two line graphs labeled (a) and (b), each depicting contrast (in dB) as a function of frequency (in Hz). Both graphs feature three data series represented by distinct line styles and colors:
- **Signal A**: Solid blue line
- **Signal B**: Dashed black line
- **Signal C**: Dotted red line
Vertical dotted lines are drawn at 50 Hz, 100 Hz, and 200 Hz to highlight specific frequency intervals.
### Components/Axes
- **X-axis**: Frequency [Hz], ranging from 0 to 200 Hz.
- **Y-axis**: Contrast [dB], ranging from -10 to 30 dB.
- **Legend**: Positioned at the top-left of each graph, with labels:
- Signal A (blue)
- Signal B (black)
- Signal C (red)
- **Annotations**: Graphs labeled (a) and (b) at the bottom.
### Detailed Analysis
#### Graph (a):
- **Signal A (blue)**: Peaks at ~50 Hz (~28 dB), ~100 Hz (~25 dB), and ~150 Hz (~22 dB).
- **Signal B (black)**: Peaks at ~75 Hz (~22 dB), ~125 Hz (~18 dB), and ~175 Hz (~15 dB).
- **Signal C (red)**: Peaks at ~25 Hz (~12 dB), ~75 Hz (~10 dB), and ~125 Hz (~8 dB).
- **Trends**: Signal A exhibits the highest contrast, with sharp peaks at regular intervals. Signal B shows lower contrast with broader peaks. Signal C has the lowest contrast, with peaks spaced further apart.
#### Graph (b):
- **Signal A (blue)**: Peaks at ~50 Hz (~25 dB), ~100 Hz (~20 dB), and ~150 Hz (~18 dB).
- **Signal B (black)**: Peaks at ~75 Hz (~18 dB), ~125 Hz (~15 dB), and ~175 Hz (~12 dB).
- **Signal C (red)**: Peaks at ~25 Hz (~10 dB), ~75 Hz (~8 dB), and ~125 Hz (~6 dB).
- **Trends**: Similar to graph (a), but with slightly reduced peak amplitudes for all signals.
### Key Observations
1. **Signal A** consistently shows the highest contrast across both graphs, with peaks aligned at 50 Hz, 100 Hz, and 150 Hz.
2. **Signal B** exhibits lower contrast than Signal A but higher than Signal C, with peaks at 75 Hz, 125 Hz, and 175 Hz.
3. **Signal C** has the lowest contrast, with peaks at 25 Hz, 75 Hz, and 125 Hz.
4. Vertical lines at 50 Hz, 100 Hz, and 200 Hz may indicate critical frequency thresholds or measurement intervals.
5. Both graphs share similar patterns, suggesting consistent behavior across different conditions or datasets.
### Interpretation
The data suggests that Signal A dominates in terms of contrast, potentially indicating stronger or more dominant waveforms at specific frequencies. Signal B and C show progressively weaker responses, which could reflect differences in signal strength, filtering, or system response characteristics. The vertical lines at 50 Hz, 100 Hz, and 200 Hz may serve as reference points for analyzing harmonic relationships or system resonances. The consistency between graphs (a) and (b) implies that the observed trends are robust across the measured conditions.
### Uncertainties
- Contrast values are approximate due to the lack of gridlines or numerical annotations on the lines themselves.
- Peak positions are estimated based on visual alignment with the x-axis.
- The exact relationship between the signals (e.g., phase differences, amplitude ratios) cannot be determined without additional context.
</details>
is observed that the deep learning-based model performs better than the sparse reconstruction below 150 Hz for 5 and 15 microphones. Above 150 Hz, both models struggle to provide sufficiently accurate RTFs to create sound zones.
In Fig. 9, the model specialized for the Listening Room with l x + U ( -1 . 0 , 1 . 0) m, is compared to the sparse reconstruction. As expected, the resulting performance is reduced for the model. However, it is observed that there is still a benefit when using 5 microphones. At 15 microphones, on the other hand, the performance is comparable for both methods.
These results indicate that sound zones could be created based on sound fields extrapolated from very few microphone positions. However, at this stage it requires models which are specialized to the particular room or a narrow range of rooms. Alternatively, it would be required to increase the number of microphones to improve the accuracy of the estimated RTFs.
## VII. CONCLUSION
In this paper, deep learning-based sound field reconstruction is evaluated using a new set of extensive mea-
FIG. 9. Contrast results for the simulated Listening Room with l x + U ( -1 . 0 , 1 . 0) m (50 different observation masks). (blue): Perfectly known TFs. (black): Deep learning model. (red): Sparse reconstruction. (dashed): ± 1 standard deviation.
<details>
<summary>Image 9 Details</summary>

### Visual Description
## Line Graphs: Contrast vs. Frequency Analysis
### Overview
The image contains two line graphs (labeled a) and b)) depicting contrast measurements in decibels (dB) across a frequency range of 0–200 Hz. Each graph includes three data series represented by distinct line styles and colors: solid blue, dashed red, and dotted black. The graphs share identical axes but differ in data series behavior.
### Components/Axes
- **X-axis**: Frequency [Hz], scaled linearly from 0 to 200 Hz with gridlines at 50, 100, and 150 Hz.
- **Y-axis**: Contrast [dB], scaled linearly from -10 to 30 dB with gridlines at 0, 10, 20, and 30 dB.
- **Legends**: Located in the top-right corner of both graphs. Colors and line styles correspond as follows:
- **Solid blue**: Primary data series (highest contrast values).
- **Dashed red**: Secondary data series (intermediate contrast values).
- **Dotted black**: Tertiary data series (lowest contrast values).
### Detailed Analysis
#### Graph (a)
- **Solid blue line**: Peaks at ~100 Hz (~25 dB), dips to ~5 dB at 50 Hz, and rises again to ~20 dB at 200 Hz. Shows moderate amplitude modulation.
- **Dashed red line**: Peaks at ~150 Hz (~15 dB), dips to ~0 dB at 50 Hz, and fluctuates between 0–10 dB. Less stable than the blue line.
- **Dotted black line**: Peaks at ~75 Hz (~8 dB), dips to ~-5 dB at 100 Hz, and stabilizes near 0 dB at 200 Hz. Most erratic trend.
#### Graph (b)
- **Solid blue line**: Peaks at ~50 Hz (~30 dB) and ~150 Hz (~28 dB), with a trough of ~10 dB at 100 Hz. Highest amplitude modulation.
- **Dashed red line**: Sharp dip to ~-5 dB at 100 Hz, peaks at ~200 Hz (~10 dB). More volatile than in graph (a).
- **Dotted black line**: Deepest dip at 100 Hz (~-10 dB), peaks at ~25 Hz (~5 dB). Most unstable series.
### Key Observations
1. **Consistent Trends**: The solid blue line consistently exhibits the highest contrast values across both graphs, suggesting it represents a dominant or stable factor.
2. **Frequency-Specific Dips**: Both graphs show significant contrast reductions at 100 Hz (black line in a), 100 Hz (red line in b), and 50 Hz (red line in a).
3. **Amplitude Variability**: The dotted black line demonstrates the greatest instability, with sharp troughs and peaks, potentially indicating noise or transient effects.
4. **Graph Discrepancies**: Graph (b) shows more pronounced amplitude modulation (e.g., 30 dB peak in blue line) compared to graph (a), suggesting differing experimental conditions.
### Interpretation
The data likely represents acoustic or signal processing measurements, where contrast (e.g., signal-to-noise ratio) varies with frequency. The solid blue line’s consistent dominance implies a primary contributor (e.g., a stable sound source), while the dashed red and dotted black lines may represent secondary factors or noise. The recurring dips at 50–100 Hz suggest resonant frequencies or interference patterns. The increased volatility in graph (b) could indicate experimental perturbations (e.g., environmental noise, equipment calibration differences). The black line’s erratic behavior in both graphs warrants further investigation into measurement artifacts or transient events.
</details>
surements from real rooms, which are released alongside the paper. The focus of the work is threefold: examine performance of simulation-based learning of magnitude reconstruction in real rooms, extend reconstruction to complex-valued sound fields, and show a sound zone application taking advantage of the reconstructed sound fields. Experiments for each of the three directions indicate promising aspects of data-driven sound field reconstruction, even with a low number of arbitrarily placed microphones.
In the future, it would be of interest to investigate whether transfer learning can help bridge the discrepancies between simulated and real data. With the addition of more rooms, some could be used in the training phase. Furthermore, three-dimensional reconstruction can be achieved using available convolutional models designed specifically to solve three-dimensional problems.
## ACKNOWLEDGMENTS
This work is part of the ISOBEL Grand Solutions project, and is supported in part by the Innovation Fund Denmark (IFD) under File No. 9069-00038A.
- 1 The data are collected under the Interactive Sound Zones for Better Living (ISOBEL) project, which aims to develop interactive sound zone systems, responding to the need for sound exposure control in dynamic real-world contexts, adapted to and tested in healthcare and homes. The ISOBEL Sound Field dataset can be accessed at https://doi.org/10.5281/zenodo.4501339 .
- 2 Further details of the experimental setup and protocol, e.g. equipment, are available in the measurement reports included with the dataset.
3 See footnote 2.
4 Room B has measurements at a single height: 1 meter above the floor.
- 5 The use case with multiple individual audio signals can be realized using superposition of this solution and one where the role of bright and dark zone are reversed.
- Ajdler, T., Sbaiz, L., and Vetterli, M. ( 2006 ). 'The Plenacoustic Function and Its Sampling,' IEEE Transactions on Signal Processing 54 (10), 3790-3804, doi: 10.1109/TSP.2006.879280 .
- Antonello, N., Sena, E. D., Moonen, M., Naylor, P. A., and van Waterschoot, T. ( 2017 ). 'Room Impulse Response Interpolation Using a Sparse Spatio-Temporal Representation of the Sound Field,' IEEE/ACM Transactions on Audio, Speech, and Language Processing 25 (10), 1929-1941, doi: 10.1109/TASLP.2017. 2730284 .
- Bertalmio, M., Sapiro, G., Caselles, V., and Ballester, C. ( 2000 ). 'Image inpainting,' in Proceedings of the 27th Annual Conference on Computer Graphics and Interactive Techniques , SIGGRAPH '00, ACM Press/Addison-Wesley Publishing Co., USA, pp. 417-424, doi: 10.1145/344779.344972 .
- Betlehem, T., Zhang, W., Poletti, M. A., and Abhayapala, T. D. ( 2015 ). 'Personal Sound Zones: Delivering interface-free audio to multiple listeners,' IEEE Signal Processing Magazine 32 (2), 81-91, doi: 10.1109/MSP.2014.2360707 .
- Cecchi, S., Carini, A., and Spors, S. ( 2018 ). 'Room Response Equalization-A Review,' Applied Sciences 8 (1), 16, doi: 10. 3390/app8010016 .
- Choi, J., and Kim, Y. ( 2002 ). 'Generation of an acoustically bright zone with an illuminated region using multiple sources,' Journal of the Acoustical Society of America 111 (4), 1695-1700.
- Dong, C., Loy, C. C., He, K., and Tang, X. ( 2016 ). 'Image SuperResolution Using Deep Convolutional Networks,' IEEE Transactions on Pattern Analysis and Machine Intelligence 38 (2), 295307, doi: 10.1109/TPAMI.2015.2439281 .
- Elliott, S. J., Cheer, J., Choi, J., and Kim, Y. ( 2012 ). 'Robustness and regularization of personal audio systems,' IEEE Transactions on Audio, Speech, and Language Processing 20 (7), 21232133.
- Farina, A. ( 2000 ). 'Simultaneous Measurement of Impulse Response and Distortion with a Swept-Sine Technique,' in Proceedings of the Audio Engineering Society Convention 108 .
- Fernandez-Grande, E. ( 2019 ). 'Sound field reconstruction in a room from spatially distributed measurements,' in 23rd International Congress on Acoustics , pp. 4961-68.
- Genovese, A. F., Gamper, H., Pulkki, V., Raghuvanshi, N., and Tashev, I. J. ( 2019 ). 'Blind Room Volume Estimation from Singlechannel Noisy Speech,' in ICASSP 2019 - 2019 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP) , pp. 231-235, doi: 10.1109/ICASSP.2019.8682951 .
- Glorot, X., and Bengio, Y. ( 2010 ). 'Understanding the difficulty of training deep feedforward neural networks,' in Proceedings of the Thirteenth International Conference on Artificial Intelligence and Statistics , pp. 249-256.
- ISO 3382-2:2008 ( 2008 ). 'Acoustics - Measurement of room acoustic parameters - Part 2: Reverberation time in ordinary rooms,' Standard.
- ITU-R BS.1116-3 ( 2015 ). 'Methods for the subjective assessment of small impairments in audio systems,' Standard.
- Jacobsen, F., and Juhl, P. M. ( 2013 ). Fundamentals of General Linear Acoustics (John Wiley & Sons).
- Karjalainen, M., Makivirta, A., Antsalo, P., and Valimaki, V. ( 2001 ). 'Low-frequency modal equalization of loudspeaker-room responses,' in Audio Engineering Society Convention 111 .
- Kim, J., El-Khamy, M., and Lee, J. ( 2020 ). 'T-GSA: Transformer with Gaussian-Weighted Self-Attention for Speech Enhancement,' in ICASSP 2020 - 2020 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP) , pp. 6649-6653, doi: 10.1109/ICASSP40776.2020.9053591 .
- Kingma, D. P., and Ba, J. ( 2014 ). 'Adam: A Method for Stochastic Optimization,' arXiv:1412.6980 [cs] .
- Ledig, C., Theis, L., Huszar, F., Caballero, J., Cunningham, A., Acosta, A., Aitken, A., Tejani, A., Totz, J., Wang, Z., and Shi, W. ( 2017 ). 'Photo-realistic single image super-resolution using a generative adversarial network,' in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR) .
- Liu, G., Reda, F. A., Shih, K. J., Wang, T.-C., Tao, A., and Catanzaro, B. ( 2018 ). 'Image Inpainting for Irregular Holes Using Partial Convolutions,' in Computer Vision - ECCV 2018 , edited by V. Ferrari, M. Hebert, C. Sminchisescu, and Y. Weiss, Lecture Notes in Computer Science, Springer International Publishing, Cham, pp. 89-105, doi: 10.1007/978-3-030-01252-6\_6 .
- Llu´ ıs, F., Mart´ ınez-Nuevo, P., Møller, M. B., and Shepstone, S. E. ( 2020 ). 'Sound field reconstruction in rooms: Inpainting meets super-resolution,' The Journal of the Acoustical Society of America 148 (2), 649-659, doi: 10.1121/10.0001687 .
- Mignot, R., Chardon, G., and Daudet, L. ( 2014 ). 'Low Frequency Interpolation of Room Impulse Responses Using Compressed Sensing,' IEEE/ACM Transactions on Audio, Speech, and Language Processing 22 (1), 205-216, doi: 10.1109/TASLP. 2013.2286922 .
- Møller, M. B., and Østergaard, J. ( 2020 ). 'A Moving Horizon Framework for Sound Zones,' IEEE/ACM Transactions on Audio, Speech, and Language Processing 28 , 256-265, doi: 10.1109/TASLP.2019.2951995 .
- Nowakowski, T., de Rosny, J., and Daudet, L. ( 2017 ). 'Robust source localization from wavefield separation including prior information,' The Journal of the Acoustical Society of America 141 (4), 2375-2386, doi: 10.1121/1.4979258 .
- Radlovic, B. D., Williamson, R. C., and Kennedy, R. A. ( 2000 ). 'Equalization in an acoustic reverberant environment: Robustness results,' IEEE Transactions on Speech and Audio Processing 8 (3), 311-319, doi: 10.1109/89.841213 .
- Ronneberger, O., Fischer, P., and Brox, T. ( 2015 ). 'U-Net: Convolutional Networks for Biomedical Image Segmentation,' in Medical Image Computing and Computer-Assisted Intervention - MICCAI 2015 , edited by N. Navab, J. Hornegger, W. M. Wells, and A. F. Frangi, Lecture Notes in Computer Science, Springer International Publishing, Cham, pp. 234-241, doi: 10.1007/978-3-319-24574-4\_28 .
- Tylka, J. G., and Choueiri, E. ( 2015 ). 'Comparison of techniques for binaural navigation of higher-order ambisonic soundfields,' in Audio Engineering Society Convention 139 .
- Verburg, S. A., and Fernandez-Grande, E. ( 2018 ). 'Reconstruction of the sound field in a room using compressive sensing,' The Journal of the Acoustical Society of America 143 (6), 3770-3779, doi: 10.1121/1.5042247 .
- Vu, T. P., and Lissek, H. ( 2020 ). 'Low frequency sound field reconstruction in a non-rectangular room using a small number of microphones,' Acta Acustica 4 (2), 5, doi: 10.1051/aacus/2020006 .
- Yang, M., Ma, M. Q., Li, D., Tsai, Y. H., and Salakhutdinov, R. ( 2020 ). 'Complex Transformer: A Framework for Modeling Complex-Valued Sequence,' in ICASSP 2020 - 2020 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP) , pp. 4232-4236, doi: 10.1109/ICASSP40776.2020. 9054008 .
- Yu, W., and Kleijn, W. B. ( 2021 ). 'Room Acoustical Parameter Estimation From Room Impulse Responses Using Deep Neural Networks,' IEEE/ACM Transactions on Audio, Speech, and Language Processing 29 , 436-447, doi: 10.1109/TASLP.2020. 3043115 .