1. Introduction
Extended Reality (XR) technologies emerged in the 1970s, and were first commercialized in the 1990s [
1]. Recent advances in computer chips, machine vision, and algorithms have made XR headsets portable, facilitating their use in medical settings. In the medical field, the application of XR is referred to as Medical Extended Reality (MXR). This term encompasses a range of technologies and applications. Head-mounted displays (HMDs) are commonly used in MXR, with virtual reality(VR) HMDs presenting virtual objects in a virtual world; while augmented reality(AR) HMDs overlay virtual objects onto the real environment, augmenting reality with virtual content.
Although the AR development has been rapid, it faces significant challenges in healthcare applications. The seamless visualization of and interaction with digital content in the real world heavily relies on multiple HMD built-in sensors, e.g., depth cameras, inertial motion units, and color and monochromatic cameras for environmental sensing. The challenges include integrating AR technology seamlessly into clinical workflows, ensuring patient safety and privacy, and overcoming technical limitations such as accuracy and field of view [
2,
3]. Despite these challenges, a growing number of AR applications have been used in training, education, telemedicine, preoperative planning and postoperative assessment, and intraoperative systems. [
4,
5]. For instance, AR HMDs enable capturing of images and videos from a surgeon’s perspective and share them with remote users [
6]. Novel clinical applications are also emerging, providing engaging environments and training for physiotherapy [
7] and pain management [
8,
9], as well as the gamification of physical therapy.
Amid this rapid development and widespread adoption of AR technologies across various applications, evaluating these technologies to assess their reliability has received comparatively less attention. Most efforts have been directed toward evaluating these applications as a whole using qualitative methods [
10,
11,
12,
13], which rely heavily on subjective user feedback and focus on whether a very specific task can be done better or worse using applications reliant on depth sensors. This predominant reliance on subjective metrics underscores the need for an alternative approach emphasizing quantitative evaluation and validation of AR technologies in healthcare settings and focusing on the performance of individual sensors and their contributions to the application’s performance as a whole.
In this work, we have developed a fully automated benchtop experimental test setup designed for quantitative evaluation of augmented reality sensors. In this paper, we aim to assess the performance of depth cameras within AR headsets under conditions closely resembling real-world scenarios, e.g. rotation and translation of user’s head. This paper initially describes the system architecture and the process of automated data collection. Subsequently, we present several evaluation studies conducted using the Intel RealSense L515 LiDAR camera as an example.
2. Materials and Methods
2.1. Experimental Design
The evaluation experimental setup, illustrated in
Figure 1(a), was designed to create a controlled and stable testing environment for evaluating range sensing technologies, with a specific focus on depth camera sensors and augmented reality. This setup aimed to assess the performance of these technologies under various conditions, including motion, position and against different test objects.
The experimental setup, mounted on an anti-vibration optical bench table measuring 6 × 3 feet
2 (model INT3-36-8-A Integrity 3 VCS by Newport Corp, Irvine, CA), is organized into two primary sections. Its left side featured the scanning hardware section, while the remaining space was dedicated to the test object mounting area, as shown in the
Figure 1(a).
To achieve precise and controlled movements, we employed a motion control system from Zaber Technologies(Vancouver, British Columbia), configured in a gantry system. Stages were stacked on top of each other. The lowest -axis of this gantry comprised two actively driven translational stages operating synchronously in a lock-step mode, while the middle -axis linear stage was mounted atop the two -axis linear stages. A rotational stage was positioned on top of the -axis linear stage. The and axes translational stages, featuring meter-long linear guides (LC40B10000) equipped with stepper motors and encoders (NEMA23), provided an accuracy of 60 m and repeatability of less than 20 m across their entire range. Of note, the custom software improvements were made to achieve performance specifications higher than the standard ones posted on the vendor’s webpage. The improved accuracy and repeatability were verified using laser interferometry. The motorized rotational stage is capable of remarkable precision, offering a repeatability of less than 0.02 degrees and an accuracy of 0.08 degrees without additional modifications. These specifications are well-suited for evaluating standalone depth camera and built-in HMD sensors, and are well beyond the typical depth resolution (on the order of 250 ) of current depth cameras on the market.
To enhance modularity and flexibility, all devices under test (DUT) are mounted on separate platforms. The preconfigured platforms could be swapped in and out in under five minutes. Each of the platforms were equipped with Ethernet, USB 3.0, and DC power connections, providing a total power delivery of 150W and several different modes of connectivity, supporting a variety of USB devices, Ethernet devices and portable computing systems such as Raspberry Pi or NVidia Jetson Nano GPU systems. We utilized solid aluminum breadboards (12 inches × 12 inches × 0.25 inches) from Base Lab Tools, USA, as our mounting platforms. Post-processing machining of the platforms aligned the central four holes arranged on a one-inch grid with four 25-mm-spaced mounting holes on the rotational stage, enabling the support of objects weighing up to 50 kg. These platforms served as a means to assemble and prepare devices under evaluation and for testing necessary software before mounting them on the rotation stage, thus facilitating independent work without requiring direct access to the evaluation experimental setup.
Each platform, featuring the preconfigured DUTs, was mounted on a rotation stage above
-
translation stages. This configuration offered three motorized independent movements: two translational and one rotational. This arrangement enabled for the evaluation of range sensing by DUTs with three degrees of freedom. Similarly, the cameras mounted on the platform could have an extra degree of freedom achieved via manual or motorized tilt stage, four total degrees of freedom, including rotation around a vertical axis and tilting around a horizontal axis, as depicted in
Figure 1(c). All stages in our setup were capable of controlled movements at varying speeds and accelerations. This flexibility was essential for assessing camera responses during motion and at different locations, allowing us to simulate a wide range of user activities and scenarios.
While our system’s current configuration could support various depth cameras and a HoloLens 2 augmented reality headset, as indicated in Figures
1(b) and
1(c), this study focused on presenting results using the Intel RealSense L515 LiDAR camera as an example.
In addition to the hardware components, we developed a series of test objects. These were intended to simulate real-world scenarios and play a crucial role in our testing procedures, a sample of which is shown in
Figure 2.
Figure 2(a) features a slanted edge and the motorized screen mounted on the optical table in the test object observation area. The slanted edge is positioned in front of the screen (TB4 - Black Hardboard, 24 x 24 inches), which acts as an adjustable background that can be positioned at various distances away from the slanted edge.
The second is a flat wall test object and shown in
Figure 2(b), and a large flat area of drywall located outside of the test object positioning area due to its’ size.
Our experimental setup with integrated components and flexible configurations was intended to provide a dependable and flexible testing environment for evaluating depth cameras and other range sensing technologies.
2.2. Modular Experimental Setup with Integrated Control and Data Collection System
The experimental setup was designed with a modular approach, enabling quick and seamless alternation of devices under investigation. These devices could be swiftly mounted on top of the rotational stage without impacting the overall system functionality. We utilized a Raspberry Pi 4 Model B Quad Core 64 Bit WiFi Bluetooth (8GB) (Raspberry Pi Foundation, United Kingdom) for synchronizing the various components. Our setup included three different depth cameras connected to the same Raspberry Pi. The Raspberry Pi was networked via an Ethernet cable and ran on an Ubuntu 20.0 Linux system. This setup supports server applications and facilitates high-speed data transfer, with Ethernet bandwidths up to 10 Gbps.
In
Figure 1(b), we show a platform with an Intel Real Sense L515 LiDAR depth camera mounted on it. This camera is powered through and is connected to the control computer via a USB 3.0 Power Delivery (PD) port on the USB hub mounted on the platform. The control computer uses Python-based software that establishes connection with, controls and retrieves data from the camera.
Alternatively, as depicted in
Figure 1(c), a different platform may accommodate the HoloLens 2 augmented reality headset. The HoloLens 2 is powered through a USB 3.0 Power Delivery (PD) port and connects to the control computer directly via WiFi. Connectivity with the HoloLens 2 is established through a server application, written in C programming language, running on the HoloLens 2 AR device. A Python-based client process retrieves real-time data from the built-in sensors of HoloLens 2.
The data collected were frames that reflect the depth measurements captured by the cameras. These frames can be saved in various formats for analysis. Users can configure the motion trajectories, speed, and acceleration of the test objects and cameras during data collection using a pre-set dataset of points, which served as input for our data acquisition software. While the system allows for dynamic positioning, in this study we focused on examples of evaluation under static conditions.
2.2.1. Systematic Calibration and Alignment Process for Depth Cameras
This calibration procedure consisted of two primary phases: intrinsic calibration and extrinsic alignment.
In the first stage, we executed intrinsic calibration in strict accordance with the manufacturer’s guidelines, as referenced in [
14]. This critical step was designed to rectify lens distortions and mitigate any sensor misalignment within the camera. It also encompassed inter-camera alignment, in which we employed a combination of advanced software algorithms and precise manual adjustments to ensure a synchronized and highly precise depth perception across multiple sensors. For an in-depth understanding of our benchtop calibration method, we encourage readers to explore the comprehensive details provided in a published paper [
15].
After intrinsic calibration, we moved to the extrinsic alignment phase, a crucial step that aligns the captured depth data with real-world coordinates custom-tailored to match our study’s specific requirements. For instance, when aligning a depth camera with a flat wall test object, our approach utilized root-mean-square error minimization within two regions of interest (ROI): the horizontal and vertical dimensions. In the horizontal ROI, we meticulously selected a set of five rows of pixels, symmetrically centered around the central pixel. By vertically binning these pixels, we effectively enhanced the signal-to-noise ratio and derived a vector of distances along a horizontal axis. This vector is then subjected to a straight-line fitting process, enabling us to extract the yaw angle. The minimization of this angle plays a pivotal role in achieving a high-precision alignment of the camera with the test object.
We repeated this measurements for the vertical ROI, affording us real-time visualization of the pitch angle. By minimizing the pitch angle, we ensured a robust and accurate alignment of the camera with the test object, further enhancing the reliability of our measurements.
In summary, our benchtop evaluation setup employed a systematic calibration procedure that adhered closely to manufacturer specifications while incorporating our in-house developed alignment techniques. This thorough calibration process ensured that our depth cameras performed optimally and delivered precise and synchronized depth perception across various sensors.
3. Examples of Evaluation Studies
The depth cameras return an array of data with three coordinates X,Y,Z, where Z is depth coordinate, and X and Y are in object plane which is perpendicular to the depth axis. The examples of evaluation metrics in following section aim at establishing accuracy and precision of numbers reported by depth cameras.
Section 3.1 measures modulation transfer function(MTF) of a sharp change in Z values created by a physical slanted edge.
Section 3.2 provides an example of evaluation metric that looks at precision and accuracy of Z measurements for fixed X and Y coordinates.
3.1. Image Plane Spatial Resolution
In this section, we explore how the spatial resolution of a depth camera varies with different contrast levels. Here, "contrast" is defined as the range of depth values detected by the camera. Our method involved using a slanted edge test object (
Figure 2(a)) positioned at a constant distance from a uniform background panel, creating a distinct difference in depth values between the test object and the panel. Images of this arrangement were captured with the back panel set at various distances 10 mm, 50 mm, and 160 mm - as illustrated in
Figure 3. We then performed 2D Modulation Transfer Function (MTF) measurements on each image to assess the spatial resolution within the image plane, utilizing an open-source software tool for standard 2D MTF analysis [
16].
The resulting MTF data, as shown in
Figure 4, were acquired using the Intel RealSense L515 LiDAR range sensing camera. Interestingly, we observed a significant reduction in resolution with decreasing contrast. This finding highlights that spatial resolution deteriorates as the distance between objects in the field of view lessens, indicating a dependency of the camera’s resolution in the
-
plane on the contrast in image depth values, with
c representing the camera.
3.2. Z-Precision and Z-Accuracy Measurements
We aimed to assess the precision and accuracy of distance measurements obtained from depth cameras, employing specific metrics to provide a comprehensive evaluation. Assessment of precision in distance measurements serves as a crucial metric, enabling the placement of error bars on the distance values returned by depth cameras. In this method, the depth camera was first positioned at a fixed distance of from a flat-wall test object. We performed careful configuration and alignment to ensure that the optical axis (-axis) of the camera was perpendicular to the test object and aligned with the -axis of the gantry system. Next, we precisely repositioned the depth camera under evaluation using an automated motion control system at M different distances from the flat wall test object. At each of these positions, we collected and saved a dataset comprised of N depth images. The acquired data was analyzed and - mean value, - root-mean-square error, - accuracy in z at j positing, and - standard deviation were computed for each position() of the camera.
-
Root-mean-square error (RMSE) between the observed distance
, obtained using a depth sensing camera and the known values
of depth camera position, were calculated using a single pixel in the center of the depth image). The RMSE is calculated as:
Here, i - represents image index, j - represents positing index, N represents the total number of images collected at fixed position of the depth camera with respect to the flat wall test object.
-
Accuracy of the z-value measurement, as measured by the depth camera as a deviation of measured mean values
and the known camera position, the ground truth. The mean value can be calculated as:
Here,
represents the values of depth returned by the depth camera for a given pixel. Accuracy in z -
can be calculated as:
Precision of the z-value measurement, as measured by the depth camera (
) of measured depth values across different positions of depth camera with respect to the flat wall test object, calculated as:
As an example, we positioned the camera at = 500 mm away from the flat wall. We confirmed the distance using external ruler and collected 1000 images to ensure good statistical accuracy in measuring of and in establishing one-to-one correspondence between encoder readings and physical distance. Later, we repositioned the depth camera to = 600 mm, 700 mm, 800 mm, 900 mm and 1000 mm, collected 100 images(N = 100) from each location, and computed the metrics mentioned above.
Figure 5(a) displays the measured distances by the central pixel for all six locations, offering insights into how distance measurements vary with camera placement. In
Figure 5(b), for each position we present the standard deviation and RMSE of the measured position. Notably, the standard deviation exhibits degradation as the camera is positioned farther away from the test object and RMSE drastically degrades due to large errors in measured mean position and actual position.
Figure 5(c) illustrates the difference between the measured distance and the position specified by precise translational stages, read from built-in encoders. This comparison highlights errors in Z-accuracy, emphasizing that calibration at a single anchor point is insufficient for ensuring accurate calibration across the entire range of distances.
Finally,
Figure 5(d) Panel D showcases six histograms of independent measurements obtained from 100 sequential depth images acquired at six known distances (
). These values were fitted to a Gaussian function, enabling us to extract mean and standard deviation values from the fit. This analysis provides valuable insights into the distribution of depth measurements at different distances, as well as accuracy and precision of distance measurements.
Through this evaluation of precision in distance measurements, our setup and analysis shed light on the variations and accuracy considerations associated with depth camera performance across varying distances.
3.3. Pearson Pixel-to-Pixel Correlation
Within the framework of the setup detailed in this paper, one can efficiently compute correlations between distance measurements obtained from distinct pairs of pixels. This correlation analysis employs the Pearson correlation coefficient [
17,
18], providing valuable insights into the relationships between these measurements.
The dataset under examination comprises 1800 sequentially measured images, equivalent to 60 seconds of data collection at a frame rate of 30 frames per second. In
Figure 6(a) Panel A, we present static distance measurements obtained from a camera positioned facing a flat wall. Specifically, three traces are displayed: one for the central pixel, another for a pixel situated 100 pixels above the center, and a third for a pixel located 100 pixels to the right. To provide a clear overview of how average depth measurements evolve over time, we incorporate a solid line to represent the averaged data. Notably, these three traces exhibit significant correlation, particularly at larger distances of approximately 100 mm within the XY plane.
In
Figure 6(b), we show the correlation map generated by correlating the central pixel with all other pixels within the dataset, enabling us to explore comprehensive patterns of correlation.
Figure 6(c) displays both vertical and horizontal slices through the central pixel. Here, a correlation coefficient value of one signifies the pixel’s correlation with itself, acting as a reference point for the analysis.
Figure 6(d) and
6(e) delve into the central region of interest, along with corresponding slices. Within this region, we observe a notable long-distance correlation, characterized by a full width at half maximum of 10 pixels and a baseline of 0.4. At a 500 mm distance, this translates to an approximate 10 mm correlation. Intriguingly, the correlation coefficient consistently remains positive and does not dip below zero, highlighting that noise is predominantly correlated rather than uncorrelated.
Through this comprehensive analysis, our experimental setup and findings offer insights into the correlation of distance measurements between pixels, particularly across varying distances and within specific regions of interest. The measurements showed that the Intel RealSense L515 LiDAR has high correlation between different pixels at long distances, e.g., 100 pixels corresponding to approximately 100 mm at a 500 mm distance. This finding might put constraints on techniques used to increase precision such as averaging over several nearby pixels. The long-distance correlation between pixels will not improve precision since distance values in nearby pixels move synchronously.
4. Current state and Future Direction
The proposed design embodies a modular and extensible approach to controlling software and hardware, opening the door to a multitude of research opportunities. Presently, our control software enables the movement of the observation platform along simple trajectories. However, it is worth noting that our design inherently supports motion along complex trajectories and mounting of different augmented reality sensors. In the future, we plan to expand the use of the evaluation setup and unlock capabilities that would potentially help us to explore numerous critical questions. These include head, hand, or object tracking; static and dynamic image registration and quality; latency in head-mounted displays; temporal noise associated with moving objects or observers; and other emerging needs in sensor characterization and testing in task-based experiments.
This study aimed to evaluate the effectiveness of the proposed metrics and their application in real-world scenarios. Looking ahead, we plan to conduct a comprehensive comparative study involving various range sensing technologies.
Author Contributions
Conceptualization, V.S.; methodology, V.S.; software, V.S.; validation, V.S. and B.G.; formal analysis, V.S. and B.G.; investigation, V.S.; resources, V.S.; data curation, V.S.; writing—original draft preparation, V.S.; writing—review and editing, V.S. and B.G.; visualization, V.S. and B.G; supervision, V.S.; project administration, V.S.; funding acquisition, V.S. All authors have read and agreed to the published version of the manuscript.
Funding
This is a contribution of the U.S. Food and Drug Administration. No funding was received for conducting this study.
Acknowledgments
The mention of commercial products, their resources or their use in connection with material reported herein is not to be construed as either an actual or implied endorsement of such products by the Department of Health and Human Services. We would like to thank Joanne Berger, FDA Library, for editing a draft of the manuscript.
This is a contribution of the U.S. Food and Drug Administration and is not subject to copyright.
Abbreviations
The following abbreviations are used in this manuscript:
AR |
Augmented Reality |
VR |
virtual reality |
XR |
Extended Reality |
MXR |
Medical Extended Reality |
HMD |
Head-mounted display |
LiDAR |
Light Detection and Ranging |
CA |
California |
DUT |
devices under test |
GPU |
graphics processing unit |
USB |
Universal Serial Bus |
PD |
Power Delivery |
ROI |
region of interest |
MTF |
Modulation Transfer Function |
RMSE |
Root-mean-square error |
U.S. |
United States |
FDA |
Food and Drug Administration |
References
- Sutherland, I.E. A head-mounted three dimensional display. AFIPS ’68 1968, pp. 757–764. [CrossRef]
- Beams, R.; Brown, E.; Cheng, W.e.a. Evaluation Challenges for the Application of Extended Reality Devices in Medicine. J Digit Imaging 2022. [CrossRef]
- Dennler, C.; Bauer, D.E.; Scheibler, A.G.; Spirig, J.; Götschi, T.; Fürnstahl, P.; Farshad, M. Augmented reality in the operating room: a clinical feasibility study. BMC Musculoskeletal Disorders 2021, 22. [CrossRef]
- William Omar Contreras López, Paula Alejandra Navarro, S.C. Intraoperative clinical application of augmented reality in neurosurgery: A systematic review. Clin Neurol Neurosurg. 2018. [CrossRef]
- Bollen, E.; Awad, L.; Langridge, B.; Butler, P.E.M. The intraoperative use of augmented and mixed reality technology to improve surgical outcomes: A systematic review. The International Journal of Medical Robotics and Computer Assisted Surgery 2022. [CrossRef]
- Huang, E.; Knight, S.; Guetter, C.; Davis, C.; Moller, M.; Slama, E.; Crandall, M. Telemedicine and telementoring in the surgical specialties: a narrative review. The American Journal of Surgery 2019, 218(4).
- Carl, E.; Stein, A.; Levihn-Coon, A.; Pogue, J.; Rothbaum, B.; Emmelkamp, P.; Asmundson, G.; Carlbring, P.; Powers, M. Virtual reality exposure therapy for anxiety and related disorders: A meta-analysis of randomized controlled trials. Journal of Anxiety Disorders 2019, 61.
- Cherkin, D.; Sherman, K.; Balderson, B.; Cook, A.; Anderson, M.; Hawkes, R.; Hansen, K.; Turner, J. Effect of mindfulness-based stress reduction vs cognitive behavioral therapy or usual care on back pain and functional limitations in adults with chronic low back pain: a randomized clinical trial. JAMA 2016, 315(12).
- Spiegel, B.; Fuller, G.; Lopez, M.; Dupuy, T.; Noah, B.; Howard, A.; Albert, M.; Tashjian, V.; Lam, R.; Ahn, J.; et al. Virtual reality for management of pain in hospitalized patients: A randomized comparative effectiveness trial. 2019. 14(8).
- Addison, A.P.; Addison, P.S.; Smit, P.; Jacquel, D.; Borg, U.R. Noncontact Respiratory Monitoring Using Depth Sensing Cameras: A Review of Current Literature. Sensors 2021, 21. [CrossRef]
- Xavier Cejnog, L.W.; Marcondes Cesar, R.; de Campos, T.E.; Meirelles Carril Elui, V. Hand range of motion evaluation for Rheumatoid Arthritis patients. In Proceedings of the 2019 14th IEEE International Conference on Automatic Face & Gesture Recognition (FG 2019), 2019, pp. 1–5. [CrossRef]
- Hernández, A.; Arteaga-Marrero, N.; Villa, E.; Fabelo, H.; Callicó, G.M.; Ruiz-Alzola, J. Automatic Segmentation Based on Deep Learning Techniques for Diabetic Foot Monitoring Through Multimodal Images. In Proceedings of the Image Analysis and Processing – ICIAP 2019; Ricci, E.; Rota Bulò, S.; Snoek, C.; Lanz, O.; Messelodi, S.; Sebe, N., Eds., Cham, 2019; pp. 414–424.
- Burger, L.; Sharan, L.; Karl, R.; Wang, C.; Karck, M.; Simone, R.D.; Ivo Wolf, G.R.; Engelhardt, S. Comparative evaluation of three commercially available markerless depth sensors for close-range use in surgical simulation. Int J CARS 2023, 18, 1109–1118. [CrossRef]
- Intel® RealSense™ LiDAR Camera L515 User Guide. https://support.intelrealsense.com/hc/en-us/articles/360051646094-Intel-RealSense-LiDAR-Camera-L515-User-Guide. (accessed: 01.08.2024).
- Herrera C., D.; Kannala, J.; Heikkilä, J. Accurate and Practical Calibration of a Depth and Color Camera Pair. In Proceedings of the Computer Analysis of Images and Patterns, Berlin, Heidelberg, 2011; pp. 437–445.
- Granton, P. Slant Edge Script. https://www.mathworks.com/matlabcentral/fileexchange/28631-slant-edge-script, 2023. MATLAB Central File Exchange. Retrieved December 7, 2023.
- Janse, R.J.; Hoekstra, T.; Jager, K.J.; Zoccali, C.; Tripepi, G.; Dekker, F.W.; van Diepen, M. Conducting correlation analysis: important limitations and pitfalls. Clinical Kidney Journal 2021, 14, 2332–2337, [https://academic.oup.com/ckj/article-pdf/14/11/2332/41100015/sfab085.pdf]. [CrossRef]
- Zar, J.H. Biostatistical analysis; Pearson Education India, 1999.
Figure 1.
The experimental setup consists of an optical table 6 feet by 3 feet (INT3-36-8-A Integrity 3 VCS by Newport Corp). The table-top is divided into two sections: on the left in Panel A - motion control system(Zaber Technologies) and on the right in Panel A - test object observation area for positioning evaluation test objects). The motion system allows controlled motion in two horizontal axes and , and in rotational axis . A platform with devices under investigation can be mounted on top of the rotational stage. The depth camera generates data consisting of points in space, each assigned coordinates is noted as , , and . Panel B shows a platform with Intel Real Sense L515 LiDAR depth camera mounted on it and Panel C shows the Microsoft HoloLens 2 augmented reality headset mounted.
Figure 1.
The experimental setup consists of an optical table 6 feet by 3 feet (INT3-36-8-A Integrity 3 VCS by Newport Corp). The table-top is divided into two sections: on the left in Panel A - motion control system(Zaber Technologies) and on the right in Panel A - test object observation area for positioning evaluation test objects). The motion system allows controlled motion in two horizontal axes and , and in rotational axis . A platform with devices under investigation can be mounted on top of the rotational stage. The depth camera generates data consisting of points in space, each assigned coordinates is noted as , , and . Panel B shows a platform with Intel Real Sense L515 LiDAR depth camera mounted on it and Panel C shows the Microsoft HoloLens 2 augmented reality headset mounted.
Figure 2.
Examples of available test objects positioned in the regulatory test object observation area, from left to right: Panel 2(a) - a slanted edge test object, 60 mm tall, slanted at 4 degrees, mounted in front of a ThorLabs Black Construction Hardboard measuring 24 × 24 inches; Panel 2(b) - a segment of a flat wall test object, 2×2 m2, used for alignment of a depth camera optical axis and later measurement of a static Pearson correlation coefficient
Figure 2.
Examples of available test objects positioned in the regulatory test object observation area, from left to right: Panel 2(a) - a slanted edge test object, 60 mm tall, slanted at 4 degrees, mounted in front of a ThorLabs Black Construction Hardboard measuring 24 × 24 inches; Panel 2(b) - a segment of a flat wall test object, 2×2 m2, used for alignment of a depth camera optical axis and later measurement of a static Pearson correlation coefficient
Figure 3.
Acquired images of the slanted edge test object with varying distances between the slanted edge and the background panel, illustrating different contrast levels in depth: (a) low contrast at 10 mm, (b) medium contrast at 50 mm, and (c) high contrast at 160 mm. The gray scale in each figure shows the range in depth measured in the images in mm. In gray-scale heat maps, the dark area always shows the distance (600 mm) to the slanted edge since it is stationary, and the white area shows distance to the background screen. Hence, the white areas in panels a, b, and c represents 610 mm, 650 mm and 760 mm distances, respectively. The physical size of displayed region of interest is 100 mm(y-axis) by 80 mm(x-axis).
Figure 3.
Acquired images of the slanted edge test object with varying distances between the slanted edge and the background panel, illustrating different contrast levels in depth: (a) low contrast at 10 mm, (b) medium contrast at 50 mm, and (c) high contrast at 160 mm. The gray scale in each figure shows the range in depth measured in the images in mm. In gray-scale heat maps, the dark area always shows the distance (600 mm) to the slanted edge since it is stationary, and the white area shows distance to the background screen. Hence, the white areas in panels a, b, and c represents 610 mm, 650 mm and 760 mm distances, respectively. The physical size of displayed region of interest is 100 mm(y-axis) by 80 mm(x-axis).
Figure 4.
Modulation Transfer Function (MTF) calculated for three varying contrast levels, determined by the distance from the slanted edge to the background panel: low contrast at 10 mm, medium contrast at 50 mm, and high contrast at 160 mm. The resulting spatial resolutions of 714.4 , 862.8 and 1298.9 were for high, medium and low contrast measurements, respectively.
Figure 4.
Modulation Transfer Function (MTF) calculated for three varying contrast levels, determined by the distance from the slanted edge to the background panel: low contrast at 10 mm, medium contrast at 50 mm, and high contrast at 160 mm. The resulting spatial resolutions of 714.4 , 862.8 and 1298.9 were for high, medium and low contrast measurements, respectively.
Figure 5.
z-accuracy measurement of Intel RealSense L515 LiDAR range sensing standalone camera. The depth camera was positioned at six fixed distances 500, 600, 700, 800, 900 and 1000mm, and 100 images were collected at each position. Panel A - measured position of depth camera from the flat wall test object on y-axis vs. set distance by a gantry system on x-axis; Panel B - standard deviation and RMSE of every 100 images collected at each fixed camera distances; Panel C - difference between measured position and ground truth provided by motion control system, , where each semi-transparant dot represents one measurement and darker color dots show overlapping measurements. The mean value and standard deviation shown on the graph corresponds to entire dataset and ideally should be zero; Panel D - histograms of distance values measured at all distances away from the wall, where the blue circles correspond to measured data, orange show a single Gaussian fit, green vertical lines show mean value and red vertical lines show ground truth. The difference between measured mean value and the truth shows the accuracy of the camera. The mean and standard deviation values for 500 mm distance are 499.6mm and 0.55mm.
Figure 5.
z-accuracy measurement of Intel RealSense L515 LiDAR range sensing standalone camera. The depth camera was positioned at six fixed distances 500, 600, 700, 800, 900 and 1000mm, and 100 images were collected at each position. Panel A - measured position of depth camera from the flat wall test object on y-axis vs. set distance by a gantry system on x-axis; Panel B - standard deviation and RMSE of every 100 images collected at each fixed camera distances; Panel C - difference between measured position and ground truth provided by motion control system, , where each semi-transparant dot represents one measurement and darker color dots show overlapping measurements. The mean value and standard deviation shown on the graph corresponds to entire dataset and ideally should be zero; Panel D - histograms of distance values measured at all distances away from the wall, where the blue circles correspond to measured data, orange show a single Gaussian fit, green vertical lines show mean value and red vertical lines show ground truth. The difference between measured mean value and the truth shows the accuracy of the camera. The mean and standard deviation values for 500 mm distance are 499.6mm and 0.55mm.
Figure 6.
Noise measurements from Intel RealSense L515 LiDAR range sensing standalone camera and Pearson correlation analysis. Panel A - three traces of depth values measured over 60 seconds for three different pixels; Panels B and D show Pearson correlation maps of central pixel vs. all other pixels with Panel D showing the central region of interest. The red square in Panel B highlights the region of interest shown in panel D; Panels C and E show vertical and horizontal slices via central pixel, corresponding to the whole correlation map (Panel B) and the central region of interest (Panel D).
Figure 6.
Noise measurements from Intel RealSense L515 LiDAR range sensing standalone camera and Pearson correlation analysis. Panel A - three traces of depth values measured over 60 seconds for three different pixels; Panels B and D show Pearson correlation maps of central pixel vs. all other pixels with Panel D showing the central region of interest. The red square in Panel B highlights the region of interest shown in panel D; Panels C and E show vertical and horizontal slices via central pixel, corresponding to the whole correlation map (Panel B) and the central region of interest (Panel D).
|
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2024 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (http://creativecommons.org/licenses/by/4.0/).