Analyze Diet
Sensors (Basel, Switzerland)2026; 26(4); 1310; doi: 10.3390/s26041310

Optimising Camera-ChArUco Geometry for Motion Compensation in Standing Equine CT: A CT-Motivated Benchtop Study.

Abstract: Standing equine computed tomography (CT) acquisitions are susceptible to residual postural sway, which can introduce view-inconsistent motion and degrade image quality. External optical tracking based on ChArUco fiducials is a promising, low-cost strategy to enable projection-wise motion compensation, yet quantitative guidance on how camera-marker geometry affects pose-estimation performance remains limited. This CT-motivated benchtop study characterizes how the relative camera-ChArUco configuration influences both the accuracy (bias with respect to ground truth) and the precision (repeatability) of pose estimates obtained from RGB images using OpenCV ChArUco detection and reprojection-error minimization to estimate the rigid camera-to-board transformation. Controlled experiments systematically varied acquisition protocol (continuous repeated estimates at fixed pose versus cyclic repositioning), viewing angle over a wide angular range at two working distances, and camera-to-board distance over multiple depth settings. Ground truth for angular configurations was defined by a stepper-motor rotation stage, while distance ground truth was obtained by ruler measurements. Performance was summarized via mean absolute error and standard deviation across repeated measurements, complemented by variance-based statistical testing with multiple-comparison correction. Cyclic repositioning did not yield evidence of increased variability relative to continuous acquisitions, supporting view-by-view sampling. Viewing angle induced a consistent accuracy-precision trade-off for rotations: frontal views minimized mean error but exhibited higher variability, whereas oblique views reduced jitter at the expense of increased bias. Increasing working distance reduced repeatability, most prominently for depth-related components. Overall, these findings provide pre-clinical guidance for selecting camera/marker placement (moderately oblique viewpoints, limited working distance, sufficient image footprint) before in-scanner and in-vivo validation for standing equine CT motion compensation.
Publication Date: 2026-02-18 PubMed ID: 41755249PubMed Central: PMC12943937DOI: 10.3390/s26041310Google Scholar: Lookup
The Equine Research Bank provides access to a large database of publicly available scientific literature. Inclusion in the Research Bank does not imply endorsement of study methods or findings by Mad Barn.
  • Journal Article

Summary

This research summary has been generated with artificial intelligence and may contain errors and omissions. Refer to the original study to confirm details provided. Submit correction.

Overview

  • This study investigates how the positioning and geometry of a camera relative to ChArUco fiducial markers affect the accuracy and precision of motion tracking used to compensate for motion during standing equine CT scans.
  • The research uses a benchtop experimental setup simulating CT conditions to provide quantitative guidance on optimal camera-marker configurations for improving motion compensation techniques.

Background and Motivation

  • Standing equine CT scans are prone to residual postural sway from the horse, causing inconsistent motion between views and degrading image quality.
  • External optical tracking systems that use ChArUco markers—specialized AprilTags combined with chessboard patterns—offer a low-cost way to track motion with high spatial accuracy.
  • Effective motion compensation requires precise pose estimation, but it is unclear how the physical arrangement (geometry) of the camera relative to these markers influences estimation performance.
  • This gap motivated a controlled benchtop study to emulate CT imaging considerations and systematically evaluate various camera-marker geometries.

Methodology

  • Pose Estimation Technique:
    • OpenCV’s ChArUco detection system was used to identify fiducial points from RGB images.
    • Rigid transformations between camera and marker board were estimated by minimizing reprojection errors.
  • Experimental Setup:
    • A benchtop rig simulated CT acquisition conditions with precise control of angles and distances.
    • Ground truth for angular orientation was provided by a stepper-motor rotation stage for precise rotational offsets.
    • Distance ground truth was measured manually using rulers for accuracy assessment.
  • Variables Tested:
    • Acquisition protocol: continuous repeated measurements at a fixed pose versus cyclic repositioning of the board.
    • Viewing angle: varied over a wide angular range, including frontal and oblique perspectives.
    • Camera-to-board distance: multiple depths were tested to emulate different working distances.
  • Performance Metrics:
    • Accuracy: measured as bias (mean absolute error) compared to ground truth pose.
    • Precision: measured as repeatability or variability (standard deviation) across repeated samples.
    • Statistical testing: variance-based tests with corrections for multiple comparisons were used to ascertain significance of differences.

Key Findings

  • Repositioning Results:
    • No significant increase in variability was found when cyclic repositioning was used versus continuous measurements at a fixed pose.
    • This supports the feasibility of acquiring view-by-view pose estimates during CT scans without loss in measurement precision.
  • Viewing Angle Effects:
    • Frontal camera views minimized the mean pose error, resulting in higher accuracy but showed higher variability (worse precision).
    • Oblique viewpoints reduced jitter (improved precision) but introduced an increase in systematic bias (lower accuracy).
    • This illustrates a trade-off between accuracy and precision depending on camera angle, suggesting that moderate oblique angles may be ideal.
  • Distance Effects:
    • Increasing camera-to-marker working distance generally reduced repeatability, especially for depth-related pose components.
    • This suggests working distances should be kept limited to improve the consistency of motion estimates.
  • Overall Optimal Configuration:
    • Moderately oblique viewing angles combined with limited working distance offer a good balance of accuracy and precision.
    • Ensuring sufficient image footprint (marker size and resolution in the camera image) is important for robust detection and pose estimation.

Implications and Future Directions

  • The study provides quantitative, pre-clinical recommendations for camera and ChArUco marker placement in standing equine CT environments.
  • Using these findings, researchers and clinicians can better design optical tracking setups to enable real-time, projection-wise motion compensation during scans.
  • Next steps include validating these optimized configurations in actual CT scanners with live equine subjects to confirm motion compensation effectiveness in vivo.
  • Improved motion tracking can lead to higher-quality CT images, aiding diagnosis and treatment planning in equine veterinary medicine.

Cite This Article

APA
Aliani C, Lorenzetto Bologna C, Francia P, Bocchi L. (2026). Optimising Camera-ChArUco Geometry for Motion Compensation in Standing Equine CT: A CT-Motivated Benchtop Study. Sensors (Basel), 26(4), 1310. https://doi.org/10.3390/s26041310

Publication

ISSN: 1424-8220
NlmUniqueID: 101204366
Country: Switzerland
Language: English
Volume: 26
Issue: 4
PII: 1310

Researcher Affiliations

Aliani, Cosimo
  • Department of Information Engineering, University of Florence, Via di Santa Marta 3, 50139 Florence, Italy.
Lorenzetto Bologna, Cosimo
  • Epica Imaginalis, Via Rodolfo Morandi 13/15, 50019 Florence, Italy.
Francia, Piergiorgio
  • Department of Information Engineering, University of Florence, Via di Santa Marta 3, 50139 Florence, Italy.
Bocchi, Leonardo
  • Department of Information Engineering, University of Florence, Via di Santa Marta 3, 50139 Florence, Italy.

MeSH Terms

  • Horses
  • Tomography, X-Ray Computed / methods
  • Animals
  • Image Processing, Computer-Assisted / methods
  • Motion
  • Algorithms

Conflict of Interest Statement

Author Cosimo Lorenzetto Bologna was employed by the company Epica Imaginalis. The remaining authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

References

This article includes 33 references
  1. Klette R. Concise Computer Vision. Volume 233 Springer; Berlin/Heidelberg, Germany: 2014.
  2. Fazakas T, Fekete RT. 3D reconstruction system for autonomous robot navigation. Proceedings of the 2010 11th International Symposium on Computational Intelligence and Informatics (CINTI), Budapest, Hungary, 18–20 November 2010; New York, NY, USA. IEEE; 2010. pp. 267–270.
  3. Tang Y, Chen M, Wang C, Luo L, Li J, Lian G, Zou X. Recognition and localization methods for vision-based fruit picking robots: A review. Front. Plant Sci. 2020;11:510.
    doi: 10.3389/fpls.2020.00510pmc: PMC7250149pubmed: 32508853google scholar: lookup
  4. Huang Y, Qiu Z, Song Z. 3D reconstruction and visualization from 2D CT images. Proceedings of the 2011 International Symposium on IT in Medicine and Education (ITME 2011), Guangzhou, China, 9–11 December 2011; New York, NY, USA. IEEE; 2011. pp. 153–157.
  5. Rengier F, Mehndiratta A, Von Tengg-Kobligk H, Zechmann CM, Unterhinninghofen R, Kauczor HU, Giesel FL. 3D printing based on imaging data: Review of medical applications. Int. J. Comput. Assist. Radiol. Surg. 2010;5:335–341.
    doi: 10.1007/s11548-010-0476-xpubmed: 20467825google scholar: lookup
  6. Ebner T, Feldmann I, Renault S, Schreer O, Eisert P. Multi-view reconstruction of dynamic real-world objects and their integration in augmented and virtual reality applications. J. Soc. Inf. Disp. 2017;25:151–157.
    doi: 10.1002/jsid.538google scholar: lookup
  7. Tang F, Wu Y, Hou X, Ling H. 3D mapping and 6D pose computation for real time augmented reality on cylindrical objects. IEEE Trans. Circuits Syst. Video Technol. 2019;30:2887–2899.
  8. Romero-Ramirez FJ, Muñoz-Salinas R, Medina-Carnicer R. Speeded up detection of squared fiducial markers. Image Vis. Comput. 2018;76:38–47.
  9. Garrido-Jurado S, Munoz-Salinas R, Madrid-Cuevas FJ, Medina-Carnicer R. Generation of fiducial marker dictionaries using mixed integer linear programming. Pattern Recognit. 2016;51:481–491.
  10. De la Escalera A, Armingol JM. Automatic chessboard detection for intrinsic and extrinsic camera parameter calibration. Sensors 2010;10:2027–2044.
    doi: 10.3390/s100302027pmc: PMC3264465pubmed: 22294912google scholar: lookup
  11. . ChArUco Marker Detection. [(accessed on 8 February 2026)]. Available online: https://docs.opencv.org/3.4/df/d4a/tutorial_charuco_detection.html.
  12. Dakin S, Lam R, Rees E, Mumby C, West C, Weller R. Technical set-up and radiation exposure for standing computed tomography of the equine head. Equine Vet. Educ. 2014;26:208–215.
    doi: 10.1111/eve.12127google scholar: lookup
  13. Davies T, Skelly C, Puggioni A, D’Helft C, Connolly S, Hoey S. Standing CT of the equine head: Reducing radiation dose maintains image quality. Vet. Radiol. Ultrasound. 2020;61:137–146.
    doi: 10.1111/vru.12823pubmed: 31823453google scholar: lookup
  14. Mathee N, Robert M, Higgerty SM, Fosgate GT, Rogers AL, d’Ablon X, Carstens A. Computed tomographic evaluation of the distal limb in the standing sedated horse: Technique, imaging diagnoses, feasibility, and artifacts. Vet. Radiol. Ultrasound. 2023;64:243–252.
    doi: 10.1111/vru.13182pubmed: 36373276google scholar: lookup
  15. Pauwels FE, Van der Vekens E, Christan Y, Koch C, Schweizer D. Feasibility, indications, and radiographically confirmed diagnoses of standing extremity cone beam computed tomography in the horse. Vet. Surg. 2021;50:365–374.
    doi: 10.1111/vsu.13560pubmed: 33421172google scholar: lookup
  16. Gaida JL, Steinberg T, Stieger-Vanegas SM, Merle R, Lischer CJ. Equine Standing Multidetector Computed Tomography of the Distal Thoracic Limb and Tarsus Has a Lower Cumulative Radiation Dose than Digital Radiography. Vet. Radiol. Ultrasound 2025;66:e70049.
    doi: 10.1111/vru.70049pmc: PMC12106850pubmed: 40420323google scholar: lookup
  17. Kim J, Nuyts J, Kyme A, Kuncic Z, Fulton R. A rigid motion correction method for helical computed tomography (CT). Phys. Med. Biol. 2015;60:2047.
    doi: 10.1088/0031-9155/60/5/2047pubmed: 25674780google scholar: lookup
  18. Brombal L, Arana Peña LM, Arfelli F, Longo R, Brun F, Contillo A, Di Lillo F, Tromba G, Di Trapani V, Donato S. Motion artifacts assessment and correction using optical tracking in synchrotron radiation breast CT. Med Phys 2021;48:5343–5355.
    doi: 10.1002/mp.15084pmc: PMC9291820pubmed: 34252212google scholar: lookup
  19. Maier J, Nitschke M, Choi JH, Gold G, Fahrig R, Eskofier BM, Maier A. Rigid and non-rigid motion compensation in weight-bearing CBCT of the knee using simulated inertial measurements. IEEE Trans. Biomed. Eng. 2021;69:1608–1619.
    doi: 10.1109/TBME.2021.3123673pmc: PMC9134858pubmed: 34714730google scholar: lookup
  20. Siena FL, Byrom B, Watts P, Breedon P. Utilising the intel realsense camera for measuring health outcomes in clinical research. J. Med Syst. 2018;42:53.
    doi: 10.1007/s10916-018-0905-xpmc: PMC5799357pubmed: 29404692google scholar: lookup
  21. House R, Lasso A, Harish V, Baum Z, Fichtinger G. Evaluation of the Intel RealSense SR300 camera for image-guided interventions and application in vertebral level localization. Proceedings of the Medical Imaging 2017: Image-Guided Procedures, Robotic Interventions, and Modeling, SPIE, Orlando, FL, USA, 14–16 February 2017; Washington, DC, USA. SPIE; 2017. pp. 771–777.
  22. Moghari MD, Noonan P, Henry D, Fulton RR, Young N, Moore K, Evanns A, Kyme A. Characterisation of the Intel RealSense D415 Stereo Depth Camera for Motion-Corrected CT Perfusion Imaging. arXiv 20242403.16490.
  23. Thies M, Wagner F, Maul N, Yu H, Goldmann M, Schneider LS, Gu M, Mei S, Folle L, Preuhs A. A gradient-based approach to fast and accurate head motion compensation in cone-beam CT. IEEE Trans. Med Imaging 2024;44:1098–1109.
    doi: 10.1109/TMI.2024.3474250pubmed: 39365718google scholar: lookup
  24. Lu A, Huang H, Hu Y, Zbijewski W, Unberath M, Siewerdsen JH, Weiss CR, Sisniega A. Vessel-targeted compensation of deformable motion in interventional cone-beam CT. Med Image Anal. 2024;97:103254.
  25. Maier J, Sawall S, Arheit M, Paysan P, Kachelrieß M. Deep learning-based cone-beam CT motion compensation with single-view temporal resolution. Med. Phys. 2025;52:e17911.
    doi: 10.1002/mp.17911pmc: PMC12258004pubmed: 40467957google scholar: lookup
  26. Berral-Soler R, Muñoz-Salinas R, Medina-Carnicer R, Marín-Jiménez MJ. DeepArUco++: Improved detection of square fiducial markers in challenging lighting conditions. Image Vis. Comput. 2024;152:105313.
  27. Kim D, Bong JH, Jeong S. Enhancing pose estimation using multiple graphical markers with spatial and temporal outlier detection. Appl. Sci. 2024;14:10225.
    doi: 10.3390/app142210225google scholar: lookup
  28. Krumstroh F, Eisenmann A, Franssen J, Rindermann M, Pfeiffer T. Proceedings of the GI Workshop 2025 on Virtual and Augmented Reality. Gesellschaft für Informatik; Bonn, Germany: 2025. Comparing Fiducial Marker Tracking Across Cameras in Virtual and Physical Environments; pp. 10–18420.
  29. de Medeiros Esper I, Smolkin O, Manko M, Popov A, From PJ, Mason A. Evaluation of RGB-D Multi-Camera Pose Estimation for 3D Reconstruction. Appl. Sci. 2022;12:4134.
    doi: 10.3390/app12094134google scholar: lookup
  30. An GH, Lee S, Seo MW, Yun K, Cheong WS, Kang SJ. Charuco Board-Based Omnidirectional Camera Calibration Method. Electronics 2018;7:421.
  31. Hu D, DeTone D, Malisiewicz T. Deep charuco: Dark charuco marker pose estimation. Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; New York, NY, USA. IEEE; 2019. pp. 8436–8444.
  32. . OpenCV: Open Source Computer Vision Library. [(accessed on 8 February 2026)]. Available online: https://opencv.org.
  33. García-Ruiz P, Romero-Ramirez FJ, Muñoz-Salinas R, Marín-Jiménez MJ, Medina-Carnicer R. Sparse Indoor Camera Positioning with Fiducial Markers. Appl. Sci. 2025;15:1855.
    doi: 10.3390/app15041855pmc: PMC11244017pubmed: 39001083google scholar: lookup

Citations

This article has been cited 0 times.