METHODS: In this study a novel system named Ceph-X is developed to computerize the manual tasks of orthodontics during cephalometric measurements. Ceph-X is developed by using image processing techniques with three main models: enhancements X-ray image model, locating landmark model, and computation model. Ceph-X was then evaluated by using X-ray images of 30 subjects (male and female) obtained from University of Malaya hospital. Three orthodontics specialists were involved in the evaluation of accuracy to avoid intra examiner error, and performance for Ceph-X, and 20 orthodontics specialists were involved in the evaluation of the usability, and user satisfaction for Ceph-X by using the SUS approach.
RESULTS: Statistical analysis for the comparison between the manual and automatic cephalometric approaches showed that Ceph-X achieved a great accuracy approximately 96.6%, with an acceptable errors variation approximately less than 0.5 mm, and 1°. Results showed that Ceph-X increased the specialist performance, and minimized the processing time to obtain cephalometric measurements of human skull. Furthermore, SUS analysis approach showed that Ceph-X has an excellent usability user's feedback.
CONCLUSIONS: The Ceph-X has proved its reliability, performance, and usability to be used by orthodontists for the analysis, diagnosis, and treatment of cephalometric.
OBJECTIVES: The main objective of the research is to develop a robust and high-performance human action recognition techniques. A combination of local and holistic feature extraction methods used through analyzing the most effective features to extract to reach the objective, followed by using simple and high-performance machine learning algorithms.
METHODS: This paper presents three robust action recognition techniques based on a series of image analysis methods to detect activities in different scenes. The general scheme architecture consists of shot boundary detection, shot frame rate re-sampling, and compact feature vector extraction. This process is achieved by emphasizing variations and extracting strong patterns in feature vectors before classification.
RESULTS: The proposed schemes are tested on datasets with cluttered backgrounds, low- or high-resolution videos, different viewpoints, and different camera motion conditions, namely, the Hollywood-2, KTH, UCF11 (YouTube actions), and Weizmann datasets. The proposed schemes resulted in highly accurate video analysis results compared to those of other works based on four widely used datasets. The First, Second, and Third Schemes provides recognition accuracies of 57.8%, 73.6%, and 52.0% on Hollywood2, 94.5%, 97.0%, and 59.3% on KTH, 94.5%, 95.6%, and 94.2% on UCF11, and 98.9%, 97.8% and 100% on Weizmann.
CONCLUSION: Each of the proposed schemes provides high recognition accuracy compared to other state-of-art methods. Especially, the Second Scheme as it gives excellent comparable results to other benchmarked approaches.