RESULTS: At present, the classifier used has achieved an accuracy of 100% based on skulls' views. Classification and identification to regions and sexes have also attained 72.5%, 87.5% and 80.0% of accuracy for dorsal, lateral, and jaw views, respectively. This results show that the shape characteristic features used are substantial because they can differentiate the specimens based on regions and sexes up to the accuracy of 80% and above. Finally, an application was developed and can be used for the scientific community.
CONCLUSIONS: This automated system demonstrates the practicability of using computer-assisted systems in providing interesting alternative approach for quick and easy identification of unknown species.
METHODS: In this study a novel system named Ceph-X is developed to computerize the manual tasks of orthodontics during cephalometric measurements. Ceph-X is developed by using image processing techniques with three main models: enhancements X-ray image model, locating landmark model, and computation model. Ceph-X was then evaluated by using X-ray images of 30 subjects (male and female) obtained from University of Malaya hospital. Three orthodontics specialists were involved in the evaluation of accuracy to avoid intra examiner error, and performance for Ceph-X, and 20 orthodontics specialists were involved in the evaluation of the usability, and user satisfaction for Ceph-X by using the SUS approach.
RESULTS: Statistical analysis for the comparison between the manual and automatic cephalometric approaches showed that Ceph-X achieved a great accuracy approximately 96.6%, with an acceptable errors variation approximately less than 0.5 mm, and 1°. Results showed that Ceph-X increased the specialist performance, and minimized the processing time to obtain cephalometric measurements of human skull. Furthermore, SUS analysis approach showed that Ceph-X has an excellent usability user's feedback.
CONCLUSIONS: The Ceph-X has proved its reliability, performance, and usability to be used by orthodontists for the analysis, diagnosis, and treatment of cephalometric.
RESULT: Images of four monogenean species namely Sinodiplectanotrema malayanus, Trianchoratus pahangensis, Metahaliotrema mizellei and Metahaliotrema sp. (undescribed) were used to develop an automated technique for identification. K-nearest neighbour (KNN) was applied to classify the monogenean specimens based on the extracted features. 50% of the dataset was used for training and the other 50% was used as testing for system evaluation. Our approach demonstrated overall classification accuracy of 90%. In this study Leave One Out (LOO) cross validation is used for validation of our system and the accuracy is 91.25%.
CONCLUSIONS: The methods presented in this study facilitate fast and accurate fully automated classification of monogeneans at the species level. In future studies more classes will be included in the model, the time to capture the monogenean images will be reduced and improvements in extraction and selection of features will be implemented.