Auflistung nach Autor:in "Schmitt, Thomas H."
1 - 3 von 3
Treffer pro Seite
Sortieroptionen
- KonferenzbeitragMachine Learning in Glass Bottle Printing Quality Control: A Collaboration with a Medium-Sized Industrial Partner(INFORMATIK 2024, 2024) Bundscherer, Maximilian; Schmitt, Thomas H.; Bocklet, TobiasIn cooperation with a medium-sized industrial partner, we developed and evaluated two ML-based approaches for quality control in glass bottle printing. Our first approach utilized various filters to suppress reflections, image quality metrics for image comparison, and supervised classification models, resulting in an accuracy of 84%. We used the ORB algorithm for image alignment and to estimate print rotations, which may indicate manufacturing anomalies. In our second approach, we fine-tuned pre-trained CNN models, which resulted in an accuracy of 87%. Utilizing Grad-CAM, an Explainable AI method, we localized and visualized frequently defective bottle print regions without explicitly training our models for this use case. These insights can be used to optimize the actual manufacturing process beyond classification. This paper also describes our general approach and the challenges we encountered in practice with data collection during ongoing production, unsupervised preselection, and labeling.
- KonferenzbeitragSegmenting Wood Rot using Computer Vision Models(INFORMATIK 2024, 2024) Kammerbauer, Roland; Schmitt, Thomas H.; Bocklet, TobiasIn the woodworking industry, a huge amount of effort has to be invested into the initial quality assessment of the raw material. In this study, we present an AI model to detect, quantify, and localize defects on wooden logs. This model aims to both automate the quality control process and provide a more consistent and reliable quality assessment. For this purpose, a dataset of 1424 sample images of wood logs is created. A total of 5 annotators possessing different levels of expertise are involved in dataset creation. An inter-annotator agreement analysis is conducted to analyze the impact of expertise on the annotation task and to highlight subjective differences in annotator judgment. We explore, train, and fine-tune the state-of-the-art InternImage and ONE-PEACE architectures for semantic segmentation. The best model created achieves an average IoU of 0.71 and shows detection and quantification capabilities close to the human annotators.
- KonferenzbeitragTraining a Computer Vision Model for Commercial Bakeries with Primarily Synthetic Images(INFORMATIK 2024, 2024) Schmitt, Thomas H.; Bundscherer, Maximilian; Bocklet, TobiasIn the food industry, reprocessing returned products is a vital step to increase resource efficiency. [SBB23] presented an AI application that automates the tracking of returned bread buns. We extend their work by creating an expanded dataset comprising 2432 images and a wider range of baked goods. To increase model robustness, we use generative models pix2pix and CycleGAN to create synthetic images. We train state-of-the-art object detection models YOLOv9 and YOLOv8 on our detection task. Our overall best-performing model achieved an average precision AP 0.5 of 90.3% on our test set.