To efficiently teach novices skilled tasks, it is necessary to analyze the difference between a novice and an expert worker. Accordingly, a method for extracting differences (on the basis of skill level) in motions of workers performing tasks is proposed. As for this method, a network (multi-stream LSTM) that estimates skill level from 3D positional information of the worker’s visual point and joints is trained, and the internal structure of the network is then analyzed. The results of an experiment indicate that a particular motion, namely, “grasping an object,” becomes different when the worker becomes skilled; in particular, the worker grasps the object without moving their visual point to the position of the part, namely, without looking at the object, and uses both hands efficiently.
When inspecting defects such as scratches due to image processing, if we can obtain an image before defect occurrence, the defect can be detected by simply comparing the image pair before and after defect occurrence. However, this idea is generally unrealistic because it is impossible to obtain an image before defect occurrence from an image after the defect occurred unless we go back in time. Therefore, we propose a method of training a generation-based model to detect scratches based on subtraction between an input image and output image during testing. We obtained a large amount of ideal image pairs before and after defect occurrence (i.e., image pairs in which only defect regions are different and the others are almost completely the same) using an image-capturing device. An image with a background texture of high reconstruction performance was generated with our method by using this dataset. Although we used image pairs of only an aluminum plate for training, the area under the curve of the receiver operating characteristic (ROC-AUC) measure was 0.9973 for a copper plate and 0.9904 for a stainless steel plate. This shows that our method is also effective for background textures different from those used for training. We also compared our generation-based method with a method of training a classificationbased model (ResNet18). We confirmed that the ROC-AUC measure with the proposed method was 0.9824 for an acrylic plate and 0.6156 for a carbon fiber sheet, which is better than those with the classification-based method (0.9650 and 0.5000, respectively). The proposed method has much higher reconstruction performance than previous training of defect-free images only, and is effective for scratch detection of background texture similar to those used for training.
In this paper, we propose using image recognition techniques to estimate the “understanding measure” in person-toperson teaching situations. The phrase “understanding measure” refers to how strongly a teacher feels a student understands a topic. First, we extract a student’s nonverbal behavior (head movement, gazes, and blinking) as the features for the estimation process. Next, we calculate the subspace from the aforementioned feature by using principal component analysis (PCA) and linear discriminant analysis (LDA). Finally, we classify unknown data as either “understood” or “did not understand” by using a kNN classifier in subspace. Our experiments confirmed that the Fmeasure of the classification “understood” by our method was 0.75 and “did not understand” was 0.60, indicating that our method improved F-measures 0.38 and 0.11, respectively, compared with previous methods.
Access to the requested content is limited to institutions that have purchased or subscribe to SPIE eBooks.
You are receiving this notice because your organization may not have SPIE eBooks access.*
*Shibboleth/Open Athens users─please
sign in
to access your institution's subscriptions.
To obtain this item, you may purchase the complete book in print or electronic format on
SPIE.org.
INSTITUTIONAL Select your institution to access the SPIE Digital Library.
PERSONAL Sign in with your SPIE account to access your personal subscriptions or to use specific features such as save to my library, sign up for alerts, save searches, etc.