Humans have a natural ability to recognize emotional cues from the facial expressions of others, as a crucial evolutionary trait to navigate social interactions. This ability likely develops through normal development and social experience, but it is unclear how much influence age and sex have in emotional facial recognition (EFR). In this study, the authors investigate EFR in children and teenagers, and look at whether accurate emotional recognition does occur more in males or females.
In this study the authors develop an app for faster chess game entry method to help chess learners improve their game. This culminated in the Augmented Reality Chess Analyzer (ARChessAnalyzer) which uses traditional image and vision techniques for chess board recognition and Convolutional Neural Networks (CNN) for chess piece recognition.
The application of machine learning techniques has facilitated the automatic annotation of behavior in video sequences, offering a promising approach for ethological studies by reducing the manual effort required for annotating each video frame. Nevertheless, before solely relying on machine-generated annotations, it is essential to evaluate the accuracy of these annotations to ensure their reliability and applicability. While it is conventionally accepted that there cannot be a perfect annotation, the degree of error associated with machine-generated annotations should be commensurate with the error between different human annotators. We hypothesized that machine learning supervised with adequate human annotations would be able to accurately predict body parts from video sequences. Here, we conducted a comparative analysis of the quality of annotations generated by humans and machines for the body parts of sheep during treadmill walking. For human annotation, two annotators manually labeled six body parts of sheep in 300 frames. To generate machine annotations, we employed the state-of-the-art pose-estimating library, DeepLabCut, which was trained using the frames annotated by human annotators. As expected, the human annotations demonstrated high consistency between annotators. Notably, the machine learning algorithm also generated accurate predictions, with errors comparable to those between humans. We also observed that abnormal annotations with a high error could be revised by introducing Kalman Filtering, which interpolates the trajectory of body parts over the time series, enhancing robustness. Our results suggest that conventional transfer learning methods can generate behavior annotations as accurate as those made by humans, presenting great potential for further research.
Using facial recognition as a use-case scenario, we attempt to identify sources of bias in a model developed using transfer learning. To achieve this task, we developed a model based on a pre-trained facial recognition model, and scrutinized the accuracy of the model’s image classification against factors such as age, gender, and race to observe whether or not the model performed better on some demographic groups than others. By identifying the bias and finding potential sources of bias, his work contributes a unique technical perspective from the view of a small scale developer to emerging discussions of accountability and transparency in AI.
Modern artificial neural networks have been remarkably successful in various applications, from speech recognition to computer vision. However, it remains less clear whether they can implement abstract concepts, which are essential to generalization and understanding. To address this problem, the authors investigated the above vs. below task, a simple concept-based task that honeybees can solve, using a conventional neural network. They found that networks achieved 100% test accuracy when a visual target was presented below a black bar, however only 50% test accuracy when a visual target was presented below a reference shape.
With the advance of technology, artificial intelligence (AI) is now applied widely in society. In the study of AI, machine learning (ML) is a subfield in which a machine learns to be better at performing certain tasks through experience. This work focuses on the convolutional neural network (CNN), a framework of ML, applied to an image classification task. Specifically, we analyzed the performance of the CNN as the type of neural activation function changes.
The purpose of the study was to determine whether graph-based machine learning techniques, which have increased prevalence in the last few years, can accurately classify data into one of many clusters, while requiring less labeled training data and parameter tuning as opposed to traditional machine learning algorithms. The results determined that the accuracy of graph-based and traditional classification algorithms depends directly upon the number of features of each dataset, the number of classes in each dataset, and the amount of labeled training data used.
Here recognizing the growing amount of plastic waste in the oceans, the authors sought to develop and test laser imaging for the identification of waste in water. They found that while possible, limitations such as increasing depth and water turbidity result in increasing blurriness in laser images. While their image processing methods were somewhat insufficient they identified recent methods to use deep learning-based techniques as a potential avenue to viability for this method.