Browse Articles

Identifying Neural Networks that Implement a Simple Spatial Concept

Zirvi et al. | Sep 13, 2022

Identifying Neural Networks that Implement a Simple Spatial Concept

Modern artificial neural networks have been remarkably successful in various applications, from speech recognition to computer vision. However, it remains less clear whether they can implement abstract concepts, which are essential to generalization and understanding. To address this problem, the authors investigated the above vs. below task, a simple concept-based task that honeybees can solve, using a conventional neural network. They found that networks achieved 100% test accuracy when a visual target was presented below a black bar, however only 50% test accuracy when a visual target was presented below a reference shape.

Read More...

The Effect of Varying Training on Neural Network Weights and Visualizations

Fountain et al. | Dec 04, 2019

The Effect of Varying Training on Neural Network Weights and Visualizations

Neural networks are used throughout modern society to solve many problems commonly thought of as impossible for computers. Fountain and Rasmus designed a convolutional neural network and ran it with varying levels of training to see if consistent, accurate, and precise changes or patterns could be observed. They found that training introduced and strengthened patterns in the weights and visualizations, the patterns observed may not be consistent between all neural networks.

Read More...

Part of speech distributions for Grimm versus artificially generated fairy tales

Arvind et al. | Nov 16, 2024

Part of speech distributions for Grimm versus artificially generated fairy tales
Image credit: Nayalia Y.

Here, the authors wanted to explore mathematical paradoxes in which there are multiple contradictory interpretations or analyses for a problem. They used ChatGPT to generate a novel dataset of fairy tales. They found statistical differences between the artificially generated text and human produced text based on the distribution of parts of speech elements.

Read More...

Tomato disease identification with shallow convolutional neural networks

Trinh et al. | Mar 03, 2023

Tomato disease identification with shallow convolutional neural networks

Plant diseases can cause up to 50% crop yield loss for the popular tomato plant. A mobile device-based method to identify diseases from photos of symptomatic leaves via computer vision can be more effective due to its convenience and accessibility. To enable a practical mobile solution, a “shallow” convolutional neural networks (CNNs) with few layers, and thus low computational requirement but with high accuracy similar to the deep CNNs is needed. In this work, we explored if such a model was possible.

Read More...

Collaboration beats heterogeneity: Improving federated learning-based waste classification

Chong et al. | Jul 18, 2023

Collaboration beats heterogeneity: Improving federated learning-based waste classification

Based on the success of deep learning, recent works have attempted to develop a waste classification model using deep neural networks. This work presents federated learning (FL) for a solution, as it allows participants to aid in training the model using their own data. Results showed that with less clients, having a higher participation ratio resulted in less accuracy degradation by the data heterogeneity.

Read More...

Large Language Models are Good Translators

Zeng et al. | Oct 16, 2024

Large Language Models are Good Translators

Machine translation remains a challenging area in artificial intelligence, with neural machine translation (NMT) making significant strides over the past decade but still facing hurdles, particularly in translation quality due to the reliance on expensive bilingual training data. This study explores whether large language models (LLMs), like GPT-4, can be effectively adapted for translation tasks and outperform traditional NMT systems.

Read More...

Rhythmic lyrics translation: Customizing a pre-trained language model using stacked fine-tuning

Chong et al. | May 01, 2023

Rhythmic lyrics translation: Customizing a pre-trained language model using stacked fine-tuning
Image credit: Pixabay

Neural machine translation (NMT) is a software that uses neural network techniques to translate text from one language to another. However, one of the most famous NMT models—Google Translate—failed to give an accurate English translation of a famous Korean nursery rhyme, "Airplane" (비행기). The authors fine-tuned a pre-trained model first with a dataset from the lyrics domain, and then with a smaller dataset containing the rhythmical properties, to teach the model to translate rhythmically accurate lyrics. This stacked fine-tuning method resulted in an NMT model that could maintain the rhythmical characteristics of lyrics during translation while single fine-tuned models failed to do so.

Read More...