Wiki Contents

About the Computer Vision (CV) Division


The Computer Vision (CV) division of the KDD Lab focuses on learning for vision tasks in the areas of object detection and classification, action classification, and scene understanding. Topic emphases include learning for organism identification in the wild (by taxa and individual identity), few-shot learning (FSL) and zero-shot learning (ZSL), inter-class transfer learning, meta-learning, and anomaly detection.

The division is currently directed by Robert Stewart (spring 2022 - present). The previous division lead was Dr. Ademola Okerinde (fall 2021 - spring 2022) and the inaugural division lead (spring 2017 - summer 2021) was Dr. Ray Luo.

Project Description

This is an umbrella project for all deep learning-based research with special emphasis on vision centric and spatiotemporal AI applications. The current sub-projects are listed as follows.

Classification of Bee Species

This project is a collaboration with the Entomology Department at K-State which deals with automated identification of bee species from images. Our current focus is on Bumble bees. We have an in house dataset of 89,000 images of bumble bees, representing 36 species in North America. We have compared different convolutional neural network based image classification models. In the future, we are looking forward to improving the accuracy of bumble bee classification as well as adding more species of bees. You can try the BeeMachine web app.

Automated Detection of Risky Tackle from Sports Video

The aim of this project is to classify safe and risky tackles from videos of American football practice. We have collected and annotated a data set of 108 videos by collaborating with experts from the Department of Kinesiology. The label 0 indicates the tackle is safe and 1 indicates the tackle is risky. Currently, we have 32 tackle videos marked as risky and 76 videos marked as safe. Our initial target is to frame the detectable region in space and time (spatiotemporal event detection), relate the "bounding box in space and time" to automatically-generated annotations, and link qualitative annotations (attention maps) to quantitative ones (analytic results such as estimated contact surface, force, angles). Also, we are trying to collect more data with a clean background.

Real-Time Cattle Face Recognition

In this ongoing study, we aim to identify the faces of livestock cattle by using deep learning to free farmers from using radio frequency identification (RFID) ear tags to identify cattle, a practice that is expensive due to sensor, ear tag, and installation labor costs.

Autonomous Agents for Precision Agriculture

This project deals with using deep reinforcement learning (DRL) for policy learning in spatiotemporal domains such as precision agriculture tasks (pesticide and fertilizer application, image gathering) and autonomous navigation. An important part of this research is its focus on multi-agent DRL (MADRL). In recent years, single agent DRL algorithms have been shown to effectively learn desired policy and exhibit super-human performance. However, practical applications of DRL naturally involve more than one agent; for example, a multi-robot precision agriculture task (involving multiple autonomous agents) such as pesticide spray in a large field or removal of harmful invasive species from farmland, requires multiple robots which must learn to collaborate amongst each other and solve the challenge jointly, including path planning, navigation and task distribution. MADRL is also necessitated by the fact that the role played by each agent may be different and hence partitioning of the problem according to these roles can improve learning efficiency.

Automated Inspection & Engineering Anomaly Detection

This project is part of a spatiotemporal anomaly detection project that combines image processing, computer vision techniques for object detection and pattern recognition, geographic information systems (GIS) and aerial visual survey. Applications include analyzing unmanned aerial systems (UAS) data to identify and geotag systems such as streetlights and insulators that are inoperative. Research in machine learning and data mining focuses on a selection of relevant visual features and training data for capturing and labeling objects from flyover images. In this study, a deep learning based system will be developed which will automatically recognize damaged electrical insulators from images taken by a drone. The project will provide the development support needed to prototype and build a complete object detection and classification application. The system has several key features:

  1. Ability to locate insulators in images with a high success rate.
  2. Ability to identify damaged insulators with over 85% success rate.
  3. Design a web application interface that allows easy use of the system.
  4. Several ongoing research efforts focus on utilizing different types of generative models to combat class imbalance issues.

Cancer Imaging & Medical Anomaly Detection

This project deals with using generative adversarial network (GAN) models for data augmentation via texture transfer and balancing. It is part of a broader project in medical informatics that seeks to apply self-supervised learning and differentiable computing (including deep neural networks) to open problems in medical imaging and image-based diagnosis.

Animal Behavior from Video

The goal of this continuing research project, conducted in collaboration with the KSU Department of Psychological Sciences under a Centers of Biomedical Research Excellence (COBRE) Phase II grant from the National Institutes of Health (NIH), is to classify the behavior of animals in videos, a process also known as grading. In computer vision, the general classification task, including the detection of organism(s) involved, is known as activity recognition.


Back to KSU KDD Lab research overview

Last updated by rotclanny on Jul 28, 2023