TY - JOUR
T1 - Multiview Summarization and Activity Recognition Meet Edge Computing in IoT Environments
AU - Hussain, Tanveer
AU - Muhammad, Khan
AU - Ullah, Amin
AU - Ser, Javier Del
AU - Gandomi, Amir H.
AU - Sajjad, Muhammad
AU - Baik, Sung Wook
AU - De Albuquerque, Victor Hugo C.
N1 - Publisher Copyright:
© 2014 IEEE.
PY - 2021/6/15
Y1 - 2021/6/15
N2 - Multiview video summarization (MVS) has not received much attention from the research community due to inter-view correlations and views' overlapping, etc. The majority of previous MVS works are offline, relying on only summary, and require additional communication bandwidth and transmission time, with no focus on foggy environments. We propose an edge intelligence-based MVS and activity recognition framework that combines artificial intelligence with Internet of Things (IoT) devices. In our framework, resource-constrained devices with cameras use a lightweight CNN-based object detection model to segment multiview videos into shots, followed by mutual information computation that helps in a summary generation. Our system does not rely solely on a summary, but encodes and transmits it to a master device using a neural computing stick for inter-view correlations computation and efficient activity recognition, an approach which saves computation resources, communication bandwidth, and transmission time. Experiments show an increase of 0.4 unit in F -measure on an MVS Office dateset and 0.2% and 2% improved accuracy for UCF-50 and YouTube 11 datesets, respectively, with lower storage and transmission times. The processing time is reduced from 1.23 to 0.45 s for a single frame and optimally 0.75 seconds faster MVS. A new dateset is constructed by synthetically adding fog to an MVS dateset to show the adaptability of our system for both certain and uncertain IoT surveillance environments.
AB - Multiview video summarization (MVS) has not received much attention from the research community due to inter-view correlations and views' overlapping, etc. The majority of previous MVS works are offline, relying on only summary, and require additional communication bandwidth and transmission time, with no focus on foggy environments. We propose an edge intelligence-based MVS and activity recognition framework that combines artificial intelligence with Internet of Things (IoT) devices. In our framework, resource-constrained devices with cameras use a lightweight CNN-based object detection model to segment multiview videos into shots, followed by mutual information computation that helps in a summary generation. Our system does not rely solely on a summary, but encodes and transmits it to a master device using a neural computing stick for inter-view correlations computation and efficient activity recognition, an approach which saves computation resources, communication bandwidth, and transmission time. Experiments show an increase of 0.4 unit in F -measure on an MVS Office dateset and 0.2% and 2% improved accuracy for UCF-50 and YouTube 11 datesets, respectively, with lower storage and transmission times. The processing time is reduced from 1.23 to 0.45 s for a single frame and optimally 0.75 seconds faster MVS. A new dateset is constructed by synthetically adding fog to an MVS dateset to show the adaptability of our system for both certain and uncertain IoT surveillance environments.
KW - Activity recognition
KW - Internet of Things (IoT)
KW - deep autoencoder
KW - deep learning
KW - multiview video summarization (MVS)
KW - sequential learning
KW - video data analytics
KW - video summarization
UR - http://www.scopus.com/inward/record.url?scp=85107497489&partnerID=8YFLogxK
U2 - 10.1109/JIOT.2020.3027483
DO - 10.1109/JIOT.2020.3027483
M3 - Article
AN - SCOPUS:85107497489
SN - 2327-4662
VL - 8
SP - 9634
EP - 9644
JO - IEEE Internet of Things Journal
JF - IEEE Internet of Things Journal
IS - 12
M1 - 9208765
ER -