Academic Positions

  • Present 2020
    image

    Assistant Professor

    University of Western Macedonia, Department of Communication and Digital Media

  • 2020 2019
    image

    Research Scientist

    University of Ioannina, Department of Computer Science & Engineering

  • 2019 2018
    image

    Adjunct Lecturer

    University of Ioannina, Department of Computer Science & Engineering

  • 2018 2016
    image

    Postdoctoral Fellow

    University of Houston, Computational Biomedicine Lab

Education

  • Ph.D. 2016
    image

    Ph.D. in Computer Science

    University of Ioannina

    Department of Computer Science and Engineering

  • M.Sc.2010
    image

    Master in Computer Science

    University of Ioannina

    Department of Computer Science

  • B.Sc.2008
    image

    Bachelor in Computer Science

    University of Ioannina

    Department of Computer Science


Honors and Awards

  • January 2024
    Excellence in Research Award
    image
    2nd prize of excellence in research of the Faculty of Social Sciences and Humanities, University of Western Macedonia. For the research work entitled “Facemask: A new image dataset for the automated identification of people wearing masks in the wild” published in the International Journal: Sensors.
  • January 2023
    Excellence in Research Award
    image
    2nd prize of excellence in research of the Faculty of Social Sciences and Humanities, University of Western Macedonia. For the research work entitled “Human activity recognition using robust adaptive privileged probabilistic learning” published in the International Journal: Pattern Analysis and Applications.
  • July 2018
    NVIDIA GPU Grant
    image
    NVIDIA Corporation supports my research with the donation of one Titan Xp GPU.
  • March 2017
    Outstanding Reviewer
    image
    Outstanding reviewer for the sicentific Journal Signal Processing: Image Communication.
  • October 2016
    Outstanding Reviewer
    image
    Outstanding reviewer for the sicentific Journal Pervasive and Mobile Computing.
  • June 2016
    Honorable Mention Paper Award
    image
    I was the recipient of the best paper award for the paper entitled “Exploiting privileged information for facial expression recognition”. One of two papers selected among 151 submitted and 52 accepted papers at IAPR/IEEE International Conference on Biometrics (ICB).
  • 2003
    Excellence Award
    image
    Excellence award from the Ministry of Education of Greece.

Currrent Teaching

  • Undergraduate Courses
  • Fall 2023
    CDM3114: Graphics Design and Visual Communication

    Instructor  For more information see the course web page here.

  • Fall 2023
    CMD5134: Interactive Multimedia

    Instructor  For more information see the course web page here.

  • Fall 2023
    CDM5295: Theory and Design of 3D Graphics for Virtual Enviroments

    Instructor  For more information see the course web page here.

  • Spring 2024
    CDM6115: Theory and Design of Animation

    Instructor  For more information see the course web page here.

  • Spring 2024
    CDM8245: Theory and Design of Virtual and Augmented Reality Applications

    Instructor  For more information see the course web page here.


  • Fall 2023
    Digital Multimedia Processing

    Instructor - Co-teaching  For more information see the course web page here.

  • Spring 2024
    Augmented and Virtual Reality

    Instructor - Co-teaching  For more information see the course web page here.


  • Fall 2023
    Didactic Utilization of ICT in Education

    Instructor - Co-teaching  For more information see the course web page here.

Teaching History

  • 2023 2020
    CDM4255: Mobile Digital Media and Diffusible Computation

    Instructor

  • 2021 2020
    CDM4145: Creative Studio and Audiovisual Productions

    Instructor

  • 2021 2020
    DMC563: 3D Digital Applications

    Instructor

  • 2021 2020
    DMC638: Interactive Communication

    Instructor

  • 2022 2020
    DMC737: Integrated Multimedia Creation

    Instructor

  • 2020 2019
    DMC425: Creative Animation

    Instructor

  • 2020 2019
    DMC628: Public Relations and Marketing on the Internet

    Instructor - Co-teaching


  • 2019 2018

    Object Oriented Programming Techniques

    Instructor

  • 2019 2018

    Digital Image Processing

    Instructor

Filter by type:

  • 2024
    (To appear)

    Accurate cell segmentation based on generative adversarial networks and nuclei guide factors

    K. Lavntaniti, M.E. Plissiti, M. Vrigkas, C. Nikou
    Conference Paper Proc. 21st IEEE International Symposium on Biomedical Imaging (ISBI), Athens Greece, May 27-30, 2024

Abstract

The accurate segmentation of cells in cervical images is crucial for the recognition of pathological situations and the estimation of their severity. In this work, we investigate the segmentation of both the nucleus and the cytoplasm of each cell based on two Generative Adversarial Networks (GANs). First, we detect the location of the nucleus with the extraction of the nucleus boundaries in each cell, which is obtained by the training of the Nucleus-GAN. The segmented nucleus area serves as a guide factor for the definition of the cell boundary, and it is used as input in the Cell-GAN, for the segmentation of the cell boundaries. As it is verified by the experimental results, the proposed method is efficient and leads to accurate nucleus and cell boundaries, presenting high performance.


BibTex

@inproceedings{Lavntaniti_etal_ISBI2024,
author    = {Kostantsa Lavntaniti and Marina E. Plissiti and Michalis Vrigkas and Christophoros Nikou},
title     = {Accurate cell segmentation based on generative adversarial networks and nuclei guide factors},
booktitle = {Proc. 21st IEEE International Symposium on Biomedical Imaging (ISBI)},
address   = {Athens, Greese},
pages     = {},
month     = {May},
year      = {2024}
}
  • 2023

    Creating location-based augmented reality games and immersive experiences for touristic Ddstination marketing and education

    A. Kleftodimos, A. Evagelou, S. Gkoutzios, M. Matsiola, M. Vrigkas, A. Yannacopoulou, A. Triantafillidou, G. Lappas
    Journal PaperComputers, vol. 12, no. 11, pp. 1-34, article number 227, November 2023.

Abstract

The aim of this paper is to present an approach that utilizes several mixed reality technologies for touristic promotion and education. More specifically, mixed reality applications and games were created to promote the mountainous areas of Western Macedonia, Greece, and to educate visitors on various aspects of these destinations, such as their history and cultural heritage. Location-based augmented reality (AR) games were designed to guide the users to visit and explore the destinations, get informed, gather points and prizes by accomplishing specific tasks, and meet virtual characters that tell stories. Furthermore, an immersive lab was established to inform visitors about the region of interest through mixed reality content designed for entertainment and education. The lab visitors can experience content and games through virtual reality (VR) and augmented reality (AR) wearable devices. Likewise, 3D content can be viewed through special stereoscopic monitors. An evaluation of the lab experience was performed with a sample of 82 visitors who positively evaluated features of the immersive experience such as the level of satisfaction, immersion, educational usefulness, the intention to visit the mountainous destinations of Western Macedonia, intention to revisit the lab, and intention to recommend the experience to others.


BibTex

@article{Computers_23,
author         = {Alexandros Kleftodimos and Athanasios Evagelou and Stefanos Gkoutzios and Maria Matsiola and Michalis Vrigkas and Anastasia Yannacopoulou and Amalia Triantafillidou and Georgios Lappas},
title          = {Creating location-based augmented reality games and immersive experiences for touristic Ddstination marketing and education},
journal        = {Computers},
volume         = {12},
number         = {11},
pages          = {1--34},
article-number = {227},
month          = {November},
year           = {2023},
doi            = {10.3390/computers12110227}
}
  • 2023

    A virtual reality 3D game: A comparison between an immersive virtual reality application and a desktop experience

    M. Vrigkas, C. Nikou
    Conference Paper Proc. IEEE International Conference on Image Processing Challenges and Workshops (ICIPCW), pp. 3725-3729, Kuala Lumpur, Malaysia, October 8-11, 2023

Abstract

The work aims to design and implement a 3D interactive and addictive object avoidance game using the Unity platform. The implementation of the immersive virtual reality application uses any smart mobile device as an input and output device, utilizing its accelerometer and compass to record the orientation and rotation data of the device in 3D space and capture the digital environment stereoscopically on the device screen. A comparative study between a virtual reality and a desktop real-time 3D game is performed to analyze the various attributes of the game and determine which medium is most effective.


BibTex

@inproceedings{Vrigkas_et_al_3DCVP_ICIP23,
author    = {Michalis Vrigkas and Christophoros Nikou},
title     = {A virtual reality 3D game: A comparison between an immersive virtual reality application and a desktop experience},
booktitle = {Proc. IEEE International Conference on Image Processing Challenges and Workshops (ICIPCW)},
address   = {Kuala Lumpur, Malaysia},
pages     = {3725--3729},
month     = {October},
year      = {2023},
doi       = {doi: 10.1109/ICIPC59416.2023.10328382}
}
  • 2023

    Spatial transformer generative adversarial network for image super-resolution

    P. Rempakos, M. Vrigkas, M.E. Plissiti, C. Nikou
    Conference Paper 22nd International Conference on Image Analysis and Processing (ICIAP '23), pp. 399-411, Udine, Italy, September 11-15, 2023

Abstract

High-resolution images play an essential role in the performance of image analysis and pattern recognition methods. However, the expensive setup required to generate them and the inherent limitations of the sensors in optics manufacturing technology leads to the restricted availability of these images. In this work, we exploit the information retrieved in feature maps using the notable VGG networks and apply a transformer network to address spatial rigid affine transformation invariances, such as translation, scaling, and rotation. To evaluate and compare the performance of the model, three publicly available datasets were used. The model achieved very gratifying and accurate performance in terms of image PSNR and SSIM metrics against the baseline method.


BibTex

@inproceedings{Rempakos_et_al_ICIAP23,
author    = {Pantelis Rempakos and Michalis Vrigkas and Marina E. Plissiti and Christophoros Nikou},
title     = {Spatial Transformer Generative Adversarial Network for Image Super-Resolution},
booktitle = {Proc. 22nd International Conference on Image Analysis and Processing},
address   = {Udine, Italy},
pages     = {399--411},
month     = {September},
year      = {2023},
doi       = {10.1007/978-3-031-43148-7_34}
}
  • 2023

    Composition of motion from video animation through learning local transformations

    M. Vrigkas, V. Tagka, M.E. Plissiti, C. Nikou
    Conference Paper IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP '23), pp. 1-5, Rhodes Island, Greece, June 4-10, 2023

Abstract

In this work, we solve the problem of motion representation in videos, according to local transformations applied to specific keypoints extracted from static the images. First, we compute the coordinates of the keypoints of the body or face through a pre-trained model, and then we introduce a convolutional neural network to estimate a dense motion field through optical flow. Next, we train a generative adversarial network that exploits the previous information to generate new images that resemble as much as possible the target frames. To reduce trembling and extract smooth movements, our model incorporates a low-pass spatio-temporal Gaussian filter. Results indicate that our method provides high performance and the movement of objects is accurate and robust.


BibTex

@inproceedings{MVrigkas_etal_ICASSP_23,
author    = {Michalis Vrigkas and Virginia Tagka and Marina E. Plissiti and Christophoros Nikou},
title     = {Composition of motion from video animation through learning local transformations},
booktitle = {Proc. IEEE International Conference on Acoustics, Speech and Signal Processing},
address   = {Rhodes Island, Greece},
pages     = {1--5},
month     = {June},
year      = {2023}
}
  • 2023

    Consumer experience and augmented reality wine label application

    A. Triantafillidou, M. Vrigkas, A. Kleftodimos, A. Yannacopoulou, M. Matsiola, S. Gkoutzios, G. Lappas
    Conference Paper 8th International Conference on Digital Economy (ICDEc '23), pp. 263–273, Braga, Portugal, May 2-4, 2023

Abstract

Augmented reality (AR) applications are regarded as effective experiential marketing practices that can help companies promote their products/services in an interactive manner and deliver exceptional consumer experiences. The purpose of the present study is to evaluate a wine-label AR mobile application by examining its impact on consumer experience dimensions, satisfaction, and re-usage intentions towards the application, as well as attitude and purchase intentions towards the wine product. Moreover, to test the effect of product-consumption related factors (consumption frequency, amount of spending, wine expertise, and attention to wine labels) and technology-related factors (consumers’ familiarity with smartphone applications, number of AR applications used in smartphone, and extent of information search for wine-related information through smartphones) on the experiential dimensions of entertainment, flow, escapism, and education. Towards this end, a wine AR label application was developed and evaluated using a quantitative survey. In total, 306 respondents answered a self-administered questionnaire after interacting with the application. Results indicate that the AR application induced the entertainment and educational dimensions of consumer experience. The AR experience was also able to increase respondents’ satisfaction with the application and in turn help them form positive attitudes and purchase intentions for the wine. Moreover, the present study revealed that respondents’ expertise for wine, attention to wine label, familiarity with smartphone applications, and information search for wine-related information through smartphones are important factors that have an impact on the experience lived by consumers when using the AR label application.


BibTex

 @inproceedings{ICDEc23,
 author    = {A. Triantafillidou and M. Vrigkas and A. Kleftodimos and A. Yannacopoulou and M. Matsiola and S. Gkoutzios and G. Lappas},
 title     = {Consumer experience and augmented reality wine label application},
 booktitle = {Proc. 8th International Conference on Digital Economy},
 address   = {Braga, Portugal},
 pages     = {263--273},
 month     = {May},
 year      = {2023},
 doi       = {10.1007/978-3-031-42788-6_16}
 }
  • 2023

    Segmentation of SEM images of multiphase materials: When Gaussian mixture models are accurate?

    M. Chatzigeorgiou, M. Vrigkas, M. Beazi-Katsioti, M. Katsiotis, N. Boukos, V. Constantoudis
    Journal PaperJournal of Microscopy, vol. 289, no. 1, pp. 58-70, January 2023.

Abstract

Scanning Electron Microscopy (SEM) have found prosperous ground in the characterization of multiphase materials. One of the fastest modes of SEM with the ability of distinguish different phases is Back-Scattered Electron (BSE) imaging. As an imaging technique, however, the application of a segmentation method is required for the extraction of quantitative results. A very common segmentation technique is based on Gaussian Mixture Models algorithm. This algorithm is able to deconvolute the image histogram into different distributions attributed to different phases of the material. In this work a systematic study on the evaluation of GMM accuracy and an investigation of its limitation is conducted. Towards this investigation, a framework of synthetic BSE image histograms has been conducted for the control of parameters that correlate sample composition and image acquisition setting. The application of this framework is realized for the calculation of the impact of collective parameters of image histogram on the accuracy of GMM deconvolution. To this end some rules of thump are extracted for the guidance of SEM user on the prediction of GMM accuracy based on the inspection of image histogram only. These rule of thump can me summarized if one can distinguish the number of peaks in the histogram equal to the number of gaussian component taking part in GMM. If this is not the case, statistical moments, kurtosis and skewness, can be used in order to differentiate a histogram suitable for an accurate GMM deconvolution from non-suitable ones.


BibTex

 @article{MChatzigeorgiou_etal_JM_23,
 author    = {Manolis Chatzigeorgiou and Michalis Vrigkas and Margarita Beazi-Katsiotiand Marios Katsiotis and Nikos Boukos and Vassilios Constantoudis},
 title     = {Segmentation of SEM images of multiphase materials: When Gaussian mixture models are accurate?},
 journal   = {Journal of Microscopy},
 volume    = {289},
 number    = {1},
 pages     = {58--70},
 month     = {January},
 year      = {2023},
 doi       = {https://doi.org/10.1111/jmi.13150}
 }
  • 2022

    An augmented reality workflow for creating “live” wine labels

    M. Vrigkas, A. Kleftodimos, G. Lappas
    Journal PaperInternational Journal of Entertainment Technology and Management, vol. 1, no. 4, pp. 311–327, January 2022.

Abstract

Augmented reality (AR) technologies are constantly developing in various fields of communication such as entertainment, education, information, and marketing and other fields such as industrial product design among others. This paper aims to present an integrated AR workflow for the labelling of wine products. The proposed wine label enhancement workflow may work as follows: The wine business comes up with the idea of creating an AR experience for its wine products, then an AR expert designs the AR experience, and develops the AR application. As a final step in the process, the application is distributed to the users with the use of various platforms and the experience can then be activated by pointing the camera of a mobile device to the bottle label. The AR content is then generated and displayed to the user who can interact with the digital product on a whole new level.


BibTex

@article{MVrigkas_etal_JM_23,
author    = {Michalis Vrigkas and Alexandros Kleftodimos and Georgios Lappas },
title     = {An augmented reality workflow for creating ``live'' wine labels},
journal   = {International Journal of Entertainment Technology and Management},
volume    = {1},
number    = {4},
pages     = {311--327},
month     = {January},
year      = {2022},
doi       = {10.1504/IJENTTM.2022.10054762}
}
  • 2022

    Taleblazer vs Metaverse: A comparative analysis of the two platforms for building AR location-based educational experiences

    A. Kleftodimos, G. Lappas, M. Vrigkas
    Journal PaperInternational Journal of Entertainment Technology and Management, vol. 1, no. 4, pp. 290–310, January 2022.

Abstract

Location-based AR games are becoming increasingly popular in education. With location-based AR games, learners can obtain knowledge by visiting places of educational value through informative digital content that is activated and displayed on their mobile devices when specific locations are reached. To create location-based AR games, there are several available authoring tools. Taleblazer and Metaverse Studio are two popular platforms that are used nowadays by many educators. This study aims to perform a comparative analysis between these platforms to provide educators interested in developing location-based AR experiences with all the information needed to make an informed decision on which platform to use. The analysis examines the designer environment and its available features, the end-user interface, the documentation that accompanies each platform, and third-party applications that are developed by these tools. Furthermore, two game prototypes have been developed to better understand the two platforms’ functionality.


BibTex

@article{Kleftodimos_etal_JM_23,
author    = {Alexandros Kleftodimos and Georgios Lappas and Michalis Vrigkas},
title     = {Taleblazer vs Metaverse: A comparative analysis of the two platforms for building AR location-based educational experiences},
journal   = {International Journal of Entertainment Technology and Management},
volume    = {1},
number    = {4},
pages     = {290--310},
month     = {January},
year      = {2022},
doi       = {10.1504/IJENTTM.2022.10054761}
}
  • 2022

    Machine learning evaluation of microscopy image segmentation methods: The case of Gaussian mixture models

    M. Chatzigeorgiou, M. Vrigkas, N. Boukos, M. Beazi-Katsioti, M. Katsiotis and V. Constantoudis
    Conference Paper 12th Hellenic Conference on Artificial Intelligence (SETN '22). Association for Computing Machinery, New York, NY, USA, Article 59, 1–4, Corfu, Greece, September 7-9, 2022

Abstract

Multiphase materials are encountered in several areas of science and technology. Their properties are determined by the fraction of the phases (material compounds) constituting the composite material. Therefore, the quantitative characterization of phase fractions is highly demanded and has been the subject of extensive studies. To this end, a widely used technique is the segmentation of top-down back-scattered electron SEM (BSE-SEM) images given that different phases are depicted with pixel collections of different luminosity. Gaussian mixture models (GMM) are one the most popular and easily implemented methods to segment the BSE-SEM images through the deconvolution of their histograms. However, the accuracy and the limitations of their application have not been fully investigated. The aim of this paper is to design a neural-network approach to fill this gap and provide a fast tool for the automatic evaluation of the accuracy of GMM predictions for all material phases based on the inspection of the measured SEM image histogram alone. The proposed tool facilitates the decision-making process of an SEM user concerning the optimum choice of a segmentation method.


BibTex

@inproceedings{MChatzigeorgiou_etal_SETN_22,
author    = {Manolis Chatzigeorgiou and Michalis Vrigkas and Nikos Boukos and Margarita Beazi-Katsioti and Marios Katsiotis and Vassilios Constantoudis},
title     = {Machine learning evaluation of microscopy image segmentation methods: The case of Gaussian mixture models},
booktitle = {Proc. 12th Hellenic Conference on Artificial Intelligence},
address   = {Corfu, Greece},
articleno = {59},
numpages  = {4},
doi       = {10.1145/3549737.3549800},
month     = {September},
year      = {2022},
series    = {SETN '22}
}
  • 2022

    FaceMask: A new image dataset for the automated identification of people wearing masks in the wild

    M. Vrigkas, E-A. Kourfalidou, M.E. Plissiti, C. Nikou
    Journal PaperSensors, vol. 22, no. 3, 896, January 2022.

Abstract

The rapid spread of the COVID-19 pandemic, in early 2020, has radically changed the lives of people. In our daily routine, the use of a face (surgical) mask is necessary, especially in public places, to prevent the spread of this disease. Furthermore, in crowded indoor areas, the automated recognition of people wearing a mask is a requisite for the assurance of public health. In this direction, image processing techniques, in combination with deep learning, provide effective ways to deal with this problem. However, it is a common phenomenon that well-established datasets containing images of people wearing masks are not publicly available. To overcome this obstacle and to assist the research progress in this field, we present a publicly available annotated image database containing images of people with and without a mask on their faces, in different environments and situations. Moreover, we tested the performance of deep learning detectors in images and videos on this dataset. The training and the evaluation were performed on different versions of the YOLO network using Darknet, which is a state-of-the-art real-time object detection system. Finally, different experiments and evaluations were carried out for each version of YOLO, and the results for each detector are presented.


BibTex

@article{MVrigkas_etal_22,
author          = {Michalis Vrigkas and Evangelia-Andriana Kourfalidou and Marina E. Plissiti and Christophoros Nikou},
title           = {FaceMask: A new image dataset for the automated identification of people wearing masks in the wild},
journal         = {Sensors},
volume          = {22},
number          = {3},
article-number  = {896},
month           = {January},
year            = {2022},
doi             = {https://doi.org/10.3390/s22030896}
}
  • 2021

    Human activity recognition using robust adaptive privileged probabilistic learning

    M. Vrigkas, E. Kazakos, C. Nikou, I.A. Kakadiaris
    Journal PaperPattern Analysis and Applications, vol. 24, no. 3, pp. 915-932, January 2021.

Abstract

In this work, a supervised probabilistic approach is proposed that integrates the learning using privileged information (LUPI) paradigm into a hidden conditional random field (HCRF) model, called HCRF+, for human action recognition. The proposed model employs a self-training technique for automatic estimation of the regularization parameters of the objective function. Moreover, the method provides robustness to outliers by modeling the conditional distribution of the privileged information by a Student's t-density function, which is naturally integrated into the HCRF+ framework. The proposed method was evaluated using different forms of privileged information on four publicly available datasets. The experimental results demonstrate its effectiveness concerning the-state-of-the-art in the LUPI framework using both hand-crafted and deep learning-based features extracted from a convolutional neural network.


BibTex

@article{MVrigkas_etal_21,
author    = {Michalis Vrigkas and Evangelos Kazakos and Christophoros Nikou and Ioannis A. Kakadiaris},
title     = {Human activity recognition using robust adaptive privileged probabilistic learning},
journal   = {Pattern Analysis and Applications},
volume    = {24},
number    = {3},
pages     = {915–-932},
month     = {January},
year      = {2021},
doi       = {https://doi.org/10.1007/s10044-020-00953-x}
}
  • 2021

    Augmented Reality for Wine Industry: Past, Present, and Future

    M. Vrigkas, G. Lappas, A. Kleftodimos, A. Triantafillidou
    Conference Paper3rd ETLTC International Conference on Information and Communications Technology, pp.04006, Aizuwakamatsu Japan, January 27-30 2021

Abstract

In this paper, we study the concepts, materials, tools, and applications that constitute what we call augmented reality (AR) for the wine industry. A comprehensive review of what are the basic multimedia content and the minimum algorithmic requirements used to implement successful AR applications for wine products is given. To this end, we provide a detailed analysis of how AR technology is used to create augmented “live” wine labels, and how digital storytelling has revolutionized wine products marketing. Also, we describe the use of AR technology to promote winemaking companies to influence consumer preferences. Finally, we report the characteristics of future research directions and some open issues and challenges on using AR for wine product promotion.


BibTex

@inproceedings{MVrigkas_etal_ETLTC_21,
author    = {Michalis Vrigkas and Georgios Lappas and Alexandros Kleftodimos and Amalia Triantafillidou},
title     = {Augmented Reality for Wine Industry: Past, Present, and Future},
booktitle = {3rd International Conference on Information and Communications Technology},
address   = {Aizuwakamatsu Japan},
pages     = {04006},
month     = {January},
year      = {2021},
doi       = {10.1051/shsconf/202110204006}
}
  • 2020

    Gender and age estimation without facial information from still images

    G. Chatzitzisi, M. Vrigkas, C. Nikou
    Conference Paper15th International Symposioum on Visual Computing, pp.488-500, October 5-7 2020

Abstract

In this paper, the task of gender and age recognition is performed on pedestrian still images, which are usually captured in-the-wild with no near face-frontal information. Moreover, another difficulty originates from the underlying class imbalance in real examples, especially for the age estimation problem. The scope of the paper is to examine how different loss functions in convolutional neural networks (CNN) perform under the class imbalance problem. For this purpose, as a backbone, we employ the Residual Network (ResNet). On top of that, we attempt to benefit from appearance-based attributes, which are inherently present in the available data. We incorporate this knowledge in an autoencoder, which we attach to our baseline CNN for the combined model to jointly learn the features and increase the classification accuracy. Finally, all of our experiments are evaluated on two publicly available datasets.


BibTex

@inproceedings{GChatzitzisi_etal_20,
author    = {Georgia Chatzitzisi and Michalis Vrigkas and Christophoros Nikou},
title     = {Gender and age estimation without facial information from still images},
booktitle = {International Symposioum on Visual Computing},
publisher = {Springer International Publishing},
address   = {San Diego, CA},
pages     = {488-500},
month     = {October},
year      = {2020},
isbn      = {978-3-030-64556-4}
}
  • 2020

    Improving 3D medical image compression efficiency using spatiotemporal coherence

    M.C. Zerva, M. Vrigkas, L.P. Kondi, C. Nikou
    Conference PaperIS&T International Symposioum on Electronic Imaging, Image Processing: Algorithms and Systems XVII, Burlingame, CA, USA, January 26-30 2020

Abstract

Advanced methodologies for transmitting compressed images, within acceptable ranges of transmission rate and loss of information, make it possible to transmit a medical image through a communication channel. Most prior works on 3D medical image compression consider volumetric images as a whole but fail to account for the spatial and temporal coherence of adjacent slices. In this paper, we set out to develop a 3D medical image compression method that extends the 3D wavelet difference reduction algorithm by computing the similarity of the pixels in adjacent slices and progressively compress only the similar slices. The proposed method achieves high-efficiency performance on publicly available datasets of MRI scans by achieving compression down to one bit per voxel with PSNR and SSIM up to 52.3 dB and 0.7578, respectively.


BibTex

@inproceedings{MZerva_etal_20,
  author    = {Matina Ch. Zerva and Michalis Vrigkas and Lisimachos P. Kondi and Christophoros Nikou},
  title     = {Improving {3D} medical image compression efficiency using spatiotemporal coherence},
  booktitle = {Proc. IS&T International Symposioum on Electronic Imaging, Image Processing: Algorithms and Systems XVII},
  pages     = {63-1-63-6},
  address   = {Burlingame, CA},
  month     = {January},
  year 	    = {2020}
}
  • 2019

    RECASPIA: Recognizing carrying actions in single images using privileged information

    C. Smailis, M. Vrigkas, I.A. Kakadiaris
    Conference Paper26th International Conference on Image Processing, pp. 26-30, Taipei, Taiwan, September 22-25 2019

Abstract

Many approaches for action recognition focus on general actions, such as “running” or “walking”. This work presents a method for recognizing carrying actions in single images, by utilizing privileged information, such as annotations, available only during training, following the learning using privileged information paradigm. In addition, we introduce a dataset for carrying actions, formed using images extracted from YouTube videos depicting several scenarios. We accompany the dataset with a variety of different annotation types that include human pose, object and scene attributes. The experimental results demonstrate that our method, boosted sample averaged F1 score performance by 15.4% and 4.15% respectively, in the validation and testing partitions of our dataset, when compared to an end-to-end CNN model, trained only with observable information.


BibTex

@inproceedings{CSmailis_ICIP19,
  author    = {Christos Smailis and Michalis Vrigkas and Ioannis A. Kakadiaris},
  title     = {RECASPIA: Recognizing carrying actions in single images using privileged information},
  booktitle = {Proc. 26th IEEE International Conference on Image Processing},
  pages     = {26--30},
  address   = {Taipei, Taiwan},
  month     = {September},
  year 	    = {2019}
}
  • 2018

    Robust incremental hidden conditional random fields for human action recognition

    M. Vrigkas, E. Mastora, C. Nikou, I.A. Kakadiaris
    Conference Paper13th International Symposium on Visual Computing, pp. 126-136, Las Vegas, NV, November 19-21 2018

Abstract

Hidden conditional random fields (HCRFs) are a powerful supervised classification system, which is able to capture the intrinsic motion patterns of a human action. However, finding the optimal number of hidden states remains a severe limitation for this model. This paper addresses this limitation by proposing a new model, called robust incremental hidden conditional random field (RI-HCRF). A hidden Markov model (HMM) is created for each observation paired with an action label and its parameters are defined by the potentials of the original HCRF graph. Starting from an initial number of hidden states and increasing their number incrementally, the Viterbi path is computed for each HMM. The method seeks for a sequence of hidden states, where each variable participates in a maximum number of optimal paths. Thereby, variables with low participation in optimal paths are rejected. In addition, a robust mixture of Student's t-distributions is imposed as a regularizer to the parameters of the model. The experimental results on human action recognition show that RI-HCRF successfully estimates the number of hidden states and outperforms all state-of-the-art models.


BibTex

@inproceedings{MVrigkas_ISVC18,
  author    = {Michalis Vrigkas and Ermioni Mastora and Christophoros Nikou and Ioannis A. Kakadiaris},
  title     = {Robust incremental hidden conditional random fields for human action recognition},
  booktitle = {Proc. 13th International Symposium on Visual Computing},
  address   = {Las Vegas, NV},
  month     = {November},
  pages	    = {126--136},
  year 	    = {2018}
}
  • 2018

    Machine learning outperforms ACC/AHA CVD risk calculator in MESA

    I.A. Kakadiaris, M. Vrigkas, A.A. Yen, T. Kuznetsova, M. Budoff, M. Naghavi
    Journal PaperJournal of the American Heart Association, vol. 7, no. 22, pp. e009476, November 2018

Abstract

The 2013 ACC/AHA Pooled Cohort Equations risk calculator has been shown to be inaccurate in certain populations. Using the same risk variables, we developed a Machine Learning-based risk calculator in the MESA (Multi-Ethnic Study of Atherosclerosis) cohort and validated in the Flemish Study on Environment, Genes and Health Outcomes (FLEMENGHO). The ML Risk Calculator outperformed the ACC/AHA Risk Calculator by recommending less drug therapy, yet missing fewer CVD events. These findings demonstrate the potential of Machine Learning to assist medical decision-making.


BibTex

@article{Kakadiaris_JAHA18,
  author  = {Ioannis A. Kakadiaris and Michalis Vrigkas and Albert A. Yen and Tatiana Kuznetsova and Matthew Budoff and Morteza Naghavi},
  title   = {Machine learning outperforms {ACC/AHA CVD} risk calculator in {MESA}},
  journal = {Journal of the American Heart Association},
  volume  = {7},
  number  = {22},
  pages   = {e009476},
  year    = {2018},
  month   = {November},
  doi     = {10.1161/JAHA.118.009476}
}
  • 2018

    Machine learning outperforms ACC/AHA CVD risk calculator in MESA offering new opportunities for short-term risk prediction and early detection of the vulnerable patient

    I.A. Kakadiaris, M. Vrigkas, A.A. Yen, T. Kuznetsova, M. Budoff, M. Naghavi
    Conference - Abstract Circulation, vol 138, no. Suppl 1, pp. A17154, American Heart Association, Scientific Sessions, Chicago, IL, November, 2018

Abstract

Introduction: Machine learning (ML) is poised to revolutionize healthcare. Current national guidelines for prediction and prevention of atherosclerotic cardiovascular disease (ASCVD) use ACC/AHA Pooled Cohort Equation Risk Calculator which relies on traditional risk factors and linear statistical models. Unfortunately, this approach yields a low level of sensitivity and specificity. The low sensitivity results in missing high-risk individuals who need intensive therapy and the low specificity results in millions of people unnecessarily recommended drugs such as statin. We aimed to utilize Machine Learning (ML) to create a more accurate predictor of ASCVD events and whom to recommend statin.
Methods: We developed and validated a ML Risk Calculator based on Support Vector Machines (SVMs) using the latest 13-year follow up dataset from MESA (Multi-Ethnic Study of Atherosclerosis) of 6,459 participants who were free of cardiovascular disease at baseline. We provided identical input to the ACC/AHA and ML risk calculators and compared their accuracy. We also validated the ML model in another longitudinal cohort: the Flemish Study on Environment, Genes and Health Outcomes (FLEMENGHO).
Results: According to the ACC/AHA Risk Calculator and a 7.5% 10-year risk threshold, 46.0% would be recommended statin. Despite this high proportion, 23.8% of the 480 “Hard CVD” events occurred in those not recommended statin, resulting in sensitivity (Sn) 0.76, specificity (Sp) 0.56, and AUC 0.71. In contrast, ML Risk Calculator recommended statin to 11.4%, and only 14.4% of “Hard CVD” events occurred in those not recommended statin, resulting in Sn 0.86, Sp 0.95, and AUC 0.92. Similar results were seen in prediction of “All CVD” events.
Conclusions: The ML Risk Calculator outperformed the ACC/AHA Risk Calculator by recommending less drug therapy, yet missing fewer events. Additional studies are underway to validate the ML model in other cohorts and to explore its ability in predicting short-term (1-5 years) events with additional biomarkers including imaging. Machine learning is paving the way for early detection of asymptomatic high-risk individuals destined to a CVD event in the near future, the Vulnerable Patient


BibTex

@inproceedings{Kakadiaris_etal18,
  author    = {Ioannis A Kakadiaris and Michail Vrigkas and Albert Yen and Tatiana Kuznetsova and Matthew Budoff and Morteza Naghavi},
  title     = {Machine learning outperforms ACC/AHA CVD risk calculator in MESA offering new opportunities for short-term risk prediction and early detection of the vulnerable patient},
  journal   = {Circulation},
  volume    = {138},
  number    = {Suppl\ 1},
  pages     = {A17154--A17154},
  year      = {2018},
  month     = {November},
  address   = {Chicago, IL},
  publisher = {American Heart Association, Inc.},
  doi       = {10.1161/circ.138.suppl\_1.17154}
}
  • 2018

    SPICE: Superpixel classification for cell detection and counting

    O. Magaña-Tellez, M. Vrigkas, C. Nikou, I.A. Kakadiaris
    Conference Paper 13th International Conference on Computer Vision Theory and Applications pp. 485-490, Funchal, Madeira, Portugal, January 27-29 2018

Abstract

An algorithm for the localization and counting of cells in histopathological images is presented. The algorithm relies on the presegmentation of an image into a number of superpixels followed by two random forests for classification. The first random forest determines if there are any cells in the superpixels at its input and the second random forest provides the number of cells in the respective superpixel. The algorithm is evaluated on a bone marrow histopathological dataset. We argue that a single random forest is not sufficient to detect all the cells in the image while a cascade of classifiers achieves higher accuracy. The results compare favorably with the state of the art but with a lower computational cost.


BibTex

@inproceedings{MOman_VISAPP18,
  author    = {Oman Maga\~{n}a-Tellez and Michalis Vrigkas and Christophoros Nikou and Ioannis A. Kakadiaris},
  title     = {SPICE: Superpixel classification for cell detection and counting},
  booktitle = {Proc. 13th International Conference on Computer Vision Theory and Applications},
  address   = {Funchal, Madeira, Portugal},
  month     = {January},
  pages	    = {485--490},
  year 	    = {2018}
}
  • 2017

    Machine learning outperformed ACC/AHA Pooled Cohort Equations Risk Calculator for detection of high-risk asymptomatic individuals and recommending treatment for prevention of cardiovascular events in the Multi-Ethnic Study of Atherosclerosis (MESA)

    I.A. Kakadiaris, M. Vrigkas, M. Budoff, A. Yen, M. Naghavi
    Conference - Abstract Circulation, vol 136, no. Suppl 1, pp. A23075, American Heart Association, Scientific Sessions, Anaheim, CA, November 2017

Abstract

Studies have shown that the status quo for atherosclerotic cardiovascular disease (ASCVD) prediction in the U.S. - using ACC/AHA Pooled Cohort Equations Risk Calculator - is inaccurate and results in overtreatment of low-risk and undertreatment of high-risk individuals. Machine Learning (ML) is poised to revolutionize healthcare. We used ML to develop a new ASCVD risk calculator and tackled the problem.


BibTex

@article{IKakadiaris_AHA17,
  author    = {Ioannis Kakadiaris and Michalis Vrigkas and Matthew Budoff and Albert Yen and Morteza Naghavi},
  title     = {Machine learning outperformed {ACC/AHA} {P}ooled {C}ohort {E}quations {R}isk {C}alculator for detection of high-risk asymptomatic individuals and recommending treatment for prevention of cardiovascular events in the {M}ulti-{E}thnic {S}tudy of {A}therosclerosis {(MESA)}},
  volume    = {136},
  number    = {Suppl 1},
  pages     = {A23075--A23075},
  year      = {2017},
  month     = {November 11-15},
  address   = {Anaheim, CA},
  publisher = {American Heart Association, Inc.},
  issn      = {0009-7322},
  URL       = {http://circ.ahajournals.org/content/136/Suppl_1/A23075},
  eprint    = {http://circ.ahajournals.org/content},
  journal   = {Circulation}
}
  • 2017

    Inferring human activities using robust privileged probabilistic learning

    M. Vrigkas, E. Kazakos, C. Nikou, I.A. Kakadiaris
    Conference PaperIEEE International Conference on Computer VisionWorkshops, pp. 2658-2665, Venice, Italy, October 22-29 2017

Abstract

Classification models may often suffer from “structure imbalance” between training and testing data that may occur due to the deficient data collection process. This imbalance can be represented by the learning using privileged information (LUPI) paradigm. In this paper, we present a supervised probabilistic classification approach that integrates LUPI into a hidden conditional random field (HCRF) model. The proposed model is called LUPI-HCRF and is able to cope with additional information that is only available during training. Moreover, the proposed method employes Student's t-distribution to provide robustness to outliers by modeling the conditional distribution of the privileged information. Experimental results in three publicly available datasets demonstrate the effectiveness of the proposed approach and improve the state-of-the-art in the LUPI framework for recognizing human activities.


BibTex

@inproceedings{MVrigkas_ICCVW17,
  author    = {Michalis Vrigkas and Evangelos Kazakos and Christophoros Nikou and Ioannis A. Kakadiaris},
  title     = {Inferring human activities using robust privileged probabilistic learning},
  booktitle = {Proc. IEEE International Conference on Computer Vision Workshops},
  year      = {2017},
  month     = {October},
  pages     = {2658--2665},
  address   = {Venice, Italy}
}
  • 2017

    Adaptive SVM+: Learning with privileged information for domain adaptation

    N. Sarafianos, M. Vrigkas, I.A. Kakadiaris
    Conference PaperIEEE International Conference on Computer VisionWorkshops, pp. 2637-2644, Venice, Italy, October 22-29 2017

Abstract

Incorporating additional knowledge in the learning process can be beneficial for several computer vision and machine learning tasks. Whether privileged information originates from a source domain that is adapted to a target domain, or as additional features available at training time only, using such privileged (i.e., auxiliary) information is of high importance as it improves the recognition performance and generalization. However, both primary and privileged information are rarely derived from the same distribution, which poses an additional challenge to the recognition task. To address these challenges, we present a novel learning paradigm that leverages privileged information in a domain adaptation setup to perform visual recognition tasks. The proposed framework, named Adaptive SVM+, combines the advantages of both the learning using privileged information (LUPI) paradigm and the domain adaptation framework, which are naturally embedded in the objective function of a regular SVM. We demonstrate the effectiveness of our approach on the publicly available Animals with Attributes and INTERACT datasets and report state-of-the-art results in both of them.


BibTex

@inproceedings{NSarafianos_ICCVW17,
  author    = {Nikolaos Sarafianos and Michalis Vrigkas and Ioannis A. Kakadiaris},
  title     = {Adaptive SVM+: Learning with privileged information for domain adaptation},
  booktitle = {Proc. IEEE International Conference on Computer Vision Workshops},
  year      = {2017},
  month     = {October},
  pages     = {2637--2644},
  address   = {Venice, Italy}
}
  • 2017

    Identifying human behaviors using synchronized audio-visual cues

    M. Vrigkas, C. Nikou, I.A. Kakadiaris
    Journal PaperIEEE Transactions on Affective Computing, vol. 8, no. 1, pp. 54-66, January-March 2017

Abstract

In this paper, a human behavior recognition method using multimodal features is presented. We focus on modeling individual and social behaviors of a subject (e.g., friendly/aggressive or hugging/kissing behaviors) with a hidden conditional random field (HCRF) in a supervised framework. Each video is represented by a vector of spatio-temporal visual features (STIP, head orientation and proxemic features) along with audio features (MFCCs). We propose a feature pruning method for removing irrelevant and redundant features based on the spatio-temporal neighborhood of each feature in a video sequence. The proposed framework assumes that human movements are highly correlated with sound emissions. For this reason, canonical correlation analysis (CCA) is employed to find correlation between the audio and video features prior to fusion. The experimental results, performed in two human behavior recognition datasets including political speeches and human interactions from TV shows, attest the advantages of the proposed method compared with several baseline and alternative human behavior recognition methods.


BibTex

@article{MVrigkas_TAffC15,
  author  = {Michalis Vrigkas and Christophoros Nikou and Ioannis A. Kakadiaris},
  title   = {Identifying human behaviors using synchronized audio-visual cues},
  journal = {IEEE Transactions on Affective Computing},
  year    = {2017},
  volume  = {8},
  number  = {1},
  pages   = {54-66},
  doi     = {10.1109/TAFFC.2015.2507168},
  month   = {January}
}
  • 2016

    Human activity recognition using conditional random fields and privileged information

    M. Vrigkas
    PhD ThesisDepartment of Computer Science and Engineering, University of Ioannina, May 2016

Abstract

This thesis solves the problem of human activity recognition from video sequences. To model human activities, conditional random fields were applied using data from heterogeneous sources. Moreover, a novel classification scheme that is based on the learning using privileged information (LUPI) paradigm was also proposed, where privileged information is given as an additional input to the classification model and it is available only during training but never during testing. Experimental results demonstrated that privileged information helps to build a stronger classifier than one would not learn without it, while it significantly increases the recognition accuracy of the model.


BibTex

@phdthesis{phdthesisMVrigkas16,
  author  = {Michalis Vrigkas},
  title   = {Human activity recognition using conditional random fields and privileged information},
  school  = {Department of Computer Science and Engineering, University of Ioannina},
  year    = {2018},
  month   = {May}
}
  • 2016

    Active privileged learning of human activities from weakly labeled samples

    M. Vrigkas, C. Nikou, I.A. Kakadiaris
    Conference Paper23rd IEEE International Conference on Image Processing, pp. 3036-3040, Phoenix, AZ, September 25-28 2016

Abstract

In many human activity recognition systems the size of the unlabeled training data may be significantly large due to expensive human effort required for data annotation. Moreover, the insufficient data collection process from heterogenous sources may cause dissimilarities between training and testing data. To address these limitations, a novel probabilistic approach that combines learning using privileged information (LUPI) and active learning is proposed. A pool-based privileged active learning approach is presented for semi-supervising learning of human activities from multimodal labeled and unlabeled data. Both uncertainty and distance from the decision boundary are used as a query inference strategies for selecting an unlabeled observation and query its label. Experimental results in four publicly available datasets demonstrate that the proposed method can identify with high accuracy complex human activities.


BibTex

@inproceedings{MVrigkas_ICIP16,
  author    = {Michalis Vrigkas and Christophoros Nikou and Ioannis A. Kakadiaris},
  title	    = {Active privileged learning of human activities from weakly labeled samples},
  booktitle = {Proc. 23rd IEEE International Conference on Image Processing},
  year      = {2016},
  month     = {September},
  pages     = {3036--3040},
  address   = {Phoenix, AZ}
}
  • 2016

    Exploiting privileged information for facial expression recognition

    M. Vrigkas, C. Nikou, I.A. Kakadiaris
    Conference Paper9th IAPR International Conference on Biometrics, pp. 1-8, Halmstad, Sweden, June 13-16 2016
    Honorable Mention Paper Award

Abstract

Most of the facial expression recognition methods consider that both training and testing data are equally distributed. As facial image sequences may contain information for heterogeneous sources, facial data may be asymmetrically distributed between training and testing, as it may be difficult to maintain the same quality and quantity of information. In this work, we present a novel classification method based on the learning using privileged information (LUPI) paradigm to address the problem of facial expression recognition. We introduce a probabilistic classification approach based on conditional random fields (CRFs) to indirectly propagate knowledge from privileged to regular feature space. Each feature space owns specific parameter settings, which are combined together through a Gaussian prior, to train the proposed t-CRF+ model and allow the different tasks to share parameters and improve classification performance. The proposed method is validated on two challenging and publicly available benchmarks on facial expression recognition and improved the state-of-the-art methods in the LUPI framework.


BibTex

@inproceedings{MVrigkas_ICIP16,
  author    = {Michalis Vrigkas and Christophoros Nikou and Ioannis A. Kakadiaris},
  title     = {Exploiting privileged information for facial expression recognition},
  booktitle = {Proc. 9th IAPR/IEEE International Conference on Biometrics},
  year      = {2016},
  month     = {June},
  pages     = {1--8},
  address   = {Halmstad, Sweden},
  doi       = {10.1109/ICB.2016.7550048},
  note      = {Honorable Mention Paper Award}
}
  • 2015

    A review of human activity recognition methods

    M. Vrigkas, C. Nikou, I.A. Kakadiaris
    Journal PaperFrontiers in Robotics and Artificial Intelligence, vol 2, no. 28, pp. 1-26, November 2015

Abstract

Recognizing human activities from video sequences or still images is a challenging task due to problems, such as background clutter, partial occlusion, changes in scale, viewpoint, lighting, and appearance. Many applications, including video surveillance systems, human-computer interaction, and robotics for human behavior characterization, require a multiple activity recognition system. In this work, we provide a detailed review of recent and state-of-the-art research advances in the field of human activity classification. We propose a categorization of human activity methodologies and discuss their advantages and limitations. In particular, we divide human activity classification methods into two large categories according to whether they use data from different modalities or not. Then, each of these categories is further analyzed into sub-categories, which reflect how they model human activities and what type of activities they are interested in. Moreover, we provide a comprehensive analysis of the existing, publicly available human activity classification datasets and examine the requirements for an ideal human activity recognition dataset. Finally, we report the characteristics of future research directions and present some open issues on human activity recognition.


BibTex

@article{MVrigkas_FRONTIERS2015,
  author  = {Michalis Vrigkas and Christophoros Nikou and Ioannis A. Kakadiaris},
  title   = {A review of human activity recognition methods},
  journal = {Frontiers in Robotics and Artificieal Inteligence},
  volume  = {2},
  number  = {28},
  pages   = {1--26},
  year    = {2015},
  url     = {http://www.frontiersin.org/vision_systems_theory,_tools_and_applications/10.3389/frobt.2015.00028/abstract},
  doi     = {10.3389/frobt.2015.00028},
  issn    = {2296-9144}
}
  • 2015

    Segmentation of cell clusters in Pap smear images using intensity variation between superpixels

    M.E. Plissiti, M. Vrigkas, C. Nikou
    Conference Paper 22nd International Conference on Systems, Signals and Image Processing, pp. 184-187, London, UK, September 10-12 2015

Abstract

The automated interpretation of Pap smear images is a challenging issue with several aspects. The accurate segmentation of the structuring elements of each cell is a crucial procedure which entails in the correct identification of pathological situations. However, the extended cell overlapping in Pap smear slides complicates the automated analysis of these cytological images. In this work, we propose an efficient algorithm for the separation of the cytoplasm area of overlapping cells. The proposed method is based on the fact that in isolated cells the pixels of the cytoplasm exhibit similar features and the cytoplasm area is homogeneous. Thus, the observation of intensity changes in extended subareas of the cytoplasm indicates the existence of overlapping cells. In the first step of the proposed method, the image is tesselated into perceptually meaningful individual regions using a superpixel algorithm. In a second step, these areas are merged into regions exhibiting the same characteristics, resulting in the identification of each cytoplasm area and the corresponding nuclei. The area of overlap is then detected using an algorithm that specifies faint changes in the intensity of the cytoplasm of each cell. The method has been evaluated on cytological images of conventional Pap smears, and the results are very promising.


BibTex

@inproceedings{MPlissiti_IWSSIP15,
  author    = {Marina E. Plissiti and Michalis Vrigkas and Christophoros Nikou},
  title     = {Segmentation of cell clusters in Pap smear images using intensity variation between superpixels},
  booktitle = {Proc. 22nd International Conference on Systems, Signals and Image Processing},
  year      = {2015},
  month     = {September},
  pages     = {184--187},
  address   = {London, UK}
}
  • 2014

    Robust maximum a posteriori image super-resolution

    M. Vrigkas, C. Nikou, L.P. Kondi
    Journal PaperJournal of Electronic Imaging, vol. 23, no. 4, pp. 043016, July 2014

Abstract

A global robust M-estimation scheme for maximum a posteriori (MAP) image super-resolution which efficiently addresses the presence of outliers in the low-resolution images is proposed. In iterative MAP image super-resolution, the objective function to be minimized involves the highly resolved image, a parameter controlling the step size of the iterative algorithm, and a parameter weighing the data fidelity term with respect to the smoothness term. Apart from the robust estimation of the high-resolution image, the contribution of the proposed method is twofold: (1) the robust computation of the regularization parameters controlling the relative strength of the prior with respect to the data fidelity term and (2) the robust estimation of the optimal step size in the update of the high-resolution image. Experimental results demonstrate that integrating these estimations into a robust framework leads to significant improvement in the accuracy of the high-resolution image.


BibTex

@article{MVrigkas_JEI14,
  author  = {Michalis Vrigkas and Christophoros Nikou and Lisimachos P. Kondi},
  title   = {Robust maximum a posteriori image super-resolution},
  journal = {Journal of Electronic Imaging},
  volume  = {23},
  number  = {4},
  pages   = {043016},
  year    = {2014},
  isbn    = {1017-9909},
  doi     = {10.1117/1.JEI.23.4.043016},
  URL     = {http://dx.doi.org/10.1117/1.JEI.23.4.043016}
}
  • 2014

    Classifying behavioral attributes using conditional random fields

    M. Vrigkas, C. Nikou, I.A. Kakadiaris
    Conference Paper8th Hellenic Conference on Artificial Intelligence, pp. 95-104, Ioannina, Greece, May 15-17 2014

Abstract

A human behavior recognition method with an application to political speech videos is presented. We focus on modeling the behavior of a subject with a conditional random field (CRF). The unary terms of the CRF employ spatiotemporal features (i.e., HOG3D, STIP and LBP). The pairwise terms are based on kinematic features such as the velocity and the acceleration of the subject. As an exact solution to the maximization of the posterior probability of the labels is generally intractable, loopy belief propagation was employed as an approximate inference method. To evaluate the performance of the model, we also introduce a novel behavior dataset, which includes low resolution video sequences depicting different people speaking in the Greek parliament. The subjects of the Parliament dataset are labeled as friendly, aggressive or neutral depending on the intensity of their political speech. The discrimination between friendly and aggressive labels is not straightforward in political speeches as the subjects perform similar movements in both cases. Experimental results show that the model can reach high accuracy in this relatively difficult dataset.


BibTex

@inproceedings{MVrigkas_SETN14,
  author    = {Michalis Vrigkas and Christophoros Nikou and Ioannis A. Kakadiaris},
  title     = {Classifying behavioral attributes using conditional random fields},
  booktitle = {Proc. 8th Hellenic Conference on Artificial Intelligence},
  year      = {2014},
  month     = {May},
  pages     = {95--104},
  volume    = {8445},
  series    = {Lecture Notes in Computer Science},
  address   = {Ioannina, Greece}
}
  • 2014

    Matching mixtures of curves for human action recognition

    M. Vrigkas, V. Karavasilis, C. Nikou, I.A. Kakadiaris
    Journal PaperComputer Vision and Image Understanding, vol. 119, pp. 27-40, February 2014

Abstract

A learning-based framework for action representation and recognition relying on the description of an action by time series of optical flow motion features is presented. In the learning step, the motion curves representing each action are clustered using Gaussian mixture modeling (GMM). In the recognition step, the optical flow curves of a probe sequence are also clustered using a GMM, then each probe sequence is projected onto the training space and the probe curves are matched to the learned curves using a non-metric similarity function based on the longest common subsequence, which is robust to noise and provides an intuitive notion of similarity between curves. Alignment between the mean curves is performed using canonical time warping. Finally, the probe sequence is categorized to the learned action with the maximum similarity using a nearest neighbor classification scheme. We also present a variant of the method where the length of the time series is reduced by dimensionality reduction in both training and test phases, in order to smooth out the outliers, which are common in these type of sequences. Experimental results on KTH, UCF Sports and UCF YouTube action databases demonstrate the effectiveness of the proposed method.


BibTex

@article{MVrigkas_CVIU14,
  author  = {Michalis Vrigkas and Vasileios Karavasilis and Christophoros Nikou and Ioannis A. Kakadiaris},
  title   = {Matching mixtures of curves for human action recognition},
  journal = {Computer Vision and Image Understanding},
  volume  = {119},
  pages   = {27--40},
  year    = {2014},
  issn    = {1077--3142},
  doi     = {http://dx.doi.org/10.1016/j.cviu.2013.11.007}
}
  • 2013

    Accurate image registration for MAP image super-resolution

    M. Vrigkas, C. Nikou, L.P. Kondi
    Journal PaperSignal Processing: Image Communication, vol. 28, no. 5, pp. 494-508, May 2013

Abstract

The accuracy of image registration plays a dominant role in image super-resolution methods and in the related literature, landmark-based registration methods have gained increasing acceptance in this framework. In this work, we take advantage of a maximum a posteriori (MAP) scheme for image super-resolution in conjunction with the maximization of mutual information to improve image registration for super-resolution imaging. Local as well as global motion in the low-resolution images is considered. The overall scheme consists of two steps. At first, the low-resolution images are registered by establishing correspondences between image features. The second step is to fine-tune the registration parameters along with the high-resolution image estimation, using the maximization of mutual information criterion. Quantitative and qualitative results are reported indicating the effectiveness of the proposed scheme, which is evaluated with different image features and MAP image super-resolution computation methods.


BibTex

@article{MVrigkas_SPIC13,
  author  = {Michalis Vrigkas and Christophoros Nikou and Lisimachos P. Kondi},
  title   = {Accurate image registration for \{MAP\} image super-resolution},
  journal = {Signal Processing: Image Communication},
  volume  = {28},
  number  = {5},
  pages   = {494--508},
  year    = {2013},
  issn    = {0923-5965},
  doi     = {10.1016/j.image.2012.12.008}
}
  • 2013

    Action recognition by matching clustered trajectories of motion vectors

    M. Vrigkas, V. Karavasilis, C. Nikou, I.A. Kakadiaris
    Conference Paper8th International Conference on Computer Vision Theory and Applications, pp. 112-117, Barcelona, Spain, February 21-24 2013

Abstract

A framework for action representation and recognition based on the description of an action by time series of optical flow motion features is presented. In the learning step, the motion curves representing each action are clustered using Gaussian mixture modeling (GMM). In the recognition step, the optical flow curves of a probe sequence are also clustered using a GMM and the probe curves are matched to the learned curves using a non-metric similarity function based on the longest common subsequence which is robust to noise and provides an intuitive notion of similarity between trajectories. Finally, the probe sequence is categorized to the learned action with the maximum similarity using a nearest neighbor classification scheme. Experimental results on common action databases demonstrate the effectiveness of the proposed method.


BibTex

@inproceedings{MVrigkas_VISAPP13,
  author    = {Michalis Vrigkas and Vasileios Karavasilis and Christophoros Nikou and Ioannis Kakadiaris},
  title     = {Action recognition by matching clustered trajectories of motion vectors},
  booktitle = {Proc. 8th International Conference on Computer Vision Theory and Applications},
  year      = {2013},
  pages     = {112--117},
  address   = {Barcelona, Spain},
  month     = {February}
}
  • 2012

    A fully robust framework for MAP image super-resolution

    M. Vrigkas, C. Nikou, L.P. Kondi
    Conference Paper19th IEEE International Conference on Image Processing, pp. 2225-2228, Orlando, FL, September 30-October 3 2012

Abstract

In this work, we propose an adaptive M-estimation scheme for robust image super-resolution. The proposed algorithm relies on a maximum a posteriori (MAP) framework and addresses the presence of outliers in the low resolution images. Moreover, apart from the robust estimation of the high resolution image, the contribution of the method is twofold: (i) the robust computation of the regularization parameters controlling the relative strength of the prior with respect to the data fidelity term and (ii) the robust estimation of the optimal step size in the update of the high resolution image. Experimental results demonstrate that integrating these estimations into a robust framework leads to significant improvement in the accuracy of the high resolution image.


BibTex

@inproceedings{MVrigkas_ICIP12,
  author    = {Michalis Vrigkas and Christophoros Nikou and Lisimachos P. Kondi},
  title     = {A fully robust framework for MAP image Super-Resolution},
  booktitle = {Proc. IEEE International Conference on Image Processing},
  year      = {2012},
  pages     = {2225--2228},
  address   = {Orlando, FL},
  month     = {September}
}
  • 2011

    On the improvement of image registration for high accuracy super-resolution

    M. Vrigkas, C. Nikou, L.P. Kondi
    Conference PaperIEEE International Conference on Acoustics, Speech and Signal Processing, pp. 981-984, Prague, Czech Republic, May 22-27 2011

Abstract

Accurate image registration plays a preponderant role in image super-resolution methods and in the related literature landmarkbased registration methods have gained increasing acceptance in this framework. However, their solution relies on point correspondences and on least squares estimation of the registration parameters necessitating further improvement. In this work, a maximum a posteriori scheme for image super-resolution is presented where the image registration part is accomplished in two steps. At first, the lowresolution images are registered by establishing correspondences between robust SIFT features. In the second step, the estimation of the registration parameters is fine-tuned along with the estimation of the high resolution image, in an iterative scheme, using the maximization of the mutual information criterion. Numerical results showed that the reconstructed image is consistently of higher quality than in standard MAP-based methods employing only landmarks.


BibTex

@inproceedings{MVrigkas_ICASSP11,
  author    = {Michalis Vrigkas and Christophoros Nikou and Lisimachos P. Kondi},
  title     = {On the improvement of image registration for high accuracy super-resolution},
  booktitle = {Proc. IEEE International Conference on Acoustics, Speech and Signal Processing},
  year      = {2011},
  pages     = {981--984},
  address   = {Prague, Czech Republic},
  month     = {May}
}

Research Interests

  • Virtual and Augmented Reality
  • Computer Vision
  • Image and Video Processing
  • Machine Learning
  • Deep Learning
  • Biometrics
  • Medical Image Analysis
  • Predictive Analytics

Datasets

  • FaceMask Dataset

    You can download the dataset here.

    The FaceMask Database consists of 4,866 images of people in diffrerent environments and situations. These images were manually collected from Google Images and annotated with LabelImg tool in YOLO format. They are divided into two categories containing people wearing face masks and people not wearing face masks. The Dataset is splited into 3 folders -the training folder, the validation folder, and the test folder.

    If you use this dataset, I would be grateful if you cite with one of the following related publications:

    Related Publications

    1. M. Vrigkas, E-A. Kourfalidou, M.E. Plissiti, C. Nikou, “FaceMask: A New Image Dataset for the Automated Identification of People Wearing Masks in the Wild.,” Sensors. 2022; 22(3):896. https://doi.org/10.3390/s22030896. [pdf] [External link] [bibtex]
  • Parliament Dataset

    You can download the dataset features here and the dataset videos here.

    The Parliament dataset is a collection of 228 video sequences, depicting political speeches in the Greek parliament, at a resolution of 320 × 240 pixels at 25 fps. All behaviors were recorded for 20 different subjects. The videos were acquired with a static camera and contain uncluttered backgrounds. The length of the video sequences is 250 frames. The video sequences were manually labeled with one of three behavioral labels: friendly (90 videos), aggressive (73 videos), or neutral (65 videos). The subjects express their opinion on a specific law proposal and they adjust their body movements and voice intensity level according to whether they agree with that or not.

    The dataset was annotated by two observers of Greek origin, who watched the videos independently and recorded their labels separately. Disagreement was resolved by a third observer. The observers were asked to categorize the videos with respect to the notions of kindness and aggressiveness according to a general perception of a political speech by a citizen with a Greek mentality as follows. (i) Subjects with large and abrupt body, head and hand movements and high speech signal amplitude are to be labeled as aggressive. This corresponds to statesmen who express strongly their disagreement with the topic discussed or a previous speech given by a political opponent. (ii) Subjects with very small variations in their motion and speech signal amplitude are to be labeled as neutral. This class includes standard political speeches only expressing a point of view without any strong indication (body motion or voice tone) of agreement or disagreement with the topic discussed. (iii) Subjects with large but smooth variations in the pose of their body and hands speaking with a normal speech signal amplitudes are to be labeled as friendly.

    If you use this dataset, I would be grateful if you cite with one of the following related publications:

    Related Publications

    1. M. Vrigkas, E. Kazakos, C. Nikou and I. A. Kakadiaris, “Human activity recognition using robust adaptive privileged probabilistic learning,” Pattern Analysis and Applications, pp. 1-18 January 2021. [pdf] [External link] [bibtex]
    2. M. Vrigkas, E. Kazakos, C. Nikou and I. A. Kakadiaris, “Inferring human activities using robust privileged probabilistic learning,” in Proc. IEEE International Conference on Computer Vision Workshops, pp. 2658-2665, Venice, Italy, October 22-29 2017. [pdf] [bibtex]
    3. M. Vrigkas, C. Nikou and I.A. Kakadiaris, “Identifying human behaviors using synchronized audio-visual cues,” IEEE Transactions on Affective Computing, vol. 8, no. 1, pp. 54-66, Jan.-March, 2017. [pdf] [Extrernal link] [bibtex]
    4. M. Vrigkas, C. Nikou and I.A. Kakadiaris, “Classifying behavioral attributes using conditional random fields,” in Proc. 8th Hellenic Conference on Artificial Intelligence, Lecture Notes in Computer Science, vol. 8445, pp. 95-104, Ioannina, Greece, May 15-17 2014. [pdf] [bibtex]

Source Code

At My Office

You can find me at my office located at University of Western Macedonia, Fourka Area, Kastoria, Greece.