Next Article in Journal
Habilitation of Executive Functions in Pediatric Congenital Heart Disease Patients through LEGO®-Based Therapy: A Quasi-Experimental Study
Previous Article in Journal
Smoking Status and Risk Awareness of Heated Tobacco Product Use among General Dental Practitioners Belonging to the Aichi Dental Association, Japan
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Automatic Identification of a Depressive State in Primary Care

1
Institute of Psychology, Chinese Academy of Sciences, Beijing 100107, China
2
Department of Psychology, University of Chinese Academy of Sciences, Beijing 101408, China
*
Author to whom correspondence should be addressed.
Healthcare 2022, 10(12), 2347; https://doi.org/10.3390/healthcare10122347
Submission received: 30 August 2022 / Revised: 4 November 2022 / Accepted: 19 November 2022 / Published: 22 November 2022

Abstract

:
The Center for Epidemiologic Studies Depression Scale (CES-D) performs well in screening depression in primary care. However, people are looking for alternatives because it screens for too many items. With the popularity of social media platforms, facial movement can be recorded ecologically. Considering that there are nonverbal behaviors, including facial movement, associated with a depressive state, this study aims to establish an automatic depression recognition model to be easily used in primary healthcare. We integrated facial activities and gaze behaviors to establish a machine learning algorithm (Kernal Ridge Regression, KRR). We compared different algorithms and different features to achieve the best model. The results showed that the prediction effect of facial and gaze features was higher than that of only facial features. In all of the models we tried, the ridge model with a periodic kernel showed the best performance. The model showed a mutual fund R-squared (R2) value of 0.43 and a Pearson correlation coefficient (r) value of 0.69 (p < 0.001). Then, the most relevant variables (e.g., gaze directions and facial action units) were revealed in the present study.

1. Introduction

Mental health is an important and growing public health concern among the general population. Considering that as time goes by, the severity of mild mental health problems may increase if they are not given timely attention or help [1], people try to integrate mental health services within primary care to provide extensive and timely screening of people’s mental states [2,3].
The Center for Epidemiologic Studies Depression Scale (CES-D) [4] is a commonly used depression screening instrument to identify depression among the general population [5]. The psychometric characteristics of the CES-D scale appear to be relatively constant across different groups [6,7]. Additionally, its effectiveness in primary care scenes has been proven many times [8,9]. Although the CES-D is a good match for depression screening in primary care, people usually use the PHQ-9 instead because the CES-D has too many items to screen for [10,11]. However, given the high sensitivity of CES-D [11,12], which is more important than specificity in primary care, a simple, but effective CES-D automatic identification system can be helpful.
People’s nonverbal behavior is related to their psychological state, and psychiatrists have always combined the patient’s nonverbal behavior to make a diagnosis [13,14,15]. Then, aided by emerging technology, researchers quantify nonverbal behavior and explore the relationship between nonverbal behavior and a depressive state [16], the therapeutic effect of depression [17], and the development of depression [18]. Recently, with the development of artificial intelligence technology, machine learning methods have been applied to a wider range of research fields [19,20]. An automatic recognition model has been proposed to realize the automatic recognition of depression through nonverbal behavior (such as by voice [21], face, etc.). Some researchers proposed a deep regression network of DepresNet to learn the depression representation with visual interpretation and realized the depression detection method based on facial image analysis [22,23]. In Ref. [24], researchers effectively recognized depression by fusing facial and voice data. Meanwhile, in Ref. [25], researchers used 3D facial expressions and spoken language to recognize depression based on the machine learning method.
Facial expression and gaze are thought to be associated with depression. Studies found that a depression group opens their eyes smaller, have a longer duration of blinking [26], and have less eye contact with others [27] compared with a general group or a cured group. Other studies used facial activities to effectively predict people’s depression levels [28,29,30]. Moreover, the integration of different features brings a better prediction effect than a single feature [31]. Considering that both facial activities and gaze are visual data that are easily accessible, using facial activities and gaze data together to predict CES-D scores may be a simple, yet effective method.
In summary, this study integrated facial activities and gaze data to establish an automatic recognition model of CES-D scores. We assume that the prediction model can predict CES-D scores effectively to indicate depressive states, and the model fusing facial activities and gaze is better than the model containing only facial activities. This study can be applied to primary care to simply identify depressive states.
In Section 2, we introduce the process of data acquisition, facial data acquisition, facial feature extraction methods, and model-building methods. In Section 3, the performance evaluation results of the depression recognition model are reported. The research results are analyzed and discussed in Section 4.

2. Method

2.1. Participants and Procedure

A total of 152 students and workers from the University of Chinese Academy of Sciences participated in this study, of which 79 were male and 73 were female. To participate in the study, participants had to meet three inclusion criteria: (1) participants needed to be at least 18 years old; (2) participants needed to be fluent in Mandarin; and (3) participants needed to be physically healthy and able to make normal facial activities.

2.2. Measures

To build a predictive model between facial expression videos and CES-D scores, this study designed a facial data collection experiment with a self-introductory situation. The self-introductory task made the subjects show more attitude toward themselves [32], and this is also common in primary healthcare. The data acquisition experiment was carried out according to the following steps in Figure 1: (1) First, the demographic information of each subject was recorded. (2) Then, the participants were asked to complete the CES-D scale. (3) During the experimental task (the self-introduction), the participants were recorded introducing themselves for one to three minutes using a high-definition camera (the frame rate was 30 Hz). The distance between the camera and the participant was controlled to be 3 m to exclude the influence of distance on the intensity of any facial movements. Meanwhile, participants were asked to stand in a bright place without direct sunlight to prevent overexposure or dark light.

2.3. Instruments

This study used the Center for Epidemiologic Studies Depression Scale (CES-D) [4] as a tool to identify depression among the general population. The scale contains 20 questions in total. Each question describes a possible or recent feeling, such as: “In the past week, I have been bothered by something that doesn’t bother me normally”, “In the past week, I don’t want to eat, and my appetite is poor”, etc. The score ranges from 20 to 80. The items primarily measure the affective and somatic aspects of depression. According to the criteria proposed by the original author, depression exists in individuals with CES-D scores of 36 or more [4].
OpenFace [33] is a state-of-the-art tool for analyzing human behavioral video data, and it can recognize the movements of 17 facial action units [34], as well as estimate 8 indices of eye gaze [35] for each video frame. The specific meaning of each visual feature is provided in Table 1.

2.4. Data Preprocessing

After obtaining the full-body videos, we first cut the videos to 500 px × 500 px facial videos containing the subjects’ whole face. Then, these facial videos were grayed to reduce irrelevant information, such as color.

2.5. Feature Extraction and Reduction

Seventeen facial features and eight gaze features for each frame were extracted from grayed facial videos using OpenFace. Then, we intercepted each feature data point from frame 300 to frame 2100, which took approximately 1 min. The purpose of only intercepting the middle data was to eliminate the preparation time before and after the self-introduction. After that, the “all features” pattern of tsfresh 0.18.0 [36] was used to extract 787 time-series characteristics from each facial expression and gaze. Thus far, we have obtained a feature file; the number of rows is the number of subjects (152), and the number of columns is the number of time-series characteristics (25 × 787). Finally, normalization was conducted to balance the range of characteristics.
To prevent overfitting, we had to select dozens of features from this great quantity of features, which was related to the sample size. We carried out feature reduction and feature selection. For feature reduction, we used principal component analysis (PCA) to extract 8 principal components from 787 time-series characteristics of each feature. Compared with reduction from the time-series characteristics of all features, extracting principal components separately helps to ensure that all facial or gaze features can be preserved after unsupervised reduction. For feature selection, we calculated the F values between 200 principal components (25 × 8) and CES-D scores, and then we selected the principal components with the largest F value. The F value is the importance of the correlation coefficients between principal components and CES-D scores, and the calculation process is shown in Formula 1. Because of the uncertainty of how many principal components could achieve the best effect, we output the top (20, 25, 30, 35, 40, 45, 50, 55, 60) principal components.
F = r 2 1 r 2 × d f
In Formula 1, r represents the Pearson correlation coefficients between CES-D scores and values of one principal component. df represents the degrees of freedom. The larger the F value is more relevant than the principal component is to CES-D scores.

2.5.1. Statistical Analysis

A regression model attached to L2 regularization and a kernel function was implemented in Python 3.7.6 [37], which was called kernel ridge regression (KRR). KRR is used for supervised learning problems where we use the multiple training features X_i (the selected principal components) to predict a target variable Y_i (the CES-D scores). Compared with other regression models, KRR has both the advantages of L2-norm regularization and the kernel function. L2-norm regularization can learn from all of the features and reduce the influence of outliers on increasing parameters to prevent overfitting [38]. Meanwhile, the kernel function can make the nonlinear relation map linear in a high-dimensional space for the principal components that are nonlinear with CES-D scores to resolve. Then, we compare the performance of KRR with linear ridge regression and kernel support vector regression (SVR).

2.5.2. Model Testing and Validation

The KRR model can utilize a 5-fold cross-validation method to split the “training set” and the “testing set”. This method can balance the false high or false low performance caused by sampling deviation. Validation of the algorithm in predicting CES-D scores was calculated using the mean absolute error (MAE), mean squared error (MSE), R squared (R2), and Pearson correlation coefficient (r). In this study, Mae and MSE were related to the difference between the true scores and the predicted scores. Thus, the smaller indices showed a better performance. R2 reflects the interpretation degree of the predicted scores to the true scores, and a larger R2 represents a better model. R is the correlation coefficient between the true scores and the predicted scores; thus, a larger correlation coefficient is expected.

3. Results

A total of 79 males (51.97%) and 73 females (48.03%) were involved in this study. Their average CES-D score was 31.44, and the standard deviation was 7.05. According to the criteria (CES-D score ≥ 36) proposed by the original author [4], 37 of them (24%) had depression.
The study developed a KRR model to simulate the association between facial expression, gaze, and depression. After adjusting the parameters, the Pearson correlation coefficient of the best-performing algorithm was 0.69 (MAE 4.73, MSE 33.62, R2 0.43), which represented a strong correlation between the predicted values and the true scores. The trend of evaluation scores (r, MAE, MSE, R2) changing with the number of features is shown in Figure 2. We found that the model performed best when the 40 most relevant principal components were input.
To obtain a better prediction model and make it more accurate for primary care, we compared two common algorithms attached with different kernel functions (Table 2). The results show that the ridge regression model with the periodic kernel function is the best.
Eyes and faces are generally considered to be important areas of depression expression. However, this study also compared depression recognition performance with face-only features and facial and gaze features. The results showed that the combination of facial expression and gaze was better for predicting depression in all four evaluation indicators (Table 3).
In the best model, the 40 most relevant principal components were used. The model may have been more explicable if we knew which features these principal components belonged to and which features were more relevant to the CES-D scores. Thus, we investigated the original features of these 40 principal components. Then, we classified these 40 principal components into their features and added their F values to obtain the correlation degree between these features and the CES-D scores. We rank these effective features by correlation degree, as shown in Figure 3.

4. Discussion

In this study, we tried a machine learning approach to identify depressive states that may be used in primary healthcare. The results proved our hypothesis after we found that using video data can effectively predict depressive states. In addition, the combination of facial expression and gaze is better than only facial features in screening for depression. Furthermore, we investigated the correlation between these features and the CES-D scores.
We collected CES-D scores to be predicted based on video data. Then, we used Mae, MSE, R2, and r to compare the performance of different algorithms, different categories of features, and different numbers of principal components. We found that the KRR model with 40-dimensional principal components had the best performance, which was inputted with both facial and gaze features. Compared with some previous studies on auto recognition of depression, the proposed model has a smaller Mae value [39,40] and a larger p-value [41]. This shows that the model performed well.
Consistent with previous studies on the nonverbal expression of depression [13,16], this study found that both facial activities and gaze were associated with a depressive state. Among facial activities, AU20 (lip stretcher) was found to be the most relevant to the CES-D scores. One potential explanation is that the articulation of depressed people is different from that of the general population [42], and articulation is related to lip stretching [43]. In addition, we noted that AU9 (Nose Wrinkler) and AU10 (Upper Lip Raiser) were also closely related to CES-D scores, which are also representative of facial activities of disgust [44]. Considering the self-introduction tasks, this is in line with the fact that depressed and self-disgust feelings emerged as closely co-occurring and consistent symptoms (≥80% of depressed patients) [32]. Among gaze activity, eye gaze direction vectors were relevant to the CES-D scores. Such a relationship was also found in the correlation between eye gaze direction vectors and schizophrenic symptom severity [45].
The results also showed that the predictive effect of spliced facial and gaze features was better than that of facial features alone. When individuals are depressed, their eyes are not fully opened, which may indicate fatigue or decreased interest [26], and their faces may express negative emotions. The combination of these two nonverbal behaviors can more comprehensively describe the visualization of a depressive state. This proves that it is better to analyze multiple features in videos. This is also consistent with previous multimodal studies [39,46], and multichannel information is helpful for depression identification.
This study identified reliable and accurate predictive models. The psychological reliability and validity test method is applied to the depression recognition model validation, and more evaluation indicators are used to evaluate the validity of the model. Our model provides information about the most important variables to predict depressive state in primary healthcare. This method can be combined with patient interrogation and does not need other operations. Thus, it is simpler and more repeatable. This method can also be combined with monitoring technology for large-scale depression monitoring in some special occupations.
Our research also has some shortcomings. First, because we do not use the recorder to collect the subjects’ pure voice, we cannot incorporate voice into the establishment of a multimodal model. In fact, voice can provide much information for depression recognition. Future research may consider integrating audio-visual behaviors to predict a depressive state in primary healthcare, as a voice sample can be easily obtained without additional equipment or tasks. Second, when fusing multimodal features, we only simply splice the features instead of applying some more complex feature fusion methods, such as calculating the inner product. Complex feature fusion methods can bring more data, but the amount of data is also extremely large. Third, in future research, we will try to further verify the effectiveness of our method on the public dataset.

Author Contributions

X.L. (First Author and Corresponding Author) contributed to the conception and design of the study and collected the data and developed the instrument; X.W. performed most of the experiment, wrote the manuscript with input from all authors, and performed part of the statistical analysis and helped draft a part of the text. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the Key Research Program of the Chinese Academy of Sciences (No. ZDRW-XH-2019-4) and supported by the Scientific Foundation of the Institute of Psychology, Chinese Academy of Sciences (No. E2CX4735YZ).

Institutional Review Board Statement

The studies involving human participants were reviewed and approved by the scientific research ethics committee of the Chinese Academy of Sciences Institute of Psychology (H15010). The patients/participants provided their written informed consent to participate in this study.

Informed Consent Statement

Informed consent was obtained from all subjects involved in the study.

Data Availability Statement

The datasets presented in this article are not readily available because raw data cannot be made public. If necessary, we can provide behavioral characteristic data. Requests to access the datasets should be directed to X.L.

Acknowledgments

We wish to thank all subjects for their participation in this study.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Fogel, J.; Eaton, W.W.; Ford, D.E. Minor depression as a predictor of the first onset of major depressive disorder over a 15-year follow-up. Acta Psychiatr. Scand. 2006, 113, 36–43. [Google Scholar] [CrossRef]
  2. Kates, N.; Mcpherson-Doe, C.; George, L. Integrating mental health services within primary care settings: The Hamilton Family Health Team. J. Ambul. Care Manag. 2011, 34, 174–182. [Google Scholar] [CrossRef]
  3. Thielke, S.; Vannoy, S.; Unützer, J. Integrating Mental Health and Primary Care. Prim. Care Clin. Off. Pract. 2007, 34, 571–592. [Google Scholar] [CrossRef]
  4. Radloff, L.S. The CES-D Scale: A Self-Report Depression Scale for Research in the General Population. Appl. Psychol. Meas. 1977, 1, 385–401. [Google Scholar] [CrossRef]
  5. Patten, S.B.; Burton, J.M.; Fiest, K.M.; Wiebe, S.; Bulloch, A.G.; Koch, M.; Dobson, K.S.; Metz, L.M.; Maxwell, C.J.; Jette, N. Validity of four screening scales for major depression in MS. Mult. Scler. J. 2015, 21, 1064–1071. [Google Scholar] [CrossRef]
  6. Devins, G.M.; Orme, C.M.; Binik, Y.M.; Frizzell, B.; Stam, H.J.; Pullin, W.M. Measuring depressive symptoms in illness populations: Psychometric properties of the Center for Epidemiologic Studies Depression (CES-D) Scale. Psychol. Health 1988, 2, 139–156. [Google Scholar] [CrossRef]
  7. Ross, C.E.; Mirowsky, J. Components of depressed mood in married men and women. The Center for Epidemiologic Studies’ Depression Scale. Am. J. Epidemiol. 1984, 119, 997–1004. [Google Scholar] [CrossRef]
  8. Lyness, J.M.; Noel, T.K.; Cox, C.; A King, D.; Conwell, Y.; Caine, E.D. Screening for depression in elderly primary care patients. A comparison of the Center for Epidemiologic Studies-Depression Scale and the Geriatric Depression Scale. Arch. Intern. Med. 1997, 157, 449–454. [Google Scholar] [CrossRef]
  9. Pickard, A.S.; Dalal, M.R.; Bushnell, D.M. A Comparison of Depressive Symptoms in Stroke and Primary Care: Applying Rasch Models to Evaluate the Center for Epidemiologic Studies-Depression Scale. Value Health 2006, 9, 59–64. [Google Scholar] [CrossRef] [Green Version]
  10. Milette, K.; Hudson, M.; Baron, M.; Thombs, B.D. Comparison of the PHQ-9 and CES-D depression scales in systemic sclerosis: Internal consistency reliability, convergent validity and clinical correlates. Rheumatology 2010, 49, 789–796. [Google Scholar] [CrossRef]
  11. Zahra, Y.; Maryam, M.; Anna, J.; Baradaran, H.R.; Khamseh, M.E.; Mojtaba, M. Comparison of the CES-D and PHQ-9 depression scales in people with type 2 diabetes in Tehran, Iran. BMC Psychiatry 2011, 11, 61. [Google Scholar]
  12. Joode, J.; Dijk, S.; Walburg, F.S.; Bosmans, J.E.; Adriaanse, M.C. Diagnostic accuracy of depression questionnaires in adult patients with diabetes: A systematic review and meta-analysis. PLoS ONE 2019, 14, e0218512. [Google Scholar] [CrossRef] [Green Version]
  13. Jones, I.H.; Pansa, M. Some nonverbal aspects of depression and schizophrenia occurring during the interview. J. Nerv. Ment. Dis. 1979, 167, 402–409. [Google Scholar] [CrossRef]
  14. Prkachin, K.M. Nonverbal communication deficits and response to performance feedback in depression. J. Abnorm. Psychol. 1977, 86, 224–234. [Google Scholar] [CrossRef]
  15. Waxer, P.H. Therapist training in nonverbal communication I: Nonverbal cues for depression. J. Clin. Psychol. 1974, 30, 215–218. [Google Scholar] [CrossRef]
  16. Mergl, R.; Mavrogiorgou, P.; Hegerl, U.; Juckel, G. Kinematical analysis of emotionally induced facial expressions: A novel tool to investigate hypomimia in patients suffering from depression. J. Neurol. Neurosurg. Psychiatry 2005, 76, 138–140. [Google Scholar] [CrossRef]
  17. Ulrich, G.; Harms, K. A video analysis of the non-verbal behaviour of depressed patients before and after treatment. J. Affect. Disord. 1985, 9, 63–67. [Google Scholar] [CrossRef]
  18. Geerts, E.; Bouhuys, A.L.; Van, D. Attunement of non-verbal behaviour between depressed patients and a psychiatrist at admission is related to persistence of depression. Acta Neuropsychiatr. 1995, 7, 64–66. [Google Scholar] [CrossRef]
  19. Mavrogiorgou, A.; Kiourtis, A.; Kyriazis, D.; Themistocleous, M. A comparative study in data mining: Clustering and classification capabilities. In European, Mediterranean, and Middle Eastern Conference on Information Systems; Springer: Cham, Switzerland, 2017; pp. 82–96. [Google Scholar]
  20. Khanam, J.J.; Foo, S.Y. A comparison of machine learning algorithms for diabetes prediction. ICT Express 2021, 7, 432–439. [Google Scholar] [CrossRef]
  21. Tasnim, M.; Stroulia, E. Detecting Depression from Voice. In Proceedings of the Canadian Conference on Artificial Intelligence, Canadian AI 2019, Kingston, ON, Canada, 28–31 May 2019. [Google Scholar]
  22. Zhou, X.; Jin, K.; Shang, Y.; Guo, G. Visually interpretable representation learning for depression recognition from facial images. IEEE Trans. Affect. Comput. 2018, 11, 542–552. [Google Scholar] [CrossRef]
  23. Zhou, X.; Wei, Z.; Xu, M.; Qu, S.; Guo, G. Facial depression recognition by deep joint label distribution and metric learning. IEEE Trans. Affect. Comput. 2020, 13, 1605–1618. [Google Scholar] [CrossRef]
  24. Jan, A.; Meng, H.; Gaus, Y.F.B.A.; Zhang, F. Artificial Intelligent System for Automatic Depression Level Analysis Through Visual and Vocal Expressions. IEEE Trans. Cogn. Dev. Syst. 2018, 10, 668–680. [Google Scholar] [CrossRef]
  25. Haque, A.; Guo, M.; Miner, A.S.; Fei-Fei, L. Measuring depression symptom severity from spoken language and 3D facial expressions. arXiv 2018, arXiv:1811.08592. [Google Scholar]
  26. Alghowinem, S.; Goecke, R.; Wagner, M.; Parker, G.; Breakspear, M. Eye Movement Analysis for Depression Detection. In Proceedings of the 2013 IEEE International Conference on Image Processing, Melbourne, VIC, Australia, 15–18 September 2013. [Google Scholar]
  27. Hinchliffe, M.K.; Lancashire, M.; Roberts, F.J. A Study of Eye-Contact Changes in Depressed and Recovered Psychiatric Patients. Br. J. Psychiatry 1971, 119, 213–215. [Google Scholar] [CrossRef] [Green Version]
  28. Douglas, K.M.; Porter, R.J. Recognition of disgusted facial expressions in severe depression. Br. J. Psychiatry 2010, 197, 156–157. [Google Scholar] [CrossRef] [Green Version]
  29. Scoralick, F.M.; Piazzolla, L.P.; Camargos, E.F.; Dias Freitas, M.P.; Guimaraes, R.M.; Laiana, C. Facial Expression May Indicate Depression in Older Adults. J. Am. Geriatr. Soc. 2012, 60, 2371–2373. [Google Scholar] [CrossRef]
  30. Zhou, X.; Huang, P.; Liu, H.; Niu, S. Learning content-adaptive feature pooling for facial depression recognition in videos. Electron. Lett. 2019, 66, 648–650. [Google Scholar] [CrossRef]
  31. Alghowinem, S.; Goecke, R.; Wagner, M.; Epps, J.; Hyett, M.; Parker, G.; Breakspear, M. Multimodal Depression Detection: Fusion Analysis of Paralinguistic, Head Pose and Eye Gaze Behaviors. IEEE Trans. Affect. Comput. 2018, 9, 478–490. [Google Scholar] [CrossRef]
  32. Zahn, R.; Lythe, K.E.; Gethin, J.A.; Green, S.; Deakin, J.F.W.; Young, A.H.; Moll, J. The role of self-blame and worthlessness in the psychopathology of major depressive disorder. J. Affect. Disord. 2015, 186, 337–341. [Google Scholar] [CrossRef]
  33. Baltrusaitis, T.; Zadeh, A.; Lim, Y.C.; Morency, L.P. OpenFace 2.0: Facial Behavior Analysis Toolkit. In Proceedings of the IEEE International Conference on Automatic Face & Gesture Recognition, Xi’an, China, 15–19 May 2018. [Google Scholar]
  34. Baltrusaitis, T.; Mahmoud, M.; Robinson, P. Cross-dataset learning and person-specific normalisation for automatic Action Unit detection. In Proceedings of the IEEE International Conference & Workshops on Automatic Face & Gesture Recognition, Ljubljana, Slovenia, 4–8 May 2015. [Google Scholar]
  35. Wood, E.; Baltrusaitis, T.; Zhang, X.; Sugano, Y.; Robinson, P.; Bulling, A. Rendering of Eyes for Eye-Shape Registration and Gaze Estimation. In Proceedings of the IEEE International Conference on Computer Vision, Santiago, Chile, 7–13 December 2015. [Google Scholar]
  36. Christ, M.; Braun, N.; Neuffer, J.; Kempa-Liehr, A.W. Time Series FeatuRe Extraction on basis of Scalable Hypothesis tests (tsfresh–A Python package). Neurocomputing 2018, 307, 72–77. [Google Scholar] [CrossRef]
  37. Pedregosa, F.; Varoquaux, G.; Gramfort, A.; Michel, V.; Thirion, B.; Grisel, O.; Blondel, M.; Prettenhofer, P.; Weiss, R.; Dubourg, V.; et al. Scikit-learn: Machine Learning in Python. J. Mach. Learn. Res. 2011, 12, 2825–2830. [Google Scholar]
  38. Chen, S.; Geng, C.; Gu, R. An efficient L2-norm regularized least-squares temporal difference learning algorithm. Knowl.-Based Syst. 2013, 45, 94–99. [Google Scholar] [CrossRef]
  39. He, L.; Jiang, D.; Sahli, H. Multimodal depression recognition with dynamic visual and audio cues. In Proceedings of the International Conference on Affective Computing & Intelligent Interaction, Xi’an, China, 21–24 September 2015. [Google Scholar]
  40. Wang, Z.; Chen, L.; Wang, L.; Diao, G. Recognition of Audio Depression Based on Convolutional Neural Network and Generative Antagonism Network Model. IEEE Access 2020, 8, 101181–101191. [Google Scholar] [CrossRef]
  41. Takahashi, Y.; Ueki, M.; Yamada, M.; Tamiya, G.; Motoike, I.N.; Saigusa, D.; Sakurai, M.; Nagami, F.; Ogishima, S.; Koshiba, S.; et al. Improved metabolomic data-based prediction of depressive symptoms using nonlinear machine learning with feature selection. Transl. Psychiatry 2020, 10, 157. [Google Scholar] [CrossRef]
  42. Flint, A.J.; Black, S.E.; Campbell-Taylor, I.; Gailey, G.F.; Levinton, C. Abnormal speech articulation, psychomotor retardation, and subcortical dysfunction in major depression. J. Psychiatr. Res. 1993, 27, 309–319. [Google Scholar] [CrossRef]
  43. Tang, L.Y.; Hannah, B.; Jongman, A.; Sereno, J.; Wang, Y.; Hamarneh, G. Ghassan Hamarneh; Examining visible articulatory features in clear and plain speech. Speech Commun. 2015, 75, 1–13. [Google Scholar] [CrossRef]
  44. Krumhuber, E.G.; Küster, D.; Namba, S.; Skora, L. Human and machine validation of 14 databases of dynamic facial expressions. Behav. Res. Methods 2021, 53, 686–701. [Google Scholar] [CrossRef]
  45. Vail, A.K.; Baltrušaitis, T.; Pennant, L.; Liebson, E.; Baker, J.; Morency, L.P. Visual Attention in Schizophrenia: Eye Contact and Gaze Aversion during Clinical Interactions. In Proceedings of the 2017 Seventh International Conference on Affective Computing and Intelligent Interaction, San Antonio, TX, USA, 23–26 October 2017; pp. 490–497. [Google Scholar]
  46. Chao, L.; Tao, J.; Yang, M.; Li, Y. Multi Task Sequence Learning for Depression Scale Prediction from Video. In Proceedings of the 2015 International Conference on Affective Computing and Intelligent Interaction, Xi’an, China, 21–24 September 2015. [Google Scholar]
Figure 1. The overall process of the experiment.
Figure 1. The overall process of the experiment.
Healthcare 10 02347 g001
Figure 2. The trend of evaluation scores (r, MAE, MSE, R2) changing with the number of features. Smaller Mae and MSE scores represent that the predicted scores are closer to the real scores, while larger R2 and r values represent that the predicted scores are more related to the real scores.
Figure 2. The trend of evaluation scores (r, MAE, MSE, R2) changing with the number of features. Smaller Mae and MSE scores represent that the predicted scores are closer to the real scores, while larger R2 and r values represent that the predicted scores are more related to the real scores.
Healthcare 10 02347 g002
Figure 3. The ranking of the correlation degree between features and CES-D scores. A correlation degree is the sum of F values of selected principal components of a feature and CES-D score.
Figure 3. The ranking of the correlation degree between features and CES-D scores. A correlation degree is the sum of F values of selected principal components of a feature and CES-D score.
Healthcare 10 02347 g003
Table 1. The specific meaning of visual features.
Table 1. The specific meaning of visual features.
Visual Features Index Meaning
Facial activities AU01_r Inner Brow Raiser
AU02_r Outer Brow Raiser
AU04_r Brow Lowerer
AU05_r Upper Lid Raiser
AU06_r Cheek Raiser
AU07_r Lid Tightener
AU09_r Nose Wrinkler
AU10_r Upper Lip Raiser
AU12_r Lip Corner Puller
AU14_r Dimpler
AU15_r Lip Corner Depressor
AU17_r Chin Raiser
AU20_r Lip stretcher
AU23_r Lip Tightener
AU25_r Lips part
AU26_r Jaw Drop
AU45_r Blink
Eye_gaze gaze_0_x Eye gaze direction vector for the left eye in the y-axis direction
gaze_0_y Eye gaze direction vector for the left eye in the z-axis direction
gaze_0_z Eye gaze direction vector for the left eye in the x-axis direction
gaze_1_x Eye gaze direction vector for the right eye in the y-axis direction
gaze_1_y Eye gaze direction vector for the right eye in the z-axis direction
gaze_1_z Eye gaze direction vector for the right eye in the x-axis direction
gaze_angle_x Eye gaze direction in radians averaged for both eyes in the x-axis direction
gaze_angle_y Eye gaze direction in radians averaged for both eyes in the y-axis direction
AU represents action unit; x, y, and z represent three-dimensional coordinates, respectively.
Table 2. Performance comparison of different models.
Table 2. Performance comparison of different models.
Algorithms Feature Number MAE MSE R2 r
KRR 40 5.413 40.948 0.305 0.628 ***
KRR-Periodic 40 4.731 33.620 0.429 0.694 ***
SVR-linear 45 4.489 34.342 0.417 0.688 ***
SVR-Periodic 45 4.488 34.328 0.418 0.692 ***
SVR-RBF 40 4.736 40.403 0.315 0.601 ***
*** p < 0.001; Bold represents the best result.
Table 3. Performance comparison of different characteristics.
Table 3. Performance comparison of different characteristics.
Feature Feature Number MAE MSE R2 R
Facial 35 4.957 36.098 0.388 0.655 ***
Facial & gaze 40 4.731 33.619 0.429 0.694 ***
*** p < 0.001.
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Liu, X.; Wang, X. Automatic Identification of a Depressive State in Primary Care. Healthcare 2022, 10, 2347. https://doi.org/10.3390/healthcare10122347

AMA Style

Liu X, Wang X. Automatic Identification of a Depressive State in Primary Care. Healthcare. 2022; 10(12):2347. https://doi.org/10.3390/healthcare10122347

Chicago/Turabian Style

Liu, Xiaoqian, and Xiaoyang Wang. 2022. "Automatic Identification of a Depressive State in Primary Care" Healthcare 10, no. 12: 2347. https://doi.org/10.3390/healthcare10122347

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop