Can acute suicidality be predicted by Instagram data? Results from qualitative and quantitative language analyses
Rebecca C. Brown aff001; Eileen Bendig aff002; Tin Fischer aff003; A. David Goldwich aff004; Harald Baumeister aff002; Paul L. Plener aff001
Authors place of work:
University of Ulm, Department of Child and Adolescent Psychiatry and Psychotherapy, Ulm, Germany
aff001; University of Ulm, Department of Clinical Psychology and Psychotherapy, Ulm
aff002; Independent Contributor, Freelancing Data Journalist, Berlin, Germany
aff003; Independent Contributor, Freelancing Software Developer, Berlin, Germany
aff004; Medical University of Vienna, Department for Child and Adolescent Psychiatry, Vienna, Austria
Published in the journal:
PLoS ONE 14(9)
Social media has become increasingly important for communication among young people. It is also often used to communicate suicidal ideation.
To investigate the link between acute suicidality and language use as well as activity on Instagram.
A total of 52 participants, aged on average around 16 years, who had posted pictures of non-suicidal self-injury on Instagram, and reported a lifetime history of suicidal ideation, were interviewed using Instagram messenger. Of those participants, 45.5% reported suicidal ideation on the day of the interview (acute suicidal ideation). Qualitative text analysis (software ATLAS.ti 7) was used to investigate experiences with expressions of active suicidal thoughts on Instagram. Quantitative text analysis of language use in the interviews and directly on Instagram (in picture captions) was performed using the Linguistic Inquiry and Word Count software. Language markers in the interviews and in picture captions, as well as activity on Instagram were added to regression analyses, in order to investigate predictors for current suicidal ideation.
Most participants (80%) had come across expressions of active suicidal thoughts on Instagram and 25% had expressed active suicidal thoughts themselves. Participants with acute suicidal ideation used significantly more negative emotion words (Cohen’s d = 0.66, 95% CI: 0.088–1.232) and words expressing overall affect (Cohen’s d = 0.57, 95% CI: 0.001–1.138) in interviews. However, activity and language use on Instagram did not predict acute suicidality.
While participants differed with regard to their use of language in interviews, differences in activity and language use on Instagram were not associated with acute suicidality. Other mechanisms of machine learning, like identifying picture content, might be more valuable.
Medicine and health sciences – Mental health and psychiatry – Suicide – Biology and life sciences – Neuroscience – Cognitive science – Cognitive psychology – Language – Psychology – Emotions – Social sciences – Sociology – Communications – Social communication – Social media – Linguistics – Cognitive linguistics – Psycholinguistics – Computer and information sciences – Network analysis – Social networks – People and places – Population groupings – Age groups – Children – Adolescents – Families
Suicide is the second leading cause of death among adolescents and young adults according to the World Health Organization . Especially suicidal ideation and suicide attempts reach high prevalence rates among adolescents. In a large study comprising 17 European countries, around one third of all adolescents reported lifetime suicidal ideation , with slightly higher rates in German school-based populations of 36.4–39.4% [3,4] and similar rates in first year college students . A concerning 7–9% of German adolescents also report a lifetime history of suicide attempts [3,4,6].
Social media, of which Instagram is the most popular platform among adolescents [7,8], has become a fundamental channel for social interaction for adolescents . Adolescents use social media as an essential communication strategy, disclosing information by generating, obtaining and sharing content [9,10]. The great use of social media among adolescents enables new approaches and perspectives to investigate suicide, as they provide big data sets of individual content [11–14], which is not influenced by laboratory settings and allows for exploration of everyday life communication. While several risk factors for suicidality have been described, prediction of suicide has not clearly progressed within the last decades, thus creating a need for new avenues, such as machine-learning based algorithms [15–17]. The analysis of social media behavior (e.g., posting pictures, connecting with others) and linguistic features of generated content (e.g., user posts) has been used to predict depression  and suicidal ideation .
Expression of acute suicidal thoughts is also a concerning phenomenon on social media, to which large providers like Facebook reacted by offering tools to help users who post online about their thoughts or plans of suicide (https://www.facebook.com/help/contact/305410456169423). In Chinese social media, Fu et al.  investigated responses to self-presented suicide attempts and expression of acute suicidal thoughts and found that written responses of other users often expressed caring or empathy, although cynical comments or shocked reactions were also common. However, little is known about how young people who express acute suicidal thoughts in social media perceive those reactions.
Language use and suicidality
There are a number of cognitive and behavioral changes that have been described in users which are progressing towards verbalizing suicidal ideation. De Choudhury et al.  investigated posts and comments in mental health-related online forums. They designed a prediction framework that incorporates linguistic features (e.g., first person pronouns), linguistic structure (e.g. readability index) and interactional patterns (e.g., number of posts and comments). The framework was used to predict the individual risk of undergoing shifts from talking about general mental health issues to expressing suicidal ideation. Different linguistic features within this framework, like heightened self-attentional focus and poor interaction with the community characterized shifts from mere discussion of mental health issues to expressing suicidal ideation .
Current literature shows that it is possible to distinguish the level of concern among suicide related posts in social media using language-based classifiers [16,17,20,21]. This field of research is enabled by the availability of computer-based text analysis tools such as Linguistic Inquiry and Word Count (LIWC ). LIWC allows for a quantitative analysis of text with a focus on psychometric properties  and psychologically meaningful linguistic categories . In the context of suicidal ideation, relevant linguistic markers include heightened self-attentional focus [18,25], a rise in negative emotion words , and changes in cognitive wording [26,27]. Additionally, authors reported poor readability (FRE; ) (i.e. the ease with which texts can be read/understood by the reader) [29–32] to be a marker for developing suicidal ideation .
Aims of the current study
This is the first study to investigate language use on Instagram, one of the most prominent social media platforms among adolescents, concerning suicidality. The current study had two major aims: (1) to investigate the experience with expressions of acute suicidal thoughts on Instagram of young people using a qualitative approach and (2) to use LIWC as a quantitative approach to analyze differences in the language and Instagram activity of vulnerable young people (using qualitative interview data and captions on Instagram) with regard to their current suicidal ideation.
Regarding the first aim it was hypothesized that the majority of participants had come across expressions of acute suicidal thoughts on Instagram and that common reactions would include showing empathy or being shocked. We further hypothesized that expressions of acute suicidal thoughts would be met with an activation of a social help system on Instagram.
Regarding the second aim of the study, we hypothesized that in comparison to participants with past suicidal ideation only, participants with current suicidal ideation would use significantly more words related to a self-attentional focus (e.g. pronouns I, me, mine), negative emotions (e.g. fear, hate), their language would be defined by a high amount of cognitive words (e.g. confine, therefore) and a lower readability (FRE).
Participants were identified from a larger data-set investigating the occurrence of non-suicidal self-injury (NSSI) on Instagram . All pictures and user accounts associated with the 16 German hashtags most commonly related to featuring pictures of NSSI wounds (i.e. #ritzen (‘#cutting’) were downloaded at an hourly rate during four weeks in April 2016. For details on how those hashtags were identified see Brown et al. . During data collection it was recorded how many followers users had, how many other users they were following, how many pictures they had posted and how many comments each picture had received. After those four weeks of Instagram data collection, a total of N = 100 randomly chosen users from this data-set were approached via Instagram messenger and asked if they were willing to participate in an interview-based study.
Interviews were conducted on Instagram messenger using chats, which allowed participants to stay anonymous. The interviews were semi-structured and consisted of 33 questions about the participants’ experiences with NSSI and suicidality on Instagram. Additionally, socio-demographic variables (i.e. gender, age) were assessed. Acute suicidality was assessed by the question: “Are you currently thinking about, or planning to, end your life?”. Lifetime suicidality was assessed by the question “Have you ever sincerely thought about ending your life?” and lifetime suicide attempts were assessed by the question “Have you ever tried to end your life?”.
Of the N = 100 users on Instagram who were initially approached, N = 64 agreed to participate in a qualitative interview regarding their experiences with suicidality and NSSI on Instagram, of which N = 59 completed the interview. Of those participants, N = 52 reported a lifetime history of suicidal ideation. Data of these 52 participants (of which n = 5 did not want to answer questions on socio-demographic variables, but completed the interview) are presented in the present paper.
Qualitative data analyses
Semi-structured qualitative interviews were analyzed using the software ATLAS.ti 7. Two independent raters were thoroughly trained. Taking an example of three interviews, and paraphrasing them, categories from those paraphrased responses were generated in order to facilitate standardized ratings. Raters were trained continuously until the first five rated interviews showed very good inter-rater-reliability of a minimum of kappa = .80. The rest of the interviews were rated under ongoing supervision. Answers to the following questions were analyzed in the current study: “Have you ever announced a future suicide attempt on Instagram?” and “What were reactions of other users to this announcement?” as well as “Have you ever come across someone announcing suicide on Instagram?” and “What were reactions to this announcement?”. Inter-rater reliability ranged from substantial agreement (kappa = .61) for the category “Worried reactions to suicide announcements online” to perfect agreement (kappa = 1.0) for the categories “My account was reported to Instagram after my suicide announcement” and “Other users reacted shocked to my suicide announcement”. Whenever there was a disagreement between two ratings, an agreement was found between both raters and one of the authors of the paper (RB).
Quantitative data analyses
Qualitative interview data and captions on Instagram were analyzed using the Linguistic Inquiry and Word Count (LIWC) software. Instagram data from N = 52 users, who answered the question about acute suicidality (N = 25 participants reported acute suicidality, N = 27 no acute suicidality, but suicidal ideation in the past) were analyzed. Features measuring linguistic style were extracted.
For word count and linguistic analysis, the German dictionary of the LIWC was used . The LIWC is a computer-based text analysis software tool whose algorithms count words according to pre-defined criteria word categories . The resulting output file from this analysis contains information in percent (frequency of specific words in relation to the total number of words).
Furthermore, we used the Flesch-Reading-Ease index (FRE; [28,30]) to calculate the ease with which one can read or understand responses given by adolescents. The FRE is normalized to values between 0 and 100 with higher values indicating high reading ease (0–30 very low reading ease, comprehensible for academics; 30–50 low, 50–60 medium, 60–70 well understandable texts; 70–80 medium understandable, 80–90 easy and 90–100 very high reading ease, comprehensible for eleven-year old pupils) . The index is calculated using the average sentence length (ASL) and the average number of syllables per word (ASW) . Based on previous research on language use and suicidality, we chose the following specific variables from the LIWC and the automated readability index for analyses:
Category of negative emotion words (e.g., sad, angry, hatred)
Category of overall affect (emotion expression)
Category of cognitive words (e.g., because, understand, but)
Category of first person pronouns (I, my, mine)
Automated readability index (FREgerman = 180 − ASL − (58,5 * ASW))
As a further feature of the quantitative analysis, activity on Instagram (number of followers, number of following others, number of pictures posted, average number of comments per picture) within the past month was taken into account.
Statistical analyses were performed with R . Differences between participants with acute vs. non-acute suicidality were calculated using t-tests. Effect sizes (Cohen’s d) were calculated for significant differences. Based on previous research , logistic regression analysis was calculated with acute vs. non-acute suicidality as dependent variable and linguistic features (expression of negative affect, pronoun, cognitive mechanism, emotion expression, readability index) as well as activity on Instagram (number of followers on Instagram, number of users on Instagram they were following, number of pictures posted within the past month, or number of comments other users posted in response to those pictures on average per picture) as independent variables.
Of the N = 47 participants providing information on socio-demographic details, N = 41 (87%) stated to be female. Participants were on average 16.6 years old (SD = 0.96, min = 16, max = 20), with a median age of 16 (interquartile range = 16 to 17 years). Most participants attended school (N = 38, 80.9%), followed by 14.9% (N = 7) who went to university or were in professional training, and 4.3% (N = 2) who were currently unemployed.
All participants reported a lifetime-history of suicidal ideation and NSSI and 45.5% (N = 25) reported suicidal ideation on the day of the interview. Around half of all participants (53.8%, N = 28) reported a lifetime history of suicide attempts, and 23.5% reported a suicide attempt within the past month (N = 12).
Of all participants, 13 (25%) reported to have announced a planned suicide on Instagram. Asked about the reaction of other users to their announcement, the following themes emerged: “People offered help” (N = 6), “People tried to talk me out of it” (N = 8), “My account was reported to Instagram (N = 2), “People suggested a joined suicide” (N = 2), “People were shocked, sad, and devastated” (N = 1), “People encouraged me to commit suicide” (N = 1). Four participants reported ‘actual’ action by other users in reaction to their announcement (calling the police, telling parents), while all other perceived reactions remained purely online.
The majority (80.8%, N = 42) of all participants reported to have come across a expression of acute suicidal thoughts online. The following themes emerged when asking about reaction to those suicide announcements: “people were worried” (N = 30), “people showed empathy” (N = 5), “people encouraged the person to commit suicide” (N = 5), “people were helpless” (N = 2), “people reported the user to Instagram” (N = 2), “people expressed to not understand the person” (N = 2), “people identified with the person” (N = 1).
Participants with suicidal ideation on the day of the interview (‘acute suicidality’, AS) were compared to participants with past, but without current suicidal ideation (‘non-acute suicidality’, NAS).
Gender, age, occupational status, or lifetime attempted suicide were not associated with acute suicidality, and neither was activity on Instagram in the past four weeks. That is, number of followers on Instagram, number of users on Instagram they were following, number of pictures posted within the past month, or number of comments other users posted in response to those pictures on average per picture did not differ between the two groups (see Table 1).
Language analyses were calculated separately for language use in interviews and language use in captions on Instagram, respectively.
Results concerning language in interviews
On average, participants in the AS group used significantly more negative emotion words (M = 1.95, SD = .52) than participants in the NAS group (M = 1.57, SD = .63). For psychological processes, participants in the AS group used significantly more words reflective of emotion expression (M = 5.71, SD = .98), than participants in the NAS group (M = 5.13, SD = 1.05). All other differences were not significant (see Table 2).
In a step-wise logistic regression analysis combining language use in interviews and Instagram activity (except language in captions), only expression of negative emotion in the interviews was significantly associated with acute suicidality (Regression-coefficient B = 1.19, p = .029, OR = 3.28 (95% CI: 1.10 to 9.77), while we did not find associations with any of the other variables (pronoun, cognitive mechanism, emotion expression, number of followers, number of users following, pictures posted, average of comments per picture) (see Table 3).
The final model with negative emotion as associated variable was applied to the data to calculate the odds for AS of each individual. Based on a cut-off for the calculated odds, participants were defined as AS or NAS. Predicted AS was compared to reported AS by participants. Accuracy, sensitivity and specificity varied depending on cut-off: Maximal accuracy of 69.23% was achieved at a cut-off of 0.7 (sensitivity = 84%, specificity = 56.56%) (see Supporting Information, S1 Table, S1 Fig).
Results concerning language in captions
Participants in the acute suicidality group did not differ from participants in the non-acute suicidality group regarding their use of language in captions on Instagram in the four weeks prior to the interview (for details see Table 3). Prediction models with pronouns, negative emotion, cognitive mechanism, and emotion expression as factors were applied to language in the captions. No significant predictors could be found (all p>.05) (Table 3).
In this sample of young Instagram users with a lifetime history of suicidal ideation and NSSI, half of all participants had attempted suicide at least once and half of them were expressing acute suicidality on the day of the interview. These characteristics constitute this group of Instagram users as a very vulnerable at-risk group for suicidality. A quarter of all participants in this study reported to have expressed acute suicidal thoughts on Instagram. Reactions by other Instagram users indicated empathy, the activation of a social help system by other users offering help, trying to talk them out of it, or indicating sadness or shock. However, only in around a third of the cases, action was taken by informing the police or parents. Additionally, around 80% of all participants reported to have come across a suicide announcement online. Again, reactions of other Instagram users were mainly trying to offer help, by showing empathy, being worried, or reporting the user to Instagram. However, in both cases (either actively posting online about their thoughts or plans of suicide or coming across a suicide announcement online), a small percentage of participants reported incidents of other users encouraging the person to commit suicide or suggesting a joint suicide. No actions by Instagram (i.e. immediate deletion of the comment) were reported by the interviewed participants. These results are in line with former studies investigating responses to expressions of acute suicidal thoughts in social media . These should comprise the discussion of ethical questions and practical implications for future suicide prevention in social media  which could result in stricter legal requirements for social media providers regarding comments in the context of suicidality.
The detection of suicide risk through social media might be an opportunity for accurate and timely identification of acute suicidality , e.g. by using language variables for predictive analytics [19,37]. In this line, the second aim of this study was to investigate whether participants with current suicidal ideation would differ in their language use as well as in their Instagram activity from participants with non-acute suicidality in this German speaking sample. In order to control for situational biases of language used in captions on Instagram, data of qualitative interviews was also used to test for differences in language use. Overall, Instagram activity did not distinguish between participants with acute versus non-acute suicidality (neither language use in captions nor number of followers, pictures posted, comments, etc.). This is somewhat contrary to a study by De Choudhury et al. , who found language use and some activity markers (e.g. length of comments posted) in mental health forums on Reddit to be predictable of suicidal expressions. However, this might be due to the different nature of Reddit (where the main content is shared in language based discussion forums) and Instagram (where the main content is shared through pictures). According to our results, in this highly vulnerable group of participants posting pictures with NSSI on Instagram, automated linguistic analyses of data shared on Instagram might not be feasible to detect persons at risk. Approaches that apply machine learning tools to Instagram photos might be more promising in this context .
In language data obtained from qualitative interviews, significant differences between participants with AS vs. NAS could be found for negative emotion words and emotion expression with medium effect sizes. A binary logistic regression model revealed that for each unit (percentage) increase of negative emotion words, the odds for acute suicidality increased about 3 times. Differences regarding self-attentional focus and cognitive words were non-significant, but indicated the same direction as previous studies. Considering the homogeneity of the participants in this study regarding past NSSI and suicidal ideation (100% reported suicidal ideation and had posted pictures of NSSI on Instagram), and the rather small sample size, it is quite remarkable that effects found in previous studies seem to be robust in the current study and point towards a rather high validity of using interview data as compared to using captions in social media for language analyses. Even though interviews were conducted on Instagram messenger, language in those interviews was quite coherent (e.g. full sentences), while data in Instagram captions was usually quite fractured (e.g. changing between English and German in the middle of a sentence, heavy use of Emojis, use of incomplete sentences or single words). Interestingly, the overall use of emotional words in both groups seemed to be twice as high in captions as compared to interviews, while participants used around twice as many pronouns and cognitive words in interviews. This may have also biased linguistic calculations regarding captions. However, analysis of Instagram data should possibly include machine learning algorithms trained on picture- rather than on language data .
The calculated accuracy of our model indicates that in 69 percent of cases, participants could be correctly assigned to acute suicidality and non-acute suicidality based on language data in interviews (Meaningful prediction of acute suicidality based on captions was not possible (all p>.05). Although this prediction model is depending on the present sample, it achieved highly similar accuracy and slightly higher specificity and sensitivity as a supervised machine learning model of Nobles et al. , which correctly assigned participants in 70 percent of cases to depression and suicidality, with a sensitivity of 81% and a specificity of 56%. Further studies could use the prediction model (see supplementary material) to test the predictive value in other samples. Machine learning algorithms trained on larger datasets and incorporating additional information, like e.g. acoustic features [39,40] might be a fruitful approach to further investigate this finding.
Methodological limitations are related to the exploratory character of this study and the small sample size. Therefore, results of this study have to be interpreted with caution and cannot be generalized to other populations. Additionally, there was no “never suicidal/NSSI” control group to which we could have compared the average word use (e.g. Instagram users who had not posted pictures of NSSI). The LIWC might be a well-validated instrument to reveal information pertaining to psychological aspects , but a major problem of the software is that there are no standard values available to compare data to the general population. Furthermore, data of participants was completely anonymous, as interviews were conducted on Instagram messenger. Therefore, socio-demographic data cannot be validated. There may have also been a self-selection bias of mainly female adolescents choosing to participate in the current study.
Social media platforms are increasingly integrating mechanisms to detect suicidal posts and have started to implement automated help suggestions. With recent advances in machine learning and data mining, massive amounts of data can be used for predictive models, opening up new avenues for detection and prevention of suicidal behavior . For example, a recent study using Twitter data showed that users posting in ‘suicidal networks’ seem to be much more closely connected than other Twitter users. Those network analyses could be interesting for future investigations of Instagram data. However, ethical challenges when analyzing mental health data of social media users have to be taken into account , and data generated by at-risk individuals might not always be accurately pointing towards a risk for suicidality. Furthermore, our findings point to the fact that language based machine-learning algorithms might be limited in their ability to detect suicidality among users when used in mostly picture based social media, like Instagram. Other mechanisms of machine learning, which are also capable identifying picture content might be more helpful. According to the reports of participants of the current study, Instagram did not take active and effective measures to prevent suicide or possible contagion effects of suicidal ideation. Overall, social media providers need to be aware of at-risk users within their networks and need to take action when necessary. Mental health care providers should be aware of their patients’ social media use, address it, and discuss benefits and risks with their patients. Reading active suicidal thoughts online might be disturbing and should be addressed accordingly. Within the network of participants posting pictures of NSSI, universal preventative measures could be implemented, as a large number of those young people seem to be at risk for suicide. Furthermore, there seems to be a potential for using social media in a protective way, as it has recently been shown that fictional peer comments, can have an impact to positively change attitudes towards recovery from NSSI .
1. World Health Organization. Suicide data. 2018 [cited 2018 03/13/2018]; http://www.who.int/mental_health/prevention/suicide/suicideprevent/en/.
2. Kokkevi A, Rotsika V, Arapaki A, Richardson C. Adolescents’ self-reported suicide attempts, self-harm thoughts and their correlates across 17 European countries. J Child Psychol Psychiatry Allied Discip. 2012;53(4):381–9.
3. Donath C, Graessel E, Baier D, Bleich S, Hillemacher T. Is parenting style a predictor of suicide attempts in a representative sample of adolescents? BMC Pediatr. 2014 Dec;14(1):113.
4. Plener PL, Libal G, Keller F, Fegert JM, Muehlenkamp JJ. An international comparison of adolescent non-suicidal self-injury (NSSI) and suicide attempts: Germany and the USA. Psychol Med. 2009;39(9):1549–58. doi: 10.1017/S0033291708005114 19171079
5. Mortier P, Auerbach RP, Alonso J, Bantjes J, Benjet C, Cuijpers P, et al. Suicidal Thoughts and Behaviors Among First-Year College Students: Results From the WMH-ICS Project. J Am Acad Child Adolesc Psychiatry [Internet]. 2018 Apr 1 [cited 2018 Aug 29];57(4):263–273.e1. Available from: https://www.sciencedirect.com/science/article/pii/S0890856718300571 29588052
6. Brunner R, Kaess M, Parzer P, Fischer G, Carli V, Hoven CW, et al. Life-time prevalence and psychosocial correlates of adolescent direct self-injurious behavior: A comparative study of findings in 11 European countries. J Child Psychol Psychiatry. 2014 Apr;55(4):337–48. doi: 10.1111/jcpp.12166 24215434
7. Guimaraes T. Revealed: the demographic trends for every social network 2014. 2014.
8. Duggan M, Smith A. Social media update 2013. 2013.
9. Adrian M, Lyon AR. Social Media Data for Online Adolescent Suicide Risk Identification: Considerations for Integration Within Platforms, Clinics, and Schools. In: Technology and Adolescent Mental Health. 2018. p. 155–70.
10. Luxton DD, June JD, Kinn JT. Technology-Based Suicide Prevention: Current Applications and Future Directions. Telemed e-Health [Internet]. 2011 Jan [cited 2018 Jul 26];17(1):50–4. Available from: http://www.ncbi.nlm.nih.gov/pubmed/21214382
11. Dyson MP, Hartling L, Shulhan J, Chisholm A, Milne A, Sundar P, et al. A Systematic Review of Social Media Use to Discuss and View Deliberate Self-Harm Acts. Seedat S, editor. PLoS One [Internet]. 2016 May 18 [cited 2018 Jul 19];11(5):e0155813. Available from: http://dx.plos.org/10.1371/journal.pone.0155813 27191728
12. Marchant A, Hawton K, Stewart A, Montgomery P, Singaravelu V, Lloyd K, et al. A systematic review of the relationship between internet use, self-harm and suicidal behaviour in young people: The good, the bad and the unknown. Adolesc Ment Heal Serv Oxford Heal NHS Found Trust [Internet]. 2017 [cited 2018 May 9];1–26. Available from: http://journals.plos.org/plosone/article/file?id=10.1371/journal.pone.0181722&type=printable
13. Robinson J, Cox G, Bailey E, Hetrick S, Rodrigues M, Fisher S, et al. Social media and suicide prevention: A systematic review. Early Interv Psychiatry. 2016;10(2):103–21. doi: 10.1111/eip.12229 25702826
14. Luxton DD, June JD, Fairall JM. Social media and suicide: a public health perspective. Am J Public Health [Internet]. 2012 May [cited 2018 Jul 26];102 Suppl 2(Suppl 2):S195–200. Available from: http://www.ncbi.nlm.nih.gov/pubmed/22401525
15. Franklin JC, Ribeiro JD, Fox KR, Bentley KH, Kleiman EM, Huang X, et al. Risk factors for suicidal thoughts and behaviors: A meta-analysis of 50 years of research. Psychol Bull [Internet]. 2017 Feb [cited 2018 Aug 23];143(2):187–232. Available from: http://www.ncbi.nlm.nih.gov/pubmed/27841450
16. Nobles AL, Glenn JJ, Kowsari K, Teachman BA, Barnes LE. Identification of Imminent Suicide Risk Among Young Adults using Text Messages. Proc 2018 CHI Conf Hum Factors Comput Syst—CHI ‘18. 2018;1–11.
17. Braithwaite SR, Giraud-Carrier C, West J, Barnes MD, Hanson CL. Validating Machine Learning Algorithms for Twitter Data Against Established Measures of Suicidality. JMIR Ment Heal. 2016;3(2):e21.
18. De Choudhury M, Gamon M, Counts S, Horvitz E. Predicting Depression via Social Media. Proc Seventh Int AAAI Conf Weblogs Soc Media [Internet]. 2013;2:128–37. http://www.aaai.org/ocs/index.php/ICWSM/ICWSM13/paper/viewFile/6124/6351
19. De Choudhury M, Kiciman E, Dredze M, Coppersmith G, Kumar M. Discovering Shifts to Suicidal Ideation from Mental Health Content in Social Media. Proc SIGCHI Conf Hum factors Comput Syst CHI Conf [Internet]. 2016 May [cited 2018 May 9];2016:2098–110. Available from: http://www.ncbi.nlm.nih.gov/pubmed/29082385
20. Fu K, Cheng Q, Wong PWC, Yip PSF. Responses to a Self-Presented Suicide Attempt in Social Media. Crisis [Internet]. 2013 Nov 1 [cited 2018 Jul 20];34(6):406–12. Available from: http://www.ncbi.nlm.nih.gov/pubmed/23871954
21. O’Dea B, Wan S, Batterham PJ, Calear AL, Paris C, Christensen H. Detecting suicidality on Twitter. Internet Interv [Internet]. 2015 May 1 [cited 2018 May 9];2(2):183–8. Available from: https://www.sciencedirect.com/science/article/pii/S2214782915000160
22. Pennebaker JW, Boyd RL, Jordan K, Blackburn K. The Development and Psychometric Properties of LIWC2015. Austin, TX Univ Texas Austin. 2015;
23. Pennebaker JW, Mehl MR, Niederhoffer KG. Psychological Aspects of Natural Language Use: Our Words, Our Selves. Annu Rev Psychol [Internet]. 2003 Feb [cited 2018 Jul 25];54(1):547–77. Available from: http://www.ncbi.nlm.nih.gov/pubmed/12185209
24. Tausczik YR, Pennebaker JW. The Psychological Meaning of Words: LIWC and Computerized Text Analysis Methods. J Lang Soc Psychol. 2010;29(1):24–54.
25. Venek V, Scherer S, Morency L-P, Rizzo AS, Pestian J. Adolescent Suicidal Risk Assessment in Clinician-Patient Interaction. IEEE Trans Affect Comput. 2017;8(2):204–15.
26. Gunn J.F. & Lester D. (2015). Twitter postings and suicide: an analysis of the postings of a fatal suicide in the 24 hours prior to death. Suicidologi, 17(3).
27. Poulin C., Shiner B., Thompson P., Vepstas L., Young-Xu Y., Goertzel B., … McAllister T. (2014). Predicting the risk of suicide by analyzing the text of clinical notes. PloS one, 9(1), e85733 doi: 10.1371/journal.pone.0085733 24489669
28. Flesch R. The art of readable reading. New York: Harper & Brothers; 1949.
29. Temnikova I, Vieweg S, Castillo C. The Case for Readability of Crisis Communications in Social Media. In: Proceedings of the 24th International Conference on World Wide Web—WWW ‘15 Companion [Internet]. New York, New York, USA: ACM Press; 2015 [cited 2018 Jul 26]. p. 1245–50. http://dl.acm.org/citation.cfm?doid=2740908.2741718
30. DuBay WH. The Principles of Readability. Online Submiss [Internet]. 2004 Aug 25 [cited 2018 Jul 26]; https://eric.ed.gov/?id=ED490073
31. Pestian J, Nasrallah H, Matykiewicz P, Bennett A, Leenaars A. Suicide Note Classification using natural language processing: A content analysis. Biomed Inf insights. 2010;6(9):19–28.
32. Batterham PJ, Ftanou M, Pirkis J, Brewer JL, Mackinnon AJ, Beautrais A, Fairweather-Schmidt AK, Christensen H. Psychological Assessment. 2015; 27(2):501–512.
33. Brown RC, Plener PL. Non-suicidal Self-Injury in Adolescence. Curr Psychiatry Rep [Internet]. 2017 Mar [cited 2018 Apr 30];19(3):20. Available from: http://www.ncbi.nlm.nih.gov/pubmed/28315191
34. Wolf M, Horn AB, Mehl MR, Haug S, Pennebaker JW, Kordy H. Computergest??tzte quantitative Textanalyse: ??quivalenz und Robustheit der deutschen Version des Linguistic Inquiry and Word Count. Diagnostica. 2008;54(2):85–98.
35. Amstad T. Wie verständlich sind unsere Zeitungen? Zürich; 1978.
36. R Core Team. R: A Language and Environment for Statistical Computing [Internet]. Vienna, Austria; 2017. https://www.r-project.org
37. Eichstaedt JC, Schwartz HA, Kern ML, Park G, Labarthe DR, Merchant RM, et al. Psychological Language on Twitter Predicts County-Level Heart Disease Mortality. Psychol Sci [Internet]. 2015 Feb 20 [cited 2018 Jul 24];26(2):159–69. Available from: http://journals.sagepub.com/doi/10.1177/0956797614557867 25605707
38. Reece AG, Danforth CM. Instagram photos reveal predictive markers of depression. EPJ Data Sci. 2017;6(1).
39. Pestian JP, Sorter M, Connolly B, Bretonnel Cohen K, McCullumsmith C, Gee JT, et al. A Machine Learning Approach to Identifying the Thought Markers of Suicidal Subjects: A Prospective Multicenter Trial. Suicide Life-Threatening Behav. 2017;47(1):112–21.
40. Venek V, Scherer S, Morency L-P, Rizzo AS, Pestian J. Adolescent Suicidal Risk Assessment in Clinician-Patient Interaction. IEEE Trans Affect Comput [Internet]. 2017;8(2):204–15. Available from: http://ieeexplore.ieee.org/document/7384418/
41. Conway M, O’Connor D. Social Media, Big Data, and Mental Health: Current Advances and Ethical Implications. Curr Opin Psychol [Internet]. 2016 Jun [cited 2018 Jul 25];9:77–82. Available from: http://www.ncbi.nlm.nih.gov/pubmed/27042689
42. Lewis SP, Seko Y, Joshi P. The impact of YouTube peer feedback on attitudes toward recovery from non-suicidal self-injury: An experimental pilot study. Digit Heal [Internet]. 2018 Jan 5 [cited 2018 Aug 23];4:205520761878049. Available from: http://journals.sagepub.com/doi/10.1177/2055207618780499