Abstract
Background
The working environment of a suicide prevention helpline requires high emotional and cognitive awareness from chat counselors. A shared opinion among counselors is that as a chat conversation becomes more difficult, it takes more effort and a longer amount of time to compose a response, which, in turn, can lead to writer’s block.
Objective
This study evaluates and then designs supportive technology to determine if a support system that provides inspiration can help counselors resolve writer’s block when they encounter difficult situations in chats with help-seekers.
Methods
A content-based recommender system with sentence embedding was used to search a chat corpus for similar chat situations. The system showed a counselor the most similar parts of former chat conversations so that the counselor would be able to use approaches previously taken by their colleagues as inspiration. In a within-subject experiment, counselors’ chat replies when confronted with a difficult situation were analyzed to determine if experts could see a noticeable difference in chat replies that were obtained in 3 conditions: (1) with the help of the support system, (2) with written advice from a senior counselor, or (3) when receiving no help. In addition, the system’s utility and usability were measured, and the validity of the algorithm was examined.
Results
A total of 24 counselors used a prototype of the support system; the results showed that, by reading chat replies, experts were able to significantly predict if counselors had received help from the support system or from a senior counselor (P=.004). Counselors scored the information they received from a senior counselor (M=1.46, SD 1.91) as significantly more helpful than the information received from the support system or when no help was given at all (M=–0.21, SD 2.26). Finally, compared with randomly selected former chat conversations, counselors rated the ones identified by the content-based recommendation system as significantly more similar to their current chats (β=.30, P<.001).
Conclusions
Support given to counselors influenced how they responded in difficult conversations. However, the higher utility scores given for the advice from senior counselors seem to indicate that specific actionable instructions are preferred. We expect that these findings will be beneficial for developing a system that can use similar chat situations to generate advice in a descriptive style, hence helping counselors through writer’s block.
Keywords: suicide prevention, content based recommender system, chat corpus, crisis line, sentence embedding, suicide, mental health
Introduction
Worldwide, helplines have been set up to help individuals who are struggling with suicidal thoughts. These helplines are a preventive service to reduce the suicidal ideation or behavior of help-seekers [1]. These help-seekers can contact trained volunteers and professionals (counselors) who can listen to them and assist them with their problems relating to suicide.
Historically, people have been able to contact these helplines over the telephone, but with the advent of the internet, chat services have become increasingly popular. Compared with telephone helplines, online chat helplines show approximately the same beneficial effects [2]. Help-seekers mention several reasons for using counseling through an online chat rather than a traditional phone call, such as privacy and the slow deliberate nature of online chatting [3-6]. In the Netherlands, the 113 Suicide Prevention service saw the number of conversations increase to more than 35,000 via telephone and more than 57,000 via online chat in 2018, an increase of 33% from 2017. However, this increase resulted in a higher need for counselors as well. Because of the difficult nature of crisis counseling, suicide prevention helplines often have difficulty retaining counselors [7].
Studies have indicated that technology can support chat line operators in executing cognitive tasks. For example, in the related field of commercial telephone and chat customer support, there are various supportive technologies developed for operators [8-10]. However, in computing research aimed at suicide prevention, most work focuses on the prediction and detection of suicidal behavior [11,12], while only a few studies have examined assisting online counselors; this could be beneficial, though. Salmi [13] has identified several difficulties that counselors encounter in their work. First, the counselor has to take in a large amount of information about the help-seeker. Here, counselors could be supported in understanding a help-seeker’s history without having to read large portions of transcripts. Dinakar et al [14], therefore, have created a support system prototype for text-based crisis counseling called Fathom. Fathom uses visualizations based on topic modeling to provide information at a glance. In comparison to a control interface without a visualization aspect, Fathom was preferred by counselors when eliciting a list of issues and a conversation summary. Another difficulty is that the counselor must be aware of the conversation quality. In this respect, Althoff et al [15] compared the chat conversations of more and less successful counselors with natural language processing techniques to discover the quality differences, defining actionable strategies to improve conversation quality. For example, they showed that more successful counselors spend a longer time exploring solutions, while less successful counselors spend more time defining the problems.
Finally, the complexity and severity of help-seekers’ situations may lead to writer’s block in counselors. Although not directly related to the suicide prevention domain, Isbister et al [16] have designed a helper agent for human-human interaction. When a conversation lags, the agent suggests topics for the conversation pair to talk about and, thereby, the agent is generally able to make positive contributions to the chat.
In situations where counselors experience writer’s block, a straightforward solution would be to approach a senior colleague for help. These senior counselors can read along and describe in as much detail as necessary how they would respond to the help-seeker. However, this requires availability and time from a colleague, and this is not always possible. Responding quickly is important in life-threatening situations, and counselors cannot always wait for somebody to become available. We also suspect that an approach such as suggesting topics to keep a conversation going [16] or providing a conversation summary [12] would not be optimal in difficult situations where counselors have to de-escalate a suicide-related crisis. This paper, therefore, presents a system that uses natural language processing techniques to provide support for counselors in difficult chat conversations. The system recommends parts of similar, previous chat situations for the counselor to draw inspiration from, which might be able to reduce their writer’s block. This paper also evaluates the designed support system by comparing it with 1) written, general advice from a senior counselor and 2) receiving no additional help during chats. The system’s usability and utility, along with the validity of the algorithm used, were also examined.
Methods
Design
We used a within-subject design to evaluate the impact and usefulness of similar chat situations that could be used as inspiration. In the study, the counselor wrote a chat reply to a simulated chat that was interrupted as a difficult situation. The counselor took part in 3 simulations: 1) the counselor received parts of similar chats from a support system, 2) the counselor received written advice from an experienced counselor, and 3) the counselor received no additional help. A questionnaire was used to measure the support system’s usability. Finally, we evaluated the validity of the similarity of the generated chats by testing the algorithm in a small additional experiment with a within-subject design.
The current study received ethical approval from the TU Delft University research ethics committee (id: 688). Before starting the data collection, the experimental setup was also preregistered on the Open Science Framework [17].
Recommender Support System
For the study, we developed a system recommending the transcripts of similar previous chat conversations to a counselor based on the content of the counselor’s current chat conversation. Figure 1 shows a chat window on the left and the support system interface on the right. The support system shows the top 10 most similar chat messages, which the counselor could click to read them in their entirety.
A corpus of chat conversations between help-seekers and counselors was used to find similar previous chat situations. We used the corpus from 113 Suicide Prevention in the Netherlands. This corpus contained 7 months of chats spanning from March 2018 to September 2018. The chat data were first filtered, removing all chats that had less than 20 interactions. In total, we used 17,773 chats. Furthermore, any special symbols in the messages were cleaned, and capital letters were replaced by lowercase letters.
Because the chats each contained multiple problems, we used a sliding window algorithm to scan for relevant chat segments instead of comparing complete chats. This algorithm created sets of chat messages, starting with the first 5 messages. The next set removed the first message in the window and added the sixth message; this process was repeated to create every possible set of 5 subsequent messages in a chat. The sliding window algorithm was then used to create the chat segments for the entire corpus.
We used an embedding algorithm to compute the similarity. For each chat segment, an embedding was created using smooth inverse frequency [18], which takes a weighted average of the word embeddings for each word in the text of the window corresponding to the inverse of the frequency of the word in the corpus. This resulted in less meaningful words receiving a lower weight. To create word embeddings, Mikolov et al [19] developed an algorithm dubbed Word2Vec, improving previous methods [20]. The word embeddings we used were obtained from the COOSTO Word2Vec model [21], a model developed using Dutch social media and blog posts. A window of 5 messages resulted in 1,286,659 embeddings, which were stored alongside the corresponding chat and window positions.
When a counselor in an ongoing conversation requested similar chat conversations, a single smooth inverse frequency embedding was created using the same steps as with the corpus, except only the last 5 messages of the ongoing conversation were used. This embedding was then compared with the corpus embeddings through a cosine similarity. Ten windows with the highest similarity were recommended to the counselor.
Difficult Chats
We used 6 chats for the experiment to cover several difficult situations: a situation where a help-seeker 1) was in a dangerous location and had withheld this from the counselor; 2) did not want to inform anybody in their environment of their suicidality because they felt like it would put a burden on others; 3) was afraid of people in their environment not understanding their problems; 4) tried to look for help but was not believed; 5) was excessively rude; and 6) had to contact a psychologist.
Participants
Counselor and expert recruitment, as well as conducting the experiment, happened at 113 Suicide Prevention. In total, 24 counselors participated. On average, the participants’ age was 27 years old, and 79% were female. Only counselors who were interns, volunteers, or trainees were eligible to participate. Each counselor met all the components and conditions of the evaluation.
Measures
The perceived utility was assessed with the following question: “How, in your opinion, did the extra information help you with coming up with your response?”. The counselors graded each support type on a fixed interval scale from –3 to 3, where –3 indicated the extra information was hindering, 0 indicated the information was neutral, and 3 indicated the information was useful.
To measure usability, the counselors were asked to fill out the System Usability Scale questionnaire [22]; this is a validated 10-item questionnaire with a 5-point scale ranging from “Strongly disagree” to “Strongly agree.”
To measure the validity of the algorithm, the counselors used a 7-point fixed interval scale to indicate how much they agreed with the following statement: “The problem in the matched chat is the same as the problem in the ongoing chat.” A score of 1 meant the counselor did not agree, whereas a score of 7 meant they did agree.
Procedure
The counselors used a test environment with simulated chats. The experiment consisted of 2 parts. Figure 2 shows a diagram of the procedure for the first part. Before the experiment, the counselor had 5 minutes to explore and familiarize themself with the support system.
Part 1 consisted of a simulated environment where the counselor read and reacted to 3 simulated chats, one after the other. The support information was contained in an extra tab called “Help.” Figure 3 shows each support type. Each counselor had the same amount of time to read the chat. To simulate a real situation, each counselor had a 2-minute window to reply. The counselor could not access the support tab before the 2-minute timer started. Directly after the counselor submitted their reply to a chat, they were asked to rate the utility of the support type. These steps were repeated for each condition. Therefore, the participants reacted to 3 chats in total. The chats, support types, and combinations were counterbalanced for the 24 participants. This part ended with the System Usability Scale questionnaire being used to measure the usability of the support system.
Part 2 recorded the measurements for evaluating the validity of the algorithm. Figure 4 shows a diagram of the procedure. The left side of the screen contained the transcript of an ongoing chat. The right side of the screen showed 10 chat segments. Half of these segments were randomly selected, and the other half was matched to the ongoing chat using the embedding algorithm. Below each of the segments was a fixed interval scale from 1 to 7 where the counselor rated the degree to which that chat segment related to the ongoing chat. To enhance generalization, the participants did this for the transcripts of 3 different ongoing chats. Therefore, in total, a participant rated 30 segments.
Data Preparation
Eight experts labeled the reply of the counselor with the type of help (condition) that the expert assumed that the counselor had received. To prevent expert bias, each expert judged all the counselor responses. Furthermore, a reliability analysis for the items of the System Usability Scale questionnaire showed an acceptable level of consistency, with a Cronbach alpha of .89. Therefore, the System Usability Scale items were compiled into a single score.
Analysis
The noticeable difference in counselor outputs was analyzed using generalized mixed-effects analyses [23] to predict the outcome variable support type based on the label the expert assigned to the counselor reply. The analyses were done by comparing 2 support type conditions at a time, thereby excluding the data from 1 of the 3 support type conditions. The models fitted on the remaining 2 conditions hence assumed a binomial distribution. Each model was compared with a null model that did not include an expert label as a fixed effect. Because the test was conducted 3 times, a Bonferroni correction [24] was used to set the significance threshold at .016. In addition, crossed random effects were used with random intercepts for the counselor and expert. Furthermore, for each support type, the utility ratings were analyzed using a one-sample t test to examine whether the rating deviated from the neutral zero score on the scale.
To examine the validity of the algorithm, a linear mixed-effects analysis was performed on the counselor’s rating of the similarity between the chat segment and the ongoing chat. As a two-level fixed effect the analysis included the recommendation method, that is, randomly selected versus selected by embedding algorithm. Furthermore, the ongoing chat was added as a three-level fixed variable because the quality of the suggestions was assumed to depend on the specific chat. As a random effect, the intercepts for counselors were used.
Anonymized data and R scripts are available online [25].
Results
Noticeable Difference in Counselor Output
Table 1 shows the effect of support type on the outcome measure of the expert label. The first row shows that the expert label significantly predicts the support type, when the data of no support condition was left out. In other words, the experts could tell the difference between replies given with the support system and replies with help from a senior counselor. Table 2 shows that when the expert labeled the counselor’s response as having received help from the support system, the counselor was 0.47 times less likely to have received the senior counselor support. This effect is further illustrated when looking at the confusion matrix of these conditions, as shown in Figure 5. However, no significant difference was found between the no support condition and any of the other conditions.
Table 1.
Outcomes data included in analysis | χ2 (df) | P value |
Support system and senior counselor written advice | 11.31 (2) | .004 |
No support and support system | 1.44 (2) | .49 |
No support and senior counselor written advice | 4.78 (2) | .09 |
Table 2.
Parameter | ORa | Standard error | z Value | P value |
Intercept | 1.29 | 0.16 | 1.54 | .12 |
Support system | 0.47 | 0.25 | –3.04 | .002 |
Senior counselor written advice | 1.01 | 0.28 | 0.049 | .96 |
aOR: odds ratio.
Utility
The results of the utility ratings are shown in Table 3. The mean score of the support system was –0.21 (SD 2.26) and did not significantly deviate from 0, indicating that there was neither a hindering nor a helping effect experienced by the counselors. However, the mean utility score of the written advice from a senior counselor was 1.46 (SD 1.91) and significantly deviated from 0. This suggests that the written advice was perceived as helpful. It is noteworthy that the support system had a high variance, suggesting that the counselor’s opinion on the utility was divided.
Table 3.
Support type | Mean (SD) | 95% CI | t df | P value |
Support system | –0.21 (2.26) | –0.84 to 0.43 | –0.6823 | .5 |
Counselor written advice | 1.46 (1.91) | 0.87 to 2.04 | 5.1723 | <.001 |
No support | –0.21 (0.95) | –0.62 to 0.2 | –1.0423 | .31 |
Usability
The mean score of the support system for the System Usability Scale questionnaire was 71, with a 95% confidence interval of 63-78. According to Bangor et al [26], this score can be classified as “good” based on an adjective rating scale.
Validity of the Algorithm
How the chat segments were selected (randomly vs by the embedding algorithm) significantly predicted the rating counselors gave on the chat segment’s similarity to the ongoing chat, β=.30, t(7.66), P<.001. This means that counselors could tell the difference between the random chats and those generated by the support system. The suggestions from the algorithm increased the similarity rating given by counselors from an average of 2.35 to an average of 3.42 (difference of 1.07).
Discussion and Conclusions
In the current study, we evaluated a prototype support system to assist chat counselors in suicide prevention helplines by providing inspiration from previous chats. The results show that counselors gave different answers depending on whether they received help from the support system or from a senior colleague. Upon inspection, the replies given by the counselors who received written advice from a senior colleague were, for the most part, copied directly and with little to no alterations made. Replies from counselors using the support system were more varied. This could be a possible explanation for the noticeable difference. However, we could not find a significant result for the no-help condition, which also had varied replies. Additionally, we observed that written advice from a senior counselor was given a significantly higher utility score than the other conditions; this suggests that the counselors value short actionable information that is highly accurate to the situation and that is given by someone with expertise. Gunaratne et al [27] have observed similar findings in their study on the effects of expert advice and social comparison on decision making for retirement savings; they showed that expert advice helped people make better decisions, whereas social comparison was seen as a useful mechanism to keep people from deviating too far from the mean and, hence, make safe decisions. However, both of these conditions outperformed a control condition where no additional information was provided.
The main contribution of our study is the idea of retrieving inspiration from a conversation corpus. Other support systems for chats [28-30] have used topics to assist the conversation. Compared with these methods, our approach for combating writer’s block in a counseling conversation is novel. Furthermore, an experimental design was used to compare this supportive technology with advice from a senior colleague, showing how the two differ.
Some limitations should be considered regarding the findings and their implications. We used chat transcripts of conversations with situations that previous counselors found difficult to evaluate. However, this might not cause writer’s block for every participant because not every counselor will have problems with the same situations. For writer’s block to occur naturally, the system would have to be tested in live chats. This was, however, not possible because of the ethical constraints of deploying an unevaluated system in a possible crisis situation. Furthermore, the specification, development, and evaluation were done in the context of counselors working at 113 Suicide Prevention in the Netherlands, with a limited number of counselors. The support system should also be tested in different helplines and with a larger sample size.
We have outlined two major directions for future work. First, the recommendation mechanism could be improved in different ways. This study, as well as other related works such as recommenders for creativity [31] and scientific writing [32], relies on topic modeling and bag-of-word models to find recommendations. Encoding text using attention-based models [33], such as BERT [34], have been shown to perform well on various natural language processing tasks, including semantic sentence similarity for conversation data [28]. These methods could be applied to improve the recommendations to find more relevant and similar examples, which we expect will increase the perceived utility. Additionally, curating the corpus can help denoise the dataset and improve the recommendations. This can also give counselors the knowledge that the information comes from a subset of quality chats, thereby acting on the persuasive principle of authority as outlined by Cialdini [29]. Lastly, there is also an opportunity to apply active learning methods by adding positive labels to the recommendations that the counselors interacted with or explicitly marked as useful [30].
Second, the findings show that the embedding algorithm found similar chats and that written advice from senior counselors had high utility. Compared to the Gunaratne et al study [27], the main difference to the setup of our study is that the social comparison condition provided information as an average; this indicates that refining the output of the support system recommendations to be more instructional could be a possible direction for improving the system. To combine both the extensive coverage of a chat corpus and the high utility of curated written advice, clustering could be used, that is, grouping similar chats together based on a similarity metric and curating the labels based on these clusters. Derrar [35] uses clustering to automate the annotation of customer service chat messages. A similar approach could be used to annotate the chat corpus to create a taxonomy of situations and advice, which then could emulate receiving written advice from a senior colleague. In other words, working together with experts, a set of advice could be formulated in advance for each specific situation. Next, a data-driven algorithm could be trained to classify chats according to categories of the taxonomy, consequently providing counselors with expert advice associated with the category and making the expert advice situation relevant. This approach would be most suitable for assisting counselors with frequently occurring tasks, as these would be the most likely cases to be included in the taxonomy. The focus of the support system might therefore shift from an inspiration source to a system that could reduce workload. Alternatively, the field of conversational information retrieval has explored multiple methods that could be applied to the task presented in this paper. For example, Qiu et al [36] combined both information retrieval methods and generation based models to create a chat bot trained using existing customer service chat logs. These techniques could potentially also be used to allow the system to generate proposal responses that counselors could consider using in their chats with help-seekers.
In conclusion, the current study shows a possible method to provide inspiration during chat counseling in a helpline for suicide prevention and how this supportive technology compares with human assistance. A support system may be a relief for counselors as they handle many cognitively difficult situations. In addition, supportive technology seems useful for helplines to better deal with busy periods, to provide a safety-net for junior counselors, and to help sustain counselors.
Acknowledgments
The authors would like to thank the counselors of 113 Suicide Prevention who participated in the experiment. The authors would also like to thank the floor managers of 113 Suicide Prevention who provided the senior counselor comments and helped select the difficult chat situations.
Footnotes
Conflicts of Interest: None declared.
References
- 1.Gould Madelyn S, Kalafat John, Harrismunfakh Jimmie Lou, Kleinman Marjorie. An evaluation of crisis hotline outcomes. Part 2: Suicidal callers. Suicide and Life-Threatening Behavior. 2007 Jun;37(3):338–52. doi: 10.1521/suli.2007.37.3.338. [DOI] [PubMed] [Google Scholar]
- 2.Mokkenstorm JK, Eikelenboom M, Huisman A, Wiebenga J, Gilissen R, Kerkhof AJFM, Smit JH. Evaluation of the 113Online Suicide Prevention Crisis Chat Service: Outcomes, Helper Behaviors and Comparison to Telephone Hotlines. Suicide and Life-Threatening Behavior. 2016 Aug 19;47(3):282–296. doi: 10.1111/sltb.12286. [DOI] [PubMed] [Google Scholar]
- 3.Evans WP, Davidson L, Sicafuse L. Someone to Listen: Increasing Youth Help-Seeking Behavior Through a Text-Based Crisis Line for Youth. J. Community Psychol. 2013 Mar 21;41(4):471–487. doi: 10.1002/jcop.21551. [DOI] [Google Scholar]
- 4.Suler John. The psychology of text relationships. Online Counseling (Second Edition) 2011:21–53. doi: 10.1016/B978-0-12-378596-1.00002-2. [DOI] [Google Scholar]
- 5.Dowling M, Rickwood D. Online Counseling and Therapy for Mental Health Problems: A Systematic Review of Individual Synchronous Interventions Using Chat. Journal of Technology in Human Services. 2013 Jan;31(1):1–21. doi: 10.1080/15228835.2012.728508. [DOI] [Google Scholar]
- 6.Fukkink R, Hermanns J. Counseling children at a helpline: chatting or calling? J. Community Psychol. 2009 Nov;37(8):939–948. doi: 10.1002/jcop.20340. [DOI] [Google Scholar]
- 7.Pratt Mary E. The future of volunteers in crisis hotline work. 2013. [2020-12-11]. http://d-scholarship.pitt.edu/16489/
- 8.Acorn Timothy L, Walden Sherry H. SMART: support management automated reasoning technology for compaq customer service. IAAI'92: Proceedings of the fourth conference on Innovative applications of artificial intelligence; The Fourth Conference on Innovative applications of artificial intelligence; 1992; San Jose, California. 1992. pp. 3–18. [Google Scholar]
- 9.Goldberg Randy G, Rosinski Richard R. Automated natural language understanding customer service system. Google patents. 1999. [2020-12-11]. https://patents.google.com/patent/US5895466A/en.
- 10.Ignacio Madrid R, Van Oostendorp H, Puerta Melguizo MC. The effects of the number of links and navigation support on cognitive load and learning with hypertext: The mediating role of reading order. Computers in Human Behavior. 2009 Jan;25(1):66–75. doi: 10.1016/j.chb.2008.06.005. [DOI] [Google Scholar]
- 11.Walsh CG, Ribeiro JD, Franklin JC. Predicting Risk of Suicide Attempts Over Time Through Machine Learning. Clinical Psychological Science. 2017 Apr 11;5(3):457–469. doi: 10.1177/2167702617691560. [DOI] [Google Scholar]
- 12.Coppersmith G, Leary R, Crutchley P, Fine A. Natural Language Processing of Social Media as Screening for Suicide Risk. Biomed Inform Insights. 2018 Aug 27;10:1178222618792860. doi: 10.1177/1178222618792860. https://journals.sagepub.com/doi/10.1177/1178222618792860?url_ver=Z39.88-2003&rfr_id=ori:rid:crossref.org&rfr_dat=cr_pub%3dpubmed. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 13.Salmi Salim. Context-based recommender system to provide cognitive support to online chat counsellors in the Helpline of 113 Suicide Prevention. 2019. Nov 26, [2020-12-11]. https://repository.tudelft.nl/islandora/object/uuid:cd6e857c-7241-48be-b71b-59371401a640.
- 14.Dinakar Karthik, Chen Jackie, Lieberman Henry, Picard Rosalind, Filbin Robert. Mixed-Initiative Real-Time Topic Modeling & Visualization for Crisis Counseling. IUI '15: Proceedings of the 20th International Conference on Intelligent User Interfaces; IUI'15 20th International Conference on Intelligent User Interfaces; 2015; Atlanta Georgia USA. 2015. [DOI] [Google Scholar]
- 15.Althoff T, Clark K, Leskovec J. Large-scale Analysis of Counseling Conversations: An Application of Natural Language Processing to Mental Health. TACL. 2016 Dec;4:463–476. doi: 10.1162/tacl_a_00111. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 16.Isbister Katherine, Nakanishi Hideyuki, Ishida Toru, Nass Cliff. Helper agent: designing an assistant for human-human interaction in a virtual meeting space. CHI '00: Proceedings of the SIGCHI conference on Human Factors in Computing Systems; CHI00: Human Factors in Computing Systems; 2000; The Hague, The Netherlands. 2000. Apr, [DOI] [Google Scholar]
- 17.Salmi Salim, Brinkman Willem-Paul. Evaluation inspiration support system for suicide crisis counseling. Open Science Framework. 2019. May 13, [2019-05-13]. https://osf.io/9gu2y.
- 18.Arora Sanjeev, Liang Yingyu, Ma Tengyu. A simple but tough-to-beat baseline for sentence embeddings. 5th International Conference on Learning Representations; 5th International Conference on Learning Representations; 2017; Toulon, France. 2017. [Google Scholar]
- 19.Mikolov Tomas, Sutskever Ilya, Chen Kai, Corrado Greg S, Dean Jeff. Distributed representations of words and phrases and their compositionality. Advances in neural information processing systems 26 (NIPS 2013); Advances in neural information processing systems 26; 2013; Lake Tahoe, Nevada, United States. 2013. pp. 3111–3119. [Google Scholar]
- 20.Baroni Marco, Dinu Georgiana, Kruszewski Germán. Don’t count, predict! A systematic comparison of context-counting vs. context-predicting semantic vectors. Proceedings of the 52nd Annual Meeting of the Association for Computational Linguistics; 2014 Association for Computational Linguistics; 2014; Baltimore, Maryland, USA. 2014. pp. 238–247. [DOI] [Google Scholar]
- 21.Nieuwenhuijse Alexander. COOSTO Dutch Word2Vec Model. GitHub. [2019-03-20]. https://github.com/coosto/dutch-word-embeddings.
- 22.Brooke John. SUS-A quick and dirty usability scale. In: Jordan Patrick W, Thomas Bruce, Weerdmeester Bernard A, McClelland Ian L., editors. Usability evaluation in industry. London: Taylor & Francis; 1996. pp. 189–194. [Google Scholar]
- 23.Bates D, Mächler M, Bolker B, Walker S. Fitting Linear Mixed-Effects Models Using lme4. J. Stat. Soft. 2015 Oct 07;67(1):1–48. doi: 10.18637/jss.v067.i01. [DOI] [Google Scholar]
- 24.Abdi Hervé. The Bonferonni and Šidák Corrections for Multiple Comparisons. In: Salkind Neil J., editor. Encyclopedia of Measurement and Statistics volume 3. Newbury Park, California, USA: SAGE Publishing; 2007. pp. 103–107. [Google Scholar]
- 25.Salmi S, Mérelle S, Gilissen R, Renske B, Brinkman WP. Data and analysis for the publication: Content-Based Recommender Support System for Counselors in a Suicide Prevention Chat Helpline: Design and Evaluation Study. 4TU.ResearchData; 2020. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 26.Bangor Aaron, Kortum Philip, Miller James. Determining What Individual SUS Scores Mean: Adding an Adjective Rating Scale. Journal of usability studies. 2009 May;4(3):114–123. [Google Scholar]
- 27.Gunaratne Junius, Nov Oded. Influencing Retirement Saving Behavior with Expert Advice and Social Comparison as Persuasive Techniques. International Conference on Persuasive Technology; International Conference on Persuasive Technology; 2015; Chicago, Illinois, USA. 2015. pp. 205–216. [DOI] [Google Scholar]
- 28.Yang Yinfei, Yuan Steve, Cer Daniel, Kong Sheng-yi, Constant Noah, Pilar Petr, Ge Heming, Sung Yun-Hsuan, Strope Brian, Kurzweil Ray. Learning Semantic Textual Similarity from Conversations. Proceedings of the 3rd Workshop on Representation Learning for NLP; The 3rd Workshop on Representation Learning for NLP; July 20, 2018; Melbourne, Australia. 2018. pp. 164–174. [DOI] [Google Scholar]
- 29.Cialdini Robert B. Influence: The psychology of persuasion. New York City, USA: Harper Collins Publishers Inc; 2007. Authority. [Google Scholar]
- 30.Cha Zhang. Tsuhan Chen An active learning framework for content-based information retrieval. IEEE Trans. Multimedia. 2002 Jun;4(2):260–268. doi: 10.1109/tmm.2002.1017738. [DOI] [Google Scholar]
- 31.Sielis George A, Mettouris Christos, Papadopoulos George A. A Context Aware Recommender System for Creativity Support Tools. Journal of Universal Computer Science. 2011 Aug 1;17(12):1743–1763. [Google Scholar]
- 32.Achakulvisut T, Acuna DE, Ruangrong T, Kording K. Science Concierge: A Fast Content-Based Recommendation System for Scientific Publications. PLoS One. 2016 Jul 6;11(7):e0158423. doi: 10.1371/journal.pone.0158423. https://dx.plos.org/10.1371/journal.pone.0158423. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 33.Vaswani A, Shazeer N, Parmar N, Uszkoreit J, Jones L, Gomez AN, Kaiser L, Polosukhin I. Attention Is All You Need. 31st Conference on Neural Information Processing Systems; December 2017; Long Beach, CA. 2017. pp. 6000–6010. https://dl.acm.org/doi/pdf/10.5555/3295222.3295349. [DOI] [Google Scholar]
- 34.Devlin J, Chang MW, Lee K, Toutanova K. BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding. Annual Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies; June 2019; Minneapolis. 2019. https://arxiv.org/pdf/1810.04805.pdf. [DOI] [Google Scholar]
- 35.Derrar Honain Mohib. Clustering for the automatic annotation of customer service chat messages. 2019. [2020-12-11]. https://trepo.tuni.fi/bitstream/handle/123456789/27114/Derrar.pdf.
- 36.Qiu Minghui, Li Feng-Lin, Wang Siyu, Gao Xing, Chen Yan, Zhao Weipeng, Chen Haiqing, Huang Jun, Chu Wei. AliMe Chat: A Sequence to Sequence and Rerank based Chatbot Engine. Proceedings of the 55th Annual Meeting of the Association for Computational Linguistics; The 55th Annual Meeting of the Association for Computational Linguistics; 2017; Vancouver, Canada. 2017. pp. 498–503. [DOI] [Google Scholar]