Skip to main content
Springer Nature - PMC COVID-19 Collection logoLink to Springer Nature - PMC COVID-19 Collection
. 2018 Jun 22;822:267–280. doi: 10.1007/978-3-319-96553-6_19

News Timeline Generation: Accounting for Structural Aspects and Temporal Nature of News Stream

Mikhail Tikhomirov 7,, Boris Dobrov 7
Editors: Leonid Kalinichenko8, Yannis Manolopoulos9, Oleg Malkov10, Nikolay Skvortsov11, Sergey Stupnikov12, Vladimir Sukhomlin13
PMCID: PMC7120637

Abstract

The number of news articles that are published daily is larger than any person can afford to study. Correct summarization of the information allows for an easy search for the event of interest. This research was designed to address the issue of constructing annotations of news story. Standard multi-document summarization approaches are not able to extract all information relevant to the event. This is due to the fact that such approaches do not take into account the variability of the event context in time. We have implemented a system that automatically builds timeline summary. We investigated impact of three factors: query extension, accounting for temporal nature and structure of news article in form of inverted pyramid. The annotations that we generate are composed of sentences sorted in chronological order, which together contain the main details of the news story. The paper shows that taking into account the described factors positively affects the quality of the annotations created.

Keywords: Timeline summarization, Extractive summarization, Multi-document summarization, Information retrieval

Introduction

Due to the explosive growth of the amount of content on the Internet, the problems of extraction and automatically summarizing useful information in the incoming data stream arises. One of such problems is the summarization of news articles on an event. The news story - is a set of news reports from various sources dedicated to describing an event. Such problems are often investigated and solved by news aggregators, for example, Google.News1 [17] or Yandex.News.2 This is due to the fact that to work with such problems the researcher needs a huge and diverse collection of news articles.

The typical “lifetime” of the news story (the time of active discussion of the event) is usually a day or two, but not all events are so short. Some news stories have a “history” in the form of a set of previous events that occurred at different moments and are more or less related to each other. Existing multi-document summarization approaches do not take into account the fact that the context, actors, geography and other event properties can vary over time.

The fact that journalists are returning to the same events, for example, with the appearance of new data, indicates that such events are important for the society. The need for a brief summary of the event raises the problem of forming a “timeline summary”. Timeline summary is a type of multi-document summary, containing the essential details of the subject matter under discussion. The construction of such annotations is a complex task, performed by journalists or analysts manually. This implies that the automation of such a process is a urgent problem.

In this paper we consider challenges and solutions for the automatic generation of temporal summaries. We consider this problem as a multi-document summarization on a query over a representative collection of news documents. The query in this case is the text of the news message. The situation corresponds to the scenario when a user would like to receive a timeline summary after reading the news document. The result should be a time-ordered list of descriptions of the key sub-events related to main event. The result consists of parts of existing sentences, since our solution refers to extractive summarization approaches.

A system was developed to automate the timeline summarization process. Experiments were conducted over a collection of 2 million Russian news for the first half of 2015. Three new factors were investigated to improve the results of constructing a timeline summary: query extension using pseudo-relevance feedback, accounting for the timing characteristics of news stories and the structure of the inverted pyramid.

This is a follow-up study of timeline summarization problem reported in previous paper [25]. In this study, we expanded the collection of standard annotations three-fold. The evaluation process was improved by dividing the collection into a training and test parts. An optimization module was added for fitting the configurations. As a result, substantial progress was achieved. Taking into account the structure of the inverted pyramid showed a significant increase in the values of metrics, which was not achieved in the previous article.

Related Work

Automatic Text Summarization Problem

Currently, there are quite a number of methods for automatic text summarization [3]. Some methods that use large linguistic ontologies [12, 15], that may be automatically supplemented during the analysis. Other methods are based on the statistical properties of texts [16] or machine learning [13].

During the generation of the annotations, the following problems occur [3, 7, 11]:

  • Ensuring the completeness of the presentation of information, including the most up-to-date information.

  • Decreasing of redundancy in the information provided.

  • Ensuring the coherence and understandability of the information provided.

To ensure the completeness of the resulting annotation, it is often necessary to find links between sentences or documents [20].

To determine the redundancy in the generated annotations, various measures of similarity between sentences are used. One of the most common approaches is clustering - the selection of content groups of sentences [6]. Another approach to reduce redundancy is to compare candidate sentence with sentences that have already been included in the summary and to evaluate novel information. Example of such approach is the Maximal Marginal Relevance (MMR) [2].

The problem of ensuring the coherence of information in the summary arises both in the methods of generating the annotation [18, 19], and in the methods of evaluation, because in order to assess the connectivity and linguistic qualities of the annotation, it is necessary to perform a manual evaluation.

Timeline Summary

The problem of timeline summary construction has a number of differences from the standard summarization problem. For example, the temporal nature of events must be taken into account [9]. Also, to ensure completeness of the information provided, it is required to find documents from all sub-events of the topic under consideration.

When constructing timeline summary, data processing is mainly carried out over huge collections. In such collections, most of the information is not relevant to the user’s request. This problem can be solved by using clustering methods [10, 14]. But the clustering methods have some issues. First, such a task should be solved many times over huge collections of documents, which affects the response time of the system. Secondly, the degree of closeness can be significantly smaller with standard measures of similarity for documents that describe far-in-time but related events. And, of course, it is required to identify the most characteristic objects [1, 9], for example, taking into account the structural features of the flow of documents [5, 8].

Statement of the Problem

General Description

The problem of constructing a timeline summary is a query-oriented. In the most general case, the user has a news document as a query. So further this problem will be considered as a problem of automatic creation of a summary on a query in the form of a text document. The output of the system is an annotation of n sentences. The connectivity between the sentences in this paper is not required. Figure 1 provides an example of a possible summary about the conflict on cemetery taken from the Interfax website.3

Fig. 1.

Fig. 1.

Timeline summary part about conflict at cemetery.

The aim of the work is to study the influence of various factors on the quality of the annotation.

Mathematical Statement of the Problem

The problem described above can be formalized in the following way. Let Inline graphic be a set of queries and an associated set of reference annotations Inline graphic be an associated set of reference annotations. The system generates a set of summary Inline graphic in response to queries Inline graphic by algorithm Inline graphic. Then the problem is reduced to maximizing the following functional:

graphic file with name M6.gif 1

where Inline graphic is the proximity function between the annotations. Optimization is carried out for all parameters of the algorithm.

Approach

Collection Processing

As mentioned earlier, the input collection contains 2 million news articles. It is not possible to work directly with such amount of information, therefore, it was decided to interact with the collection through a search engine. Search engine allows:

  • Get a list of documents by text request.

  • For a given document from collection, get the basic information: text, index, meta-information.

Studied Features

In this paper the following factors were investigated:

  • Query extension strategy.

  • Accounting for the temporal nature of news stories.

  • Accounting for the structure of a news article in the form of an inverted pyramid.

Query Extension Strategy

Information that can be obtained from a query document is basically not enough to effectively build this type of annotation. This fact is a consequence of the fact that most news articles are not a general description of the event, but a discussion of some particular incident or fact.

To avoid this problem, it is necessary to use the query extension techniques. The developed algorithm uses the idea of pseudo-relevance feedback, which is widely used in information retrieval problems [21]. For the query-document, the algorithm includes the following steps:

  1. The most significant K-terms are chosen on the basis of tf-idf weights forming thus the first-level query.

  2. Further on the basis of the first-level query documents are retrieved.

  3. The extracted cluster of documents is analyzed to find the most important terms forming thus the second-level query:

    1. For each document, the most significant Inline graphic terms are considered.
    2. For each term, it is calculated how often it was in the top Inline graphic terms on all cluster.
    3. The list of terms is sorted by frequency, the best Inline graphic terms are selected.
  4. Steps 2–3 are repeated (A double query extension process that forms a third-level query).

  5. Output of the algorithm is a vector of Inline graphic terms representing to some extent the semantics of the input document.

Note that Inline graphic are parameters of the algorithm and they must be configured. As an example of the work of the query extension module, consider the algorithm steps on a news article about the terrorist attack in Paris (Table 1).

Table 1.

Query extension algorithm stages example.

Entity name Content
Initial doc President François Hollande of France called it a display of extraordinary “barbarism” that was “without a doubt” an act of terrorism. According to the latest information, as a result of shooting, 11 people were killed, four more are in critical condition. …
First-level query Posten, Jyllands-posten, Jyllands, Herbo, Charlie, Hollande
Second-level query Reprint, Scandal, Weekly, Caricature, Hollande, Satirical, Terrorist act, Charlie, Herbo
Third-level query (double query extension) Journal, Muhammad, Satirical, Attack, Prophet, Terrorist act, Paris, Caricature, Hollande, Herbo, Charlie

The table shows that a higher-level query has more significant terms for this event.

Temporal Nature of News Stories

Since any event depends on time, the content of publications and their number also depend on the time. As an example, Fig. 2 shows a graph of the time dependence of publications on the “Earthquake in Nepal” event.

Fig. 2.

Fig. 2.

Dependence of the number of publications per day for an event

To take into account this factor, for the set Inline graphic of extracted documents the following procedure is undertaken:

  1. The entire timeline of the event is divided into days with labels Inline graphic.

  2. Each document receives a label from Inline graphic based on the publication date Inline graphic.

  3. Documents published on days with a number of publications less then the Inline graphic threshold (2) are discarded.

    graphic file with name M18.gif 2
  4. The output is a sorted list of collections Inline graphic, where each collection Inline graphic contains only documents with the label Inline graphic.

Inverted Pyramid

The strategy of writing a high-quality news article often relies on the structure of the “inverted pyramid” (Fig. 3). The greatest interest is the upper and lower parts of the pyramid:

  • The upper part contains the most concentrated information about the event under discussion.

  • The lower part may contain references to important related events in the past.

Fig. 3.

Fig. 3.

Inverted pyramid on the example of an article. (https://themoscowtimes.com/articles/moscow-museum-takes-you-inside-north-korea-60240)

This structure is taken into account in two ways:

  1. Inter-document feature based on the graph approach.

  2. Intra-document feature, which increases the weight of sentences located in the upper and lower parts of the inverted pyramid.

Inter-Document Feature.

This feature is taken into account in the following way:

  1. For a set of documents Inline graphic, a similarity matrix between the upper and lower parts of the documents is constructed. If the specified similarity threshold is exceeded, it is considered that there is a link between the documents Inline graphic and Inline graphic.

  2. The importance of documents is calculated by using the LexRank algorithm over the constructed graph [4].

  3. For documents whose weight is greater than a certain threshold, the previously described procedure for expanding the query is performed.

As a result, the output is a ranked list of documents Inline graphic and a set of Inline graphic new queries, which further, together with accounting for the temporal nature of the news story, will help in sentence ranking algorithm. Among other things, document weights will also be taken into account in the ranking functions.

Intra-Document Feature.

To this feature into account for the following procedure is undertaken: during the ranking of sentences, the weight of the sentence is multiplied by a coefficient that lowers the weight of sentences in the middle of the document.

Also, after described inter-document procedure, all constructed extended queries Inline graphic are mapped to Inline graphic labels from Inline graphic (Fig. 4).

Fig. 4.

Fig. 4.

Query mapping.

Similarity of Sentences

At various stages of the algorithm, there are a number of points where the measure of closeness between sentences is calculated. For these purpose a cosine measure of similarity (3) is used in all cases.

graphic file with name M30.gif 3

The choice of representation of a sentence plays an important role for calculating similarity. In this article we used the standard tf-idf representation. But to calculate the similarity between sentences when searching for links between documents, word2vec [24] representation was used. To achieve this, the resulting sentence vector is represented as a weighted mean of word2vec word representations. Weighing was carried out by tf-idf.

Word2vec model was trained on the entire collection of 2 million news articles. During preprocessing removal of stop words and lemmatization were applied. The width of the window was chosen to be 5, and the length of the vector was 100.

Sentence Ranking Module

This module deals with the ranking of sentences. The ranking is a modified version of the Inline graphic algorithm – Inline graphic (4) taking into account all the factors described in Sect. 4.2:

graphic file with name M33.gif 4

where Inline graphic – is a term describing the positive part of the formula, which depends on the similarity of the sentence to the query, the weight of the document from which the sentence is taken, and the sentence number in the document.

graphic file with name M35.gif 5
graphic file with name M36.gif 6

The parameters Inline graphic and Inline graphic are configurable parameters of the algorithm, Inline graphic – is document weight Inline graphic, which includes a sentence under the index Inline graphic, Inline graphic – is estimated sentence under the index Inline graphic with label Inline graphic, Inline graphic – query for this time label, Inline graphic – multiplier, which reduce the weight of sentences from the middle of the document.

Inline graphic is the penalty term. It depends on the similarity to the already extracted sentences:

graphic file with name M48.gif 7

where Inline graphic is one of the extracted sentences, Inline graphic is the set of all already extracted sentences.

Processing of sentences occurs in chronological order with a restriction on the maximum number of sentences per day.

System Diagram

The features described in Subsect. 4.2 are realized at various stages of the system. The general scheme of the algorithm is shown on Fig. 5.

Fig. 5.

Fig. 5.

Working scheme.

Evaluation

Metrics for Evaluation

The system was evaluated using several metrics: ROUGE-1, ROUGE-2, and Sentence Recall Inline graphic:

graphic file with name M52.gif 8

where Inline graphic is the set of n-grams for the constructed annotations, Inline graphic is the set of n-grams for the reference (gold) annotations.

graphic file with name M55.gif 9

where Inline graphic is the set of sentences from the constructed annotations, Inline graphic is the set of sentences from the reference annotations. Operator ≡ denotes the following: the result of the Inline graphic is a subset of Inline graphic such that their semantic equivalent is present in Inline graphic.

Data Preparation

Since a test set of annotations is required for evaluating procedure, in the course of the research, timeline summaries were manually prepared. The procedure for the formation of such a collection was as follows:

  1. At the first stage with the help of Wikipedia there high-profile events were selected, which were actively covered in the press for the beginning of 2015.

  2. Further, for most of the events on the site “Interfax”, the search for the corresponding story was carried out. On the basis of documents corresponding to the story, a timeline summary was created.

  3. If there is no corresponding story on the “Interfax”, the materials were studied on the topic and a timeline summary was created on the basis of the documents read.

As a result, 45 annotations on 15 news stories were created (Table 2).

Table 2.

News stories on which the reference annotations are made.

Story ID Description
Story 1 The flight of the space probe DAWN to Ceres
Story 2 Fire in Khakassia
Story 3 Terrorist act in Paris
Story 4 The plane crash in Taiwan
Story 5 Earthquake in Nepal
Story 6 Fire on the Orel submarine.
Story 7 Attack on the synagogue in Copenhagen
Story 8 Fire in the mall «Admiral»
Story 9 Terrorist act in Sydney
Story 10 Ice Hockey Championship
Story 11 The case of Svetlana Davydova
Story 12 The Murder of Boris Nemtsov
Story 13 Dangerous coronavirus in Korea
Story 14 Protest in Yerevan
Story 15 The failure of the spaceship “Progress”

Optimization of Algorithm Parameters

Since the system contains a large number of parameters (total 23 parameters), some of which are presented in Table 3, there was a need to optimize the choice of the values of these parameters.

Table 3.

Some system parameters.

Parameter name Description Opt. value
KeepL The number of lemmas to choose when building a first-level query 6
DocCount The maximum number of documents retrieved using a search engine 400
QuerrySize The size of the resulting expanded query 14
TopLemms The number of the most significant lemmas extracted in the work of the algorithm for constructing an extended query 18
MinLinkScore The minimum value of similarity between the top and bottom of the documents to identify the reference 0.5
Lambda The value of the parameter Inline graphic for Inline graphic 0.58
Alpha The value of the parameter Inline graphic for Inline graphic 0.45
MaxDaily AnswerSize The maximum number of sentences per day 15
Doc Boundary Threshold of the importance of documents for building new queries 0.61

To achieve this, the entire collection of the reference annotations was divided into train and test parts with the ratio 2 to 1. Further, the functional (1) was implemented in Python using an open hyperopt [22] package based on machine learning. This package uses the technique of Sequential model-based optimization (SMBO) [23] for the parameters selection. The parameters were trained on the training part. After that, the final evaluation of the configurations took place on the test part.

Results

In order to evaluate the contribution of the considered features, a fitting and evaluation of the following 6 configurations was made:

  1. baseline – a simple approach to summarization, without taking into account the factors considered, using MMR as ranking algorithm.

  2. querry-ex – adding a query extension strategy feature to baseline (Sect. 4.3), but without double query extension.

  3. double-exquerry-ex + double query extension (Sect. 4.3).

  4. temporaldouble-ex + accounting for the temporal nature of news stories (Sect. 4.4).

  5. importancetemporal + accounting for the structure of a news article in the form of an inverted pyramid, when tf-idf representation is used (Sect. 4.5).

  6. w2v-impimportance, but using w2v for computing sentence similarity when accounting for the structure of a news article (Sect. 4.6).

The result of evaluation of the configurations can be seen in Table 4. This table shows that each of the features considered gives a positive contribution to the quality of generation of timeline summary. As an example of the final annotation, one can consider a fragment of the annotation on the previously mentioned incident of the crash in Taiwan in Table 5.

Table 4.

Evaluation results.

R1 R2 Inline graphic
baseline 0.320 0.127 0.162
querry-ex 0.326 0.133 0.175
double-ex 0.356 0.176 0.233
temporal 0.384 0.178 0.240
importance 0.399 0.176 0.246
w2v-imp 0.403 0.181 0.254

Table 5.

The generated timeline summary fragment about the plane crash in Taiwan.

Date Sentence
11.02.2015 Transasia Airways will pay relatives of victims of a plane crash in Taiwan for 470 thousand
11.02.2015 The tragedy in Taiwan, one-fifth of the pilots of Taiwan’s Transasia airline have not passed the proficiency test
12.02.2015 Rescuers completed the search operation for victims of the crash of an airline to Transasia Airways, which crashed on February 4 in Taiwan
01.07.2015 Crew crashed in Taiwan aircraft Transasia Airways shut off the engines after a loss of power
02.07.2015 The Transasia plane crashed on February 4 in Taiwan, because the pilot accidentally turned off the running engine when the second engine stalled

Conclusions and Future Work

In this article we presented an approach for building a timeline summary. The conducted research shows that the problem of constructing the timeline summary differs from the standard MDS problem. The effectiveness of using the following features was shown:

  • Query extension strategy.

  • Accounting for the temporal nature of news stories.

  • Accounting for the structure of a news article in the form of an inverted pyramid.

Extending the query, as expected, has a positive effect on the event representation discussed in the document. But the interesting fact is that re-extension the query (double query extension) has a much greater effect. This is because the documents that are retrieved on the first-level query are not sufficient for a good presentation of the event.

The fact that accounting for the temporal nature of news stories improves the quality of the annotation is an obvious consequence of the fact that news stories and events have temporal characteristics.

Taking into account the structure of the inverted pyramid gives an improvement. Increase the values of metrics on the w2v-imp configuration means that the correctness of the recognized links between the documents plays a significant role. This fact raises challenges for future research.

Using structural features of news articles make it possible to obtain information, the use of which can significantly improve the quality of generated annotations.

Footnotes

Contributor Information

Leonid Kalinichenko, Email: leonidandk@gmail.com.

Yannis Manolopoulos, Email: manolopo@csd.auth.gr.

Oleg Malkov, Email: malkov@inasan.ru.

Nikolay Skvortsov, Email: nskv@mail.ru.

Sergey Stupnikov, Email: sstupnikov@ipiran.ru.

Vladimir Sukhomlin, Email: sukhomlin@gmail.com.

Mikhail Tikhomirov, Email: tikhomirov.mm@gmail.com.

Boris Dobrov, Email: dobrov_bv@srcc.msu.ru.

References

  • 1.Binh, T.G., Alrifai, M., Quoc Nguyen, D.: Predicting relevant news events for timeline summaries. In: Proceedings of the 22nd International Conference on World Wide Web, pp. 91–92. ACM (2013)
  • 2.Carbonell, J., Goldstein, J.: The use of MMR, diversity-based reranking for reordering documents and producing summaries. In: Proceedings of the 21st Annual International ACM SIGIR Conference on Research and Development in Information Retrieval, pp. 335–336. ACM (1998)
  • 3.Dang, H.T.: Overview of DUC 2006. In: Proceedings of the Document Understanding Workshop, Presented at HLT-NAACL 2006 (2006). http://duc.nist.gov/pubs/2006papers/duc2006.pdf
  • 4.Erkan G, Radev DR. Lexrank: graph-based lexical centrality as salience in text summarization. J. Artif. Intell. Res. 2004;22:457–479. [Google Scholar]
  • 5.Hu P, Huang ML, Zhu XY. Exploring the interactions of storylines from informative news events. J. Comput. Sci. Technol. 2014;29(3):502–518. doi: 10.1007/s11390-014-1445-6. [DOI] [Google Scholar]
  • 6.Radev, D., Jing, H., Budzikowska, M.: Centroid-based summarization of multiple docuemtns: sentence extraction, utility-based evaluation, and user studies. In: Proceedings of the 2000 NAACL-ANLP Workshop on Automatic Summarization, Seattle, pp. 21–30 (2000)
  • 7.Radev D, McKeown K, Hovy E. Introduction to the special issue on summarization. Comput. Linguist. 2002;28(4):399–408. doi: 10.1162/089120102762671927. [DOI] [Google Scholar]
  • 8.Shahaf D, Guestrin C. Connecting two (or less) dots: discovering structure in news articles. ACM Trans. Knowl. Discov. Data (TKDD) 2012;5(4):24–54. [Google Scholar]
  • 9.Tran G, Alrifai M, Herder E. Timeline summarization from relevant headlines. In: Hanbury A, Kazai G, Rauber A, Fuhr N, editors. Advances in Information Retrieval; Cham: Springer; 2015. pp. 245–256. [Google Scholar]
  • 10.Yan, R., et al.: Evolutionary timeline summarization: a balanced optimization framework via iterative substitution. In: Proceedings of the 34th International ACM SIGIR Conference on Research and Development in Information Retrieval, Beijing, China, 24–28 July 2011, pp. 745–754. ACM (2011). 10.1145/2009916.2010016
  • 11.Wu Z, Lei L, Li G, Huang H, Zheng C, Chen E, Xu G. A topic modeling based approach to novel document automatic summarization. Expert Syst. Appl. 2017;84:12–23. doi: 10.1016/j.eswa.2017.04.054. [DOI] [Google Scholar]
  • 12.Hennig, L., Umbrath, W., Wetzker, R.: An ontology-based approach to text summarization. In: Proceedings of the 2008 IEEE/WIC/ACM International Conference on Web Intelligence and Intelligent Agent Technology, vol. 3, pp. 291–294 (2008)
  • 13.Nallapati, R., Zhou, B., Gulcehre, C., Xiang, B.: Abstractive text summarization using sequence-to-sequence RNNs and beyond. arXiv preprint arXiv:1602.06023 (2016)
  • 14.Wei T, Lu Y, Chang H, Zhou Q, Bao X. A semantic approach for text clustering using WordNet and lexical chains. Expert Syst. Appl. 2015;42(4):2264–2275. doi: 10.1016/j.eswa.2014.10.023. [DOI] [Google Scholar]
  • 15.Allahyari, M., Pouriyeh, S., Assefi, M., Safaei, S., Trippe, E.D., Gutierrez, J.B., Kochut, K.: Text summarization techniques: a brief survey. arXiv preprint arXiv:1707.02268 (2017)
  • 16.Rush, A.M., Chopra, S., Weston, J.: A neural attention model for abstractive sentence summarization. arXiv preprint arXiv:1509.00685 (2015)
  • 17.Hertzfeld, A.: Introducing Google News Timeline. https://news.googleblog.com/2009/04/introducing-google-news-timeline.html. Accessed 10 Jan 2018
  • 18.Christensen, J., Mausam, S.S., Soderland, S., Etzioni, O.: Towards Coherent Multi-Document Summarization. In: HLT-NAACL, pp. 1163–1173 (2013)
  • 19.Nishikawa, H., Arita, K., Tanaka, K., Hirao, T., Makino, T., Matsuo, Y.: Learning to generate coherent summary with discriminative hidden semi-markov model. In: COLING, pp. 1648–1659 (2014)
  • 20.Barzilay, R., Elhadad, M.: Using lexical chains for text summarization. In: Advances in Automatic Text Summarization, pp. 111–121 (1999)
  • 21.Jiang, L., Mitamura, T., Yu, S.I., Hauptmann, A.G.: Zero-example event search using multimodal pseudo relevance feedback. In: Proceedings of International Conference on Multimedia Retrieval, p. 297 (2014)
  • 22.Bergstra J, Komer B, Eliasmith C, Yamins D, Cox DD. Hyperopt: a python library for model selection and hyperparameter optimization. Comput. Sci. Discov. 2015;8(1):014008. doi: 10.1088/1749-4699/8/1/014008. [DOI] [Google Scholar]
  • 23.Hutter F, Hoos Holger H, Leyton-Brown K. Sequential model-based optimization for general algorithm configuration. In: Coello CA, editor. Learning and Intelligent Optimization; Heidelberg: Springer; 2011. pp. 507–523. [Google Scholar]
  • 24.Goldberg, Y., Levy, O.: word2vec Explained: deriving Mikolov et al.’s negative-sampling word-embedding method. arXiv preprint arXiv:1402.3722 (2014)
  • 25.Tikhomirov, M.M., Dobrov, B.V.: Using news corpora for temporal summary formation (in Russian). In: Selected Papers of the XIX International Conference on Data Analytics and Management in Data Intensive Domains (DAMDID/RCDL 2017), CEUR Workshop Proceedings, Moscow, Russia, vol. 2022, pp. 165–171 (2017)

Articles from Data Analytics and Management in Data Intensive Domains are provided here courtesy of Nature Publishing Group

RESOURCES