Skip to main content
PeerJ Computer Science logoLink to PeerJ Computer Science
. 2018 Mar 26;4:e148. doi: 10.7717/peerj-cs.148

Indexing labeled sequences

Tatiana Rocher 1, Mathieu Giraud 1,, Mikaël Salson 1,
Editor: Rahul Shah
PMCID: PMC7924554  PMID: 33816803

Abstract

Background

Labels are a way to add some information on a text, such as functional annotations such as genes on a DNA sequences. V(D)J recombinations are DNA recombinations involving two or three short genes in lymphocytes. Sequencing this short region (500 bp or less) produces labeled sequences and brings insight in the lymphocyte repertoire for onco-hematology or immunology studies.

Methods

We present two indexes for a text with non-overlapping labels. They store the text in a Burrows–Wheeler transform (BWT) and a compressed label sequence in a Wavelet Tree. The label sequence is taken in the order of the text (TL-index) or in the order of the BWT (TLBW-index). Both indexes need a space related to the entropy of the labeled text.

Results

These indexes allow efficient text–label queries to count and find labeled patterns. The TLBW-index has an overhead on simple label queries but is very efficient on combined pattern–label queries. We implemented the indexes in C++ and compared them against a baseline solution on pseudo-random as well as on V(D)J labeled texts.

Discussion

New indexes such as the ones we proposed improve the way we index and query labeled texts as, for instance, lymphocyte repertoire for hematological and immunological studies.

Keywords: Data structures, Text indexing, Burrows–Wheeler transform, Wavelet Tree, V(D)J recombination

Introduction

Labels are a way to add some information on a text, as the semantics of words on an English sentence or functional annotations such as genes on a DNA sequences. Can we build an index that saves a labeled text like ACGCC…TTGA (of size 96), which have the label L1 in the positions 12–41 and the label L2 in the positions 56–96? We consider here the case where a same label can be given to different (but similar) patterns and can occur several times in the text. We introduce two indexes which store a labeled text and answers to position–label association queries. Those indexes share some ideas with the RL-FMI (Mäkinen & Navarro, 2004) which uses a Burrows–Wheeler transform (BWT) and a Wavelet Tree (WT). Using a somewhat similar organization, we index a labeled text. The following sections present the TL- and TLBW-indexes (text–label indexes) and their associated queries. The last section presents experimental results on simulated and real genomic data.

Let T = t0 t1tn − 1 be a text of length n over an alphabet of size σ. The text may be composed of several sequences, each sequence ending with the symbol $. Let L = {L0, L1Ll − 1} be a set of labels. A labeled text (T, A) is here a text with non-overlapping labels: a letter should have at most one label. Each position i of the text is labeled by exactly one label aiL ∪ {ε}, where the special ε label is put on every letter without label. A = a0 a1an − 1 is the label string. Figure 1 shows the text which will be used all over this article.

Figure 1. The text T = AACAGC$ATCAAC$AGCTTT$, with three sequences, is labeled with the label string A = L1.2 L1.2 L1.2 L2 L2 L2 ε L3 L3 L3 L1.1 L1.1 L1.1 ε L2 L2 L2 ε ε ε ε.

Figure 1

Given a bit vector B = B[0]B[1] … B[n − 1], where each B[i] is either 0 or 1, we define as B[i, j] the vector B[i]B[i + 1] … B[j]. Let us call rank(b, i, B) the number of times the bit b ∈ {0, 1} appears in the prefix B[0, i] and select (b, j, B) the position i of the jth appearance of the bit b in B. Such a bit vector B can be stored in nH0(B) + o(n) bits to support rank and select in O(1), where H0(B) is the zeroth order entropy of B (Raman, Raman & Rao, 2002).

The BWT (Burrows & Wheeler, 1994) is a reversible algorithm which reorganizes the letters of a text. The transformed text, BWT(T), is the concatenation of the last letters of the lexicographically sorted rotations of the text. BWT(T) is easier to compress and can be stored using nHk(T) + o(n) bits, where Hk is the kth order empirical entropy of the text T. The FM-index (Ferragina & Manzini, 2000) uses the BWT, a table C, where C[α] gives the number of letters lexicographically smaller than α, and a function Occ(α, i) giving the number of occurrences of α in BWT(T)[0, i]. The FM-index allows to efficiently search a pattern using backward search.

A WT (Grossi, Gupta & Vitter, 2003) is a binary tree storing a text, where each symbol from the alphabet corresponds to a leaf. The root is a bit vector where every position corresponds to the element it has to index. Any position marked 0 (respectively 1) corresponds to an element whose leaf is on the left (respectively right) descendant of the node. The process is repeated recursively until the leaves. For a text A of length a in an alphabet of size l, the construction of a balanced WT needs O(alogl/loga) time (Munro, Nekrich & Vitter, 2016) and requires nH0(A) + o(a log l) bits when bit vectors are zero-order compressed (Navarro & Mäkinen, 2007).

The usual full-text indexes such as the FM-index (Ferragina & Manzini, 2000) or the LZ-index (Kärkkäinen & Ukkonen, 1996) do not index labeled texts. Some recent researches allow bidimensional range queries: Arroyuelo et al. (2015) represent an XML tree structure as a compressed bit vector and allow queries on structured patterns.

We focus here on non-overlapping labels and look for ways to efficiently store such labels along with sequences as well as to query them. The basic idea is that consecutive labels in A are often the same: we will thus store compressed label sequences.

Materials and Methods

TL-index: indexing labels over a text

Given a labeled text (T, A), we define the TL-index as, using a FM-index built on a BWT U to index the text, a bit vector BA marking the positions in the text where the labels change, and a WT WA indexing a compressed label sequence (Fig. 2A).

Figure 2. TL- and TLBW-indexes store a text of length 21 with four unique labels.

Figure 2

(A) TL-index. The WT WA has four internal nodes and five leaves. We have BA[4] = 0 because a4 = a3 = L2, and BA[3] = 1 because a3a2. The label a3 = L2 is thus stored in A′, at position 1, hence W〈1〉 = a1 = L2. A′ has two occurrences of the label L2: W−1〈L2〉 = {1, 6}, corresponding to the six positions {3, 4, 5, 14, 15, 16} in A. (B) TLBW-index. The root of the WT D′ is now built in the order of the BWT U. The WT WD has four internal nodes and five leaves. The label d6 = L1.2 is stored in D′, at position 6, hence W〈6〉= d6 = L1.2. D′ has two occurrences of the label L1.2: W−1〈L1.2〉 = {6, 11}, corresponding to the three positions {6, 7, 13} in U. In both cases, the label sequences (A, D) and the compressed label sequences (A′, D′) are not stored in the indexes.

BWT U. Let U = u0 u1un − 1 be the BWT of T = t0 t1tn − 1. As usually done, the FM-index samples every log1+ε n values of a suffix array to retrieve the text positions of any occurrence (Navarro & Mäkinen, 2007).

Bit vector BA. Let BA a compressed bit vector of length n such that BA[0] = 1, and, for i ≥ 1, BA[i] = 0 if ai = ai − 1, and otherwise BA[i] = 1.

Wavelet Tree WA. Let A′ = 〈ai | BA[i] = 1〉. A′ = a0 a1aa − 1 is called the compressed label sequence. It is a subsequence of A, of length a, containing only the changing labels according to the positions in A. The compressed label sequence A′ is stored using a WT WA. WA is defined as (Q, q0), where Q is the set of nodes of WA and q0Q is the root node. Each node qQ is q = (q.val, q.label, q.left, q.right, q.parent), where q.labelL ∪ {ε} is the label of q and q.parent is the parent node of q (q0.parent being the null value ⊥). Both q.left and q.right are child nodes in Q ∪ {⊥}. A leaf is a node q where q.right and q.left are ⊥. Each leaf q is linked to a label q.labelL ∪ {ε}. The l + 1 leaves are exactly the labels of L ∪ {ε} and we define leaf (q.label) = q. On a leaf q, we have q.val = ⊥. Let q be a non-leaf node: we have q.label = q.val is the bit vector rooted at q in the WT. We explain in “Shaping the WT for a Label Hierarchy” how WA can be further shaped depending on a label hierarchy.

WA is part of the index. This WT is used to answer efficiently bidimensional range queries where the labels and the label positions are the two dimensions (Mäkinen & Navarro, 2007). A balanced WT has a height of log l, with l, the number of leaves. The accessor Wi〉 returns ai, in O(log l) time. This is a classical query within a WT. Given a label LxL, the function selectWLx, i〉 gives the position of the ith Lx label in A′ in O(log l) time. The accessor W−1Lx〉 gives the list of positions in A′ where ai = Lx. It runs in O(log l × occ) time, with occ the number of occurrences of Lx in A′.

TLBW-index: indexing labels in the order of the BWT

The BWT tends to store text repetitions consecutively. As those repetitions may have the same labels, it would be interesting that the labels benefit from the BWT reordering. Hence, labels can also be stored in the order of U.

Given a labeled text (T, A), the TLBW-index is defined as (U, BD, WD) (Fig. 2B). The BWT U is built in the same way as the TL-index. Let D = d0 d1dn − 1 the labels in the order of U. The bit vector BD of size n is such that BD[0] = 1, and, for i ≥ 1, BD[i] = 0 if di = di − 1, and otherwise BD[i] = 1. Let D′ = 〈di | BD[i] = 1〉. D′ = d0, d1dd − 1 is a compressed label sequence of length d, subsequence of D. The WT WD now indexes the compressed label sequence D′. The TLBW-index will be slower to build than the TL-index as it needs D. On the other side, as it is aware of the order of letters in the BWT, it will be able to support faster text/label queries.

Queries

The indexes allow the following classical queries.

  • label(ti) (also called access(ai))—Which label is on the letter ti? This query is done is O(log l) time in the TL-index, and in O(log1+ε n + log l) time in the TLBW-index since it has to convert positions in U order (see Fig. 3).

  • findP(P) P—Which are the occurrences of a pattern P? It is solved with the FM-index alone (Ferragina & Manzini, 2000). This query runs in O(|P| + occ × log1+ε n) time in both indexes, where occ is the number of occurrences of P in T.

  • findL(Lx)—Which text positions are labeled Lx? The query runs in O(y × log l) time in the TL-index, and in O(y(log l + log1+ε n)) time in the TLBW-index, where Y = W−1Lx〉 and y = |Y|. See Fig. 4.

Figure 3. Finding the label of a letter in a TLBW-index.

Figure 3

The letter u7 corresponds to a 0-bit in BD. The previous 1-bit in BD is the bit at position 6. It’s the 7th 1-bit of BD, it corresponds to the bit at position 6 in WD’s root, which label is W〈6〉 = L1.2.

Figure 4. Finding the letters which have a L1.2 label in a TLBW-index.

Figure 4

Y = W−1L1.2〉 = {6, 11}. For every bit of Y, we access to the 1-bit corresponding in BD ({6, 13}), and all the 0-bits which follow it ({7}). The corresponding letters of U are labeled L1.3, at positions {6, 7, 13}.

The three previous queries are well known in text indexes. The two next queries search for a pattern and a label at the same time.

  • countPL(P, Lx)—How many text positions are labeled Lx and begin an occurrence of a pattern P? As in the findL(Lx) query, the occurrences of P are found in U, in the positions from i to j.
    • TL-index: We translate all these occ occurrences to have the corresponding positions in the text. For each of them, we run the query label(ti). The total time is O(|P| + occ (log n1+ε + log l)).
    • TLBW-index: See Algorithm 1. i and j correspond to the positions i′ = rank(1, i, BD) and j′ = rank(1, j, BD) in the root q0 of WD. We then use an accessor customized from the rangeLocate function of (Mäkinen & Navarro, 2007), simulating a two-dimensional range query on [Lx, Lx] × [i′, j′] in WD to give the set of positions Z = {z | az = Lx and i′ ≤ zj′ } in O(|Z| × log l) time. This accessor first traverses the WT from the root to the leaf Lx and then traverses it back to the root to find the positions in q0.val which match the label in the given range. For every position found, we find the corresponding positions in BD and expand them to the following 0-bits in BD. This query runs in O(|P| + |Z| × log l) time.
  • findPL(P, Lx)—Which text positions are labeled Lx and begin an occurrence of a pattern P?
    • TL-index: This query is exactly the same as countPL(P, Lx).
    • TLBW-index: We use the countPL(P, Lx) query detailed in Algorithm 1, replacing the counter cnt with a list Y holding the matching positions in U. The positions are converted at the end in the text order. This query runs in O(|P| + |Z| × log l + y × log1+ε n) time, with |Y| = y.

Algorithm 1. countPL(P, Lx): Count the positions starting a pattern P and labeled with Lx.

(i,j) from findP(P)             ▹ starting and ending positions of occurrences of P in U
i′ = rank(1, i, BD)
j′ = rank(1, j, BD)
C = path(Lx)              ▹ bit vector representing the path from the root to leaf(Lx)
node = q0
for p in 0 to |C| − 2 do       ▹ loop corresponding to rangeLocate in (Mäkinen & Navarro, 2007)
i′ = rank(C[p], i′ − 1, node.val)
j′ = rank(C[p], j′, node.val) − 1
node = (C[p] == 0)? node.left : node.right
if i′ > jthen return 0
cnt = 0
for k in i′ to jdo
k′ = selectWLx, rank(C[|C| − 1], k, node.val)〉              ▹ ith Lx label in A′
i″ = select(1, k′, BD)
j″ = select(1, k′ + 1, BD)
cnt = cnt + j″ − i″                 ▹ positions [i″, j″ − 1] in U taken into account
return cnt

As |Z| ≤ yocc, the countPL() and the findPL() queries may thus be faster on the TLBW-index, |Z| depending of the compression BD can do on Y.

Note that the countPL(P, Lx) query could be faster if the WT was directly built on the labels, without the intermediate bit vector BD (BA): the answer would be known while reaching the leaf of the WT in O(|P| + log l) time. We chose to favor the execution time of the findPL(P, Lx) query, as well as the size of the structure (when A can be compressed).

The findPL(P, Lx) can vary to find the patterns P which have the label Lx on the ith position of the pattern, or in any pattern’s position. Adapting this queries is easy in the TL-index as we find the positions of the pattern in the BWT, translate them in the text order and then find the label of the ith position following each of them. In the TLBW-index, we find the patterns’ positions in the BWT, access to the ith letter (we need to sample the BWT to read the patterns in the forward way), and find the label as usual. To have the label in any pattern’s position, in the TL-index we need to find a label Lx between the first and last letter of the pattern (with only two access in the WT) but in the TLBW-index, we look for the label of all the pattern’s letters.

Construction and space

We recall that the text (T, A) is of length n and is labeled with l unique labels. As defined above, the indexes store U in nHk(T) + o(n) bits. The TL-index stores the bit vector with rank and select capabilities in nH0(BA) + o(n) bits. The size of WA depends on the compressed label sequence A′, of length a. WA takes a H0(A′) + o(a log l) bits. Similarly, the TLBW-index stores BD in nH0(BD) + o(n) bits and WD takes d H0(D′) + o(d log l) bits, where d is the length of D′. The BWT can be built in linear time while using little space (Belazzougui et al., 2016; Munro, Navarro & Nekrich, 2017). BA is built while reading A in O(n) time. To make BD, we need to read the labels in the order of the original data file in O(n) time. To make WA, we find the occurrence of each label, corresponding to a 1-bit in BA, in O(a) time. Then we form the shape of WA in O(l). The labels corresponding to a 1-bit are extracted to make the WT’s root q0. For each node containing at least two labels, we separate them by following the shape previously calculated, in O(alogl/loga). We build WD the same way. The TL-index has thus a size of nHk(T) + nH0(BA) + a H0(A′) + o(n log l) bits, assuming σ = O(l), and is built in O(n+l+alogl/loga) time. The TLBW-index has a size of nHk(T) + d H0(D′) + o(n log l) bits and is built in O(n+l+dlogl/logd) time.

Shaping the WT for a label hierarchy

Labels may be further organized into a label hierarchy, given an additional set F = {F0, …, Ff − 1} of label families (Fig. 5A). Both TL- and TLBW-indexes can be adapted: The WT W (either WA or WD) will be shaped according to the hierarchy, and internal nodes q of W may have non-empty q.label values belonging to F. For example, in Fig. 5B, one can set on either index q1.label = L1, where L1 is the label family gathering L1.1, L1.2 and L1.3.

Figure 5. A label n-ary hierarchy (A) can be represented with a binary tree shaping the Wavelet Tree (B).

Figure 5

The label family L1 has here three descendants, L1.1, L1.2 and L1.3.

The findL() and findPL() queries naturally extend to label families. With the hierarchy depicted in Fig. 5, findL(L1) has to find the positions that have a label L1.1, L1.2 or L1.3. Such a query does not need to be iterated on every label of the family L1, but rather directly starts at the corresponding internal node (q1 on Fig. 5B).

Shaping W for a label hierarchy may increase the height w of the WT to O(l) in the worst case. To have a lower space consumption and a faster average query time, one can give a Huffman shape to W (Huffman, 1952). A leaf which corresponds to a frequently used label will be higher in the tree than a leaf of a rarely used label. Depending on the label hierarchy, the average height of the tree is H0(A′) in the best case while in the worst case it could be O(l). If no label hierarchy is given, the average height w will be H0(A′) (Mäkinen & Navarro, 2005).

Results and Discussion

HT-index: a baseline index

We compared the TL- and TLBW-indexes with a baseline solution called HT-index, indexing the text T with a BWT. The labels are stored in a map linking each label to the list of its (start,end) positions. We also store the labels in the text order with the compressed bit vector BA and, stored in plain form, A′. This enables the findL(Lx) query in O(y), where Y′ is the list of pairs (start,end) which represent the occurrences and y = |Y′|. Note that Y (in the TL-index) and Y′ represent the same information as the labels are stored in the text order in both indexes. The label(ti) query runs in O(1) time. This solution is not space-efficient with repeated labels: it needs nHk(T) + nH0(BA) + a + l′ + o(n) bits, where a is the size of A′ and l′ the number of labeled factors. The query times are summarized in Table 1.

Table 1. Query time complexities for HT-index, TL-index and TLBW-index.

Requests HT-index TL-index TLBW-index
label(i) O(1) O(log l) O(log1+ε n + log l)
findP(P) O(|P| + occP × log1+ε n)
findL(Lx) O(y) O(y × log l) O(y × (log l + log1+ε n))
countPL(P, Lx) O(|P| + occP × log1+ε n) O(|P| + occp × (log1+ε n + log l)) O(|P| + |Z| × log l)
findPL(P, Lx) O(|P| + |Z| × log l + y × log1+ε n)

Note:

Note that we have |Z| ≤ y ≤ occp. The label(i) and findL(Lx) queries are faster in the HT-index and the TL-index as the HT-index needs a sampling time. However, the countPL(P, Lx) and findPL(P, Lx) are faster in the HT-index.

Evaluation procedure

The three indexes were implemented in C++. We used the SDSL-Lite library (Gog et al., 2014) to build the bit vectors and the WT. We used the RopeBWT2 library (Li, 2014), which builds a BWT in O(n log n) time on small DNA sequences, as it is very efficient for storing and appending sequences corresponding to our projected application (Cox et al., 2012). As RopeBWT2 does not sample the suffix array, we iterate over the text until we find a $ symbol. To have results close to the usual FM-index sampling in O(log1+ε n) steps, we use sequences of length 50, which is similar to the sampling distance usually chosen in practice. The real files have longer sequences, thus longer sampling distances. The queries relying on the BWT will be slower and therefore cannot be compared between real and simulated files. We build the BA (or BD) bit vectors, compress them using the rrr_vector class of SDSL-Lite, and finally build WA (or WD) using a shape we added in SDSL-Lite, which depends of the label hierarchy.

We evaluated the build time, the index size, as well as the run time of three of the queries detailed in “Queries”—findP(P) behaving similarly in all indexes and countPL(P, Lx) being very similar to findPL(P, Lx). The three indexes were tested on various datasets of labeled texts, each with 100 M characters (Table 2). Datasets and code are available at http://www.vidjil.org/data/#2018-peerjcs.

Table 2. Size, build and query times of three indexes indexing labeled texts, on three simulated files and on a genomic DNA sequence file with immunologic labels.

Random Fixed #1 Fixed #2 Genomic
Sequence size 50 50 50 264
Lab. (t/u) 3.7 M / 1,000 4 M / 100 4 M / 10 852 K / 355
Lab. avg size 12.6 25 5 110.3
Lab. letters (%) 47 100 20 92
Variation (%) 100 5 50 ??
a = …/d = … 0.06n/0.77n 0.06n/0.16n 0.08n/0.36n 0.016n/0.26n
TL TLBW HT TL TLBW HT TL TLBW HT TL TLBW HT
Size (MB) 104 184 217 33 38 114 99 116 209 13 33 35
Time (s) 18 80 15 13.2 77 11 14.6 73.8 13.2 10.2 65 10.2
label(ti) (μs) 3.84 21.54 0.73 4.50 20.02 0.55 1.22 21.2 0.67 2.98 81.1 0.37
findL(L) (μs/l) 0.4 34 0.4 0.12 13.1 0.20 0.40 21.41 0.21 0.04 52.5 0.17
findPL(P, L) (s) 34.7 5.13 29.7 26.0 3.19 20.9 30.64 3.00 28.91 131.1 3.83 127.8

Notes:

All files have 100 M characters, and differ by the number of total and unique labels (“Lab (t/u)”) and their size (“Lab. avg size”), by the ratio of labeled letters (“Lab. letters”), and by the variation between sequences labeled by the same label (“Variation”). Queries use patterns P with three letters. Times were averaged on 1 M launches (label()) or at least five launches (other queries). Times for findL(L) are reported per letter. Best or close-to-the-best results are in bold.

  • Simulated files with random sequences and random labels. All sequences and labels are random (d ∼ 0.8n).

  • Simulated files, with random sequences but fixed labels. Here a given label is associated to the same pattern, possibly with some variation, and we alter the proportion of labeled letters (5–100%), the variation in the label’s pattern (0–50%, more variations giving random labels), the number of unique labels (10–1,000), the length of the labels (5–100 letters). The dataset has 546 files, two of those files are shown in Table 2.

  • Genomic sequences with immunologic labels. A person’s immune system can be described with a set of labeled DNA sequences with V(D)J recombinations. The dataset, detailed below, uses 838 K labels from 355 unique labels, with d ∼ 0.26n.

A dataset of DNA sequences with immunologic labels

The adaptive immunity is the mechanism thanks to which our body defends itself against infections. When B and T-cells, or lymphocytes, are under maturation, their DNA undergo a recombination which allows several billions possibilities from a register of a thousand genes (Tonegawa, 1983). For example, the V(D)J recombination V4*02 4/ACGT/0 J1*12 means that the DNA sequence is composed from the V4*02 gene without the four last letters, then the ACGT sequence, then the J1*12 gene.

A person’s immune system can thus be described with a set of labeled DNA sequences encoding V(D)J recombinations (Fig. 6). These sequences can be sampled by next-generation sequencing with bioinformatics analysis (Bystry et al., 2016, Duez et al., 2016).

Figure 6. V(D)J recombinations.

Figure 6

The first sequence is an immunoglobulin “light chain,” that is a VJ recombination with two labels (one V gene, positions 0–9, and one J gene, positions 17–22). The second sequence is a “heavy chain,” that is a V(D)J recombination.

The tested DNA sequences come from patients 09, 12, 14 and 63 from a public dataset on a study on acute lymphoblastic leukemia (Salson et al., 2017). They have 100 M letters and 838 K labels from 355 unique labels, making a 117 MB file. Each DNA sequence has between 100 and 350 letters and two or three labels, each label having a size between 5 and 200 letters (Fig. 7). For a given label, the labeled subsequences may vary up to 15% due to amplification and sequencing errors.

Figure 7. Two annotated sequences from the dataset.

Figure 7

Results

Index sizes. The Table 1 shows the results. As expected, the size of U, B (either BA or BD) and W (either WA or WD) grows linearly with the number of indexed elements (data not shown). The TL-index is the smallest index, and the TLBW-index is generally slightly larger. The compression is directly related to a and d. The file with random labels (d = 0.77n) is hard to compress, whereas the files with a low d/n ratio give a 2× to 7× compression. Figure 8 further shows how these sizes vary: As expected, the indexes are larger when there are less consecutive identical labels in T or in U, thus more 1-bits in B. Note that when there are more labeled letters, the text is more similar (as labels are placed on similar substrings), hence a decrease in the BWT size (A, C). W increases while the number of unique labels increases (D), the height of W increasing logarithmically with the number of unique labels.

Figure 8. Size of the indexes and size of the underlying BWT in the 546 files with fixed labels.

Figure 8

The additional size in TL- and TLBW-indexes mostly depends from the size of the compressed label sequences A′ and D′. They grow when there are more labeled letters (A), more variation in the labels (B), or when the number of distinct labels increase (D). Note that when all the letters are labeled (A, 100%), there is a small decrease in the index size because there is no random letters between the patterns. The indexes shrink when the labels grow (C), as there are more common suffixes in the label sequences.

Build time. Most of the build time of TLBW-index is devoted to build D′. For TL-index, the building of U takes most of the total time.

Queries. The label() query is faster on the HT-index. As expected, the TLBW-index needs more time for label() and findL(), as it needs to translate positions from the BWT back to the text. Note that locating the positions in the text takes about the same time as label(ti) in the TL-index. However, for the complex findPL(P, L) query, the TLBW-index is the fastest solution because the position translation is only done on the letters which have both the label and the pattern. For the TL-index and HT-index, the actual time of the findPL(P, L) queries is more affected by the number of pattern occurrences than the number of final occurrences (between 0 and 100 K depending on the file).

On the genomic dataset, the sequences are longer: The TLBW-index suffers here even more from the missing suffix array sampling of the implementation for queries label() and findL(). However, on the findPL(P, L) query, the other indexes are penalized due to the sparsity of the sampling, bringing a more than 30× difference with TLBW-index.

Conclusion

The TL-index and TLBW-index store a labeled text and allow efficient queries. They can be constructed from files of some MB in a few seconds. Experiments confirm that the indexes stay small when the text is redundant (thus a smaller U), when each label describes a pattern with few variations (many 0-bits in B, thus a smaller W), and when few letters are labeled (thus a small W). However, the TL-index and TLBW-index are robust even to non-redundant text with almost random labels everywhere. The TLBW-index needs more memory space than the TL-index but is more efficient in combined label/pattern queries. Those structures might be used on any labeled data, such as DNA sequences, but also on natural language texts or on music sheets with some semantic annotations.

Perspectives include improvement of the implementation, with label families queries or parameterizing the distance between samples in the FM-index to offer a space-time trade off. Within SDSL we could use the sd_vector bit vector instead of the rrr_vector bit vector which should improve space consumption when the bit vectors are very sparse. However, this would only minimally improve the global space consumption of the index. We plan to use one of the indexes in a clone database for hematological malignancies: It will allow comparison of V(D)J recombinations between different samples of a patient or between several patients.

Supplemental Information

Supplemental Information 1. Code.
DOI: 10.7717/peerj-cs.148/supp-1
Supplemental Information 2. Raw Data.
DOI: 10.7717/peerj-cs.148/supp-2

Acknowledgments

We thank anonymous reviewers for their insightful comments on earlier versions of this article, as well as the EuroClonality-NGS consortium for insightful discussions.

Funding Statement

This work was supported by Université de Lille, SIRIC ONCOLille, and Région Hauts-de-France. The funders had no role in study design, data collection and analysis, decision to publish, or preparation of the manuscript.

Contributor Information

Mathieu Giraud, Email: mathieu.giraud@univ-lille.fr.

Mikaël Salson, Email: mikael.salson@univ-lille.fr.

Additional Information and Declarations

Competing Interests

The authors declare that they have no competing interests.

Author Contributions

Tatiana Rocher conceived and designed the experiments, performed the experiments, analyzed the data, prepared figures and/or tables, performed the computation work, authored or reviewed drafts of the paper.

Mathieu Giraud conceived and designed the experiments, analyzed the data, authored or reviewed drafts of the paper.

Mikaël Salson conceived and designed the experiments, analyzed the data, authored or reviewed drafts of the paper.

Data Availability

The following information was supplied regarding data availability:

Code and data are available as Supplemental Dataset Files and from http://www.vidjil.org/data/#2018-peerjcs.

References

  • Arroyuelo et al. (2015).Arroyuelo D, Claude F, Maneth S, Mäkinen V, Navarro G, Nguyen K, Sirén J, Välimäki N. Fast in-memory XPath search using compressed indexes. Software: Practice and Experience. 2015;45(3):399–434. doi: 10.1002/spe.2227. [DOI] [Google Scholar]
  • Belazzougui et al. (2016).Belazzougui D, Cunial F, Kärkkäinen J, Mäkinen V. Linear-time string indexing and analysis in small space. arXiv Preprint. 20161609.06378
  • Burrows & Wheeler (1994).Burrows M, Wheeler DJ. Digital Equipment Corporation. SRC Research Report; 1994. A block-sorting lossless data compression algorithm; p. 124. [Google Scholar]
  • Bystry et al. (2016).Bystry V, Reigl T, Krejci A, Demko M, Hanakova B, Grioni A, Knecht H, Schlitt M, Dreger P, Sellner L, Herrmann D, Pingeon M, Boudjoghra M, Rijntjes J, Pott C, Langerak AW, Groenen PJTA, Davi F, Brüggemann M, Darzentas N. ARResT/Interrogate: an interactive immunoprofiler for IG/TR NGS data. Bioinformatics. 2016;33(3):435–437. doi: 10.1093/bioinformatics/btw634. [DOI] [PubMed] [Google Scholar]
  • Cox et al. (2012).Cox AJ, Bauer MJ, Jakobi T, Rosone G. Large-scale compression of genomic sequence databases with the Burrows–Wheeler transform. Bioinformatics. 2012;28(11):1415–1419. doi: 10.1093/bioinformatics/bts173. [DOI] [PubMed] [Google Scholar]
  • Duez et al. (2016).Duez M, Giraud M, Herbert R, Rocher T, Salson M, Thonier F. Vidjil: a web platform for analysis of high-throughput repertoire sequencing. PLOS ONE. 2016;11(11):e0166126. doi: 10.1371/journal.pone.0166126. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • Ferragina & Manzini (2000).Ferragina P, Manzini G. Opportunistic data structures with applications. Foundations of Computer Science (FOCS 2000); Piscataway: IEEE; 2000. pp. 390–398. [Google Scholar]
  • Gog et al. (2014).Gog S, Beller T, Moffat A, Petri M. From theory to practice: plug and play with succinct data structures. Symposium on experimental algorithms (SEA 2014); 2014. pp. 326–337. [Google Scholar]
  • Grossi, Gupta & Vitter (2003).Grossi R, Gupta A, Vitter J. High-order entropy-compressed text indexes. Symposium on discrete algorithms (SODA 2003); New York: ACM; 2003. pp. 841–850. [Google Scholar]
  • Huffman (1952).Huffman DA. A method for the construction of minimum-redundancy codes. Proceedings of the IRE. 1952;40(9):1098–1101. [Google Scholar]
  • Kärkkäinen & Ukkonen (1996).Kärkkäinen J, Ukkonen E. Lempel-Ziv parsing and sublinear-size index structures for string matching. South American workshop on string processing (WSP 96).1996. [Google Scholar]
  • Li (2014).Li H. Fast construction of FM-index for long sequence reads. Bioinformatics. 2014;30(22):3274–3275. doi: 10.1093/bioinformatics/btu541. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • Mäkinen & Navarro (2004).Mäkinen V, Navarro G. Run-length FM-index. Proceedings of the DIMACS workshop: the Burrows-Wheeler Transform: ten years later; 2004. pp. 17–19. [Google Scholar]
  • Mäkinen & Navarro (2005).Mäkinen V, Navarro G. Succinct suffix arrays based on run-length encoding. Combinatorial Pattern Matching (CPM 2005); 2005. pp. 45–56. [Google Scholar]
  • Mäkinen & Navarro (2007).Mäkinen V, Navarro G. Rank and select revisited and extended. Theoretical Computer Science. 2007;387(3):332–347. doi: 10.1016/j.tcs.2007.07.013. [DOI] [Google Scholar]
  • Munro, Navarro & Nekrich (2017).Munro JI, Navarro G, Nekrich Y. Space-efficient construction of compressed indexes in deterministic linear time. Symposium on discrete algorithms (SODA 2017); 2017. pp. 408–424. [Google Scholar]
  • Munro, Nekrich & Vitter (2016).Munro JI, Nekrich Y, Vitter JS. Fast construction of wavelet trees. Theoretical Computer Science. 2016;638:91–97. doi: 10.1016/j.tcs.2015.11.011. [DOI] [Google Scholar]
  • Navarro & Mäkinen (2007).Navarro G, Mäkinen V. Compressed full-text indexes. ACM Computing Surveys (CSUR) 2007;39(1):2. doi: 10.1145/1216370.1216372. [DOI] [Google Scholar]
  • Raman, Raman & Rao (2002).Raman R, Raman V, Rao SS. Succinct indexable dictionaries with applications to encoding k-ary trees and multisets. Symposium on discrete algorithms (SODA 2012); 2002. pp. 233–242. [Google Scholar]
  • Salson et al. (2017).Salson M, Giraud M, Caillault A, Ferret Y, Duployez N, Duez M, Sebda S, Quief S, Villenet C, Figeac M, Preudhomme C, Grardel N. High-throughput sequencing in acute lymphoblastic leukemia: Follow-up of minimal residual disease and emergence of new clones. Leukemia Research. 2017;53:1–7. doi: 10.1016/j.leukres.2016.11.009. [DOI] [PubMed] [Google Scholar]
  • Tonegawa (1983).Tonegawa S. Somatic generation of antibody diversity. Nature. 1983;302(5909):575–581. doi: 10.1038/302575a0. [DOI] [PubMed] [Google Scholar]

Associated Data

This section collects any data citations, data availability statements, or supplementary materials included in this article.

Supplementary Materials

Supplemental Information 1. Code.
DOI: 10.7717/peerj-cs.148/supp-1
Supplemental Information 2. Raw Data.
DOI: 10.7717/peerj-cs.148/supp-2

Data Availability Statement

The following information was supplied regarding data availability:

Code and data are available as Supplemental Dataset Files and from http://www.vidjil.org/data/#2018-peerjcs.


Articles from PeerJ Computer Science are provided here courtesy of PeerJ, Inc

RESOURCES