-
Book Overview & Buying
-
Table Of Contents
-
Feedback & Rating
NoSQL Data Models
By :
As discussed in section 7.3.2, TDV computation is a very time-consuming process. We need to extract similarities between all pairs of items in a collection for N times: one without all the terms, and N times by removing a term to compute the density of this term. Consequently, this process cannot be done in real time and must be evaluated in parallel with the filtering process. Initially, it took two days to compute 10M items. Even if a TDV does not evolve much over time, it relies on the evolution of the presence of a term compared with all of them, on all items. The computation step needs to start from the beginning each time we need to provide new TDV.
Our work to enhance the computation of TDV updates is twofold: i) adapt TDV computation techniques in an incremental process and ii) adapt our algorithm in a distributed context in order to scale up.
TDV computation is a heavy process that requires evaluation of the density of a collection...
Change the font size
Change margin width
Change background colour