Skip to content

Commit

Permalink
minor fixes
Browse files Browse the repository at this point in the history
  • Loading branch information
manathan1984 committed Jun 21, 2024
1 parent 9d3cd31 commit f570847
Showing 1 changed file with 8 additions and 7 deletions.
15 changes: 8 additions & 7 deletions hdms/2024/program.html
Original file line number Diff line number Diff line change
Expand Up @@ -366,10 +366,10 @@ <h4> Download the program
<td class="program-sessions">1:10 - 1:20pm</td>
<td class="program-sessions"> <i>Chimp: Efficient Lossless Floating Point Compression for Time Series Databases</i>
<br> Liakos, Panagiotis; Papakonstantinopoulou, Katia; Kotidis, Yannis
<!-- <br> <a class="toggle-abstract" onclick="toggleAbstract('abstract-16')">Click to display the abstract</a>
<br> <a class="toggle-abstract" onclick="toggleAbstract('abstract-16')">Click to display the abstract</a>
<span class="paper-abstract" id="abstract-16">
</span> -->
Applications in diverse domains such as astronomy, economics and industrial monitoring, increasingly press the need for analyzing massive collections of time series data. The sheer size of the latter hinders our ability to efficiently store them and also yields significant storage costs. Applying general purpose compression algorithms would effectively reduce the size of the data, at the expense of introducing significant computational overhead. Time Series Management Systems that have emerged to address the challenge of handling this overwhelming amount of information, cannot suffer the ingestion rate restrictions that such compression algorithms would cause. Data points are usually encoded using faster, streaming compression approaches. However, the techniques that contemporary systems use do not fully utilize the compression potential of time series data, with implications in both storage requirements and access times. In this work, we propose a novel streaming compression algorithm, suitable for floating point time series data. We empirically establish properties exhibited by a diverse set of time series and harness these features in our proposed encodings. Our experimental evaluation demonstrates that our approach readily outperforms competing techniques, attaining compression ratios that are competitive with slower general purpose algorithms, and on average around 50% of the space required by state-of-the-art streaming approaches. Moreover, our algorithm outperforms all earlier techniques with regards to both compression and access time, offering a significantly improved trade-off between space and speed. The aforementioned benefits of our approach - in terms of all space requirements, compression time and read access - significantly improve the efficiency in which we can store and analyze time series data.
</span>
</td>
</tr>
<tr>
Expand All @@ -382,7 +382,7 @@ <h4> Download the program
</tr>
<tr>
<td class="program-sessions">3:30 - 4:00pm</td>
<td class="program-sessions"><b>Keynote 2:</b> Verena Kantere </td>
<td class="program-sessions"><b>Keynote 2:</b> Verena Kantere (Univeristy of Ottawa) </td>
</tr>
<!-- <tr>
<td class="program-sessions">3:45 - 4:00pm</td>
Expand Down Expand Up @@ -435,7 +435,7 @@ <h4> Download the program
</tr>
<tr>
<td class="program-sessions">9:30 - 10:00am</td>
<td class="program-sessions"> <b> keynote 3: Data Management innovation at Amazon Web Services </b> Ippokratis Pandis – Vice President/Distinguished Engineer, Amazon Web Services </td>
<td class="program-sessions"> <b> Keynote 3: <i>Data Management innovation at Amazon Web Services </i></b> Ippokratis Pandis (Amazon Web Services) </td>
</tr>
<tr>
<td colspan="2" class="program-session-titles"> <b> Session 1: "Query Processing & Execution" </b> </tr>
Expand Down Expand Up @@ -693,9 +693,10 @@ <h4> Download the program
<td class="program-sessions">12:50 - 1:00pm</td>
<td class="program-sessions"> <i> Proportionality on Spatial Data with Context </i>
<br> Fakas, George; Kalamatianos, Georgios;
<!-- <br> <a class="toggle-abstract" onclick="toggleAbstract('abstract-30')">Click to display the abstract</a>
<br> <a class="toggle-abstract" onclick="toggleAbstract('abstract-30')">Click to display the abstract</a>
<span class="paper-abstract" id="abstract-30">
</span> -->
More often than not, spatial objects are associated with some context, in the form of text, descriptive tags (e.g., points of interest, flickr photos), or linked entities in semantic graphs (e.g., Yago2, DBpedia). Hence, location-based retrieval should be extended to consider not only the locations but also the context of the objects, especially when the retrieved objects are too many and the query result is overwhelming. In this article, we study the problem of selecting a subset of the query result, which is the most representative. We argue that objects with similar context and nearby locations should proportionally be represented in the selection. Proportionality dictates the pairwise comparison of all retrieved objects and hence bears a high cost. We propose novel algorithms which greatly reduce the cost of proportional object selection in practice. In addition, we propose pre-processing, pruning, and approximate computation techniques that their combination reduces the computational cost of the algorithms even further. We theoretically analyze the approximation quality of our approaches. Extensive empirical studies on real datasets show that our algorithms are effective and efficient. A user evaluation verifies that proportional selection is more preferable than random selection and selection based on object diversification.
</span>
</td>
</tr>
<tr>
Expand Down

0 comments on commit f570847

Please sign in to comment.