To be held in conjunction with HPDC-2011, 8 June 2011, San Jose, CA
There has been a lot of effort in managing and distributing tasks where computational loads are dominant. Such applications have after all, been historically the drivers of "grid" computing. There has, however, been relatively less effort on tasks where the computational load is matched by the data load, or even dominated by the data load. For such tasks to be able to operate at scale, there are conceptually simple run-time trade-offs that need to be made, such as determining whether to move data to compute versus keeping data localized and move computational tasks to operate on the data in situ, or possibly neither, and with data regenerated on-the-fly. Due to fluctuating resource availability and capabilities, as well as insufficient prior information about application requirements, such decisions must be made at run-time. Furthermore, resource, connectivity and/or storage constraints may require the data to be manipulated in-transit so that it is "made-right" for the consumer. Currently it is very difficult to implement these dynamic decisions or the underlying mechanisms in a general-purpose and scalable fashion.
Although the increasing volumes and complexity of data will make many problems data load dominated, the computational requirements will still be high. In practice, data-intensive applications will encompass data-driven applications. For example, many data-driven applications will involve computational activities triggered as a consequence of independently created data; thus it is imperative for an application to be able to respond to unplanned changes in data load or content. Therefore, understanding how to support dynamic computations is a fundamental, but currently missing element in data-intensive computing.
This workshop will operate at the triple point of dynamic and distributed and data-intensive (3D) attributes. This workshop will operate at the triple point of dynamic, distributed and data-intensive (3D) attributes. It will also focus on innovative approaches for scalability in the end-to-end real-time processing of scientific data. We refer to 3D applications as those are data-intensive, need to support and respond to dynamic data, and, either are fundamentally, or need to be, distributed. We are interested in papers that span the spectrum from the design of cyberinfrastructure to support 3D applications, to novel application examples. We are also looking to bring researchers together to look at holistic, rather than piecewise, approaches to the end-to-end processing and managing of scientific data.
3DAPAS builds upon a 3-year research theme on Distributed Programming Abstractions (DPA), which has held a series of related workshops (see: DPA Past Events) including but not limited to e-Science2008, EuroPar 2008 and the CLADE series. 3DAPAS will also draw on ideas from the ongoing 3DPAS Research Theme funded by the NSF and UK EPSRC.
Topics of interest include but are not limited to:
Authors are invited to submit technical papers of at most 8 pages in PDF format, including all figures and references. Papers should be formatted in the ACM Proceedings Style and submitted via EasyChair. Accepted papers will appear in the conference proceedings, and will be incorporated into the ACM Digital Library.
Submission of a paper implies that at least one author will attend the workshop to present the paper, if it is accepted.
Papers must be self-contained and provide the technical substance required for the program committee to evaluate the paper's contribution. Papers should thoughtfully address all related work. Submitted papers must be original work that has not appeared in and is not under consideration for another conference or a journal. See the ACM Prior Publication Policy for more details.
Submissions Due: 31 Jan 2011
Paper Decisions Announced: 28 Feb 2011
Final Camera-Ready Papers Due: 24 Mar 2011
Workshop Date: 8 June 2011
(all dates are firm)
Program Committee Members: