Data Grid
9 Followers
Recent papers in Data Grid
In this paper medical applications on a Grid infrastructure, the MAGIC-5 Project, are presented and discussed. MAGIC-5 aims at developing Computer Aided Detection (CADe) software for the analysis of medical images on distributed databases... more
A high throughput Basic Local Alignment Search Tool (BLAST) system based on Web services is implemented. It provides an alternative BLAST service and allows users to perform multiple BLAST queries at one run in a distributed, parallel... more
- by jiren wang
The Network Common Data Form (netCDF) is one of the primary methods of self-documenting community and recent evolution toward tion via messages in the defacto standard XML language. XML is a text-based language while netCDF is based on a... more
- by Ben Domenico
Data replication is the creation and maintenance of multiple copies of the same data. Replication is used in Data Grid to enhance data availability and fault tolerance. One of the main objectives of replication strategies is reducing... more
Abstract. The Social Informatics Data Grid is a new infrastructure designed to transform how social and behavioral scientists collect and annotate data, collaborate and share data, and analyze and mine large data repositories. An... more
High resolution climatology-towards climate change services
Large distributed systems such as Computational/Data Grids require large amounts of data to be colocated with the computing facilities for processing. Ensuring that the data is there in time for the computation in today's Internet is a... more
Selecting optimal resources for submitting jobs on a computational Grid or accessing data from a data grid is one of the most important tasks of any Grid middleware. Most modern Grid software today satisfies this responsibility and gives... more
Selecting optimal resources for submitting jobs on a computational Grid or accessing data from a data grid is one of the most important tasks of any Grid middleware. Most modern Grid software today satisfies this responsibility and gives... more
Grid computing has recently gained in popularity. Grid applications can be very demanding of the data storage facilities in the Grid. The existing data grid services are often insufficient and additional optimization of the data access is... more
The Sloan Digitial Sky Surveys (SDSS) have been collecting imaging and spectoscopic data since 1998. These data as well as their derived data products are made publicly available through regular data releases, of which the 13th took place... more
- by Ani Thakar
- Physics, QC, Sky
In data grids, data replication on variant nodes can change some problems such as response time and availability. Also, in data replication, there are some challenges to finding the best replica efficiently in relation to performance and... more
We propose a novel approach for joint denoising and interpolation of noisy Bayer-patterned data acquired from a digital imaging sensor (e.g., CMOS, CCD). The aim is to obtain a full-resolution RGB noiseless image. The proposed technique... more
Abstract: Recently, there have been many efforts to develop middlewares supporting applications in managing distributed data in Grid computing environment. Although these research activities address various requirements in Grid data... more
In different scientific disciplines, large-scale data are generated with enormous storage requirements. Therefore, effective data management is a critical issue in distributed systems such as the cloud. As tasks can access a nearby site... more
GriPhyN (Grid Physics Network) is a large US collaboration to build grid services for large physics experiments, one of which is LIGO, a gravitational-wave observatory. This paper explains the physics and computing challenges of LIGO, and... more
The success of grid computing depends on the existence of grid middleware that provides core services such as security, data management, resource information, and resource brokering and scheduling. Current general-purpose grid resource... more
The Grid provides mechanisms to share dynamic, heterogeneous, distributed resources spanned across multiple administrative domains. Resources required to execute a job are identified from the resource pool based on the desired set of... more
Current trend of network-based multimedia storage, distributed scientific simulations and distributed geographic information system applications are require both compute intensive and data intensive. This is unavoidable since grid and... more
The SHAMAN project targets a framework integrating advances in the data grid, digital library, and persistent archives communities in order to archive a longterm preservation environment. Within the project we identified several... more
Data warehouse view maintenance is an important issue due to the growing use of warehouse technology for information integration and data analysis. Given the dynamic nature of modern distributed environments, both data updates and schema... more
Some of the most recently proposed algorithms for the incremental maintenance of materialized data warehouses (DW), such as SWEEP and PSWEEP, offer several significant advantages over previous solutions, such as high-performance, no... more
High resolution climatology-towards climate change services
We describe a Grid market for exchanging data mining services based on the Catallactic market mechanism proposed by von Hayek. This market mechanism allows selection between multiple instances of services based on operations required in a... more
Nowadays, the process of data mining is one of the most important topics in scientific and business problems. There is a huge amount of data that can help to solve many of these problems. However, data is geographically distributed in... more
The Italica Project is the implementation of an Electronic Health Record system at the Italian Hospital of Buenos Aires. The present work shows the implementation of a Medical Signal Grid Repository module and its integration to the... more
Data integration over multiple heterogeneous data sources has become increasingly important for modern applications. The integrated data is usually stored in materialized views for better access, performance and high availability. Such... more
Data integration over multiple heterogeneous data sources has become increasingly important for modern applications. The integrated data is usually stored in materialized views (MV) to allow better access, performance and high... more
Data integration over multiple heterogeneous data sources has become increasingly important for modern applications. The integrated data is usually stored as materialized views to allow better access, performance, and high availability.... more
The Grid-based Virtual Laboratory AMsterdam (VLAM-G) provides a science portal for distributed analysis in applied scientific research. By facilitating access to distributed compute and information resources held by multiple... more
Data Grid environment is a geographically distributed that deal with date-intensive application in scientific and enterprise computing. Dealing with large amount of data makes the requirement for efficiency in data access more critical.... more
Recently, Grid computing activities in Asia-Pacific have been drawn attention, includes in high learning education institutes in Malaysia. Many university and institute in Malaysia are started to build their campus grid infrastructure.... more
Summary: Timely worldwide distribution of biosequence and bioinformatics data depends on high performance networking and advances in Internet transport methods. The Bio-Mirror project focuses on providing up-to-date distribution of this... more
Data Grids deal with geographically-distributed large-scale data-intensive applications. Schemes scheduled for data grids attempt to not only improve data access time, but also aim to improve the ratio of data availability to a node,... more
An SDDS-2000 server currently manages only buckets in its RAM storage [C01]. Several buckets can coexist. When many files are created however, RAM storage space may not be sufficient for all the buckets simultaneously. When an application... more
The CMS experiment at CERN is setting up a Grid infrastructure required to fulfil the needs imposed by Terabyte scale productions for the next few years. The goal is to automate the production and at the same time allow the users to... more