A Pseudo-Parallel Python Environment for Database Curation

Research output: Contribution to journalArticle

Abstract

One of the major challenges providing large databases like the WFCAM Science Archive (WSA) is to minimize ingest times for pixel/image metadata and catalogue data. In this article we describe how the pipeline processed data are ingested into the database as the first stage in building a release database which will be succeeded by advanced processing (source merging, seaming, detection quality flagging etc.). To accomplish the ingestion procedure as fast as possible we use a mixed Python/C++ environment and run the required tasks in a simple parallel modus operandi where the data are split into daily chunks and then processed on different computers. The created data files can be ingested into the database immediately as they are available. This flexible way of handling the data allows the most usage of the available CPUs as the comparison with sequential processing shows.
Original languageEnglish
Pages (from-to)403
JournalAstronomical Society of the Pacific Conference Series
Volume394
Publication statusPublished - 1 Aug 2008

Fingerprint Dive into the research topics of 'A Pseudo-Parallel Python Environment for Database Curation'. Together they form a unique fingerprint.

Cite this