Abstract / Description of output
One of the major challenges providing large databases like the WFCAM
Science Archive (WSA) is to minimize ingest times for pixel/image
metadata and catalogue data. In this article we describe how the
pipeline processed data are ingested into the database as the first
stage in building a release database which will be succeeded by advanced
processing (source merging, seaming, detection quality flagging etc.).
To accomplish the ingestion procedure as fast as possible we use a mixed
Python/C++ environment and run the required tasks in a simple parallel
modus operandi where the data are split into daily chunks and then
processed on different computers. The created data files can be ingested
into the database immediately as they are available. This flexible way
of handling the data allows the most usage of the available CPUs as the
comparison with sequential processing shows.
Original language | English |
---|---|
Pages (from-to) | 403 |
Journal | Astronomical Society of the Pacific Conference Series |
Volume | 394 |
Publication status | Published - 1 Aug 2008 |