Versions Compared

Key

  • This line was added.
  • This line was removed.
  • Formatting was changed.

...

The raw data consists of event records from the CODA data acquisition system. These records contain the digitized hit information from the various detectors elements, including, for example, drift times from tracking chambers, hodoscope hits, scaler values, etc. These data will be stored onsite at Fermilab in the experimental counting house on a RAID disk array. In addition, new data are copied daily to the Fermilab STKEN Enstore system (which is located in a separate building) for additional protection against data loss. 

The raw data are then decoded and stored in a MySQL database. The decoding takes the information in the raw CODA data records and translates them into a more user-friendly format, for example, assigning specific wires numbers in tracking chambers to digitized drift time information or hodoscope numbers to hits. Further processing then occurs on these data to change the hits into reconstructed tracks and events that are also stored in the MySQL database. The MySQL database is also hosted on site in the SpinQuest counting house. For ease of access and data security, the MySQL database is mirrored off site at the University of Illinois on a RAID system, and possibly other collaboration sites in the future. The source code, related calibrations, alignment data, etc. needed to translate the raw data to the MySQL database is under Subversion (SVN) version control. A second copy of this information is also maintained at the University of Illinois. MySQL is an open source database system that is widely available and well supported.

It is the SpinQuest Collaboration’s policy that these raw data and processed MySQL data are available to collaboration members for use in collaboration-approved scientific studies and analyses. Completed analyses will be submitted for publication and shared with outside researchers. SpinQuest will maintain the ability to access these data for a minimum of 7 years after the completion of the experiment.

...

Databases: Database servers are managed by IT SpinQuest and regular snapshots of the database content are stored along with the tools and documentation required for their recovery.

Log Books: Jefferson lab uses SpinQuest uses an electronic logbook system SpinQuest ECL with a database back-end. Calibration and Geometry databases: Running conditions, as well as the detector calibration constants and detector geometries are stored in a database at Fermilab.

Analysis software source: Data analysis software is developed within the SpinQuest reconstruction and analysis package. Contributions to the package are from several sources, university faculty, Fermilab users, off-site lab collaborators and third parties. Locally written software source code and build files, along with contributions from collaborators are stored in a version management system, git. Third party software is managed by software maintainers under oversight of the Analysis Working Group. Source code repositories and managed third party packages are continually backed up to the University of Virginia Rivanna storage.

Documentation: Documentation is available online in the form of content either maintained by a content management system (CMS) such as a Wiki in Githut or Confluence pagers or as static web pages. This content is backed up continually. Other documentation for the software is distributed via wiki pages, and consists of a combination of html and pdf files.

The Fermilab E-1039/SpinQuest experiment is expecting to collect approximately 20 Tb of raw data between commissioning and the end of data acquisition in 2022. The raw data is subsequently processed and stored in a MySQL database. The MySQL database will be approximately twice the size of the raw data.  In addition, there is a substantial volume of simulated, Monte Carlo events produced and stored at collaborating institutions and universities.

The raw data consists of event records from the CODA data acquisition system. These records contain the digitized hit information from the various detectors elements, including, for example, drift times from tracking chambers, hodoscope hits, scaler values, etc. These data will be stored onsite at Fermilab in the experimental counting house on a RAID disk array. In addition, new data are copied daily to the Fermilab STKEN Enstore system (which is located in a separate building) for additional protection against data loss.

The raw data are then decoded and stored in a MySQL database. The decoding takes the information in the raw CODA data records and translates them into a more user-friendly format, for example, assigning specific wires numbers in tracking chambers to digitized drift time information or hodoscope numbers to hits. Further processing then occurs on these data to change the hits into reconstructed tracks and events that are also stored in the MySQL database. The MySQL database is also hosted on site in the SpinQuest counting house. For ease of access and data security, the MySQL database is mirrored off site at the University of Illinois on a RAID system, and possibly other collaboration sites in the future. The source code, related calibrations, alignment data, etc. needed to translate the raw data to the MySQL database is under Subversion (SVN) version control. A second copy of this information is also maintained at the University of Illinois. MySQL is an open source database system that is widely available and well supported.

It is the SpinQuest Collaboration’s policy that these raw data and processed MySQL data are available to collaboration members for use in collaboration-approved scientific studies and analyses. Completed analyses will be submitted for publication and shared with outside researchers. SpinQuest will maintain the ability to access these data for a minimum of 7 years after the completion of the experiment.

Contribution from Fermilab Scientific Computing Devision:

...

Provide appropriate networking at NM4 hall including WiFi in both the counting area and detector hall for commissioning, data transfers to mass storage, network access for users’ laptops, etc. Provide firewalls/bridges which Fermilab deems necessary to isolate the experiment’s network from the general Fermilab network.

...

Provide “General Computing” accounts for collaborators. Primary analysis and Monte Carlo computing will be done on LINUX-based PC’s provided by the collaboration.

...

Provide storage for 50 TB of raw data. The collaboration also plans to keep a second copy of the raw data on a separate disk system.

...

Support for 4 virtual machines.

...