Difference between revisions of "Atlas:SC4"

Un article de lcgwiki.
Jump to: navigation, search
(Information from dCache monitoring (provided by Lyon))
(VOBOX Configuration)
Ligne 70: Ligne 70:
* 4 processors 3 GHz
* 4 processors 3 GHz
* 4 GB of memory ( 2 GB dedicated to SWAP)
* 4 GB of memory ( 2 GB dedicated to SWAP)
== Disk space availability ==
* [http://gridice2.cnaf.infn.it:50080/gridice/site/site_details.php?siteName=GRIF&visibility=SE GRIF]
* [http://gridice2.cnaf.infn.it:50080/gridice/site/site_details.php?siteName=IN2P3-LPC&visibility=SE LPC]
* [http://gridice2.cnaf.infn.it:50080/gridice/site/site_details.php?siteName=IN2P3-LAPP&visibility=SE LAPP]
* [http://gridice2.cnaf.infn.it:50080/gridice/site/site_details.php?siteName=TOKYO-LCG2&visibility=SE TOKYO]
== Daily news ==
== Daily news ==

Version du 17:29, 7 juillet 2006

Bienvenue sur la page Atlas SC4 LCG-France

Welcome to the LCG-France Atlas SC4 page

Twiki page : SC4 ATLAS

Compte rendu de la réunion SC4 ATLAS au CERN du 9 Juin (S.Jézéquel, G. Rahal) (written in french)

  • T0 Role(CERN)
    • Produce dummy files with 1 to 2 GB size(RAW, ESD et AOD) (see T0 Twiki)
    • Initiate T0->T1 transfers
    • FTS server sents files to Lyon choosing between 'TAPE' (RAW 43,2 Mo/s) or 'DISK' (ESD,AOD 23+20 Mo/s) areas
  • T1 Role (CCIN2P3)
    • Get files from T0 (dedicated dcache area: L. Schwarz)
    • Provides LFC (lfc-atlas.in2p3.fr) and FTS service (cclcgftsli01.in2p3.fr) (D. Bouvet)
    • Send all AODs to each T2 (20 Mo/s) using Lyon FTS server
    • Regurlarly cleanup files
    • Get files from T1 (Lyon). Files on the T2 are written in /home/atlas/sc4tier0/...
    • Clean-up the files (?)
  • Other roles
    • ATLAS (S. Jézéquel) : Initiate T1->T2 transfers

Information from DDM monitoring


Information from FTS monitoring

  • T1->T2 :
    • 15 concurrent files and 10 streams for LYON-TOKYO
    • 5 concurrent files and 5 streams for LYON-BEIJING (SE not enough powerful for 15/15 )
    • 10 concurrent files and 10 streams for LYON-French T2s (LAL, LAPP, LPNHE, LPC, SACLAY)

Information from dCache monitoring (provided by Lyon)

VOBOX Configuration

  • 4 processors 3 GHz
  • 4 GB of memory ( 2 GB dedicated to SWAP)

Disk space availability

Daily news

  • 20 June 2006: Mail from Miguel Branco (DDM responsible)

Today we started deploying DQ2 on the remaining T1 sites (not all sites still available).
Attached is the result of a (nice) ramp up, easily beating SC3's record (on the 1st day of export of SC4) peaking at ~ 270 MB/s. Each 'step' in the graph is an additional T1 being added to the export.
Dataset subscriptions are now slowing down and will resume tomorrow. Our DQ2 monitoring has been turned off and we expect to have it back tomorrow! Still a long way to go until we have a reasonable understanding of the limiting factors..

Atlas day-jpeg.jpg

  • 22 June: General power cut at CERN at 2:pm.
  • 24 June : Dataset T0.D.run000949.ESD transfered from Lyon to LAL and TOKYO. Tranfering the same dataset to LAPP and LPC failed because these sites have same domain name (*.in2p3.fr) as Lyon.
  • 25 June : Almost no transfer from CERN to T1s during the week-end.
  • 26 June : SC4 transfers restarted with working DDM monitoring. Successfull transfers to LAL, SACLAY and TOKYO. Technical problem (domain name) for LAPP, LPNHE and LPC : under investigation. Contact BEIJING.


  • 28 June :
    • Problem of domain name solved for LAPP, LPC and LPNHE. First transfers to these sites have been done.
    • Increase the number of LFC connection to 40 (advice from CERN-IT and DDM).

AODs were transfered to all T2s associated to Lyon except BEIJING (looks like a FTS problem)

    • Transfer all AODs to LAL. Problems to transfer AODs to LAPP (one dcache server crashed)
  • 29 June :
    • Transfer of AODs to TOKYO
  • 1 July
    • Test transfers from LYONDISK to T2s


  • 5 July :
    • First DDM transfer to BEIJING