Результаты поиска по 'WLCG':
Найдено статей: 4
  1. Smirnova O., Kónya B., Cameron D., Nilsen J.K., Filipčič A.
    ARC-CE: updates and plans
    Computer Research and Modeling, 2015, v. 7, no. 3, pp. 407-414

    ARC Compute Element is becoming more popular in WLCG and EGI infrastructures, being used not only in the Grid context, but also as an interface to HPC and Cloud resources. It strongly relies on community contributions, which helps keeping up with the changes in the distributed computing landscape. Future ARC plans are closely linked to the needs of the LHC computing, whichever shape it may take. There are also numerous examples of ARC usage for smaller research communities through national computing infrastructure projects in different countries. As such, ARC is a viable solution for building uniform distributed computing infrastructures using a variety of resources.

  2. Zarochentsev A.K., Stiforov G.G.
    ALICE computing update before start of RUN2
    Computer Research and Modeling, 2015, v. 7, no. 3, pp. 415-419

    The report presents a number of news and updates of the ALICE computing for RUN2 and RUN3.

    This includes:

    – implementation in production of a new system EOS;

    – migration to the file system CVMFS to be used for storage of the software;

    – the plan for solving the problem of “Long-Term Data Preservation”;

    – overview of the concept of “O square”, combining offline and online data processing;

    – overview of the existing models to use the virtual clouds for ALICE data processing. Innovations are shown on the example of the Russian sites.

    Views (last year): 2.
  3. Kiryanov A.K.
    GridFTP frontend with redirection for DMlite
    Computer Research and Modeling, 2015, v. 7, no. 3, pp. 543-547

    One of the most widely used storage solutions in WLCG is a Disk Pool Manager (DPM) developed and supported by SDC/ID group at CERN. Recently DPM went through a massive overhaul to address scalability and extensibility issues of the old code.

    New system was called DMLite. Unlike the old DPM that was based on daemons, DMLite is arranged as a library that can be loaded directly by an application. This approach greatly improves performance and transaction rate by avoiding unnecessary inter-process communication via network as well as threading bottlenecks.

    DMLite has a modular architecture with its core library providing only the very basic functionality. Backends (storage engines) and frontends (data access protocols) are implemented as plug-in modules. Doubtlessly DMLite wouldn't be able to completely replace DPM without GridFTP as it is used for most of the data transfers in WLCG.

    In DPM GridFTP support was implemented in a Data Storage Interface (DSI) module for Globus’ GridFTP server. In DMLite an effort was made to rewrite a GridFTP module from scratch in order to take advantage of new DMLite features and also implement new functionality. The most important improvement over the old version is a redirection capability.

    With old GridFTP frontend a client needed to contact SRM on the head node in order to obtain a transfer URL (TURL) before reading or writing a file. With new GridFTP frontend this is no longer necessary: a client may connect directly to the GridFTP server on the head node and perform file I/O using only logical file names (LFNs). Data channel is then automatically redirected to a proper disk node.

    This renders the most often used part of SRM unnecessary, simplifies file access and improves performance. It also makes DMLite a more appealing choice for non-LHC VOs that were never much interested in SRM.

    With new GridFTP frontend it's also possible to access data on various DMLite-supported backends like HDFS, S3 and legacy DPM.

    Views (last year): 1.
  4. Berezhnaya A.Ya., Velikhov V.E., Lazin Y.A., Lyalin I.N., Ryabinkin E.A., Tkachenko I.A.
    The Tier-1 resource center at the National Research Centre “Kurchatov Institute” for the experiments, ALICE, ATLAS and LHCb at the Large Hadron Collider (LHC)
    Computer Research and Modeling, 2015, v. 7, no. 3, pp. 621-630

    The review of the distributed computing infrastructure of the Tier-1 sites for the Alice, ATLAS, LHCb experiments at the LHC is given. The special emphasis is placed on the main tasks and services of the Tier-1 site, which operates in the Kurchatov Institute in Moscow.

    Views (last year): 2.

Indexed in Scopus

Full-text version of the journal is also available on the web site of the scientific electronic library eLIBRARY.RU

The journal is included in the Russian Science Citation Index

The journal is included in the RSCI

International Interdisciplinary Conference "Mathematics. Computing. Education"