Результаты поиска по 'CERN':
Найдено статей: 6
  1. Gavrilov V.B., Golutvin I.A., Kodolova O.L., Korenkov V.V., Levchuk L.G., Shmatov S.V., Tikhonenko E.A., Zhiltsov V.E.
    RDMS CMS computing: current status and plans
    Computer Research and Modeling, 2015, v. 7, no. 3, pp. 395-398

    The Compact Muon Solenoid (CMS) is a high-performance general-purpose detector at the Large Hadron Collider (LHC) at CERN. More than twenty institutes from Russia and Joint Institute for Nuclear Research (JINR) are involved in Russia and Dubna Member States (RDMS) CMS Collaboration. A proper computing grid-infrastructure has been constructed at the RDMS institutes for the participation in the running phase of the CMS experiment. Current status of RDMS CMS computing and plans of its development to the next LHC start in 2015 are presented.

    Views (last year): 2.
  2. Zarochentsev A.K., Stiforov G.G.
    ALICE computing update before start of RUN2
    Computer Research and Modeling, 2015, v. 7, no. 3, pp. 415-419

    The report presents a number of news and updates of the ALICE computing for RUN2 and RUN3.

    This includes:

    – implementation in production of a new system EOS;

    – migration to the file system CVMFS to be used for storage of the software;

    – the plan for solving the problem of “Long-Term Data Preservation”;

    – overview of the concept of “O square”, combining offline and online data processing;

    – overview of the existing models to use the virtual clouds for ALICE data processing. Innovations are shown on the example of the Russian sites.

    Views (last year): 2.
  3. Astakhov N.S., Baginyan A.S., Belov S.D., Dolbilov A.G., Golunov A.O., Gorbunov I.N., Gromova N.I., Kashunin I.A., Korenkov V.V., Mitsyn V.V., Shmatov S.V., Strizh T.A., Tikhonenko E.A., Trofimov V.V., Voitishin N.N., Zhiltsov V.E.
    JINR TIER-1-level computing system for the CMS experiment at LHC: status and perspectives
    Computer Research and Modeling, 2015, v. 7, no. 3, pp. 455-462

    The Compact Muon Solenoid (CMS) is a high-performance general-purpose detector at the Large Hadron Collider (LHC) at CERN. A distributed data analysis system for processing and further analysis of CMS experimental data has been developed and this model foresees the obligatory usage of modern grid-technologies. The CMS Computing Model makes use of the hierarchy of computing centers (Tiers). The Joint Institute for Nuclear Research (JINR) takes an active part in the CMS experiment. In order to provide a proper computing infrastructure for the CMS experiment at JINR and for Russian institutes collaborating in CMS, Tier-1 center for the CMS experiment is constructing at JINR. The main tasks and services of the CMS Tier-1 at JINR are described. The status and perspectives of the Tier1 center for the CMS experiment at JINR are presented.

    Views (last year): 3. Citations: 2 (RSCI).
  4. Kiryanov A.K.
    GridFTP frontend with redirection for DMlite
    Computer Research and Modeling, 2015, v. 7, no. 3, pp. 543-547

    One of the most widely used storage solutions in WLCG is a Disk Pool Manager (DPM) developed and supported by SDC/ID group at CERN. Recently DPM went through a massive overhaul to address scalability and extensibility issues of the old code.

    New system was called DMLite. Unlike the old DPM that was based on daemons, DMLite is arranged as a library that can be loaded directly by an application. This approach greatly improves performance and transaction rate by avoiding unnecessary inter-process communication via network as well as threading bottlenecks.

    DMLite has a modular architecture with its core library providing only the very basic functionality. Backends (storage engines) and frontends (data access protocols) are implemented as plug-in modules. Doubtlessly DMLite wouldn't be able to completely replace DPM without GridFTP as it is used for most of the data transfers in WLCG.

    In DPM GridFTP support was implemented in a Data Storage Interface (DSI) module for Globus’ GridFTP server. In DMLite an effort was made to rewrite a GridFTP module from scratch in order to take advantage of new DMLite features and also implement new functionality. The most important improvement over the old version is a redirection capability.

    With old GridFTP frontend a client needed to contact SRM on the head node in order to obtain a transfer URL (TURL) before reading or writing a file. With new GridFTP frontend this is no longer necessary: a client may connect directly to the GridFTP server on the head node and perform file I/O using only logical file names (LFNs). Data channel is then automatically redirected to a proper disk node.

    This renders the most often used part of SRM unnecessary, simplifies file access and improves performance. It also makes DMLite a more appealing choice for non-LHC VOs that were never much interested in SRM.

    With new GridFTP frontend it's also possible to access data on various DMLite-supported backends like HDFS, S3 and legacy DPM.

    Views (last year): 1.
  5. Bondyakov A.S.
    Basic directions of information technology in National Academy of Sciences of Azerbaijan
    Computer Research and Modeling, 2015, v. 7, no. 3, pp. 657-660

    Grid is a new type of computing infrastructure, is intensively developed in today world of information technologies. Grid provides global integration of information and computing resources. The essence Conception of GRID in Azerbaijan is to create a set of standardized services to provide a reliable, compatible, inexpensive and secure access to geographically distributed high-tech information and computing resources a separate computer, cluster and supercomputing centers, information storage, networks, scientific tools etc.

    Views (last year): 6. Citations: 1 (RSCI).
  6. Kazymov A.I., Kotov V.M., Mineev M.A., Russakovich N.A., Yakovlev A.V.
    Using CERN cloud technologies for the further ATLAS TDAQ software development and for its application for the remote sensing data processing in the space monitoring tasks
    Computer Research and Modeling, 2015, v. 7, no. 3, pp. 683-689

    The CERN cloud technologies (the CernVM project) give a new possibility for the software developers. The participation of the JINR ATLAS TDAQ working group in the software development for distributed data acquisition and processing system (TDAQ) of the ATLAS experiment (CERN) involves the work in the condition of the dynamically developing system and its infrastructure. The CERN cloud technologies, especially CernVM, provide the most effective access as to the TDAQ software as to the third-part software used in ATLAS. The access to the Scientific Linux environment is provided by CernVM virtual machines and the access software repository — by CernVM-FS. The problem of the functioning of the TDAQ middleware in the CernVM environment was studied in this work. The CernVM usage is illustrated on three examples: the development of the packages Event Dump and Webemon, and the adaptation of the data quality auto checking system of the ATLAS TDAQ (Data Quality Monitoring Framework) for the radar data assessment.

    Views (last year): 2.

Indexed in Scopus

Full-text version of the journal is also available on the web site of the scientific electronic library eLIBRARY.RU

The journal is included in the Russian Science Citation Index

The journal is included in the RSCI

International Interdisciplinary Conference "Mathematics. Computing. Education"