Результаты поиска по 'measure':
Найдено статей: 111
  1. Matveev A.V.
    Modeling the kinetics of radiopharmaceuticals with iodine isotopes in nuclear medicine problems
    Computer Research and Modeling, 2020, v. 12, no. 4, pp. 883-905

    Radiopharmaceuticals with iodine radioisotopes are now widely used in imaging and non-imaging methods of nuclear medicine. When evaluating the results of radionuclide studies of the structural and functional state of organs and tissues, parallel modeling of the kinetics of radiopharmaceuticals in the body plays an important role. The complexity of such modeling lies in two opposite aspects. On the one hand, excessive simplification of the anatomical and physiological characteristics of the organism when splitting it to the compartments that may result in the loss or distortion of important clinical diagnosis information, on the other – excessive, taking into account all possible interdependencies of the functioning of the organs and systems that, on the contrary, will lead to excess amount of absolutely useless for clinical interpretation of the data or the mathematical model becomes even more intractable. Our work develops a unified approach to the construction of mathematical models of the kinetics of radiopharmaceuticals with iodine isotopes in the human body during diagnostic and therapeutic procedures of nuclear medicine. Based on this approach, three- and four-compartment pharmacokinetic models were developed and corresponding calculation programs were created in the C++ programming language for processing and evaluating the results of radionuclide diagnostics and therapy. Various methods for identifying model parameters based on quantitative data from radionuclide studies of the functional state of vital organs are proposed. The results of pharmacokinetic modeling for radionuclide diagnostics of the liver, kidney, and thyroid using iodine-containing radiopharmaceuticals are presented and analyzed. Using clinical and diagnostic data, individual pharmacokinetic parameters of transport of different radiopharmaceuticals in the body (transport constants, half-life periods, maximum activity in the organ and the time of its achievement) were determined. It is shown that the pharmacokinetic characteristics for each patient are strictly individual and cannot be described by averaged kinetic parameters. Within the framework of three pharmacokinetic models, “Activity–time” relationships were obtained and analyzed for different organs and tissues, including for tissues in which the activity of a radiopharmaceutical is impossible or difficult to measure by clinical methods. Also discussed are the features and the results of simulation and dosimetric planning of radioiodine therapy of the thyroid gland. It is shown that the values of absorbed radiation doses are very sensitive to the kinetic parameters of the compartment model. Therefore, special attention should be paid to obtaining accurate quantitative data from ultrasound and thyroid radiometry and identifying simulation parameters based on them. The work is based on the principles and methods of pharmacokinetics. For the numerical solution of systems of differential equations of the pharmacokinetic models we used Runge–Kutta methods and Rosenbrock method. The Hooke–Jeeves method was used to find the minimum of a function of several variables when identifying modeling parameters.

  2. Elaraby A.E., Nechaevskiy A.V.
    An effective segmentation approach for liver computed tomography scans using fuzzy exponential entropy
    Computer Research and Modeling, 2021, v. 13, no. 1, pp. 195-202

    Accurate segmentation of liver plays important in contouring during diagnosis and the planning of treatment. Imaging technology analysis and processing are wide usage in medical diagnostics, and therapeutic applications. Liver segmentation referring to the process of automatic or semi-automatic detection of liver image boundaries. A major difficulty in segmentation of liver image is the high variability as; the human anatomy itself shows major variation modes. In this paper, a proposed approach for computed tomography (CT) liver segmentation is presented by combining exponential entropy and fuzzy c-partition. Entropy concept has been utilized in various applications in imaging computing domain. Threshold techniques based on entropy have attracted a considerable attention over the last years in image analysis and processing literatures and it is among the most powerful techniques in image segmentation. In the proposed approach, the computed tomography (CT) of liver is transformed into fuzzy domain and fuzzy entropies are defined for liver image object and background. In threshold selection procedure, the proposed approach considers not only the information of liver image background and object, but also interactions between them as the selection of threshold is done by find a proper parameter combination of membership function such that the total fuzzy exponential entropy is maximized. Differential Evolution (DE) algorithm is utilizing to optimize the exponential entropy measure to obtain image thresholds. Experimental results in different CT livers scan are done and the results demonstrate the efficient of the proposed approach. Based on the visual clarity of segmented images with varied threshold values using the proposed approach, it was observed that liver segmented image visual quality is better with the results higher level of threshold.

  3. Smolyak S.A.
    Valuation of machines at the random process of their degradation and premature sales
    Computer Research and Modeling, 2024, v. 16, no. 3, pp. 797-815

    The model of the process of using machinery and equipment is considered, which takes into account the probabilistic nature of the process of their operation and sale. It takes into account the possibility of random hidden failures, after which the condition of the machine deteriorates abruptly, as well as the randomly arising need for premature (before the end of its service life) sale of the machine, which requires, generally speaking, random time. The model is focused on assessing the market value and service life of machines in accordance with International Valuation Standards. Strictly speaking, the market value of a used machine depends on its technical condition, but in practice, appraisers only take into account its age, since generally accepted measures of the technical condition of machines do not yet exist. As a result, the market value of a used machine is assumed to be equal to the average market value of similar machines of the corresponding age. For these purposes, appraisers use coefficients that reflect the influence of the age of machines on their market value. Such coefficients are not always justified and do not take into account either the degradation of the machine or the probabilistic nature of the process of its use. The proposed model is based on the anticipation of benefits principle. In it, we characterize the state of the machine by the intensity of the benefits it brings. The machine is subjected to a complex Poisson failure process, and after failure its condition abruptly worsens and may even reach its limit. Situations also arise that preclude further use of the machine by its owner. In such situations, the owner puts the machine up for sale before the end of its service life (prematurely), and the sale requires a random timing. The model allows us to take into account the influence of such situations and construct an analytical relationship linking the market value of a machine with its condition, and calculate the average coefficients of change in the market value of machines with age. At the same time, it is also possible to take into account the influence of inflation and the scrap cost of the machine. We have found that the rate of prematurely sales has a significant impact on the cost of new and used machines. The model also allows us to take into account the influence of inflation and the scrap value of the machine. We have found that the rate of premature sales has a significant impact on the service life and market value of new and used machines. At the same time, the dependence of the market value of machines on age is largely determined by the coefficient of variation of the service life of the machines. The results obtained allow us to obtain more reasonable estimates of the market value of machines, including for the purposes of the system of national accounts.

  4. Shestoperov A.I., Ivchenko A.V., Fomina E.V.
    Changepoint detection in biometric data: retrospective nonparametric segmentation methods based on dynamic programming and sliding windows
    Computer Research and Modeling, 2024, v. 16, no. 5, pp. 1295-1321

    This paper is dedicated to the analysis of medical and biological data obtained through locomotor training and testing of astronauts conducted both on Earth and during spaceflight. These experiments can be described as the astronaut’s movement on a treadmill according to a predefined regimen in various speed modes. During these modes, not only the speed is recorded but also a range of parameters, including heart rate, ground reaction force, and others, are collected. In order to analyze the dynamics of the astronaut’s condition over an extended period, it is necessary to perform a qualitative segmentation of their movement modes to independently assess the target metrics. This task becomes particularly relevant in the development of an autonomous life support system for astronauts that operates without direct supervision from Earth. The segmentation of target data is complicated by the presence of various anomalies, such as deviations from the predefined regimen, arbitrary and varying duration of mode transitions, hardware failures, and other factors. The paper includes a detailed review of several contemporary retrospective (offline) nonparametric methods for detecting multiple changepoints, which refer to sudden changes in the properties of the observed time series occurring at unknown moments. Special attention is given to algorithms and statistical measures that determine the homogeneity of the data and methods for detecting change points. The paper considers approaches based on dynamic programming and sliding window methods. The second part of the paper focuses on the numerical modeling of these methods using characteristic examples of experimental data, including both “simple” and “complex” speed profiles of movement. The analysis conducted allowed us to identify the preferred methods, which will be further evaluated on the complete dataset. Preference is given to methods that ensure the closeness of the markup to a reference one, potentially allow the detection of both boundaries of transient processes, as well as are robust relative to internal parameters.

  5. Kazaryan A.M., Shapoval A.B.
    Timeclusterring of stock indicies’ big fall
    Computer Research and Modeling, 2012, v. 4, no. 3, pp. 631-638

    The paper estimates the recurrence rate of stock indicies S&P100, CAC40, DAX, FTSE, AMEX, ATX, NASDAQ, BEL20. The introduced qunatitative measure of the recurrence rate underlies type I and type II errors. We show that more than three quarters of the indicies’ falls occur on average during the first quarter of the time between them. This result expands from sufficiently large falls, which are observed on average two times a year, over smaller falls, which occur approximately once 1.5–2 months.

    Views (last year): 2.
  6. Shpitonkov M.I.
    Application of correlation adaptometry technique to sports and biomedical research
    Computer Research and Modeling, 2017, v. 9, no. 2, pp. 345-354

    The paper outlines the approaches to mathematical modeling correlation adaptometry techniques widely used in biology and medicine. The analysis is based on models employed in descriptions of structured biological systems. It is assumed that the distribution density of the biological population numbers satisfies the equation of Kolmogorov-Fokker-Planck. Using this technique evaluated the effectiveness of treatment of patients with obesity. All patients depending on the obesity degree and the comorbidity nature were divided into three groups. Shows a decrease in weight of the correlation graph computed from the measured in the patients of the indicators that characterizes the effectiveness of the treatment for all studied groups. This technique was also used to assess the intensity of the training loads in academic rowing three age groups. It was shown that with the highest voltage worked with athletes for youth group. Also, using the technique of correlation adaptometry evaluated the effectiveness of the treatment of hormone replacement therapy in women. All the patients depending on the assigned drug were divided into four groups. In the standard analysis of the dynamics of mean values of indicators, it was shown that in the course of the treatment were observed normalization of the averages for all groups of patients. However, using the technique of correlation adaptometry it was found that during the first six months the weight of the correlation graph was decreasing and during the second six months the weight increased for all study groups. This indicates the excessive length of the annual course of hormone replacement therapy and the practicality of transition to a semiannual rate.

    Views (last year): 10.
  7. Shumov V.V.
    Mathematical models of combat and military operations
    Computer Research and Modeling, 2020, v. 12, no. 4, pp. 907-920

    Modeling the fight against terrorist, pirate and robbery acts at sea is an urgent scientific task due to the prevalence of force acts and the insufficient number of works on this issue. The actions of pirates and terrorists are diverse. Using a base ship, they can attack ships up to 450–500 miles from the coast. Having chosen the target, they pursue it and use the weapons to board the ship. Actions to free a ship captured by pirates or terrorists include: blocking the ship, predicting where pirates might be on the ship, penetrating (from board to board, by air or from under water) and cleaning up the ship’s premises. An analysis of the special literature on the actions of pirates and terrorists showed that the act of force (and actions to neutralize it) consists of two stages: firstly, blocking the vessel, which consists in forcing it to stop, and secondly, neutralizing the team (terrorist groups, pirates), including penetration of a ship (ship) and its cleaning. The stages of the cycle are matched by indicators — the probability of blocking and the probability of neutralization. The variables of the act of force model are the number of ships (ships, boats) of the attackers and defenders, as well as the strength of the capture group of the attackers and the crew of the ship - the victim of the attack. Model parameters (indicators of naval and combat superiority) were estimated using the maximum likelihood method using an international database of incidents at sea. The values of these parameters are 7.6–8.5. Such high values of superiority parameters reflect the parties' ability to act in force acts. An analytical method for calculating excellence parameters is proposed and statistically substantiated. The following indicators are taken into account in the model: the ability of the parties to detect the enemy, the speed and maneuverability characteristics of the vessels, the height of the vessel and the characteristics of the boarding equipment, the characteristics of weapons and protective equipment, etc. Using the Becker model and the theory of discrete choice, the probability of failure of the force act is estimated. The significance of the obtained models for combating acts of force in the sea space lies in the possibility of quantitative substantiation of measures to protect the ship from pirate and terrorist attacks and deterrence measures aimed at preventing attacks (the presence of armed guards on board the ship, assistance from warships and helicopters).

  8. Kalachin S.V.
    Fuzzy modeling of human susceptibility to panic situations
    Computer Research and Modeling, 2021, v. 13, no. 1, pp. 203-218

    The study of the mechanism for the development of mass panic in view of its extreme importance and social danger is an important scientific task. Available information about the mechanism of her development is based mainly on the work of psychologists and belongs to the category of inaccurate. Therefore, the theory of fuzzy sets has been chosen as a tool for developing a mathematical model of a person's susceptibility to panic situations. As a result of the study, an fuzzy model was developed, consisting of blocks: “Fuzzyfication”, where the degree of belonging of the values of the input parameters to fuzzy sets is calculated; “Inference” where, based on the degree of belonging of the input parameters, the resulting function of belonging of the output value to an odd model is calculated; “Defuzzyfication”, where using the center of gravity method, the only quantitative value of the output variable characterizing a person's susceptibility to panic situations is determined Since the real quantitative values for linguistic variables mental properties of a person are unknown, then to assess the quality of the developed model, without endangering people, it is not possible. Therefore, the quality of the results of fuzzy modeling was estimated by the calculated value of the determination coefficient R2, which showed that the developed fuzzy model belongs to the category of good quality models $(R^2 = 0.93)$, which confirms the legitimacy of the assumptions made during her development. In accordance with to the results of the simulation, human susceptibility to panic situations for sanguinics and cholerics can be attributed to “increased” (0.88), and for phlegmatics and melancholics — to “moderate” (0.38). This means that cholerics and sanguinics can become epicenters of panic and the initiators of stampede, and phlegmatics and melancholics — obstacles to evacuation routes. What should be taken into account when developing effective evacuation measures, the main task of which is to quickly and safely evacuate people from adverse conditions. In the approved methods, the calculation of normative values of safety parameters is based on simplified analytical models of human flow movement, because a large number of factors have to be taken into account, some of which are quantitatively uncertain. The obtained result in the form of quantitative estimates of a person's susceptibility to panic situations will increase the accuracy of calculations.

  9. Vasyukov A.V., Beklemysheva K.A., Onuchin E.S., Tovarnova N.A., Petrov I.B.
    Calculation of transverse wave speed in preloaded fibres under an impact
    Computer Research and Modeling, 2022, v. 14, no. 4, pp. 887-897

    The paper considers the problem of transverse impact on a thin preloaded fiber. The commonly accepted theory of transverse impact on a thin fiber is based on the classical works of Rakhmatulin and Smith. The simple relations obtained from the Rakhmatulin – Smith theory are widely used in engineering practice. However, there are numerous evidences that experimental results may differ significantly from estimations based on these relations. A brief overview of the factors that cause the differences is given in this article.

    This paper focuses on the shear wave velocity, as it is the only feature that can be directly observed and measured using high-speed cameras or similar methods. The influence of the fiber preload on the wave speed is considered. This factor is important, since it inevitably arises in the experimental results. The reliable fastening and precise positioning of the fiber during the experiments requires its preload. This work shows that the preload significantly affects the shear wave velocity in the impacted fiber.

    Numerical calculations were performed for Kevlar 29 and Spectra 1000 yarns. Shear wave velocities are obtained for different levels of initial tension. A direct comparison of numerical results and analytical estimations with experimental data is presented. The speed of the transverse wave in free and preloaded fibers differed by a factor of two for the setup parameters considered. This fact demonstrates that measurements based on high-speed imaging and analysis of the observed shear waves should take into account the preload of the fibers.

    This paper proposes a formula for a quick estimation of the shear wave velocity in preloaded fibers. The formula is obtained from the basic relations of the Rakhmatulin – Smith theory under the assumption of a large initial deformation of the fiber. The formula can give significantly better results than the classical approximation, this fact is demonstrated using the data for preloaded Kevlar 29 and Spectra 1000. The paper also shows that direct numerical calculation has better corresponding with the experimental data than any of the considered analytical estimations.

  10. Zavodskikh R.K., Efanov N.N.
    Performance prediction for chosen types of loops over one-dimensional arrays with embedding-driven intermediate representations analysis
    Computer Research and Modeling, 2023, v. 15, no. 1, pp. 211-224

    The method for mapping of intermediate representations (IR) set of C, C++ programs to vector embedding space is considered to create an empirical estimation framework for static performance prediction using LLVM compiler infrastructure. The usage of embeddings makes programs easier to compare due to avoiding Control Flow Graphs (CFG) and Data Flow Graphs (DFG) direct comparison. This method is based on transformation series of the initial IR such as: instrumentation — injection of artificial instructions in an instrumentation compiler’s pass depending on load offset delta in the current instruction compared to the previous one, mapping of instrumented IR into multidimensional vector with IR2Vec and dimension reduction with t-SNE (t-distributed stochastic neighbor embedding) method. The D1 cache miss ratio measured with perf stat tool is considered as performance metric. A heuristic criterion of programs having more or less cache miss ratio is given. This criterion is based on embeddings of programs in 2D-space. The instrumentation compiler’s pass developed in this work is described: how it generates and injects artificial instructions into IR within the used memory model. The software pipeline that implements the performance estimation based on LLVM compiler infrastructure is given. Computational experiments are performed on synthetic tests which are the sets of programs with the same CFGs but with different sequences of offsets used when accessing the one-dimensional array of a given size. The correlation coefficient between performance metric and distance to the worst program’s embedding is measured and proved to be negative regardless of t-SNE initialization. This fact proves the heuristic criterion to be true. The process of such synthetic tests generation is also considered. Moreover, the variety of performance metric in programs set in such a test is proposed as a metric to be improved with exploration of more tests generators.

Pages: « first previous next last »

Indexed in Scopus

Full-text version of the journal is also available on the web site of the scientific electronic library eLIBRARY.RU

The journal is included in the Russian Science Citation Index

The journal is included in the RSCI

International Interdisciplinary Conference "Mathematics. Computing. Education"