Результаты поиска по 'principal':
Найдено статей: 16
  1. Malsagov M.X., Ougolnitsky G.A., Usov A.B.
    Struggle against economic corruption in resource allocation
    Computer Research and Modeling, 2019, v. 11, no. 1, pp. 173-185

    A dynamic game theoretic model of struggle against corruption in resource allocation is considered. It is supposed that the system of resource allocation includes one principal, one or several supervisors, and several agents. The relations between them are hierarchical: the principal influences to the supervisors, and they in turn exert influence on the agents. It is assumed that the supervisor can be corrupted. The agents propose bribes to the supervisor who in exchange allocates additional resources to them. It is also supposed that the principal is not corrupted and does not have her own purposes. The model is investigated from the point of view of the supervisor and the agents. From the point of view of agents a non-cooperative game arises with a set of Nash equilibria as a solution. The set is found analytically on the base of Pontryagin maximum principle for the specific class of model functions. From the point of view of the supervisor a hierarchical Germeyer game of the type Г2t is built, and the respective algorithm of its solution is proposed. The punishment strategy is found analytically, and the reward strategy is built numerically on the base of a discrete analogue of the initial continuous- time model. It is supposed that all agents can change their strategies in the same time instants only a finite number of times. Thus, the supervisor can maximize his objective function of many variables instead of maximization of the objective functional. A method of qualitatively representative scenarios is used for the solution. The idea of this method consists in that it is possible to choose a very small number of scenarios among all potential ones that represent all qualitatively different trajectories of the system dynamics. These scenarios differ in principle while all other scenarios yield no essentially new results. Then a complete enumeration of the qualitatively representative scenarios becomes possible. After that, the supervisor reports to the agents the rewardpunishment control mechanism.

    Views (last year): 33. Citations: 1 (RSCI).
  2. When modeling turbulent flows in practical applications, it is often necessary to carry out a series of calculations of bodies of similar topology. For example, bodies that differ in the shape of the fairing. The use of convolutional neural networks allows to reduce the number of calculations in a series, restoring some of them based on calculations already performed. The paper proposes a method that allows to apply a convolutional neural network regardless of the method of constructing a computational mesh. To do this, the flow field is reinterpolated to a uniform mesh along with the body itself. The geometry of the body is set using the signed distance function and masking. The restoration of the flow field based on part of the calculations for similar geometries is carried out using a neural network of the UNet type with a spatial attention mechanism. The resolution of the nearwall region, which is a critical condition for turbulent modeling, is based on the equations obtained in the nearwall domain decomposition method.

    A demonstration of the method is given for the case of a flow around a rounded plate by a turbulent air flow with different rounding at fixed parameters of the incoming flow with the Reynolds number $Re = 10^5$ and the Mach number $M = 0.15$. Since flows with such parameters of the incoming flow can be considered incompressible, only the velocity components are studied directly. The flow fields, velocity and friction profiles obtained by the surrogate model and numerically are compared. The analysis is carried out both on the plate and on the rounding. The simulation results confirm the prospects of the proposed approach. In particular, it was shown that even if the model is used at the maximum permissible limits of its applicability, friction can be obtained with an accuracy of up to 90%. The work also analyzes the constructed architecture of the neural network. The obtained surrogate model is compared with alternative models based on a variational autoencoder or the principal component analysis using radial basis functions. Based on this comparison, the advantages of the proposed method are demonstrated.

  3. Kalitin K.Y., Nevzorov A.A., Spasov A.A., Mukha O.Y.
    Deep learning analysis of intracranial EEG for recognizing drug effects and mechanisms of action
    Computer Research and Modeling, 2024, v. 16, no. 3, pp. 755-772

    Predicting novel drug properties is fundamental to polypharmacology, repositioning, and the study of biologically active substances during the preclinical phase. The use of machine learning, including deep learning methods, for the identification of drug – target interactions has gained increasing popularity in recent years.

    The objective of this study was to develop a method for recognizing psychotropic effects and drug mechanisms of action (drug – target interactions) based on an analysis of the bioelectrical activity of the brain using artificial intelligence technologies.

    Intracranial electroencephalographic (EEG) signals from rats were recorded (4 channels at a sampling frequency of 500 Hz) after the administration of psychotropic drugs (gabapentin, diazepam, carbamazepine, pregabalin, eslicarbazepine, phenazepam, arecoline, pentylenetetrazole, picrotoxin, pilocarpine, chloral hydrate). The signals were divided into 2-second epochs, then converted into $2000\times 4$ images and input into an autoencoder. The output of the bottleneck layer was subjected to classification and clustering using t-SNE, and then the distances between resulting clusters were calculated. As an alternative, an approach based on feature extraction with dimensionality reduction using principal component analysis and kernel support vector machine (kSVM) classification was used. Models were validated using 5-fold cross-validation.

    The classification accuracy obtained for 11 drugs during cross-validation was $0.580 \pm 0.021$, which is significantly higher than the accuracy of the random classifier $(0.091 \pm 0.045, p < 0.0001)$ and the kSVM $(0.441 \pm 0.035, p < 0.05)$. t-SNE maps were generated from the bottleneck parameters of intracranial EEG signals. The relative proximity of the signal clusters in the parametric space was assessed.

    The present study introduces an original method for biopotential-mediated prediction of effects and mechanism of action (drug – target interaction). This method employs convolutional neural networks in conjunction with a modified selective parameter reduction algorithm. Post-treatment EEGs were compressed into a unified parameter space. Using a neural network classifier and clustering, we were able to recognize the patterns of neuronal response to the administration of various psychotropic drugs.

  4. Panteleev M.A., Bershadsky E.S., Shibeko A.M., Nechipurenko D.Y.
    Current issues in computational modeling of thrombosis, fibrinolysis, and thrombolysis
    Computer Research and Modeling, 2024, v. 16, no. 4, pp. 975-995

    Hemostasis system is one of the key body’s defense systems, which is presented in all the liquid tissues and especially important in blood. Hemostatic response is triggered as a result of the vessel injury. The interaction between specialized cells and humoral systems leads to the formation of the initial hemostatic clot, which stops bleeding. After that the slow process of clot dissolution occurs. The formation of hemostatic plug is a unique physiological process, because during several minutes the hemostatic system generates complex structures on a scale ranging from microns for microvessel injury or damaged endothelial cell-cell contacts, to centimeters for damaged systemic arteries. Hemostatic response depends on the numerous coordinated processes, which include platelet adhesion and aggregation, granule secretion, platelet shape change, modification of the chemical composition of the lipid bilayer, clot contraction, and formation of the fibrin mesh due to activation of blood coagulation cascade. Computer modeling is a powerful tool, which is used to study this complex system at different levels of organization. This includes study of intracellular signaling in platelets, modelling humoral systems of blood coagulation and fibrinolysis, and development of the multiscale models of thrombus growth. There are two key issues of the computer modeling in biology: absence of the adequate physico-mathematical description of the existing experimental data due to the complexity of the biological processes, and high computational complexity of the models, which doesn’t allow to use them to test physiologically relevant scenarios. Here we discuss some key unresolved problems in the field, as well as the current progress in experimental research of hemostasis and thrombosis. New findings lead to reevaluation of the existing concepts and development of the novel computer models. We focus on the arterial thrombosis, venous thrombosis, thrombosis in microcirculation and the problems of fibrinolysis and thrombolysis. We also briefly discuss basic types of the existing mathematical models, their computational complexity, and principal issues in simulation of thrombus growth in arteries.

  5. Sukhov E.A., Chekina E.A.
    Software complex for numerical modeling of multibody system dynamics
    Computer Research and Modeling, 2024, v. 16, no. 1, pp. 161-174

    This work deals with numerical modeling of motion of the multibody systems consisting of rigid bodies with arbitrary masses and inertial properties. We consider both planar and spatial systems which may contain kinematic loops.

    The numerical modeling is fully automatic and its computational algorithm contains three principal steps. On step one a graph of the considered mechanical system is formed from the userinput data. This graph represents the hierarchical structure of the mechanical system. On step two the differential-algebraic equations of motion of the system are derived using the so-called Joint Coordinate Method. This method allows to minimize the redundancy and lower the number of the equations of motion and thus optimize the calculations. On step three the equations of motion are integrated numerically and the resulting laws of motion are presented via user interface or files.

    The aforementioned algorithm is implemented in the software complex that contains a computer algebra system, a graph library, a mechanical solver, a library of numerical methods and a user interface.

  6. The article discusses the problem of the influence of the research goals on the structure of the multivariate model of regression analysis (in particular, on the implementation of the procedure for reducing the dimension of the model). It is shown how bringing the specification of the multiple regression model in line with the research objectives affects the choice of modeling methods. Two schemes for constructing a model are compared: the first does not allow taking into account the typology of primary predictors and the nature of their influence on the performance characteristics, the second scheme implies a stage of preliminary division of the initial predictors into groups, in accordance with the objectives of the study. Using the example of solving the problem of analyzing the causes of burnout of creative workers, the importance of the stage of qualitative analysis and systematization of a priori selected factors is shown, which is implemented not by computing means, but by attracting the knowledge and experience of specialists in the studied subject area. The presented example of the implementation of the approach to determining the specification of the regression model combines formalized mathematical and statistical procedures and the preceding stage of the classification of primary factors. The presence of this stage makes it possible to explain the scheme of managing (corrective) actions (softening the leadership style and increasing approval lead to a decrease in the manifestations of anxiety and stress, which, in turn, reduces the severity of the emotional exhaustion of the team members). Preclassification also allows avoiding the combination in one main component of controlled and uncontrolled, regulatory and controlled feature factors, which could worsen the interpretability of the synthesized predictors. On the example of a specific problem, it is shown that the selection of factors-regressors is a process that requires an individual solution. In the case under consideration, the following were consistently used: systematization of features, correlation analysis, principal component analysis, regression analysis. The first three methods made it possible to significantly reduce the dimension of the problem, which did not affect the achievement of the goal for which this task was posed: significant measures of controlling influence on the team were shown. allowing to reduce the degree of emotional burnout of its participants.

Pages: previous

Indexed in Scopus

Full-text version of the journal is also available on the web site of the scientific electronic library eLIBRARY.RU

The journal is included in the Russian Science Citation Index

The journal is included in the RSCI

International Interdisciplinary Conference "Mathematics. Computing. Education"