Результаты поиска по 'optimization':
Найдено статей: 222
  1. Varshavsky L.E.
    Studying indicators of development of oligopolistic markets on the basis of operational calculus
    Computer Research and Modeling, 2019, v. 11, no. 5, pp. 949-963

    The traditional approach to computing optimal game strategies of firms on oligopolistic markets and of indicators of such markets consists in studying linear dynamical games with quadratic criteria and solving generalized matrix Riccati equations.

    The other approach proposed by the author is based on methods of operational calculus (in particular, Z-transform). This approach makes it possible to achieve economic meaningful decisions under wider field of parameter values. It characterizes by simplicity of computations and by necessary for economic analysis visibility. One of its advantages is that in many cases important for economic practice, it, in contrast to the traditional approach, provides the ability to make calculations using widespread spreadsheets, which allows to study the prospects for the development of oligopolistic markets to a wide range of professionals and consumers.

    The article deals with the practical aspects of determining the optimal Nash–Cournot strategies of participants in oligopolistic markets on the basis of operational calculus, in particular the technique of computing the optimal Nash–Cournot strategies in Excel. As an illustration of the opportinities of the proposed methods of calculation, examples close to the practical problems of forecasting indicators of the markets of high-tech products are studied.

    The results of calculations obtained by the author for numerous examples and real economic systems, both using the obtained relations on the basis of spreadsheets and using extended Riccati equations, are very close. In most of the considered practical problems, the deviation of the indicators calculated in accordance with the two approaches, as a rule, does not exceed 1.5–2%. The highest value of relative deviations (up to 3–5%) is observed at the beginning of the forecasting period. In typical cases, the period of relatively noticeable deviations is 3–5 moments of time. After the transition period, there is almost complete agreement of the values of the required indicators using both approaches.

  2. Khorkov A.V., Khorkov A.V.
    Linear and nonlinear optimization models of multiple covering of a bounded plane domain with circles
    Computer Research and Modeling, 2019, v. 11, no. 6, pp. 1101-1110

    Problems of multiple covering ($k$-covering) of a bounded set $G$ with equal circles of a given radius are well known. They are thoroughly studied under the assumption that $G$ is a finite set. There are several papers concerned with studying this problem in the case where $G$ is a connected set. In this paper, we study the problem of minimizing the number of circles that form a $k$-covering, $k \geqslant 1$, provided that $G$ is a bounded convex plane domain.

    For the above-mentioned problem, we state a 0-1 linear model, a general integer linear model, and a nonlinear model, imposing a constraint on the minimum distance between the centers of covering circles. The latter constraint is due to the fact that in practice one can place at most one device at each point. We establish necessary and sufficient solvability conditions for the linear models and describe one (easily realizable) variant of these conditions in the case where the covered set $G$ is a rectangle.

    We propose some methods for finding an approximate number of circles of a given radius that provide the desired $k$-covering of the set $G$, both with and without constraints on distances between the circles’ centers. We treat the calculated values as approximate upper bounds for the number of circles. We also propose a technique that allows one to get approximate lower bounds for the number of circles that is necessary for providing a $k$-covering of the set $G$. In the general linear model, as distinct from the 0-1 linear model, we require no additional constraint. The difference between the upper and lower bounds for the number of circles characterizes the quality (acceptability) of the constructed $k$-covering.

    We state a nonlinear mathematical model for the $k$-covering problem with the above-mentioned constraints imposed on distances between the centers of covering circles. For this model, we propose an algorithm which (in certain cases) allows one to find more exact solutions to covering problems than those calculated from linear models.

    For implementing the proposed approach, we have developed computer programs and performed numerical experiments. Results of numerical experiments demonstrate the effectiveness of the method.

  3. Kiselev M.V.
    Exploration of 2-neuron memory units in spiking neural networks
    Computer Research and Modeling, 2020, v. 12, no. 2, pp. 401-416

    Working memory mechanisms in spiking neural networks consisting of leaky integrate-and-fire neurons with adaptive threshold and synaptic plasticity are studied in this work. Moderate size networks including thousands of neurons were explored. Working memory is a network ability to keep in its state the information about recent stimuli presented to the network such that this information is sufficient to determine which stimulus has been presented. In this study, network state is defined as the current characteristics of network activity only — without internal state of its neurons. In order to discover the neuronal structures serving as a possible substrate of the memory mechanism, optimization of the network parameters and structure using genetic algorithm was carried out. Two kinds of neuronal structures with the desired properties were found. These are neuron pairs mutually connected by strong synaptic links and long tree-like neuronal ensembles. It was shown that only the neuron pairs are suitable for efficient and reliable implementation of working memory. Properties of such memory units and structures formed by them are explored in the present study. It is shown that characteristics of the studied two-neuron memory units can be set easily by the respective choice of the parameters of its neurons and synaptic connections. Besides that, this work demonstrates that ensembles of these structures can provide the network with capability of unsupervised learning to recognize patterns in the input signal.

  4. Ostroukhov P.A., Kamalov R.A., Dvurechensky P.E., Gasnikov A.V.
    Tensor methods for strongly convex strongly concave saddle point problems and strongly monotone variational inequalities
    Computer Research and Modeling, 2022, v. 14, no. 2, pp. 357-376

    In this paper we propose high-order (tensor) methods for two types of saddle point problems. Firstly, we consider the classic min-max saddle point problem. Secondly, we consider the search for a stationary point of the saddle point problem objective by its gradient norm minimization. Obviously, the stationary point does not always coincide with the optimal point. However, if we have a linear optimization problem with linear constraints, the algorithm for gradient norm minimization becomes useful. In this case we can reconstruct the solution of the optimization problem of a primal function from the solution of gradient norm minimization of dual function. In this paper we consider both types of problems with no constraints. Additionally, we assume that the objective function is $\mu$-strongly convex by the first argument, $\mu$-strongly concave by the second argument, and that the $p$-th derivative of the objective is Lipschitz-continous.

    For min-max problems we propose two algorithms. Since we consider strongly convex a strongly concave problem, the first algorithm uses the existing tensor method for regular convex concave saddle point problems and accelerates it with the restarts technique. The complexity of such an algorithm is linear. If we additionally assume that our objective is first and second order Lipschitz, we can improve its performance even more. To do this, we can switch to another existing algorithm in its area of quadratic convergence. Thus, we get the second algorithm, which has a global linear convergence rate and a local quadratic convergence rate.

    Finally, in convex optimization there exists a special methodology to solve gradient norm minimization problems by tensor methods. Its main idea is to use existing (near-)optimal algorithms inside a special framework. I want to emphasize that inside this framework we do not necessarily need the assumptions of strong convexity, because we can regularize the convex objective in a special way to make it strongly convex. In our article we transfer this framework on convex-concave objective functions and use it with our aforementioned algorithm with a global linear convergence and a local quadratic convergence rate.

    Since the saddle point problem is a particular case of the monotone variation inequality problem, the proposed methods will also work in solving strongly monotone variational inequality problems.

  5. Skorik S.N., Pirau V.V., Sedov S.A., Dvinskikh D.M.
    Comparsion of stochastic approximation and sample average approximation for saddle point problem with bilinear coupling term
    Computer Research and Modeling, 2023, v. 15, no. 2, pp. 381-391

    Stochastic optimization is a current area of research due to significant advances in machine learning and their applications to everyday problems. In this paper, we consider two fundamentally different methods for solving the problem of stochastic optimization — online and offline algorithms. The corresponding algorithms have their qualitative advantages over each other. So, for offline algorithms, it is required to solve an auxiliary problem with high accuracy. However, this can be done in a distributed manner, and this opens up fundamental possibilities such as, for example, the construction of a dual problem. Despite this, both online and offline algorithms pursue a common goal — solving the stochastic optimization problem with a given accuracy. This is reflected in the comparison of the computational complexity of the described algorithms, which is demonstrated in this paper.

    The comparison of the described methods is carried out for two types of stochastic problems — convex optimization and saddles. For problems of stochastic convex optimization, the existing solutions make it possible to compare online and offline algorithms in some detail. In particular, for strongly convex problems, the computational complexity of the algorithms is the same, and the condition of strong convexity can be weakened to the condition of $\gamma$-growth of the objective function. From this point of view, saddle point problems are much less studied. Nevertheless, existing solutions allow us to outline the main directions of research. Thus, significant progress has been made for bilinear saddle point problems using online algorithms. Offline algorithms are represented by just one study. In this paper, this example demonstrates the similarity of both algorithms with convex optimization. The issue of the accuracy of solving the auxiliary problem for saddles was also worked out. On the other hand, the saddle point problem of stochastic optimization generalizes the convex one, that is, it is its logical continuation. This is manifested in the fact that existing results from convex optimization can be transferred to saddles. In this paper, such a transfer is carried out for the results of the online algorithm in the convex case, when the objective function satisfies the $\gamma$-growth condition.

  6. The creation of a virtual laboratory stand that allows one to obtain reliable characteristics that can be proven as actual, taking into account errors and noises (which is the main distinguishing feature of a computational experiment from model studies) is one of the main problems of this work. It considers the following task: there is a rectangular waveguide in the single operating mode, on the wide wall of which a technological hole is cut, through which a sample for research is placed into the cavity of the transmission line. The recovery algorithm is as follows: the laboratory measures the network parameters (S11 and/or S21) in the transmission line with the sample. In the computer model of the laboratory stand, the sample geometry is reconstructed and an iterative process of optimization (or sweeping) of the electrophysical parameters is started, the mask of this process is the experimental data, and the stop criterion is the interpretive estimate of proximity (or residual). It is important to note that the developed computer model, along with its apparent simplicity, is initially ill-conditioned. To set up a computational experiment, the Comsol modeling environment is used. The results of the computational experiment with a good degree of accuracy coincided with the results of laboratory studies. Thus, experimental verification was carried out for several significant components, both the computer model in particular and the algorithm for restoring the target parameters in general. It is important to note that the computer model developed and described in this work may be effectively used for a computational experiment to restore the full dielectric parameters of a complex geometry target. Weak bianisotropy effects can also be detected, including chirality, gyrotropy, and material nonreciprocity. The resulting model is, by definition, incomplete, but its completeness is the highest of the considered options, while at the same time, the resulting model is well conditioned. Particular attention in this work is paid to the modeling of a coaxial-waveguide transition, it is shown that the use of a discrete-element approach is preferable to the direct modeling of the geometry of a microwave device.

  7. Saade M.G.
    Modeling the impact of epidemic spread and lockdown on economy
    Computer Research and Modeling, 2025, v. 17, no. 2, pp. 339-363

    Epidemics severely destabilize economies by reducing productivity, weakening consumer spending, and overwhelming public infrastructure, often culminating in economic recessions. The COVID-19 pandemic underscored the critical role of nonpharmaceutical interventions, such as lockdowns, in containing infectious disease transmission. This study investigates how the progression of epidemics and the implementation of lockdown policies shape the economic well-being of populations. By integrating compartmental ordinary differential equation (ODE) models, the research analyzes the interplay between epidemic dynamics and economic outcomes, particularly focusing on how varying lockdown intensities influence both disease spread and population wealth. Findings reveal that epidemics inflict significant economic damage, but timely and stringent lockdowns can mitigate healthcare system overload by sharply reducing infection peaks and delaying the epidemic’s trajectory. However, carefully timed lockdown relaxation is equally vital to prevent resurgent outbreaks. The study identifies key epidemiological thresholds—such as transmission rates, recovery rates, and the basic reproduction number $(\mathfrak{R}0)$ — that determine the effectiveness of lockdowns. Analytically, it pinpoints the optimal proportion of isolated individuals required to minimize total infections in scenarios where permanent immunity is assumed. Economically, the analysis quantifies lockdown impacts by tracking population wealth, demonstrating that economic outcomes depend heavily on the fraction of isolated individuals who remain economically productive. Higher proportions of productive individuals during lockdowns correlate with better wealth retention, even under fixed epidemic conditions. These insights equip policymakers with actionable frameworks to design balanced lockdown strategies that curb disease spread while safeguarding economic stability during future health crises.

  8. Kuznetsov M.B., Kolobov A.V.
    Optimization of proton therapy with radiosensitizing nanoparticles and antiangiogenic therapy via mathematical modeling
    Computer Research and Modeling, 2025, v. 17, no. 4, pp. 697-715

    Optimization of antitumor radiotherapy represents an urgent issue, as approximately half of the patients diagnosed with cancer undergo radiotherapy during their treatment. Proton therapy is potentially more efficient than traditional X-ray radiotherapy due to fundamental differences in physics of dose deposition, leading to better targeting of tumors and less collateral damage to healthy tissue. There is increasing interest in the use of non-radioactive radiosensitizing tumor-specific nanoparticles the use of which can boost the performance of proton therapy. Such nanoparticles are small volumes of a sensitizer, such as boron-10 or various metal oxides, enclosed in a polymer layer containing tumor-specific antibodies, which allows for their targeted delivery to malignant cells. Furthermore, a combination of proton therapy with antiangiogenic therapy that normalizes tumor-associated microvasculature may yield further synergistic increase in overall treatment efficacy.

    We have developed a spatially distributed mathematical model simulating the growth of a non-invasive tumor undergoing treatment by fractionated proton therapy with nanosensitizers and antiangiogenic therapy. The modeling results suggest that the most effective way to combine these treatment modalities should strongly depend on the tumor cells’ proliferation rate and their intrinsic radiosensitivity. Namely, a combination of antiangiogenic therapy with proton therapy, regardless of whether radiosensitizing nanoparticles are used, benefits treatment efficacy of rapidly growing tumors as well as radioresistant tumors with moderate growth rate. In these cases, administration of proton therapy simultaneously with antiangiogenic drugs after the initial single injection of nanosensitizers is the most effective option among those analyzed. Conversely, for slowly growing tumors, maximization of the number of nanosensitizer injections without antiangiogenic therapy proves to be a more efficient option, with enhancement in treatment efficacy growing with the increase of tumor radiosensitivity. However, the results also show that the overall efficacy of proton therapy is likely to increase only modestly with the addition of nanosensitizers and antiangiogenic drugs.

  9. Fedorova E.A.
    The mathematical optimization model based on several quality criteria
    Computer Research and Modeling, 2011, v. 3, no. 4, pp. 489-502

    An effective regional policy in order to stabilize production is impossible without an analysis of the dynamics of economic processes taking place. This article focuses on developing a mathematical model reflecting the interaction of several economic agents with regard to their interests. Developing such a model and its study can be considered as an important step in solving theoretical and practical problems of managing growth.

    Views (last year): 7.
  10. Shumov V.V.
    Protection of biological resources in the coastal area: the mathematical model
    Computer Research and Modeling, 2015, v. 7, no. 5, pp. 1109-1125

    Protection of aquatic biological resources in the coastal area has significant features (a large number of small fishing vessels, the dynamism of the situation, the use of coastal protection), by virtue of which stands in a class of applications. A mathematical model of protection designed for the determination of detection equipment and means of violators of the situation in order to ensure the function of deterrence of illegal activities. Resolves a tactical game-theoretic problem - find the optimal line patrol (parking) means of implementation (guard boats) and optimal removal of seats from the shore fishing violators. Using the methods of the theory of experimental design, linear regression models to assess the contribution of the main factors affecting the results of the simulation.

    In order to enhance the sustainability and adequacy of the model is proposed to use the mechanism of rankings means of protection, based on the borders and the rank and Pareto allows to take into account the principles of protection and further means of protection. To account for the variability of the situation offered several scenarios in which it is advisable to perform calculations.

    Views (last year): 1. Citations: 1 (RSCI).
Pages: « first previous next last »

Indexed in Scopus

Full-text version of the journal is also available on the web site of the scientific electronic library eLIBRARY.RU

The journal is included in the Russian Science Citation Index

The journal is included in the RSCI

International Interdisciplinary Conference "Mathematics. Computing. Education"