Результаты поиска по 'algorithm':
Найдено статей: 320
  1. Gesture recognition is an urgent challenge in developing systems of human-machine interfaces. We analyzed machine learning methods for gesture classification based on electromyographic muscle signals to identify the most effective one. Methods such as the naive Bayesian classifier (NBC), logistic regression, decision tree, random forest, gradient boosting, support vector machine (SVM), $k$-nearest neighbor algorithm, and ensembles (NBC and decision tree, NBC and gradient boosting, gradient boosting and decision tree) were considered. Electromyography (EMG) was chosen as a method of obtaining information about gestures. This solution does not require the location of the hand in the field of view of the camera and can be used to recognize finger movements. To test the effectiveness of the selected methods of gesture recognition, a device was developed for recording the EMG signal, which includes three electrodes and an EMG sensor connected to the microcontroller and the power supply. The following gestures were chosen: clenched fist, “thumb up”, “Victory”, squeezing an index finger and waving a hand from right to left. Accuracy, precision, recall and execution time were used to evaluate the effectiveness of classifiers. These parameters were calculated for three options for the location of EMG electrodes on the forearm. According to the test results, the most effective methods are $k$-nearest neighbors’ algorithm, random forest and the ensemble of NBC and gradient boosting, the average accuracy of ensemble for three electrode positions was 81.55%. The position of the electrodes was also determined at which machine learning methods achieve the maximum accuracy. In this position, one of the differential electrodes is located at the intersection of the flexor digitorum profundus and flexor pollicis longus, the second — above the flexor digitorum superficialis.

  2. Makarov I.S., Bagantsova E.R., Iashin P.A., Kovaleva M.D., Gorbachev R.A.
    Development of and research on an algorithm for distinguishing features in Twitter publications for a classification problem with known markup
    Computer Research and Modeling, 2023, v. 15, no. 1, pp. 171-183

    Social media posts play an important role in demonstration of financial market state, and their analysis is a powerful tool for trading. The article describes the result of a study of the impact of social media activities on the movement of the financial market. The top authoritative influencers are selected. Twitter posts are used as data. Such texts usually include slang and abbreviations, so methods for preparing primary text data, including Stanza, regular expressions are presented. Two approaches to the representation of a point in time in the format of text data are considered. The difference of the influence of a single tweet or a whole package consisting of tweets collected over a certain period of time is investigated. A statistical approach in the form of frequency analysis is also considered, metrics defined by the significance of a particular word when identifying the relationship between price changes and Twitter posts are introduced. Frequency analysis involves the study of the occurrence distributions of various words and bigrams in the text for positive, negative or general trends. To build the markup, changes in the market are processed into a binary vector using various parameters, thus setting the task of binary classification. The parameters for Binance candlesticks are sorted out for better description of the movement of the cryptocurrency market, their variability is also explored in this article. Sentiment is studied using Stanford Core NLP. The result of statistical analysis is relevant to feature selection for further binary or multiclass classification tasks. The presented methods of text analysis contribute to the increase of the accuracy of models designed to solve natural language processing problems by selecting words, improving the quality of vectorization. Such algorithms are often used in automated trading strategies to predict the price of an asset, the trend of its movement.

  3. Stonyakin F.S., Savchuk O.S., Baran I.V., Alkousa M.S., Titov A.A.
    Analogues of the relative strong convexity condition for relatively smooth problems and adaptive gradient-type methods
    Computer Research and Modeling, 2023, v. 15, no. 2, pp. 413-432

    This paper is devoted to some variants of improving the convergence rate guarantees of the gradient-type algorithms for relatively smooth and relatively Lipschitz-continuous problems in the case of additional information about some analogues of the strong convexity of the objective function. We consider two classes of problems, namely, convex problems with a relative functional growth condition, and problems (generally, non-convex) with an analogue of the Polyak – Lojasiewicz gradient dominance condition with respect to Bregman divergence. For the first type of problems, we propose two restart schemes for the gradient type methods and justify theoretical estimates of the convergence of two algorithms with adaptively chosen parameters corresponding to the relative smoothness or Lipschitz property of the objective function. The first of these algorithms is simpler in terms of the stopping criterion from the iteration, but for this algorithm, the near-optimal computational guarantees are justified only on the class of relatively Lipschitz-continuous problems. The restart procedure of another algorithm, in its turn, allowed us to obtain more universal theoretical results. We proved a near-optimal estimate of the complexity on the class of convex relatively Lipschitz continuous problems with a functional growth condition. We also obtained linear convergence rate guarantees on the class of relatively smooth problems with a functional growth condition. For a class of problems with an analogue of the gradient dominance condition with respect to the Bregman divergence, estimates of the quality of the output solution were obtained using adaptively selected parameters. We also present the results of some computational experiments illustrating the performance of the methods for the second approach at the conclusion of the paper. As examples, we considered a linear inverse Poisson problem (minimizing the Kullback – Leibler divergence), its regularized version which allows guaranteeing a relative strong convexity of the objective function, as well as an example of a relatively smooth and relatively strongly convex problem. In particular, calculations show that a relatively strongly convex function may not satisfy the relative variant of the gradient dominance condition.

  4. Aksenov A.A., Pokhilko V.I., Moryak A.P.
    Usage of boundary layer grids in numerical simulations of viscous phenomena in of ship hydrodynamics problems
    Computer Research and Modeling, 2023, v. 15, no. 4, pp. 995-1008

    Numerical simulation of hull flow, marine propellers and other basic problems of ship hydrodynamics using Cartesian adaptive locally-refined grids is advantageous with respect to numerical setup and makes an express analysis very convenient. However, when more accurate viscous phenomena are needed, they condition some problems including a sharp increase of cell number due to high levels of main grid adaptation needed to resolve boundary layers and time step decrease in simulations with a free surface due to decrease of transit time in adapted cells. To avoid those disadvantages, additional boundary layer grids are suggested for resolution of boundary layers. The boundary layer grids are one-dimensional adaptations of main grid layers nearest to a wall, which are built along a normal direction. The boundary layer grids are additional (or chimerical), their volumes are not subtracted from main grid volumes. Governing equations of flow are integrated in both grids simultaneously, and the solutions are merged according to a special algorithm. In simulations of ship hull flow boundary layer grids are able to provide sufficient conditions for low-Reynolds turbulence models and significantly improve flow structure in continues boundary layers along smooth surfaces. When there are flow separations or other complex phenomena on a hull surface, it can be subdivided into regions, and the boundary layer grids should be applied to the regions with simple flow only. This still provides a drastic decrease of computational efforts. In simulations of marine propellers, the boundary layer grids are able to provide refuse of wall functions on blade surfaces, what leads to significantly more accurate hydrodynamic forces. Altering number and configuration of boundary grid layers, it is possible to vary a boundary layer resolution without change of a main grid. This makes the boundary layer grids a suitable tool to investigate scale effects in both problems considered.

  5. Nechaevskiy A.V., Streltsova O.I., Kulikov K.V., Bashashin M.V., Butenko Y.A., Zuev M.I.
    Development of a computational environment for mathematical modeling of superconducting nanostructures with a magnet
    Computer Research and Modeling, 2023, v. 15, no. 5, pp. 1349-1358

    Now days the main research activity in the field of nanotechnology is aimed at the creation, study and application of new materials and new structures. Recently, much attention has been attracted by the possibility of controlling magnetic properties using a superconducting current, as well as the influence of magnetic dynamics on the current–voltage characteristics of hybrid superconductor/ferromagnet (S/F) nanostructures. In particular, such structures include the S/F/S Josephson junction or molecular nanomagnets coupled to the Josephson junctions. Theoretical studies of the dynamics of such structures need processes of a large number of coupled nonlinear equations. Numerical modeling of hybrid superconductor/magnet nanostructures implies the calculation of both magnetic dynamics and the dynamics of the superconducting phase, which strongly increases their complexity and scale, so it is advisable to use heterogeneous computing systems.

    In the course of studying the physical properties of these objects, it becomes necessary to numerically solve complex systems of nonlinear differential equations, which requires significant time and computational resources.

    The currently existing micromagnetic algorithms and frameworks are based on the finite difference or finite element method and are extremely useful for modeling the dynamics of magnetization on a wide time scale. However, the functionality of existing packages does not allow to fully implement the desired computation scheme.

    The aim of the research is to develop a unified environment for modeling hybrid superconductor/magnet nanostructures, providing access to solvers and developed algorithms, and based on a heterogeneous computing paradigm that allows research of superconducting elements in nanoscale structures with magnets and hybrid quantum materials. In this paper, we investigate resonant phenomena in the nanomagnet system associated with the Josephson junction. Such a system has rich resonant physics. To study the possibility of magnetic reversal depending on the model parameters, it is necessary to solve numerically the Cauchy problem for a system of nonlinear equations. For numerical simulation of hybrid superconductor/magnet nanostructures, a computing environment based on the heterogeneous HybriLIT computing platform is implemented. During the calculations, all the calculation times obtained were averaged over three launches. The results obtained here are of great practical importance and provide the necessary information for evaluating the physical parameters in superconductor/magnet hybrid nanostructures.

  6. Sukhov E.A., Chekina E.A.
    Software complex for numerical modeling of multibody system dynamics
    Computer Research and Modeling, 2024, v. 16, no. 1, pp. 161-174

    This work deals with numerical modeling of motion of the multibody systems consisting of rigid bodies with arbitrary masses and inertial properties. We consider both planar and spatial systems which may contain kinematic loops.

    The numerical modeling is fully automatic and its computational algorithm contains three principal steps. On step one a graph of the considered mechanical system is formed from the userinput data. This graph represents the hierarchical structure of the mechanical system. On step two the differential-algebraic equations of motion of the system are derived using the so-called Joint Coordinate Method. This method allows to minimize the redundancy and lower the number of the equations of motion and thus optimize the calculations. On step three the equations of motion are integrated numerically and the resulting laws of motion are presented via user interface or files.

    The aforementioned algorithm is implemented in the software complex that contains a computer algebra system, a graph library, a mechanical solver, a library of numerical methods and a user interface.

  7. Krivovichev G.V.
    Difference splitting schemes for the system of one-dimensional equations of hemodynamics
    Computer Research and Modeling, 2024, v. 16, no. 2, pp. 459-488

    The work is devoted to the construction and analysis of difference schemes for a system of hemodynamic equations obtained by averaging the hydrodynamic equations of a viscous incompressible fluid over the vessel cross-section. Models of blood as an ideal and as a viscous Newtonian fluid are considered. Difference schemes that approximate equations with second order on the spatial variable are proposed. The computational algorithms of the constructed schemes are based on the method of splitting on physical processes. According to this approach, at one time step, the model equations are considered separately and sequentially. The practical implementation of the proposed schemes at each time step leads to a sequential solution of two linear systems with tridiagonal matrices. It is demonstrated that the schemes are $\rho$-stable under minor restrictions on the time step in the case of sufficiently smooth solutions.

    For the problem with a known analytical solution, it is demonstrated that the numerical solution has a second order convergence in a wide range of spatial grid step. The proposed schemes are compared with well-known explicit schemes, such as the Lax – Wendroff, Lax – Friedrichs and McCormack schemes in computational experiments on modeling blood flow in model vascular systems. It is demonstrated that the results obtained using the proposed schemes are close to the results obtained using other computational schemes, including schemes constructed by other approaches to spatial discretization. It is demonstrated that in the case of different spatial grids, the time of computation for the proposed schemes is significantly less than in the case of explicit schemes, despite the need to solve systems of linear equations at each step. The disadvantages of the schemes are the limitation on the time step in the case of discontinuous or strongly changing solutions and the need to use extrapolation of values at the boundary points of the vessels. In this regard, problems on the adaptation of splitting schemes for problems with discontinuous solutions and in cases of special types of conditions at the vessels ends are perspective for further research.

  8. Salem N., Al-Tarawneh K., Hudaib A., Salem H., Tareef A., Salloum H., Mazzara M.
    Generating database schema from requirement specification based on natural language processing and large language model
    Computer Research and Modeling, 2024, v. 16, no. 7, pp. 1703-1713

    A Large Language Model (LLM) is an advanced artificial intelligence algorithm that utilizes deep learning methodologies and extensive datasets to process, understand, and generate humanlike text. These models are capable of performing various tasks, such as summarization, content creation, translation, and predictive text generation, making them highly versatile in applications involving natural language understanding. Generative AI, often associated with LLMs, specifically focuses on creating new content, particularly text, by leveraging the capabilities of these models. Developers can harness LLMs to automate complex processes, such as extracting relevant information from system requirement documents and translating them into a structured database schema. This capability has the potential to streamline the database design phase, saving significant time and effort while ensuring that the resulting schema aligns closely with the given requirements. By integrating LLM technology with Natural Language Processing (NLP) techniques, the efficiency and accuracy of generating database schemas based on textual requirement specifications can be significantly enhanced. The proposed tool will utilize these capabilities to read system requirement specifications, which may be provided as text descriptions or as Entity-Relationship Diagrams (ERDs). It will then analyze the input and automatically generate a relational database schema in the form of SQL commands. This innovation eliminates much of the manual effort involved in database design, reduces human errors, and accelerates development timelines. The aim of this work is to provide a tool can be invaluable for software developers, database architects, and organizations aiming to optimize their workflow and align technical deliverables with business requirements seamlessly.

  9. Varshavsky L.E.
    Iterative decomposition methods in modelling the development of oligopolistic markets
    Computer Research and Modeling, 2025, v. 17, no. 6, pp. 1237-1256

    One of the principles of forming a competitive market environment is to create conditions for economic agents to implement Nash – Cournot optimal strategies. With the standard approach to determining Nash – Cournot optimal market strategies, economic agents must have complete information about the indicators and dynamic characteristics of all market participants. Which is not true.

    In this regard, to find Nash – Cournot optimal solutions in dynamic models, it is necessary to have a coordinator who has complete information about the participants. However, in the case of a large number of game participants, even if the coordinator has the necessary information, computational difficulties arise associated with the need to solve a large number of coupled equations (in the case of linear dynamic games — Riccati matrix equations).

    In this regard, there is a need to decompose the general problem of determining optimal strategies for market participants into private (local) problems. Approaches based on the iterative decomposition of coupled matrix Riccati equations and the solution of local Riccati equations were studied for linear dynamic games with a quadratic criterion. This article considers a simpler approach to the iterative determination of the Nash – Cournot equilibrium in an oligopoly, by decomposition using operational calculus (operator method).

    The proposed approach is based on the following procedure. A virtual coordinator, which has information about the parameters of the inverse demand function, forms prices for the prospective period. Oligopolists, given fixed price dynamics, determine their strategies in accordance with a slightly modified optimality criterion. The optimal volumes of production of the oligopolists are sent to the coordinator, who, based on the iterative algorithm, adjusts the price dynamics at the previous step.

    The proposed procedure is illustrated by the example of a static and dynamic model of rational behavior of oligopoly participants who maximize the net present value (NPV). Using the methods of operational calculus (and in particular, the inverse Z-transformation), conditions are found under which the iterative procedure leads to equilibrium levels of price and production volumes in the case of linear dynamic games with both quadratic and nonlinear (concave) optimization criteria.

    The approach considered is used in relation to examples of duopoly, triopoly, duopoly on the market with a differentiated product, duopoly with interacting oligopolists with a linear inverse demand function. Comparison of the results of calculating the dynamics of price and production volumes of oligopolists for the considered examples based on coupled equations of the matrix Riccati equations in Matlab (in the table — Riccati), as well as in accordance with the proposed iterative method in the widely available Excel system shows their practical identity.

    In addition, the application of the proposed iterative procedure is illustrated by the example of a duopoly with a nonlinear demand function.

  10. Vetchanin E.V., Tenenev V.A., Shaura A.S.
    Motion control of a rigid body in viscous fluid
    Computer Research and Modeling, 2013, v. 5, no. 4, pp. 659-675

    We consider the optimal motion control problem for a mobile device with an external rigid shell moving along a prescribed trajectory in a viscous fluid. The mobile robot under consideration possesses the property of self-locomotion. Self-locomotion is implemented due to back-and-forth motion of an internal material point. The optimal motion control is based on the Sugeno fuzzy inference system. An approach based on constructing decision trees using the genetic algorithm for structural and parametric synthesis has been proposed to obtain the base of fuzzy rules.

    Views (last year): 2. Citations: 1 (RSCI).
Pages: « first previous next last »

Indexed in Scopus

Full-text version of the journal is also available on the web site of the scientific electronic library eLIBRARY.RU

The journal is included in the Russian Science Citation Index

The journal is included in the RSCI

International Interdisciplinary Conference "Mathematics. Computing. Education"