Результаты поиска по 'design':
Найдено статей: 99
  1. Shepelev V.D., Kostyuchenkov N.V., Shepelev S.D., Alieva A.A., Makarova I.V., Buyvol P.A., Parsin G.A.
    The development of an intelligent system for recognizing the volume and weight characteristics of cargo
    Computer Research and Modeling, 2021, v. 13, no. 2, pp. 437-450

    Industrial imaging or “machine vision” is currently a key technology in many industries as it can be used to optimize various processes. The purpose of this work is to create a software and hardware complex for measuring the overall and weight characteristics of cargo based on an intelligent system using neural network identification methods that allow one to overcome the technological limitations of similar complexes implemented on ultrasonic and infrared measuring sensors. The complex to be developed will measure cargo without restrictions on the volume and weight characteristics of cargo to be tariffed and sorted within the framework of the warehouse complexes. The system will include an intelligent computer program that determines the volume and weight characteristics of cargo using the machine vision technology and an experimental sample of the stand for measuring the volume and weight of cargo.

    We analyzed the solutions to similar problems. We noted that the disadvantages of the studied methods are very high requirements for the location of the camera, as well as the need for manual operations when calculating the dimensions, which cannot be automated without significant modifications. In the course of the work, we investigated various methods of object recognition in images to carry out subject filtering by the presence of cargo and measure its overall dimensions. We obtained satisfactory results when using cameras that combine both an optical method of image capture and infrared sensors. As a result of the work, we developed a computer program allowing one to capture a continuous stream from Intel RealSense video cameras with subsequent extraction of a three-dimensional object from the designated area and to calculate the overall dimensions of the object. At this stage, we analyzed computer vision techniques; developed an algorithm to implement the task of automatic measurement of goods using special cameras and the software allowing one to obtain the overall dimensions of objects in automatic mode.

    Upon completion of the work, this development can be used as a ready-made solution for transport companies, logistics centers, warehouses of large industrial and commercial enterprises.

  2. Shardyko I.V., Kopylov V.M., Volnyakov K.A.
    Design, modeling, and control of a variable stiffness joint based on a torsional magnetic spring
    Computer Research and Modeling, 2023, v. 15, no. 5, pp. 1323-1347

    Industrial robots have made it possible for robotics to become a worldwide discipline both in economy and in science. However, their capabilities are limited, especially regarding contact tasks where it is required to regulate or at least limit contact forces. At one point, it was noticed that elasticity in the joint transmission, which was treated as a drawback previously, is actually helpful in this regard. This observation led to the introduction of elastic joint robots that are well-suited to contact tasks and cooperative behavior in particular, so they become more and more widespread nowadays. Many researchers try to implement such devices not with trivial series elastic actuators (SEA) but with more sophisticated variable stiffness actuators (VSA) that can regulate their own mechanical stiffness. All elastic actuators demonstrate shock robustness and safe interaction with external objects to some extent, but when stiffness may be varied, it provides additional benefits, e. g., in terms of energy efficiency and task adaptability. Here, we present a novel variable stiffness actuator with a magnetic coupler as an elastic element. Magnetic transmission is contactless and thus advantageous in terms of robustness to misalignment. In addition, the friction model of the transmission becomes less complex. It also has milder stiffness characteristic than typical mechanical nonlinear springs, moreover, the stiffness curve has a maximum after which it descends. Therefore, when this maximum torque is achieved, the coupler slips, and a new pair of poles defines the equilibrium position. As a result, the risk of damage is smaller for this design solution. The design of the joint is thoroughly described, along with its mathematical model. Finally, the control system is also proposed, and simulation tests confirm the design ideas.

  3. Vasiliev A.N., Karp V.P.
    Modeling self-regulation of active neuron in the network
    Computer Research and Modeling, 2012, v. 4, no. 3, pp. 613-619

    A model of the behavior of the active neuron, which was the development of the model described in Shamis A.L. [Shamis, 2006], is designed. Proposed topology is locally connected matrix of the active neural network and the structure integration of information from different sources. An example of the script behavior robot controlled by this neural network is described. The results of experiments with the software implementation of a neural network are presented.

    Views (last year): 1.
  4. Makarov I.S., Bagantsova E.R., Iashin P.A., Kovaleva M.D., Gorbachev R.A.
    Development of and research on an algorithm for distinguishing features in Twitter publications for a classification problem with known markup
    Computer Research and Modeling, 2023, v. 15, no. 1, pp. 171-183

    Social media posts play an important role in demonstration of financial market state, and their analysis is a powerful tool for trading. The article describes the result of a study of the impact of social media activities on the movement of the financial market. The top authoritative influencers are selected. Twitter posts are used as data. Such texts usually include slang and abbreviations, so methods for preparing primary text data, including Stanza, regular expressions are presented. Two approaches to the representation of a point in time in the format of text data are considered. The difference of the influence of a single tweet or a whole package consisting of tweets collected over a certain period of time is investigated. A statistical approach in the form of frequency analysis is also considered, metrics defined by the significance of a particular word when identifying the relationship between price changes and Twitter posts are introduced. Frequency analysis involves the study of the occurrence distributions of various words and bigrams in the text for positive, negative or general trends. To build the markup, changes in the market are processed into a binary vector using various parameters, thus setting the task of binary classification. The parameters for Binance candlesticks are sorted out for better description of the movement of the cryptocurrency market, their variability is also explored in this article. Sentiment is studied using Stanford Core NLP. The result of statistical analysis is relevant to feature selection for further binary or multiclass classification tasks. The presented methods of text analysis contribute to the increase of the accuracy of models designed to solve natural language processing problems by selecting words, improving the quality of vectorization. Such algorithms are often used in automated trading strategies to predict the price of an asset, the trend of its movement.

  5. Zharkova V.V., Schelyaev A.E., Fisher J.V.
    Numerical simulation of sportsman's external flow
    Computer Research and Modeling, 2017, v. 9, no. 2, pp. 331-344

    Numerical simulation of moving sportsman external flow is presented. The unique method is developed for obtaining integral aerodynamic characteristics, which were the function of the flow regime (i.e. angle of attack, flow speed) and body position. Individual anthropometric characteristics and moving boundaries of sportsman (or sports equipment) during the race are taken into consideration.

    Numerical simulation is realized using FlowVision CFD. The software is based on the finite volume method, high-performance numerical methods and reliable mathematical models of physical processes. A Cartesian computational grid is used by FlowVision, the grid generation is a completely automated process. Local grid adaptation is used for solving high-pressure gradient and object complex shape. Flow simulation process performed by solutions systems of equations describing movement of fluid and/or gas in the computational domain, including: mass, moment and energy conservation equations; state equations; turbulence model equations. FlowVision permits flow simulation near moving bodies by means of computational domain transformation according to the athlete shape changes in the motion. Ski jumper aerodynamic characteristics are studied during all phases: take-off performance in motion, in-run and flight. Projected investigation defined simulation method, which includes: inverted statement of sportsman external flow development (velocity of the motion is equal to air flow velocity, object is immobile); changes boundary of the body technology defining; multiple calculations with the national team member data projecting. The research results are identification of the main factors affected to jumping performance: aerodynamic forces, rotating moments etc. Developed method was tested with active sportsmen. Ski jumpers used this method during preparations for Sochi Olympic Games 2014. A comparison of the predicted characteristics and experimental data shows a good agreement. Method versatility is underlined by performing swimmer and skater flow simulation. Designed technology is applicable for sorts of natural and technical objects.

    Views (last year): 29.
  6. Tomonin Y.D., Tominin V.D., Borodich E.D., Kovalev D.A., Dvurechensky P.E., Gasnikov A.V., Chukanov S.V.
    On Accelerated Methods for Saddle-Point Problems with Composite Structure
    Computer Research and Modeling, 2023, v. 15, no. 2, pp. 433-467

    We consider strongly-convex-strongly-concave saddle-point problems with general non-bilinear objective and different condition numbers with respect to the primal and dual variables. First, we consider such problems with smooth composite terms, one of which has finite-sum structure. For this setting we propose a variance reduction algorithm with complexity estimates superior to the existing bounds in the literature. Second, we consider finite-sum saddle-point problems with composite terms and propose several algorithms depending on the properties of the composite terms. When the composite terms are smooth we obtain better complexity bounds than the ones in the literature, including the bounds of a recently proposed nearly-optimal algorithms which do not consider the composite structure of the problem. If the composite terms are prox-friendly, we propose a variance reduction algorithm that, on the one hand, is accelerated compared to existing variance reduction algorithms and, on the other hand, provides in the composite setting similar complexity bounds to the nearly-optimal algorithm which is designed for noncomposite setting. Besides, our algorithms allow one to separate the complexity bounds, i. e. estimate, for each part of the objective separately, the number of oracle calls that is sufficient to achieve a given accuracy. This is important since different parts can have different arithmetic complexity of the oracle, and it is desired to call expensive oracles less often than cheap oracles. The key thing to all these results is our general framework for saddle-point problems, which may be of independent interest. This framework, in turn is based on our proposed Accelerated Meta-Algorithm for composite optimization with probabilistic inexact oracles and probabilistic inexactness in the proximal mapping, which may be of independent interest as well.

  7. Bogdanov A.V., Zaya K., P. Sone K. Ko
    Improvement of computational abilities in computing environments with virtualization technologies
    Computer Research and Modeling, 2015, v. 7, no. 3, pp. 499-504

    In this paper, we illustrates the ways to improve abilities of the computing environments by using virtualization, single system image (SSI) and hypervisor technologies’ collaboration for goal to improve computational abilities. Recently cloud computing as a new service concept has become popular to provide various services to user such as multi-media sharing, online office software, game and online storage. The cloud computing is bringing together multiple computers and servers in a single environment designed to address certain types of tasks, such as scientific problems or complex calculations. By using virtualization technologies, cloud computing environment is able to virtualize and share resources among different applications with the objective for better server utilization, better load balancing and effectiveness.

    Views (last year): 3.
  8. Kalachin S.V.
    Fuzzy modeling the mechanism of transmitting panic state among people with various temperament species
    Computer Research and Modeling, 2021, v. 13, no. 5, pp. 1079-1092

    A mass congestion of people always represents a potential danger and threat for their lives. In addition, every year in the world a very large number of people die because of the crush, the main cause of which is mass panic. Therefore, the study of the phenomenon of mass panic in view of her extreme social danger is an important scientific task. Available information, about the processes of her occurrence and spread refers to the category inaccurate. Therefore, the theory of fuzzy sets has been chosen as a tool for developing a mathematical model of the mechanism of transmitting panic state among people with various temperament species.

    When developing an fuzzy model, it was assumed that panic, from the epicenter of the shocking stimulus, spreads among people according to the wave principle, passing at different frequencies through different environments (types of human temperament), and is determined by the speed and intensity of the circular reaction of the mechanism of transmitting panic state among people. Therefore, the developed fuzzy model, along with two inputs, has two outputs — the speed and intensity of the circular reaction. In the block «Fuzzyfication», the degrees of membership of the numerical values of the input parameters to fuzzy sets are calculated. The «Inference» block at the input receives degrees of belonging for each input parameter and at the output determines the resulting function of belonging the speed of the circular reaction and her derivative, which is a function of belonging for the intensity of the circular reaction. In the «Defuzzyfication» block, using the center of gravity method, a quantitative value is determined for each output parameter. The quality assessment of the developed fuzzy model, carried out by calculating of the determination coefficient, showed that the developed mathematical model belongs to the category of good quality models.

    The result obtained in the form of quantitative assessments of the circular reaction makes it possible to improve the quality of understanding of the mental processes occurring during the transmission of the panic state among people. In addition, this makes it possible to improve existing and develop new models of chaotic humans behaviors. Which are designed to develop effective solutions in crisis situations, aimed at full or partial prevention of the spread of mass panic, leading to the emergence of panic flight and the appearance of human casualties.

     

  9. Aksenov A.A., Kalugina M.D., Lobanov A.I., Kashirin V.S.
    Numerical simulation of fluid flow in a blood pump in the FlowVision software package
    Computer Research and Modeling, 2023, v. 15, no. 4, pp. 1025-1038

    A numerical simulation of fluid flow in a blood pump was performed using the FlowVision software package. This test problem, provided by the Center for Devices and Radiological Health of the US. Food and Drug Administration, involved considering fluid flow according to several design modes. At the same time for each case of calculation a certain value of liquid flow rate and rotor speed was set. Necessary data for calculations in the form of exact geometry, flow conditions and fluid characteristics were provided to all research participants, who used different software packages for modeling. Numerical simulations were performed in FlowVision for six calculation modes with the Newtonian fluid and standard $k-\varepsilon$ turbulence model, in addition, the fifth mode with the $k-\omega$ SST turbulence model and with the Caro rheological fluid model were performed. In the first stage of the numerical simulation, the convergence over the mesh was investigated, on the basis of which a final mesh with a number of cells of the order of 6 million was chosen. Due to the large number of cells, in order to accelerate the study, part of the calculations was performed on the Lomonosov-2 cluster. As a result of numerical simulation, we obtained and analyzed values of pressure difference between inlet and outlet of the pump, velocity between rotor blades and in the area of diffuser, and also, we carried out visualization of velocity distribution in certain cross-sections. For all design modes there was compared the pressure difference received numerically with the experimental data, and for the fifth calculation mode there was also compared with the experiment by speed distribution between rotor blades and in the area of diffuser. Data analysis has shown good correlation of calculation results in FlowVision with experimental results and numerical simulation in other software packages. The results obtained in FlowVision for solving the US FDA test suggest that FlowVision software package can be used for solving a wide range of hemodynamic problems.

  10. Cox M.A., Reed R.G., Mellado B.
    The development of an ARM system on chip based processing unit for data stream computing
    Computer Research and Modeling, 2015, v. 7, no. 3, pp. 505-509

    Modern big science projects are becoming highly data intensive to the point where offline processing of stored data is infeasible. High data throughput computing, or Data Stream Computing, for future projects is required to deal with terabytes of data per second which cannot be stored in long-term storage elements. Conventional data-centres based on typical server-grade hardware are expensive and are biased towards processing power. The overall I/O bandwidth can be increased with massive parallelism, usually at the expense of excessive processing power and high energy consumption. An ARM System on Chip (SoC) based processing unit may address the issue of system I/O and CPU balance, affordability and energy efficiency since ARM SoCs are mass produced and designed to be energy efficient for use in mobile devices. Such a processing unit is currently in development, with a design goal of 20 Gb/s I/O throughput and significant processing power. The I/O capabilities of consumer ARM System on Chips are discussed along with to-date performance and I/O throughput tests.

    Views (last year): 1.
Pages: « first previous next

Indexed in Scopus

Full-text version of the journal is also available on the web site of the scientific electronic library eLIBRARY.RU

The journal is included in the Russian Science Citation Index

The journal is included in the RSCI

International Interdisciplinary Conference "Mathematics. Computing. Education"