All issues
- 2024 Vol. 16
- 2023 Vol. 15
- 2022 Vol. 14
- 2021 Vol. 13
- 2020 Vol. 12
- 2019 Vol. 11
- 2018 Vol. 10
- 2017 Vol. 9
- 2016 Vol. 8
- 2015 Vol. 7
- 2014 Vol. 6
- 2013 Vol. 5
- 2012 Vol. 4
- 2011 Vol. 3
- 2010 Vol. 2
- 2009 Vol. 1
-
Visualization of three-dimensional scenes. Technology for data storing and manipulating
Computer Research and Modeling, 2009, v. 1, no. 2, pp. 119-127This article is devoted to some problems of declaring and storing information for objects' visualization. The storage structure and resources control technology can be applied for real-time visualization of three-dimensional scenes. Such instruments as Sample Framework from DirectX SDK and Direct3D Extension Library (D3DX) were used in the implementation.
Keywords: 3D visualization, information storing.Views (last year): 2. Citations: 2 (RSCI). -
Analytical solution and computer simulation of the task of Rician distribution’s parameters in limiting cases of large and small values of signal-to-noise ratio
Computer Research and Modeling, 2015, v. 7, no. 2, pp. 227-242Views (last year): 2.The paper provides a solution of a task of calculating the parameters of a Rician distributed signal on the basis of the maximum likelihood principle in limiting cases of large and small values of the signal-tonoise ratio. The analytical formulas are obtained for the solution of the maximum likelihood equations’ system for the required signal and noise parameters for both the one-parameter approximation, when only one parameter is being calculated on the assumption that the second one is known a-priori, and for the two-parameter task, when both parameters are a-priori unknown. The direct calculation of required signal and noise parameters by formulas allows escaping the necessity of time resource consuming numerical solving the nonlinear equations’ s system and thus optimizing the duration of computer processing of signals and images. There are presented the results of computer simulation of a task confirming the theoretical conclusions. The task is meaningful for the purposes of Rician data processing, in particular, magnetic-resonance visualization.
-
Theoretical substantiation of the mathematical techniques for joint signal and noise estimation at rician data analysis
Computer Research and Modeling, 2016, v. 8, no. 3, pp. 445-473Views (last year): 2. Citations: 2 (RSCI).The paper provides a solution of the two-parameter task of joint signal and noise estimation at data analysis within the conditions of the Rice distribution by the techniques of mathematical statistics: the maximum likelihood method and the variants of the method of moments. The considered variants of the method of moments include the following techniques: the joint signal and noise estimation on the basis of measuring the 2-nd and the 4-th moments (MM24) and on the basis of measuring the 1-st and the 2-nd moments (MM12). For each of the elaborated methods the explicit equations’ systems have been obtained for required parameters of the signal and noise. An important mathematical result of the investigation consists in the fact that the solution of the system of two nonlinear equations with two variables — the sought for signal and noise parameters — has been reduced to the solution of just one equation with one unknown quantity what is important from the view point of both the theoretical investigation of the proposed technique and its practical application, providing the possibility of essential decreasing the calculating resources required for the technique’s realization. The implemented theoretical analysis has resulted in an important practical conclusion: solving the two-parameter task does not lead to the increase of required numerical resources if compared with the one-parameter approximation. The task is meaningful for the purposes of the rician data processing, in particular — the image processing in the systems of magnetic-resonance visualization. The theoretical conclusions have been confirmed by the results of the numerical experiment.
-
Development of anisotropic nonlinear noise-reduction algorithm for computed tomography data with context dynamic threshold
Computer Research and Modeling, 2019, v. 11, no. 2, pp. 233-248Views (last year): 21.The article deals with the development of the noise-reduction algorithm based on anisotropic nonlinear data filtering of computed tomography (CT). Analysis of domestic and foreign literature has shown that the most effective algorithms for noise reduction of CT data use complex methods for analyzing and processing data, such as bilateral, adaptive, three-dimensional and other types of filtrations. However, a combination of such techniques is rarely used in practice due to long processing time per slice. In this regard, it was decided to develop an efficient and fast algorithm for noise-reduction based on simplified bilateral filtration method with three-dimensional data accumulation. The algorithm was developed on C ++11 programming language in Microsoft Visual Studio 2015. The main difference of the developed noise reduction algorithm is the use an improved mathematical model of CT noise, based on the distribution of Poisson and Gauss from the logarithmic value, developed earlier by our team. This allows a more accurate determination of the noise level and, thus, the threshold of data processing. As the result of the noise reduction algorithm, processed CT data with lower noise level were obtained. Visual evaluation of the data showed the increased information content of the processed data, compared to original data, the clarity of the mapping of homogeneous regions, and a significant reduction in noise in processing areas. Assessing the numerical results of the algorithm showed a decrease in the standard deviation (SD) level by more than 6 times in the processed areas, and high rates of the determination coefficient showed that the data were not distorted and changed only due to the removal of noise. Usage of newly developed context dynamic threshold made it possible to decrease SD level on every area of data. The main difference of the developed threshold is its simplicity and speed, achieved by preliminary estimation of the data array and derivation of the threshold values that are put in correspondence with each pixel of the CT. The principle of its work is based on threshold criteria, which fits well both into the developed noise reduction algorithm based on anisotropic nonlinear filtration, and another algorithm of noise-reduction. The algorithm successfully functions as part of the MultiVox workstation and is being prepared for implementation in a single radiological network of the city of Moscow.
-
Signal and noise parameters’ determination at rician data analysis by method of moments of lower odd orders
Computer Research and Modeling, 2017, v. 9, no. 5, pp. 717-728Views (last year): 10. Citations: 1 (RSCI).The paper develops a new mathematical method of the joint signal and noise parameters determination at the Rice statistical distribution by method of moments based upon the analysis of data for the 1-st and the 3-rd raw moments of the random rician value. The explicit equations’ system have been obtained for required parameters of the signal and noise. In the limiting case of the small value of the signal-to-noise ratio the analytical formulas have been derived that allow calculating the required parameters without the necessity of solving the equations numerically. The technique having been elaborated in the paper ensures an efficient separation of the informative and noise components of the data to be analyzed without any a-priori restrictions, just based upon the processing of the results of the signal’s sampled measurements. The task is meaningful for the purposes of the rician data processing, in particular in the systems of magnetic-resonance visualization, in ultrasound visualization systems, at the optical signals’ analysis in range measuring systems, in radio location, etc. The results of the investigation have shown that the two parameter task solution of the proposed technique does not lead to the increase in demanded volume of computing resources compared with the one parameter task being solved in approximation that the second parameter of the task is known a-priori There are provided the results of the elaborated technique’s computer simulation. The results of the signal and noise parameters’ numerical calculation have confirmed the efficiency of the elaborated technique. There has been conducted the comparison of the accuracy of the sought-for parameters estimation by the technique having been developed in this paper and by the previously elaborated method of moments based upon processing the measured data for lower even moments of the signal to be analyzed.
-
Optimization of a hull form for decrease ship resistance to movement
Computer Research and Modeling, 2017, v. 9, no. 1, pp. 57-65Views (last year): 10. Citations: 1 (RSCI).Optimization of hull lines for the minimum resistance to movement is a problem of current interest in ship hydrodynamics. In practice, lines design is still to some extent an art. The usual approaches to decrease the ship resistance are based on the model experiment and/or CFD simulation, following the trial and error method. The paper presents a new method of in-detail hull form design based on the wave-based optimization approach. The method provides systematic variation of the hull geometrical form, which corresponds to alteration of longitudinal distribution of the hull volume, while its vertical volume distribution is fixed or highly controlled. It’s well known from the theoretical studies that the vertical distribution can't be optimized by condition of minimum wave resistance, thus it can be neglected for the optimization procedures. The method efficiency was investigated by application to the foreship of KCS, the well-known test object from the workshop Gothenburg-2000. The variations of the longitudinal distribution of the volume were set on the sectional area curve as finite volume increments and then transferred to the lines plan with the help of special frame transformation methods. The CFD towing simulations were carried out for the initial hull form and the six modified variants. According to the simulation results, examined modifications caused the resistance increments in the range 1.3–6.5 %. Optimization process was underpinned with the respective data analysis based on the new hypothesis, according to which, the resistance increments caused by separate longitudinal segments of hull form meet the principle of superposition. The achieved results, which are presented as the optimum distribution of volume present in the optimized designed hull form, which shows the interesting characteristics that its resistance has decrease by 8.9 % in respect to initial KCS hull form. Visualization of the wave patterns showed an attenuation of the transversal wave components, and the intensification of the diverging wave components.
-
High-speed penetration. Discrete-element simulation and experiments
Computer Research and Modeling, 2017, v. 9, no. 6, pp. 937-944Views (last year): 13. Citations: 4 (RSCI).The paper presents the results of numerical simulation and experimental data on the high-speed penetration of the impactor into the obstacle. In the calculations, a discrete-element model has been used, based on the representation of the impactor and the target by a set of close packed interconnected particles. This class of models finds an increasingly wide application in the problems of high-speed interaction of bodies. In the previous works of the authors, the questions of application of the discrete-element model to the problem of the penetration of spherical impactors into massive targets were considered. On the basis of a comparative analysis of the data of computational and physical experiments, it was found out that for a wide class of high-speed penetration problems, a high accuracy of discrete-element modeling can be achieved using the two-parameter Lennard–Jones potential. The binding energy was identified as a function of the dynamic hardness of materials. It was shown that the use of this approach makes it possible to describe accurately the penetration process in the range of impactor velocities 500–2500 m/c.
In this paper, we compare the results of discrete-element modeling with experimental data on penetration of high-strength targets of different thickness by steel impactors. The use of computational parallelization technologies on graphic processors in combination with 3D visualization and animation of the results makes it possible to obtain detailed spatio-temporal patterns of the penetration process and compare them with experimental data.
A comparative analysis of the experimental and calculated data has shown a sufficiently high accuracy of discrete-element modeling for a wide range of target thicknesses: for thin targets pierced with preservation of the integrity of the deformed impactor, for targets of medium thickness, pierced with practically complete fragmentation of the impactor at the exit from the target, and for thick impenetrable targets.
-
Numerical Simulation, Parallel Algorithms and Software for Performance Forecast of the System “Fractured-Porous Reservoir – Producing Well” During its Commissioning Into Operation
Computer Research and Modeling, 2019, v. 11, no. 6, pp. 1069-1075The mathematical model, finite-difference schemes and algorithms for computation of transient thermoand hydrodynamic processes involved in commissioning the unified system including the oil producing well, electrical submersible pump and fractured-porous reservoir with bottom water are developed. These models are implemented in the computer package to simulate transient processes with simultaneous visualization of their results along with computations. An important feature of the package Oil-RWP is its interaction with the special external program GCS which simulates the work of the surface electric control station and data exchange between these two programs. The package Oil-RWP sends telemetry data and current parameters of the operating submersible unit to the program module GCS (direct coupling). The station controller analyzes incoming data and generates the required control parameters for the submersible pump. These parameters are sent to Oil-RWP (feedback). Such an approach allows us to consider the developed software as the “Intellectual Well System”.
Some principal results of the simulations can be briefly presented as follows. The transient time between inaction and quasi-steady operation of the producing well depends on the well stream watering, filtration and capacitive parameters of oil reservoir, physical-chemical properties of phases and technical characteristics of the submersible unit. For the large time solution of the nonstationary equations governing the nonsteady processes is practically identical to the inverse quasi-stationary problem solution with the same initial data. The developed software package is an effective tool for analysis, forecast and optimization of the exploiting parameters of the unified oil-producing complex during its commissioning into the operating regime.
-
Tasks and algorithms for optimal clustering of multidimensional objects by a variety of heterogeneous indicators and their applications in medicine
Computer Research and Modeling, 2024, v. 16, no. 3, pp. 673-693The work is devoted to the description of the author’s formal statements of the clustering problem for a given number of clusters, algorithms for their solution, as well as the results of using this toolkit in medicine.
The solution of the formulated problems by exact algorithms of implementations of even relatively low dimensions before proving optimality is impossible in a finite time due to their belonging to the NP class.
In this regard, we have proposed a hybrid algorithm that combines the advantages of precise methods based on clustering in paired distances at the initial stage with the speed of methods for solving simplified problems of splitting by cluster centers at the final stage. In the development of this direction, a sequential hybrid clustering algorithm using random search in the paradigm of swarm intelligence has been developed. The article describes it and presents the results of calculations of applied clustering problems.
To determine the effectiveness of the developed tools for optimal clustering of multidimensional objects according to a variety of heterogeneous indicators, a number of computational experiments were performed using data sets including socio-demographic, clinical anamnestic, electroencephalographic and psychometric data on the cognitive status of patients of the cardiology clinic. An experimental proof of the effectiveness of using local search algorithms in the paradigm of swarm intelligence within the framework of a hybrid algorithm for solving optimal clustering problems has been obtained.
The results of the calculations indicate the actual resolution of the main problem of using the discrete optimization apparatus — limiting the available dimensions of task implementations. We have shown that this problem is eliminated while maintaining an acceptable proximity of the clustering results to the optimal ones. The applied significance of the obtained clustering results is also due to the fact that the developed optimal clustering toolkit is supplemented by an assessment of the stability of the formed clusters, which allows for known factors (the presence of stenosis or older age) to additionally identify those patients whose cognitive resources are insufficient to overcome the influence of surgical anesthesia, as a result of which there is a unidirectional effect of postoperative deterioration of complex visual-motor reaction, attention and memory. This effect indicates the possibility of differentiating the classification of patients using the proposed tools.
-
Methodology and program for the storage and statistical analysis of the results of computer experiment
Computer Research and Modeling, 2013, v. 5, no. 4, pp. 589-595Views (last year): 1. Citations: 5 (RSCI).The problem of accumulation and the statistical analysis of computer experiment results are solved. The main experiment program is considered as the data source. The results of main experiment are collected on specially prepared sheet Excel with pre-organized structure for the accumulation, statistical processing and visualization of the data. The created method and the program are used at efficiency research of the scientific researches which are carried out by authors.
Indexed in Scopus
Full-text version of the journal is also available on the web site of the scientific electronic library eLIBRARY.RU
The journal is included in the Russian Science Citation Index
The journal is included in the RSCI
International Interdisciplinary Conference "Mathematics. Computing. Education"