Issue 2 (188), article 6

Kibern. vyčisl. teh., 2017, Issue 2 (188), pp.


May 23, 2017 the 80th anniversary of Vladimir Ilyich Gritsenko, known scientist in computer science, information technologies and its applications in economics, industrial and technological field, biological and medical cybernetics, computer technology training, director of the International Scientific and Training Center for Information Technologies and Systems. He is an initiator of development of a new class of high technologies — intelligent information technologies. Gritsenko V.I. is a member of a number of leading international and state councils of Ukraine on informatics, Permanent Representative of Ukraine to the Council of UNESCO Intergovernmental Programme on the information and communications, heads the UNESCO Chair “New Information Technologies in Education for All”, the chief editor of the scientific journals “Control Systems and Machines” and “Cybernetics”.

Download full text (ua)!

Issue 2 (188), article 5


Kibern. vyčisl. teh., 2017, Issue 2 (188), pp.

Rudenko А.V.1, Professor, Corresponding member NAS of Ukraine,
First deputy of director
Nastenko I.А.1,2, Doctor (in Biology), PhD (in Technics),
Head of the Department of Information technologies and mathematical modeling
of physiological processes, Head of the Department of Biomedical Cybernetics
Zhurba O.А.1, Cardiovascular surgeon
Nosovets О.K.2, (in Technics),
Senior lecturer of the Department of Biomedical Cybernetics
Shardukova Y.V.1, Researcher at the Department of Information technology
and mathematical modeling of physiological processes
Lasoryshinets V.V.1, Professor, (in Medicine), Director
1 National Technical University of Ukraine «Igor Sikorsky Kyiv Polytechnic Institute», 37, Pobedi st., Kyiv, 03056, Ukraine
2 Government Facility “M.M. Amosov National Institute of Cardiovascular Surgery of National Academy of Medical Sciences of Ukraine”, 6, Amosov st., Kyiv, 03110, Ukraine


Introduction. The planned beating heart coronary aortic bypass grafting operations (BH CABG) prepares with and without parallel perfusion circulatory support. In second case the necessity of emergent use of circulatory support can appear. In these situations, the frequency of postoperative complications in circulatory system increases.

The decision about the planned use of circulatory support makes with analysis of big number of clinical data. This causes necessity to create the computer decision support systems to minimize the risk of emergent use of parallel perfusion.

The purpose of the article is to analyse statistically the risk factors for BH CABG operations on a working heart with the aim to minimize the risk of circulatory support emergent use.

Clinical material. 972 patients which undergone the BH CABG without circulatory support, 178 patients with planned use of circulatory support and 90 patients with emergent use of circulatory support. The 67 clinical parameters with package IBM SPSS Statistics 21.0 were analyzed.

Methods. The algorithm of binary logistic regression (BLR) for decision support systems development were used.

Results and discussion. The association between clinical values in groups without circulatory support, their planned and emergent use as well as their odds ratios and confidence intervals were analyzed. Then BLR algorithm to build the models for estimation of probability of planned and emergent use of circulatory support was used. The strategies of compulsory and stepwise inclusion of parameters were applied. The sensitivity, specificity and accuracy of the models obtained for learning and examination samples. The best models were chosen were calculated.

Conclusions. The created predictive models have a high sensitivity, specificity and accuracy, and can reduce the influence of subjective factors on medical decisions making regarding the use of the artificial circulatory support during off-pump coronary artery bypass surgery. Analysis of the variables included in the model, can contribute to a better understanding of the structure of existing pathogenic factors. The additional estimation the risk of cardiopulmonary bypass emergent use can reduce the likelihood of such situations occurrence.

Keywords: the beating heart coronary aortic bypass grafting, circulatory support with cardiopulmonary bypass, predicting algorithms.

Download full text (ru)!


1 Mykheev A.A., Klyuzhev V.M., Karpun N.A. Surgery on coronary arteries on a working heart without artificial circulation in IHD patients. M.: Medicine, 2001. 43 p. (in Russian).

2 Allen B. S., Rosenkranz E.R., Buckberg G.D. Studies of controlled reperfusion after ischemia. VII. High oxygen requirements of dyskinetic cardiac muscle. Journal of Thoracic and Cardiovascular Surgery. 1986. No92. P. 543–552.

3 Mo A., Lin H., Wen Z.Efficacy and safety of on-pump beating heart surgery. Ann Thorac Surg. 2008. No 86. P. 1914–1918.

4 Puskas J. Presidential Address, 2009: ISMICS Means Innovation. Innovations: Technology & Techniques in Cardiothoracic & Vascular Surgery. 2009. No 4. P. 240–247.

5 Shabalkyn B.V., Zhbanov Y.V. Minimally invasive myocardial revascularization or aortocoronary bypass without artificial circulation? Bulletin of the Center Bakulev RAMS. V All-Russian Congress of Cardiovascular Surgeons. Novosibirsk, 1999. 152 p. (in Russian).

6 Beauford R.B., Goldstein D.J., Sardari F.F. Multivessel off-pump revascularization in octogenarians: early and midterm outcomes. Ann. Thorac. Surg. 2003. Vol. 76. P. 12–17.

7 Stamou S., Bail A., Boyce S.Coronary revascularization of the circumflex. Ann. Thorac. Surg. 2000. Vol. 70. P. 1371–1377.

8 Witten Ian H., Frank Eibe, Hall Mark A. Data Mining: Practical Machine Learning Tools and Techniques. [3rd Ed.]. Morgan Kaufmann, 2011. P. 664.

9 McHugh M. L. The odds ratio: calculation, usage, and interpretation. Biochemia Medica. 2009. No19 (2). P. 120–126.

10 Sperandei S. Understanding logistic regression analysis. Biochemia Medica. 2014. 24(1). P. 12–18.

Recieved 03.04.2017

Issue 2 (188), article 4


Kibern. vyčisl. teh., 2017, Issue 2 (188), pp.

Grygoryan R.D., Dr (in biology),
Chief of Department of human systems modeling and reliability
Aksenova T.V., Junior-researcher
Degoda A.G., Senior-researcher, PhD (in math.)
Institute of software systems of National Аcademy of Sciences of Ukraine,
40, Acad.Glushkov ave., corp.5, Kiev, 052187, Ukraine


Introduction. Human organism is a too complex system to be empirically examined and comprehended: there is no method for simultaneously measuring or integrally analyzing of billions of multi-scale life variables. Experts need models and information technologies that causally incorporate cell-scale and organism-scale biophysical and physiological data.

The purpose of the article is to describe a specialized simulator (SS) providing users of additional information concerning scenarios and multi-level mechanisms of energy optimization in the human organism.

Results. Multi-scale mechanisms providing cell energy balance (CEB) are in the basis of SS. At cell-level, providers of CEB form a battery of autonomous mechanisms (BAM). Under energy lack (EL), BAM increases the rate of ATP synthesis using local opportunities. If EL remains, extracellular providers of aerobic synthesis of ATP enlarge current potentials of the EL-cell. The SS provides simulation scenarios increasing the lung ventilation, the erythropoiesis, blood concentrations of carbohydrates, and of other nutrients for biogenesis of mitochondria. The role of the cardiovascular system is in regulating of blood incomes to EL-cells.

Conclusion. The SS is a novel informational technology of dual goals. Explaining the function of mechanisms-providers of CEB, the SS also can be used by applied physiologists and physicians for the planning of optimal scenarios for physical empowering of healthy people.

Keywords: mathematical models, mitochondria, glucose, integrative physiology, medicine.

Download full text (ru)!


1 Skulachev V.A.,Bogachev A.V., Kasparinsky F.O. Principles of Bioenergetics. Springer-Verlag, Berlin Heidelberg, 2013. 436 p.

2 Michiels C. Physiological and Pathological Responses to Hypoxia. Am J Pathol. 2004. No164. P. 1875–1882.

3 Kandel J., Angelin A.A., Wallace D.C. Mitochondrial respiration is sensitive to cytoarchitectural breakdown. Integr. Biol. (Camb). 2016, 8 (11). P. 1170–1182.

4 Finsterer J. Hematological manifestations of primary mitochondrial disorders. Acta Haematol. 2007. 118 (2). P. 88–98.

5 Mali V.R., pan G., Deshpande M. Cardiac Mitochondrial Respiratory Dysfunction and Tissue Damage in Chronic Hyperglycemia Correlate with Reduced Aldehyde Dehydrogenase-2 Activity. PLoS One. 2016.11 (10):e0163158.

6 Grygoryan R.D. The Energy Basis of Reversible Adaptation. N.Y.: Nova Science, 2012. 252 p.

7 Grygoryan R.D., Lyabakh K.G. Arterial pressure: a comprehension. Kyiv: ISS of National Academy of Sciences of Ukraine. 2015. 458 p. (In Russian).

8 Grygoryan R.D. The “floating” arterial pressure paradigm: a concept of physiological supersystems. Dusseldorf: Palmarium Academic Publishing. 2016. 417 p. (In Russian).

9 Grygoryan R.D., Deriev I.I., Aksionova T.V. A software simulator of aerobe cell’s responses to energy imbalance. Problems in programming. 2014. No 1. P. 90–98. (In Russian).

10 Grygoryan R.D., Aksionova T.V., Markevich R.A. A software simulator of pancreas. Problems in programming. 2013. No 1. P. 100–106. (In Russian).

11 Grygoryan R.D., Aksionova T.V., Degoda A.G. Modeling of mechanisms and hemodynamic effects of heart hypertrophy. Cybernetics and computer engineering. 2016. Issue.184. P. 72–83. (In Russian).

12 Grygoryan R.D., Aksionova T.V. Modeling of organism-scale mechanisms fighting against energy shortage in cells. Bulletin of University “Ukraine”. Series: Informatics, computers, and cybernetics. 2016. P. 91–99. (In Russian).

13 Aksionova T.V. A software technology providing simulations of mathematical models of physiological systems. Problems in programming. 2012. No1. P. 110–120. (In Russian).

14 Chada S.R., Hollenbeck P.J. Nerve growth factor signaling regulates motility and docking of axonal mitochondria. Curr. Biol. 2004. V.14. P. 1272–1276.

15 Ramamurthy S., Ronnett G. AMP-activated protein kinase (AMPK) and energy-sensing in the brain. Neurobiol. 2012. 21, No 2. P. 52–60.

Recieved 15.03.2017

Issue 2 (188), article 3


Kibern. vyčisl. teh., 2017, Issue 2 (188), pp.

Aralova N.I., senior researcher of department of optimization of controlled processes
Institute of cybernetics of National Academy of Science of Ukraine,
Acad. Glushkov ave., 40, Kiev, 03680 GSP, Ukraine


Introduction. Under conditions of the physical exertion and human presence on the altitude, the oxygen deficiency in tissues occurs. For a theoretical study of the role of various mechanisms in the regulation of the respiratory system, the use of the mathematical model for the transport of respiratory gases in the body was proposed.

Purpose. Researches of the role of hypoxia, hypercapnia and hypometabolism in external and internal disturbances, based on the mathematical model of the respiratory system.

Results. On the mathematical model of respiratory gas transport in the dynamics of the respiratory cycle, as control parameters, pulmonary ventilation, minute blood volume and local blood flow, as well as self-regulation mechanisms — respiratory muscles, cardiac muscle and smooth muscle vessels — were used. It resolved the conflict situation that arises between the managers and the executive tissues in the fight for oxygen. An analysis of the results of numerical experiments in simulating hypoxia and hypoxic hypoxia and their comparison with experimental data was made.

Conclusion. The proposed approach can be useful in assessing the role of hypoxia, hypercapnia and hypometabolism in the disturbances of the internal and external environment in the process of human vital activity under extreme conditions and leads to the formulation of new tasks in the physiology of sports, work and leisure.

Keywords: Mathematical model of respiratory gas transport, load hypoxia, hypoxic hypoxia, regulation of the respiratory system, disturbing effects, oxygen deficiency.

Download full text (ru)!


  1. Polinkevich K.B., Onopchuk Yu.N. Conflict situations of regulating ofthe basic function of the body’s breathing and mathematical models for their resolution. Cyberne
    tics. 1986. № 3. P. 100–104 (in Russian).
  2. Bioecomedicine. Unified informative space / ed. IN AND. Grytsenko. Kyiv: Nauk. dumka, 2001. 314 p. (in Russian).
  3. Filippov M.M. Modes of mass transfer of oxygen and carbon dioxide in muscle activity. Special and clinical physiology of hypoxic conditions. Kiev: Nauk. dumka, 1979. 3. P. 208–214 (in Russian).
  4. Secondary tissue hypoxia / under the general. Ed. A.Z. Kolchinskaya. K.: Nauk. dumka. 1983. 253 p. (in Russian).
  5. Kolchinskaya A.Z. On the classification of hypoxic states. Pathological physiology and experiment. Therapy. 1981. Issue 4. P. 9–10 (in Russian).
  6. Mudrik V.I. Features of the development of oxygen deficiency in humans under the influence of muscular activity of moderate intensity. Special and clinical physiology of hypoxic conditions. Kiev: Nauk.duma, 1979. 3. P. 173–178 (in Russian).
  7. Lyabakh E.G. Study of hypoxia in skeletal muscle on a mathematical model. Special and clinical physiology of hypoxic states. Kiev: Nauk. dumka, 1979. 2. P. 189–194 (in Russian).
  8. Kolchinskaya A.Z., Misyura I.N., Mankovskaya A.G. Breathing and oxygen regimes of dolphins. Kiev: Nauk. dumka, 1980. 332 p. (in Russian).
  9. Filippov A.F. Differential equations with discontinuous right-hand side. M.: Nauka, 1985. 224p. (in Russian).
  10. Breslav IS Physiology of respiration. Physiologist. Journal of the USSR. 1979. 15. No 1. P. 3–14 (in Russian).
  11. Molchanova N.I., Marchenko D.I. On the role of hypoxia in the self-regulation of the respiratory system at the physical load. Cybernetics. 1986. No 4. P. 93–96 (in Russian).
  12. Onopchuk Yu.N. Equilibrium systems and transient processes in the systems of external respiration and circulation. Research on the mathematical model. Cybernetics. 1991.
    No 1. P. 136–139. (in Russian).
  13. Grabova NI, Onopchuk Yu.N., Portnichenko V.I. Mathematical models of hypometa-bolism as a mechanism of adaptation of the functional state of the human body. Computer Mathematics. 2009. № 1. P.120–129 (in Russian).
  14. Marchuk G.I. Mathematical modeling in the problem of the environment. M . : Nauka, 1982. 320 p. (in Russian).
  15. Bobriakova I.L. Mathematical modeling of the hypometabolism of the human body. Cybernetics and computer technology. 2014. Issue.178. P. 64–69.

Recieved 29.03.2017

Issue 2 (188), article 2


Kibern. vyčisl. teh., 2017, Issue 2 (188), pp.

Khorozov O.A., Ph.D (Phys-Math), Leading Researcher
Institute of Telecommunications and Global Information Space of the National Academy of Sciences of Ukraine,
Chokolovskiy ave., 13, Kiev 04186, Ukraine


Introduction. The telemonitoring system of patient’s vital signs for primary diagnosis and detection of abnormal values biophysical indicators is described. Expert estimates inherent in fuzzy logic rules are compared with the measured values of the vital signs for disease risk counting. The system is implemented at the Arduino with code for fuzzy logic controller. The structure of distributed management of the warning system is represented.

The purpose of the article is to develop an expert system based on fuzzy logic rules to calculate the risk level of the patient and use feedback control in decision-making.

Method. Expert estimates inherent in fuzzy logic rules are compared with the measured values of the vital signs for disease risk estimation.

Results. Expert system was considered for determination of patient’s health risk level. The fuzzy logic rules was formed for determination of belonging variables to risk groups and used for reflect the input to the decision making process. The application detects anomalous values of monitoring data, generates a medical report and sends it to the server for decision-making. The system includes monitors vital signs of the patients, warning services based on Fuzzy Logic techniques with the objective of reducing the risk from the slow provision of health care. The architecture of the integrated mHealth platform with functional models was proposed.

Conclusions. Telemedicine system was designed for primary diagnosis and monitoring of patients on the basis of fuzzy logic. The method based on expert knowledge, which are incorporated in the rules of fuzzy logic to compare the values of the input parameters of patients and disease risk prediction was used. The technique is common in detecting abnormal values biophysical indicators for disease risk assessment.

Download full text (ua)!


  1. S.Dutta, A.Maeder, J.Basilakis, Using Fuzzy Logic for Decision Support in Vital Signs Monitoring Jjint Workshop Proceedings, 26th Australasian Joint Conference on Artificial Intelligence, 2013, p. 29–33.
  2. M. Mayilvaganan, K. Rajeswari, Risk Factor Analysis to Patient Based on Fuzzy Logic Control System. International Journal of Engineering Research and General Science. 2014. Vol. 2. Issue 5. P. 185–190.
  3. M.K. Choudhury, N. Baruah, A Fuzzy Logic Based Expert System for Denermination of Health Risk Level of Patient. International Journal of research in Engineering and Technology. 2015. Vol. 4. Issue 5. P.261–267.
  4. A.Povoroznjuk, E.Kharchenko, The use of fuzzy logic in computer systems medical diagnostics. Vestnik National Technical University. 2015. № 33. P. 125–133.
  5. Aj O. Alves URL:
  6. S. Sriparasa. JavaScript and JSON Essentials, 2013. URL:

Recieved 09.03.2017

Issue 2 (188), article 1


Kibern. vyčisl. teh., 2017, Issue 2 (188), pp.

Grytsenko V.I., Corresponding Member of NAS of Ukraine, Director
International Research and Training Center for Information Technologies and Systems of the NAS of Ukraine and of Ministry of Education and Science of Ukraine,
av. Acad. Glushkova, 40, Kiev, 03680, Ukraine

Rachkovskij D.A., Doctor of Engineering, Leading Researcher,
Dept. of Neural Information Processing Technologies,
International Research and Training Center for Information Technologies and Systems of the NAS of Ukraine and of Ministry of Education and Science of Ukraine,
av. Acad. Glushkova, 40, Kiev, 03680, Ukraine

Frolov A.A., Doctor of Biology, Professor,
Faculty of Electrical Engineering and Computer Science FEI,
Technical University of Ostrava, 17 listopadu 15, 708 33 Ostrava-Poruba, Czech Republic

Gayler R., PhD,
Independent Researcher,
Melbourne, VIC, Australia

Kleyko D., PhD post graduated,
Department of Computer Science, Electrical and Space Engineering,
Lulea University of Technology, 971 87 Lulea, Sweden

Osipov E., PhD, Professor,
Department of Computer Science, Electrical and Space Engineering,
Lulea University of Technology, 971 87 Lulea, Sweden


Introduction. Neural network models of autoassociative, distributed memory allow storage and retrieval of many items (vectors) where the number of stored items can exceed the vector dimension (the number of neurons in the network). This opens the possibility of a sublinear time search (in the number of stored items) for approximate nearest neighbors among vectors of high dimension.

The purpose of the paper is to review models of autoassociative, distributed memory that can be naturally implemented by neural networks (mainly with local learning rules and iterative dynamics based on information locally available to neurons).

Scope. The survey is focused mainly on the networks of Hopfield, Willshaw, and Potts, that have connections between pairs of neurons and operate on sparse binary vectors. We discuss not only autoassociative memory, but also the generalization properties of these networks. We also consider neural networks with higher-order connections, and networks with a bipartite graph structure for non-binary data with linear constraints.

Conclusions. In conclusion we discuss the relations to similarity search, advantages and drawbacks of these techniques, and topics for further research. An interesting and still not completely resolved question is whether neural autoassociative memories can search for approximate nearest neighbors faster than other index structures for similarity search, in particular for the case of very high dimensional vectors.

Keywords: distributed associative memory, sparse binary vector, Hopfield network, Willshaw memory, Potts model, nearest neighbor, similarity search

Download full text!


1 Abbott L.F., Arian Y. Storage capacity of generalized networks. Physical Review A. 1987. Vol. 36, N 10. P. 5091–5094.

2 Ahle T.D. Optimal las vegas locality sensitive data structures. arXiv:1704.02054. 6 Apr 2017.

3 Aliabadi B. K., Berrou C., Gripon V., Jiang X. Storing sparse messages in networks of neural cliques. IEEE Trans. NNLS. 2014. Vol. 25. P. 980–989.

4 Amari S. Characteristics of sparsely encoded associative memory. Neural Networks. 1989. Vol. 2, N 6. P. 451–457.

5 Amari S., Maginu K. Statistical neurodynamics of associative memory. Neural Networks. 1988. Vol. 1. P. 63–73.

6 Amit D.J. Modeling brain function: the world of attractor neural networks. Cambridge: Cambridge University Press, 1989. 554 p.

7 Amit D.J., Fusi S. Learning in neural networks with material synapses. Neural Computation. 1994. V. 6, N 5. P. 957–982.

8 Amit D.J., Gutfreund H., Sompolinsky H. Statistical mechanics of neural networks near saturation. Annals of Physics. 1987. Vol. 173. P. 30–67.

9 Amosov N. M. Modelling of thinking and the mind. New York: Spartan Books. 1967.

10 Anderson J. A. A theory for the recognition of items from short memorized lists. Psychological Review. 1973. Vol. 80, N 6. P. 417–438.

11 Anderson J. A. Cognitive and psychological computation with neural models. IEEE trans. Systems, Man, and Cybernetics. 1983. Vol. 13, N 5. P. 799–814.

12 Anderson J.A., Murphy G.L. Psychological concepts in a parallel system. Physica D. 1986. Vol. 22, N 1–3. P. 318–336.

13 Anderson J.A., Silverstein J.W., Ritz S.A., Jones R.S. Distinctive features, categorical perception and probability learning: Some applications of a neural model. Psychological Review. 1977. V. 84. P. 413–451.

14 Andoni A., Laarhoven T., Razenshteyn I., Waingarten E. Optimal hashing-based time-space trade-offs for approximate near neighbors. Proc. SODA’17. 2017. P. 47–66.

15 Baidyk T.N., Kussul E.M. Structure of neural assembly. Proc. RNNS/IEEE symposium on neuroinformatics and neurocomputers. 1992. P. 423–434.

16 Baidyk T.N., Kussul E.M., Rachkovskij D.A. Numerical-analytical method for neural network investigation. Proc. NEURONET’90. 1990. P. 217–219.

17 Baldi, P. and Venkatesh, S.S. Number of stable points for spin-glasses and neural networks of higher orders. Physical Review Letters. 1987. Vol. 58, N 9. P. 913–916.

18 Becker A., Ducas L., Gama N., Laarhoven T. New directions in nearest neighbor searching with applications to lattice sieving. Proc. SODA’16. 2016. P. 10–24.

19 Boguslawski B., Gripon V., Seguin F., Heitzmann F. Twin neurons for efficient real-world data distribution in networks of neural cliques: Applications in power management in electronic circuits. IEEE Trans. NNLS. 2016. Vol. 27, N 2. P. 375–387.

20 Bovier A. Sharp upper bounds on perfect retrieval in the Hopfield model. J. Appl. Probab. 1999. Vol. 36, N 3. P. 941–950.

21 Braitenberg V. Cell assemblies in the cerebral cortex. In Theoretical approaches to complex systems. Berlin: Springer-Verlag. 1978. P. 171–188.

22 Broder A., Mitzenmacher M. Network applications of Bloom filters: A survey. Internet mathematics. 2004. Vol. 1, N 4. P. 485–509.

23 Brunel N., Carusi F., Fusi S. Slow stochastic Hebbian learning of classes of stimuli in a recurrent neural network. Network. 1998. Vol. 9. P. 123–152.

24 Buckingham J., Willshaw D. On setting unit thresholds in an incompletely connected associative net. Network. 1993. Vol. 4. P. 441–459.

25 Burshtein D. Non-direct convergence radius and number of iterations of the Hopfield associative memory. IEEE Trans. Inform. Theory. 1994. Vol. 40. P. 838–847.

26 Burshtein D. Long-term attraction in higher order neural networks. IEEE Trans. Neural Networks. 1998. Vol. 9, N 1. P. 42–50.

27 Christiani T., Pagh R. Set similarity search beyond MinHash. Proc. STOC’17. 2017.

28 Cole R., Gottlieb L.-A., Lewenstein M. Dictionary matching and indexing with errors and don’t cares. Proc. STOC’04. 2004. P. 91–100.

29 Dahlgaard S., Knudsen M.B.T., Thorup M. Fast similarity sketching. arXiv:1704.04370. 14 Apr 2017.

30 Demircigil M., Heusel J., Lowe M., Upgang S. Vermet F. On a model of associative memory with huge storage capacity. J. Stat. Phys. doi:10.1007/s10955-017-1806-y. 2017.

31 Donaldson R., Gupta A, Plan Y., Reimer T. Random mappings designed for commercial search engines. arXiv:1507.05929. 21 Jul 2015.

32 Feigelman M.V., Ioffe L.B. The augmented models of associative memory – asymmetric interaction and hierarchy of patterns. Int. Journal of Modern Physics B. 1987. Vol. 1, N 1, P. 51–68.

33 Ferdowsi S., Voloshynovskiy S., Kostadinov D., Holotyak T. Fast content identification in highdimensional feature spaces using sparse ternary codes. Proc. WIFS’16. 2016. P. 1–6.

34 Frolov A.A., Husek D., Muraviev I.P. Information capacity and recall quality in sparsely encoded Hopfield-like neural network: Analytical approaches and computer simulation. Neural Networks. 1997. Vol. 10, N 5. P. 845–855.

35 Frolov A.A., Husek D., Muraviev I.P. Informational efficiency of sparsely encoded Hopfield-like associative memory. Optical Memory & Neural Networks. 2003. Vol. 12, N 3. P. 177–197.

36 Frolov A.A., Husek D., Muraviev I.P., Polyakov P. Boolean factor analysis by attractor neural network. IEEE Trans. Neural Networks. 2007. Vol. 18, N 3. P. 698–707.

37 Frolov A.A., Husek D., Polyakov P.Y. Recurrent neural-network-based boolean factor analysis and its application to word clustering. IEEE Trans. Neural Networks. 2009.Vol. 20, N 7. P. 1073–1086.

38 Frolov A. A., Husek D., Rachkovskij. Time of searching for similar binary vectors in associative memory. Cybernetics and Systems Analysis. 2006. Vol. 42, N 5. P. 615–623.

39 Frolov A.A., Muraviev I.P. Neural models of associative memory. Moscow: Nauka, 1987. 161 p.

40 Frolov A.A., Muraviev I.P. Information characteristics of neural networks. Moscow: Nauka, 1988. 160 p.

41 Frolov A.A., Muraviev I.P. Information characteristics of neural networks capable of associative learning based on Hebbian plasticity. Network. 1993. Vol. 4, N 4. P. 495–536.

42 Frolov A., Kartashov A., Goltsev A., Folk R. Quality and efficiency of retrieval for Willshaw-like autoassociative networks. Correction. Network. 1995. Vol. 6. P. 513–534.

43 Frolov A., Kartashov A., Goltsev A., Folk R. Quality and efficiency of retrieval for Willshaw-like autoassociative networks. Recognition. Network. 1995. Vol. 6. P. 535–549.

44 Frolov A.A., Rachkovskij D.A., Husek D. On information characteristics of Willshaw-like auto-associative memory. Neural Network World. 2002. Vol. 12, N 2. P. 141–157.

45 Gallant S.I., Okaywe T.W. Representing objects, relations, and sequences. Neural Computation. 2013. Vol. 25, N 8. P. 2038–2078.

46 Gardner E. 1987. Multiconnected neural network models. Journal of Physics A. 1998. Vol. 20, N 11. P.3453–3464.

47 Gardner E. The space of interactions in neural-network models. J. Phys. A. 1988. Vol. 21. P. 257–270.

48 Gibson W. G., Robinson J. Statistical analysis of the dynamics of a sparse associative memory. Neural Networks. 1992. Vol. 5. P. 645–661.

49 Golomb D., Rubin N., Sompolinsky H. Willshaw model: Associative memory with sparse coding and low firing rates 1990. Phys Rev A. Vol. 41, N 4. P. 1843–1854.

50 Goltsev A. An assembly neural network for texture segmentation. Neural Networks. 1996. Vol. 9, N 4. P. 643–653.

51 Goltsev A. Secondary learning in the assembly neural network. Neurocomputing. 2004. Vol. 62. P. 405–426.

52 Goltsev A., Husek D. Some properties of the assembly neural networks. Neural Network World. 2002. Vol. 12, N 1. P. 15–32.

53 Goltsev A., Wunsch D.C. Generalization of features in the assembly neural networks. International Journal of Neural Systems. 2004. Vol. 14, N 1. P. 1–18.

54 Goltsev A.D. Neural networks with the assembly organization. Kiev: Naukova Dumka, 2005. 200 p.

55 Goltsev A., Gritsenko V. Modular neural networks with Hebbian learning rule. Neurocomputing. 2009. Vol. 72. P. 2477–2482.

56 Goltsev A., Gritsenko V. Modular neural networks with radial neural columnar architecture. Biologically Inspired Cognitive Architectures. 2015. Vol. 13, P. 63–74.

57 Goswami M., Pagh R., Silvestri F., Sivertsen J. Distance sensitive bloom filters without false negatives. Proc. SODA’17. 2017. P. 257–269.

58 Gripon V., Berrou C. Sparse neural networks with large learning diversity. IEEE Trans. on Neural Networks. 2011. Vol. 22, N 7. P. 1087–1096.

59 Gripon V., Heusel J., Lowe M., Vermet F. A comparative study of sparse associative memories. Journal of Statistical Physics. 2016. Vol. 164. P. 105–129.

60 Gripon V., Lowe M., Vermet F. Associative memories to accelerate approximate nearest neighbor search. ArXiv:1611.05898. 10 Nov 2016.

61 Gritsenko V.I., Rachkovskij D.A., Goltsev A.D., Lukovych V.V., Misuno I.S., Revunova E.G., Slipchenko S.V., Sokolov A.M., Talayev S.A. Neural distributed representation for intelligent information technologies and modeling of thinking. Cybernetics and Computer Engineering. 2013. Vol. 173. P. 7–24.

62 Guo J. K., Brackle D. V., Lofaso N., Hofmann M. O. Vector representation for sub-graph encoding to resolve entities. Procedia Computer Science. 2016. Vol. 95. P. 327–334.

63 Gutfreund H. Neural networks with hierarchically correlated patterns. Physical Review A. 1988. Vol. 37, N 2. P. 570–577.

64 Hacene G. B., Gripon V., Farrugia N., Arzel M., Jezequel M. Finding all matches in a database using binary neural networks. Proc. COGNITIVE’17. 2017. P. 59–64.

65 Hebb D.O. The Organization of Behavior: A Neuropsychological Theory. New York: Wiley, 1949. 335 p.

66 Herrmann M., Ruppin E., Usher M. A neural model of the dynamic activation of memory. Biol. Cybern. 1993. Vol. 68. P. 455–463.

67 Heusel J., Lowe M., Vermet F. On the capacity of an associative memory model based on neural cliques. Statist. Probab. Lett. 2015. Vol. 106. P. 256–261.

68 Hopfield J.J. Neural networks and physical systems with emergent collective computational abilities // Proc. of the Nat. Acad. Sci. USA. 1982. Vol. 79, N 8. P. 2554–2558.

69 Hopfield J.J., Feinstein D.I., Palmer R.G. “Unlearning” has a stabilizing effect in collective memories. Nature. 1983. Vol. 304. P. 158–159.

70 Horn D., Usher M. Capacities of multiconnected memory models. Journal de Physique, 1988. Vol. 49, N 3. P. 389–395.

71 Horner H., Bormann D., Frick M., Kinzelbach H., Schmidt A. Transients and basins of attraction in neutral network models. Z. Physik B. 1989. Vol. 76. P. 381–398.

72 Howard M. W., Kahana M. J. A distributed representation of temporal context. Journal of Mathematical Psychology. 2002. Vol. 46. P. 269–299.

73 Iscen A., Furon T., Gripon V., Rabbat M., Jegou H. Memory vectors for similarity search in high-dimensional spaces. arXiv:1412.3328. 1 Mar 2017.

74 Kakeya H., Kindo T. Hierarchical concept formation in associative memory composed of neuro-window elements. Neural Networks. 1996. Vol. 9, N 7. P. 1095–1098.

75 Kanerva P. Sparse Distributed Memory. Cambridge: MIT Press, 1988. 155 p.

76 Kanerva P. Hyperdimensional computing: an introduction to computing in distributed representation with high-dimensional random vectors. Cognitive Computation. 2009. Vol. 1, N 2. P. 139–159.

77 Kanter I. Potts-glass models of neural networks. Physical Rev. A. 1988. V. 37, N 7. P. 2739–2742.

78 Karbasi A., Salavati A. H., Shokrollahi A. Iterative learning and denoising in convolutional neural associative memories. Proc. ICML’13. 2013. P. 445–453.

79 Kartashov A., Frolov A., Goltsev A., Folk R. Quality and efficiency of retrieval for Willshaw-like autoassociative networks. Willshaw–Potts model. Network. 1997. Vol. 8, N 1. P. 71–86.

80 Kinzel W. Learning and pattern recognition in spin glass models. Z. Physik B. 1985. Vol. 60. P. 205–213.

81 Knoblauch A., Palm G., Sommer F. T. Memory capacities for synaptic and structural plasticity. Neural Computation. 2010. Vol. 22, N 2. P. 289–341.

82 Kleyko D., Khan S., Osipov E., Yong S. P. Modality Classification of Medical Images with Distributed Representations based on Cellular Automata Reservoir Computing. Proc. ISBI’17. 2017. P. 1–4.

83 Kleyko D., Lyamin N., Osipov E., Riliskis L. Dependable MAC layer architecture based on holographic data representation using hyperdimensional binary spatter codes. Proc. MACOM’12. 2012. P. 134–145

84 Kleyko D., Osipov E. Brain-like classifier of temporal patterns. Proc. ICCOINS’14. 2014. P. 1–6.

85 Kleyko D., Osipov E. On bidirectional transitions between localist and distributed representations: The case of common substrings search using vector symbolic architecture. Procedia Computer Science. 2014. Vol. 41. P. 104–113.

86 Kleyko D., Osipov E., Gayler R. W. Recognizing permuted words with Vector Symbolic Architectures: A Cambridge test for machines. Procedia Computer Science. 2016. Vol. 88. P. 169–175.

87 Kleyko D., Osipov E., Gayler R. W., Khan A. I., Dyer A. G. Imitation of honey bees’ concept learning processes using vector symbolic architectures. Biologically Inspired Cognitive Architectures. 2015. Vol. 14. P. 55–72.

88 Kleyko D., Osipov E., Papakonstantinou N., Vyatkin V., Mousavi A. Fault detection in the hyperspace: Towards intelligent automation systems. Proc. INDIN’15. 2015. P. 1219–1224.

89 Kleyko D., Osipov E., Senior A., Khan A. I., Sekercioglu Y. A. Holographic Graph Neuron: a bio-inspired architecture for pattern processing. IEEE Trans. Neural Networks and Learning Systems. 2017. Vol. 28, N 6. P. 1250–1262.

90 Kleyko D., Osipov E., Rachkovskij D. Modification of holographic graph neuron using sparse distributed representations. Procedia Computer Science. 2016. Vol. 88. P. 39–45.

91 Kleyko D., Rahimi A., Rachkovskij D.A., Osipov E., Rabaey J.M. Classification and recall with binary hyperdimensional computing: trade-offs in choice of density and mapping characteristics (2017, Submitted).

92 Kleyko D., Rahimi A, Osipov E. Autoscaling Bloom Filter: controlling trade-off between true and false. arXiv:1705.03934. 10 May 2017

93 Kohonen T. Content-Addressable Memories. Berlin: Springer, 1987. 388 p.

94 Kohring G.A. Neural networks with many-neuron interactions. Journal de Physique. 1990. Vol. 51, N 2. P. 145–155.

95 Krotov D., Hopfield J.J. Dense associative memory for pattern recognition. Proc. NIPS’16. 2016. P. 1172–1180.

96 Krotov D., Hopfield J.J. Dense associative memory is robust to adversarial inputs. arXiv:1701.00939. 4 Jan 2017

97 Kryzhanovsky B.V., Mikaelian A.L., Fonarev A.B. Vector neural net identifing many strongly distorted and correlated patterns. Proc. SPIE. 2005. Vol. 5642. 124–133.

98 Kussul E. M. Associative neuron-like structures. Kiev: Naukova Dumka, 1992.

99 Kussul E. M., Baidyk T. N. A modular structure of associative-projective neural networks. Preprint 93-6. Kiev, Ukraine: GIC, 1993.

100 Kussul E.M., Fedoseyeva T.V. On audio signals recognition in neural assembly structures. Preprint 87-28. Kiev: Inst. of Cybern. 1987. 21 pp.

101 Kussul E., Makeyev O., Baidyk T., Calderon Reyes D. Neural network with ensembles. Proc. IJCNN’10. 2010. P. 2955–2961.

102 Kussul E.M., Rachkovskij D.A. Multilevel assembly neural architecture and processing of sequences. In Neurocomputers and Attention, V. II: Connectionism and neurocomputers. Manchester and New York: Manchester University Press, 1991. P.577–590.

103 Kussul E.M., Rachkovskij D.A., Wunsch D.C. The random subspace coarse coding scheme for real-valued vectors. Proc. IJCNN’99. 1999. P. 450-455.

104 Lansner A. Associative memory models: From the cell assembly theory to biophysically detailed cortex simulations. Trends in Neurosciences. 2009. Vol. 32, N 3. P. 178–186.

105 Lansner A., Ekeberg O. Reliability and speed of recall in an associative network. IEEE Trans. on Pattern Analysis and Machine Intelligence. 1985. Vol. 7. P. 490–498.

106 Levy S. D., Gayler R. Vector Symbolic Architectures: A new building material for artificial general intelligence. Proc. AGI’08. 2008. P. 414–418.

107 Lowe M. On the storage capacity of Hopfield models with correlated patterns. The Annals of Applied Probability. 1998. Vol. 8, N 4. P. 1216–1250.

108 Lowe M., Vermet F. The storage capacity of the Hopfield model and moderate deviations. Statistics and Probability Letters. 2005. Vol. 75. P. 237–248.

109 Lowe M., Vermet F. The capacity of q-state Potts neural networks with parallel retrieval dynamics. Statistics and Probability Letters. 2007. Vol. 77, N 4. P. 1505–1514.

110 Mazumdar A., Rawat A.S. Associative memory via a sparse recovery model. Proc. NIPS’15. 2015. P. 2683–2691.

111 Mazumdar A., Rawat A.S. Associative memory using dictionary learning and expander decoding. Proc. AAAI’17. 2017.

112 McEliece R.J., Posner E.C., Rodemich E.R., Venkatesh S.S. The capacity of the Hopfield associative memory. IEEE Trans. Information Theory. 1987. Vol. 33, N 4. P. 461–482.

113 Misuno I.S., Rachkovskij D.A., Slipchenko S.V. Vector and distributed representations reflecting semantic relatedness of words. Math. machines and systems. 2005. N 3. P. 50–67.

114 Misuno I.S., Rachkovskij D.A., Slipchenko S.V., Sokolov A.M. Searching for text information with the help of vector representations. Probl. Progr. 2005. N 4. P. 50–59.

115 Nadal J.-P. Associative memory: on the (puzzling) sparse coding limit. J. Phys. A. 1991. Vol. 24. P. 1093–1101.

116 Norouzi M., Punjani A., Fleet D. J. Fast exact search in Hamming space with multi-index hashing. IEEE Trans. PAMI. 2014. Vol. 36, N 6. P. 1107–1119.

117 Onizawa N., Jarollahi H., Hanyu T., Gross W.J. Hardware implementation of associative memories based on multiple-valued sparse clustered networks. IEEE Journal on Emerging and Selected Topics in Circuits and Systems. 2016. Vol. 6, N 1. P. 13–24.

118 Palm G. On associative memory. Biological Cybernetics. 1980. Vol. 36. P. 19–31.

119 Palm G. Memory capacity of local rules for synaptic modification. Concepts in Neuroscience. 1991. Vol. 2, N 1. P. 97–128.

120 Palm G. Neural associative memories and sparse coding. Neural Networks. 2013. Vol. 37. P. 165–171.

121 Palm G., Knoblauch A., Hauser F., Schuz A. Cell assemblies in the cerebral cortex. Biol. Cybern. 2014. Vol. 108, N 5. P. 559–572

122 Palm G., Sommer F. T. Information capacity in recurrent Mc.Culloch-Pitts networks with sparsely coded memory states. Network. 1992, Vol. 3, P. 177–186.

123 Parga N., Virasoro M.A. The ultrametric organization of memories in a neural network. J. Mod. Phys. 1986. Vol. 47, N 11. P. 1857–1864.

124 Peretto P., Niez J.J. Long term memory storage capacity of multiconnected neural networks. Biol. Cybern. 1986. Vol. 54, N 1. P. 53–63.

125 Personnaz L., Guyon I., Dreyfus G. Collective computational properties of neural networks: New learning mechanisms. Phys. Rev. A. 1986. Vol. 34, N 5. P. 4217–4228.

126 Plate T. Holographic reduced representation: Distributed representation for cognitive structures. Stanford: CSLI Publications, 2003. 300 p.

127 Rachkovskij D.A. Representation and processing of structures with binary sparse distributed codes. IEEE Transactions on Knowledge and Data Engineering. 2001. Vol. 13, N 2. P. 261–276.

128 Rachkovskij D.A. Some approaches to analogical mapping with structure sensitive distributed representations. Journal of Experimental and Theoretical Artificial Intelligence. 2004. Vol. 16, N 3. P. 125–145.

129 Rachkovskij D.A. Formation of similarity-reflecting binary vectors with random binary projections. Cybernetics and Systems Analysis. 2015. Vol. 51, N 2. P. 313–323.

130 Rachkovskij D.A. Estimation of vectors similarity by their randomized binary projections. Cybernetics and Systems Analysis. 2015. Vol. 51, N 5. P. 808–818.

131 Rachkovskij D.A. Binary vectors for fast distance and similarity estimation. Cybernetics and Systems Analysis. 2017. Vol. 53, N 1. P. 138–156.

132 Rachkovskij D.A. Distance-based index structures for fast similarity search. Cybernetics and Systems Analysis. 2017. Vol. 53, N 4.

133 Rachkovskij D.A. Index structures for fast similarity search of binary vectors. Cybernetics and Systems Analysis. 2017. Vol. 53, N 5.

134 Rachkovskij D.A., Kussul E.M., Baidyk T.N. Building a world model with structure-sensitive sparse binary distributed representations. BICA. 2013. Vol. 3. P. 64–86.

135 Rachkovskij D.A., Misuno I.S., Slipchenko S.V. Randomized projective methods for construction of binary sparse vector representations. Cybernetics and Systems Analysis. 2012. Vol. 48, N 1. P. 146–156.

136 Rachkovskij D.A., Slipchenko S.V. Similarity-based retrieval with structure-sensitive sparse binary distributed representations. Computational Intelligence. 2012. Vol. 28, N 1. P. 106–129.

137 Rachkovskij D.A., Slipchenko S.V., Kussul E.M., Baidyk T.N. Sparse binary distributed encoding of scalars. J. of Automation and Inf. Sci. 2005. Vol. 37, N 6. P. 12–23.
Automat Inf Scien.v37.i6.20

138 Rachkovskij D.A., Slipchenko S.V., Kussul E.M., Baidyk T.N. Properties of numeric codes for the scheme of random subspaces RSC. Cybernetics and Systems Analysis. 2005. 41, N 4. P. 509–520.

139 Rachkovskij D.A., Slipchenko S.V., Misuno I.S., Kussul E.M., Baidyk T.N. Sparse binary distributed encoding of numeric vectors. J. of Automation and Inf. Sci. 2005. Vol. 37, N 11. P. 47–61.
Automat Inf Scien.v37.i11.60

140 Reznik A.M., Sitchov A.S., Dekhtyarenko O.K., Nowicki D.W. Associative memories with killed neurons: the methods of recovery. Proc. IJCNN’03. 2003. P. 2579–2582.

141 Rizzuto D.S., Kahana M.J. An autoassociative neural network model of paired-associate learning. Neural Computation. 2001. Vol. 13. P. 2075–2092.

142 Rolls E. T. Advantages of dilution in the connectivity of attractor networks in the brain. Biologically Inspired Cognitive Architectures. 2012. Vol. 1, P. 44–54.

143 Romani S., Pinkoviezky I., Rubin A., Tsodyks M. Scaling laws of associative memory retrieval. Neural Computation. 2013. Vol. 25, N 10. P. 2523–2544.

144 Rosenfeld R., Touretzky D.S. Coarse-coded symbol memories and their properties. Complex Systems. 1988. Vol. 2, N 4. P. 463–484.

145 Salavati A.H., Kumar K.R., Shokrollahi A. Nonbinary associative memory with exponential pattern retrieval capacity and iterative learning. IEEE Trans. Neural Networks and Learning Systems. 2014. Vol. 25, N 3. P. 557–570.

146 Schwenker F., Sommer F.T., Palm G. Iterative retrieval of sparsely coded associative memory patterns. Neural Networks. 1996. Vol. 9. P. 445-455.

147 Shrivastava A., Li P. In defense of minhash over simhash. Proc. AISTATS’14. 2014. P. 886–894.

148 Slipchenko S.V., Rachkovskij D.A. Analogical mapping using similarity of binary distributed representations. International Journal Information Theories and Applications. 2009. Vol. 16, N 3. P. 269–290.

149 KO5. Tarkoma S., Rothenberg C. E., Lagerspetz E. Theory and Practice of Bloom Filters for Distributed Systems. IEEE Communications Surveys and Tutorials. 2012. Vol. 14, N 1. P. 131–155.

150 Tsodyks M. Associative memory in asymmetric diluted network with low level of activity. Europhysics Letters. 1988. Vol. 7, N 3. 203–208.

151 Tsodyks M. Associative memory in neural networks with the hebbian learning rule. Modern Physics Letters B. 1989. Vol. 3, N 7. P. 555–560.

152 Tsodyks M.V. Associative memory in neural networks with binary synapses. Mod. Phys. Lett. 1990. Vol. B4. P. 713–716.

153 Tsodyks M.V. Hierarchical associative memory in neural networks with low activity level. Modern Physics Letters B. 1990. Vol. 4, N 4. P. 259–265.

154 Tsodyks M., Feigelman M. The enhanced storage capacity in neural networks with low activity level. Europhysics Letters. 1988. Vol. 6, N 2. P. 101–105.

155 Vedenov A. A., Ezhov A.A., Knizhnikova L.A., Levchenko E.B. “Spurious memory” in model neural networks. Preprint IAE-4395/1. 1987. Moscow: KIAE.

156 Willshaw D. Holography, associative memory and inductive generalization. In Parallel Models of Associative Memory. Hillside: Lawrence Erlbaum Associates. 1981. P. 83–104.

157 Willshaw D. J., Buneman O. P., Longuet-Higgins H. C. Non-holographic associative memory. Nature. Vol. 1969. Vol. 222. P. 960–962.

158 Yang X., Vernitski A., Carrea L. An approximate dynamic programming approach for improving accuracy of lossy data compression by Bloom filters. European Journal of Operational Research. 2016. Vol. 252, N 3. P 985–994.

159 Yu C., Gripon V., Jiang X., Jegou H. Neural associative memories as accelerators for binary vector search. Proc. COGNITIVE’15. 2015. P. 85–89.

Recieved 15.04.2017