Work place: Ivan Franko National University of Lviv, Lviv, 79000, Ukraine
E-mail: Lyubomyr.Chyrun@lnu.edu.ua
Website: https://orcid.org/0000-0002-9448-1751
Research Interests: Machine Learning, Natural Language Processing, Data Science, Web Technologies
Biography
Lyubomir Chyrun is an Associate Professor at the Ivan Franko National University of Lviv. Since 2001, he worked at the Lviv Polytechnic University at the Institute of Computer Sciences and Information Technologies. in the position of associate professor of the Department of Information Systems and Networks. In 2007 defended his PhD thesis on May 1, 2002 - "Mathematical modeling and computational methods". He is the co-author of the monograph "Continuous Fractions and Complex Numbers". Author of more than 120 scientific publications. His areas of scientific interest are attern recognition, application of numerical methods in information technologies, object-oriented programming, web technologies, fake identification, natural language processing, computer linguistics, data science, system analysis, information technologies, and machine learning.
By Danylo Levkivskyi Victoria Vysotska Lyubomyr Chyrun Yuriy Ushenko Dmytro Uhryn Cennuo Hu
DOI: https://doi.org/10.5815/ijieeb.2025.02.01, Pub. Date: 8 Apr. 2025
Research devoted to the categorization and creation of semantic annotations for scientific articles stands out as an essential direction of development in the context of the growing volume of scientific literature. The application of machine learning and natural language processing in this field allows you to effectively organize and provide access to scientific information. The article discusses methods of automatic annotation of texts. Based on the review, the use of the constraint propagation model is proposed to improve the technique of text relationship maps. The developed software system is aimed at automating the process of analysis and categorization of scientific materials, which opens the way to improving the speed and accuracy of searching for the necessary information for researchers. The use of advanced machine learning models, such as roBERTa and RAG, ensures the highest quality of data processing and creation of semantic annotations. The accuracy of predicting article categories after improving the model reached 88%. The novelty of the approach is the combination of categorization and semantic annotation to increase the convenience and speed of searching for scientific information. The software system opens up opportunities for future expansion and improvement through the use of advanced technologies and machine learning models. This study is noted for its relevance, originality of approach and potential for practical application in the field of scientific research and development of science as a whole. The proposed approach contributes to the development of the Information Engineering and Electronic Business industry through the following key aspects: automation of categorization and annotation of scientific articles, improving the accuracy of information search, increasing the efficiency of scientific research, and the flexibility and scalability of the solution.
[...] Read more.By Dmytro Uhryn Victoria Vysotska Lyubomyr Chyrun Sofia Chyrun Cennuo Hu Yuriy Ushenko
DOI: https://doi.org/10.5815/ijisa.2025.02.05, Pub. Date: 8 Apr. 2025
During the development and implementation of the software system for text analysis, attention was focused on the morphological, syntactic and stylistic levels of the language, which made it possible to develop detailed profiles of authorship for various writers. The main goal of the system is to automate the process of identifying authorship and detecting plagiarism, which ensures the protection of intellectual property and contributes to the preservation of cultural heritage. The scientific novelty of the research was manifested in the development of specific algorithms adapted to the peculiarities of the natural language, as well as in the use of advanced technologies, such as deep learning and big data. The introduction of the interdisciplinary approach, which combines computer science, linguistics, and literary studies, has opened up new perspectives for the detailed analysis of scholarly works. The results of the work confirm the high efficiency and accuracy of the system in authorship identification, which can serve as an essential tool for scientists, publishers, and law enforcement agencies. In addition to technical aspects, it is vital to take into account ethical issues related to confidentiality and copyright protection, which puts under control not only the technological side of the process but also moral and legal norms. Thus, the work revealed the importance and potential of using modern text processing methods for improving literary analysis and protecting cultural heritage, which makes it significant for further research and practical use in this area.
[...] Read more.By Lyubomyr Chyrun Victoria Vysotska Sofia Chyrun Zhengbing Hu Yuriy Ushenko Dmytro Uhryn
DOI: https://doi.org/10.5815/ijigsp.2025.02.01, Pub. Date: 8 Apr. 2025
The study considers the methodology of using continued fractions to approximate transfer functions in speech synthesis systems. The main results of the research are an increase in the accuracy of approximation, acceleration of calculations, and a new method of convergence analysis. The use of continued fractions allowed for a reduction in the error of approximation of transfer functions compared to classical methods. With an error of 1.0E-06, the continued fraction method requires only 3–13 terms, while the power series requires 3–15 terms. The use of continued fractions reduced the time for calculating transfer functions by 2–3%. It was determined that the most effective for calculating the values of continued fractions are the Δ-algorithm and the α-algorithm. A new criterion for the convergence of continued fractions is proposed, which allows the sum fractions that are "divergent" in the classical sense. The graphs used to classify different types of continued fractions allowed us to better understand their structure and potential for application in speech synthesis. Software for calculating transfer function values based on continued fraction decomposition has been developed and tested. It has allowed automation of the approximation process and increased the efficiency of speech synthesis systems. The results obtained have allowed improving the quality of synthesised speech while simultaneously reducing the complexity of calculations. Systems using continued fractions consume less memory and provide more accurate voice reproduction. In summary, the work presents a new approach to the approximation of transfer functions, which is essential for optimising speech synthesis systems.
[...] Read more.By Zhengbing Hu Victoria Vysotska Lyubomyr Chyrun Roman Romanchuk Yuriy Ushenko Dmytro Uhryn Cennuo Hu
DOI: https://doi.org/10.5815/ijmecs.2025.02.02, Pub. Date: 8 Apr. 2025
The main goal of the work is to create an intelligent system that uses NLP methods and machine learning algorithms to analyse and classify textual content authorship. The following machine learning models for English and Ukrainian publications were tested and trained on the dataset: Support Vector Classifier, Random Forest, Naive Bayes, Logistic Regression and Neuron Networks. For English, the accuracy of the models was higher due to the more significant amount of text data available. The results for English fiction publication show that the Neuron Networks classifier outperforms the other models in all evaluated metrics, achieving the highest accuracy (0.97), recall (0.96), F1 score (0.98), and precision (0.96). It shows that Neuron Networks is particularly effective in capturing distinctive features of the writing styles of different English authors in scientific and technical texts. For the Ukrainian language, there is a drop in accuracy by 5-10% due to the smaller number of corpora of texts for teaching. The results for scientific and technical Ukrainian publications show that the Random Forest classifier outperforms the other models in all evaluated metrics, achieving the highest accuracy (0.88), recall (0.87), F1 score (0.87), and precision (0.87). It shows that Random Forest is particularly effective in capturing distinctive features of the writing styles of different Ukrainian authors in scientific and technical texts. Much worse accuracy results were shown by other models such as Support Vector Classifier (77%), Logistic Regression (73%) and Naive Bayes (70%). The results for the Ukrainian fiction publication show that the Random Forest classifier outperforms the other models in all evaluated metrics, achieving the highest accuracy (0.85), recall (0.84), F1 score (0.84), and precision (0.84). Much worse accuracy results were shown by other models such as Support Vector Classifier (77%), Logistic Regression (73%) and Naive Bayes (70%)
[...] Read more.By Lyubomyr Chyrun Victoria Vysotska Stepan Tchynetskyi Yuriy Ushenko Dmytro Uhryn
DOI: https://doi.org/10.5815/ijisa.2024.06.03, Pub. Date: 8 Dec. 2024
The goal of designing and implementing an intelligent information system for the recognition and classification of sound signals is to create an effective solution at the software level, which would allow analysis, recognition, classification and forecasting of sound signals in megacities and smart cities using machine learning methods. This system can help people in various fields to simplify their lives, for example, it can help farmers protect their crops from animals, in the military it can help with the identification of weapons and the search for flying objects, such as drones or missiles, in the future there is a possibility for recognizing the distance to sound, also, in cities can help with security, so a preventive response system can be built, which can check if everything is in order based on sounds. Also, it can make life easier for people with impaired hearing to detect danger in everyday life. In the part of the comparison of analogues of the developed product, 4 analogues were found: Shazam, sound recognition from Apple, Vocapia, and SoundHound. A table of comparisons was made for these analogues and the product under development. Also, after comparing analogues, a table for evaluating the effects of the development was built. During the system analysis section, a variety of audio research materials were developed to indicate the characteristics that can be used for this design: period, amplitude, and frequency, and, as an example, an article on real-world audio applications is shown. A precedent scenario is described using the RUP methodology and UML diagrams are constructed: Diagram of use cases; Class diagram; Activity chart; Sequence diagram; Diagram of components; and Deployment diagram. Also, sound data analysis was performed, sound data was visualized as spectrograms and sound waves, which clearly show that the data are different, so it is possible to classify them using machine learning methods. An experimental selection of the machine learning method as staandart clasificers for building a sound recognition model was made. The best method turned out to be SVC, the accuracy of which reflects more than 30 per cent. A neural network was also implemented to improve the obtained results. The result of training a model based on a neural network during 100 epochs achieved a result of 97.7% accuracy for training data and 47.8% accuracy when checking performance on test data. This result should be higher, so it is necessary to consider improving recognition algorithms, increasing the amount of data, and changing the recognition method. Testing of the project was carried out, showing its operation and pointing out shortcomings that need to be corrected in the future.
[...] Read more.By Yevgen Burov Victoria Vysotska Lyubomyr Chyrun Yuriy Ushenko Dmytro Uhryn Zhengbing Hu
DOI: https://doi.org/10.5815/ijieeb.2024.05.01, Pub. Date: 8 Oct. 2024
The use of ontological models for intelligent systems construction allows for improved quality characteristics at all stages of the life cycle of a software product. The main source of improvement in quality characteristics is the possibility of reusing the conceptualization and code provided by the corresponding models. Due to the use of a single conceptualization when creating various software products, the degree of interoperability and code portability increases. The new-generation electronic business analytics systems implementation is based on the use of active models for business processes (BP). Such models, on the one hand, reflect the BPs taking place in the organization on a real-time scale, and on the other hand, embody corporate and other regulatory rules and restrictions and monitor their compliance. The purpose of this article is to research the methods of presenting and building active executable BP models, determining the methods of their execution and coordination, and building the resulting intelligent network of BP models. In the process of its implementation, such a network ensures the implementation, support of decision-making and compliance with regulatory rules in the relevant real BPs. A formal specification of an intelligent system for modelling a complex of BPs of the enterprise using models has been proposed. A hierarchical approach to the introduction of intelligent functions into the modelling system has been proposed. The simulation system is designed to be used for the design and management of complex intelligent systems. Achieving the set goal involves solving several development tasks: methods of presenting BP models for different types of such models; methods of analysis and display of time relations and attributes in BP models; ways of presenting the association of artefacts, and business analytics models with individual BP operations; metric ratios for evaluating the quality of process execution; methods of interaction of various BPs and coordination of their implementation. The purpose of functioning an intelligent model-driven software system is achieved through the interaction of a large number of simple models. At the same time, each model encapsulates a certain aspect of the expert's knowledge about the subject area. To apply executable conceptual models in the field of modelling BPes, it is necessary to determine the types of conceptual models used, their purpose and functions, and the role they play in the operation of an intelligent system. Models used in modelling BPes can be classified according to various characteristics. At the same time, the same model can be included in different classifications.
[...] Read more.By Victoria Vysotska Krzysztof Przystupa Lyubomyr Chyrun Serhii Vladov Yuriy Ushenko Dmytro Uhryn Zhengbing Hu
DOI: https://doi.org/10.5815/ijcnis.2024.05.06, Pub. Date: 8 Oct. 2024
A new method of propaganda analysis is proposed to identify signs and change the dynamics of the behaviour of coordinated groups based on machine learning at the processing disinformation stages. In the course of the work, two models were implemented to recognise propaganda in textual data - at the message level and the phrase level. Within the framework of solving the problem of analysis and recognition of text data, in particular, fake news on the Internet, an important component of NLP technology (natural language processing) is the classification of words in text data. In this context, classification is the assignment or assignment of textual data to one or more predefined categories or classes. For this purpose, the task of binary text classification was solved. Both models are built based on logistic regression, and in the process of data preparation and feature extraction, such methods as vectorisation using TF-IDF vectorisation (Term Frequency – Inverse Document Frequency), the BOW model (Bag-of-Words), POS marking (Part-Of-Speech), word embedding using the Word2Vec two-layer neural network, as well as manual feature extraction methods aimed at identifying specific methods of political propaganda in texts are used. The analogues of the project under development are analysed the subject area (the propaganda used in the media and the basis of its production methods) is studied. The software implementation is carried out in Python, using the seaborn, matplotlib, genism, spacy, NLTK (Natural Language Toolkit), NumPy, pandas, scikit-learn libraries. The model's score for propaganda recognition at the phrase level was obtained: 0.74, and at the message level: 0.99. The implementation of the results will significantly reduce the time required to make the most appropriate decision on the implementation of counter-disinformation measures concerning the identified coordinated groups of disinformation generation, fake news and propaganda. Different classification algorithms for detecting fake news and non-fakes or fakes identification accuracy from Internet resources ana social mass media are used as the decision tree (for non-fakes identification accuracy 0.98 and fakes identification accuracy 0.9903), the k-nearest neighbours (0.83/0.999), the random forest (0.991/0.933), the multilayer perceptron (0.9979/0.9945), the logistic regression (0.9965/0.9988), and the Bayes classifier (0.998/0.913). The logistic regression (0.9965) the multilayer perceptron (0.9979) and the Bayesian classifier (0.998) are more optimal for non-fakes news identification. The logistic regression (0.9988), the multilayer perceptron (0.9945), and k-nearest neighbours (0.999) are more optimal for identifying fake news identification.
[...] Read more.Subscribe to receive issue release notifications and newsletters from MECS Press journals