Journal:Informatica
Volume 35, Issue 3 (2024), pp. 453–481
Abstract
The article focuses on the presentation and comparison of selected heuristic algorithms for solving the inverse problem for the anomalous diffusion model. Considered mathematical model consists of time-space fractional diffusion equation with initial boundary conditions. Those kind of models are used in modelling the phenomena of heat flow in porous materials. In the model, Caputo’s and Riemann-Liouville’s fractional derivatives were used. The inverse problem was based on identifying orders of the derivatives and recreating fractional boundary condition. Taking into consideration the fact that inverse problems of this kind are ill-conditioned, the problem should be considered as hard to solve. Therefore,to solve it, metaheuristic optimization algorithms popular in scientific literature were used and their performance were compared: Group Teaching Optimization Algorithm (GTOA), Equilibrium Optimizer (EO), Grey Wolf Optimizer (GWO), War Strategy Optimizer (WSO), Tuna Swarm Optimization (TSO), Ant Colony Optimization (ACO), Jellyfish Search (JS) and Artificial Bee Colony (ABC). This paper presents computational examples showing effectiveness of considered metaheuristic optimization algorithms in solving inverse problem for anomalous diffusion model.
Pub. online:17 Jun 2024Type:Research ArticleOpen Access
Journal:Informatica
Volume 35, Issue 3 (2024), pp. 483–507
Abstract
Spam reviews are a pervasive problem on online platforms due to its significant impact on reputation. However, research into spam detection in data streams is scarce. Another concern lies in their need for transparency. Consequently, this paper addresses those problems by proposing an online solution for identifying and explaining spam reviews, incorporating data drift adaptation. It integrates (i) incremental profiling, (ii) data drift detection & adaptation, and (iii) identification of spam reviews employing Machine Learning. The explainable mechanism displays a visual and textual prediction explanation in a dashboard. The best results obtained reached up to 87% spam F-measure.
Journal:Informatica
Volume 35, Issue 3 (2024), pp. 509–528
Abstract
This paper attempts to demystify the stability of CoCoSo ranking method via a comprehensive simulation experiment. In the experiment, matrices of different dimensions are generated via Python with fuzzy data. Stability is investigated via adequacy and partial adequacy tests. The test passes if the ranking order does not change even after changes are made to entities, and the partial pass signifies that the top ranked alternative remains intact. Results infer that CoCoSo method has better stability with respect to change of alternatives compared to criteria; and CoCoSo method shows better stability with respect to partial adequacy test for criteria.
Pub. online:22 May 2024Type:Research ArticleOpen Access
Journal:Informatica
Volume 35, Issue 3 (2024), pp. 529–556
Abstract
Artificial Intelligence (AI) in the price management process is being applied in business practice and research to a variety of pricing use cases that can be augmented or automated, providing opportunities as a forecasting tool or for price optimization. However, the complexity of evaluating the technology to prioritize implementation is challenging, especially for small and medium enterprises (SMEs), and guidance is sparse. Which are the relevant stakeholder criteria for a sustainable implementation of AI for pricing purpose? Which type of AI supported price functions meet these criteria best? Theoretically motivated by the hedonic price theory and advances in AI research, we identify nine criteria and eight AI supported price functions (AISPF). A multiple attribute decision model (MADM) using the fuzzy Best Worst Method (BWM) and fuzzy combined compromise solution (CoCoSo) is set up and evaluated by pricing experts from Germany and Spain. To validate our results and model stability, we carried out several random sensitivity analyses based on the weight of criteria exchange. The results suggest accuracy and reliability as the most prominent attribute to evaluate AISPF, while ethical and sustainable criteria are sorted as least important. The AISPF which best meet the criteria are financial prices followed by procurement prices.
Pub. online:16 May 2024Type:Research ArticleOpen Access
Journal:Informatica
Volume 35, Issue 3 (2024), pp. 557–576
Abstract
Ontologies are used to semantically enrich different types of information systems (IS), ensure a reasoning on their content and integrate heterogeneous IS at the semantical level. On the other hand, fuzzy theory is employed in IS for handling the uncertainty and fuzziness of their attributes, resulting in a fully fuzzy IS. As such, ontology- and fuzzy-based IS (i.e. ontology and fuzzy IS) are being developed. So, in this paper, we present a bibliometric analysis of the ontology and fuzzy IS concept to grasp its main ideas, and to increase its body of knowledge by providing a concept map for ontology and fuzzy IS. The main results obtained show that by adding ontologies and fuzzy theory to traditional ISs, they evolve into intelligent ISs capable of managing fuzzy and semantically rich (ontological) information and ensuring knowledge recognition in various fields of application. This bibliometric analysis would enable practitioners and researchers gain a comprehensive understanding of the ontology and fuzzy IS concept that they can eventually adopt for development of intelligent IS in their work.
Journal:Informatica
Volume 35, Issue 3 (2024), pp. 577–600
Abstract
Frequent gradual pattern extraction is an important problem in computer science widely studied by the data mining community. Such a pattern reflects a co-variation between attributes of a database. The applications of the extraction of the gradual patterns concern several fields, in particular, biology, finances, health and metrology. The algorithms for extracting these patterns are greedy in terms of memory and computational resources. This clearly poses the problem of improving their performance. This paper proposes a new approach for the extraction of gradual and frequent patterns based on the reduction of candidate generation and processing costs by exploiting frequent itemsets whose size is a power of two to generate all candidates. The analysis of the complexity, in terms of CPU time and memory usage, and the experiments show that the obtained algorithm outperforms the previous ones and confirms the interest of the proposed approach. It is sometimes at least 5 times faster than previous algorithms and requires at most half the memory.
Pub. online:19 Aug 2024Type:Research ArticleOpen Access
Journal:Informatica
Volume 35, Issue 3 (2024), pp. 601–616
Abstract
One of the main trends for the monitoring and control of business processes is to implement these processes via private blockchain systems. These systems must ensure data privacy and verifiability for the entire network here denoted by ‘Net’. In addition, every business activity should be declared to a trusted third party (TTP), such as an Audit Authority (AA), for tax declaration and collection purposes.
We present a solution for a confidential and verifiable realization of transactions based on the Unspent Transaction Output (UTxO) paradigm. This means that the total sum of transaction inputs (incomes) $In$ must be equal to the total sum of transaction outputs (expenses) $Ex$, satisfying the balance equation $In=Ex$. Privacy in a private blockchain must be achieved through the encryption of actual transaction values. However, it is crucial that all participants in the network be able to verify the validity of the transaction balance equation. This poses a challenge with probabilistically encrypted data. Moreover, the inputs and outputs are encrypted with different public keys. With the introduction of the AA, the number of different public keys for encryption can be reduced to two. Incomes are encrypted with the Receiver’s public key and expenses with the AA’s public key.
The novelty of our realization lies in taking additively-multiplicative, homomorphic ElGamal encryption and integrating it with a proposed paradigm of modified Schnorr identification providing a non-interactive zero-knowledge proof (NIZKP) using a cryptographically secure h-function. Introducing the AA as a structural element in a blockchain system based on the UTxO enables effective verification of encrypted transaction data for the Net. This is possible because the proposed NIZKP is able to prove the equivalency of two ciphertexts encrypted with two different public keys and different actors.
This integration allows all users on the Net to check the UTxO-based transaction balance equation on encrypted data. The security considerations of the proposed solution are presented.
Pub. online:14 May 2024Type:Research ArticleOpen Access
Journal:Informatica
Volume 35, Issue 3 (2024), pp. 617–648
Abstract
This work introduces ALMERIA, a decision-support tool for drug discovery. It estimates compound similarities and predicts activity, considering conformation variability. The methodology spans from data preparation to model selection and optimization. Implemented using scalable software, it handles large data volumes swiftly. Experiments were conducted on a distributed computer cluster using the DUD-E database. Models were evaluated on different data partitions to assess generalization ability with new compounds. The tool demonstrates excellent performance in molecular activity prediction (ROC AUC: 0.99, 0.96, 0.87), indicating good generalization properties of the chosen data representation and modelling. Molecular conformation sensitivity is also evaluated.
Journal:Informatica
Volume 35, Issue 3 (2024), pp. 649–669
Abstract
The growing popularity of mobile and cloud computing raises new challenges related to energy efficiency. This work evaluates four various SQL and NoSQL database solutions in terms of energy efficiency. Namely, Cassandra, MongoDB, Redis, and MySQL are taken into consideration. This study measures energy efficiency of the chosen data storage solutions on a selected set of physical and virtual computing nodes by leveraging Intel RAPL (Running Average Power Limit) technology. Various database usage scenarios are considered in this evaluation including both local usage and remote offloading. Different workloads are benchmarked through the use of YCSB (Yahoo! Cloud Serving Benchmark) tool. Extensive experimental results show that (i) Redis and MongoDB are more efficient in energy consumption under most usage scenarios, (ii) remote offloading saves energy if the network latency is low and destination CPU is significantly more powerful, and (iii) computationally weaker CPUs may sometimes demonstrate higher energy efficiency in terms of J/ops. An energy efficiency measurement framework is proposed in order to evaluate and compare different database solutions based on the obtained experimental results.
Journal:Informatica
Volume 35, Issue 3 (2024), pp. 671–686
Abstract
The development of various digital social network platforms has caused public opinion to play an increasingly important role in the policy making process. However, due to the fact that public opinion hotspots usually change rapidly (such as the phenomenon of public opinion inversion), both the behaviour feature and demand feature of netizens included in the public opinion often vary over time. Therefore, this paper focuses on the feature identification problem of public opinion simultaneously considering the multiple observation time intervals and key time points, in order to support the management of policy-focused online public opinion. According to the variable-scale data analysis theory, the temporal scale space model is established to describe candidate temporal observation scales, which are organized following the time points of relevant policy promulgation (policy time points). After proposing the multi-scale temporal data model, a temporal variable-scale clustering method (T-VSC) is put forward. Compared to the traditional numerical variable-scale clustering method, the proposed T-VSC enables to combine the subjective attention of decision-makers and objective timeliness of public opinion data together during the scale transformation process. The case study collects 48552 raw public opinion data on the double-reduction education policy from Sina Weibo platform during Jan 2023 to Nov 2023. Experimental results indicate that the proposed T-VSC method could divide netizens that participate in the dissemination of policy-focused public opinion into clusters with low behavioural granularity deviation on the satisfied observation time scales, and identify the differentiated demand feature of each netizen cluster at policy time points, which could be applied to build the timely and efficient digital public dialogue mechanism.