IJISA Vol. 9, No. 10, Oct. 2017
Cover page and Table of Contents: PDF (size: 232KB)
One of the major challenges for researchers and governments across the world is reducing resources-waste or loss. Resources loss can happen if there is not a capable control system that contributes to environmental change. The specific aim is to create user-friendly control and monitoring system to reduce the waste in resources. New Artificial intelligence techniques have been introduced to play an important part in developing such systems.
In oilfields, the oil is extracted then distributed via oil pipes until it reaches the end consumer. This operation will occur without a full and complete monitoring for the oil in the pipeline’s journey to the provider. Although, the existing oilfield monitoring systems can communicate locally but they will not send information back to the main provider. That means the provider is not aware of the whole circumstances happened in the transportation process. That gives the provider no control on the process. For example, a sudden decision from the main provider to stop transporting to a specific destination or knowing where the leakage is and which pipe is leaking in the pipelines grid.
This paper, introduces for the first-time oilfield pipeline Neuro-fuzzy (NF) supervision system using Simscape simulation software package. This system can be the first step solution to keep real time communication between the main provider and the oil transportation process in the oilfields and enables the provider to have full supervision on the oil pipes grid. The simulation supervision system illustrates a clear real-time oilfield pipeline grid that gives the provider the ability to control and monitor pipeline grid and prioritise the recovering process. The two parameters selected for control and monitoring were volume and pressure. The results in this paper show full control for the NF supervision system on the transportation process.
A problem of scheduling freight trains in rail-rail transshipment yards is considered. It is solved at a deeper level compared to original papers dedicated to this problem: besides scheduling service slots for trains, this article additionally solves a problem of assigning every train to a railway track. A mathematical model and a solving method for this problem are given. A key feature of the given mathematical model is that it doesn’t use Boolean variables but rather operates with combinatorial objects (tuples of permutations). The solution method is also based on generation of combinatorial sets, which is quite an unusual approach for solving such problems.[...] Read more.
Feature selection plays a very important role in all pattern recognition tasks. It has several benefits in terms of reduced data collection effort, better interpretability of the models and reduced model building and execution time. A lot of problems in feature selection have been shown to be NP – Hard. There has been significant research in feature selection in last three decades. However, the problem of feature selection for clustering is still quite an open area. The main reason is unavailability of target variable as compared to supervised tasks. In this paper, five properties or metafeatures like entropy, skewness, kurtosis, coefficient of variation and average correlation of the features have been studied and analysed. An extensive study has been conducted over 21 publicly available datasets, to evaluate viability of feature elimination strategy based on the values of the metafeatures for feature selection in clustering. A strategy to select the most appropriate metafeatures for a particular dataset has also been outlined. The results indicate that the performance decrease is not statistically significant.[...] Read more.
WWW is a huge repository of information and the amount of information available on the web is growing day by day in an exponential manner. End users make use of search engines like Google, Yahoo, and Bingo etc. for retrieving information. Search engines use web crawlers or spiders which crawl through a sequence of web pages in order to locate the relevant pages and provide a set of links ordered by relevancy. Those indexed web pages are part of surface web. Getting data from deep web requires form submission and is not performed by search engines. Data analytics and data mining applications depend on data from deep web pages and automatic extraction of data from deep web is cumbersome due to diverse structure of web pages. In the proposed work, a heuristic algorithm for automatic navigation and information extraction from journal’s home page has been devised. The algorithm is applied to many publishers website such as Nature, Elsevier, BMJ, Wiley etc. and the experimental results show that the heuristic technique provides promising results with respect to precision and recall values.[...] Read more.
Data clustering is a basic technique to show the structure of a data set. K-means clustering is a widely acceptable method of data clustering, which follow a partitioned approach for dividing the given data set into non-overlapping groups. Unfortunately, it has the pitfall of randomly choosing the initial cluster centers. Due to its gradient nature, this algorithm is highly sensitive to the initial seed value. In this paper, we propose a kernel density-based method to compute an initial seed value for the k-means algorithm. The idea is to select an initial point from the denser region because they truly reflect the property of the overall data set. Subsequently, we are avoiding the selection of outliers as an initial seed value. We have verified the proposed method on real data sets with the help of different internal and external validity measures. The experimental analysis illustrates that the proposed method has better performance over the k-means, k-means++ algorithm, and other recent initialization methods.[...] Read more.
Data Warehouse is the cornerstone for organizations that base their strategic decisions on the large scale processing of numerical data. The success of the organization depends on these decisions and hence it becomes extremely important to have a quality data warehouse. Conceptual models have been widely recognized as a key determinant of data warehouse quality during the early stages of design. Recently, metrics have been proposed by authors based on hierarchies to quantify the complexity and inturn quality of the conceptual models of data warehouse. They have formally corroborated the measures against Briand’s property based framework to ensure their validity. However, Briand’s set of properties for software measures are a set of necessary but not sufficient measure axioms. They are advantageous to refute software metrics but not to validate them. Thus, we focus on the theoretical validation of the data warehouse conceptual model metrics using the Distance framework whose sufficiency is ensured by the measurement theory. The results indicate that the metrics are valid measures of the complexity of data warehouse conceptual models. Besides, validation by Distance framework assures that the metrics are in the ratio scale which further aids in data analysis.[...] Read more.
The paper is dedicated to the problem of efficiency increasing in case of applying multilayer perceptron in context of parameters estimation for technical systems. It is shown that the increase of efficiency is possible by adaptation of structure of the multilayer perceptron to the problem specification set. It is revealed that the structure adaptation lies in the determination the following parameters:
1. The number of hidden neuron layers;
2. The number of neurons within each layer.
In terms of the paper, we introduce mathematical apparatus that allows conducting the structure adaptation for minimization of the relative error of the neuro-network model generalization. A numerical experiment to demonstrate efficiency of the mathematical apparatus was developed and described in terms of the article. Further research in this sphere lies in the development of a method for calculation of optimum relationship between the number of the hidden neuron layers and the number of hidden neurons within each layer.
Most of the biosensing applications involving analysis and detection of a particular specimen demands fast, easy to use, less expensive, highly reliable and sensitive method for the recognition of biomolecules. The reason behind this increasing demand is that most of the available laboratory equipment require large space, are highly expensive and have other preconditions. Most of the viscometers available for measuring the rheological properties of blood require cleaning after each use which can be challenging due to the capillary geometry. The substitute to this is microcantilever that has emerged as an ideal candidate for biosensing applications. Microcantilever is capable of being used in air, vacuum or liquid medium. This paper consists of seven sections in which working principle of a cantilever, different modes of vibration, their comparative analysis, analytical equations of hydrodynamic equations exerted by the fluid on the cantilever and their impact on the resonant frequency and quality factor, applications of microcantilever in liquid medium specifically in biomedical field are discussed.[...] Read more.