Anomalies in data become ubiquitous and often unavoidable. Despite it might be caused by various errors during the data collection or transportation, some anomalies potentially give important information such as indic...
详细信息
ISBN:
(数字)9781728158624
ISBN:
(纸本)9781728158631
Anomalies in data become ubiquitous and often unavoidable. Despite it might be caused by various errors during the data collection or transportation, some anomalies potentially give important information such as indication of a new underlying process. The aim of anomaly detection task is to determine all such instances in the input dataset in a data-driven fashion. In the past decade, the proliferation of anomaly data in many domains have raised research interest resulted in a plethora of anomaly detection methods A vast number of previous study reports suggested that there is no single model will achieve the best performance for every dataset. This paper presents empiric results on the effect of several kernel functions to performance of One-Class Support Vector Machine (OC-SVM) as an anomaly detector model. The proposed model is tested using financial transactions of microfinance service dataset from an Indonesian Bank. The empiric results showed that OC-SVM model with Sigmoid and RBF kernels achieve the best statistically significant value of training, validation, and testing accuracies than the OC-SVM model with no-kernel, linear kernel and polynomial (degree 3,4,5,6) kernels.
Airport runway designs must consider wind climatology to reduce the potential for crosswind and tailwind events that can cause accidents and aircraft delays. The large potential of crosswind and tailwind on a runway a...
Airport runway designs must consider wind climatology to reduce the potential for crosswind and tailwind events that can cause accidents and aircraft delays. The large potential of crosswind and tailwind on a runway at the airport is very detrimental to aircraft passengers and airlines. Therefore, it is necessary to know the potential of crosswind and tailwind on the runway of Soekarno Hatta Airport and find the right parameters for the approximate direction and speed of the wind on the airport runway. Analysis was carried out using the direction and wind speed data from the Soekarno Hatta meteorological station in 2007-2017 to determine the potential for crosswind and tailwind on the runway of Soekarno Hatta airport and run a Weather Research and Forecasting (WRF) weather model to determine the parameterization of the right weather model for wind forecast at Soekarno Hatta airport. The highest maximum crosswind component in the 2007-2017 period occurred in August 30.53 knots, while the highest maximum tailwind of 25 knots occurred in January for a 250-degree runway and 24 knots occurred in August for a 070-degree runway. The potential for crosswind and tailwind is different every month. The scheme produces forecasts of wind direction and speed with a strong level of correlation with observations of wind direction and speed from the Soekarno Hatta meteorological station, the correlation value is 0.61. The most significant social impact of plane delays is time loss for passengers. While the economic impact on one of the crosswind events at Soekarno Hatta airport was an economic loss for one airline of US $ 43,392.
Nowadays, every company knows that when making a decision that has a potential in affecting their assets, an accurately processed report is necessary in order to support the reasoning behind their decision. Generating...
Nowadays, every company knows that when making a decision that has a potential in affecting their assets, an accurately processed report is necessary in order to support the reasoning behind their decision. Generating a report for stakeholders quickly and accurately is highly required in assisting them making a data-driven decision. By developing a data warehouse, it is possible for a company to do a data-driven decision making to appeal to their customer segments. This paper proposes a data warehouse model design to analyse the sales data contained in the database. The method that was implemented for this particular data warehouse development is the nine-step methodology designed by Kimball. The results are then presented in pdf form and an interactive dashboard.
Wind power has been experiencing a quick improvement. Without a doubt, wind is a variable asset that is hard to forecast. For instance, traditionally time series, extra holds are distributed to deal with this uncertai...
详细信息
This study proposes a multiparty mediated quantum secret sharing (MQSS) protocol that allows n restricted quantum users to share a secret via the assistance of a dishonest third-party (TP) with full quantum capabiliti...
详细信息
The results of observations showed that the book about Tais produced by Secretariat de Estado da Arte e Cultura (SEAC) has not been able to meet the needs of users because the information is less interesting and inter...
详细信息
Today big data has become the basis of discussion for the organizations. The big task associated with big data stream is coping with its various challenges and performing the appropriate testing for the optimal analys...
Today big data has become the basis of discussion for the organizations. The big task associated with big data stream is coping with its various challenges and performing the appropriate testing for the optimal analysis of the data which may benefit the processing of various activities, especially from a business perspective. Big data term follows the massive volume of data, (might be in units of petabytes or exabytes) exceeding the processing and analytical capacity of the conventional systems and thereby raising the need for analyzing and testing the big data before applications can be put into use. Testing such huge data coming from the various number of sources like the internet, smartphones, audios, videos, media, etc. is a challenge itself. The most favourable solution to test big data follows the automated/programmed approach. This paper outlines the big data characteristics, and various challenges associated with it followed by the approach, strategy, and proposed framework for testing big data applications.
Second order graph Laplacian regularization has the limitation that the solution remains biased towards a constant which restricts its extrapolation capability. The lack of extrapolation results in poor generalization...
Second order graph Laplacian regularization has the limitation that the solution remains biased towards a constant which restricts its extrapolation capability. The lack of extrapolation results in poor generalization. An additional penalty factor is needed on the function to avoid its over-fitting on seen unlabeled training instances. The third order derivative based technique identifies the sharp variations in the function and accurately penalizes them to avoid overfitting. The resultant function leads to a more accurate and generic model that exploits the twist and curvature variations on the manifold. Extensive experiments on synthetic and real-world data set clearly shows that the additional regularization increases accuracy and generic nature of model.
COVID-19 has received tremendous attention from scholars worldwide and even being labelled as a black swan event that has disrupted every aspect of human life. Within a short time span of the pandemic, a large volume ...
COVID-19 has received tremendous attention from scholars worldwide and even being labelled as a black swan event that has disrupted every aspect of human life. Within a short time span of the pandemic, a large volume of research pertaining to COVID-19 has been published in diverse research fields. This paper adopts a bibliometric analysis to systematically evaluate the research development in the application of optimization and simulation methods to address COVID-19 physical distancing policy (OSPhyD) using Bibliometrix R package. A textual query on Scopus database using the combination of four classes of keywords; covid-19, optimization, simulation, and physical distance has returned a total of 299 original research articles and reviews published in English. Appropriate visualizations were generated to describe the collaborations between different authors, countries, and institutions, whilst co-word analysis that uses text mining technique has produced a conceptual cluster via co-occurrence network map to underscore the emerging themes in the current research interest. The main findings pinpoint that: 1) OSPhyD as a scientific research field is an emerging multidisciplinary research topic that is growing progressively and steadily in the fields of medicine, engineering, social sciences, mathematics and decision science, 2) The field has attracted the attention of scholars from all over the world particularly from United States of America, United Kingdom, European countries including authors from Asia, and 3) Three dominant themes or research front emerged from the publications including COVID-19 mainstream, medical education and undergraduate.
Convolutional neural networks (ConvNet or CNN) are deep learning algorithms that can process input images, assign meaning to various aspects or objects in the image (biases and learnable weight) and recognize one imag...
Convolutional neural networks (ConvNet or CNN) are deep learning algorithms that can process input images, assign meaning to various aspects or objects in the image (biases and learnable weight) and recognize one image from another. The bigger kernel size will take more time to process the *** present a novelty way to use a 4D rank tensor to improve a convolutional process. At the early stage of the Convolve4D development, the edge detection with 3×3 kernel and The Laplacian of Gaussian (LoG) with 5×5 kernel size was used to demonstrate the convolutional process improvement. The Convolve4D needs more elaboration to be used into a CNN algorithm. The advantage of convolve4D is only need 9 loops to calculate 81 outputs, whereas convolve2D need 9 × 9 × 3 × 1 × 7 × 7 = 11.907 loops. The result is 18.5% shorter when using a 5×5 kernel; it reduces from 0.54 seconds to 0.44 seconds for the edge detection convolution process.
暂无评论