Abstract
Machine-learning techniques are a handy tool for deriving insights from data extracted from the web. Because of the structure of web data extracted by web crawlers there is need for preprocessing the data to extract features that can be used to train a machine learning classifier. The number of available features that can be linked to a website is huge. Narrowing down to a minimum number of features required to drive a classifier has huge benefits. This paper presents a workflow that uses a set of metrics that can be used to reduce the numbers of features for training a support vector machine (SVM) for classifying webpages as fraudulent or not. The paper reports that a three quarter reduction in feature set size only incurs a 5% reduction in classification accuracy which has huge computational benefits.
Original language | English |
---|---|
Title of host publication | IEEE 4th International Conference on Soft Computing & Machine Intelligence (ISCMI), 2017 |
Publisher | IEEE |
Pages | 85-88 |
Number of pages | 4 |
ISBN (Electronic) | 9781538613146 |
DOIs | |
Publication status | Published - 5 Feb 2018 |
Keywords
- feature selection
- machine learning
- internet
- support vector machine
- feature extraction
- information
- web crawling
ASJC Scopus subject areas
- Computational Mathematics
- Artificial Intelligence
- Computer Vision and Pattern Recognition
- Modelling and Simulation