Open Access Open Access  Restricted Access Subscription Access
Open Access Open Access Open Access  Restricted Access Restricted Access Subscription Access

An Improved Expectation Maximization based Semi-Supervised Text Classification using Naïve Bayes and Support Vector Machine


Affiliations
1 Department of Computer Engineering, Chandubhai S Patel Institute of Technology, Changa, Petlad, India
2 Department of Information and Technology, Chandubhai S Patel Institute of Technology, Changa, Petlad, India
3 U & PU Patel Department of Computer Engineering, Chandubhai S Patel Institute of Technology, Changa, Petlad, India
     

   Subscribe/Renew Journal


With the development of Internet and the emergence of a large number of text resources, the automatic text classification has become a research hotspot. As number of training documents increases, accuracy of Text Classification increases. Traditional classifiers (Supervised learning) use only labeled data for training. Labeled instances are often difficult, expensive, or time consuming to obtain. Meanwhile unlabeled data may be relatively easy to collect. Semi-Supervised Learning makes use of both labeled and unlabeled data. Several researchers have given algorithms for Text Classification using Semi-Supervised Learning. But still improving accuracy of Text Classification using Semi-Supervised Learning is a challenge. In the iterative process in the standard Expectation Maximization (EM) based semi-supervised learning, some unlabeled samples are misclassified by the current classifier because the initial labeled samples are not enough. To overcome this limitation, an EM based Semi-Supervised Learning algorithm using Naïve Bayesian and Support vector machine is proposed in this paper to improve accuracy of text classification using semi-supervised learning.

Keywords

Expectation Maximization (EM), Naïve Bayes (NB), Support Vector Machine (SVM), Semi-Supervised Machine (SSL).
User
Subscription Login to verify subscription
Notifications
Font Size

Abstract Views: 205

PDF Views: 3




  • An Improved Expectation Maximization based Semi-Supervised Text Classification using Naïve Bayes and Support Vector Machine

Abstract Views: 205  |  PDF Views: 3

Authors

Purvi Rekh
Department of Computer Engineering, Chandubhai S Patel Institute of Technology, Changa, Petlad, India
Amit Thakkar
Department of Information and Technology, Chandubhai S Patel Institute of Technology, Changa, Petlad, India
Amit Ganatra
U & PU Patel Department of Computer Engineering, Chandubhai S Patel Institute of Technology, Changa, Petlad, India

Abstract


With the development of Internet and the emergence of a large number of text resources, the automatic text classification has become a research hotspot. As number of training documents increases, accuracy of Text Classification increases. Traditional classifiers (Supervised learning) use only labeled data for training. Labeled instances are often difficult, expensive, or time consuming to obtain. Meanwhile unlabeled data may be relatively easy to collect. Semi-Supervised Learning makes use of both labeled and unlabeled data. Several researchers have given algorithms for Text Classification using Semi-Supervised Learning. But still improving accuracy of Text Classification using Semi-Supervised Learning is a challenge. In the iterative process in the standard Expectation Maximization (EM) based semi-supervised learning, some unlabeled samples are misclassified by the current classifier because the initial labeled samples are not enough. To overcome this limitation, an EM based Semi-Supervised Learning algorithm using Naïve Bayesian and Support vector machine is proposed in this paper to improve accuracy of text classification using semi-supervised learning.

Keywords


Expectation Maximization (EM), Naïve Bayes (NB), Support Vector Machine (SVM), Semi-Supervised Machine (SSL).