Diese Präsentation wurde erfolgreich gemeldet.
Wir verwenden Ihre LinkedIn Profilangaben und Informationen zu Ihren Aktivitäten, um Anzeigen zu personalisieren und Ihnen relevantere Inhalte anzuzeigen. Sie können Ihre Anzeigeneinstellungen jederzeit ändern.

Tweets Classification using Naive Bayes and SVM

26.683 Aufrufe

Veröffentlicht am

Tweets Classification for classifying test tweet into several Wikipedia categories.

Veröffentlicht in: Ingenieurwesen, Technologie, Bildung
  • Loggen Sie sich ein, um Kommentare anzuzeigen.

Tweets Classification using Naive Bayes and SVM

  1. 1. Tweet Classification Mentor: Romil Bansal GROUP NO-37 Manish Jindal(201305578) Trilok Sharma(201206527) Yash Shah (201101127) Guided by : Dr. Vasudeva Varma
  2. 2. Problem Statement : To automatically classify Tweets from Twitter into various genres based on predefined Wikipedia Categories. Motivation: o Twitter is a major social networking service with over 200 million tweets made every day. o Twitter provides a list of Trending Topics in real time, but it is often hard to understand what these trending topics are about. o It is important and necessary to classify these topics into general categories with high accuracy for better information retrieval.
  3. 3. Data Dataset : o Input Data is the static / real-time data consisting of the user tweets. o Training dataset : Fetched from twitter with twitter4j api. Final Deliverable: o It will return list of all categories to which the input tweet belongs. o It will also give the accuracy of the algorithm used for classifying tweets.
  4. 4. Categories We took following categories into consideration for classifying twitter data. 1)Business 5)Law 9)Politics 2)Education 6)Lifestyle 10)Sports 3)Entertainment 7)Nature 11)Technology 4)Health 8)Places
  5. 5. Concepts used for better performance  Outliers removal  To remove low frequent and high frequent words using Bag of words approach .  Stop words removal  To remove most common words, such as the, is, at, which, and on.  Keyword Stemming  To reduce inflected words to their stem, base or root form using porter stemming  Cleaning crawl data
  6. 6. Other Concepts used ..  Spelling Correction  To correct spellings using Edit distance method.  Named Entity Recognition:  For ranking result category and finding most appropriate.  Synonym form  If feature(word) of test query not found as one of dimension in feature space than replace that word with its synonym. Done using WordNet.
  7. 7. Tweets Classification Algorithms We used 3 algorithms for classification 1) Naïve based 2) SVM based Supervised 3) Rule based
  8. 8. Crawl tweeter data Tweets Cleaning, Stop word removal Create Index file Of feature vector Extract Features (Unique wordlist) Create feature vector for each tweet Edit Distance, WordNet (synonyms) Test Query/ Tweet Create Index file Of feature vectors Create / Apply Model files Output Category Training Testing Remove Outliers Tweets Cleaning, Stop word removal Create feature vector for test tweet Apply Named Entity Recognition Rank result category
  9. 9. Main idea for Supervised Learning  Assumption: training set consists of instances of different classes described cj as conjunctions of attributes values  Task: Classify a new instance d based on a tuple of attribute values into one of the classes cj C  Key idea: assign the most probable class using supervised learning algorithm.
  10. 10. Method 1 : Bayes Classifier  Bayes rule states :  We used “WEKA” library for machine learning in Bayes Classifier for our project. Normalization Constant Likelihood Prior
  11. 11. Method 2 : SVM Classifier (Support Vector Machine)  Given a new point x, we can score its projection onto the hyperplane normal:  I.e., compute score: wTx + b = Σαiyixi Tx + b  Decide class based on whether < or > 0  Can set confidence threshold t. 11 -1 0 1 Score > t: yes Score < -t: no Else: don’t know
  12. 12. 12 Multi-class SVM
  13. 13. 13 Multi-class SVM Approaches  1-against-all Each of the SVMs separates a single class from all remaining classes (Cortes and Vapnik, 1995)  1-against-1 Pair-wise. k(k-1)/2, k Y SVMs are trained. Each SVM separates a pair of classes (Fridman, 1996)
  14. 14. Advantages of SVM  High dimensional input space  Few irrelevant features (dense concept)  Sparse document vectors (sparse instances)  Text categorization problems are linearly separable  For linearly inseparable data we can use kernels to map data into high dimensional space, so that it becomes linearly separable with hyperplane.
  15. 15. Method 3 : Rule Based  We defined set of rule to classify a tweet based on term frequency.  a. Extract the features of a tweet.  b. Count term frequency of each feature , the feature having maximum term frequency from all categories mentioned above will be our first classification.  c. As it cannot be right all time so now we maintain count of categories in which tweet falls , category which is near to tweet will be our next classification.
  16. 16. Example-  Tweet=sachin is a good player, who eats apple and banana which is good for health.  Feature- sachin,player,eats,apple,health,banana  Stop word-is,a,good,he,was,for,which,and,who  Classification- Feature-category term-frequency sachin-sports 2000 player-sports 900 eating-health 500 apple-technology 1000 health-health 800 banana-health 700
  17. 17.  Max term-frequency - sachin  So our category is - sports  2nd approximation -  Max feature is laying in health i.e. 3 times ,  So our second approximation would be health.  If both of these are in same category then we have only one category. i.e. if here max feature would be laying in sports than we have only one result that is sports.
  18. 18. Cross-validation (Accuracy)  Steps for k-fold cross-validation : Step 1: split data into k subsets of equal size Step 2 : use each subset in turn for testing, the remainder for training  Often the subsets are stratified before the cross- validation is performed  The error estimates are averaged to yield an overall error estimate
  19. 19. Accuracy Results ( 10 folds) Accuracy of Algorithm in % Categories Algo. SVM Naïve Rule Business 86.6 81.44 98.30 Education 85.71 76.07 81.8 Entertainment 86.8 79.1 87.49 Health 95.67 84.62 90.93 Law 81.17 73.38 75.25 Lifestyle 93.27 89.71 82.42 Nature 87.0 78.64 84.24 Places 81.01 75.35 80.73 Politics 81.91 81.88 76.31 Sports 87.11 83.57 81.87 Technology 83.64 82.44 77.05
  20. 20. Unique features  Worked on latest Crawled tweeter data using tweeter4j api  Worked on Eleven different Categories.  Applied three different method of supervised learning to classify in different categories.  Achieved high performance speed with accuracy in range of 85 to 95 %  Done Tweets Cleaning , Stemming , Stop Word removal.  Used Edit distance for spelling correction.  Used Named entity recognition for ranking.  Used WordNet for Query Expansion and Synonyms finding.  Validated using CrossFold (10 fold) validation.
  21. 21. Snapshot
  22. 22. Result
  23. 23. Accuracy
  24. 24. Thank You!