My naive (ha!) Gaussian Naive Bayes classifier is too slow












0












$begingroup$


I am trying to build a film review classifier where I determine if a given review is positive or negative (w/ Python). I'm trying to avoid any other ML libraries so that I can better understand the processes. Here is my approach and the problems that I am facing:




  1. I mine thousands of film reviews as training sets and classify them as positive or negative.

  2. I parse through my training set and for each class, I build an array of unique words.

  3. For each document, I build a vector of TF-IDF values where the vector size is my number of unique words.

  4. I use a Gaussian classifier to determine: $$P(C_i|w)=P(C_i)P(w|C)=P(C_i)*dfrac{1}{sqrt{2pi}sigma_i}e^{-(1/2)(w-mu_i)^Tsigma_i^{-1}(w-mu_i)}$$ where $w$ is the my document in a vector, $C_i$ is a particular class, $mu_i$ is the mean vector and $sigma_i$ is my covariance matrix.


This approach seems to makes sense. My problem is that my algorithm is much too slow. As an example, I have sampled over 1,500 documents and I have determined over 40,000 unique words. This mean that each of my document vectors has 40,000 entries and if I were to build a covariance matrix, it would have dimensions 40,000 by 40,000. Even I were able to generate the entirety of $sigma_i$, but then I would have to compute the matrix product in the exponent, which will take an extraordinarily long time just to classify one document.



I have experimented with a multinomial approach, which is working well. I very curious on how to make this work more efficiently. I realise the matrix multiplication runtime can't be improved, and I was hoping for insight on how others are able to do this.



Some things I have tried:




  • Filtered any stop words (but this still leaves me with tens of thousands of words)

  • Estimated $sigma_i$ by summing over a couple of documents.










share|improve this question







New contributor




Ayumu Kasugano is a new contributor to this site. Take care in asking for clarification, commenting, and answering.
Check out our Code of Conduct.







$endgroup$

















    0












    $begingroup$


    I am trying to build a film review classifier where I determine if a given review is positive or negative (w/ Python). I'm trying to avoid any other ML libraries so that I can better understand the processes. Here is my approach and the problems that I am facing:




    1. I mine thousands of film reviews as training sets and classify them as positive or negative.

    2. I parse through my training set and for each class, I build an array of unique words.

    3. For each document, I build a vector of TF-IDF values where the vector size is my number of unique words.

    4. I use a Gaussian classifier to determine: $$P(C_i|w)=P(C_i)P(w|C)=P(C_i)*dfrac{1}{sqrt{2pi}sigma_i}e^{-(1/2)(w-mu_i)^Tsigma_i^{-1}(w-mu_i)}$$ where $w$ is the my document in a vector, $C_i$ is a particular class, $mu_i$ is the mean vector and $sigma_i$ is my covariance matrix.


    This approach seems to makes sense. My problem is that my algorithm is much too slow. As an example, I have sampled over 1,500 documents and I have determined over 40,000 unique words. This mean that each of my document vectors has 40,000 entries and if I were to build a covariance matrix, it would have dimensions 40,000 by 40,000. Even I were able to generate the entirety of $sigma_i$, but then I would have to compute the matrix product in the exponent, which will take an extraordinarily long time just to classify one document.



    I have experimented with a multinomial approach, which is working well. I very curious on how to make this work more efficiently. I realise the matrix multiplication runtime can't be improved, and I was hoping for insight on how others are able to do this.



    Some things I have tried:




    • Filtered any stop words (but this still leaves me with tens of thousands of words)

    • Estimated $sigma_i$ by summing over a couple of documents.










    share|improve this question







    New contributor




    Ayumu Kasugano is a new contributor to this site. Take care in asking for clarification, commenting, and answering.
    Check out our Code of Conduct.







    $endgroup$















      0












      0








      0





      $begingroup$


      I am trying to build a film review classifier where I determine if a given review is positive or negative (w/ Python). I'm trying to avoid any other ML libraries so that I can better understand the processes. Here is my approach and the problems that I am facing:




      1. I mine thousands of film reviews as training sets and classify them as positive or negative.

      2. I parse through my training set and for each class, I build an array of unique words.

      3. For each document, I build a vector of TF-IDF values where the vector size is my number of unique words.

      4. I use a Gaussian classifier to determine: $$P(C_i|w)=P(C_i)P(w|C)=P(C_i)*dfrac{1}{sqrt{2pi}sigma_i}e^{-(1/2)(w-mu_i)^Tsigma_i^{-1}(w-mu_i)}$$ where $w$ is the my document in a vector, $C_i$ is a particular class, $mu_i$ is the mean vector and $sigma_i$ is my covariance matrix.


      This approach seems to makes sense. My problem is that my algorithm is much too slow. As an example, I have sampled over 1,500 documents and I have determined over 40,000 unique words. This mean that each of my document vectors has 40,000 entries and if I were to build a covariance matrix, it would have dimensions 40,000 by 40,000. Even I were able to generate the entirety of $sigma_i$, but then I would have to compute the matrix product in the exponent, which will take an extraordinarily long time just to classify one document.



      I have experimented with a multinomial approach, which is working well. I very curious on how to make this work more efficiently. I realise the matrix multiplication runtime can't be improved, and I was hoping for insight on how others are able to do this.



      Some things I have tried:




      • Filtered any stop words (but this still leaves me with tens of thousands of words)

      • Estimated $sigma_i$ by summing over a couple of documents.










      share|improve this question







      New contributor




      Ayumu Kasugano is a new contributor to this site. Take care in asking for clarification, commenting, and answering.
      Check out our Code of Conduct.







      $endgroup$




      I am trying to build a film review classifier where I determine if a given review is positive or negative (w/ Python). I'm trying to avoid any other ML libraries so that I can better understand the processes. Here is my approach and the problems that I am facing:




      1. I mine thousands of film reviews as training sets and classify them as positive or negative.

      2. I parse through my training set and for each class, I build an array of unique words.

      3. For each document, I build a vector of TF-IDF values where the vector size is my number of unique words.

      4. I use a Gaussian classifier to determine: $$P(C_i|w)=P(C_i)P(w|C)=P(C_i)*dfrac{1}{sqrt{2pi}sigma_i}e^{-(1/2)(w-mu_i)^Tsigma_i^{-1}(w-mu_i)}$$ where $w$ is the my document in a vector, $C_i$ is a particular class, $mu_i$ is the mean vector and $sigma_i$ is my covariance matrix.


      This approach seems to makes sense. My problem is that my algorithm is much too slow. As an example, I have sampled over 1,500 documents and I have determined over 40,000 unique words. This mean that each of my document vectors has 40,000 entries and if I were to build a covariance matrix, it would have dimensions 40,000 by 40,000. Even I were able to generate the entirety of $sigma_i$, but then I would have to compute the matrix product in the exponent, which will take an extraordinarily long time just to classify one document.



      I have experimented with a multinomial approach, which is working well. I very curious on how to make this work more efficiently. I realise the matrix multiplication runtime can't be improved, and I was hoping for insight on how others are able to do this.



      Some things I have tried:




      • Filtered any stop words (but this still leaves me with tens of thousands of words)

      • Estimated $sigma_i$ by summing over a couple of documents.







      machine-learning python naive-bayes-classifier tfidf






      share|improve this question







      New contributor




      Ayumu Kasugano is a new contributor to this site. Take care in asking for clarification, commenting, and answering.
      Check out our Code of Conduct.











      share|improve this question







      New contributor




      Ayumu Kasugano is a new contributor to this site. Take care in asking for clarification, commenting, and answering.
      Check out our Code of Conduct.









      share|improve this question




      share|improve this question






      New contributor




      Ayumu Kasugano is a new contributor to this site. Take care in asking for clarification, commenting, and answering.
      Check out our Code of Conduct.









      asked 13 mins ago









      Ayumu KasuganoAyumu Kasugano

      101




      101




      New contributor




      Ayumu Kasugano is a new contributor to this site. Take care in asking for clarification, commenting, and answering.
      Check out our Code of Conduct.





      New contributor





      Ayumu Kasugano is a new contributor to this site. Take care in asking for clarification, commenting, and answering.
      Check out our Code of Conduct.






      Ayumu Kasugano is a new contributor to this site. Take care in asking for clarification, commenting, and answering.
      Check out our Code of Conduct.






















          0






          active

          oldest

          votes











          Your Answer





          StackExchange.ifUsing("editor", function () {
          return StackExchange.using("mathjaxEditing", function () {
          StackExchange.MarkdownEditor.creationCallbacks.add(function (editor, postfix) {
          StackExchange.mathjaxEditing.prepareWmdForMathJax(editor, postfix, [["$", "$"], ["\\(","\\)"]]);
          });
          });
          }, "mathjax-editing");

          StackExchange.ready(function() {
          var channelOptions = {
          tags: "".split(" "),
          id: "557"
          };
          initTagRenderer("".split(" "), "".split(" "), channelOptions);

          StackExchange.using("externalEditor", function() {
          // Have to fire editor after snippets, if snippets enabled
          if (StackExchange.settings.snippets.snippetsEnabled) {
          StackExchange.using("snippets", function() {
          createEditor();
          });
          }
          else {
          createEditor();
          }
          });

          function createEditor() {
          StackExchange.prepareEditor({
          heartbeatType: 'answer',
          autoActivateHeartbeat: false,
          convertImagesToLinks: false,
          noModals: true,
          showLowRepImageUploadWarning: true,
          reputationToPostImages: null,
          bindNavPrevention: true,
          postfix: "",
          imageUploader: {
          brandingHtml: "Powered by u003ca class="icon-imgur-white" href="https://imgur.com/"u003eu003c/au003e",
          contentPolicyHtml: "User contributions licensed under u003ca href="https://creativecommons.org/licenses/by-sa/3.0/"u003ecc by-sa 3.0 with attribution requiredu003c/au003e u003ca href="https://stackoverflow.com/legal/content-policy"u003e(content policy)u003c/au003e",
          allowUrls: true
          },
          onDemand: true,
          discardSelector: ".discard-answer"
          ,immediatelyShowMarkdownHelp:true
          });


          }
          });






          Ayumu Kasugano is a new contributor. Be nice, and check out our Code of Conduct.










          draft saved

          draft discarded


















          StackExchange.ready(
          function () {
          StackExchange.openid.initPostLogin('.new-post-login', 'https%3a%2f%2fdatascience.stackexchange.com%2fquestions%2f45067%2fmy-naive-ha-gaussian-naive-bayes-classifier-is-too-slow%23new-answer', 'question_page');
          }
          );

          Post as a guest















          Required, but never shown

























          0






          active

          oldest

          votes








          0






          active

          oldest

          votes









          active

          oldest

          votes






          active

          oldest

          votes








          Ayumu Kasugano is a new contributor. Be nice, and check out our Code of Conduct.










          draft saved

          draft discarded


















          Ayumu Kasugano is a new contributor. Be nice, and check out our Code of Conduct.













          Ayumu Kasugano is a new contributor. Be nice, and check out our Code of Conduct.












          Ayumu Kasugano is a new contributor. Be nice, and check out our Code of Conduct.
















          Thanks for contributing an answer to Data Science Stack Exchange!


          • Please be sure to answer the question. Provide details and share your research!

          But avoid



          • Asking for help, clarification, or responding to other answers.

          • Making statements based on opinion; back them up with references or personal experience.


          Use MathJax to format equations. MathJax reference.


          To learn more, see our tips on writing great answers.




          draft saved


          draft discarded














          StackExchange.ready(
          function () {
          StackExchange.openid.initPostLogin('.new-post-login', 'https%3a%2f%2fdatascience.stackexchange.com%2fquestions%2f45067%2fmy-naive-ha-gaussian-naive-bayes-classifier-is-too-slow%23new-answer', 'question_page');
          }
          );

          Post as a guest















          Required, but never shown





















































          Required, but never shown














          Required, but never shown












          Required, but never shown







          Required, but never shown

































          Required, but never shown














          Required, but never shown












          Required, but never shown







          Required, but never shown







          Popular posts from this blog

          How to label and detect the document text images

          Vallis Paradisi

          Tabula Rosettana