I am a computer science student with interest in problem solving and automating things using python, machine learning and other technologies.
Reach me at 📧:
- Github: https://github.com/AryanRastogi7767
- Gmail: [email protected]
- Linkedin: https://www.linkedin.com/in/aryan-rastogi-b38ba1190/
This markdown file is to give you a detailed and easy-to-understand explaination of WHAT the famous Naive Bayes algorithm is and WHY it is used widely in the field of machine learning. Read the file to the end for better understanding.
Go through the Jupyter Notebook by me for more information on the implementation of the algorithm.
For and doubts and queries, contact me on the specified platforms. 😄
- Naive Bayes is a probability based machine learning algorithm that is used for large volumes of data.
- It gives very good results when it comes to NLP tasks such as sentimental analysis.
- It is a fast and uncomplicated classification algorithm.
- It is based on the Bayes Theorem with an assumption of independence among predictors.
- Real Time Prediction: As Naive Bayes has low Time Complexity, it can be easily deployed and used for real time applications.
- Multi-Class Prediction: The algorithm can be used in multi-class prediction scenario as it can predict the posterior probability of mulltiple classes of the target variable.
- Natural Language Processing: Because of their good performance in multi-class settings, Naive Bayes tends to be very useful in NLP applications like:
- Text Classification
- Spam Filtering
- Sentiment Analysis, etc.
- Recommendation Systems: It helps in recommendation system because of probabilistic prediction ability and low time complexity.
- Gaussian Naive Bayes:
- Used in Classification
- Assumes that features follow normal distribution
- Multinomial Naive Bayes:
- Suitable for classification with discrete features, e.g., word count for text classification.
- Bernoulli Naive Bayes:
- Used for data that is distributed according to multivariate Bernoulli distributions.
- Features should be independent booleans.
- Naive Bayes (NB) is ‘naive’ because it makes the assumption that features of a measurement are independent of each other.
- Highly extensible and fast algorithm , Time Complexity:O(nK), where n is the number of features and K is the number of class lables.
- Performs well in both binary and multi-class settings.
- When assumption of independence holds, a Naive Bayes classifier performs better compare to other models like logistic regression and you need less training data.
- Performs well whin data is categorical or normally distributed.
- Can be easily trained on small datasets and can be used for large ones as well.
- The main disadvantage of the NB is considering all the variables independent that contributes to the probability.
- If categorical variable has a category (in test data set), which was not observed in training data set, then model will assign a 0 (zero) probability and will be unable to make a prediction. This is often known as “Zero Frequency”. To solve this, we can use the smoothing techniques like Laplace or Additive smoothing.