Fake_News_Detection_Machine_learning_project.rar
8.3 MB
Fake news Detection Machine Learning Project with 92%Accuracy
it contain compressed file in which "jupyter notebook file and dataset"✅
it contain compressed file in which "jupyter notebook file and dataset"✅
❤2👍1
dice_roll.py
445 B
🎲Dice_roll_Simulator_Gui with python in 2 minute 😊
Dimensionality reduction techniques
Singular Value Decomposition (SVD)
Principal Component Analysis (PCA)
Linear Discriminant Analysis (LDA)
T-distributed Stochastic Neighbor Embedding (t-SNE)
Autoencoders
Fourier and Wavelet Transforms
Singular Value Decomposition (SVD)
Principal Component Analysis (PCA)
Linear Discriminant Analysis (LDA)
T-distributed Stochastic Neighbor Embedding (t-SNE)
Autoencoders
Fourier and Wavelet Transforms
What is the curse of dimensionality? Why do we care about it?
Data in only one dimension is relatively tightly packed. Adding a dimension stretches the points across that dimension, pushing them further apart. Additional dimensions spread the data even further making high dimensional data extremely sparse. We care about it, because it is difficult to use machine learning in sparse spaces.
Data in only one dimension is relatively tightly packed. Adding a dimension stretches the points across that dimension, pushing them further apart. Additional dimensions spread the data even further making high dimensional data extremely sparse. We care about it, because it is difficult to use machine learning in sparse spaces.
👍1
K-means vs DBScan ML Algorithm
DBScan is more robust to noise.
DBScan is better when the amount of clusters is difficult to guess.
K-means has a lower complexity, i.e. it will be much faster, especially with a larger amount of points.
DBScan is more robust to noise.
DBScan is better when the amount of clusters is difficult to guess.
K-means has a lower complexity, i.e. it will be much faster, especially with a larger amount of points.
❤1
Data Science & Machine Learning
Fake_News_Detection_Machine_learning_project.rar
Start working on any project if you are a beginner and want to grow your career as a data scientist
You will learn much more as you practice and work on projects from yourself
You can find dataset in this channel or go to kaggle to find any random dataset and just work on it
Learning concepts is fine but most of the learnings come from projects
I know that might feel boring at first time but as you move forward, it become interesting
You will learn much more as you practice and work on projects from yourself
You can find dataset in this channel or go to kaggle to find any random dataset and just work on it
Learning concepts is fine but most of the learnings come from projects
I know that might feel boring at first time but as you move forward, it become interesting
❤1
👉The Ultimate Guide to the Pandas Library for Data Science in Python
👇👇
https://www.freecodecamp.org/news/the-ultimate-guide-to-the-pandas-library-for-data-science-in-python/amp/
A Visual Intro to NumPy and Data Representation
.
Link : 👇👇
https://jalammar.github.io/visual-numpy/
Matplotlib Cheatsheet 👇👇
https://github.com/rougier/matplotlib-cheatsheet
SQL Cheatsheet 👇👇
https://websitesetup.org/sql-cheat-sheet/
👇👇
https://www.freecodecamp.org/news/the-ultimate-guide-to-the-pandas-library-for-data-science-in-python/amp/
A Visual Intro to NumPy and Data Representation
.
Link : 👇👇
https://jalammar.github.io/visual-numpy/
Matplotlib Cheatsheet 👇👇
https://github.com/rougier/matplotlib-cheatsheet
SQL Cheatsheet 👇👇
https://websitesetup.org/sql-cheat-sheet/
Seeing Theory : A visual introduction to probability and statistics
Link :👇👇
https://seeing-theory.brown.edu/
“The Projects You Should Do to Get a Data Science Job” by Ken Jee
👇👇
https://link.medium.com/Q2DnxSGRO6
Link :👇👇
https://seeing-theory.brown.edu/
“The Projects You Should Do to Get a Data Science Job” by Ken Jee
👇👇
https://link.medium.com/Q2DnxSGRO6
Type-2 error is?
Anonymous Quiz
18%
True Positive
32%
True Negative
15%
False Positive
35%
False Negative
Precision is one indicator of a machine learning model's performance – the quality of a positive prediction made by the model. Its formula would be?
Anonymous Quiz
43%
True Positive divided by actual yes
10%
True Positive divided by actual no
43%
True Positive divided by predicted yes
4%
True Positive divided by predicted no
Scatter plot is used to?
Anonymous Quiz
27%
Find Correlation between two variables
12%
Detect Outliers
14%
Compare large numbers of data points without regard to time
47%
All of the above
👍2
👉A handy notebook on handling missing values
Link : 👇👇
https://www.kaggle.com/parulpandey/a-guide-to-handling-missing-values-in-python
A list of NLP Tutorials
Link : 👇👇
https://github.com/lyeoni/nlp-tutorial
“An Implementation and Explanation of the Random Forest in Python” by Will Koehrsen 👇👇
https://link.medium.com/GCWFv81v95
“How to analyse 100s of GBs of data on your laptop with Python” by Jovan Veljanoski 👇👇
https://link.medium.com/V8xS82Cax6
Link : 👇👇
https://www.kaggle.com/parulpandey/a-guide-to-handling-missing-values-in-python
A list of NLP Tutorials
Link : 👇👇
https://github.com/lyeoni/nlp-tutorial
“An Implementation and Explanation of the Random Forest in Python” by Will Koehrsen 👇👇
https://link.medium.com/GCWFv81v95
“How to analyse 100s of GBs of data on your laptop with Python” by Jovan Veljanoski 👇👇
https://link.medium.com/V8xS82Cax6
👍1
Skills Required For A Data Analyst in 2021
Basic Excel-
https://www.youtube.com/playlist?list=PLmQAMKHKeLZ_ADx6nJcoTM5t2S1bmsMdm
Advanced Excel-
https://www.youtube.com/playlist?list=PLmQAMKHKeLZ_e9xmZNPACsLdgie3Tkaxf
SQL-
https://www.youtube.com/watch?v=5JCyiutyu_o&list=PLmQAMKHKeLZ-kD9VN0prfKCByr9pa4jw6
SQL- (Khan Academy)-
https://www.khanacademy.org/computing/computer-programming/sql
Python Programming Language-
https://www.youtube.com/watch?v=bPrmA1SEN2k&list=PLZoTAELRMXVNUL99R4bDlVYsncUNvwUBB
Stats Lectures-
https://www.youtube.com/watch?v=zRUliXuwJCQ&list=PLZoTAELRMXVMhVyr3Ri9IQ-t5QPBtxzJO
Stats Lectures(Khans Academy)-
https://www.khanacademy.org/math/statistics-probability
Python EDA-
https://www.youtube.com/playlist?list=PLZoTAELRMXVPQyArDHyQVjQxjj_YmEuO9
Python Feature Engineering-
https://www.youtube.com/playlist?list=PLZoTAELRMXVPwYGE2PXD3x0bfKnR0cJjN
Tableau-
https://www.tableau.com/academic/student-Iron-Viz
Basic Excel-
https://www.youtube.com/playlist?list=PLmQAMKHKeLZ_ADx6nJcoTM5t2S1bmsMdm
Advanced Excel-
https://www.youtube.com/playlist?list=PLmQAMKHKeLZ_e9xmZNPACsLdgie3Tkaxf
SQL-
https://www.youtube.com/watch?v=5JCyiutyu_o&list=PLmQAMKHKeLZ-kD9VN0prfKCByr9pa4jw6
SQL- (Khan Academy)-
https://www.khanacademy.org/computing/computer-programming/sql
Python Programming Language-
https://www.youtube.com/watch?v=bPrmA1SEN2k&list=PLZoTAELRMXVNUL99R4bDlVYsncUNvwUBB
Stats Lectures-
https://www.youtube.com/watch?v=zRUliXuwJCQ&list=PLZoTAELRMXVMhVyr3Ri9IQ-t5QPBtxzJO
Stats Lectures(Khans Academy)-
https://www.khanacademy.org/math/statistics-probability
Python EDA-
https://www.youtube.com/playlist?list=PLZoTAELRMXVPQyArDHyQVjQxjj_YmEuO9
Python Feature Engineering-
https://www.youtube.com/playlist?list=PLZoTAELRMXVPwYGE2PXD3x0bfKnR0cJjN
Tableau-
https://www.tableau.com/academic/student-Iron-Viz
👍2
Data Science & Machine Learning
What are precision, recall, and F1-score? Precision and recall are classification evaluation metrics: P = TP / (TP + FP) and R = TP / (TP + FN). Where TP is true positives, FP is false positives and FN is false negatives In both cases the score of 1 is…
Here is the explanation for the quiz
The F-beta score is the weighted harmonic mean of precision and recall, reaching its optimal value at 1 and its worst value at 0. The highest possible value of an F-score is 1.0, indicating perfect precision and recall, and the lowest possible value is 0, if either the precision or the recall is zero. The F-score is commonly used for evaluating information retrieval systems such as search engines, and also for many kinds of machine learning models, in particular in natural language processing.
The F-beta score is the weighted harmonic mean of precision and recall, reaching its optimal value at 1 and its worst value at 0. The highest possible value of an F-score is 1.0, indicating perfect precision and recall, and the lowest possible value is 0, if either the precision or the recall is zero. The F-score is commonly used for evaluating information retrieval systems such as search engines, and also for many kinds of machine learning models, in particular in natural language processing.
What is the full form of LSTM?
Hint- LSTM algorithm is used for processing and making predictions based on time series data
Hint- LSTM algorithm is used for processing and making predictions based on time series data
Anonymous Quiz
7%
Long story total memory
72%
Long short-term memory
16%
Long short-term machine
4%
None of three
❤1
Which of the following project seems attractive to you?
Anonymous Poll
30%
Fake News Detection
26%
Speech Emotion Recognition
28%
Chatbot Project in Python
27%
Movie Recommendation System
21%
Customer Segmentation
31%
Sentiment Analysis
21%
Handwritten Character Recognition
12%
Digit Recognition System
35%
Face Detection System
5%
None of these
Data Science & Machine Learning
Which of the following project seems attractive to you?
Amazing response from you guys in this poll
Lets start with project #1
Fake News Detection
This is an example of text classification since we need to classify whether a news is real or fake
You can refer dataset from kaggle to work on such an amazing project
https://bit.ly/3FGcyoJ
Or
https://www.kaggle.com/c/fake-news/data
Before you work on this project, you should have fair understanding of below topics
Concepts: Stopwords, Porter Stemmer, Tokenisation, Tfid Vectorizer, LSTM, NLP
Libraries: Pandas, NumPy, Matplotlib, Seaborn, Scikit-learn, re, nltk, Tensorflow
Steps:
1. Go through dataset
2. Import the libraries
3. Exploratory Data Analysis [EDA]
4. Data Visualization
5. Data Preparation using Tokenisation and padding
6. Apply theoretical concepts to reduce unnecessary words using Stopwords and Porter Stemmer. Convert text to vector using Count Vectorizer.
7. Split dataset into training and testing
8. Build and train the model using ML Algorithms
9. Model Evaluation using accuracy, recall, precision, confusion matrix and other metrics concepts
Algorithms you can apply:
Logistic Regression, Support Vector Machine, Multilayer Perceptron, KNN, Random Forest, Linear SVM, etc.
ENJOY LEARNING 👍👍
Lets start with project #1
Fake News Detection
This is an example of text classification since we need to classify whether a news is real or fake
You can refer dataset from kaggle to work on such an amazing project
https://bit.ly/3FGcyoJ
Or
https://www.kaggle.com/c/fake-news/data
Before you work on this project, you should have fair understanding of below topics
Concepts: Stopwords, Porter Stemmer, Tokenisation, Tfid Vectorizer, LSTM, NLP
Libraries: Pandas, NumPy, Matplotlib, Seaborn, Scikit-learn, re, nltk, Tensorflow
Steps:
1. Go through dataset
2. Import the libraries
3. Exploratory Data Analysis [EDA]
4. Data Visualization
5. Data Preparation using Tokenisation and padding
6. Apply theoretical concepts to reduce unnecessary words using Stopwords and Porter Stemmer. Convert text to vector using Count Vectorizer.
7. Split dataset into training and testing
8. Build and train the model using ML Algorithms
9. Model Evaluation using accuracy, recall, precision, confusion matrix and other metrics concepts
Algorithms you can apply:
Logistic Regression, Support Vector Machine, Multilayer Perceptron, KNN, Random Forest, Linear SVM, etc.
ENJOY LEARNING 👍👍
👍2