Data Science & Machine Learning – Telegram
Data Science & Machine Learning
73.2K subscribers
792 photos
2 videos
68 files
691 links
Join this channel to learn data science, artificial intelligence and machine learning with funny quizzes, interesting projects and amazing resources for free

For collaborations: @love_data
Download Telegram
Seaborn Cheatsheet
👍41
🚀 Top 10 Tools Data Scientists Love! 🧠

In the ever-evolving world of data science, staying updated with the right tools is crucial to solving complex problems and deriving meaningful insights.

🔍 Here’s a quick breakdown of the most popular tools:

1. Python 🐍: The go-to language for data science, favored for its versatility and powerful libraries.
2. SQL 🛠️: Essential for querying databases and manipulating data.
3. Jupyter Notebooks 📓: An interactive environment that makes data analysis and visualization a breeze.
4. TensorFlow/PyTorch 🤖: Leading frameworks for deep learning and neural networks.
5. Tableau 📊: A user-friendly tool for creating stunning visualizations and dashboards.
6. Git & GitHub 💻: Version control systems that every data scientist should master.
7. Hadoop & Spark 🔥: Big data frameworks that help process massive datasets efficiently.
8. Scikit-learn 🧬: A powerful library for machine learning in Python.
9. R 📈: A statistical programming language that is still a favorite among many analysts.
10. Docker 🐋: A must-have for containerization and deploying applications.

Like if you need similar content 😄👍
👍93
Important Topics to become a data scientist
[Advanced Level]
👇👇

1. Mathematics

Linear Algebra
Analytic Geometry
Matrix
Vector Calculus
Optimization
Regression
Dimensionality Reduction
Density Estimation
Classification

2. Probability

Introduction to Probability
1D Random Variable
The function of One Random Variable
Joint Probability Distribution
Discrete Distribution
Normal Distribution

3. Statistics

Introduction to Statistics
Data Denoscription
Random Samples
Sampling Distribution
Parameter Estimation
Hypotheses Testing
Regression

4. Programming

Python:

Python Basics
List
Set
Tuples
Dictionary
Function
NumPy
Pandas
Matplotlib/Seaborn

R Programming:

R Basics
Vector
List
Data Frame
Matrix
Array
Function
dplyr
ggplot2
Tidyr
Shiny

DataBase:
SQL
MongoDB

Data Structures

Web scraping

Linux

Git

5. Machine Learning

How Model Works
Basic Data Exploration
First ML Model
Model Validation
Underfitting & Overfitting
Random Forest
Handling Missing Values
Handling Categorical Variables
Pipelines
Cross-Validation(R)
XGBoost(Python|R)
Data Leakage

6. Deep Learning

Artificial Neural Network
Convolutional Neural Network
Recurrent Neural Network
TensorFlow
Keras
PyTorch
A Single Neuron
Deep Neural Network
Stochastic Gradient Descent
Overfitting and Underfitting
Dropout Batch Normalization
Binary Classification

7. Feature Engineering

Baseline Model
Categorical Encodings
Feature Generation
Feature Selection

8. Natural Language Processing

Text Classification
Word Vectors

9. Data Visualization Tools

BI (Business Intelligence):
Tableau
Power BI
Qlik View
Qlik Sense

10. Deployment

Microsoft Azure
Heroku
Google Cloud Platform
Flask
Django
6👍2
Top 5 Case Studies for Data Analytics: You Must Know Before Attending an Interview

1. Retail: Target's Predictive Analytics for Customer Behavior
Company: Target
Challenge: Target wanted to identify customers who were expecting a baby to send them personalized promotions.
Solution:
Target used predictive analytics to analyze customers' purchase history and identify patterns that indicated pregnancy.
They tracked purchases of items like unscented lotion, vitamins, and cotton balls.
Outcome:
The algorithm successfully identified pregnant customers, enabling Target to send them relevant promotions.
This personalized marketing strategy increased sales and customer loyalty.

2. Healthcare: IBM Watson's Oncology Treatment Recommendations
Company: IBM Watson
Challenge: Oncologists needed support in identifying the best treatment options for cancer patients.
Solution:
IBM Watson analyzed vast amounts of medical data, including patient records, clinical trials, and medical literature.
It provided oncologists with evidencebased treatment recommendations tailored to individual patients.
Outcome:
Improved treatment accuracy and personalized care for cancer patients.
Reduced time for doctors to develop treatment plans, allowing them to focus more on patient care.

3. Finance: JP Morgan Chase's Fraud Detection System
Company: JP Morgan Chase
Challenge: The bank needed to detect and prevent fraudulent transactions in realtime.
Solution:
Implemented advanced machine learning algorithms to analyze transaction patterns and detect anomalies.
The system flagged suspicious transactions for further investigation.
Outcome:
Significantly reduced fraudulent activities.
Enhanced customer trust and satisfaction due to improved security measures.

4. Sports: Oakland Athletics' Use of Sabermetrics
Team: Oakland Athletics (Moneyball)
Challenge: Compete with larger teams with higher budgets by optimizing player performance and team strategy.
Solution:
Used sabermetrics, a form of advanced statistical analysis, to evaluate player performance and potential.
Focused on undervalued players with high onbase percentages and other key metrics.
Outcome:
Achieved remarkable success with a limited budget.
Revolutionized the approach to team building and player evaluation in baseball and other sports.

5. Ecommerce: Amazon's Recommendation Engine
Company: Amazon
Challenge: Enhance customer shopping experience and increase sales through personalized recommendations.
Solution:
Implemented a recommendation engine using collaborative filtering, which analyzes user behavior and purchase history.
The system suggests products based on what similar users have bought.
Outcome:
Increased average order value and customer retention.
Significantly contributed to Amazon's revenue growth through crossselling and upselling.

Like if it helps 😄
👍12
What 𝗠𝗟 𝗰𝗼𝗻𝗰𝗲𝗽𝘁𝘀 are commonly asked in 𝗱𝗮𝘁𝗮 𝘀𝗰𝗶𝗲𝗻𝗰𝗲 𝗶𝗻𝘁𝗲𝗿𝘃𝗶𝗲𝘄𝘀?

These are fair game in interviews at 𝘀𝘁𝗮𝗿𝘁𝘂𝗽𝘀, 𝗰𝗼𝗻𝘀𝘂𝗹𝘁𝗶𝗻𝗴 & 𝗹𝗮𝗿𝗴𝗲 𝘁𝗲𝗰𝗵.

𝗙𝘂𝗻𝗱𝗮𝗺𝗲𝗻𝘁𝗮𝗹𝘀
- Supervised vs. Unsupervised Learning
- Overfitting and Underfitting
- Cross-validation
- Bias-Variance Tradeoff
- Accuracy vs Interpretability
- Accuracy vs Latency

𝗠𝗟 𝗔𝗹𝗴𝗼𝗿𝗶𝘁𝗵𝗺𝘀
- Logistic Regression
- Decision Trees
- Random Forest
- Support Vector Machines
- K-Nearest Neighbors
- Naive Bayes
- Linear Regression
- Ridge and Lasso Regression
- K-Means Clustering
- Hierarchical Clustering
- PCA

𝗠𝗼𝗱𝗲𝗹𝗶𝗻𝗴 𝗦𝘁𝗲𝗽𝘀
- EDA
- Data Cleaning (e.g. missing value imputation)
- Data Preprocessing (e.g. scaling)
- Feature Engineering (e.g. aggregation)
- Feature Selection (e.g. variable importance)
- Model Training (e.g. gradient descent)
- Model Evaluation (e.g. AUC vs Accuracy)
- Model Productionization

𝗛𝘆𝗽𝗲𝗿𝗽𝗮𝗿𝗮𝗺𝗲𝘁𝗲𝗿 𝗧𝘂𝗻𝗶𝗻𝗴
- Grid Search
- Random Search
- Bayesian Optimization

𝗠𝗟 𝗖𝗮𝘀𝗲𝘀
- [Capital One] Detect credit card fraudsters
- [Amazon] Forecast monthly sales
- [Airbnb] Estimate lifetime value of a guest

Like if you need similar content 😄👍
👍32
Relatable? 😂
😁24🤔1
When you start making good money, do this:

1. Buy fewer clothes, but wear the highest quality.
2. Eat premium food, not junk.
3. Hire a helper for household chores. Buy back your time.
4. Upgrade your mattress. Sleep changes everything.
5. Invest in experiences, not just stuff.
6. Upgrade your financial adviser. The one who got you here won’t get you to the next level.
7. Surround yourself with high-value people.

Small shifts. Big impact.
14👍6
How much Statistics must I know to become a Data Scientist?

This is one of the most common questions

Here are the must-know Statistics concepts every Data Scientist should know:

𝗣𝗿𝗼𝗯𝗮𝗯𝗶𝗹𝗶𝘁𝘆

Bayes' Theorem & conditional probability
Permutations & combinations
Card & die roll problem-solving

𝗗𝗲𝘀𝗰𝗿𝗶𝗽𝘁𝗶𝘃𝗲 𝘀𝘁𝗮𝘁𝗶𝘀𝘁𝗶𝗰𝘀 & 𝗱𝗶𝘀𝘁𝗿𝗶𝗯𝘂𝘁𝗶𝗼𝗻𝘀

Mean, median, mode
Standard deviation and variance
  Bernoulli's, Binomial, Normal, Uniform, Exponential distributions

𝗜𝗻𝗳𝗲𝗿𝗲𝗻𝘁𝗶𝗮𝗹 𝘀𝘁𝗮𝘁𝗶𝘀𝘁𝗶𝗰𝘀

A/B experimentation
T-test, Z-test, Chi-squared tests
Type 1 & 2 errors
Sampling techniques & biases
Confidence intervals & p-values
Central Limit Theorem
Causal inference techniques

𝗠𝗮𝗰𝗵𝗶𝗻𝗲 𝗹𝗲𝗮𝗿𝗻𝗶𝗻𝗴

Logistic & Linear regression
Decision trees & random forests
Clustering models
Feature engineering
Feature selection methods
Model testing & validation
Time series analysis

Join our WhatsApp channel for more Statistics Resources
👇👇
https://whatsapp.com/channel/0029Vat3Dc4KAwEcfFbNnZ3O

Like if you need similar content 😄👍
👍6
10 great Python packages for Data Science not known to many:

1️⃣ CleanLab

Cleanlab helps you clean data and labels by automatically detecting issues in a ML dataset.

2️⃣ LazyPredict

A Python library that enables you to train, test, and evaluate multiple ML models at once using just a few lines of code.

3️⃣ Lux

A Python library for quickly visualizing and analyzing data, providing an easy and efficient way to explore data.

4️⃣ PyForest

A time-saving tool that helps in importing all the necessary data science libraries and functions with a single line of code.

5️⃣ PivotTableJS

PivotTableJS lets you interactively analyse your data in Jupyter Notebooks without any code 🔥

6️⃣ Drawdata

Drawdata is a python library that allows you to draw a 2-D dataset of any shape in a Jupyter Notebook.

7️⃣ black

The Uncompromising Code Formatter

8️⃣ PyCaret

An open-source, low-code machine learning library in Python that automates the machine learning workflow.

9️⃣ PyTorch-Lightning by LightningAI

Streamlines your model training, automates boilerplate code, and lets you focus on what matters: research & innovation.

🔟 Streamlit

A framework for creating web applications for data science and machine learning projects, allowing for easy and interactive data viz & model deployment.

I have curated the best interview resources to crack Data Science Interviews
👇👇
https://whatsapp.com/channel/0029VaiM08SDuMRaGKd9Wv0L

Like if you need similar content 😄👍
4👍3
Data Science Interview Questions

Question 1 : How would you approach building a recommendation system for personalized content on Facebook? Consider factors like scalability and user privacy.

   - Answer: Building a recommendation system for personalized content on Facebook would involve collaborative filtering or content-based methods. Scalability can be achieved using distributed computing, and user privacy can be preserved through techniques like federated learning.


Question 2 : Describe a situation where you had to navigate conflicting opinions within your team. How did you facilitate resolution and maintain team cohesion?

   - Answer: In navigating conflicting opinions within a team, I facilitated resolution through open communication, active listening, and finding common ground. Prioritizing team cohesion was key to achieving consensus.


Question 3 : How would you enhance the security of user data on Facebook, considering the evolving landscape of cybersecurity threats?

   - Answer: Enhancing the security of user data on Facebook involves implementing robust encryption mechanisms, access controls, and regular security audits. Ensuring compliance with privacy regulations and proactive threat monitoring are essential.

Question 4 : Design a real-time notification system for Facebook, ensuring timely delivery of notifications to users across various platforms.

   - Answer: Designing a real-time notification system for Facebook requires technologies like WebSocket for real-time communication and push notifications. Ensuring scalability and reliability through distributed systems is crucial for timely delivery.

I have curated the best interview resources to crack Data Science Interviews
👇👇
https://whatsapp.com/channel/0029Va8v3eo1NCrQfGMseL2D

Like if you need similar content 😄👍
👍6
Data Analyst vs Data Scientist 👆
👍7
😂😂
😁19
Data Science Interview Questions

1: How would you preprocess and tokenize text data from tweets for sentiment analysis? Discuss potential challenges and solutions.

- Answer: Preprocessing and tokenizing text data for sentiment analysis involves tasks like lowercasing, removing stop words, and stemming or lemmatization. Handling challenges like handling emojis, slang, and noisy text is crucial. Tools like NLTK or spaCy can assist in these tasks.


2: Explain the collaborative filtering approach in building recommendation systems. How might Twitter use this to enhance user experience?

- Answer: Collaborative filtering recommends items based on user preferences and similarities. Techniques include user-based or item-based collaborative filtering and matrix factorization. Twitter could leverage user interactions to recommend tweets, users, or topics.


3: Write a Python or Scala function to count the frequency of hashtags in a given collection of tweets.

- Answer (Python):
   
     def count_hashtags(tweet_collection):
         hashtags_count = {}
         for tweet in tweet_collection:
             hashtags = [word for word in tweet.split() if word.startswith('#')]
             for hashtag in hashtags:
                 hashtags_count[hashtag] = hashtags_count.get(hashtag, 0) + 1
         return hashtags_count
    


4: How does graph analysis contribute to understanding user interactions and content propagation on Twitter? Provide a specific use case.

- Answer: Graph analysis on Twitter involves examining user interactions. For instance, identifying influential users or detecting communities based on retweet or mention networks. Algorithms like PageRank or Louvain Modularity can aid in these analyses.

I have curated the best interview resources to crack Data Science Interviews
👇👇
https://whatsapp.com/channel/0029Va8v3eo1NCrQfGMseL2D

Like if you need similar content 😄👍
👍71
👍12🥰2👏1
If I Were to Start My Data Science Career from Scratch, Here's What I Would Do 👇

1️⃣ Master Advanced SQL

Foundations: Learn database structures, tables, and relationships.

Basic SQL Commands: SELECT, FROM, WHERE, ORDER BY.

Aggregations: Get hands-on with SUM, COUNT, AVG, MIN, MAX, GROUP BY, and HAVING.

JOINs: Understand LEFT, RIGHT, INNER, OUTER, and CARTESIAN joins.

Advanced Concepts: CTEs, window functions, and query optimization.

Metric Development: Build and report metrics effectively.


2️⃣ Study Statistics & A/B Testing

Denoscriptive Statistics: Know your mean, median, mode, and standard deviation.

Distributions: Familiarize yourself with normal, Bernoulli, binomial, exponential, and uniform distributions.

Probability: Understand basic probability and Bayes' theorem.

Intro to ML: Start with linear regression, decision trees, and K-means clustering.

Experimentation Basics: T-tests, Z-tests, Type 1 & Type 2 errors.

A/B Testing: Design experiments—hypothesis formation, sample size calculation, and sample biases.


3️⃣ Learn Python for Data

Data Manipulation: Use pandas for data cleaning and manipulation.

Data Visualization: Explore matplotlib and seaborn for creating visualizations.

Hypothesis Testing: Dive into scipy for statistical testing.

Basic Modeling: Practice building models with scikit-learn.


4️⃣ Develop Product Sense

Product Management Basics: Manage projects and understand the product life cycle.

Data-Driven Strategy: Leverage data to inform decisions and measure success.

Metrics in Business: Define and evaluate metrics that matter to the business.


5️⃣ Hone Soft Skills

Communication: Clearly explain data findings to technical and non-technical audiences.

Collaboration: Work effectively in teams.

Time Management: Prioritize and manage projects efficiently.

Self-Reflection: Regularly assess and improve your skills.


6️⃣ Bonus: Basic Data Engineering

Data Modeling: Understand dimensional modeling and trade-offs in normalization vs. denormalization.

ETL: Set up extraction jobs, manage dependencies, clean and validate data.

Pipeline Testing: Conduct unit testing and ensure data quality throughout the pipeline.

I have curated the best interview resources to crack Data Science Interviews
👇👇
https://whatsapp.com/channel/0029Va8v3eo1NCrQfGMseL2D

Like if you need similar content 😄👍
👍101
Complete Data Science Roadmap 
👇👇 

1. Introduction to Data Science 
   - Overview and Importance 
   - Data Science Lifecycle 
   - Key Roles (Data Scientist, Analyst, Engineer) 

2. Mathematics and Statistics 
   - Probability and Distributions 
   - Denoscriptive/Inferential Statistics 
   - Hypothesis Testing 
   - Linear Algebra and Calculus Basics 

3. Programming Languages 
   - Python: NumPy, Pandas, Matplotlib 
   - R: dplyr, ggplot2 
   - SQL: Joins, Aggregations, CRUD 

4. Data Collection & Preprocessing 
   - Data Cleaning and Wrangling 
   - Handling Missing Data 
   - Feature Engineering 

5. Exploratory Data Analysis (EDA) 
   - Summary Statistics 
   - Data Visualization (Histograms, Box Plots, Correlation) 

6. Machine Learning 
   - Supervised (Linear/Logistic Regression, Decision Trees) 
   - Unsupervised (K-Means, PCA) 
   - Model Selection and Cross-Validation 

7. Advanced Machine Learning 
   - SVM, Random Forests, Boosting 
   - Neural Networks Basics 

8. Deep Learning 
   - Neural Networks Architecture 
   - CNNs for Image Data 
   - RNNs for Sequential Data 

9. Natural Language Processing (NLP) 
   - Text Preprocessing 
   - Sentiment Analysis 
   - Word Embeddings (Word2Vec) 

10. Data Visualization & Storytelling 
   - Dashboards (Tableau, Power BI) 
   - Telling Stories with Data 

11. Model Deployment 
   - Deploy with Flask or Django 
   - Monitoring and Retraining Models 

12. Big Data & Cloud 
   - Introduction to Hadoop, Spark 
   - Cloud Tools (AWS, Google Cloud) 

13. Data Engineering Basics 
   - ETL Pipelines 
   - Data Warehousing (Redshift, BigQuery) 

14. Ethics in Data Science 
   - Ethical Data Usage 
   - Bias in AI Models 

15. Tools for Data Science 
   - Jupyter, Git, Docker 

16. Career Path & Certifications 
   - Building a Data Science Portfolio 

Like if you need similar content 😄👍
👍10
©How fresher can get a job as a data scientist?©

1. Education: Obtain a degree in a relevant field such as computer science, statistics, mathematics, or data science. Consider pursuing additional certifications or specialized courses in data science to enhance your skills.

2. Build a strong foundation: Develop a strong understanding of key concepts in data science such as statistics, machine learning, programming languages (such as Python or R), and data visualization.

3. Hands-on experience: Gain practical experience by working on projects, participating in hackathons, or internships. Building a portfolio of projects showcasing your data science skills can be beneficial when applying for jobs.

4. Networking: Attend industry events, conferences, and meetups to network with professionals in the field. Networking can help you learn about job opportunities and make valuable connections.

5. Apply for entry-level positions: Look for entry-level positions such as data analyst, research assistant, or junior data scientist roles to gain experience and start building your career in data science.

6. Prepare for interviews: Practice common data science interview questions, showcase your problem-solving skills, and be prepared to discuss your projects and experiences related to data science.

7. Continuous learning: Data science is a rapidly evolving field, so it's important to stay updated on the latest trends, tools, and techniques. Consider taking online courses, attending workshops, or joining professional organizations to continue learning and growing in the field.
👍61
Many data scientists don't know how to push ML models to production. Here's the recipe 👇

𝗞𝗲𝘆 𝗜𝗻𝗴𝗿𝗲𝗱𝗶𝗲𝗻𝘁𝘀

🔹 𝗧𝗿𝗮𝗶𝗻 / 𝗧𝗲𝘀𝘁 𝗗𝗮𝘁𝗮𝘀𝗲𝘁 - Ensure Test is representative of Online data
🔹 𝗙𝗲𝗮𝘁𝘂𝗿𝗲 𝗘𝗻𝗴𝗶𝗻𝗲𝗲𝗿𝗶𝗻𝗴 𝗣𝗶𝗽𝗲𝗹𝗶𝗻𝗲 - Generate features in real-time
🔹 𝗠𝗼𝗱𝗲𝗹 𝗢𝗯𝗷𝗲𝗰𝘁 - Trained SkLearn or Tensorflow Model
🔹 𝗣𝗿𝗼𝗷𝗲𝗰𝘁 𝗖𝗼𝗱𝗲 𝗥𝗲𝗽𝗼 - Save model project code to Github
🔹 𝗔𝗣𝗜 𝗙𝗿𝗮𝗺𝗲𝘄𝗼𝗿𝗸 - Use FastAPI or Flask to build a model API
🔹 𝗗𝗼𝗰𝗸𝗲𝗿 - Containerize the ML model API
🔹 𝗥𝗲𝗺𝗼𝘁𝗲 𝗦𝗲𝗿𝘃𝗲𝗿 - Choose a cloud service; e.g. AWS sagemaker
🔹 𝗨𝗻𝗶𝘁 𝗧𝗲𝘀𝘁𝘀 - Test inputs & outputs of functions and APIs
🔹 𝗠𝗼𝗱𝗲𝗹 𝗠𝗼𝗻𝗶𝘁𝗼𝗿𝗶𝗻𝗴 - Evidently AI, a simple, open-source for ML monitoring

𝗣𝗿𝗼𝗰𝗲𝗱𝘂𝗿𝗲

𝗦𝘁𝗲𝗽 𝟭 - 𝗗𝗮𝘁𝗮 𝗣𝗿𝗲𝗽𝗮𝗿𝗮𝘁𝗶𝗼𝗻 & 𝗙𝗲𝗮𝘁𝘂𝗿𝗲 𝗘𝗻𝗴𝗶𝗻𝗲𝗲𝗿𝗶𝗻𝗴

Don't push a model with 90% accuracy on train set. Do it based on the test set - if and only if, the test set is representative of the online data. Use SkLearn pipeline to chain a series of model preprocessing functions like null handling.

𝗦𝘁𝗲𝗽 𝟮 - 𝗠𝗼𝗱𝗲𝗹 𝗗𝗲𝘃𝗲𝗹𝗼𝗽𝗺𝗲𝗻𝘁

Train your model with frameworks like Sklearn or Tensorflow. Push the model code including preprocessing, training and validation noscripts to Github for reproducibility.

𝗦𝘁𝗲𝗽 𝟯 - 𝗔𝗣𝗜 𝗗𝗲𝘃𝗲𝗹𝗼𝗽𝗺𝗲𝗻𝘁 & 𝗖𝗼𝗻𝘁𝗮𝗶𝗻𝗲𝗿𝗶𝘇𝗮𝘁𝗶𝗼𝗻

Your model needs a "/predict" endpoint, which receives a JSON object in the request input and generates a JSON object with the model score in the response output. You can use frameworks like FastAPI or Flask. Containzerize this API so that it's agnostic to server environment

𝗦𝘁𝗲𝗽 𝟰 - 𝗧𝗲𝘀𝘁𝗶𝗻𝗴 & 𝗗𝗲𝗽𝗹𝗼𝘆𝗺𝗲𝗻𝘁

Write tests to validate inputs & outputs of API functions to prevent errors. Push the code to remote services like AWS Sagemaker.

𝗦𝘁𝗲𝗽 𝟱 - 𝗠𝗼𝗻𝗶𝘁𝗼𝗿𝗶𝗻𝗴

Set up monitoring tools like Evidently AI, or use a built-in one within AWS Sagemaker. I use such tools to track performance metrics and data drifts on online data.

Data Science Resources
👇👇
https://whatsapp.com/channel/0029Va8v3eo1NCrQfGMseL2D

Like if you need similar content 😄👍
👍121
Here is the list of few projects (found on kaggle). They cover Basics of Python, Advanced Statistics, Supervised Learning (Regression and Classification problems) & Data Science

Please also check the discussions and notebook submissions for different approaches and solution after you tried yourself.

1. Basic python and statistics

Pima Indians :- https://www.kaggle.com/uciml/pima-indians-diabetes-database
Cardio Goodness fit :- https://www.kaggle.com/saurav9786/cardiogoodfitness
Automobile :- https://www.kaggle.com/toramky/automobile-dataset

2. Advanced Statistics

Game of Thrones:-https://www.kaggle.com/mylesoneill/game-of-thrones
World University Ranking:-https://www.kaggle.com/mylesoneill/world-university-rankings
IMDB Movie Dataset:- https://www.kaggle.com/carolzhangdc/imdb-5000-movie-dataset

3. Supervised Learning

a) Regression Problems

How much did it rain :- https://www.kaggle.com/c/how-much-did-it-rain-ii/overview
Inventory Demand:- https://www.kaggle.com/c/grupo-bimbo-inventory-demand
Property Inspection predictiion:- https://www.kaggle.com/c/liberty-mutual-group-property-inspection-prediction
Restaurant Revenue prediction:- https://www.kaggle.com/c/restaurant-revenue-prediction/data
IMDB Box office Prediction:-https://www.kaggle.com/c/tmdb-box-office-prediction/overview

b) Classification problems

Employee Access challenge :- https://www.kaggle.com/c/amazon-employee-access-challenge/overview
Titanic :- https://www.kaggle.com/c/titanic
San Francisco crime:- https://www.kaggle.com/c/sf-crime
Customer satisfcation:-https://www.kaggle.com/c/santander-customer-satisfaction
Trip type classification:- https://www.kaggle.com/c/walmart-recruiting-trip-type-classification
Categorize cusine:- https://www.kaggle.com/c/whats-cooking

4. Some helpful Data science projects for beginners

https://www.kaggle.com/c/house-prices-advanced-regression-techniques

https://www.kaggle.com/c/digit-recognizer

https://www.kaggle.com/c/titanic

5. Intermediate Level Data science Projects

Black Friday Data : https://www.kaggle.com/sdolezel/black-friday

Human Activity Recognition Data : https://www.kaggle.com/uciml/human-activity-recognition-with-smartphones

Trip History Data : https://www.kaggle.com/pronto/cycle-share-dataset

Million Song Data : https://www.kaggle.com/c/msdchallenge

Census Income Data : https://www.kaggle.com/c/census-income/data

Movie Lens Data : https://www.kaggle.com/grouplens/movielens-20m-dataset

Twitter Classification Data : https://www.kaggle.com/c/twitter-sentiment-analysis2

Share with credits: https://news.1rj.ru/str/sqlproject

ENJOY LEARNING 👍👍
👍5