Machine Learning & Artificial Intelligence | Data Science Free Courses – Telegram
Machine Learning & Artificial Intelligence | Data Science Free Courses
64.5K subscribers
557 photos
2 videos
98 files
426 links
Perfect channel to learn Data Analytics, Data Sciene, Machine Learning & Artificial Intelligence

Admin: @coderfun
Download Telegram
Complete Roadmap to learn Machine Learning and Artificial Intelligence
👇👇

Week 1-2: Introduction to Machine Learning
- Learn the basics of Python programming language (if you are not already familiar with it)
- Understand the fundamentals of Machine Learning concepts such as supervised learning, unsupervised learning, and reinforcement learning
- Study linear algebra and calculus basics
- Complete online courses like Andrew Ng's Machine Learning course on Coursera

Week 3-4: Deep Learning Fundamentals
- Dive into neural networks and deep learning
- Learn about different types of neural networks like Convolutional Neural Networks (CNNs) and Recurrent Neural Networks (RNNs)
- Implement deep learning models using frameworks like TensorFlow or PyTorch
- Complete online courses like Deep Learning Specialization on Coursera

Week 5-6: Natural Language Processing (NLP) and Computer Vision
- Explore NLP techniques such as tokenization, word embeddings, and sentiment analysis
- Dive into computer vision concepts like image classification, object detection, and image segmentation
- Work on projects involving NLP and Computer Vision applications

Week 7-8: Reinforcement Learning and AI Applications
- Learn about Reinforcement Learning algorithms like Q-learning and Deep Q Networks
- Explore AI applications in fields like healthcare, finance, and autonomous vehicles
- Work on a final project that combines different aspects of Machine Learning and AI

Additional Tips:
- Practice coding regularly to strengthen your programming skills
- Join online communities like Kaggle or GitHub to collaborate with other learners
- Read research papers and articles to stay updated on the latest advancements in the field

Pro Tip: Roadmap won't help unless you start working on it consistently. Start working on projects as early as possible.

2 months are good as a starting point to get grasp the basics of ML & AI but mastering it is very difficult as AI keeps evolving every day.

Best Resources to learn ML & AI 👇

Learn Python for Free

Prompt Engineering Course

Prompt Engineering Guide

Data Science Course

Google Cloud Generative AI Path

Unlock the power of Generative AI Models

Machine Learning with Python Free Course

Machine Learning Free Book

Deep Learning Nanodegree Program with Real-world Projects

AI, Machine Learning and Deep Learning

Join @free4unow_backup for more free courses

ENJOY LEARNING👍👍
👍71👌1🤣1
Complete Roadmap to become a data scientist in 5 months

Free Resources to learn Data Science: https://whatsapp.com/channel/0029Va8v3eo1NCrQfGMseL2D

Week 1-2: Fundamentals
- Day 1-3: Introduction to Data Science, its applications, and roles.
- Day 4-7: Brush up on Python programming.
- Day 8-10: Learn basic statistics and probability.

Week 3-4: Data Manipulation and Visualization
- Day 11-15: Pandas for data manipulation.
- Day 16-20: Data visualization with Matplotlib and Seaborn.

Week 5-6: Machine Learning Foundations
- Day 21-25: Introduction to scikit-learn.
- Day 26-30: Linear regression and logistic regression.

Work on Data Science Projects: https://news.1rj.ru/str/pythonspecialist/29

Week 7-8: Advanced Machine Learning
- Day 31-35: Decision trees and random forests.
- Day 36-40: Clustering (K-Means, DBSCAN) and dimensionality reduction.

Week 9-10: Deep Learning
- Day 41-45: Basics of Neural Networks and TensorFlow/Keras.
- Day 46-50: Convolutional Neural Networks (CNNs) and Recurrent Neural Networks (RNNs).

Week 11-12: Data Engineering
- Day 51-55: Learn about SQL and databases.
- Day 56-60: Data preprocessing and cleaning.

Week 13-14: Model Evaluation and Optimization
- Day 61-65: Cross-validation, hyperparameter tuning.
- Day 66-70: Evaluation metrics (accuracy, precision, recall, F1-score).

Week 15-16: Big Data and Tools
- Day 71-75: Introduction to big data technologies (Hadoop, Spark).
- Day 76-80: Basics of cloud computing (AWS, GCP, Azure).

Week 17-18: Deployment and Production
- Day 81-85: Model deployment with Flask or FastAPI.
- Day 86-90: Containerization with Docker, cloud deployment (AWS, Heroku).

Week 19-20: Specialization
- Day 91-95: NLP or Computer Vision, based on your interests.

Week 21-22: Projects and Portfolios
- Day 96-100: Work on personal data science projects.

Week 23-24: Soft Skills and Networking
- Day 101-105: Improve communication and presentation skills.
- Day 106-110: Attend online data science meetups or forums.

Week 25-26: Interview Preparation
- Day 111-115: Practice coding interviews on platforms like LeetCode.
- Day 116-120: Review your projects and be ready to discuss them.

Week 27-28: Apply for Jobs
- Day 121-125: Start applying for entry-level data scientist positions.

Week 29-30: Interviews
- Day 126-130: Attend interviews, practice whiteboard problems.

Week 31-32: Continuous Learning
- Day 131-135: Stay updated with the latest trends in data science.

Week 33-34: Accepting Offers
- Day 136-140: Evaluate job offers and negotiate if necessary.

Week 35-36: Settling In
- Day 141-150: Start your new data science job, adapt to the team, and continue learning on the job.

ENJOY LEARNING 👍👍
👍61🤣1
Many data scientists don't know how to push ML models to production. Here's the recipe 👇

𝗞𝗲𝘆 𝗜𝗻𝗴𝗿𝗲𝗱𝗶𝗲𝗻𝘁𝘀

🔹 𝗧𝗿𝗮𝗶𝗻 / 𝗧𝗲𝘀𝘁 𝗗𝗮𝘁𝗮𝘀𝗲𝘁 - Ensure Test is representative of Online data
🔹 𝗙𝗲𝗮𝘁𝘂𝗿𝗲 𝗘𝗻𝗴𝗶𝗻𝗲𝗲𝗿𝗶𝗻𝗴 𝗣𝗶𝗽𝗲𝗹𝗶𝗻𝗲 - Generate features in real-time
🔹 𝗠𝗼𝗱𝗲𝗹 𝗢𝗯𝗷𝗲𝗰𝘁 - Trained SkLearn or Tensorflow Model
🔹 𝗣𝗿𝗼𝗷𝗲𝗰𝘁 𝗖𝗼𝗱𝗲 𝗥𝗲𝗽𝗼 - Save model project code to Github
🔹 𝗔𝗣𝗜 𝗙𝗿𝗮𝗺𝗲𝘄𝗼𝗿𝗸 - Use FastAPI or Flask to build a model API
🔹 𝗗𝗼𝗰𝗸𝗲𝗿 - Containerize the ML model API
🔹 𝗥𝗲𝗺𝗼𝘁𝗲 𝗦𝗲𝗿𝘃𝗲𝗿 - Choose a cloud service; e.g. AWS sagemaker
🔹 𝗨𝗻𝗶𝘁 𝗧𝗲𝘀𝘁𝘀 - Test inputs & outputs of functions and APIs
🔹 𝗠𝗼𝗱𝗲𝗹 𝗠𝗼𝗻𝗶𝘁𝗼𝗿𝗶𝗻𝗴 - Evidently AI, a simple, open-source for ML monitoring

𝗣𝗿𝗼𝗰𝗲𝗱𝘂𝗿𝗲

𝗦𝘁𝗲𝗽 𝟭 - 𝗗𝗮𝘁𝗮 𝗣𝗿𝗲𝗽𝗮𝗿𝗮𝘁𝗶𝗼𝗻 & 𝗙𝗲𝗮𝘁𝘂𝗿𝗲 𝗘𝗻𝗴𝗶𝗻𝗲𝗲𝗿𝗶𝗻𝗴

Don't push a model with 90% accuracy on train set. Do it based on the test set - if and only if, the test set is representative of the online data. Use SkLearn pipeline to chain a series of model preprocessing functions like null handling.

𝗦𝘁𝗲𝗽 𝟮 - 𝗠𝗼𝗱𝗲𝗹 𝗗𝗲𝘃𝗲𝗹𝗼𝗽𝗺𝗲𝗻𝘁

Train your model with frameworks like Sklearn or Tensorflow. Push the model code including preprocessing, training and validation noscripts to Github for reproducibility.

𝗦𝘁𝗲𝗽 𝟯 - 𝗔𝗣𝗜 𝗗𝗲𝘃𝗲𝗹𝗼𝗽𝗺𝗲𝗻𝘁 & 𝗖𝗼𝗻𝘁𝗮𝗶𝗻𝗲𝗿𝗶𝘇𝗮𝘁𝗶𝗼𝗻

Your model needs a "/predict" endpoint, which receives a JSON object in the request input and generates a JSON object with the model score in the response output. You can use frameworks like FastAPI or Flask. Containzerize this API so that it's agnostic to server environment

𝗦𝘁𝗲𝗽 𝟰 - 𝗧𝗲𝘀𝘁𝗶𝗻𝗴 & 𝗗𝗲𝗽𝗹𝗼𝘆𝗺𝗲𝗻𝘁

Write tests to validate inputs & outputs of API functions to prevent errors. Push the code to remote services like AWS Sagemaker.

𝗦𝘁𝗲𝗽 𝟱 - 𝗠𝗼𝗻𝗶𝘁𝗼𝗿𝗶𝗻𝗴

Set up monitoring tools like Evidently AI, or use a built-in one within AWS Sagemaker. I use such tools to track performance metrics and data drifts on online data.
👍8🤣1
Essential Python Libraries for Data Analytics 😄👇

Python Free Resources: https://news.1rj.ru/str/pythondevelopersindia

1. NumPy:
- Efficient numerical operations and array manipulation.

2. Pandas:
- Data manipulation and analysis with powerful data structures (DataFrame, Series).

3. Matplotlib:
- 2D plotting library for creating visualizations.

4. Scikit-learn:
- Machine learning toolkit for classification, regression, clustering, etc.

5. TensorFlow:
- Open-source machine learning framework for building and deploying ML models.

6. PyTorch:
- Deep learning library, particularly popular for neural network research.

7. Django:
- High-level web framework for building robust, scalable web applications.

8. Flask:
- Lightweight web framework for building smaller web applications and APIs.

9. Requests:
- HTTP library for making HTTP requests.

10. Beautiful Soup:
- Web scraping library for pulling data out of HTML and XML files.

As a beginner, you can start with Pandas and Numpy libraries for data analysis. If you want to transition from Data Analyst to Data Scientist, then you can start applying ML libraries like Scikit-learn, Tensorflow, Pytorch, etc. in your data projects.

Share with credits: https://news.1rj.ru/str/sqlspecialist

Hope it helps :)
👍41😁1
Complete Roadmap to learn Data Science

1. Foundational Knowledge

Mathematics and Statistics

- Linear Algebra: Understand vectors, matrices, and tensor operations.
- Calculus: Learn about derivatives, integrals, and optimization techniques.
- Probability: Study probability distributions, Bayes' theorem, and expected values.
- Statistics: Focus on denoscriptive statistics, hypothesis testing, regression, and statistical significance.

Programming

- Python: Start with basic syntax, data structures, and OOP concepts. Libraries to learn: NumPy, pandas, matplotlib, seaborn.
- R: Get familiar with basic syntax and data manipulation (optional but useful).
- SQL: Understand database querying, joins, aggregations, and subqueries.

2. Core Data Science Concepts

Data Wrangling and Preprocessing

- Cleaning and preparing data for analysis.
- Handling missing data, outliers, and inconsistencies.
- Feature engineering and selection.

Data Visualization

- Tools: Matplotlib, seaborn, Plotly.
- Concepts: Types of plots, storytelling with data, interactive visualizations.

Machine Learning

- Supervised Learning: Linear regression, logistic regression, decision trees, random forests, support vector machines, k-nearest neighbors.
- Unsupervised Learning: K-means clustering, hierarchical clustering, PCA.
- Advanced Techniques: Ensemble methods, gradient boosting (XGBoost, LightGBM), neural networks.
- Model Evaluation: Train-test split, cross-validation, confusion matrix, ROC-AUC.


3. Advanced Topics

Deep Learning

- Frameworks: TensorFlow, Keras, PyTorch.
- Concepts: Neural networks, CNNs, RNNs, LSTMs, GANs.

Natural Language Processing (NLP)

- Basics: Text preprocessing, tokenization, stemming, lemmatization.
- Advanced: Sentiment analysis, topic modeling, word embeddings (Word2Vec, GloVe), transformers (BERT, GPT).

Big Data Technologies

- Frameworks: Hadoop, Spark.
- Databases: NoSQL databases (MongoDB, Cassandra).

4. Practical Experience

Projects

- Start with small datasets (Kaggle, UCI Machine Learning Repository).
- Progress to more complex projects involving real-world data.
- Work on end-to-end projects, from data collection to model deployment.

Competitions and Challenges

- Participate in Kaggle competitions.
- Engage in hackathons and coding challenges.

5. Soft Skills and Tools

Communication

- Learn to present findings clearly and concisely.
- Practice writing reports and creating dashboards (Tableau, Power BI).

Collaboration Tools

- Version Control: Git and GitHub.
- Project Management: JIRA, Trello.

6. Continuous Learning and Networking

Staying Updated

- Follow data science blogs, podcasts, and research papers.
- Join professional groups and forums (LinkedIn, Kaggle, Reddit, DataSimplifier).

7. Specialization

After gaining a broad understanding, you might want to specialize in areas such as:
- Data Engineering
- Business Analytics
- Computer Vision
- AI and Machine Learning Research
👍7🤣1
Hey Guys👋,

The Average Salary Of a Data Scientist is 14LPA 

𝐁𝐞𝐜𝐨𝐦𝐞 𝐚 𝐂𝐞𝐫𝐭𝐢𝐟𝐢𝐞𝐝 𝐃𝐚𝐭𝐚 𝐒𝐜𝐢𝐞𝐧𝐭𝐢𝐬𝐭 𝐈𝐧 𝐓𝐨𝐩 𝐌𝐍𝐂𝐬😍

We help you master the required skills.

Learn by doing, build Industry level projects

👩‍🎓 1500+ Students Placed
💼 7.2 LPA Avg. Package
💰 41 LPA Highest Package
🤝 450+ Hiring Partners

Apply for FREE👇 :
https://tracking.acciojob.com/g/PUfdDxgHR

( Limited Slots )
👍71🤣1
Machine Learning types
👍31🤣1
If you know all these, you know most things in Generative AI 👇👇
https://news.1rj.ru/str/generativeai_gpt/266
👍2🎉1
Some essential concepts every data scientist should understand:

### 1. Statistics and Probability
- Purpose: Understanding data distributions and making inferences.
- Core Concepts: Denoscriptive statistics (mean, median, mode), inferential statistics, probability distributions (normal, binomial), hypothesis testing, p-values, confidence intervals.

### 2. Programming Languages
- Purpose: Implementing data analysis and machine learning algorithms.
- Popular Languages: Python, R.
- Libraries: NumPy, Pandas, Scikit-learn (Python), dplyr, ggplot2 (R).

### 3. Data Wrangling
- Purpose: Cleaning and transforming raw data into a usable format.
- Techniques: Handling missing values, data normalization, feature engineering, data aggregation.

### 4. Exploratory Data Analysis (EDA)
- Purpose: Summarizing the main characteristics of a dataset, often using visual methods.
- Tools: Matplotlib, Seaborn (Python), ggplot2 (R).
- Techniques: Histograms, scatter plots, box plots, correlation matrices.

### 5. Machine Learning
- Purpose: Building models to make predictions or find patterns in data.
- Core Concepts: Supervised learning (regression, classification), unsupervised learning (clustering, dimensionality reduction), model evaluation (accuracy, precision, recall, F1 score).
- Algorithms: Linear regression, logistic regression, decision trees, random forests, support vector machines, k-means clustering, principal component analysis (PCA).

### 6. Deep Learning
- Purpose: Advanced machine learning techniques using neural networks.
- Core Concepts: Neural networks, backpropagation, activation functions, overfitting, dropout.
- Frameworks: TensorFlow, Keras, PyTorch.

### 7. Natural Language Processing (NLP)
- Purpose: Analyzing and modeling textual data.
- Core Concepts: Tokenization, stemming, lemmatization, TF-IDF, word embeddings.
- Techniques: Sentiment analysis, topic modeling, named entity recognition (NER).

### 8. Data Visualization
- Purpose: Communicating insights through graphical representations.
- Tools: Matplotlib, Seaborn, Plotly (Python), ggplot2, Shiny (R), Tableau.
- Techniques: Bar charts, line graphs, heatmaps, interactive dashboards.

### 9. Big Data Technologies
- Purpose: Handling and analyzing large volumes of data.
- Technologies: Hadoop, Spark.
- Core Concepts: Distributed computing, MapReduce, parallel processing.

### 10. Databases
- Purpose: Storing and retrieving data efficiently.
- Types: SQL databases (MySQL, PostgreSQL), NoSQL databases (MongoDB, Cassandra).
- Core Concepts: Querying, indexing, normalization, transactions.

### 11. Time Series Analysis
- Purpose: Analyzing data points collected or recorded at specific time intervals.
- Core Concepts: Trend analysis, seasonal decomposition, ARIMA models, exponential smoothing.

### 12. Model Deployment and Productionization
- Purpose: Integrating machine learning models into production environments.
- Techniques: API development, containerization (Docker), model serving (Flask, FastAPI).
- Tools: MLflow, TensorFlow Serving, Kubernetes.

### 13. Data Ethics and Privacy
- Purpose: Ensuring ethical use and privacy of data.
- Core Concepts: Bias in data, ethical considerations, data anonymization, GDPR compliance.

### 14. Business Acumen
- Purpose: Aligning data science projects with business goals.
- Core Concepts: Understanding key performance indicators (KPIs), domain knowledge, stakeholder communication.

### 15. Collaboration and Version Control
- Purpose: Managing code changes and collaborative work.
- Tools: Git, GitHub, GitLab.
- Practices: Version control, code reviews, collaborative development.

Best Data Science & Machine Learning Resources: https://topmate.io/coding/914624

ENJOY LEARNING 👍👍
👍31🤣1
☄️ What is an Artificial Neural Networks?

Artificial neural networks (ANN) give machines the ability to process data similar to the human brain and make decisions or take actions based on the data. While there’s still more to develop before machines have similar imaginations and reasoning power as humans, ANNs help machines complete and learn from the tasks they perform.
▶️Content:


001 What is Deep Learning
002 Plan of Attack
003 The Neuron

004 The Activation Function
005 How do Neural Networks work
006 How do Neural Networks learn
007 Gradient Descent
008 Stochastic Gradient Descent
009 Back propagation

👇👇👇👇
Deep Learning Complete Course
🤣1
6 Tips for Building a Robust Machine Learning Model

1. Understand the problem thoroughly before jumping into the model.
➝ Taking time to understand the problem helps build a solution aligned with business needs and goals.

2. Focus on feature engineering to improve accuracy.
➝ Well-engineered features make a big difference in model performance. Collaborating with data engineers on clean and well-structured data can simplify feature engineering.

3. Start simple, test assumptions, and iterate.
➝ Begin with straightforward models to test ideas quickly. Iteration and experimentation will lead to stronger results.

4. Keep track of versions for reproducibility. 
➝  Documenting versions of data and code helps maintain consistency, making it easier to reproduce results.

5. Regularly validate your model with new data.
➝ Models should be updated and validated as new data becomes available to avoid performance degradation.

6. Always prioritize interpretability alongside accuracy.
➝ Building interpretable models helps stakeholders understand and trust your results, making insights more actionable.

Like if you need similar content 😄👍
🤣1
Use this checklist to see if you’re truly JOB-READY. The more items you complete, the closer you are to landing your dream data science job! 😎

Check Your Skills with This Checklist!

Python:-
Master Python fundamentals
Understand Pandas for data manipulation
Learn data visualization with Matplotlib and Seaborn
Practice error handling and debugging

Statistics:-
Grasp probability theory
Know denoscriptive and inferential statistics
Learn statistical machine learning concepts

Exploratory Data Analysis (EDA):-
Perform data summarization
Work on data cleaning and transformation
Visualize data effectively

SQL:-
Understand the BIG 6 SQL statements
Practice joins and common table expressions (CTEs)
Use window functions
Learn to write stored procedures

Machine Learning:-
Master feature engineering
Understand regression and classification techniques
Learn clustering methods

Model Evaluation:-
Work with confusion matrices
Understand precision, recall, and F1-score
Practice cross-validation
Learn about overfitting and underfitting

Deep Learning:-
Get familiar with Convolutional Neural Networks (CNNs)
Understand transformers
Learn PyTorch or TensorFlow basics
Practice model training and optimization

Resume:-
Ensure your resume is ATS-friendly
Customize for the job denoscription
Use the STAR method to highlight achievements
Include a link to your portfolio

AI-Enabled Mindset:-
Develop Googling skills
Use AI tools like ChatGPT or Bard for learning
Commit to continuous learning
Hone problem-solving abilities

Communication:-
Practice presenting insights clearly
Write professional emails
Manage stakeholder communication
Utilize project management tools

LinkedIn:-
Have a good profile picture and banner
Get 10+ endorsed skills
Collect at least 3 recommendations
Link your portfolio in your profile

Portfolio:-
Include 4+ business-related projects
Showcase one project per tool you know
Create an insights desk
Prepare a video presentation

Like if you need similar content 😄👍
5👍3🤣1
Data Science Minimum: 10 Essential Skills You Need to Know to Start Doing Data Science
👍2🤣1
Skills for Data Scientists 👆
2👍1🤣1
🔗 Machine Learning libraries
🥰4🤣2
Creating a one-month data analytics roadmap requires a focused approach to cover essential concepts and skills. Here's a structured plan along with free resources:

🗓️Week 1: Foundation of Data Analytics

Day 1-2: Basics of Data Analytics
Resource: Khan Academy's Introduction to Statistics
Focus Areas: Understand denoscriptive statistics, types of data, and data distributions.

Day 3-4: Excel for Data Analysis
Resource: Microsoft Excel tutorials on YouTube or Excel Easy
Focus Areas: Learn essential Excel functions for data manipulation and analysis.

Day 5-7: Introduction to Python for Data Analysis
Resource: Codecademy's Python course or Google's Python Class
Focus Areas: Basic Python syntax, data structures, and libraries like NumPy and Pandas.

🗓️Week 2: Intermediate Data Analytics Skills

Day 8-10: Data Visualization
Resource: Data Visualization with Matplotlib and Seaborn tutorials
Focus Areas: Creating effective charts and graphs to communicate insights.

Day 11-12: Exploratory Data Analysis (EDA)
Resource: Towards Data Science articles on EDA techniques
Focus Areas: Techniques to summarize and explore datasets.

Day 13-14: SQL Fundamentals
Resource: Mode Analytics SQL Tutorial or SQLZoo
Focus Areas: Writing SQL queries for data manipulation.

🗓️Week 3: Advanced Techniques and Tools

Day 15-17: Machine Learning Basics
Resource: Andrew Ng's Machine Learning course on Coursera
Focus Areas: Understand key ML concepts like supervised learning and evaluation metrics.

Day 18-20: Data Cleaning and Preprocessing
Resource: Data Cleaning with Python by Packt
Focus Areas: Techniques to handle missing data, outliers, and normalization.

Day 21-22: Introduction to Big Data
Resource: Big Data University's courses on Hadoop and Spark
Focus Areas: Basics of distributed computing and big data technologies.


🗓️Week 4: Projects and Practice

Day 23-25: Real-World Data Analytics Projects
Resource: Kaggle datasets and competitions
Focus Areas: Apply learned skills to solve practical problems.

Day 26-28: Online Webinars and Community Engagement
Resource: Data Science meetups and webinars (Meetup.com, Eventbrite)
Focus Areas: Networking and learning from industry experts.


Day 29-30: Portfolio Building and Review
Activity: Create a GitHub repository showcasing projects and code
Focus Areas: Present projects and skills effectively for job applications.

👉Additional Resources:
Books: "Python for Data Analysis" by Wes McKinney, "Data Science from Scratch" by Joel Grus.
Online Platforms: DataSimplifier, Kaggle, Towards Data Science

Tailor this roadmap to your learning pace and adjust the resources based on your preferences. Consistent practice and hands-on projects are crucial for mastering data analytics within a month. Good luck!
2👍2🤣1
You don't need to spend several $𝟭𝟬𝟬𝟬𝘀 to learn Data Science.

Stanford University, Harvard University & Massachusetts Institute of Technology is providing free courses.💥

Here's 8 free Courses that'll teach you better than the paid ones:


1. CS50’s Introduction to Artificial Intelligence with Python (Harvard)

https://pll.harvard.edu/course/cs50s-introduction-artificial-intelligence-python

2. Data Science: Machine Learning (Harvard)

https://pll.harvard.edu/course/data-science-machine-learning

3. Artificial Intelligence (MIT)

https://lnkd.in/dG5BCPen

4. Introduction to Computational Thinking and Data Science (MIT)

https://lnkd.in/ddm5Ckk9

5. Machine Learning (MIT)

https://lnkd.in/dJEjStCw

6. Matrix Methods in Data Analysis, Signal Processing, and Machine Learning (MIT)

https://lnkd.in/dkpyt6qr

7. Statistical Learning (Stanford)

https://online.stanford.edu/courses/sohs-ystatslearning-statistical-learning

8. Mining Massive Data Sets (Stanford)

📍https://online.stanford.edu/courses/soe-ycs0007-mining-massive-data-sets

ENJOY LEARNING
👍42🤣1
“The Best Public Datasets for Machine Learning and Data Science” by Stacy Stanford

https://datasimplifier.com/best-data-analyst-projects-for-freshers/

https://toolbox.google.com/datasetsearch

https://www.kaggle.com/datasets

http://mlr.cs.umass.edu/ml/

https://www.visualdata.io/

https://guides.library.cmu.edu/machine-learning/datasets

https://www.data.gov/

https://nces.ed.gov/

https://www.ukdataservice.ac.uk/

https://datausa.io/

https://www.cs.toronto.edu/~delve/data/boston/bostonDetail.html

https://www.kaggle.com/xiuchengwang/python-dataset-download

https://www.quandl.com/

https://data.worldbank.org/

https://www.imf.org/en/Data

https://markets.ft.com/data/

https://trends.google.com/trends/?q=google&ctab=0&geo=all&date=all&sort=0

https://www.aeaweb.org/resources/data/us-macro-regional

http://xviewdataset.org/#dataset

http://labelme.csail.mit.edu/Release3.0/browserTools/php/dataset.php

http://image-net.org/

http://cocodataset.org/

http://visualgenome.org/

https://ai.googleblog.com/2016/09/introducing-open-images-dataset.html?m=1

http://vis-www.cs.umass.edu/lfw/

http://vision.stanford.edu/aditya86/ImageNetDogs/

http://web.mit.edu/torralba/www/indoor.html

http://www.cs.jhu.edu/~mdredze/datasets/sentiment/

http://ai.stanford.edu/~amaas/data/sentiment/

http://nlp.stanford.edu/sentiment/code.html

http://help.sentiment140.com/for-students/

https://www.kaggle.com/crowdflower/twitter-airline-sentiment

https://hotpotqa.github.io/

https://www.cs.cmu.edu/~./enron/

https://snap.stanford.edu/data/web-Amazon.html

https://aws.amazon.com/datasets/google-books-ngrams/

http://u.cs.biu.ac.il/~koppel/BlogCorpus.htm

https://code.google.com/archive/p/wiki-links/downloads

http://www.dt.fee.unicamp.br/~tiago/smsspamcollection/

https://www.yelp.com/dataset

https://news.1rj.ru/str/DataPortfolio/2

https://archive.ics.uci.edu/ml/datasets/Spambase

https://bdd-data.berkeley.edu/

http://apolloscape.auto/

https://archive.org/details/comma-dataset

https://www.cityscapes-dataset.com/

http://aplicaciones.cimat.mx/Personal/jbhayet/ccsad-dataset

http://www.vision.ee.ethz.ch/~timofter/traffic_signs/

http://cvrr.ucsd.edu/LISA/datasets.html

https://hci.iwr.uni-heidelberg.de/node/6132

http://www.lara.prd.fr/benchmarks/trafficlightsrecognition

http://computing.wpi.edu/dataset.html

https://mimic.physionet.org/

Best Telegram channels to get free coding & data science resources
https://news.1rj.ru/str/addlist/4q2PYC0pH_VjZDk5

Free Courses with Certificate:
https://news.1rj.ru/str/free4unow_backup
👍8🤣1