Data science/ML/AI – Telegram
Data science/ML/AI
13.3K subscribers
539 photos
2 videos
111 files
317 links
Data science and machine learning hub

Python, SQL, stats, ML, deep learning, projects, PDFs, roadmaps and AI resources.

For beginners, data scientists and ML engineers
👉 https://rebrand.ly/bigdatachannels

DMCA: @disclosure_bds
Contact: @mldatascientist
Download Telegram
📚 Data Science Riddle - Feature Engineering

A model's performance drops because some features have extreme outliers. What helps most?
Anonymous Quiz
18%
Label smoothing
43%
Robust scaling
18%
Bagging
21%
Increasing k-fold splits
4
🧵 Thread Series on:

Mastering Pandas for Data Manipulation!


Pandas is the go-to library for handling tabular data in Python. Whether you're analyzing sales, surveys, or logs, start every project the same way:

import pandas as pd

# Load CSV
df = pd.read_csv('sales_data.csv')

# Quick look
df.head()     # First 5 rows
df.info()     # Structure & data types
df.describe() # Basic stats


Next up 👉 Selecting Columns & Rows
4🔥3
Selecting Columns & Rows

Need specific columns or rows? Pandas makes selection intuitive and fast:

# Single column (Series)
df['name']

# Multiple columns (DataFrame)
df[['name', 'age', 'sales']]

# Row selection with .loc (label-based)
df.loc[0:5]                    # Rows 0 to 5
df.loc[df['sales'] > 1000]     # Conditional

# .iloc (position-based)
df.iloc[0:5, 1:4]              # Rows 0-4, columns 1-3


Next up 👉 Filtering and Querying
7
Filtering and Querying

Want to zoom in on specific data?

Filtering in Pandas is incredibly powerful. Check the code below:

# Multiple conditions
high_sales = df[(df['sales'] > 1000) & (df['region'] == 'West')]

# Using .query() – cleaner syntax!
high_performers = df.query("sales > 1000 and region == 'West'")

# Find missing values
df[df['email'].isna()]

# Contains substring
df[df['product'].str.contains('Pro', case=False)]


Next up 👉 Adding and Removing Columns
5
Adding and Removing Columns

DataFrames are flexible! Easily create new columns or remove unnecessary ones:

# Add new column
df['revenue'] = df['sales'] * df['price']

# From existing columns
df['full_name'] = df['first_name'] + ' ' + df['last_name']

# Drop columns
df.drop(columns=['temp_col'], inplace=True)

# Or create a new DF without modifying original
clean_df = df.drop(columns=['old_col1', 'old_col2'])


Next up 👉 Dealing with Missing Values
7
Dealing with Missing Values

Real-world data is messy, missing values are common.

Here's how to handle them cleanly:

# Check for nulls
df.isnull().sum()

# Drop rows with any missing values
df_clean = df.dropna()

# Fill missing values
df['age'].fillna(df['age'].median(), inplace=True)
df['category'].fillna('Unknown', inplace=True)

# Forward or backward fill (great for time series)
df['value'].ffill()


Next up 👉 Using GroupBy
8
Using GroupBy

GroupBy is where Pandas shines brightest. It summarizes data by categories in one line.

# Total sales by region
df.groupby('region')['sales'].sum()

# Multiple aggregations
df.groupby('region').agg({
    'sales': 'sum',
    'customer_id': 'nunique',
    'order_date': 'max'
})

# Group by multiple columns
df.groupby(['region', 'product'])['sales'].mean()


Next up 👉 Sorting and Ranking
3
📚 Data Science Riddle - Evaluation

You're measuring performance on a dataset with heavy class imbalance. What metric is most reliable?
Anonymous Quiz
20%
Accuracy
44%
F1 Score
19%
Precision
18%
AUC
Sorting and Ranking

Order matters! Sort your data to find top performers or trends:

# Sort by one column
df.sort_values('sales', ascending=False)

# Sort by multiple columns
df.sort_values(['region', 'sales'], ascending=[True, False])

# Reset index after sorting
df = df.sort_values('sales', ascending=False).reset_index(drop=True)

# Add rank
df['sales_rank'] = df['sales'].rank(ascending=False)


Next up 👉 Merging and Joining Data
3
Media is too big
VIEW IN TELEGRAM
OnSpace Mobile App builder: Build AI Apps in minutes

With OnSpace, you can build website or AI Mobile Apps by chatting with AI, and publish to PlayStore or AppStore.

🔥 What will you get:
🤖 Create app or website by chatting with AI;
🧠 Integrate with Any top AI power just by giving order (like Sora2, Nanobanan Pro & Gemini 3 Pro);
📦 Download APK,AAB file, publish to AppStore.
💳 Add payments and monetize like in-app-purchase and Stripe.
🔐 Functional login & signup.
🗄 Database + dashboard in minutes.
🎥 Full tutorial on YouTube and within 1 day customer service

🌐 Visit website:
👉 https://www.onspace.ai/?via=tg_bigdata

📲 Or Download app:
👉 https://onspace.onelink.me/za8S/h1jb6sb9?c=bigdata
4
Merging and Joining Data

Working with multiple datasets? Combine them just like SQL:

# Inner join (default)
merged = pd.merge(df_sales, df_customers, on='customer_id')

# Left join
pd.merge(df_sales, df_customers, on='customer_id', how='left')

# Concatenate vertically
all_data = pd.concat([df_2023, df_2024], ignore_index=True)

# Join on index
df1.join(df2, on='date')


This wraps up our Data Manipulation Using Pandas Series.

Hit ❤️ if you liked this series. It will help us tailor more content based on what you like.

👉Join @datascience_bds for more
Part of the @bigdataspecialist family
7
SQL Joins Explained Visually
4
📚 Data Science Riddle - Regularization

A linear model starts performing worse on unseen data right after its training loss keeps decreasing. Which fix is moat appropriate ?
Anonymous Quiz
10%
Increase epochs
59%
Add L2 penalty
16%
Shuffle data again
15%
Raise Learning rate
Vector Databases: Searching by Meaning, Not Keywords

Traditional databases retrieve exact matches.
Vector databases retrieve conceptual similarity.

They store high-dimensional embeddings(mathematical representations of meaning) and search by finding the closest vectors in that space. This is how modern systems power semantic search, personalized recommendations, and AI memory retrieval.

Instead of asking “Does this word appear?”, you ask:
👉 “Is this idea close to what I’m looking for?”

It’s a shift from storing text to storing understanding.
And it’s becoming the backbone of LLM-powered applications.
7
📚 Data Science Riddle - Data Quality

Your dataset's numeric features contain silently corrupted values. What detection method helps?
Anonymous Quiz
33%
Min-max scaling
30%
Range validation
10%
Learning rate warmup
28%
Dropout masks
Artificial Intelligence vs Machine Learning
5
Robotic Process Automation (RPA) Basics You Should Know 🤖⚙️

Robotic Process Automation (RPA) is a technology that uses software robots to automate repetitive, rule based digital tasks normally performed by humans.

🔹 1. What is RPA? 
RPA is a form of automation where software bots mimic human actions to perform structured and repetitive tasks across applications.

🔹 2. How RPA Works: 
→ Bot logs into applications 
→ Reads and processes data 
→ Applies predefined rules 
→ Performs actions like clicking, typing, copying 
→ Completes tasks without human intervention 

🔹 3. Common Use Cases: 
• Invoice processing 
• Data entry and migration 
• Payroll and HR operations 
• Customer support automation 
• Report generation 

🔹 4. Key Benefits of RPA: 
• Reduces manual work 
• Improves accuracy 
• Increases productivity 
• Works 24x7 
• Faster business processes 

🔹 5. Popular RPA Tools: 
• UiPath 
• Automation Anywhere 
• Blue Prism 
• Microsoft Power Automate 

🔹 6. RPA vs Traditional Automation: 
• RPA works at UI level 
• No need to change existing systems 
• Faster deployment 
• Lower development cost 

🔹 7. Industries Using RPA: 
• Banking and finance 
• Healthcare 
• Insurance 
• E commerce 
• Telecom 

🔹 8. Limitations of RPA: 
• Not suitable for unstructured data 
• Depends on application stability 
• Limited decision making ability 
• Breaks if UI changes 

🔹 9. RPA + AI (Intelligent Automation): 
• AI handles decision making 
• RPA handles execution 
• Enables automation of complex processes 

🔹 10. Future of RPA: 
• More intelligent bots 
• Integration with AI and ML 
• End to end process automation 
• Higher enterprise adoption 

💡 Learning RPA helps you understand how automation is transforming modern businesses.

💬 Tap ❤️ for more!
6
Pandas vs SQL: Most Common Operations Comparison
5👍1
AI Ethics Basics You Should Know 🧠⚖️

AI Ethics focuses on ensuring that artificial intelligence systems are developed and used in a responsible, fair, and transparent manner.

🔹 1. What is AI Ethics? 
AI Ethics is the study of moral principles and practices that guide the development, deployment, and use of AI technologies.

🔹 2. Why AI Ethics is Important: 
• AI systems impact millions of people 
• Prevents bias and discrimination 
• Ensures trust and accountability 
• Protects user privacy and rights 

🔹 3. Key Principles of AI Ethics: 
Fairness: Avoid bias and discrimination 
Transparency: AI decisions should be explainable 
Accountability: Humans must be responsible for AI outcomes 
Privacy: Protect user data and personal information 
Safety: AI should not cause harm 

🔹 4. Common Ethical Issues in AI: 
• Biased algorithms 
• Data privacy violations 
• Surveillance misuse 
• Job displacement due to automation 
• Misinformation and deepfakes 

🔹 5. Real World Use Cases: 
• Fair hiring systems 
• Ethical facial recognition 
• Responsible healthcare AI 
• Bias detection in financial systems 

🔹 6. Examples of AI Bias: 
• Gender bias in resume screening 
• Racial bias in face recognition 
• Language bias in NLP models 

🔹 7. How to Build Ethical AI: 
• Use diverse and representative datasets 
• Regularly audit models for bias 
• Maintain human oversight 
• Clearly document AI decisions 

🔹 8. AI Ethics vs AI Governance: 
• AI Ethics focuses on moral values 
• AI Governance focuses on rules and regulations 
• Both work together for responsible AI 

🔹 9. Who is Responsible for AI Ethics? 
• Developers 
• Companies 
• Governments 
• Researchers 
• End users 

🔹 10. Future of AI Ethics: 
• Stronger regulations 
• Ethical AI certifications 
• More transparent AI systems 
• Human centered AI development 

💡 Learning AI Ethics is essential for building trustworthy and responsible AI systems.

💬 Tap ❤️ for more!
4
Feature Leakage: When Your Model Quietly Cheats 🫠

Feature leakage is one of the most dangerous failures in machine learning because your model looks excellent on paper. Accuracy jumps, losses drop, cross-validation smiles at you… and yet the model is learning information it should never have access to.

Leakage hides in subtle places; columns updated after an event happens, IDs that encode outcome patterns, or features computed using future timestamps. Nothing looks suspicious, but the model is essentially borrowing tomorrow’s truth to predict today.

The only real defense is time awareness. Before allowing any feature into training, ask:

Would this value truly exist at the moment of prediction?


If the answer is no, the model isn’t learning. It’s cheating.
3
Mastering SQL
7