Ayush

1.6K posts

Ayush banner
Ayush

Ayush

@TensorThrottleX

Crafting objective proof from vast and ambiguous datasets. Validating a chosen path to enable confident, decisive action. #100DaysOfML

nowhere Tham gia Nisan 2025
535 Đang theo dõi478 Người theo dõi
Tweet ghim
Ayush
Ayush@TensorThrottleX·
Embarking on a 100-Day Challenge. Starting from 30th December. This time, it's fundamentally different. We @TensorThrottleX, @BinaryBlaze16, @CodeAyushD committing to unflinching transparency, no curated highlights, no polished outcomes. Only the raw work itself.
Ayush tweet media
English
2
0
5
1.9K
Ayush
Ayush@TensorThrottleX·
Day 311 : DataScience Journey Random Forest is essentially an improved version of Decision Trees designed to solve their biggest weakness : overfitting. A single decision tree tends to memorize the training data, which means it has low bias but high variance and performs poorly
Ayush tweet mediaAyush tweet media
English
1
0
1
1K
Ayush
Ayush@TensorThrottleX·
Day 310 : DataScience Journey Worked through a full Titanic pipeline from messy real-world data to a functioning Random Forest model and the interesting part wasn’t just the model, but the small practical issues along the way.Handling missing values
Ayush tweet mediaAyush tweet mediaAyush tweet mediaAyush tweet media
English
1
0
4
4.3K
Ayush
Ayush@TensorThrottleX·
Day 309 : DataScience Journey After the data, The main effort went into preprocessing: removing irrelevant features, handling missing values using median and mode focused on understanding its structure identifying missing values, checking feature types, and observing how the
Ayush tweet mediaAyush tweet mediaAyush tweet mediaAyush tweet media
English
1
0
5
4.7K
Ayush
Ayush@TensorThrottleX·
Day 308 : DataScience Journey While training models using bagging (like Random Forest), not every data point is used in building each tree. On average, about one-third of the data remains unused for a given tree these are called out-of-bag samples.
Ayush tweet mediaAyush tweet mediaAyush tweet media
English
1
0
5
5.8K
Ayush
Ayush@TensorThrottleX·
Day 307 : DataScience Journey So a single decision tree… it tends to overfit pretty easily and can give unstable results. Bagging basically fixes this by training a bunch of trees on slightly different parts of the data and then combining their outputs. Since each tree sees a
Ayush tweet mediaAyush tweet media
English
1
0
3
175
Ayush
Ayush@TensorThrottleX·
Day 306 : DataScience Journey Gradient Boosting works by building models step by step, where each new tree tries to fix the mistakes made by the previous ones. Initially, a simple model is trained on the data.Then, we calculate the residuals (errors)
Ayush tweet mediaAyush tweet media
English
2
0
5
1.3K
Ayush
Ayush@TensorThrottleX·
Day 305 : DataScience Journey Instead of depending on a single algorithm, we combine multiple models like Logistic Regression, Random Forest, and SVM. Each one has its own way of making mistakes, so when we combine them using a Voting Classifier, the final result becomes more
Ayush tweet mediaAyush tweet media
English
2
0
6
3.9K
Ayush
Ayush@TensorThrottleX·
Day 304 : DataScience Journey Bagging and Pasting in Scikit-Learn are powerful ensemble techniques that improve model performance by reducing variance and enhancing generalization. Instead of relying on a single model, they train multiple models (typically Decision Trees) on diff
Ayush tweet mediaAyush tweet media
English
1
0
4
1.4K
Ayush
Ayush@TensorThrottleX·
Day 303 : DataScience Journey Random Patches and Random Subspaces are techniques used to increase diversity in ensemble models. Instead of training each model on the full dataset, we randomly sample either data points, features, or both. When both samples and features are random
Ayush tweet mediaAyush tweet media
English
2
0
3
3.5K