Open
Conversation
This file contains hidden or bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Add this suggestion to a batch that can be applied as a single commit.This suggestion is invalid because no changes were made to the code.Suggestions cannot be applied while the pull request is closed.Suggestions cannot be applied while viewing a subset of changes.Only one suggestion per line can be applied in a batch.Add this suggestion to a batch that can be applied as a single commit.Applying suggestions on deleted lines is not supported.You must change the existing code in this line in order to create a valid suggestion.Outdated suggestions cannot be applied.This suggestion has been applied or marked resolved.Suggestions cannot be applied from pending reviews.Suggestions cannot be applied on multi-line comments.Suggestions cannot be applied while the pull request is queued to merge.Suggestion cannot be applied right now. Please check back later.
ssaurav@hawk.iit.edu (FNU Saurav) - A20536122
psavant@hawk.iit.edu (Pallavi Savant) - A20540976
ssinha20@hawk.iit.edu(Satwik Sinha)-A20547790
akutre@hawk.iit.edu (Aditya Ramchandra Kutre)-A20544809
tyerra@hawk.iit.edu(Tejaswi Yerra) -A20545536
MODEL SELECTION:
This project implements two model selection techniques:
k-Fold Cross-Validation:
Evaluates a machine learning model by splitting the dataset into ( k ) folds and using each fold as a validation set while training on the remaining ( k-1 ) folds.
The average loss (e.g., Mean Squared Error) across all folds is calculated to estimate the model’s predictive performance.
Bootstrapping:
Evaluates a model by generating multiple bootstrap samples (random sampling with replacement) from the dataset and using the out-of-bag (OOB) samples for validation.
The average error across all bootstrap iterations is computed to measure model performance.
Both methods are implemented for general-purpose models that provide fit() and predict() methods.