Skip to content

Commit

Permalink
Merge pull request #1516 from RB137/patch-2
Browse files Browse the repository at this point in the history
Added README.md file inside Advanced-house-prediction model section
  • Loading branch information
sanjay-kv authored Oct 22, 2024
2 parents 76607ce + fe462e4 commit 78589e0
Showing 1 changed file with 53 additions and 0 deletions.
53 changes: 53 additions & 0 deletions Prediction Models/Advanced House Price Predictions/README.md
Original file line number Diff line number Diff line change
@@ -0,0 +1,53 @@
# Advanced House Price Prediction Model

This project focuses on predicting house prices using advanced machine learning techniques. The model is trained on a dataset containing various features related to houses (e.g., square footage, number of rooms, location) to estimate the selling price.

## Features

- *Preprocessing*: Handles missing values, outliers, and categorical data.
- *Feature Engineering*: Adds relevant new features to improve model performance.
- *Modeling*: Uses multiple models including Linear Regression, Decision Trees, Random Forest, and Gradient Boosting.
- *Evaluation*: Assesses the model's performance using metrics like RMSE (Root Mean Squared Error) and R² (coefficient of determination).

## Dataset

The dataset contains features such as:

- *Lot Area*: Size of the lot in square feet
- *Year Built*: Year the house was constructed
- *Overall Quality*: Material and finish quality
- *Total Rooms*: Number of rooms excluding bathrooms
- *Neighborhood*: The physical location of the property
- *Sale Price*: The target variable for prediction

## Setup

1. Clone the repository:
bash
git clone https://github.com/recodehive/house-price-prediction.git
2. Navigate to the project directory:
bash
cd house-price-prediction
3. Install dependencies:
bash
pip install -r requirements.txt
4. Run the model:
bash
python main.py

## Model Training

The model uses the following steps:

1. Data Preprocessing: Imputing missing values, scaling numerical features, encoding categorical variables.
2. Feature Selection: Selecting the most important features to reduce overfitting.
3. Training: Training multiple models to compare their performances.
3. Evaluation: Checking accuracy using cross-validation and other metrics.

## Results

The model achieved a RMSE of X and R² of Y on the test dataset. Gradient Boosting performed the best, followed by Random Forest.

## Contribution

Feel free to fork the repository and contribute! Submit pull requests with clear descriptions of the changes made.

0 comments on commit 78589e0

Please sign in to comment.