Updated Readme

This commit is contained in:
devoalda 2024-03-17 22:42:14 +08:00
parent 59a90be0f5
commit bb411245ee
2 changed files with 73 additions and 10 deletions

BIN
CNN.keras

Binary file not shown.

View File

@ -1,18 +1,81 @@
# CSC 3105 Project # README
This project involves data preprocessing and dimensionality reduction using Principal Component Analysis (PCA). ## Project Overview
## File Structure This Python notebook is a comprehensive machine learning project that performs data preprocessing, feature engineering, model training, and model evaluation. It uses various machine learning algorithms such as Random Forest, Support Vector Machine (SVM), Logistic Regression, Gradient Boosting, K-Nearest Neighbors (KNN), Naive Bayes, K-Means Clustering, Convolutional Neural Network (CNN), and Multi-Layer Perceptron (MLP) for classification tasks. The project also includes data visualization and model performance evaluation.
The project consists of the following files and directories: ## Getting Started
- `Project.ipynb`: This is the main Jupyter notebook file where all the data processing. ### Prerequisites
- `./UWB-LOS-NLOS-Data-Set/dataset`: This directory contains the dataset used in the project.
# Installation and Setup - Python 3.7 or later
- pip (Python Package Installer)
### Virtual Environment Setup
It's recommended to create a virtual environment to keep the dependencies required by this project separate from your other Python projects.
For Unix or MacOS, run:
```bash
python3 -m venv env
```
For Windows, run:
```bash
py -m venv env
```
This will create a new virtual environment in a folder named `env`.
To activate the virtual environment, on Unix or MacOS, run:
```bash
source env/bin/activate
```
On Windows, run:
```bash
.\env\Scripts\activate
```
### Dependencies Installation
After activating the virtual environment, you can install the necessary dependencies by running:
```bash ```bash
git clone <REPO_URL>
cd <REPO_NAME>
pip install -r requirements.txt pip install -r requirements.txt
``` ```
This command installs all the necessary libraries used in this project, including:
- pandas
- numpy
- matplotlib
- seaborn
- scikit-learn
- tensorflow
## Usage
After setting up the environment and installing the dependencies, you can start the Jupyter notebook by running:
```bash
jupyter notebook
```
Then, navigate to the notebook file (.ipynb) in the Jupyter notebook web interface and open it.
## Notebook Structure
The notebook is structured as follows:
1. **Data Loading and Cleaning**: The notebook starts by loading the data from a pickle file. If the file doesn't exist, it loads the data from a dataset directory, cleans it, and saves it to a pickle file.
2. **Data Visualization**: The notebook visualizes the data using various plots.
3. **Model Training and Evaluation**: The notebook trains various machine learning models on the data and evaluates their performance. It also saves the trained models to a file for future use.
4. **Model Visualization**: The notebook visualizes the performance of the trained models using various plots and metrics.