This repository contains code for a semantic segmentation model trained on the Cityscapes dataset. The model is designed to be used for self-driving cars to identify the different objects in urban environments.
To install the necessary packages, run the following command:
pip install -r requirements.txt
To use the segmentation model, run the following command:
python segment.py <image_path>
The model will segment the image and output an image with the different objects in the image highlighted with different colors.
The segmentation model is a U-Net architecture that was trained on the Cityscapes dataset. The model is able to segment various objects in urban environments, such as roads, buildings, and vehicles.
The trained model can be downloaded from the following link:
Download the file and place it in the same directory as segment.py
.
The dataset used to train the model is the Cityscapes dataset. The dataset contains images of urban environments with pixel-level annotations for 30 different object classes. The model was trained on 9 categories.
Contributions to this project are welcome. If you find a bug or have a feature request, please open an issue. If you would like to contribute code, please open a pull request.
This project is licensed under the MIT License.