View on GitHub

datasetGAN2_release

datasetGAN2_release

Enhanced generation of automatically labeled image segmentation datasets by advanced style interpreter deep architectures

[Pattern Recognition Letters 2025] [paper]

Experiment framework

Proposed Architecture

Main results

Requirements

This code is tested with:

Replicate the environment

The experiments were conducted using a hybrid environment to solve dependency issues. To reproduce it, first create and acivate a Conda environment as follows:

conda env create -f environment_conda.yml
conda activate edgn_c

Then, create a virtual environment to manage the remaining libraries with pip as follows:

python3 -m venv edgan_p
source edgan_p/bin/activate
python3 -m pip install -r requirements_edgan_p.txt

Download the datasets

Important: our code requires that the images inside datasetGAN2_release/datasetGAN2/dataset_release/annotation/training_data/<dataset_name>/ are in PNG format to work.

1. Dropout Architecture experiments

cd datasetGAN2_release/datasetGAN2

1.1. Interpreter 4-fold cross-validation

python interpreter_cross_validation.py --n_processes <number_of_GPUs> --fold <fold_index> --exp experiments/<experiment_name>.json --classifier_param experiments/<dropout_configuration>.json --exp_dir model_dir/<dataset_folder>/<experiment_folder>

If you are running the experiments for the first time, it is recommended to execute the code above for only one fold and wait for the execution to conclude. Then, the remaining three fold could be executed in parallel. This recommendation is because in the first execution the feature maps will be stored in the disk. After the feature maps are already in disk, the code will not generate them again.

Important: The fold index must be a number within [0, 1, 2, 3]. All folds must be executed to perform the 4-fold cross-validation.

Example:

python interpreter_cross_validation.py --n_processes 1 --fold 0 --exp experiments/face_34.json --classifier_param experiments/nn_params_dropout_0604.json --exp_dir model_dir/face_34/dropout_60_40

2. Proposed (Convolutional) Architecture experiments

2.1. Interpreter 4-fold cross-validation

python train_conv_interpreter_opt.py --exp experiments/<experiment_name>.json  --n_gpus <number_of_GPUs> --fold_run <fold_index> --seed <seed_of_the_network> --float32 True --crossvalidate True

Example:

python train_conv_interpreter_opt.py --exp experiments/bedroom_10_conv.json  --n_gpus 1 --fold_run 0 --seed 0 --float32 True --crossvalidate True

Important:

2.2. Generate qualitative results

The results are generated for each network for each epoch.

python train_conv_interpreter_opt.py --exp experiments/<experiment_name>.json  --results_epoch <epoch_wth_the_best_results> --n_gpus <number_of_GPUs> --fold_run -1 --float32 True --generate_qualitative_results True --output_dir <path_to_save_the_results>

Example:

python train_conv_interpreter_opt.py --exp experiments/cat_16_conv.json  --results_epoch 1 --n_gpus 1 --fold_run -1 --float32 True --generate_qualitative_results True --output_dir /your/local/path/

Disclaimer

Due to inherent non-determinism in training (e.g., random seed initialization, hardware variability, number of GPUs), results may slightly vary on reruns.

Citations

Please include cite us if you use our data or code:

Pending

Acknowledgments

This work is partially supported by the Ministry of Science and Innovation of Spain [grant number PID2022-136764OA-I00], project name Automated Detection of Non Lesional Focal Epilepsy by Probabilistic Diffusion Deep Neural Models. It is also partially supported by the Autonomous Government of Andalusia (Spain) under project UMA18-FEDERJA-084, project name Detection of anomalous behavior agents by deep learning in low-cost video surveillance intelligent systems. All of them include funds from the European Regional Development Fund (ERDF). The authors thankfully acknowledge the computer resources, technical expertise, and assistance provided by the SCBI (Supercomputing and Bioinformatics) center of the University of Málaga. The authors acknowledge the funding from the Universidad de Málaga. No conflict of interest has been declared by the authors.

License

For any code dependency related to Stylegan, the license is under the Creative Commons BY-NC 4.0 license by NVIDIA Corporation. To view a copy of this license, visit LICENSE.

The code dependency related to DatasetGAN is under the MIT license. See LICENSE for additional details.

For the datasets of DatasetGAN, the license is under the Creative Commons BY-NC 4.0 license by NVIDIA Corporation. You can use, redistribute, and adapt the material for non-commercial purposes, as long as you give appropriate credit by citing our paper and indicating any changes that you’ve made.