@@ -23,27 +23,30 @@ If you are interested in the new nomenclature of a smaller number of classes (Bi
We provide code and model weights for the following deep learning models that have been pre-trained on BigEarthNet with the original Level-3 class nomenclature of CLC 2018 for scene classification:
The results provided in the [BigEarthNet paper](http://bigearth.net/static/documents/BigEarthNet_IGARSS_2019.pdf) are different from those given above due to the selection of different train, validation and test sets.
The results provided in the [BigEarthNet paper](http://bigearth.net/static/documents/BigEarthNet_IGARSS_2019.pdf) are different from those obtained by the models given above due to the selection of different train, validation and test sets.
The TensorFlow code for these models can be found [here](https://gitlab.tu-berlin.de/rsim/bigearthnet-models-tf).
The PyTorch code for these models can be found [here](https://gitlab.tu-berlin.de/rsim/bigearthnet-models-pytorch).
# Generation of Training/Test/Validation Splits
After downloading the raw images from https://www.bigearth.net, they need to be prepared for your ML application. We provide the script `prep_splits.py` for this purpose. It generates consumable data files (i.e., TFRecord) for training, validation and test splits which are suitable to use with TensorFlow. Suggested splits can be found with corresponding csv files under `splits` folder. The following command line arguments for `prep_splits.py` can be specified:
After downloading the raw images from https://www.bigearth.net, they need to be prepared for your ML application. We provide the script `prep_splits.py` for this purpose. It generates consumable data files (i.e., TFRecord) for training, validation and test splits which are suitable to use with TensorFlow or PyTorch. Suggested splits can be found with corresponding csv files under `splits` folder. The following command line arguments for `prep_splits.py` can be specified:
*`-r` or `--root_folder`: The root folder containing the raw images you have previously downloaded.
*`-o` or `--out_folder`: The output folder where the resulting files will be created.
*`-n` or `--splits`: A list of CSV files each of which contains the patch names of corresponding split.
*`-l` or `--library`: A flag to indicate for which ML library data files will be prepared: TensorFlow or PyTorch.
To run the script, either the GDAL or the rasterio package should be installed. The TensorFlow package should also be installed. The script is tested with Python 2.7, TensorFlow 1.3 and Ubuntu 16.04.
To run the script, either the GDAL or the rasterio package should be installed. The TensorFlow package should also be installed. The script is tested with Python 2.7, TensorFlow 1.3, PyTorch 1.2 and Ubuntu 16.04.
**Note**: BigEarthNet patches with high density snow, cloud and cloud shadow are not included in the training, test and validation sets constructed by the provided scripts (see the list of patches with seasonal snow [here](http://bigearth.net/static/documents/patches_with_seasonal_snow.csv) and that of cloud and cloud shadow [here](http://bigearth.net/static/documents/patches_with_cloud_and_shadow.csv)).
The BigEarthNet Archive is licensed under the **Community Data License Agreement – Permissive, Version 1.0** ([Text](https://cdla.io/permissive-1-0/)).