4 KiB
Creating your own dataset
Overview of the pipeline
The training script initiates a bunch of variables and classes before starting the training on a dataset. Here are the initialization steps:
-
Create an instance of the
Config
class. This instance will hold all the parameters defining the network. -
Create an instance of your dataset class. This instance will handle the data, and the input pipeline. This is the class you have to implement to train our network on your own data.
-
Load the input point cloud in memory. Most datasets will fit in a 32GB RAM computer. If you don't have enough memory for your dataset, you will have to redesign the input pipeline.
-
Initialize the tensorflow input pipeline, which is a
tf.dataset
object that will create and feed the input batches to the network. -
Create an instance of the network model class. This class contains the tensorflow operations defining the network.
-
Create an instance of our generic
ModelTrainer
class. This class handles the training of the model
Then the training can start.
The dataset class
This class has several roles. First this is where you define your dataset parameters (class names, data path, nature of the data...). Then this class will hold the point clouds loaded in memory. Eventually, it also defines the Tensorflow input pipeline. For efficiency, our implementation uses a parallel input queue, feeding batches to the network.
Here we give you a description of each essential method that need to be implemented in your new dataset class. For more details, follow the implementation of the current datasets, which contains a lot of indications as comments.
-
The __init__ method: Here you have to define the parameters of your dataset. Notice that your dataset class has to be a child of the common
Dataset
class, where generic methods are implemented. Their are a few thing that has to be defined here:- The labels: define a dictionary
self.label_to_names
, call theself.init_labels()
method, and define which label should be ignored inself.ignored_labels
. - The network model: the type of model that will be used on this dataset ("classification", "segmentation", "multi_segmentation" or "cloud_segmentation").
- The number of CPU threads used in the parallel input queue.
- Data paths and splits: you can manage your data as you wish, these variables are only used in methods that you will implement, so you do not have to follow exactly the notations of the other dataset classes.
- The labels: define a dictionary
-
The load_subsampled_clouds method: Here you load your data in memory. Depending on your dataset (if this is a classification or segmentation task, 3D scenes or 3D models) you will not have to load the same variables. Just follow the implementation of the existing datasets.
-
The get_batch_gen method: This method should return a python generator. This will be the base generator for the
tf.dataset
object. It is called in the genericself.init_input_pipeline
orself.init_test_input_pipeline
methods. Along with the generator, it also has to return the generated types and shapes. You can redesign the generators or used the ones we implemented. The generator returns np.arrays, but from this point of the pipeline, they will be converted to tensorflow tensors. -
The get_tf_mapping method: This method return a mapping function that takes the generated batches and creates all the variables for the network. Remember that from this point we are defining a tensorflow graph of operations. There is not much to implement here as most of the work is done by two generic function
self.tf_augment_input
andself.tf_xxxxxxxxx_inputs
where xxxxxxxxx can be "classification" of "segmentation" depending on the task. The only important thing to do here is to define the features that will be fed to the network.
The training script and configuration class
In the training script you have to create a class that inherits from the Config
class. This is where you will define
all the network parameters by overwriting the attributes