Before you go, check out these stories!

Hackernoon logoRethNet Model: Object-by-Object Learning for Detecting Facial Skin Problems by@shohbek

RethNet Model: Object-by-Object Learning for Detecting Facial Skin Problems

Author profile picture

@shohbekShohrukh Bekmirzaev

In August 2019, a group of researchers from lululab Inc propose the state-of-the-art concept using a semantic segmentation method to detect the most common facial skin problems accurately. The work is accepted to ICCV 2019 Workshop.


1. The Concept of Object-by-Object Learning

2. Dataset

3. REthinker Blocks

4. REthinker Modules VS SENet Modules

5. RethNeT

6. Results

1. The Concept of Object-by-Object Learning

In fact, skin lesion objects have visual relations between each other where it helps easily humanity to judge what type of skin lesions they are.

Precisely, there are the region-region, object-region and object-object interactions between some skin lesions where the detection of an object class helps to detect another by their co-occurrence interactions (e.g., wrinkle and age spots, or papule and whitehead, etc.). The detection decisions about individual skin lesions can be switched dynamically through contextual relations among object classes. This cognitive process is denoted as object-by-object decision-making.

Simply, object-by-object learning is observed when an object can be identified by looking at other objects.

 2. Dataset

The researches prepared a dataset called ”Multi-type Skin Lesion Labeled Database” (MSLD) with pixel-wise labeling of frontal face images. Reported that the designing of MSLD is unique in ML community where it is not available such a dataset with the labeling of multi-type skin lesions of facial images.

412 images have been annotated with the labeling of 11 common types of facial skin lesions and 6 additional classes. The skin lesions are whitehead, papule, pustule, freckle, age spots, PIH, flush, seborrheic, dermatitis, wrinkle and black dot. The additional classes are normal skin, hair, eyes/mouth/eyebrow, glasses, mask/scarf and background. Reported that they do not disclose their (MSLD) dataset as the user’s privacy is taken under the responsibility.

3. The REthinker Block

Proposed a REthinker module based on the SENet module and locally constructed convLSTM/conv3D unit to increase the network’s sensitivity in local and global contextual representation that helps to capture ambiguously appeared objects and co-occurrence interactions between object classes.

4. REthinker Modules VS SENet Modules

REthinker modules consist of the SENet module and locally constructed convLSTM/conv3D layers as an one-shot attention mechanism, which are both responsible for extracting contextual relations from features.

Precisely, as the global pooling of SENet module aggregates the global spatial information, the SE module passes more embedded higher-level contextual information across large neighborhoods of each feature map. Whereas, the locally constructed convLSTM/conv3D layers encode lower-level contextual information across local neighborhoods elements of fragmented feature map (patches) while further take spatial correlation into consideration distributively over patches.

def image_to_patches(image, patch_size=[1,4,4,1]):
    Batch_size, H, W, C = image.get_shape()
    patches = tf.image.extract_image_patches(image, patch_size, patch_size, [1, 1, 1, 1], 'VALID')
    patches = tf.reshape(patches, [Batch_size ,patch_size[1]*patch_size[1], patch_size[1], patch_size[1], C])
    return patches

def patches_to_image(patches, image_shape):
    _, N, H, W, C = patches.get_shape()
    patches = tf.reshape(patches, [-1, image_shape[0], image_shape[1], C])
    rec_new = tf.space_to_depth(patches, H)
    rec_new = tf.reshape(rec_new, [1, image_shape[0], image_shape[1], C])
    return rec_new

def contructed_convLSTM(inputs, depth,kernel_size, rate):
    _, H, W, C = inputs.get_shape()
    net = image_to_patches(inputs, patch_size=[1,4,4,1])
    net = tf.keras.layers.ConvLSTM2D(depth, kernel_size=kernel_size, padding='same', dilation_rate=rate, return_sequences=True)(net)
    # net = tf.keras.layers.Conv3D(depth, kernel_size=kernel_size, padding='same', dilation_rate=rate)(net)
    net = patches_to_image(net, image_shape=[H, W])
    return net

def RethBlock(inputs, depth,kernel_size, rate, prefix=None):

    residual = inputs
    inputs = contructed_convLSTM(inputs,depth,kernel_size,rate=rate)
    _, H, W, C = inputs.get_shape()
    residual = tf.keras.layers.GlobalAveragePooling2D(name='Globalpooling_SeNet' + prefix)(residual)
    residual = tf.keras.layers.Dense(C.value // 8, activation='relu', name='fc_SeNet' + prefix + '_squ')(residual)
    residual = tf.keras.layers.Dense(C.value, activation='sigmoid', name='fc_seNet' + prefix + '_exc')(residual)
    residual = tf.keras.layers.Reshape([1, 1, C.value])(residual)
    outputs  = tf.keras.layers.Multiply(name='scale' + prefix)([residual, inputs])
    return outputs
Note that this is a fast implementation of REthinker blocks that support only when the input size has the root of a number such as 256x256, 64x64, 16x16, etc ...

5. RethNeT

An Encoder Search: In practice, proposed REthinker blocks are applicable in any standard CNNs. However, in this work, Xception is considered as the current state of the art network and powered with REthinker blocks. Furthermore, there are lightweight versions of RehNet based on MobileNet v2 and IGCV3. The use of the REthinker modules forces networks to capture the contextual relationships between object classes regardless of similar texture and ambiguous appearance they have.

A Decoder Search: In fact, the rich contextual information is a key to capturing ambiguously appeared objects and co-occurrence interactions between object classes where that is usually obtained in encoders. Therefore, the decoder path is not considered and just used decoder of DeepLabv3+ to recover object segmentation details of individual skin lesions.

RethNet: Simply investigated RethNet with the combining of the Xception module and REthinker modules. Xception is modified as follows:

  1. The REthinker module added after each Xception block without spatial loss of feature maps.
  2. The final block of the entry flow of Xception is removed.
  3. The patches size as 4x4 in each REthinker module is kept in order to ”see” future maps wider in ConvLSTM/conv3D with simply increasing time steps.
  4. The number of parameters is minimized in the middle flow and exit flow of Xception.
  5. The max-pooling operation is replaced by the depthwise separable convolutions with striding and the batch normalization and ReLU is applied after each 3 x 3 depthwise convolution of the Xception module.

6. Results

The inference results of RethNet seem pretty good if you check the title image of the blog post that is the inference result of RethNet in real test images. According to the comparison results, it shows significant improvements in the facial skin lesion detection task of the MSLD dataset, representing a 15.34% improvement over Deeplab v3+ (MIoU of 64.12%).

If you find the work useful for your research, please consider citing our paper:

    author = {Bekmirzaev, Shohrukh and Oh, Seoyoung and Yo, Sangwook},
    title = {RethNet: Object-by-Object Learning for Detecting Facial Skin Problems},
    booktitle = {The IEEE International Conference on Computer Vision (ICCV) Workshops},
    month = {Oct},
    year = {2019}


Join Hacker Noon

Create your free account to unlock your custom reading experience.