Data Intelligence Just Accepted MS.

Data Intelligence Just Accepted MS.
https://doi.org/10.1162/dint_a_00204

Rezension

Submitted: November 18, 2022; Überarbeitet: Dezember 12, 2022; Accepted February 7, 2023

Zitat: Tao, Z.C, Lyu, S.F. A Survey on automatic delineation of radiotherapy target volume

based on machine learning. Datenintelligenz 5(2023). doi: https://doi.org/10.1162/dint_a_00204

A Survey on Automatic Delineation of Radiotherapy Target Volume based on

Machine Learning

Zhenchao Tao, 1, 2, 3, Shengfei Lyu*, 3

1. School of Data Science, University of Science and Technology of China, Hefei,

Anhui 230026, China

2. The First Affiliated Hospital of USTC, School of Life Sciences and Medicine,

University of Science and Technology of China, Hefei 230031, China

3. Nanyang Technological University, 639798, Singapur

* Korrespondierender Autor. Email: shengfei.lyu@ntu.edu.sg

Schlüsselwörter: Automatic delineation, Machine learning, Radiotherapy target volume,

Medical image matching, Cancer

Abstrakt

Radiotherapy is one of the main treatment methods for cancer, and the delineation

of the radiotherapy target area is the basis and premise of precise treatment. Artificial

intelligence technology represented by machine learning has done a lot of research in

this area, improving the accuracy and efficiency of target delineation. This article will

l

D
Ö
w
N
Ö
A
D
e
D

F
R
Ö
M
H

T
T

P

:
/
/

D
ich
R
e
C
T
.

M

ich
T
.

e
D
u
D
N

/

ich

T
/

l

A
R
T
ich
C
e

P
D

F
/

D
Ö

ich
/

ich

T

/

.

/

1
0
1
1
6
2
D
N
_
A
_
0
0
2
0
4
2
1
2
9
4
4
2
D
N
_
A
_
0
0
2
0
4
P
D

T

.

/

ich

F

B
j
G
u
e
S
T

T

Ö
N
0
7
S
e
P
e
M
B
e
R
2
0
2
3

© 2023 Chinesische Akademie der Wissenschaft. Veröffentlicht unter einer Creative Commons Namensnennung 4.0 International (CC BY 4.0) Lizenz.

Data Intelligence Just Accepted MS.
https://doi.org/10.1162/dint_a_00204

review the applications and research of machine learning in medical image matching,

normal organ delineation and treatment target delineation according to the procudures

of doctors to delineate the target volume, and give an outlook on the development

prospects.

Einführung

To estimate the global burden of Cancer-based on the cancer and mortality

information provided by the International Agency for Research on Cancer in

GLOBOCAN 2020 [1], von 2020, Globally, there are an estimated 19.3 million new

cancer cases (18.1 million excluding non-melanoma skin cancer) and nearly 10 Million

cancer deaths (9.9 million excluding non-melanoma skin cancer). The global cancer

patients will be expected to reach 28.4 million cases by 2040, A 47% increase from 2020.

Malignant tumors will surpass all other chronic diseases and become thenumber one

killerthat threatens human life and health. Radiotherapy is one of the main treatments

for malignant tumors. Its principle is to use the high-energy ionizing radiation to kill

cells of tumors. About 60%-70% of tumor patients need to receive radiotherapy.

According to statistics, the current average progression-free survival rate of malignant

tumors is about 55%, of which radiotherapy contributes 40% of the tumor cure [2], Und

the therapeutic effect has been widely recognized in clinical practice. The rapid

development of artificial intelligence represented by machine learning can be applied

to all aspects of clinical practice of radiotherapy [3-6], making radiotherapy decision-

making more simplified, individualized and precise, and improving the automation of

the entire process of radiotherapy. The precise determination of radiotherapy target

volume is the basis and premise of precision radiotherapy. The automatic delineation

of radiotherapy target volume based on machine learning is essential in the research of

artificial intelligence in the field of radiotherapy application, which greatly improves

the efficiency and accuracy of target volume delineation [7]. This article will review

medical image matching, normal organ delineation and treatment target delineation.

1. Application of machine learning in radiotherapy

Kürzlich, with the development and progress of medical and computer technology,

radiotherapy has entered a new era of precision radiotherapy, and more and more

precision radiotherapy technologies have entered the practice of clinical tumor

l

D
Ö
w
N
Ö
A
D
e
D

F
R
Ö
M
H

T
T

P

:
/
/

D
ich
R
e
C
T
.

M

ich
T
.

e
D
u
D
N

/

ich

T
/

l

A
R
T
ich
C
e

P
D

F
/

D
Ö

ich
/

ich

T

/

/

.

1
0
1
1
6
2
D
N
_
A
_
0
0
2
0
4
2
1
2
9
4
4
2
D
N
_
A
_
0
0
2
0
4
P
D

T

/

.

ich

F

B
j
G
u
e
S
T

T

Ö
N
0
7
S
e
P
e
M
B
e
R
2
0
2
3

Data Intelligence Just Accepted MS.
https://doi.org/10.1162/dint_a_00204

treatment. Precision radiotherapy is playing an increasingly important role in improving

curative effect, delaying disease progression, improving prognosis and improving

Patienten’ quality of life [8-9].

In the 1930s, radiation technology has been used to treat tumor patients [10], und in

the 1960s with the widespread applications of medical linear accelerators [11]. Jedoch,

X-ray simulation localization is used for tumor localization during radiotherapy in this

Zeitraum. The doctor obtains the location of the tumor from the patient’s fluoroscopic

Bild, and marks the irradiation range on the patient’s body surface according to the

localization image, and performs treatment through the body surface projection field.

Due to the failure to clearly define the tumor and normal tissue, and the poor uniformity

of radiation dose distribution, it is easy to miss the tumor or normal tissue is irradiated

with a higher dose, resulting in a lower cure rate and higher complications. In 1959,

Takahashi et al [12] proposed the concept of three-dimensional conformal radiation

therapy (3D-CRT). The prototype is based on the three-dimensional morphological

structure of the tumor, using lead blocks to irradiate in multiple radiation directions

through the blocking part field, so that the shape of the irradiated area is the same as

that of the tumor target, while reducing the radiation dose received by the blocked area.

In the 1970s, the widespread application of computer systems and the emergence of

computed tomography (CT), magnetic resonance imaging (MRT) and other equipment

promoted radiotherapy to three-dimensional space, enabling 3D-CRT to be realized.

In recent years, three-dimensional digital precise radiotherapy technology has

gradually replaced traditional two-dimensional radiotherapy technology, and has

become an important development direction of tumor radiotherapy in the 21st century.

The three-dimensional digital precise radiotherapy technology focuses on precise

positioning and precise treatment, and performs conformal or intensity-modulated

radiotherapy at the three-dimensional level through dose segmentation, so that the

internal irradiation dose of the lesion in the target area is the largest, and the surrounding

normal tissue is the smallest, the irradiation dose is evenly distributed, and has the

advantages of high precision, high efficacy and low damage [13]. In addition to 3D-CRT,

the currently recognized precision radiotherapy techniques also include stereotactic

body radiotherapy (SBRT), intensity modulated radiotherapy (IMRT), and image

guided radiation therapy (IGRT), usw. The technique system of precise radiotherapy for

tumor is gradually perfection, and the treatment accuracy is increasingly improved.

At present, the steps of precise radiotherapy are to first obtain the anatomical

l

D
Ö
w
N
Ö
A
D
e
D

F
R
Ö
M
H

T
T

P

:
/
/

D
ich
R
e
C
T
.

M

ich
T
.

e
D
u
D
N

/

ich

T
/

l

A
R
T
ich
C
e

P
D

F
/

D
Ö

ich
/

ich

/

/

.

T

1
0
1
1
6
2
D
N
_
A
_
0
0
2
0
4
2
1
2
9
4
4
2
D
N
_
A
_
0
0
2
0
4
P
D

T

/

.

ich

F

B
j
G
u
e
S
T

T

Ö
N
0
7
S
e
P
e
M
B
e
R
2
0
2
3

Data Intelligence Just Accepted MS.
https://doi.org/10.1162/dint_a_00204

images of the patient on the treatment couch by simulated positioning, then manually

delineate the target area and organs at risk by the doctor, and then configure the

radiation dose, number of fields, field angle and other parameters can be used to

generate a radiotherapy plan suitable for the shape and dose of the tumor target. Endlich,

after the radiotherapy plan is verified and correct, the treatment can be carried out.

Among them, target delineation is the core work of radiotherapy physicians. Accurate

target delineation is the premise and crucial step of precise tumor radiotherapy. Der

quality of delineation has a great impact on the treatment effect of patients and the

occurrence of complications [14]. If the treatment target volume is too large, it will

increase the radiation dose received by the surrounding organs, thereby increasing the

probability of complications [15]. Umgekehrt, if the tumor area is not completely

covered, it will lead to insufficient doses to kill all cancer cells, greatly increasing the

possibility of recurrence after treatment [16].

Currently, the therapeutic target volume that needs to be manually delineated by

radiologists mainly includes the gross tumor volume (GTV) visible on the image; Die

clinical target volume (CTV) is delineated based on the knowledge of tumor pathology,

tumor invasion range, and lymph node metastasis pathway. Zusätzlich, the target area

of the organ at risk (OAR) within the irradiation range also needs to be accurately

delineated to avoid over-irradiation of the OARs, causing serious side effects and

complications of radiotherapy [17]. The above-mentioned delineation quality of the

therapeutic target volume and OARs completely depends on the professional

knowledge and experience of the doctor, and certain errors will occur. Darüber hinaus, diese

large-scale structures are delineated manually layer by layer for the radiologists, Und

the time cost is also very high. With the development of artificial intelligence

Technologie, deep learning methods based on the big data of radiotherapy patient images

can automatically delineate the therapeutic target area and OARs of patients. The speed

and accuracy are greatly improved, which helps to reduce the workload of doctors and

reduce manual delineation. uncertainty, further improving the precision of radiotherapy

[18-19].

As the main method in the field of artificial intelligence, machine learning can be

divided into supervised learning, unsupervised learning, and semi-supervised learning

which combines the two [20-22]. Specifically in the field of radiotherapy, supervised

learning-assisted radiotherapy is mainly used [23]. Combining multiple simple machine

learning models to obtain an ensemble learning model with better performance can

l

D
Ö
w
N
Ö
A
D
e
D

F
R
Ö
M
H

T
T

P

:
/
/

D
ich
R
e
C
T
.

M

ich
T
.

e
D
u
D
N

/

ich

T
/

l

A
R
T
ich
C
e

P
D

F
/

D
Ö

ich
/

ich

/

.

T

/

1
0
1
1
6
2
D
N
_
A
_
0
0
2
0
4
2
1
2
9
4
4
2
D
N
_
A
_
0
0
2
0
4
P
D

T

/

.

ich

F

B
j
G
u
e
S
T

T

Ö
N
0
7
S
e
P
e
M
B
e
R
2
0
2
3

Data Intelligence Just Accepted MS.
https://doi.org/10.1162/dint_a_00204

design a combination scheme for specific machine learning problems to get a better

solution [24]. Neural networks are a form of machine learning inspired by the way the

brain works, referencing the connection structure of neurons [25-27]. When the neural

network has many hidden layers, it is defined as a deep neural network. Deep learning

methods use deep neural networks to solve various classification and prediction

problems. Compared with traditional machine learning methods, deep learning methods

have the advantage of being able to automatically learn features in data and avoid

manual feature selection. A large amount of data accumulation and the improvement of

hardware computing power have made deep learning methods more and more applied

in the medical field, and they have shown better performance than traditional machine

learning methods [28-31].

2. Medical image registration based on machine learning

The electron density of CT images is linearly related to the density of the human

Körper, which can be directly used to calculate the radiation dose, and has become the

most commonly used radiotherapy positioning equipment. It has a good effect on bone

and lung tissue observation, while soft tissue MRI images have better observation

Effekte, and PET images can indicate areas with strong metabolism. daher, multi-

modal imaging registration is often used in clinical assessment of disease. Medical

image registration is to find the optimal spatial transformation between the source

image and the target image to match all the feature points or at least all the

corresponding points with diagnostic significance on the two images, and provide

doctors with more abundant clinical information. Common registration methods

include rigid registration and non-rigid registration.

2.1 Rigid registration

Rigid deformation can be described by a few transformation parameters. In the

field of radiotherapy, rigid registration is very common and highly accepted, Und

clinicians will fuse images of different modalities through this transformation to obtain

more information about areas of interest. The registration method is to align the two

images by finding the rotation-translation transformation matrix between the fixed

image and the moving image [32]. The methods used include linear transformations such

as translation and rotation, which can ensure that the overall structure or line parallelism

of the image remains unchanged after spatial transformation. Gleichzeitig, es hat

the advantages of simple calculation and low time complexity, and is suitable for

l

D
Ö
w
N
Ö
A
D
e
D

F
R
Ö
M
H

T
T

P

:
/
/

D
ich
R
e
C
T
.

M

ich
T
.

e
D
u
D
N

/

ich

T
/

l

A
R
T
ich
C
e

P
D

F
/

D
Ö

ich
/

ich

T

/

/

.

1
0
1
1
6
2
D
N
_
A
_
0
0
2
0
4
2
1
2
9
4
4
2
D
N
_
A
_
0
0
2
0
4
P
D

.

T

/

ich

F

B
j
G
u
e
S
T

T

Ö
N
0
7
S
e
P
e
M
B
e
R
2
0
2
3

Data Intelligence Just Accepted MS.
https://doi.org/10.1162/dint_a_00204

images with little deformation.

Rigid registration not only provides a prerequisite for further non-rigid registration

and saves the calculation time of image optimization iterations, but also can intuitively

display the anatomical structure differences between images between different

modalities, assisting doctors in accurate delineation. Traditional registration methods

include surface-based methods, point-based methods (usually based on anatomical

markers), and voxel-based methods [33]. Among them, voxel-based methods have been

widely used by virtue of the rapid development of computer technology. The goal of

this method is to obtain geometric transformation parameters by computing the

similarity between two input images without pre-extracting features [34]. Jedoch, diese

traditional registration methods often require iterative calculation of similarity

measures such as mean square error, mutual information and normalized mutual

Information, usw. Due to the non-convexity of similarity measures in parameter space,

the registration process is relatively expensive. sometimes with poor robustness [35].

Besides, other methods such as intensity-based feature selection algorithms perform

image registration by extracting image features corresponding to the intensity, Jedoch,

the extracted features are difficult to correspond well in anatomy [36].

2.2 Non-rigid registration

Since medical images are affected by factors such as imaging time, Bildgebung

Ausrüstung, and patient posture, it is difficult to spatially register multimodal images. In

addition, the internal tissue structure of the human body is complicated and has time-

varying characteristics. Zum Beispiel, the tissues and organs in the lung scan images will

move with the patient’s breathing. For the deformation of the images with large

differences in each direction, the rigid registration method cannot meet the requirements.

In this case, a non-rigid registration technology needs to be used, and the same parts of

different images are corresponding to each other by means of the spatial registration

deformation field. The entire registration process will also introduce different degrees

of registration errors due to the chosen optimization method.

Non-rigid transformation includes translation, rotation, scaling, and affine

transformation based on an affine matrix and other linear and nonlinear transformation

Formen. Compared with rigid transformation, it has better deformation accuracy, aber die

calculation speed is slower. Gu et al. [37] proposed a B-spline affine transformation

registration method, using affine transformation to replace the traditional displacement

of each B-spline control point, and using a two-way distance cost function to replace

l

D
Ö
w
N
Ö
A
D
e
D

F
R
Ö
M
H

T
T

P

:
/
/

D
ich
R
e
C
T
.

M

ich
T
.

e
D
u
D
N

/

ich

T
/

l

A
R
T
ich
C
e

P
D

F
/

D
Ö

ich
/

ich

/

T

.

/

1
0
1
1
6
2
D
N
_
A
_
0
0
2
0
4
2
1
2
9
4
4
2
D
N
_
A
_
0
0
2
0
4
P
D

.

/

T

ich

F

B
j
G
u
e
S
T

T

Ö
N
0
7
S
e
P
e
M
B
e
R
2
0
2
3

Data Intelligence Just Accepted MS.
https://doi.org/10.1162/dint_a_00204

the traditional one-way distance cost function to achieve bidirectional registration of

two images. Pradhan et al. [38] used a P-spline function with a penalty added to the B-

spline for brain image registration. The method based on the physical model regards

the deformation of the floating image as the physical change caused by the external

force, takes the original image as the input, and calculates the result of the image that

is changed by the external force under the physical rules through the physical model.

The physical models used are mainly viscous fluid models and optical flow field models.

Wodzinski et al. [39] applied the algorithm of the optical flow field model to breast

cancer tumor localization, compared it with the B-spline method, and obtained a better

registration effect.

With the development of deep learning technology, significant progress has been

made in the field of image processing, mainly including the use of unsupervised or self-

supervised deep learning to calculate deformation parameters and similarity measures.

Zum Beispiel, Hessam et al. [40] used a large number of artificially generated

displacement vector fields for training to integrate image content from multiple scales,

thereby directly estimating the displacement vector field from the input image.

Hongming et al. [41] proposed a new non-rigid image registration algorithm based on a

fully convolutional network, and optimized and learned the spatial transformation

process between images through a self-supervised learning framework. Jedoch, until

Jetzt, the non-rigid registration algorithm is still not mature enough compared with the

rigid registration algorithm, and the algorithm acceptance is not enough [42].

3. Automatic delineation of normal tissue based on machine learning

3.1 Atlas based automatic contouring

After multimodal image registration, clinicians will delineate contour information

on the planned CT. The delineated targets mainly included therapeutic targets and

OARs. The shape of OARs is relatively definite, and the location generally does not

change much. In terms of automatically delineating OARs, the most widely used

clinically is the automatic segmentation technology based on the atlas library [43]. Atlas

refers to medical images and their corresponding binary delineation results, since even

among different groups of people, the relative spatial positions and spatial shapes of

normal organs in the body are similar, and the image textures have the same

characteristics. The delineation principle is to pre-establish one or several sets of OARs

templates, and machine learning methods automatically match the appropriate

l

D
Ö
w
N
Ö
A
D
e
D

F
R
Ö
M
H

T
T

P

:
/
/

D
ich
R
e
C
T
.

M

ich
T
.

e
D
u
D
N

/

ich

T
/

l

A
R
T
ich
C
e

P
D

F
/

D
Ö

ich
/

ich

T

.

/

/

1
0
1
1
6
2
D
N
_
A
_
0
0
2
0
4
2
1
2
9
4
4
2
D
N
_
A
_
0
0
2
0
4
P
D

/

.

T

ich

F

B
j
G
u
e
S
T

T

Ö
N
0
7
S
e
P
e
M
B
e
R
2
0
2
3

Data Intelligence Just Accepted MS.
https://doi.org/10.1162/dint_a_00204

templates [44].

The delineation methods based on atlas libraries can be basically divided into two

categories: delineation methods based on single atlases and delineation methods based

on multiple atlases [45]. The delineation method based on a single map can be regarded

as a deformation registration problem. Erste, the atlas is registered to the image to be

delineated, and the transformation matrix and deformation field are obtained. All the

delineated organs in the atlas will be deformed and mapped according to the same

transformation parameters, and the result of the mapping is the delineation result.

Jedoch, the single-atlas library delineation method may have a large difference

between the input patient images and the average atlas, resulting in unsatisfactory

delineation results.

The accuracy of the method based on a single atlas library depends heavily on the

accuracy of image registration. When the atlas used is very different from the image to

be delineated, it is difficult for the registration algorithm to achieve good results,

resulting in a significant reduction in delineation accuracy. In order to improve this

phenomenon, Aljabar et al. [46] proposed a multi-atlas method, which registered and

fused multiple sets of reference atlases with the images to be delineated, erhalten

multiple sets of alternative delineation schemes, and used an algorithm to synthesize

the alternative plans to form the final delineation. The performance of the multi-atlas

library is often more stable than that of the single-atlas library, because the poor

mapping results of some atlases in the multi-atlas will be corrected by other better-

performing atlases, so that each part can be relatively reasonable. While multi-map-

based methods improve the robustness of delineation compared to single-map-based

Methoden, they are prone to topological errors because voxel voting does not necessarily

result in closed surfaces. Such topological errors have a great impact on the formulation

of radiation therapy plans, and are also difficult to detect, requiring time-consuming

review and manual editing by clinicians [47].

3.2 Deep learning based automatic contouring

The atlas library is essentially the operation of registering the target image and the

template image through morphological features, das ist, the process of searching for the

most approximate shape in the atlas library. But if the shape difference of the template

image OARs is too large, the volume is too small or automatically delineated

inappropriate choice of deformation algorithm will affect the registration accuracy [48].

The multi-atlas library can improve the accuracy of delineation, but the amount of

l

D
Ö
w
N
Ö
A
D
e
D

F
R
Ö
M
H

T
T

P

:
/
/

D
ich
R
e
C
T
.

M

ich
T
.

e
D
u
D
N

/

ich

T
/

l

A
R
T
ich
C
e

P
D

F
/

D
Ö

ich
/

ich

.

/

T

/

1
0
1
1
6
2
D
N
_
A
_
0
0
2
0
4
2
1
2
9
4
4
2
D
N
_
A
_
0
0
2
0
4
P
D

.

T

/

ich

F

B
j
G
u
e
S
T

T

Ö
N
0
7
S
e
P
e
M
B
e
R
2
0
2
3

Data Intelligence Just Accepted MS.
https://doi.org/10.1162/dint_a_00204

calculation increases and the time-consuming increases, so a balance between accuracy

and speed must be balanced.

Automatic delineation based on deep learning does not require the above trade-offs.

Since the key advantage of deep learning is to automatically extract labelled features

through the learning of generalized features in training samples to identify new scenes,

the more input templates, the more accurate the learned features [49]. Dolz et al. [50] gebraucht

the support vector machine (SVM) algorithm to successfully achieve automatic

segmentation of the brainstem on the MRI image of brain tumors, and then used another

deep learning algorithm to segment the optic nerve, optic chiasm, pituitary and small

organs such as pituitary stalk are automatically segmented, and the similarity

coefficient reaches 76-83% [51]. They also used hand-extracted features, combined with

unsupervised stacked denoising autoencoders for brainstem segmentation, and the

classification speed was about 70 times faster than that based on SVM methods,

reducing segmentation time [52]. Liang et al. [53] performed automatic segmentation on

CT images based on deep learning, with a sensitivity of 0.997~1 for automatic

segmentation of most organs, which can effectively improve nasopharyngeal cancer

radiotherapy planning.

Currently, deep learning networks, especially convolutional neural networks

(CNN), have become a common method for medical image analysis [54]. CNN is capable

of processing multi-dimensional and multi-channel data, capturing complex nonlinear

mappings between input and output, with advantages for image processing and

classification. A Stanford University study used a CNN model to automatically segment

head and neck OARs for the first time. In the automatic segmentation of organs such as

bone, pharynx, larynx, eyeball and optic nerve, it is better than or equivalent to the

current best technology. But for organs such as parotid gland, submandibular gland and

optic chiasm whose boundaries are not easy to identify on CT images, the delineated

results are not satisfactory [55]. Lu et al. [56] used a 3D CNN to automatically segment

the liver, combined with a graph cut algorithm to refine the segmentation. Der

advantage is that no manual initialization is required, and the segmentation process can

be performed by non-professionals. Also using 3D CNN for liver segmentation, Hu et

al. [57] combined deep learning with global and local shape prior information, Und

evaluated on the same dataset, and all error indicators were significantly reduced. In einem

follow-up study, the target was extended to abdominal multi-organ segmentation, verwenden

3D CNN to perform pixel-to-pixel dense prediction with higher accuracy and shorter

l

D
Ö
w
N
Ö
A
D
e
D

F
R
Ö
M
H

T
T

P

:
/
/

D
ich
R
e
C
T
.

M

ich
T
.

e
D
u
D
N

/

ich

T
/

l

A
R
T
ich
C
e

P
D

F
/

D
Ö

ich
/

ich

/

/

T

.

1
0
1
1
6
2
D
N
_
A
_
0
0
2
0
4
2
1
2
9
4
4
2
D
N
_
A
_
0
0
2
0
4
P
D

T

.

/

ich

F

B
j
G
u
e
S
T

T

Ö
N
0
7
S
e
P
e
M
B
e
R
2
0
2
3

Data Intelligence Just Accepted MS.
https://doi.org/10.1162/dint_a_00204

segmentation time [58].

daher, the outline processing of OARs is a complex project, and it is often

difficult to use a set of models to achieve the expected accuracy for different parts of

the body or different modalities. In actual situations, it is necessary to combine specific

factors to make certain improvements to deep neural networks.

4. Therapeutic target segmentation based on deep learning

4.1 GTV automatic delineation

As with normal tissue delineation, deep learning-assisted tumor target delineation

helps improve execution efficiency. Jedoch, since it is often difficult to distinguish

the boundary between the tumor and the surrounding tissue, the clinical information,

pathological sections, and images of the patient will become the reference data for GTV

delineation. Various techniques are used to aid in identification. In the Multimodal

Brain Tumor Image Segmentation Challenge (BraTS) In 2013, Pereira et al. [59] gebraucht

CNN to automatically segment brain tumor MRI images, which improved the network

accuracy and ranked first. Since then, Kamnitsas et al. [60] proposed a dual-channel 3D

CNN network for brain injury (including traumatic brain injury, brain tumor, ischemic

stroke) segmentation, the first time to use fully connected conditional randomization on

medical data. Both of the above studies used neural networks with small convolution

kernels to make the network structure deeper without increasing the computational cost.

Men et al. [61] used big data to train deep dilated residual network (DD-ResNet) für

breast tumor segmentation, and the results were better than deep dilated convolutional

neural networks (DDCNN) and distributed deep neural networks (DDNN), similar to

Dice The dice similarity coefficient (DSC) War 91%, which was higher than the result

hand-drawn by experts [62].

Zusätzlich, for the above-mentioned basic network types, studies have also shown

that the improved network in [63] can improve the accuracy of network segmentation

and has stronger robustness. Lin et al. [64] trained a 3D CNN to delineate the GTV of

nasopharyngeal carcinoma on MRI images, and the similarity with the GTV delineated

by experts was high, with the DSC reaching 79%. With the help of machine learning,

doctors reduced their time by 39.4% and improved their accuracy. 3D CNN not only

utilizes the CT image information of each layer extracted by traditional CNN, but also

utilizes the information between layers, the information utilization rate is high, und das

accuracy is improved to a certain extent. Qi et al. [65] used convolutional neural networks

l

D
Ö
w
N
Ö
A
D
e
D

F
R
Ö
M
H

T
T

P

:
/
/

D
ich
R
e
C
T
.

M

ich
T
.

e
D
u
D
N

/

ich

T
/

l

A
R
T
ich
C
e

P
D

F
/

D
Ö

ich
/

ich

.

/

/

T

1
0
1
1
6
2
D
N
_
A
_
0
0
2
0
4
2
1
2
9
4
4
2
D
N
_
A
_
0
0
2
0
4
P
D

.

T

/

ich

F

B
j
G
u
e
S
T

T

Ö
N
0
7
S
e
P
e
M
B
e
R
2
0
2
3

Data Intelligence Just Accepted MS.
https://doi.org/10.1162/dint_a_00204

to delineate the target volume of nasopharyngeal carcinoma based on multimodal

Bildgebung (CT and MRI). The results show that the target area is delineated with high

precision. Li et al. [66] used the U-Net to automatically delineate the target volume of

nasopharyngeal carcinoma based on CT images. The results showed that the

segmentation accuracy of the automatically delineated target volume was high. Li et al.

[67] based on the four-dimensional computed tomography data of patients with non-

small cell lung cancer, used transfer learning to automatically delineate the tumor area,

which improved the accuracy and shortened the retraining time of the network. Wann

the breathing range was 5-10 mm, the matching index improved by 36.1% on average

compared with the comprehensive elastic deformation registration technique. In a

recent study [68], the authors used fuzzy c-means clustering (FCM), artificial neural

network(ANN), and SVM algorithms to automatically segment GTV of solid, Boden-

glass, and mixed lung cancer lesions, jeweils. It is considered that the results of

the FCM model are more accurate and efficient, and can be reliably applied to SBRT.

Delineating GTV based on deep learning can improve the work efficiency of

clinicians, but this method cannot completely replace manual delineation. On the basis

of automatic delineation, manual correction is still required to achieve accurate

delineation effects [69].

4.2 CTV automatic delineation

CTV should be given a certain dose of radiation to the subclinical foci formed by

infiltration around the primary tumor and the path of regional lymph node metastasis

according to the requirements of radiobiology and the factors of tumor occurrence and

metastasis. It is the basis for tumor regional radiotherapy to control recurrence and

metastasis. The delineation needs to be judged in combination with the specific

pathological conditions and the possible invasion or metastasis range of the diseased

tissue, and the delineation results of different types of tumors and different stages are

completely different.

Speziell, Men et al. [70] used a DDCNN model to attempt automatic

segmentation of CTV and OARs in 218 rectal cancer patients, and the results were

accurate and efficient. Among them, the DSC of CTV reaches 87.7%, the DSC of

bladder and bilateral femoral head is more than 90%, and the delineation of small

l

D
Ö
w
N
Ö
A
D
e
D

F
R
Ö
M
H

T
T

P

:
/
/

D
ich
R
e
C
T
.

M

ich
T
.

e
D
u
D
N

/

ich

T
/

l

A
R
T
ich
C
e

P
D

F
/

D
Ö

ich
/

ich

/

/

.

T

1
0
1
1
6
2
D
N
_
A
_
0
0
2
0
4
2
1
2
9
4
4
2
D
N
_
A
_
0
0
2
0
4
P
D

.

T

/

ich

F

B
j
G
u
e
S
T

T

Ö
N
0
7
S
e
P
e
M
B
e
R
2
0
2
3

Data Intelligence Just Accepted MS.
https://doi.org/10.1162/dint_a_00204

intestine and colon is not accurate enough, and the DSC is 65.3% Und 61.8%,

jeweils. It is possibly related with that they are both air-containing hollow organs.

Based on deep learning with Area-aware reweight strategy and Recursive refinement

strategy, called RA-CTVNet, Shi et al. [71] segment the CTV from cervical cancer CT

Bilder. Their experimental results show that RA-CTVNet improves DSC compared

with different network architectures. Compared with three clinical experts, RA-

CTVNet performed better than the two experts while comparably to the third expert.

Shen et al. [72] modified the U-net model by incorporating the contours of gross tumor

volume of lymph node (GTVnd) and designed the DiUnet model for the automatic

delineation of lung cancer CTV. The results showed that the DSC of most lymph node

regions was up to 70%, which was not significantly different from manual delineation.

Zusätzlich, our team [73] collected CT images of 53 cervical cancer patients. Von

modifying the U-net model and the training process according to the task, the automatic

segmentation of images of cervical cancer CTV region and normal tissue is realized.

By testing the prediction accuracy of the model and the number of required dialogue

rounds, the recall rate, accuracy rate, DSC, Intersection over Union (IoU), usw. of the

results were evaluated. The results show that the proposed model has good performance

in all the indicators outlined in the target area. And compared with commonly used deep

learning neural network models such as mask region-based convolution neural network

(Mask R-CNN), speech enhancement generative adversarial network (SegAN), and U-

net, the segmentation boundary of the proposed model is clearer and smoother, und das

recall rate is obviously better than that of other models. Darüber hinaus, because of its very

light weight, it can be adapted to the dataset size-limited case.

Due to the involvement of subclinical lesions and lymph node drainage areas, CTV

automatic delineation is relatively more difficult, and the performance of deep learning

delineation is still far from that of experts [74-76]. In the future, relying on the disease-

specific big data platform to integrate multimodal radiotherapy data, Bildgebung, genetic

and other multi-omics data, as well as the experience data of senior radiotherapy

physicians, physicists, and technicians, it is expected to be useful in the prediction of

efficacy and complication risk. Guided by the results, individualized CTV range

decisions are provided.

l

D
Ö
w
N
Ö
A
D
e
D

F
R
Ö
M
H

T
T

P

:
/
/

D
ich
R
e
C
T
.

M

ich
T
.

e
D
u
D
N

/

ich

T
/

l

A
R
T
ich
C
e

P
D

F
/

D
Ö

ich
/

ich

T

.

/

/

1
0
1
1
6
2
D
N
_
A
_
0
0
2
0
4
2
1
2
9
4
4
2
D
N
_
A
_
0
0
2
0
4
P
D

/

.

T

ich

F

B
j
G
u
e
S
T

T

Ö
N
0
7
S
e
P
e
M
B
e
R
2
0
2
3

Data Intelligence Just Accepted MS.
https://doi.org/10.1162/dint_a_00204

5. Abschluss

The research of machine learning methods in the field of radiotherapy has been

fully rolled out and achieved phased results, among which the automatic delineation of

normal tissues and tumor target areas has always been a research hotspot [77-79]. Most

of the existing deep learning models are based on natural images, and there is a lack of

deep learning models dedicated to medical, especially radiation oncology-related

Bilder. The difference between medical images and natural images is that medical

images are grayscale images and generally have continuity [80-81]. In image

segmentation, not only the regional structure of an image, but also the spatial structure

of 3D data must be considered [82]. Zusätzlich, local and global prior information needs

to be considered before it can further contribute to the segmentation of OARs and

therapeutic target volume [83]. Darüber hinaus, multimodal image registration is often

required to further identify the extent of tumor invasion [84-85].

Besides, radiotherapy is one of the links in tumor treatment. How to determine the

appropriate radiotherapy target range and irradiation dose is a complex issue that

requires system integration, such as disease characteristics and overall treatment mode,

even the cross-scale issues from molecular cells to tissues and organs, and the spatio-

temporal relationship of biomolecules and other factors need to be comprehensively

analyzed. So that the radiotherapy plan obtained in this way is more in line with the

principle of precise individualized treatment. The integration of automatic radiotherapy

target delineation with artificial intelligence knowledge maps and causal analysis may

play an important role in the formulation of clinical radiotherapy targets[86].

At present, most of the current applications are in the preclinical research stage, Aber

there are still some problems in clinical application. Erste, high-quality clinical data is

the basis for artificial intelligence to learn and judge, but the current standardization of

relevant medical data for automatic target area delineation is not high. The quality of

labeling is uneven, and the data of major medical centers lack a joint construction and

sharing mechanism. There are data barriers, which seriously hinder the effective use of

data and product development. Zweitens, it is still difficult to accurately define the

treatment target area. Based on the current CT, MRT, PET-CT and other means, es ist

generally not difficult to determine the GTV, but some lesions are still difficult to

identify, such as soft tissue invasion, bone destruction degree and scope, usw. The doses

of CTV are different according to the risk of recurrence and metastasis. There is no

l

D
Ö
w
N
Ö
A
D
e
D

F
R
Ö
M
H

T
T

P

:
/
/

D
ich
R
e
C
T
.

M

ich
T
.

e
D
u
D
N

/

ich

T
/

l

A
R
T
ich
C
e

P
D

F
/

D
Ö

ich
/

ich

/

/

T

.

1
0
1
1
6
2
D
N
_
A
_
0
0
2
0
4
2
1
2
9
4
4
2
D
N
_
A
_
0
0
2
0
4
P
D

/

.

T

ich

F

B
j
G
u
e
S
T

T

Ö
N
0
7
S
e
P
e
M
B
e
R
2
0
2
3

Data Intelligence Just Accepted MS.
https://doi.org/10.1162/dint_a_00204

relevant research on how to determine high-, medium-, and low-risk CTV. Zusätzlich,

the clinical application of artificial intelligence is directly related to life and health, Und

faces many ethical and legal challenges. Jedoch, the automatic delineation of

radiotherapy target volume based on machine learning will be an important

development direction of artificial intelligence in the medical field in the future.

Danksagungen

This work was financially supported by

1. Scientific Research Project of Anhui Provincial Health Commission (NEIN.

AHWJ2022b058)

2. Joint Fund for Medical Artificial Intelligence of the First Affiliated Hospital of

USTC (NEIN. MAI2022Q009)

3. Student Innovation and Entrepreneurship Fund of USTC (NEIN. WK5290000003)

4. China Scholarship Council (NEIN. 202206340057)

Autorenbeiträge

Zhenchao Tao

(zctao@mail.ustc.edu.cn, 0000-0001-8142-9164): Methodik,

Writingoriginal draft preparation, Akquise von Fördermitteln.

Shengfei Lyu (shengfei.lyu@ntu.edu.sg, 0000-0002-1843-6836): Writingoriginal

draft preparation, Writingreview and editing.

Reference

1. Sung, H., Ferlay, J,, Siegel, R.L., et al.:. Global Cancer Statistics 2020:

GLOBOCAN Estimates of Incidence and Mortality Worldwide for 36 Cancers in

185 Countries. CA: A Cancer Journal for Clinicians 71(3), 209-249 (2021).

2. Devita, V.T.Jr., Rosenberg, S.A.: Two hundred years of cancer research. The New

England Journal of Medicine 366 (23), 2207-14 (2012).

3. Avanzo, M., Stancanello, J., Pirrone, G., et al.: Radiomics and deep learning in lung

cancer. Strahlentherapie und Onkologie 196(10), 879-887 (2020).

4. Howard, F.M., Kochanny, S., Koshy, M., et al.: Machine Learning-Guided

Adjuvant Treatment of Head and Neck Cancer. JAMA Network Open 3 (11),

e2025881 (2020).

5. Xingyu, W., Zhenchao, T., Bingbing, J., et al.: Domain knowledge-enhanced

l

D
Ö
w
N
Ö
A
D
e
D

F
R
Ö
M
H

T
T

P

:
/
/

D
ich
R
e
C
T
.

M

ich
T
.

e
D
u
D
N

/

ich

T
/

l

A
R
T
ich
C
e

P
D

F
/

D
Ö

ich
/

ich

/

T

/

.

1
0
1
1
6
2
D
N
_
A
_
0
0
2
0
4
2
1
2
9
4
4
2
D
N
_
A
_
0
0
2
0
4
P
D

T

/

.

ich

F

B
j
G
u
e
S
T

T

Ö
N
0
7
S
e
P
e
M
B
e
R
2
0
2
3

Data Intelligence Just Accepted MS.
https://doi.org/10.1162/dint_a_00204

variable selection for biomedical data analysis, Information Sciences 606, 469-488

(2022).

6. Yang, Z., Olszewski, D., Er, C., et al.: Machine learning and statistical prediction

of patient quality-of-life after prostate radiation therapy. Computers in Biology and

Medicine 129(Feb), 104127 (2021).

7. Schmied, A.G., Petersen, J., Terrones-Campos, C., et al.: RootPainter3D: Interactive-

machine-learning enables rapid and accurate contouring for radiotherapy. Medical

Physik 49(1), 461-473 (2022).

8. Baskar, R., Lee, K.A., Yeo, R., et al.: Cancer and radiation therapy: current

advances and future directions. International Journal of Medical Sciences 9(3), 193-

9 (2012).

9. Zhenchao, T., Jun, Q., Yangyang, Z., et al.: Endostar plus chemoradiotherapy

versus chemoradiotherapy alone for patients with advanced nonsmall cell lung

cancer: A systematic review and meta-analysis. International Journal of Radiation

Forschung 19(1), 1-12 (2021).

10. Coutard, H. Principles of X-ray therapy of malignant disease. The Lancet 224

(5784), 1-4 (1934).

11. Thariat, J., Hannoun-Levi, J.M., Sun-Myint, A., et al.: Past, present, and future of

radiotherapy for the benefit of patients. Nature Reviews Clinical Oncology 10(1),

52-60 (2013).

12. Takahashi, S.: Conformation radiotherapy. Rotation techniques as applied to

radiography and radiotherapy of cancer. Acta Radiologica Diagnosis 242, 11-17

(1965).

13. Chang, J.Y., Senan, S., Paul, M.A. et al.: Stereotactic ablative radiotherapy versus

lobectomy for operable stage I non-small-cell lung cancer: a pooled analysis of two

randomised trials. The Lancet Oncology 16(6), 630-637 (2015).

14. Basson, L., Jarraya, H., Escande, A., et al.: Chest Magnetic Resonance Imaging

Decreases Inter-observer Variability of Gross Target Volume for Lung Tumors.

Frontiers in Oncology 9:690 (2019).

15. Sun, Y., Shi, H., Zhang, S., et al.: Accurate and rapid CT image segmentation of the

eyes and surrounding organs for precise radiotherapy. Medical Physics 46(5), 2214-

2222 (2019).

16. Batra, R., Kuecuekkaya, A., Zeevi, T., et al.: Proof-Of-Concept Use Of Machine

l

D
Ö
w
N
Ö
A
D
e
D

F
R
Ö
M
H

T
T

P

:
/
/

D
ich
R
e
C
T
.

M

ich
T
.

e
D
u
D
N

/

ich

T
/

l

A
R
T
ich
C
e

P
D

F
/

D
Ö

ich
/

ich

/

T

.

/

1
0
1
1
6
2
D
N
_
A
_
0
0
2
0
4
2
1
2
9
4
4
2
D
N
_
A
_
0
0
2
0
4
P
D

T

.

/

ich

F

B
j
G
u
e
S
T

T

Ö
N
0
7
S
e
P
e
M
B
e
R
2
0
2
3

Data Intelligence Just Accepted MS.
https://doi.org/10.1162/dint_a_00204

Learning To Predict Tumor Recurrence Of Early-Stage Hepatocellular Carcinoma

Before Therapy Using Baseline Magnetic Resonance Imaging. Transplantation

104(S3), S43-S44 (2020).

17. Wong, J., Fong, A., McVicar, N., et al.: Comparing deep learning-based auto-

segmentation of organs at risk and clinical target volumes to expert inter-observer

variability in radiotherapy planning. Radiotherapy And Oncology 144, 152-158

(2020).

18. Men, K., Geng, H., Cheng, C., et al.: Technical Note: More accurate and efficient

segmentation of organs-at-risk in radiotherapy with convolutional neural networks

cascades. Medical Physics 46(1), 286-292 (2019).

19. Fu, Y., Mazur, T.R., Wu, X., et al.: A novel MRI segmentation method using CNN-

based correction network for MRI-guided adaptive radiotherapy. Medical Physics

45(11), 5129-5137 (2018).

20. LeCun, Y., Bengio, Y., Hinton, G.: Deep learning. Natur 521(7553), 436-444

(2015).

21. Jiang, B., Wu, X., Zhou, X., et al.: Semi-Supervised Multiview Feature Selection

With Adaptive Graph Learning. IEEE Transactions Neural Networks and Learning

Systeme, PP (2022).

22. Li, L., Yan, M., Tao, Z., et al.: Semi-Supervised Graph Pattern Matching and

Rematching for Expert Community Location. ACM Transactions on Knowledge

Discovery from Data (TKDD), (2022).

23. Valdes, G., Simone, C.B., Chen, J., et al.: Clinical decision support of radiotherapy

treatment planning: A data-driven machine learning strategy for patient-specific

dosimetric decision making. Radiother Oncol 125(3), 392-397 (2017).

24. Guo, H.Y., Wang, D.Z.: A Multilevel Optimal Feature Selection and Ensemble

Learning for a Specific CAD System-Pulmonary Nodule Detection. Applied

Mechanics and Materials & Materials 380, 1593-1599 (2013).

25. Zhenyu, L., Chaohong, L., Haiwei, H., et al.: Hierarchical Multi-Granularity

Attention-Based Hybrid Neural Network for Text Classification. IEEE Access

2020(8), 149362-149371 (2020).

26. Zhao, X., Chen, H., Xing, Z., et al.: Brain-Inspired Search Engine Assistant Based

on Knowledge Graph. IEEE Transactions on Neural Networks and Learning

Systeme, PP (2021).

27. Huang, B., Zhu, Y., Usman, M., et al.: Graph Neural Networks for Missing Value

l

D
Ö
w
N
Ö
A
D
e
D

F
R
Ö
M
H

T
T

P

:
/
/

D
ich
R
e
C
T
.

M

ich
T
.

e
D
u
D
N

/

ich

T
/

l

A
R
T
ich
C
e

P
D

F
/

D
Ö

ich
/

ich

/

T

.

/

1
0
1
1
6
2
D
N
_
A
_
0
0
2
0
4
2
1
2
9
4
4
2
D
N
_
A
_
0
0
2
0
4
P
D

.

T

/

ich

F

B
j
G
u
e
S
T

T

Ö
N
0
7
S
e
P
e
M
B
e
R
2
0
2
3

Data Intelligence Just Accepted MS.
https://doi.org/10.1162/dint_a_00204

Classification in a Task-driven Metric Space. IEEE Transactions on Knowledge and

Data Engineering, (2022).

28. Zhao, X., Chen, L., Chen, H.: A Weighted Heterogeneous Graph-Based Dialog

System. IEEE Trans Neural Netw Learn Syst, PP (2021).

29. Yuan, B., Chen, H., Yao, X.: Toward efficient design space exploration for fault-

tolerant multiprocessor systems. IEEE Transactions on Evolutionary Computation

24(1), 157-169 (2019).

30. Lyu, S., Tian, X., Li, Y., et al.: Multiclass probabilistic classification vector machine.

IEEE Transactions on Neural Networks and Learning Systems 31(10), 3906-3919

(2019).

31. Yu, K., Liu, L., Li, J., et al.: Mining Markov Blankets Without Causal Sufficiency.

IEEE Transactions on Neural Networks and Learning Systems 29(12), 6333-6347

(2018).

32. Rahunathan, S., Stredney, D., Schmalbrock, P., et al.: Image registration using rigid

registration and maximization of mutual information. The 13th Annual Medicine

Meets Virtual Reality Conference, (2005).

33. Viergever, M.A., Maintz, J.B.A., Klein, S., et al.: A survey of medical image

registrationunder review. Medical Image Analysis 33, 140-144 (2016).

34. Oliveira, F.P., Tavares, J.M.: Medical image registration: a review. Computer

Methods in Biomechanics & Biomedizintechnik 17(2), 73-93 (2014).

35. Mahapatra, D., Antony, B., Sedai, S., et al.: Deformable medical image registration

using generative adversarial networks. 2018 IEEE 15th International Symposium

on Biomedical Imaging (ISBI 2018). IEEE, 1449-1453 (2018).

36. Kearney, V., Haaf, S., Sudhyadhom, A., et al.: An unsupervised convolutional

neural network-based algorithm for deformable image registration. Physics in

Medicine and Biology 63(18), 185017 (2018).

37. Gu, S., Meng, X., Sciurba, F.C., et al.: Bidirectional elastic image registration using

B-spline affine transformation. Computerized Medical Imaging and Graphics 38(4),

306-314 (2014).

38. Pradhan, S., Patra, D.: RMI based non-rigid image registration using BF-QPSO

optimization and P-spline. AEU-International Journal of Electronics and

Kommunikation 69(3): 609-621 (2015).

39. Wodzinski, M., Skalski, A., Ciepiela, ICH., et al.: Application of demons image

registration algorithms in resected breast cancer lodge localization. 2017 Signal

l

D
Ö
w
N
Ö
A
D
e
D

F
R
Ö
M
H

T
T

P

:
/
/

D
ich
R
e
C
T
.

M

ich
T
.

e
D
u
D
N

/

ich

T
/

l

A
R
T
ich
C
e

P
D

F
/

D
Ö

ich
/

ich

T

/

/

.

1
0
1
1
6
2
D
N
_
A
_
0
0
2
0
4
2
1
2
9
4
4
2
D
N
_
A
_
0
0
2
0
4
P
D

T

.

/

ich

F

B
j
G
u
e
S
T

T

Ö
N
0
7
S
e
P
e
M
B
e
R
2
0
2
3

Data Intelligence Just Accepted MS.
https://doi.org/10.1162/dint_a_00204

Processing: Algorithms, Architectures, Arrangements, and Applications (SPA),

400-405 (2017).

40. Sokooti, H., Vos, B., Berendsen, F., et al.: Nonrigid image registration using multi-

scale 3D convolutional neural networks. International conference on medical image

computing and computer-assisted intervention. Springer, Cham, 232-239 (2017).

41. Hongming, L., Yong, F.: Non-Rigid Image Registration Using Self-Supervised

Fully Convolutional Networks Without Training Data. IEEE 15th International

Symposium on Biomedical Imaging, 1075-1078 (2018).

42. Rong, Y., Rosu-Bubulac, M., Benedict, SH., et al.: Rigid and Deformable Image

Registration for Radiation Therapy: A Self-Study Evaluation Guide for NRG

Oncology Clinical Trial Participation. Practical Radiation Oncology 11(4), 282-298

(2021).

43. Zhang, T., Yang, Y., Wang, J., et al.: Comparison between atlas and convolutional

neural network based automatic segmentation of multiple organs at risk in non-

small cell lung cancer. Medicine (Baltimore) 99(34), e21800 (2020).

44. Lustberg, T., Soest, J.V., Gooding, M., et al.: Clinical evaluation of atlas and deep

learning based automatic contouring for lung cancer. Radiotherapy and Oncology

126(2), 312-317 (2018).

45. Cabezas, M., Oliver, A., Lladó, X., et al.: A review of atlas-based segmentation for

magnetic resonance brain

Bilder. Computer Methods and Programs

In

Biomedicine 104(3), e158-77 (2011).

46. Aljabar, P., Heckemann, R.A., Hammers, A., et al.: Multi-atlas based segmentation

of brain images: atlas selection and its effect on accuracy. Neurobild 46(3), 726-

738 (2009).

47. Sharp, G., Fritscher, K.D., Pekar, V., et al.: Vision 20/20: perspectives on automated

image segmentation for radiotherapy. Medical Physics 41(5), 050902 (2014).

48. Valerio, F., René, F.V., Fedde, V.D.L, et al.: Tissue segmentation of head and neck

CT images for treatment planning: a multiatlas approach combined with intensity

modeling. Medical Physics 40(7), 071905 (2013).

49. Li, Y., Zhang, H., Xue, X., et al.: Deep learning for remote sensing image

classification: A survey. Wiley Interdisciplinary Reviews: Data Mining and

Knowledge Discovery 8(6), e1264 (2018).

50. Dolz, J., Laprie, A., Ken, S., et al.: Supervised machine learning-based

classification scheme to segment the brainstem on MRI in multicenter brain tumor

l

D
Ö
w
N
Ö
A
D
e
D

F
R
Ö
M
H

T
T

P

:
/
/

D
ich
R
e
C
T
.

M

ich
T
.

e
D
u
D
N

/

ich

T
/

l

A
R
T
ich
C
e

P
D

F
/

D
Ö

ich
/

ich

.

/

T

/

1
0
1
1
6
2
D
N
_
A
_
0
0
2
0
4
2
1
2
9
4
4
2
D
N
_
A
_
0
0
2
0
4
P
D

T

/

.

ich

F

B
j
G
u
e
S
T

T

Ö
N
0
7
S
e
P
e
M
B
e
R
2
0
2
3

Data Intelligence Just Accepted MS.
https://doi.org/10.1162/dint_a_00204

treatment context. International Journal of Computer Assisted Radiology Surgery

11(1), 43-51 (2016).

51. Dolz, J., Reyns, N., Betrouni, N., et al.: A deep learning classification scheme based

on augmented-enhanced features to segment organs at risk on the optic region in

brain cancer patients. arXiv preprint arXiv:1703.10480 (2017).

52. Dolz, J., Betrouni, N., Quidet, M., et al.: Stacking denoising auto-encoders in a deep

network to segment the brainstem on MRI in brain cancer patients: A clinical study.

Computerized Medical Imaging and Graphics 52, 8-18 (2016).

53. Liang, S., Tang, F., Huang, X., et al.: Deep-learning-based detection and

segmentation of organs at risk in nasopharyngeal carcinoma computed tomographic

images for radiotherapy planning. European Radiology 29(4), 1961-1967 (2019).

54. Litjens, G., Kooi, T., Bejnordi, B.E., et al.: A survey on deep learning in medical

image analysis. Medical Image Analysis 42, 60-88 (2017).

55. Ibragimov, B., Xing, L.: Segmentation of organs-at-risks in head and neck CT

images using convolutional neural networks. Medical Physics 44(2), 547-557

(2017).

56. Lu, F., Wu, F., Hu, P., et al.: Automatic 3D liver location and segmentation via

convolutional neural network and graph cut. International Journal of Computer

Assisted Radiology Surgery 12(2), 171-182 (2017).

57. Hu, P., Wu, F., Peng, J., et al.: Automatic 3D liver segmentation based on deep

learning and globally optimized surface evolution. Physics in Medicine and Biology

61(24), 8676-8698 (2016).

58. Hu, P., Wu, F., Peng, J., et al.: Automatic abdominal multi-organ segmentation

using deep convolutional neural network and time-implicit level sets. International

Journal of Computer Assisted Radiology Surgery 12(3), 399-411 (2017).

59. Pereira, S., Pinto, A., Alves, V., et al.: Brain Tumor Segmentation Using

Convolutional Neural Networks in MRI Images. IEEE Transactions on Medical

Imaging 35(5), 1240-1251 (2016).

60. Kamnitsas, K., Ledig, C., Newcombe, V.F.J., et al.: Efficient multi-scale 3D CNN

with fully connected CRF for accurate brain lesion segmentation. Medical Image

Analyse 36, 61-78 (2017).

61. Men, K., Zhang, T., Chen, X., et al.: Fully automatic and robust segmentation of

the clinical target volume for radiotherapy of breast cancer using big data and deep

learning. Physica Medica 50, 13-19 (2018).

l

D
Ö
w
N
Ö
A
D
e
D

F
R
Ö
M
H

T
T

P

:
/
/

D
ich
R
e
C
T
.

M

ich
T
.

e
D
u
D
N

/

ich

T
/

l

A
R
T
ich
C
e

P
D

F
/

D
Ö

ich
/

ich

.

/

/

T

1
0
1
1
6
2
D
N
_
A
_
0
0
2
0
4
2
1
2
9
4
4
2
D
N
_
A
_
0
0
2
0
4
P
D

.

T

/

ich

F

B
j
G
u
e
S
T

T

Ö
N
0
7
S
e
P
e
M
B
e
R
2
0
2
3

Data Intelligence Just Accepted MS.
https://doi.org/10.1162/dint_a_00204

62. Mourik, A.M.V, Elkhuizen, P.H.M., Minkema, D., et al.: Multiinstitutional study

on target volume delineation variation in breast radiotherapy in the presence of

guidelines. Radiotherapy & Oncology 94(3), 286-91 (2010).

63. Oktay, O., Ferrante, E., Kamnitsas, K., et al.: Anatomically Constrained Neural

Netzwerke (ACNNs): Application to Cardiac Image Enhancement and Segmentation.

IEEE Transactions on Medical Imaging 37(2), 384-395 (2018).

64. Lin, L., Dou, Q., Jin, Y.M., et al.: Deep Learning for Automated Contouring of

Primary Tumor Volumes by MRI for Nasopharyngeal Carcinoma. Radiology

291(3), 677-686 (2019).

65. Qi, Y., Li, J., Chen, H., et al.: Computer-aided diagnosis and regional segmentation

of nasopharyngeal carcinoma based on multi-modality medical

Bilder.

International Journal of Computer Assisted Radiology Surgery 16(6), 871-882

(2021).

66. Li, S., Xiao, J., Er, L., et al.: The Tumor Target Segmentation of Nasopharyngeal

Cancer in CT Images Based on Deep Learning Methods. Technology in Cancer

Forschung & Treatment 18:1533033819884561 (2019).

67. Xiadong, L., Ziheng, D., Qinghua, D., et al.: A novel deep learning framework for

internal gross target volume definition from 4D computed tomography of lung

cancer patients. IEEE Access 6, 37775-37783 (2018).

68. Kawata, Y., Arimura, H., Ikushima, K., et al.: Impact of pixel-based machine-

learning techniques on automated frameworks for delineation of gross tumor

volume regions for stereotactic body radiation therapy. Physica Medica 42, 141-

149 (2017).

69. Chen, M., Wu, S., Zhao, W., et al.: Application of deep learning to auto-delineation

of target volumes and organs at risk in radiotherapy. Cancer Radiotherapie 26(3),

494-501 (2022).

70. Men, Kuo., Dai, Jianrong., Li, Yexiong.: Automatic segmentation of the clinical

target volume and organs at risk in the planning CT for rectal cancer using deep

dilated convolutional neural networks. Medical Physics 44(12), 6377-6389 (2017).

71. Shi, J., Ding, X., Liu, X., et al.: Automatic clinical target volume delineation for

cervical cancer in CT images using deep learning. Medical Physics 48(7), 3968-

3981 (2021).

72. Shen, J., Zhang, F., Aus, M., et al.: Clinical target volume automatic segmentation

based on lymph node stations for lung cancer with bulky lump lymph nodes.

l

D
Ö
w
N
Ö
A
D
e
D

F
R
Ö
M
H

T
T

P

:
/
/

D
ich
R
e
C
T
.

M

ich
T
.

e
D
u
D
N

/

ich

T
/

l

A
R
T
ich
C
e

P
D

F
/

D
Ö

ich
/

ich

/

/

.

T

1
0
1
1
6
2
D
N
_
A
_
0
0
2
0
4
2
1
2
9
4
4
2
D
N
_
A
_
0
0
2
0
4
P
D

.

T

/

ich

F

B
j
G
u
e
S
T

T

Ö
N
0
7
S
e
P
e
M
B
e
R
2
0
2
3

Data Intelligence Just Accepted MS.
https://doi.org/10.1162/dint_a_00204

Thoracic Cancer 13(20), 2897-2903 (2022).

73. Yizhan, F., Zhenchao, T., Jun, L., et al.: An Encoder-Decoder Network for

Automatic Clinical Target Volume Target Segmentation of Cervical Cancer in CT

Images. International Journal of Crowd Science 6(3), 111-116 (2022).

74. Chen, X., Sun, S., Bai, N., et al.: A deep learning-based auto-segmentation system

for organs-at-risk on whole-body computed tomography images for radiation

therapy. Radiotherapy and Oncology 160, 175-184 (2021).

75. Ju, Z., Guo, W., Gu, S., et al.: CT based automatic clinical target volume delineation

using a dense-fully connected convolution network for cervical Cancer radiation

therapy. BMC Cancer 21(1), 1-10 (2021).

76. Unkelbach, J., Bortfeld, T., Cardenas, C.E., et al.: The role of computational

methods for automating and improving clinical target volume definition.

Radiotherapy and Oncology 153, 15-25 (2020).

77. Min, H., Dowling, J., Jameson, M.G., et al.: Automatic radiotherapy delineation

quality assurance on prostate MRI with deep learning in a multicentre clinical trial.

Physics in Medicine Biology 66(19), 195008 (2021).

78. Robert, C., Munoz, A., Moreau, D., et al.: Clinical implementation of deep-learning

based auto-contouring tools-Experience of three French radiotherapy centers.

Cancer Radiotherapie 25(6-7), 607-616 (2021).

79. Piazzese, C., Evans, E., Thomas, B., et al.: FIELDRT: an open-source platform for

the assessment of target volume delineation in radiation therapy. British Journal of

Radiology 94(1126), 20210356 (2021).

80. Ogura, A., Kamakura, A., Kaneko, Y., et al.: Comparison of grayscale and color-

scale renderings of digital medical

images for diagnostic

Deutung.

Radiological Physics and Technology 10(3), 359-363 (2017).

81. Huang, Y., Hu, G., Ji, C., et al.: Glass-cutting medical images via a mechanical

image segmentation method based on crack propagation. Nature Communications

11(1), 5669 (2020).

82. Pizer, S.M., Fletcher, P.T., Joshi, S., et al.: Deformable M-Reps for 3D Medical

Image Segmentation. International Journal of Computer Vision 55(2-3), 85-106

(2003).

83. Mansoor, A., Bagci, U., Foster, B., et al.: Segmentation and Image Analysis of

Abnormal Lungs at CT: Current Approaches, Challenges, and Future Trends.

Radiographics 35(4), 1056-1076 (2015).

l

D
Ö
w
N
Ö
A
D
e
D

F
R
Ö
M
H

T
T

P

:
/
/

D
ich
R
e
C
T
.

M

ich
T
.

e
D
u
D
N

/

ich

T
/

l

A
R
T
ich
C
e

P
D

F
/

D
Ö

ich
/

ich

.

/

T

/

1
0
1
1
6
2
D
N
_
A
_
0
0
2
0
4
2
1
2
9
4
4
2
D
N
_
A
_
0
0
2
0
4
P
D

T

/

.

ich

F

B
j
G
u
e
S
T

T

Ö
N
0
7
S
e
P
e
M
B
e
R
2
0
2
3

Data Intelligence Just Accepted MS.
https://doi.org/10.1162/dint_a_00204

84. Perkuhn, M., Stavrinou, P., Thiele, F., et al.: Clinical Evaluation of a

Multiparametric Deep Learning Model for Glioblastoma Segmentation Using

Heterogeneous Magnetic Resonance Imaging Data From Clinical Routine.

Investigative Radiology 53(11), 647-654 (2018).

85. Li, L., Zhao, X., Lu, W., et al.: Deep Learning for Variational Multimodality Tumor

Segmentation in PET/CT. Neurocomputing 392, 277-295 (2020).

86. Lei, L., Xun, Du., Zan, Zhang., et al.: Fuzzy-Constrained Graph Pattern Matching

in Medical Knowledge Graphs. Datenintelligenz 4(3), 599–619 (2022).

Author biography

Zhenchao Tao received the M.M. degree in oncology from Anhui

Medical University, In 2012, and he is currently pursuing the Ph.D.

degree with University of Science and Technology of China, Hefei. Er

is the deputy chief physician of the Department of Radiation Oncology

at the First Affiliated Hospital of the University of Science and Technology of China.

His research interests mainly include clinical tumor radiation therapy, medical artificial

intelligence, materials science and other directions.

Shengfei Lyu is now a research fellow of Nanyang Technological

Universität (NTU). He received the Ph.D. degree with the School of

Computer Science and Technology, University of Science and

Technology of China (USTC) In 2020. He received the B.Sc. degree in

Information Management and Information Systems from Hefei University of

Technologie (HFUT) In 2015. His research interests include relation extraction, Daten

mining and machine learning.

l

D
Ö
w
N
Ö
A
D
e
D

F
R
Ö
M
H

T
T

P

:
/
/

D
ich
R
e
C
T
.

M

ich
T
.

e
D
u
D
N

/

ich

T
/

l

A
R
T
ich
C
e

P
D

F
/

D
Ö

ich
/

ich

/

.

/

T

1
0
1
1
6
2
D
N
_
A
_
0
0
2
0
4
2
1
2
9
4
4
2
D
N
_
A
_
0
0
2
0
4
P
D

/

T

.

ich

F

B
j
G
u
e
S
T

T

Ö
N
0
7
S
e
P
e
M
B
e
R
2
0
2
3Data Intelligence Just Accepted MS. Bild
Data Intelligence Just Accepted MS. Bild
Data Intelligence Just Accepted MS. Bild

PDF Herunterladen