<?xml version="1.0" encoding="utf-8"?>
<!DOCTYPE article PUBLIC "-//NLM//DTD Journal Publishing DTD v3.0 20080202//EN" "journalpublishing3.dtd">
<article xml:lang="en" article-type="research-article" xmlns:xlink="http://www.w3.org/1999/xlink">
<?release-delay 0|0?>
<front>
<journal-meta>
<journal-id journal-id-type="publisher-id">OL</journal-id>
<journal-title-group>
<journal-title>Oncology Letters</journal-title>
</journal-title-group>
<issn pub-type="ppub">1792-1074</issn>
<issn pub-type="epub">1792-1082</issn>
<publisher>
<publisher-name>D.A. Spandidos</publisher-name>
</publisher>
</journal-meta>
<article-meta>
<article-id pub-id-type="doi">10.3892/ol.2018.9761</article-id>
<article-id pub-id-type="publisher-id">OL-0-0-9761</article-id>
<article-categories>
<subj-group>
<subject>Articles</subject>
</subj-group>
</article-categories>
<title-group>
<article-title>Diagnosis of mesothelioma with deep learning</article-title>
</title-group>
<contrib-group>
<contrib contrib-type="author"><name><surname>Hu</surname><given-names>Xue</given-names></name>
<xref rid="af1-ol-0-0-9761" ref-type="aff"/></contrib>
<contrib contrib-type="author"><name><surname>Yu</surname><given-names>Zebo</given-names></name>
<xref rid="af1-ol-0-0-9761" ref-type="aff"/>
<xref rid="c1-ol-0-0-9761" ref-type="corresp"/></contrib>
</contrib-group>
<aff id="af1-ol-0-0-9761">Department of Blood Transfusion, The First Affiliated Hospital of Chongqing Medical University, Chongqing 400016, P.R. China</aff>
<author-notes>
<corresp id="c1-ol-0-0-9761"><italic>Correspondence to</italic>: Professor Zebo Yu, Department of Blood Transfusion, The First Affiliated Hospital of Chongqing Medical University, 1 Youyi Road, Yuzhong, Chongqing 400016, P.R. China, E-mail: <email>zeboyuchongqing@163.com</email></corresp>
</author-notes>
<pub-date pub-type="ppub">
<month>02</month>
<year>2019</year></pub-date>
<pub-date pub-type="epub">
<day>26</day>
<month>11</month>
<year>2018</year></pub-date>
<volume>17</volume>
<issue>2</issue>
<fpage>1483</fpage>
<lpage>1490</lpage>
<history>
<date date-type="received"><day>26</day><month>03</month><year>2018</year></date>
<date date-type="accepted"><day>03</day><month>10</month><year>2018</year></date>
</history>
<permissions>
<copyright-statement>Copyright: &#x00A9; Hu et al.</copyright-statement>
<copyright-year>2019</copyright-year>
<license license-type="open-access">
<license-p>This is an open access article distributed under the terms of the <ext-link ext-link-type="uri" xlink:href="https://creativecommons.org/licenses/by-nc-nd/4.0/">Creative Commons Attribution-NonCommercial-NoDerivs License</ext-link>, which permits use and distribution in any medium, provided the original work is properly cited, the use is non-commercial and no modifications or adaptations are made.</license-p></license>
</permissions>
<abstract>
<p>Malignant mesothelioma (MM) is a rare but aggressive cancer. The definitive diagnosis of MM is critical for effective treatment and has important medicolegal significance. However, the definitive diagnosis of MM is challenging due to its composite epithelial/mesenchymal pattern. The aim of the current study was to develop a deep learning method to automatically diagnose MM. A retrospective analysis of 324 participants with or without MM was performed. Significant features were selected using a genetic algorithm (GA) or a ReliefF algorithm performed in MATLAB software. Subsequently, the current study constructed and trained several models based on a backpropagation (BP) algorithm, extreme learning machine algorithm and stacked sparse autoencoder (SSAE) to diagnose MM. A confusion matrix, F-measure and a receiver operating characteristic (ROC) curve were used to evaluate the performance of each model. A total of 34 potential variables were analyzed, while the GA and ReliefF algorithms selected 19 and 5 effective features, respectively. The selected features were used as the inputs of the three models. SSAE and GA&#x002B;SSAE demonstrated the highest performance in terms of classification accuracy, specificity, F-measure and the area under the ROC curve. Overall, the GA&#x002B;SSAE model was the preferred model since it required a shorter CPU time and fewer variables. Therefore, the SSAE with GA feature selection was selected as the most accurate model for the diagnosis of MM. The deep learning methods developed based on the GA&#x002B;SSAE model may assist physicians with the diagnosis of MM.</p>
</abstract>
<kwd-group>
<kwd>deep learning</kwd>
<kwd>stacked sparse autoencoder</kwd>
<kwd>malignant mesothelioma</kwd>
<kwd>diagnosis</kwd>
</kwd-group>
</article-meta>
</front>
<body>
<sec sec-type="intro">
<title>Introduction</title>
<p>Malignant mesothelioma (MM) is a rare but aggressive cancer (<xref rid="b1-ol-0-0-9761" ref-type="bibr">1</xref>). The prognosis of patients with MM is poor since the majority of patients are diagnosed at an advanced stage and MM is resistant to current treatment options, including chemotherapy, surgery, radiotherapy and immunotherapy (<xref rid="b2-ol-0-0-9761" ref-type="bibr">2</xref>). The estimated median overall survival for advanced MM is 1 year following diagnosis (<xref rid="b3-ol-0-0-9761" ref-type="bibr">3</xref>). MM has a strong association with exposure to asbestos, a mineral extensively used worldwide in the 1970-80s. Although the use of asbestos has been prohibited in the 21st century, the incidence rate of MM has continued to increase worldwide due to the long latency period of MM (<xref rid="b2-ol-0-0-9761" ref-type="bibr">2</xref>,<xref rid="b4-ol-0-0-9761" ref-type="bibr">4</xref>).</p>
<p>Diagnosis of MM primarily relies on histopathological examination supported by clinical and radiological evidence (<xref rid="b5-ol-0-0-9761" ref-type="bibr">5</xref>). The definitive diagnosis of MM is a crucial step prior to appropriate treatment and has important medicolegal significance due to diagnosis-related compensation issues (<xref rid="b6-ol-0-0-9761" ref-type="bibr">6</xref>). However, the definitive diagnosis of MM can be complex, particularly during early stages. This is due to significant variation between cases and the presence of traits that mimic other cancers (particularly adenocarcinoma) or benign/reactive processes (<xref rid="b6-ol-0-0-9761" ref-type="bibr">6</xref>). Furthermore, given the low frequency of MM, it is commonly misdiagnosed or not identified due to a lack of experienced pathologists (<xref rid="b1-ol-0-0-9761" ref-type="bibr">1</xref>,<xref rid="b6-ol-0-0-9761" ref-type="bibr">6</xref>).</p>
<p>In the last two decades, the use of diagnosis support systems (DSSs) has gradually increased (<xref rid="b7-ol-0-0-9761" ref-type="bibr">7</xref>&#x2013;<xref rid="b9-ol-0-0-9761" ref-type="bibr">9</xref>). A DSS for MM may enable pathologists to rapidly examine medical data in considerable detail. More importantly, it may reduce the variability that occurs with different pathologists. Previous studies regarding computer-aided diagnosis of MM mainly focused on developing automatic image processing approaches, including methods that can automatically detect and quantitatively assess pleural thickening in thoracic computed tomography images (<xref rid="b10-ol-0-0-9761" ref-type="bibr">10</xref>&#x2013;<xref rid="b12-ol-0-0-9761" ref-type="bibr">12</xref>). However, pleural thickening does not exclusively signify an asbestos-related disease (<xref rid="b3-ol-0-0-9761" ref-type="bibr">3</xref>), therefore previous methods possess limitations for identifying MM. To increase the accuracy of the DSS, it is necessary to effectively combine multi-feature data, including clinical, laboratory and radiological characteristics of MM. Er <italic>et al</italic> (<xref rid="b7-ol-0-0-9761" ref-type="bibr">7</xref>) collected multi-feature data to diagnose MM. The authors achieved 96.30&#x0025; classification accuracy by using a probabilistic neural network (PNN). However, multi-feature data may not contribute equally to the identification of MM. Feature selection methods may be utilized to remove the redundant or irrelevant features from the original feature set (<xref rid="b9-ol-0-0-9761" ref-type="bibr">9</xref>); this may aid the diagnosis model to focus on the most discriminative features in order to achieve a higher accuracy and decrease the learning time.</p>
<p>Disease diagnosis is a major classification issue; therefore, the classifier is critical to the DSS. Previously, a number of machine leaning methods have been used as classifiers to assist the diagnosis of disease, including support vector machine (<xref rid="b13-ol-0-0-9761" ref-type="bibr">13</xref>,<xref rid="b14-ol-0-0-9761" ref-type="bibr">14</xref>), extreme learning machine (ELM) (<xref rid="b15-ol-0-0-9761" ref-type="bibr">15</xref>) and deep learning (DL; also termed Hierarchical Learning) (<xref rid="b16-ol-0-0-9761" ref-type="bibr">16</xref>,<xref rid="b17-ol-0-0-9761" ref-type="bibr">17</xref>). DL facilitates computational models, consisting of multiple processing layers, to learn representations of data with multiple levels of abstraction (<xref rid="b18-ol-0-0-9761" ref-type="bibr">18</xref>). DL is a family of computational methods, which includes the restricted Boltzmann machine (<xref rid="b19-ol-0-0-9761" ref-type="bibr">19</xref>), stacked autoencoder (<xref rid="b20-ol-0-0-9761" ref-type="bibr">20</xref>), deep belief networks (<xref rid="b21-ol-0-0-9761" ref-type="bibr">21</xref>) and convolutional neural networks (<xref rid="b22-ol-0-0-9761" ref-type="bibr">22</xref>). These algorithms markedly improve speech recognition, object detection and visual object recognition (<xref rid="b16-ol-0-0-9761" ref-type="bibr">16</xref>,<xref rid="b23-ol-0-0-9761" ref-type="bibr">23</xref>).</p>
<p>An autoencoder is a type of neural network including three layers: input layer, hidden layer, and output layer (<xref rid="b20-ol-0-0-9761" ref-type="bibr">20</xref>). <xref rid="f1-ol-0-0-9761" ref-type="fig">Fig. 1</xref> presents the architecture of a basic autoencoder with &#x2018;encoder&#x2019; and &#x2018;decoder&#x2019; networks. An extension of an autoencoder, the sparse autoencoder (SAE) introduces a spare constraint on the hidden layer (<xref rid="b20-ol-0-0-9761" ref-type="bibr">20</xref>). Its algorithm steps are as follows: The given dataset, <italic>X</italic>={<italic>x</italic>(<xref rid="b1-ol-0-0-9761" ref-type="bibr">1</xref>), <italic>x</italic>(<xref rid="b2-ol-0-0-9761" ref-type="bibr">2</xref>),&#x2026;, <italic>x</italic>(<italic>i</italic>)&#x2026; <italic>x</italic>(<italic>N</italic>)}, <italic>x</italic>(<italic>i</italic>) &#x2208;<italic>R</italic><sup>M</sup> is mapped to the hidden layer with a nonlinear activation function:</p>
<disp-formula>
<alternatives>
<mml:math id="umml1" display="block"><mml:mrow><mml:mi>Z</mml:mi><mml:mspace width=".16em" /><mml:mo>=</mml:mo><mml:mspace width=".16em" /><mml:mi>f</mml:mi><mml:mo stretchy="false">(</mml:mo><mml:msub><mml:mi>W</mml:mi><mml:mn>1</mml:mn></mml:msub><mml:mspace width=".16em" /><mml:mi>X</mml:mi><mml:mspace width=".16em" /><mml:mo>&#x002B;</mml:mo><mml:mspace width=".16em" /><mml:msub><mml:mi>b</mml:mi><mml:mn>1</mml:mn></mml:msub><mml:mo stretchy="false">)</mml:mo><mml:mspace width=".16em" /><mml:mo>=</mml:mo><mml:mspace width=".16em" /><mml:mfrac><mml:mn>1</mml:mn><mml:mrow><mml:mn>1</mml:mn><mml:mspace width=".16em" /><mml:mo>&#x002B;</mml:mo><mml:mspace width=".16em" /><mml:mo>exp</mml:mo><mml:mo stretchy="false">[</mml:mo><mml:mo>-</mml:mo><mml:mspace width=".16em" /><mml:mo stretchy="false">(</mml:mo><mml:msub><mml:mi>W</mml:mi><mml:mn>1</mml:mn></mml:msub><mml:mspace width=".16em" /><mml:mi>X</mml:mi><mml:mspace width=".16em" /><mml:mo>&#x002B;</mml:mo><mml:mspace width=".16em" /><mml:msub><mml:mi>b</mml:mi><mml:mn>1</mml:mn></mml:msub><mml:mo stretchy="false">)</mml:mo><mml:mo stretchy="false">]</mml:mo></mml:mrow></mml:mfrac></mml:mrow></mml:math>
<graphic xlink:href="ol-17-02-1483-g00.tif"/>
</alternatives>
</disp-formula>
<p>Then, the resulting hidden representation <italic>Z</italic> is mapped back to a reconstructed vector <italic>h</italic><sub>w,b</sub>(<italic>x</italic>) in the input space.</p>
<disp-formula>
<alternatives>
<mml:math id="umml2" display="block"><mml:mrow><mml:msub><mml:mi>h</mml:mi><mml:mrow><mml:mi>w</mml:mi><mml:mo>,</mml:mo><mml:mi>b</mml:mi></mml:mrow></mml:msub><mml:mspace width=".16em" /><mml:mo stretchy="false">(</mml:mo><mml:mi>x</mml:mi><mml:mo stretchy="false">)</mml:mo><mml:mspace width=".16em" /><mml:mo>=</mml:mo><mml:mspace width=".16em" /><mml:mi>f</mml:mi><mml:mo stretchy="false">(</mml:mo><mml:msub><mml:mi>W</mml:mi><mml:mn>2</mml:mn></mml:msub><mml:mspace width=".16em" /><mml:mi>Z</mml:mi><mml:mspace width=".16em" /><mml:mo>&#x002B;</mml:mo><mml:mspace width=".16em" /><mml:msub><mml:mi>b</mml:mi><mml:mn>2</mml:mn></mml:msub><mml:mo stretchy="false">)</mml:mo><mml:mspace width=".16em" /><mml:mo>=</mml:mo><mml:mspace width=".16em" /><mml:mfrac><mml:mn>1</mml:mn><mml:mrow><mml:mn>1</mml:mn><mml:mspace width=".16em" /><mml:mo>&#x002B;</mml:mo><mml:mspace width=".16em" /><mml:mo>exp</mml:mo><mml:mo stretchy="false">[</mml:mo><mml:mo>-</mml:mo><mml:mspace width=".16em" /><mml:mo stretchy="false">(</mml:mo><mml:msub><mml:mi>W</mml:mi><mml:mn>2</mml:mn></mml:msub><mml:mspace width=".16em" /><mml:mi>Z</mml:mi><mml:mspace width=".16em" /><mml:mo>&#x002B;</mml:mo><mml:mspace width=".16em" /><mml:msub><mml:mi>b</mml:mi><mml:mn>2</mml:mn></mml:msub><mml:mo stretchy="false">)</mml:mo><mml:mo stretchy="false">]</mml:mo></mml:mrow></mml:mfrac></mml:mrow></mml:math>
<graphic xlink:href="ol-17-02-1483-g01.tif"/>
</alternatives>
</disp-formula>
<p>In the aforementioned formulas, <italic>X</italic> represents the feature expression of the original data, <italic>R</italic> represents real numbers; <italic>N</italic> represents the number of data samples and <italic>M</italic> represents the length of each data sample. <italic>W</italic><sub>1</sub> represents the weight matrix and <italic>b</italic><sub>1</sub> represents the bias of the encoder. <italic>f</italic> represents the encoder activation function. Subsequently, the parameter set, <italic>W</italic><sub>1</sub>, <italic>b</italic><sub>1</sub>, <italic>W</italic><sub>2</sub>, <italic>b</italic><sub>2</sub>, are optimized by minimizing error between the input and reconstructed data. The cost function of the SAE can be written as:</p>
<disp-formula>
<alternatives>
<mml:math id="umml3" display="block"><mml:mrow><mml:mi>J</mml:mi><mml:mo stretchy="false">(</mml:mo><mml:mi>W</mml:mi><mml:mo>,</mml:mo><mml:mspace width=".16em" /><mml:mi>b</mml:mi><mml:mo stretchy="false">)</mml:mo><mml:mspace width=".16em" /><mml:mo>=</mml:mo><mml:mspace width=".16em" /><mml:mfrac><mml:mn>1</mml:mn><mml:mi>N</mml:mi></mml:mfrac><mml:mspace width=".16em" /><mml:munderover><mml:mo>&#x2211;</mml:mo><mml:mrow><mml:mi>i</mml:mi><mml:mo>=</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mi>N</mml:mi></mml:munderover><mml:mrow><mml:mrow><mml:mo>(</mml:mo><mml:mrow><mml:mfrac><mml:mn>1</mml:mn><mml:mn>2</mml:mn></mml:mfrac><mml:msup><mml:mrow><mml:mrow><mml:mo>&#x2016;</mml:mo><mml:mrow><mml:msub><mml:mi>h</mml:mi><mml:mrow><mml:mi>w</mml:mi><mml:mo>,</mml:mo><mml:mi>b</mml:mi></mml:mrow></mml:msub><mml:mspace width=".16em" /><mml:mo stretchy="false">(</mml:mo><mml:mi>x</mml:mi><mml:mo stretchy="false">(</mml:mo><mml:mi>i</mml:mi><mml:mo stretchy="false">)</mml:mo><mml:mo stretchy="false">)</mml:mo><mml:mspace width=".16em" /><mml:mo>-</mml:mo><mml:mspace width=".16em" /><mml:mi>x</mml:mi><mml:mo stretchy="false">(</mml:mo><mml:mi>i</mml:mi><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mo>&#x2016;</mml:mo></mml:mrow></mml:mrow><mml:mn>2</mml:mn></mml:msup></mml:mrow><mml:mo>)</mml:mo></mml:mrow><mml:mspace width=".16em" /><mml:mo>&#x002B;</mml:mo><mml:mspace width=".16em" /><mml:mfrac><mml:mi>&#x03BB;</mml:mi><mml:mn>2</mml:mn></mml:mfrac></mml:mrow><mml:mspace width=".16em" /><mml:munderover><mml:mo>&#x2211;</mml:mo><mml:mrow><mml:mi>l</mml:mi><mml:mo>=</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mn>2</mml:mn></mml:munderover><mml:mrow><mml:munderover><mml:mo>&#x2211;</mml:mo><mml:mrow><mml:mi>i</mml:mi><mml:mo>=</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mrow><mml:msub><mml:mi>S</mml:mi><mml:mi>l</mml:mi></mml:msub></mml:mrow></mml:munderover><mml:mrow><mml:munderover><mml:mo>&#x2211;</mml:mo><mml:mrow><mml:mi>j</mml:mi><mml:mo>=</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mrow><mml:msub><mml:mi>S</mml:mi><mml:mrow><mml:mi>l</mml:mi><mml:mo>&#x002B;</mml:mo><mml:mn>1</mml:mn></mml:mrow></mml:msub></mml:mrow></mml:munderover><mml:mrow><mml:msup><mml:mrow><mml:mrow><mml:mo>(</mml:mo><mml:mrow><mml:msubsup><mml:mi>W</mml:mi><mml:mrow><mml:mi>j</mml:mi><mml:mi>i</mml:mi></mml:mrow><mml:mi>l</mml:mi></mml:msubsup></mml:mrow><mml:mo>)</mml:mo></mml:mrow></mml:mrow><mml:mn>2</mml:mn></mml:msup></mml:mrow><mml:mspace width=".16em" /><mml:mo>&#x002B;</mml:mo><mml:mspace width=".16em" /><mml:mi>&#x03B2;</mml:mi><mml:munderover><mml:mo>&#x2211;</mml:mo><mml:mrow><mml:mi>j</mml:mi><mml:mo>=</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mrow><mml:msub><mml:mi>S</mml:mi><mml:mn>2</mml:mn></mml:msub></mml:mrow></mml:munderover><mml:mrow><mml:mi>K</mml:mi><mml:mi>L</mml:mi><mml:mspace width=".16em" /><mml:mo stretchy="false">(</mml:mo><mml:mi>&#x03C1;</mml:mi><mml:mspace width=".16em" /><mml:mrow><mml:mo>&#x2016;</mml:mo><mml:mrow><mml:msub><mml:mover accent="true"><mml:mi>&#x03C1;</mml:mi><mml:mo>&#x02C6;</mml:mo></mml:mover><mml:mi>j</mml:mi></mml:msub></mml:mrow></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow></mml:mrow></mml:mrow></mml:math>
<graphic xlink:href="ol-17-02-1483-g02.tif"/>
</alternatives>
</disp-formula>
<p>In this formula, <italic>N</italic> represents the number of data samples, <italic>h</italic><sub>w,b</sub><italic>(x(i))</italic> represents the output feature vector, <italic>x(i)</italic> represents the input feature vector, &#x03BB; represents the weight decay parameter, <italic>S</italic><sub>l</sub> represents the number of neurons in layer <italic>l</italic>, &#x03B2; represents the weight of sparsity penalty term, <italic>W</italic><sup>l</sup><sub>ji</sub> represents the weight on the connection between neuron <italic>j</italic> in layer <italic>l</italic>&#x002B;1 and neuron <italic>i</italic> in layer <italic>l, &#x03C1;</italic> represents the sparsity parameter defining the sparsity level and <italic>&#x03C1;<sub>j</sub></italic> represents the average activation of hidden neuron <italic>j</italic>. <sup><inline-formula><mml:math id="umml5" display="block"><mml:mrow><mml:mi>K</mml:mi><mml:mi>L</mml:mi><mml:mo stretchy="false">(</mml:mo><mml:mi>&#x03C1;</mml:mi><mml:mrow><mml:mo>&#x2016;</mml:mo><mml:mrow><mml:mover accent="true"><mml:mi>&#x03C1;</mml:mi><mml:mo>&#x02C6;</mml:mo></mml:mover><mml:mi mathvariant="normal">i</mml:mi></mml:mrow></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:math></inline-formula></sup> represents the Kullback-Leibler divergence between <sup>&#x03C1;</sup> and <sup><inline-formula><mml:math id="umml6" display="block"><mml:mrow><mml:mover accent="true"><mml:mi>&#x03C1;</mml:mi><mml:mo>&#x02C6;</mml:mo></mml:mover><mml:mi mathvariant="normal">i</mml:mi></mml:mrow></mml:math></inline-formula></sup>.</p>
<p>By minimizing <italic>J(W,b)</italic> the optimal parameters of <italic>W</italic> and <italic>b</italic> are updated in the process of coding. The parameters <italic>W</italic> and <italic>b</italic> in each iteration can be updated as:</p>
<disp-formula>
<alternatives>
<mml:math id="umml4" display="block"><mml:mrow><mml:msubsup><mml:mi>W</mml:mi><mml:mrow><mml:mi>j</mml:mi><mml:mi>i</mml:mi></mml:mrow><mml:mi>l</mml:mi></mml:msubsup><mml:mspace width=".16em" /><mml:mo>=</mml:mo><mml:mspace width=".16em" /><mml:msubsup><mml:mi>W</mml:mi><mml:mrow><mml:mi>j</mml:mi><mml:mi>i</mml:mi></mml:mrow><mml:mi>l</mml:mi></mml:msubsup><mml:mspace width=".16em" /><mml:mo>-</mml:mo><mml:mspace width=".16em" /><mml:mi>&#x03B5;</mml:mi><mml:mspace width=".16em" /><mml:mfrac><mml:mo>&#x2202;</mml:mo><mml:mrow><mml:mo>&#x2202;</mml:mo><mml:msubsup><mml:mi>W</mml:mi><mml:mrow><mml:mi>j</mml:mi><mml:mi>i</mml:mi></mml:mrow><mml:mi>l</mml:mi></mml:msubsup></mml:mrow></mml:mfrac><mml:mspace width=".16em" /><mml:mi>J</mml:mi><mml:mo stretchy="false">(</mml:mo><mml:mi>W</mml:mi><mml:mo>,</mml:mo><mml:mspace width=".16em" /><mml:mi>b</mml:mi><mml:mo stretchy="false">)</mml:mo><mml:mspace width=".16em" /><mml:mspace width=".16em" /><mml:mtext>and</mml:mtext><mml:mspace width=".16em" /><mml:mspace width=".16em" /><mml:msup><mml:mi>b</mml:mi><mml:mi>l</mml:mi></mml:msup><mml:mspace width=".16em" /><mml:mo>=</mml:mo><mml:mspace width=".16em" /><mml:msup><mml:mi>b</mml:mi><mml:mi>l</mml:mi></mml:msup><mml:mspace width=".16em" /><mml:mo>-</mml:mo><mml:mi>&#x03B5;</mml:mi><mml:mfrac><mml:mo>&#x2202;</mml:mo><mml:mrow><mml:mo>&#x2202;</mml:mo><mml:msup><mml:mi>b</mml:mi><mml:mi>l</mml:mi></mml:msup></mml:mrow></mml:mfrac><mml:mspace width=".16em" /><mml:mi>J</mml:mi><mml:mo stretchy="false">(</mml:mo><mml:mi>W</mml:mi><mml:mo>,</mml:mo><mml:mspace width=".16em" /><mml:mi>b</mml:mi><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:math>
<graphic xlink:href="ol-17-02-1483-g03.tif"/>
</alternatives>
</disp-formula>
<p>In these formulas, parameter &#x03B5; represents the learning rate; <italic>l</italic> represents the <italic>l</italic><sup>th</sup> layer of the network; <italic>i</italic> and <italic>j</italic> denote the <italic>i</italic><sup>th</sup> and <italic>j</italic><sup>th</sup> neurons of two neighboring layers, respectively. <italic>J(W,b)</italic> represents the cost function of the SAE and &#x2202; represents seeking partial derivative.</p>
<p>Stacked sparse autoencoder (SSAE) is a newly developed DL algorithm and a feed-forward neural network consisting of multiple layers of basic autoencoder. SSAE has successfully been employed to identify visual features in computer vision (<xref rid="b24-ol-0-0-9761" ref-type="bibr">24</xref>,<xref rid="b25-ol-0-0-9761" ref-type="bibr">25</xref>), predict protein solvent accessibility and contact number (<xref rid="b26-ol-0-0-9761" ref-type="bibr">26</xref>), predict protein-protein interactions (<xref rid="b27-ol-0-0-9761" ref-type="bibr">27</xref>) and construct an electronic nose system (<xref rid="b28-ol-0-0-9761" ref-type="bibr">28</xref>). Similar to other DL algorithms, SSAE can efficiently use a deep or layered architecture to identify potential representations from original clinical features, therefore enhancing the classification accuracy (<xref rid="b20-ol-0-0-9761" ref-type="bibr">20</xref>).</p>
<p>However, the performance of classification algorithms may differ from one dataset to another (<xref rid="b29-ol-0-0-9761" ref-type="bibr">29</xref>). Different classification algorithms may be used to develop and compare several models, allowing the best solution to be selected based on the dataset (<xref rid="b29-ol-0-0-9761" ref-type="bibr">29</xref>). In the current study, the GA and ReliefF methods were used to select high discriminative features and three commonly used machine leaning algorithms, BP, ELM and SSAE were selected as the classification algorithms. The performance of these models was compared using evaluation metrics, including classification accuracy, specificity, F-measure and AUC. The model with the best performance may be considered as a classifier to be used when developing a DSS to diagnose MM.</p>
</sec>
<sec sec-type="materials|methods">
<title>Materials and methods</title>
<sec>
<title/>
<sec>
<title>Data source</title>
<p>To facilitate comparisons with previous studies, the current study used a dataset obtained from the University of California (Irvine, CA, USA) machine learning database (<xref rid="b7-ol-0-0-9761" ref-type="bibr">7</xref>). The original dataset includes 324 participants. For each participant, 34 physiological variables were recorded, as presented in <xref rid="tI-ol-0-0-9761" ref-type="table">Table I</xref>.</p>
</sec>
<sec>
<title>Data preprocessing and feature selection</title>
<p>The original dataset was classified into two groups by experienced pathologists: 97 patients with MM and 227 healthy participants. For each participant, 34 variables were recorded and no data were missing. Multiplicity of the features may lead to over-training a model. Therefore, a frequently used preprocessing method is feature selection, in which irrelevant, weakly relevant or less discriminative features are removed. Feature selection can increase the accuracy of the resulting model (<xref rid="b30-ol-0-0-9761" ref-type="bibr">30</xref>,<xref rid="b31-ol-0-0-9761" ref-type="bibr">31</xref>). Previously, a variety of feature selection algorithms have been developed to perform feature selection (<xref rid="b32-ol-0-0-9761" ref-type="bibr">32</xref>,<xref rid="b33-ol-0-0-9761" ref-type="bibr">33</xref>). The current study used the GA (<xref rid="b34-ol-0-0-9761" ref-type="bibr">34</xref>) and the ReliefF algorithm (<xref rid="b30-ol-0-0-9761" ref-type="bibr">30</xref>) with MATLAB software (R2016b 9.1.0.441655); MathWorks, Natick, MA, USA). The parameters of GA were set as follows: Number of generations=100, population size=20, encoding length=34 and the crossover rate=2. While ReliefF methods were used to select high discriminative features, the parameters of ReliefF were set as follows: Number of iterations=323 and the number of close samples=95. The important features selected by GA and ReliefF were used as inputs to train the diagnosis models. For training these diagnosis models, the dataset was randomly divided into training (70&#x0025;) and testing (30&#x0025;) data.</p>
</sec>
<sec>
<title>Construction of the diagnosis models</title>
<p>Previous studies have demonstrated that the most effective model is not always easily identified. Therefore, testing various diagnosis models is required to select the most effective one. In the current study, a number of diagnosis models, including BP, ELM and SSAE, were developed using MATLAB software. Selected features were used as the inputs for these recognition models.</p>
<p>The algorithm principle for BP can be reviewed in a previous study (<xref rid="b35-ol-0-0-9761" ref-type="bibr">35</xref>). In the current study, the parameters of BP were set as follows: Size of hidden layers=50, transfer function of hidden layers=&#x2018;tansig&#x2019;, transfer function of output layer=&#x2018;purelin&#x2019;, training function=&#x2018;trainlm&#x2019;, epoch =1,000, goal of training accuracy=0.1 and the learning rate=0.1. Other associated parameters were based on the default values of MATLAB.</p>
<p>In addition, an ELM algorithm was selected to build a pattern recognition model. ELM is a learning algorithm, which is developed on the basis of a single-hidden layer feed-forward neural network (<xref rid="b15-ol-0-0-9761" ref-type="bibr">15</xref>). ELM can achieve a faster learning speed and improved generalization capability compared with other pattern recognition models by distributing the input weights and hidden layer biases randomly and determining the output weights through a Moore-Penrose generalized inverse operation of the weight matrices in hidden layers (<xref rid="b36-ol-0-0-9761" ref-type="bibr">36</xref>). Further information regarding ELM can be reviewed in a previous study (<xref rid="b37-ol-0-0-9761" ref-type="bibr">37</xref>). In the current study, the number of hidden nodes was set to 30 and the sigmoidal function was adopted as the transfer function.</p>
<p>The SSAE is a feed-forward neural network consisting of multiple layers of basic autoencoder (<xref rid="b28-ol-0-0-9761" ref-type="bibr">28</xref>). To reduce the dimensions and extract high-level abstract features from the input without labels, two autoencoders were stacked to generate an SSAE in the current study. When using an SSAE, features calculated by the first autoencoder are used as the inputs for the second autoencoder&#x0027;s hidden layer, as presented in <xref rid="f2-ol-0-0-9761" ref-type="fig">Fig. 2</xref>. When the expected error rate is achieved by the first autoencoder, the high-level abstract features are extracted by the second autoencoder (<xref rid="b38-ol-0-0-9761" ref-type="bibr">38</xref>,<xref rid="b39-ol-0-0-9761" ref-type="bibr">39</xref>). In the current study, a supervised model, the Softmax classifier (SMC), was connected to the end of the trained SSAE to identify the classification task. To train the SMC, the high-level abstract features extracted by the SSAE were used as the inputs for the softmax layer. Following training, the SMC was stacked with SSAE to establish a deep neural network (<xref rid="f2-ol-0-0-9761" ref-type="fig">Fig. 2</xref>). Further information regarding the SMC can be reviewed in previous studies (<xref rid="b38-ol-0-0-9761" ref-type="bibr">38</xref>,<xref rid="b40-ol-0-0-9761" ref-type="bibr">40</xref>).</p>
<p>The parameters of the first autoencoder were set as follows: When using raw data as the input the size of the hidden representation=30, when using data generated by dimension reducing processing as the input the size of the hidden representation=20, weight regularization for loss function=0.001, sparsity regularization for loss function=4 and sparsity proportion=0.05. The parameters of the second autoencoder were set as follows: Size of the hidden representation=10, weight regularization for loss function=0.001, sparsity regularization for loss function=4 and sparsity proportion=10. In addition, the current study selected the linear transfer function.</p>
</sec>
<sec>
<title>Performance measure</title>
<p>In order to evaluate and select the most accurate diagnostic model, confusion matrixes were adopted to calculate the sensitivity, specificity, precision and accuracy of each diagnostic model. Occasionally, contradictions occurred between the sensitivity and precision, therefore the F-measure was used to weight and average the two values: F=(2 &#x00D7; precision &#x00D7; sensitivity)/(precision&#x002B;sensitivity).</p>
<p>In addition, a receiver operating characteristic (ROC) curve and the area under the ROC curve (AUC) were used to evaluate the performance of the diagnostic models. ROC analysis evaluated the capacity of the diagnostic models to distinguish between MM and healthy participants. All calculations were based on standard equations published in previous studies (<xref rid="b31-ol-0-0-9761" ref-type="bibr">31</xref>,<xref rid="b41-ol-0-0-9761" ref-type="bibr">41</xref>). Furthermore, the central processing unit (CPU) time was recorded and compared between the SSAE and GA&#x002B;SSAE diagnostic models to indicate the computational complexities of each model.</p>
</sec>
</sec>
</sec>
<sec sec-type="results">
<title>Results</title>
<sec>
<title/>
<sec>
<title>Identifying variables that are important for MM diagnosis by feature selection</title>
<p>The original dataset contained 34 variables for each participant. GA and ReliefF were performed to select the most significant features for diagnosing MM. 19 relevant features were selected as the feature subset based on GA. However, only five relevant features with a weight value &#x003E;0.10 were selected as the feature subset based on the ReliefF algorithm. The feature subset based on GA included: Age, gender, city, duration of asbestos exposure, diagnosis method, cytology, ache on chest, weakness, smoking habit, white blood cell level, hemoglobin level, blood lactose dehydrogenase level, albumin level, glucose level, pleural lactose dehydrogenase level, pleural protein level, pleural glucose level, cell count and pleural level of acidity.</p>
</sec>
<sec>
<title>Comparing the diagnostic models</title>
<p>The current study compared the performance of the diagnosis models using four evaluation measures, as illustrated in <xref rid="f3-ol-0-0-9761" ref-type="fig">Figs. 3</xref>&#x2013;<xref rid="f6-ol-0-0-9761" ref-type="fig">6</xref>. Overall, SSAE and GA&#x002B;SSAE demonstrated a higher performance compared with the other diagnosis models evaluated.</p>
<p>When all 34 variables of the original data set were used to diagnosis MM, the SSAE model demonstrated the highest accuracy (100.00&#x0025;), while other models demonstrated an accuracy &#x003C;95&#x0025;. When the 19 variables selected by GA were used as the input variables, both the GA&#x002B;BP and GA&#x002B;ELM models achieved an accuracy of 98.00&#x0025;. However, GA&#x002B;SSAE demonstrated the highest accuracy (100.00&#x0025;). When the five variables selected by the ReliefF algorithm were used as the input variables, the accuracy of the ReliefF&#x002B;SSAE model decreased to 98.00&#x0025;, while the other models demonstrated accuracies &#x003C;92&#x0025; (<xref rid="f3-ol-0-0-9761" ref-type="fig">Fig. 3</xref>).</p>
<p>The SSAE model demonstrated a specificity of 100.00&#x0025; when all 34 variables of the original data set were used as the input variables. Similarly, when the 19 variables selected by GA were used as the inputs, the GA&#x002B;SSAE model demonstrated a specificity of 100.00&#x0025;. By contrast, the ReliefF&#x002B;BP model achieved the lowest specificity of 89.10&#x0025; (<xref rid="f4-ol-0-0-9761" ref-type="fig">Fig. 4</xref>).</p>
<p>When all 34 variables of the original data set were used as the inputs, the SSAE model demonstrated the highest F-measure value (100.00&#x0025;), while ELM achieved the lowest F-measure value (90.91&#x0025;). When the 19 variables selected by GA were used as the input variables, both GA&#x002B;BP and GA&#x002B;ELM models demonstrated an F-measure value of 97.10&#x0025;. With the same input variables, the GA&#x002B;SSAE model achieved the highest F-measure value (100.00&#x0025;). However, when the five variables selected by the ReliefF algorithm were used as the input variables, the F-measure value of the ReliefF&#x002B;ELM model decreased markedly to 5.48&#x0025;, while the F-measure value of the ReliefF&#x002B;SSAE model was 97.12&#x0025; (<xref rid="f5-ol-0-0-9761" ref-type="fig">Fig. 5</xref>).</p>
<p>An evaluation was performed to compare the discriminatory power of the BP, ELM and SSAE models, with or without feature selection, as illustrated in <xref rid="f6-ol-0-0-9761" ref-type="fig">Figs. 6</xref> and <xref rid="f7-ol-0-0-9761" ref-type="fig">7</xref>. An ROC curve and the AUC were used to indicate the effectiveness of a diagnostic model to discriminate between patients with MM and healthy participants. The current study revealed that the SSAE and GA&#x002B;SSAE models demonstrated the highest diagnostic power compared with the other models, as presented in <xref rid="f7-ol-0-0-9761" ref-type="fig">Fig. 7</xref>.</p>
</sec>
<sec>
<title>Performing feature selection does not exclusively improve the performance of diagnostic models</title>
<p>As demonstrated in <xref rid="f3-ol-0-0-9761" ref-type="fig">Figs. 3</xref>&#x2013;<xref rid="f7-ol-0-0-9761" ref-type="fig">7</xref>, performing GA markedly enhanced the classification performance of BP. Compared with BP alone, GA&#x002B;BP achieved a higher accuracy (98.00 vs. 94.40&#x0025;), specificity (98.4 vs. 95.30&#x0025;), F-measure (97.10 vs. 92.73&#x0025;) and AUC (97.75 vs. 94.72&#x0025;). Whereas, compared with BP alone, performing a ReliefF algorithm generated a lower accuracy (90.80 vs. 94.90&#x0025;), specificity (89.10 vs. 95.30&#x0025;), F-measure (87.69 vs. 92.73&#x0025;) and AUC (80 vs. 94.72&#x0025;).</p>
<p>As demonstrated in <xref rid="f4-ol-0-0-9761" ref-type="fig">Fig. 4</xref>, performing GA or ReliefF prior to training the ELM classifiers markedly improved the specificity. Compared with ELM alone, the GA&#x002B;ELM model revealed a higher accuracy (98.00&#x0025;), specificity (98.40&#x0025;), F-measure (97.10&#x0025;) and AUC value (97.75&#x0025;). However, comparison of the accuracy, F-measure and AUC demonstrated that performing the ReliefF algorithm reduced the performance of the ELM model on the same dataset.</p>
<p>The performance of SSAE in diagnosing MM was compared with that of the GA&#x002B;SSAE and ReliefF&#x002B;SSAE models. In comparison with SSAE without feature selection, performing the ReliefF algorithm achieved a worse performance. However, both SSAE and SSAE&#x002B;GA achieved an accuracy, specificity, F-measure and AUC of 100&#x0025;. The average CPU time of the SSAE diagnostic model following GA feature selection was compared with the average CPU time of the baseline SSAE diagnostic model without feature selection. Each model was performed ten times. the mean CPU time of SSAE was 33.2 sec. However, following GA feature selection, the mean CPU time of SSAE decreased to 28.8 sec.</p>
</sec>
</sec>
</sec>
<sec sec-type="discussion">
<title>Discussion</title>
<p>The definitive diagnosis of MM is significant at both the individual and public health level, and has important medicolegal significance due to diagnosis-related compensation issues (<xref rid="b6-ol-0-0-9761" ref-type="bibr">6</xref>). However, the definitive diagnosis of MM is complicated due to its composite epithelial/mesenchymal pattern and its low occurrence frequency (<xref rid="b6-ol-0-0-9761" ref-type="bibr">6</xref>). To improve the diagnosis of MM at an early stage, the current study designed and implemented a diagnostic model based on SSAE algorithms.</p>
<p>To the best of our knowledge, no previous studies have developed MM diagnostic models using SSAE algorithms. However, previous studies have successfully applied SSAE in processing diagnosis information (<xref rid="b38-ol-0-0-9761" ref-type="bibr">38</xref>,<xref rid="b40-ol-0-0-9761" ref-type="bibr">40</xref>,<xref rid="b42-ol-0-0-9761" ref-type="bibr">42</xref>). A number of previous studies have compared the performance of different statistical and machine learning techniques in the field of medical diagnosis; the results revealed that the performance of SSAE is higher than several other similar techniques (<xref rid="b42-ol-0-0-9761" ref-type="bibr">42</xref>,<xref rid="b43-ol-0-0-9761" ref-type="bibr">43</xref>), a finding that is consistent with the current study.</p>
<p>However, the performance and precision of a classification algorithm may vary depending on the dataset (<xref rid="b44-ol-0-0-9761" ref-type="bibr">44</xref>). For a specific dataset, it is difficult to identify which classification algorithm has the highest performance without performing a comparison. In addition, feature selection methods exhibit a marked influence on the performance of a classification algorithm; a feature subset that improves the performance of one classification algorithm may not improve the performance of a different classification algorithm. To generate a diagnostic model for MM and evaluate its performance, three different algorithms with or without feature selection were applied on the same training dataset and their performances were compared in the current study.</p>
<p>The results indicated that the SSAE and GA&#x002B;SSAE models exhibited the highest overall performance; the accuracy, specificity, F-measure and AUC for both models were 100&#x0025;. However, following feature selection with GA a decrease was identified in the CPU time for training the SSAE diagnostic model compared with the baseline SSAE diagnostic model without GA. Furthermore, the GA&#x002B;SSAE model required fewer variables to achieve the same performance as SSAE. The ReliefF&#x002B;ELM diagnosis model exhibited the worst performance on the dataset used in the current study, with an accuracy, F-measure and AUC of 65.30, 5.48 and 50.00&#x0025;, respectively. In addition, performing feature selection did not exclusively produce an improved performance. The accuracy of the ReliefF&#x002B;ELM model was 65.3&#x0025;, while the accuracy of ELM without feature selection was 93&#x0025;. Similarly, the ReliefF algorithm reduced the performance of the SSAE model.</p>
<p>The results from the current study were compared with a previous study that used the same training dataset (<xref rid="b7-ol-0-0-9761" ref-type="bibr">7</xref>). The previous study compared the performance of three classification methods and identified that the classification accuracies were 96.30, 94.41 and 91.14&#x0025; for PNN, multilayer neural network and learning vector quantization structures, respectively. The highest classification accuracy was obtained using a PNN (with 3 fully connected layers, response surface method) structure (96.30&#x0025;) (<xref rid="b7-ol-0-0-9761" ref-type="bibr">7</xref>,<xref rid="b45-ol-0-0-9761" ref-type="bibr">45</xref>). In comparison, the SSAE and GA&#x002B;SSAE methods achieved a classification accuracy of 100&#x0025; on the same training dataset, in the current study.</p>
<p>The current study demonstrates the effectiveness of DL in the diagnosis of cancer. A pathologist&#x0027;s clinical impression and diagnosis of MM is based on contextual factors, whereas a GA&#x002B;SSAE model has the ability to diagnose MM with high accuracy and may augment clinical decision-making. Furthermore, this fast and scalable method may be applied to other clinical datasets for the diagnosis or prediction of other cancer types.</p>
<p>In conclusion, the current study has aimed to improve the diagnosis of MM by evaluating three deep learning algorithms, using a dataset containing 97 patients with MM and 227 healthy participants. To avoid over-training and improve the classification accuracy of a diagnosis model, ReliefF and GA feature selection algorithms were implemented to remove the irrelevant or weakly relevant features. The current study identified that the GA&#x002B;SSAE algorithm exhibited the highest performance in all evaluation criteria and required the smallest number of variables. The GA&#x002B;SSAE algorithm-based DSS may contribute to the definitive diagnosis of MM. Consequently, the current study may assist pathologists with the diagnosis of MM by providing a system that can achieve optimal diagnostic performance. In addition, it may facilitate the screening of high-risk individuals in regions where asbestos exposure is common.</p>
</sec>
</body>
<back>
<ack>
<title>Acknowledgements</title>
<p>Not applicable.</p>
</ack>
<sec>
<title>Funding</title>
<p>The current study was supported by the Foundation and Frontier Research Project of Chongqing (grant no. cstc2016jcyjA0526), the Science and Technology Research Project of Chongqing Municipal Education Committee (grant no. KJ1600519), the Technology Innovation Project of Social Undertakings and Livelihood Security of Chongqing (grant no. cstc2017shmsA30016) and the Postgraduate Science and Technology Innovation Project of Chongqing (grant nos. CYS17215 and CYS17203).</p>
</sec>
<sec>
<title>Availability of data and materials</title>
<p>The datasets analyzed during the current study are available in the University of California (Irvine, CA, USA) machine learning database, <uri xlink:href="https://archive.ics.uci.edu/ml/datasets/Mesothelioma">https://archive.ics.uci.edu/ml/datasets/Mesothelioma</uri> &#x0025;C3&#x0025;A2&#x0025;E2&#x0025;82&#x0025;AC&#x0025;E2&#x0025;84&#x0025;A2s&#x002B;disease&#x002B;data&#x002B;set&#x002B;.</p>
</sec>
<sec>
<title>Authors&#x0027; contributions</title>
<p>ZY and XH contributed to the conception of the study. XH contributed to analysis and manuscript preparation. ZY and XH performed the data analyses and wrote the manuscript. ZY and XH helped perform the analysis with constructive discussions.</p>
</sec>
<sec>
<title>Ethics approval and consent to participate</title>
<p>The current study used a dataset obtained from a public database. The data submitters have gain informed consent for publication of the dataset from participants at the point of recruitment to the trial.</p>
</sec>
<sec>
<title>Patient consent for publication</title>
<p>Not applicable.</p>
</sec>
<sec>
<title>Competing interests</title>
<p>The authors declare that they have no competing interests.</p>
</sec>
<ref-list>
<title>References</title>
<ref id="b1-ol-0-0-9761"><label>1</label><element-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Remon</surname><given-names>J</given-names></name><name><surname>Lianes</surname><given-names>P</given-names></name><name><surname>Martinez</surname><given-names>S</given-names></name><name><surname>Velasco</surname><given-names>M</given-names></name><name><surname>Querol</surname><given-names>R</given-names></name><name><surname>Zanui</surname><given-names>M</given-names></name></person-group><article-title>Malignant mesothelioma: New insights into a rare disease</article-title><source>Cancer Treat Rev</source><volume>39</volume><fpage>584</fpage><lpage>591</lpage><year>2013</year><pub-id pub-id-type="doi">10.1016/j.ctrv.2012.12.005</pub-id><pub-id pub-id-type="pmid">23276688</pub-id></element-citation></ref>
<ref id="b2-ol-0-0-9761"><label>2</label><element-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Remon</surname><given-names>J</given-names></name><name><surname>Reguart</surname><given-names>N</given-names></name><name><surname>Corral</surname><given-names>J</given-names></name><name><surname>Lianes</surname><given-names>P</given-names></name></person-group><article-title>Malignant pleural mesothelioma: New hope in the horizon with novel therapeutic strategies</article-title><source>Cancer Treat Rev</source><volume>41</volume><fpage>27</fpage><lpage>34</lpage><year>2015</year><pub-id pub-id-type="doi">10.1016/j.ctrv.2014.10.007</pub-id><pub-id pub-id-type="pmid">25467107</pub-id></element-citation></ref>
<ref id="b3-ol-0-0-9761"><label>3</label><element-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Thellung</surname><given-names>S</given-names></name><name><surname>Favoni</surname><given-names>RE</given-names></name><name><surname>Wurth</surname><given-names>R</given-names></name><name><surname>Nizzari</surname><given-names>M</given-names></name><name><surname>Pattarozzi</surname><given-names>A</given-names></name><name><surname>Daga</surname><given-names>A</given-names></name><name><surname>Florio</surname><given-names>T</given-names></name><name><surname>Barbieri</surname><given-names>F</given-names></name></person-group><article-title>Molecular pharmacology of malignant spleural amesothelioma: Challenges and perspectives from preclinical and clinical studies</article-title><source>Curr Drug Targets</source><volume>17</volume><fpage>824</fpage><lpage>849</lpage><year>2016</year><pub-id pub-id-type="doi">10.2174/1389450116666150804110714</pub-id><pub-id pub-id-type="pmid">26240051</pub-id></element-citation></ref>
<ref id="b4-ol-0-0-9761"><label>4</label><element-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Zucali</surname><given-names>PA</given-names></name><name><surname>Ceresoli</surname><given-names>GL</given-names></name><name><surname>De Vincenzo</surname><given-names>F</given-names></name><name><surname>Simonelli</surname><given-names>M</given-names></name><name><surname>Lorenzi</surname><given-names>E</given-names></name><name><surname>Gianoncelli</surname><given-names>L</given-names></name><name><surname>Santoro</surname><given-names>A</given-names></name></person-group><article-title>Advances in the biology of malignant pleural mesothelioma</article-title><source>Cancer Treat Rev</source><volume>37</volume><fpage>543</fpage><lpage>558</lpage><year>2011</year><pub-id pub-id-type="doi">10.1016/j.ctrv.2011.01.001</pub-id><pub-id pub-id-type="pmid">21288646</pub-id></element-citation></ref>
<ref id="b5-ol-0-0-9761"><label>5</label><element-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Ascoli</surname><given-names>V</given-names></name><name><surname>Murer</surname><given-names>B</given-names></name><name><surname>Nottegar</surname><given-names>A</given-names></name><name><surname>Luchini</surname><given-names>C</given-names></name><name><surname>Carella</surname><given-names>R</given-names></name><name><surname>Calabrese</surname><given-names>F</given-names></name><name><surname>Lunardi</surname><given-names>F</given-names></name><name><surname>Cozzi</surname><given-names>I</given-names></name><name><surname>Righi</surname><given-names>L</given-names></name></person-group><article-title>What&#x0027;s new in mesothelioma</article-title><source>Pathologica</source><volume>110</volume><fpage>12</fpage><lpage>28</lpage><year>2018</year><pub-id pub-id-type="pmid">30259910</pub-id></element-citation></ref>
<ref id="b6-ol-0-0-9761"><label>6</label><element-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Ascoli</surname><given-names>V</given-names></name></person-group><article-title>Pathologic diagnosis of malignant mesothelioma: Chronological prospect and advent of recommendations and guidelines</article-title><source>Ann Ist super sanita</source><volume>51</volume><fpage>52</fpage><lpage>59</lpage><year>2015</year><pub-id pub-id-type="pmid">25857384</pub-id></element-citation></ref>
<ref id="b7-ol-0-0-9761"><label>7</label><element-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Er</surname><given-names>O</given-names></name><name><surname>Tanrikulu</surname><given-names>AC</given-names></name><name><surname>Abakay</surname><given-names>A</given-names></name><name><surname>Temurtas</surname><given-names>F</given-names></name></person-group><article-title>An approach based on probabilistic neural network for diagnosis of Mesothelioma&#x0027;s disease</article-title><source>Comput Electr Eng</source><volume>38</volume><fpage>75</fpage><lpage>81</lpage><year>2012</year><pub-id pub-id-type="doi">10.1016/j.compeleceng.2011.09.001</pub-id></element-citation></ref>
<ref id="b8-ol-0-0-9761"><label>8</label><element-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Avci</surname><given-names>D</given-names></name><name><surname>Dogantekin</surname><given-names>A</given-names></name></person-group><article-title>An expert diagnosis system for parkinson disease based on genetic algorithm-wavelet kernel-extreme learning machine</article-title><source>Parkins Dis</source><volume>2016</volume><fpage>5264743</fpage><year>2016</year></element-citation></ref>
<ref id="b9-ol-0-0-9761"><label>9</label><element-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Liu</surname><given-names>X</given-names></name><name><surname>Wang</surname><given-names>X</given-names></name><name><surname>Su</surname><given-names>Q</given-names></name><name><surname>Zhang</surname><given-names>M</given-names></name><name><surname>Zhu</surname><given-names>Y</given-names></name><name><surname>Wang</surname><given-names>Q</given-names></name><name><surname>Wang</surname><given-names>Q</given-names></name></person-group><article-title>A hybrid classification system for heart disease diagnosis based on the RFRS method</article-title><source>Comput Math Methods Med</source><volume>2017</volume><fpage>8272091</fpage><year>2017</year><pub-id pub-id-type="doi">10.1155/2017/8272091</pub-id><pub-id pub-id-type="pmid">28127385</pub-id></element-citation></ref>
<ref id="b10-ol-0-0-9761"><label>10</label><element-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Chaisaowong</surname><given-names>K</given-names></name><name><surname>Aach</surname><given-names>T</given-names></name><name><surname>Jager</surname><given-names>P</given-names></name><name><surname>Vogel</surname><given-names>S</given-names></name><name><surname>Knepper</surname><given-names>A</given-names></name><name><surname>Kraus</surname><given-names>T</given-names></name></person-group><article-title>Computer-assisted diagnosis for early stage pleural mesothelioma: Towards automated detection and quantitative assessment of pleural thickening from thoracic CT images</article-title><source>Methods Inf Med</source><volume>46</volume><fpage>324</fpage><lpage>331</lpage><year>2007</year><pub-id pub-id-type="doi">10.1160/ME9050</pub-id><pub-id pub-id-type="pmid">17492119</pub-id></element-citation></ref>
<ref id="b11-ol-0-0-9761"><label>11</label><element-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Chen</surname><given-names>M</given-names></name><name><surname>Helm</surname><given-names>E</given-names></name><name><surname>Joshi</surname><given-names>N</given-names></name><name><surname>Gleeson</surname><given-names>F</given-names></name><name><surname>Brady</surname><given-names>M</given-names></name></person-group><article-title>Computer-aided volumetric assessment of malignant pleural mesothelioma on CT using a random walk-based method</article-title><source>Int J Comput Assist Radiol Surg</source><volume>12</volume><fpage>529</fpage><lpage>538</lpage><year>2017</year><pub-id pub-id-type="doi">10.1007/s11548-016-1511-3</pub-id><pub-id pub-id-type="pmid">28028655</pub-id></element-citation></ref>
<ref id="b12-ol-0-0-9761"><label>12</label><element-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Armato</surname><given-names>SG</given-names><suffix>III</suffix></name><name><surname>Sensakovic</surname><given-names>WF</given-names></name></person-group><article-title>Automated lung segmentation for thoracic CT impact on computer-aided diagnosis</article-title><source>Acad Radiol</source><volume>11</volume><fpage>1011</fpage><lpage>1021</lpage><year>2004</year><pub-id pub-id-type="doi">10.1016/j.acra.2004.06.005</pub-id><pub-id pub-id-type="pmid">15350582</pub-id></element-citation></ref>
<ref id="b13-ol-0-0-9761"><label>13</label><element-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Jiang</surname><given-names>R</given-names></name><name><surname>You</surname><given-names>R</given-names></name><name><surname>Pei</surname><given-names>XQ</given-names></name><name><surname>Zou</surname><given-names>X</given-names></name><name><surname>Zhang</surname><given-names>MX</given-names></name><name><surname>Wang</surname><given-names>TM</given-names></name><name><surname>Sun</surname><given-names>R</given-names></name><name><surname>Luo</surname><given-names>DH</given-names></name><name><surname>Huang</surname><given-names>PY</given-names></name><name><surname>Chen</surname><given-names>QY</given-names></name><etal/></person-group><article-title>Development of a ten-signature classifier using a support vector machine integrated approach to subdivide the M1 stage into M1a and M1b stages of nasopharyngeal carcinoma with synchronous metastases to better predict patients&#x0027; survival</article-title><source>Oncotarget</source><volume>7</volume><fpage>3645</fpage><lpage>3657</lpage><year>2016</year><pub-id pub-id-type="pmid">26636646</pub-id></element-citation></ref>
<ref id="b14-ol-0-0-9761"><label>14</label><element-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Huang</surname><given-names>MW</given-names></name><name><surname>Chen</surname><given-names>CW</given-names></name><name><surname>Lin</surname><given-names>WC</given-names></name><name><surname>Ke</surname><given-names>SW</given-names></name><name><surname>Tsai</surname><given-names>CF</given-names></name></person-group><article-title>SVM and SVM ensembles in breast cancer prediction</article-title><source>PLoS One</source><volume>12</volume><fpage>e0161501</fpage><year>2017</year><pub-id pub-id-type="doi">10.1371/journal.pone.0161501</pub-id><pub-id pub-id-type="pmid">28060807</pub-id></element-citation></ref>
<ref id="b15-ol-0-0-9761"><label>15</label><element-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Huang</surname><given-names>G</given-names></name><name><surname>Huang</surname><given-names>GB</given-names></name><name><surname>Song</surname><given-names>S</given-names></name><name><surname>You</surname><given-names>K</given-names></name></person-group><article-title>Trends in extreme learning machines: A review</article-title><source>Neural Netw</source><volume>61</volume><fpage>32</fpage><lpage>48</lpage><year>2015</year><pub-id pub-id-type="doi">10.1016/j.neunet.2014.10.001</pub-id><pub-id pub-id-type="pmid">25462632</pub-id></element-citation></ref>
<ref id="b16-ol-0-0-9761"><label>16</label><element-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Ortiz</surname><given-names>A</given-names></name><name><surname>Munilla</surname><given-names>J</given-names></name><name><surname>Gorriz</surname><given-names>JM</given-names></name><name><surname>Ramirez</surname><given-names>J</given-names></name></person-group><article-title>Ensembles of deep learning architectures for the early diagnosis of the Alzheimer&#x0027;s disease</article-title><source>Int J Neural Syst</source><volume>26</volume><fpage>1650025</fpage><year>2016</year><pub-id pub-id-type="doi">10.1142/S0129065716500258</pub-id><pub-id pub-id-type="pmid">27478060</pub-id></element-citation></ref>
<ref id="b17-ol-0-0-9761"><label>17</label><element-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Komura</surname><given-names>D</given-names></name><name><surname>Ishikawa</surname><given-names>S</given-names></name></person-group><article-title>Machine learning methods for histopathological image analysis</article-title><source>Comput Struct Biotechnol J</source><volume>16</volume><fpage>34</fpage><lpage>42</lpage><year>2018</year><pub-id pub-id-type="doi">10.1016/j.csbj.2018.01.001</pub-id><pub-id pub-id-type="pmid">30275936</pub-id></element-citation></ref>
<ref id="b18-ol-0-0-9761"><label>18</label><element-citation publication-type="journal"><person-group person-group-type="author"><name><surname>LeCun</surname><given-names>Y</given-names></name><name><surname>Bengio</surname><given-names>Y</given-names></name><name><surname>Hinton</surname><given-names>G</given-names></name></person-group><article-title>Deep learning</article-title><source>Nature</source><volume>521</volume><fpage>436</fpage><lpage>444</lpage><year>2015</year><pub-id pub-id-type="doi">10.1038/nature14539</pub-id><pub-id pub-id-type="pmid">26017442</pub-id></element-citation></ref>
<ref id="b19-ol-0-0-9761"><label>19</label><element-citation publication-type="journal"><person-group person-group-type="author"><name><surname>van Tulder</surname><given-names>G</given-names></name><name><surname>de Bruijne</surname><given-names>M</given-names></name></person-group><article-title>Combining generative and discriminative representation learning for lung CT analysis with convolutional restricted boltzmann machines</article-title><source>IEEE Trans Med Imag</source><volume>35</volume><fpage>1262</fpage><lpage>1272</lpage><year>2016</year><pub-id pub-id-type="doi">10.1109/TMI.2016.2526687</pub-id></element-citation></ref>
<ref id="b20-ol-0-0-9761"><label>20</label><element-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Shin</surname><given-names>HC</given-names></name><name><surname>Orton</surname><given-names>MR</given-names></name><name><surname>Collins</surname><given-names>DJ</given-names></name><name><surname>Doran</surname><given-names>SJ</given-names></name><name><surname>Leach</surname><given-names>MO</given-names></name></person-group><article-title>Stacked autoencoders for unsupervised feature learning and multiple organ detection in a pilot study using 4D patient data</article-title><source>IEEE Trans Pattern Anal Mach Intell</source><volume>35</volume><fpage>1930</fpage><lpage>1943</lpage><year>2013</year><pub-id pub-id-type="doi">10.1109/TPAMI.2012.277</pub-id><pub-id pub-id-type="pmid">23787345</pub-id></element-citation></ref>
<ref id="b21-ol-0-0-9761"><label>21</label><element-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Brosch</surname><given-names>T</given-names></name><name><surname>Tam</surname><given-names>R</given-names></name></person-group><article-title>Efficient training of convolutional deep belief networks in the frequency domain for application to high-resolution 2D and 3D images</article-title><source>Neural Comput</source><volume>27</volume><fpage>211</fpage><lpage>227</lpage><year>2015</year><pub-id pub-id-type="doi">10.1162/NECO_a_00682</pub-id><pub-id pub-id-type="pmid">25380341</pub-id></element-citation></ref>
<ref id="b22-ol-0-0-9761"><label>22</label><element-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Esteva</surname><given-names>A</given-names></name><name><surname>Kuprel</surname><given-names>B</given-names></name><name><surname>Novoa</surname><given-names>RA</given-names></name><name><surname>Ko</surname><given-names>J</given-names></name><name><surname>Swetter</surname><given-names>SM</given-names></name><name><surname>Blau</surname><given-names>HM</given-names></name><name><surname>Thrun</surname><given-names>S</given-names></name></person-group><article-title>Dermatologist-level classification of skin cancer with deep neural networks</article-title><source>Nature</source><volume>542</volume><fpage>115</fpage><lpage>118</lpage><year>2017</year><pub-id pub-id-type="doi">10.1038/nature21056</pub-id><pub-id pub-id-type="pmid">28117445</pub-id></element-citation></ref>
<ref id="b23-ol-0-0-9761"><label>23</label><element-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Grewal</surname><given-names>PS</given-names></name><name><surname>Oloumi</surname><given-names>F</given-names></name><name><surname>Rubin</surname><given-names>U</given-names></name><name><surname>Tennant</surname><given-names>MTS</given-names></name></person-group><article-title>Deep learning in ophthalmology: A review</article-title><source>Can J Ophtalmol</source><volume>53</volume><fpage>309</fpage><lpage>313</lpage><year>2018</year><pub-id pub-id-type="doi">10.1016/j.jcjo.2018.04.019</pub-id></element-citation></ref>
<ref id="b24-ol-0-0-9761"><label>24</label><element-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Yan</surname><given-names>Y</given-names></name><name><surname>Tan</surname><given-names>Z</given-names></name><name><surname>Su</surname><given-names>N</given-names></name><name><surname>Zhao</surname><given-names>C</given-names></name></person-group><article-title>Building extraction based on an optimized stacked sparse autoencoder of structure and training samples using LIDAR DSM and optical images</article-title><source>Sensors (Basel)</source><volume>17</volume><fpage>E1957</fpage><comment>2017.25</comment><pub-id pub-id-type="doi">10.3390/s17091957</pub-id><pub-id pub-id-type="pmid">28837118</pub-id></element-citation></ref>
<ref id="b25-ol-0-0-9761"><label>25</label><element-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Jia</surname><given-names>W</given-names></name><name><surname>Yang</surname><given-names>M</given-names></name><name><surname>Wang</surname><given-names>SH</given-names></name></person-group><article-title>Three-category classification of magnetic resonance hearing loss images based on deep autoencoder</article-title><source>J Med Syst</source><volume>41</volume><fpage>165</fpage><year>2017</year><pub-id pub-id-type="doi">10.1007/s10916-017-0814-4</pub-id><pub-id pub-id-type="pmid">28895033</pub-id></element-citation></ref>
<ref id="b26-ol-0-0-9761"><label>26</label><element-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Deng</surname><given-names>L</given-names></name><name><surname>Fan</surname><given-names>C</given-names></name><name><surname>Zeng</surname><given-names>Z</given-names></name></person-group><article-title>A sparse autoencoder-based deep neural network for protein solvent accessibility and contact number prediction</article-title><source>BMC Bioinformatics</source><volume>18</volume><fpage>569</fpage><year>2017</year><pub-id pub-id-type="doi">10.1186/s12859-017-1971-7</pub-id><pub-id pub-id-type="pmid">29297299</pub-id></element-citation></ref>
<ref id="b27-ol-0-0-9761"><label>27</label><element-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Wang</surname><given-names>YB</given-names></name><name><surname>You</surname><given-names>ZH</given-names></name><name><surname>Li</surname><given-names>X</given-names></name><name><surname>Jiang</surname><given-names>TH</given-names></name><name><surname>Chen</surname><given-names>X</given-names></name><name><surname>Zhou</surname><given-names>X</given-names></name><name><surname>Wang</surname><given-names>L</given-names></name></person-group><article-title>Predicting protein-protein interactions from protein sequences by a stacked sparse autoencoder deep neural network</article-title><source>Mol BioSyst</source><volume>13</volume><fpage>1336</fpage><lpage>1344</lpage><year>2017</year><pub-id pub-id-type="doi">10.1039/C7MB00188F</pub-id><pub-id pub-id-type="pmid">28604872</pub-id></element-citation></ref>
<ref id="b28-ol-0-0-9761"><label>28</label><element-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Zhao</surname><given-names>W</given-names></name><name><surname>Meng</surname><given-names>QH</given-names></name><name><surname>Zeng</surname><given-names>M</given-names></name><name><surname>Qi</surname><given-names>PF</given-names></name></person-group><article-title>Stacked sparse auto-encoders (SSAE) based electronic nose for chinese liquors classification</article-title><source>Sensors (Basel)</source><volume>17</volume><fpage>E2855</fpage><comment>2017.29</comment><pub-id pub-id-type="doi">10.3390/s17122855</pub-id><pub-id pub-id-type="pmid">29292772</pub-id></element-citation></ref>
<ref id="b29-ol-0-0-9761"><label>29</label><element-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Podolsky</surname><given-names>MD</given-names></name><name><surname>Barchuk</surname><given-names>AA</given-names></name><name><surname>Kuznetcov</surname><given-names>VI</given-names></name><name><surname>Gusarova</surname><given-names>NF</given-names></name><name><surname>Gaidukov</surname><given-names>VS</given-names></name><name><surname>Tarakanov</surname><given-names>SA</given-names></name></person-group><article-title>Evaluation of machine learning algorithm utilization for lung cancer classification based on gene expression levels</article-title><source>Asian Pacific J Cancer Prev</source><volume>17</volume><fpage>835</fpage><lpage>838</lpage><year>2016</year><pub-id pub-id-type="doi">10.7314/APJCP.2016.17.2.835</pub-id></element-citation></ref>
<ref id="b30-ol-0-0-9761"><label>30</label><element-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Wang</surname><given-names>L</given-names></name><name><surname>Wang</surname><given-names>Y</given-names></name><name><surname>Chang</surname><given-names>Q</given-names></name></person-group><article-title>Feature selection methods for big data bioinformatics: A survey from the search perspective</article-title><source>Methods</source><volume>111</volume><fpage>21</fpage><lpage>31</lpage><year>2016</year><pub-id pub-id-type="doi">10.1016/j.ymeth.2016.08.014</pub-id><pub-id pub-id-type="pmid">27592382</pub-id></element-citation></ref>
<ref id="b31-ol-0-0-9761"><label>31</label><element-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Chen</surname><given-names>L</given-names></name><name><surname>Hao</surname><given-names>Y</given-names></name></person-group><article-title>Feature extraction and classification of EHG between pregnancy and labour group using Hilbert-Huang transform and extreme learning machine</article-title><source>Comput Math Methods Med</source><volume>2017</volume><fpage>7949507</fpage><year>2017</year><pub-id pub-id-type="doi">10.1155/2017/7949507</pub-id><pub-id pub-id-type="pmid">28316639</pub-id></element-citation></ref>
<ref id="b32-ol-0-0-9761"><label>32</label><element-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Urbanowicz</surname><given-names>RJ</given-names></name><name><surname>Meeker</surname><given-names>M</given-names></name><name><surname>La Cava</surname><given-names>W</given-names></name><name><surname>Olson</surname><given-names>RS</given-names></name><name><surname>Moore</surname><given-names>JH</given-names></name></person-group><article-title>Relief-based feature selection: Introduction and review</article-title><source>J Biomed Inform</source><volume>85</volume><fpage>189</fpage><lpage>203</lpage><year>2018</year><pub-id pub-id-type="doi">10.1016/j.jbi.2018.07.014</pub-id><pub-id pub-id-type="pmid">30031057</pub-id></element-citation></ref>
<ref id="b33-ol-0-0-9761"><label>33</label><element-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Urbanowicz</surname><given-names>RJ</given-names></name><name><surname>Olson</surname><given-names>RS</given-names></name><name><surname>Schmitt</surname><given-names>P</given-names></name><name><surname>Meeker</surname><given-names>M</given-names></name><name><surname>Moore</surname><given-names>JH</given-names></name></person-group><article-title>Benchmarking relief-based feature selection methods for bioinformatics data mining</article-title><source>J Biomed Inform</source><volume>85</volume><fpage>168</fpage><lpage>188</lpage><year>2018</year><pub-id pub-id-type="doi">10.1016/j.jbi.2018.07.015</pub-id><pub-id pub-id-type="pmid">30030120</pub-id></element-citation></ref>
<ref id="b34-ol-0-0-9761"><label>34</label><element-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Dai</surname><given-names>Q</given-names></name><name><surname>Cheng</surname><given-names>JH</given-names></name><name><surname>Sun</surname><given-names>DW</given-names></name><name><surname>Zeng</surname><given-names>XA</given-names></name></person-group><article-title>Advances in feature selection methods for hyperspectral image processing in food industry applications: A review</article-title><source>Crit Rev Food Sci Nutr</source><volume>55</volume><fpage>1368</fpage><lpage>1382</lpage><year>2015</year><pub-id pub-id-type="doi">10.1080/10408398.2013.871692</pub-id><pub-id pub-id-type="pmid">24689555</pub-id></element-citation></ref>
<ref id="b35-ol-0-0-9761"><label>35</label><element-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Cao</surname><given-names>J</given-names></name><name><surname>Cui</surname><given-names>H</given-names></name><name><surname>Shi</surname><given-names>H</given-names></name><name><surname>Jiao</surname><given-names>L</given-names></name></person-group><article-title>Big data: A parallel particle swarm optimization-back-propagation neural network algorithm based on mapreduce</article-title><source>PLoS One</source><volume>11</volume><fpage>e0157551</fpage><year>2016</year><pub-id pub-id-type="doi">10.1371/journal.pone.0157551</pub-id><pub-id pub-id-type="pmid">27304987</pub-id></element-citation></ref>
<ref id="b36-ol-0-0-9761"><label>36</label><element-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Song</surname><given-names>Y</given-names></name><name><surname>Zhang</surname><given-names>J</given-names></name></person-group><article-title>Automatic recognition of epileptic EEG patterns via extreme learning machine and multiresolution feature extraction</article-title><source>Expert Syst Appl</source><volume>40</volume><fpage>5477</fpage><lpage>5489</lpage><year>2013</year><pub-id pub-id-type="doi">10.1016/j.eswa.2013.04.025</pub-id></element-citation></ref>
<ref id="b37-ol-0-0-9761"><label>37</label><element-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Song</surname><given-names>Y</given-names></name><name><surname>Crowcroft</surname><given-names>J</given-names></name><name><surname>Zhang</surname><given-names>J</given-names></name></person-group><article-title>Automatic epileptic seizure detection in EEGs based on optimized sample entropy and extreme learning machine</article-title><source>J Neurosci Methods</source><volume>210</volume><fpage>132</fpage><lpage>146</lpage><year>2012</year><pub-id pub-id-type="doi">10.1016/j.jneumeth.2012.07.003</pub-id><pub-id pub-id-type="pmid">22824535</pub-id></element-citation></ref>
<ref id="b38-ol-0-0-9761"><label>38</label><element-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Xu</surname><given-names>J</given-names></name><name><surname>Xiang</surname><given-names>L</given-names></name><name><surname>Liu</surname><given-names>Q</given-names></name><name><surname>Gilmore</surname><given-names>H</given-names></name><name><surname>Wu</surname><given-names>J</given-names></name><name><surname>Tang</surname><given-names>J</given-names></name><name><surname>Madabhushi</surname><given-names>A</given-names></name></person-group><article-title>Stacked sparse autoencoder (SSAE) for nuclei detection on breast cancer histopathology images</article-title><source>IEEE Trans Med Imaging</source><volume>35</volume><fpage>119</fpage><lpage>130</lpage><year>2016</year><pub-id pub-id-type="doi">10.1109/TMI.2015.2458702</pub-id><pub-id pub-id-type="pmid">26208307</pub-id></element-citation></ref>
<ref id="b39-ol-0-0-9761"><label>39</label><element-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Langkvist</surname><given-names>M</given-names></name><name><surname>Loutfi</surname><given-names>A</given-names></name></person-group><article-title>Learning feature representations with a cost-relevant sparse autoencoder</article-title><source>Int J Neural Syst</source><volume>25</volume><fpage>1450034</fpage><year>2015</year><pub-id pub-id-type="doi">10.1142/S0129065714500348</pub-id><pub-id pub-id-type="pmid">25515941</pub-id></element-citation></ref>
<ref id="b40-ol-0-0-9761"><label>40</label><element-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Tsinalis</surname><given-names>O</given-names></name><name><surname>Matthews</surname><given-names>PM</given-names></name><name><surname>Guo</surname><given-names>Y</given-names></name></person-group><article-title>Automatic sleep stage scoring using time-frequency analysis and stacked sparse autoencoders</article-title><source>Ann Biomed Eng</source><volume>44</volume><fpage>1587</fpage><lpage>1597</lpage><year>2016</year><pub-id pub-id-type="doi">10.1007/s10439-015-1444-y</pub-id><pub-id pub-id-type="pmid">26464268</pub-id></element-citation></ref>
<ref id="b41-ol-0-0-9761"><label>41</label><element-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Chen</surname><given-names>LL</given-names></name><name><surname>Hao</surname><given-names>YR</given-names></name></person-group><article-title>Discriminating pregnancy and labour in electrohysterogram by sample entropy and support vector machine</article-title><source>J Med Imaging Health Inform</source><volume>7</volume><fpage>584</fpage><lpage>591</lpage><year>2017</year><pub-id pub-id-type="doi">10.1166/jmihi.2017.2065</pub-id></element-citation></ref>
<ref id="b42-ol-0-0-9761"><label>42</label><element-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Janowczyk</surname><given-names>A</given-names></name><name><surname>Basavanhally</surname><given-names>A</given-names></name><name><surname>Madabhushi</surname><given-names>A</given-names></name></person-group><article-title>Stain normalization using sparse auto encoders (StaNoSA): Application to digital pathology</article-title><source>Comput Med Imaging Graph</source><volume>57</volume><fpage>50</fpage><lpage>61</lpage><year>2017</year><pub-id pub-id-type="doi">10.1016/j.compmedimag.2016.05.003</pub-id><pub-id pub-id-type="pmid">27373749</pub-id></element-citation></ref>
<ref id="b43-ol-0-0-9761"><label>43</label><element-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Su</surname><given-names>H</given-names></name><name><surname>Xing</surname><given-names>F</given-names></name><name><surname>Kong</surname><given-names>X</given-names></name><name><surname>Xie</surname><given-names>Y</given-names></name><name><surname>Zhang</surname><given-names>S</given-names></name><name><surname>Yang</surname><given-names>L</given-names></name></person-group><article-title>Robust cell detection and segmentation in histopathological images using sparse reconstruction and stacked denoising auto encoders</article-title><source>Med Image Comput Comput Assist Interv</source><volume>9351</volume><fpage>383</fpage><lpage>390</lpage><year>2015</year><pub-id pub-id-type="pmid">27796013</pub-id></element-citation></ref>
<ref id="b44-ol-0-0-9761"><label>44</label><element-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Paydar</surname><given-names>K</given-names></name><name><surname>Niakan Kalhori</surname><given-names>SR</given-names></name><name><surname>Akbarian</surname><given-names>M</given-names></name><name><surname>Sheikhtaheri</surname><given-names>A</given-names></name></person-group><article-title>A clinical decision support system for prediction of pregnancy outcome in pregnant women with systemic lupus erythematosus</article-title><source>Int J Med Inform</source><volume>97</volume><fpage>239</fpage><lpage>246</lpage><year>2017</year><pub-id pub-id-type="doi">10.1016/j.ijmedinf.2016.10.018</pub-id><pub-id pub-id-type="pmid">27919382</pub-id></element-citation></ref>
<ref id="b45-ol-0-0-9761"><label>45</label><element-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Somu</surname><given-names>N</given-names></name><name><surname>M R</surname><given-names>RG</given-names></name><name><surname>V</surname><given-names>K</given-names></name><name><surname>Kirthivasan</surname><given-names>K</given-names></name><name><surname>V S</surname><given-names>SS</given-names></name></person-group><article-title>An improved robust heteroscedastic probabilistic neural network based trust prediction approach for cloud service selection</article-title><source>Neural Netw</source><volume>108</volume><fpage>339</fpage><lpage>354</lpage><year>2018</year><pub-id pub-id-type="doi">10.1016/j.neunet.2018.08.005</pub-id><pub-id pub-id-type="pmid">30245433</pub-id></element-citation></ref>
</ref-list>
</back>
<floats-group>
<fig id="f1-ol-0-0-9761" position="float">
<label>Figure 1.</label>
<caption><p>Architecture of a basic autoencoder. V (1,2&#x2026;n), entire training data; X (1,2&#x2026;n), set of training data; b<sub>h</sub>, bias of hidden layer; b<sub>x</sub>, bias of input layer; h (1,2&#x2026;m), feature vector at hidden layer; x&#x02C6;, reconstruction of input x.</p></caption>
<graphic xlink:href="ol-17-02-1483-g04.tif"/>
</fig>
<fig id="f2-ol-0-0-9761" position="float">
<label>Figure 2.</label>
<caption><p>Architecture of sparse autoencoder. V (1,2&#x2026;n), entire training data; X (1,2&#x2026;n), set of training data; b<sub>h</sub>, bias of hidden layer; and b<sub>x</sub>, bias of input layer; h<sup>(l)</sup> (1,2&#x2026;m), feature vector at layer l; h<sup>(<xref rid="b2-ol-0-0-9761" ref-type="bibr">2</xref>)</sup> (1,2&#x2026;m), feature vector at layer 2.</p></caption>
<graphic xlink:href="ol-17-02-1483-g05.tif"/>
</fig>
<fig id="f3-ol-0-0-9761" position="float">
<label>Figure 3.</label>
<caption><p>Comparison of the accuracy achieved by various diagnostic models. BP, backpropagation; GA, genetic algorithm; ELM, extreme learning machine; SSAE, stacked sparse autoencoder.</p></caption>
<graphic xlink:href="ol-17-02-1483-g06.tif"/>
</fig>
<fig id="f4-ol-0-0-9761" position="float">
<label>Figure 4.</label>
<caption><p>Comparison of the specificity achieved by various diagnostic models. BP, backpropagation; GA, genetic algorithm; ELM, extreme learning machine; SSAE, stacked sparse autoencoder.</p></caption>
<graphic xlink:href="ol-17-02-1483-g07.tif"/>
</fig>
<fig id="f5-ol-0-0-9761" position="float">
<label>Figure 5.</label>
<caption><p>Comparison of the F-measure for various diagnostic models. BP, backpropagation; GA, genetic algorithm; ELM, extreme learning machine; SSAE, stacked sparse autoencoder.</p></caption>
<graphic xlink:href="ol-17-02-1483-g08.tif"/>
</fig>
<fig id="f6-ol-0-0-9761" position="float">
<label>Figure 6.</label>
<caption><p>ROC curves to assess the capacity of the diagnostic models to distinguish between patients with malignant mesothelioma and healthy participants. ROC, receiver operating characteristic; BP, backpropagation; GA, genetic algorithm; ELM, extreme learning machine; SSAE, stacked sparse autoencoder.</p></caption>
<graphic xlink:href="ol-17-02-1483-g09.tif"/>
</fig>
<fig id="f7-ol-0-0-9761" position="float">
<label>Figure 7.</label>
<caption><p>Comparison of the area under the receiver operating characteristic curves for various diagnostic models. AUC, area under the curve; BP, backpropagation; GA, genetic algorithm; ELM, extreme learning machine; SSAE, stacked sparse autoencoder.</p></caption>
<graphic xlink:href="ol-17-02-1483-g10.tif"/>
</fig>
<table-wrap id="tI-ol-0-0-9761" position="float">
<label>Table I.</label>
<caption><p>Variables for diagnosing malignant mesothelioma.</p></caption>
<table frame="hsides" rules="groups">
<thead>
<tr>
<th align="left" valign="bottom">Variable</th>
<th align="center" valign="bottom">Value type</th>
</tr>
</thead>
<tbody>
<tr>
<td align="left" valign="top">Age</td>
<td align="left" valign="top">Quantitative</td>
</tr>
<tr>
<td align="left" valign="top">Sex</td>
<td align="left" valign="top">Qualitative</td>
</tr>
<tr>
<td align="left" valign="top">City</td>
<td align="left" valign="top">Qualitative</td>
</tr>
<tr>
<td align="left" valign="top">Asbestos exposure</td>
<td align="left" valign="top">Qualitative</td>
</tr>
<tr>
<td align="left" valign="top">Type of malignant mesothelioma</td>
<td align="left" valign="top">Qualitative</td>
</tr>
<tr>
<td align="left" valign="top">Duration of asbestos exposure</td>
<td align="left" valign="top">Quantitative</td>
</tr>
<tr>
<td align="left" valign="top">Diagnosis method</td>
<td align="left" valign="top">Qualitative</td>
</tr>
<tr>
<td align="left" valign="top">Keep side</td>
<td align="left" valign="top">Qualitative</td>
</tr>
<tr>
<td align="left" valign="top">Cytology</td>
<td align="left" valign="top">Qualitative</td>
</tr>
<tr>
<td align="left" valign="top">Duration of symptoms</td>
<td align="left" valign="top">Quantitative</td>
</tr>
<tr>
<td align="left" valign="top">Dyspnoea</td>
<td align="left" valign="top">Qualitative</td>
</tr>
<tr>
<td align="left" valign="top">Ache on chest</td>
<td align="left" valign="top">Qualitative</td>
</tr>
<tr>
<td align="left" valign="top">Weakness</td>
<td align="left" valign="top">Qualitative</td>
</tr>
<tr>
<td align="left" valign="top">Smoker</td>
<td align="left" valign="top">Qualitative</td>
</tr>
<tr>
<td align="left" valign="top">Performance status</td>
<td align="left" valign="top">Qualitative</td>
</tr>
<tr>
<td align="left" valign="top">White blood cell count</td>
<td align="left" valign="top">Quantitative</td>
</tr>
<tr>
<td align="left" valign="top">Hemoglobin</td>
<td align="left" valign="top">Quantitative</td>
</tr>
<tr>
<td align="left" valign="top">Platelet count</td>
<td align="left" valign="top">Quantitative</td>
</tr>
<tr>
<td align="left" valign="top">Sedimentation rate</td>
<td align="left" valign="top">Quantitative</td>
</tr>
<tr>
<td align="left" valign="top">Blood lactose dehydrogenase</td>
<td align="left" valign="top">Quantitative</td>
</tr>
<tr>
<td align="left" valign="top">Alkaline phosphatase</td>
<td align="left" valign="top">Quantitative</td>
</tr>
<tr>
<td align="left" valign="top">Total protein</td>
<td align="left" valign="top">Quantitative</td>
</tr>
<tr>
<td align="left" valign="top">Albumin</td>
<td align="left" valign="top">Quantitative</td>
</tr>
<tr>
<td align="left" valign="top">Glucose</td>
<td align="left" valign="top">Quantitative</td>
</tr>
<tr>
<td align="left" valign="top">Pleural lactose dehydrogenase</td>
<td align="left" valign="top">Quantitative</td>
</tr>
<tr>
<td align="left" valign="top">Pleural protein</td>
<td align="left" valign="top">Quantitative</td>
</tr>
<tr>
<td align="left" valign="top">Pleural albumin</td>
<td align="left" valign="top">Quantitative</td>
</tr>
<tr>
<td align="left" valign="top">Pleural glucose</td>
<td align="left" valign="top">Quantitative</td>
</tr>
<tr>
<td align="left" valign="top">Mortality</td>
<td align="left" valign="top">Qualitative</td>
</tr>
<tr>
<td align="left" valign="top">Pleural effusion</td>
<td align="left" valign="top">Qualitative</td>
</tr>
<tr>
<td align="left" valign="top">Pleural thickness on tomography</td>
<td align="left" valign="top">Qualitative</td>
</tr>
<tr>
<td align="left" valign="top">Pleural pH</td>
<td align="left" valign="top">Qualitative</td>
</tr>
<tr>
<td align="left" valign="top">Cell count</td>
<td align="left" valign="top">Quantitative</td>
</tr>
<tr>
<td align="left" valign="top">C-reactive protein</td>
<td align="left" valign="top">Quantitative</td>
</tr>
</tbody>
</table>
</table-wrap>
</floats-group>
</article>
