↓ Skip to main content

Deep-learning: investigating deep neural networks hyper-parameters and comparison of performance to shallow methods for modeling bioactivity data

Overview of attention for article published in Journal of Cheminformatics, June 2017
Altmetric Badge

About this Attention Score

  • In the top 25% of all research outputs scored by Altmetric
  • High Attention Score compared to outputs of the same age (82nd percentile)
  • Average Attention Score compared to outputs of the same age and source

Mentioned by

twitter
16 X users

Citations

dimensions_citation
238 Dimensions

Readers on

mendeley
354 Mendeley
You are seeing a free-to-access but limited selection of the activity Altmetric has collected about this research output. Click here to find out more.
Title
Deep-learning: investigating deep neural networks hyper-parameters and comparison of performance to shallow methods for modeling bioactivity data
Published in
Journal of Cheminformatics, June 2017
DOI 10.1186/s13321-017-0226-y
Pubmed ID
Authors

Alexios Koutsoukas, Keith J. Monaghan, Xiaoli Li, Jun Huan

Abstract

In recent years, research in artificial neural networks has resurged, now under the deep-learning umbrella, and grown extremely popular. Recently reported success of DL techniques in crowd-sourced QSAR and predictive toxicology competitions has showcased these methods as powerful tools in drug-discovery and toxicology research. The aim of this work was dual, first large number of hyper-parameter configurations were explored to investigate how they affect the performance of DNNs and could act as starting points when tuning DNNs and second their performance was compared to popular methods widely employed in the field of cheminformatics namely Naïve Bayes, k-nearest neighbor, random forest and support vector machines. Moreover, robustness of machine learning methods to different levels of artificially introduced noise was assessed. The open-source Caffe deep-learning framework and modern NVidia GPU units were utilized to carry out this study, allowing large number of DNN configurations to be explored. We show that feed-forward deep neural networks are capable of achieving strong classification performance and outperform shallow methods across diverse activity classes when optimized. Hyper-parameters that were found to play critical role are the activation function, dropout regularization, number hidden layers and number of neurons. When compared to the rest methods, tuned DNNs were found to statistically outperform, with p value <0.01 based on Wilcoxon statistical test. DNN achieved on average MCC units of 0.149 higher than NB, 0.092 than kNN, 0.052 than SVM with linear kernel, 0.021 than RF and finally 0.009 higher than SVM with radial basis function kernel. When exploring robustness to noise, non-linear methods were found to perform well when dealing with low levels of noise, lower than or equal to 20%, however when dealing with higher levels of noise, higher than 30%, the Naïve Bayes method was found to perform well and even outperform at the highest level of noise 50% more sophisticated methods across several datasets.

X Demographics

X Demographics

The data shown below were collected from the profiles of 16 X users who shared this research output. Click here to find out more about how the information was compiled.
Mendeley readers

Mendeley readers

The data shown below were compiled from readership statistics for 354 Mendeley readers of this research output. Click here to see the associated Mendeley record.

Geographical breakdown

Country Count As %
Germany 1 <1%
Unknown 353 100%

Demographic breakdown

Readers by professional status Count As %
Student > Ph. D. Student 70 20%
Student > Master 43 12%
Researcher 40 11%
Student > Bachelor 28 8%
Professor > Associate Professor 12 3%
Other 45 13%
Unknown 116 33%
Readers by discipline Count As %
Computer Science 59 17%
Engineering 41 12%
Chemistry 34 10%
Biochemistry, Genetics and Molecular Biology 22 6%
Pharmacology, Toxicology and Pharmaceutical Science 14 4%
Other 59 17%
Unknown 125 35%
Attention Score in Context

Attention Score in Context

This research output has an Altmetric Attention Score of 11. This is our high-level measure of the quality and quantity of online attention that it has received. This Attention Score, as well as the ranking and number of research outputs shown below, was calculated when the research output was last mentioned on 11 November 2021.
All research outputs
#3,329,925
of 25,834,578 outputs
Outputs from Journal of Cheminformatics
#302
of 982 outputs
Outputs of similar age
#57,424
of 329,533 outputs
Outputs of similar age from Journal of Cheminformatics
#8
of 16 outputs
Altmetric has tracked 25,834,578 research outputs across all sources so far. Compared to these this one has done well and is in the 87th percentile: it's in the top 25% of all research outputs ever tracked by Altmetric.
So far Altmetric has tracked 982 research outputs from this source. They typically receive more attention than average, with a mean Attention Score of 10.0. This one has gotten more attention than average, scoring higher than 69% of its peers.
Older research outputs will score higher simply because they've had more time to accumulate mentions. To account for age we can compare this Altmetric Attention Score to the 329,533 tracked outputs that were published within six weeks on either side of this one in any source. This one has done well, scoring higher than 82% of its contemporaries.
We're also able to compare this research output to 16 others from the same source and published within six weeks on either side of this one. This one has gotten more attention than average, scoring higher than 50% of its contemporaries.