Machine Learning for Healthcare Applications. Группа авторов
Чтение книги онлайн.

Читать онлайн книгу Machine Learning for Healthcare Applications - Группа авторов страница 26

Название: Machine Learning for Healthcare Applications

Автор: Группа авторов

Издательство: John Wiley & Sons Limited

Жанр: Программы

Серия:

isbn: 9781119792598

isbn:

СКАЧАТЬ of 26 volunteers and then applied Approximate Entropy on it for inter-subject evaluation of data as the part of a retrospective approach [19] while adding truthfulness to Entropy windows for its stable distribution. This filter is very extensively being used in Signal processing which led us to adopt it.

      The study [20] is an experiment on ECG signals of 26 participants where approximate entropy method is implemented for examining the concentration. Approximation entropy window was taken less for intra-patient comparing to inter-patient and for filtering the noisy signals S-Golay method was implemented.

      They have innovatively preprocessed the ECG signal using S–Golay filter technique [21]. With both quadratic degrees of smoothing and differentiation filter methods combinedly has processed ECG signals having sampling rate 500 Hz with seventeen points length.

      A very unique “double-class motor imaginary Brain Computer Interface” was implemented with Recurrent Quantum Neural Network model for filtering EEG signals [22].

      In the paper [23] using the S-Golay filter, the artifacts due to blinking of eyes are found out and it is eliminated adapting a noise removal method.

      3.3.1 Bagging Decision Tree Classifier

      Among the many Machine Learning algorithms, this method forms a group of algorithms where several instances are created of black-box estimators on variable subsets from the base training set after which we aggregate their solo predictions to form a resultant prediction. This process is used as a path to minimize the variance of the foundation estimator i.e. a decision tree by including randomization within its creation process and building an ensemble from it. In multiple scenarios, this method consists a simple path to improve with regard to a single model, avoids making it a necessity to acclimatize to a foundation algorithm. It works best with fully developed decision trees as it reduces overfitting in comparison to boosting methods which generally work best in shallow decision trees. This classifier comes in many flavors but majorly differ from each other by the path that they draw variable subsets of the training set. In our case samples were extracted with replacement called as Bagging.

      3.3.2 Gaussian Naïve Bayes Classifier

      3.3.3 Kernel Support Vector Machine (Sigmoid)

      For training purposes, we use the SVC class of the library. The difference is in the values for the Kernel parameters of SVC class. In simple SVM’s we use “Linear” for Kernel parameters but in K-SVM we use Gaussian, Sigmoid, Polynomial, etc. wherein we have used Sigmoid.

      The only limitation observed in our case is that though this method achieves the highest accuracy but not up to the mark. Hence more advanced models like Deep Learning may be applied in near future for more concrete results.

      3.3.4 Random Decision Forest Classifier

      It is a variant of supervised machine learning algorithm founded on the schematic of ensembled learning. Ensemble learning is an algorithm where you join multiple or single algorithm into multiple types of algorithms of multiple or same variant to create a complex and advanced prediction model. It also combines many algorithms of same variant as decision trees, forest trees, etc. so the name “Random Forest”. It is used for regression and classification tasks.

      The way it works is it picks a part of the dataset and builds a decision tree on these records, and after selection of number of trees you want this process is repeated. Each tree represents the prediction in that category for which the new record belongs. The only limitation here is that there forte lies in their complexity and for that we need substantial computing resources when huge number of decision trees can be brought together which in turn will better train themselves.

Schematic illustration of brain map structure and Equipment used. Schematic illustration of workflow diagram.