• Wyszukiwanie zaawansowane
  • Kategorie
  • Kategorie BISAC
  • Książki na zamówienie
  • Promocje
  • Granty
  • Książka na prezent
  • Opinie
  • Pomoc
  • Załóż konto
  • Zaloguj się

Modern Data Mining Algorithms in C++ and Cuda C: Recent Developments in Feature Extraction and Selection Algorithms for Data Science » książka

zaloguj się | załóż konto
Logo Krainaksiazek.pl

koszyk

konto

szukaj
topmenu
Księgarnia internetowa
Szukaj
Książki na zamówienie
Promocje
Granty
Książka na prezent
Moje konto
Pomoc
 
 
Wyszukiwanie zaawansowane
Pusty koszyk
Bezpłatna dostawa dla zamówień powyżej 20 złBezpłatna dostawa dla zamówień powyżej 20 zł

Kategorie główne

• Nauka
 [2949965]
• Literatura piękna
 [1857847]

  więcej...
• Turystyka
 [70818]
• Informatyka
 [151303]
• Komiksy
 [35733]
• Encyklopedie
 [23180]
• Dziecięca
 [617748]
• Hobby
 [139972]
• AudioBooki
 [1650]
• Literatura faktu
 [228361]
• Muzyka CD
 [398]
• Słowniki
 [2862]
• Inne
 [444732]
• Kalendarze
 [1620]
• Podręczniki
 [167233]
• Poradniki
 [482388]
• Religia
 [509867]
• Czasopisma
 [533]
• Sport
 [61361]
• Sztuka
 [243125]
• CD, DVD, Video
 [3451]
• Technologie
 [219309]
• Zdrowie
 [101347]
• Książkowe Klimaty
 [123]
• Zabawki
 [2362]
• Puzzle, gry
 [3791]
• Literatura w języku ukraińskim
 [253]
• Art. papiernicze i szkolne
 [7933]
Kategorie szczegółowe BISAC

Modern Data Mining Algorithms in C++ and Cuda C: Recent Developments in Feature Extraction and Selection Algorithms for Data Science

ISBN-13: 9781484259870 / Angielski / Miękka / 2020 / 228 str.

Timothy Masters
Modern Data Mining Algorithms in C++ and Cuda C: Recent Developments in Feature Extraction and Selection Algorithms for Data Science Masters, Timothy 9781484259870 Apress - książkaWidoczna okładka, to zdjęcie poglądowe, a rzeczywista szata graficzna może różnić się od prezentowanej.

Modern Data Mining Algorithms in C++ and Cuda C: Recent Developments in Feature Extraction and Selection Algorithms for Data Science

ISBN-13: 9781484259870 / Angielski / Miękka / 2020 / 228 str.

Timothy Masters
cena 262,25 zł
(netto: 249,76 VAT:  5%)

Najniższa cena z 30 dni: 250,57 zł
Termin realizacji zamówienia:
ok. 22 dni roboczych
Bez gwarancji dostawy przed świętami

Darmowa dostawa!
Kategorie:
Informatyka, Bazy danych
Kategorie BISAC:
Computers > Data Science - Data Analytics
Mathematics > Prawdopodobieństwo i statystyka
Computers > Languages - General
Wydawca:
Apress
Język:
Angielski
ISBN-13:
9781484259870
Rok wydania:
2020
Ilość stron:
228
Waga:
0.42 kg
Wymiary:
25.4 x 17.78 x 1.3
Oprawa:
Miękka
Wolumenów:
01
Dodatkowe informacje:
Wydanie ilustrowane

"This is an excellent book directed toward those who are already working in data mining." (Anthony J. Duben, Computing Reviews, May 5, 2021)

1) Introduction 7

2) Forward Selection Component Analysis 11
A) Introduction to Forward Selection Component Analysis 12
B) The Mathematics and Code Examples 16
Maximizing the Explained Variance 18
Code for the Variance Maximization Criterion 20
Backward Refinement 24
Multi-Threading Backward Refinement 28
Orthogonalizing Ordered Components 36
C) Putting It All Together 39
Components From a Forward-Only Subset 44
Components From a Backward Refined Subset 46
D) An Example With Contrived Variables 48

3) Local Feature Selection 53
A) Intuitive Overview of the Algorithm 54
What This Algorithm Reports 60
B) A Brief Detour: the Simplex Algorithm 62
The Linear Programming Problem 63
Interfacing to the Simplex Class 64
A Little More Detail 67
C) A More Rigorous Approach to LFS 69
Intra-Class and Inter-Class Separation 73
Computing the Weights 77
Maximizing Inter-Class Separation 81
Minimizing Intra-Class Separation 86
Testing a Trial Beta 88
A Quick Note on Threads 93
D) CUDA Computation of Weights 94
Integrating the CUDA Code Into the Algorithm 95
Initializing the CUDA Hardware 97
Computing Differences from the Current Case 100
Computing the Distance Matrix 102
Computing the Minimum Distances 104
Computing the Terms for the Weight Equation 112
Transposing the Term Matrix 113
Summing the Terms For the Weights 114
Moving the Weights to the Host 116
E) An Example of Local Feature Selection 117
F) A Note on Run Time 118

4) Memory in Time Series Features 119
A) A Gentle Mathematical Overview 122
The Forward Algorithm 123
The Backward Algorithm 128
Correct Alpha and Beta, For Those Who Care 131
B) Some Mundane Computations 136
Means and Covariances 136
Densities 138
The Multivariate Normal Density Function 139
C) Starting Parameters 141
Outline of the Initialization Algorithm 141
Perturbing Means 142
Perturbing Covariances 143
Perturbing Transition Probabilities 144
A Note on Random Number Generators 145
D) The Complete Optimization Algorithm 146
Computing State Probabilities 147
Updating the Means and Covariances 151
Updating Initial and Transition Probabilities 153
E) Assessing HMM Memory in a Time Series 159
F) Linking Features to a Target 164
Linking HMM States to the Target 173
A Contrived and Inappropriate Example 183
A Sensible and Practical Example 186

5) Stepwise Selection on Steroids 189
A) The Feature Evaluation Model 192
Code For the Foundation Model 193
B) The Cross-Validated Performance Measure 198
C) The Stepwise Algorithm 201
Finding the First Variable 207
Adding a Variable to an Existing Model 210
D) Demonstrating the Algorithm Three Ways 214

6) Nominal-to-Ordinal Conversion 217
A) Implementation Overview 221
B) Testing For a Legitimate Relationship 222
C) An Example From Equity Price Changes 223
D) Code for Nominal-to-Ordinal Conversion 227
The Constructor 228
Printing the Table of Counts 232
Computing the Mapping Function 234
Monte-Carlo Permutation Tests 237

7) Index 353

Timothy Masters has a PhD in statistics and is an experienced programmer. His dissertation was in image analysis. His career moved in the direction of signal processing, and for the last 25 years he's been involved in the development of automated trading systems in various financial markets.  


As a serious data miner you will often be faced with thousands of candidate features for your prediction or classification application, with most of the features being of little or no value. You’ll know that many of these features may be useful only in combination with certain other features while being practically worthless alone or in combination with most others. Some features may have enormous predictive power, but only within a small, specialized area of the feature space. The problems that plague modern data miners are endless. This book helps you solve this problem by presenting modern feature selection techniques and the code to implement them. Some of these techniques are:

    Forward selection component analysis
  • Local feature selection
  • Linking features and a target with a hidden Markov model
  • Improvements on traditional stepwise selection
  • Nominal-to-ordinal conversion
All algorithms are intuitively justified and supported by the relevant equations and explanatory material. The author also presents and explains complete, highly commented source code. 

The example code is in C++ and CUDA C but Python or other code can be substituted; the algorithm is important, not the code that's used to write it.  

You will:

  • Combine principal component analysis with forward and backward stepwise selection to identify a compact subset of a large collection of variables that captures the maximum possible variation within the entire set.
  • Identify features that may have predictive power over only a small subset of the feature domain. Such features can be profitably used by modern predictive models but may be missed by other feature selection methods.
  • Find an underlying hidden Markov model that controls the distributions of feature variables and the target simultaneously. The memory inherent in this method is especially valuable in high-noise applications such as prediction of financial markets.
  • Improve traditional stepwise selection in three ways: examine a collection of 'best-so-far' feature sets; test candidate features for inclusion with cross validation to automatically and effectively limit model complexity; and at each step estimate the probability that our results so far could be just the product of random good luck. We also estimate the probability that the improvement obtained by adding a new variable could have been just good luck.
  • Take a potentially valuable nominal variable (a category or class membership) that is unsuitable for input to a prediction model, and assign to each category a sensible numeric value that can be used as a model input.
  • Masters, Timothy David Aronson is a pioneer in machine learning and... więcej >


    Udostępnij

    Facebook - konto krainaksiazek.pl



    Opinie o Krainaksiazek.pl na Opineo.pl

    Partner Mybenefit

    Krainaksiazek.pl w programie rzetelna firma Krainaksiaze.pl - płatności przez paypal

    Czytaj nas na:

    Facebook - krainaksiazek.pl
    • książki na zamówienie
    • granty
    • książka na prezent
    • kontakt
    • pomoc
    • opinie
    • regulamin
    • polityka prywatności

    Zobacz:

    • Księgarnia czeska

    • Wydawnictwo Książkowe Klimaty

    1997-2025 DolnySlask.com Agencja Internetowa

    © 1997-2022 krainaksiazek.pl
         
    KONTAKT | REGULAMIN | POLITYKA PRYWATNOŚCI | USTAWIENIA PRYWATNOŚCI
    Zobacz: Księgarnia Czeska | Wydawnictwo Książkowe Klimaty | Mapa strony | Lista autorów
    KrainaKsiazek.PL - Księgarnia Internetowa
    Polityka prywatnosci - link
    Krainaksiazek.pl - płatnośc Przelewy24
    Przechowalnia Przechowalnia