Deep Learning with H2O & R
Course Overview
Gain firsthand expertise on installing, configuring, and deploying H2O and make use of the R APIs to handle billions of rows of data without sampling and get accurate predictions faster. Through practical guided exercises leverage the built-in machine learning algos such as generalized linear modeling (linear regression, logistic regression, etc.), Naïve Bayes, principal components analysis, time series, k-means clustering, Random Forest, Gradient Boosting, and Deep Learning at scale.
In cloud labs, practice implementing GBM, Random Forest, GLM, GLRM and become familiar with concepts such as Stacking and Super Learning.
Be the expert of deploying complex deep learning models using H2O with R.
At the end of the training, participants will be able to:
- Install and configure H2O to work with R, Python, Cloud Providers
- Gain a deep understanding of built-in Machine Learning models and usage Access
- H2O features through APIs Build and Train multiple models on a single node or in a cluster
- Train a generalized linear model, generalized low rank models Install and use H2O
- Ensemble to load, train, evaluate model performance Using Storm with H2O for real time prediction
- Deliver scalable models that can work on the complex and large datasets
Pre-requisite
Required: Working knowledge of Java, R, Storm, Machine Learning, Deep Learning Models
Duration
3 days
Course Outline
- Data Science
- H2O
- Building a Smarter Application
- Combining applications with models
- Deploying models into production
- Downloading and Unzipping H2O Package
- Installing H2O from within R
- Installing H2O from within Python
- H2O Quickstart with R
- H2O Cloud Integration
- H2O R Package
- Start H2O
- Decision Boundaries
- Cover Type Dataset
- – Exploratory Data Analysis
- – Deep Learning Model
- – Hyper-Parameter Search
- – Checkpointing
- – Cross Validation
- – Model Save & Load
- Regression and Binary Classification
- Unsupervised Anomaly Detection
- Decision Trees
- Random Forest
- Gradient Boosted Machines
- H2O Implementation
- Cover Type
- Multinomial Model
- Binomial Model
- Introduction
- Basic Model Building(Example)
- Plotting Archetypal Features
- Imputing Missing Values
- Load Training & Test Data
- Create Models
- Export the best model as POJO
- Compile the H2O model as part of the UDF Project
- Copy the UDF to the cluster and load into Hive
- Score with your UDF
- Bagging
- Boosting
- Stacking / Super Learning
- Install H2O Ensemble
- Higgs Demo
- Start H2O Cluster
- Load Data into H2O Cluster
- Specify Base Learner & Metalearner
- Train an Ensemble
- Evaluate Model Performance
- Predict
- Installing the required software
- A brief discussion of the data
- Using R to build a gbm model in H2O
- Exporting the gbm model as a Java POJO
- Copying the generated POJO files into a Storm bolt build environment
- Building Storm and the bolt for the model
- Running a Storm topology with your model deployed
- Watching predictions in real-time