All Projects → Lapis-Hong → Wide_deep

Lapis-Hong / Wide_deep

Licence: mit
Wide and Deep Learning for CTR Prediction in tensorflow

Programming Languages

python
139335 projects - #7 most used programming language

Wide and Deep Learning for CTR Prediction in tensorflow

Overview

A general Wide and Deep Joint Learning Framework. Deep part can be a simple Dnn, Dnn Variants(ResDnn, DenseDnn), MultiDnn or even combine with Cnn (Dnn-Cnn).

Here, we use the wide and deep model to predict the click labels. The wide model is able to memorize interactions with data with a large number of features but not able to generalize these learned interactions on new data. The deep model generalizes well but is unable to learn exceptions within the data. The wide and deep model combines the two models and is able to generalize while learning exceptions.

The code uses the high level tf.estimator.Estimator API. This API is great for fast iteration and quickly adapting models to your own datasets without major code overhauls. It allows you to move from single-worker training to distributed training, and it makes it easy to export model binaries for prediction.

The input function for the Estimator uses tf.data.Dataset API, which creates a Dataset object. The Dataset API makes it easy to apply transformations (map, batch, shuffle, etc.) to the data. Read more here.

The code is based on the TensorFlow wide and deep tutorial.

Extensions

  1. provide very flexible feature configuration and train configuration.
  2. scalable to arbitrarily train data size in production environment.
  3. support multi value feature input (multihot).
  4. support distributed tensorflow
  5. support custom dnn network (arbitrary connections between layers) with flexible options.
  6. add BN layer; activation_fn; l1,l2 reg; weight decay lr options for training.
  7. support dnn, multidnn joint learning, even combine with cnn.
  8. support 3 types normalization for continuous features.
  9. support weight column for imbalance sample.
  10. provide tensorflow serving for tf.estimator.
  11. provide scripts to do data proprocess using pyspark (generate continuous features from category features).

Running the code

Setup

cd conf
vim feature.yaml
vim model.yaml
vim train.yaml
...

Training

You can run the code locally as follows:

cd python
python train.py

or use shell scripts as follows:

cd scripts
bash train.sh

Testing

python eval.py

or use shell scripts as follows:

bash test.sh

Distributed Training

run the code on ps as follows:

cd scripts
bash run_ps.sh

TensorBoard

Run TensorBoard to inspect the details about the graph and training progression.

tensorboard --logdir=./model/wide_deep
Note that the project description data, including the texts, logos, images, and/or trademarks, for each open source project belongs to its rightful owner. If you wish to add or remove any projects, please contact us at [email protected].