A tensorflow
implementation of all the compared models for the CIKM 2018 paper: Learning Multi-touch Conversion Attribution with Dual-attention Mechanisms for Online Advertising.
Paper Link: https://arxiv.org/abs/1808.03737.
Slides: http://www.saying.ren/slides/deep-conv-attr.pdf.
We have uploaded a tiny data sample for training and evaluation in this repository.
The full dataset for this project has been published here.
After downloading please replace the sample data in data/ folder with the full data files.
Our raw data is Criteo Attribution Modeling for Bidding Dataset . You can download it and read its description on this page.
Below are the descriptions of our data preprocessing.
-
We group all the impressions by user_id+conversion_id ( regard as one sequence ), shuffle the whole dataset, and then divide it into trainset and testset ( ratio: train 0.8, test 0.2) with negative down sampling (ratio 0.7) at the meanwhile.
-
We create mapping from features from certain fields ([campaign, cat1, cat2, …, cat9]) to index.
-
We turn every line into such format: “time click campaign cat1 cat2 … cat9”
TensorFlow(>=1.2) and dependant packages (e.g., numpy
and sklearn
) should be pre-installed before running the code.
After package installation, you can simple run the code with the demo tiny dataset.
python LR.py [learning rate] # for LR
python SP.py # for Simple Probablistic
python AH.py # for AdditiveHazard
python AMTA.py [learning rate] [batchsize] # for AMTA
python ARNN.py [learning rate] [batchsize] [mu] # for ARNN
python DARNN.py [learning rate] [batchsize] [mu]# for DARNN
We have set default hyperparameters in the model implementation. So the parameter arguments are optional for running the code.
@inproceedings{ren2018learning,
title={Learning Multi-touch Conversion Attribution with Dual-attention Mechanisms for Online Advertising},
author={Ren, Kan and Fang, Yuchen and Zhang, Weinan and Liu, Shuhao and Li, Jiajun and Zhang, Ya and Yu, Yong and Wang, Jun},
booktitle={Proceedings of the 27th ACM International Conference on Information and Knowledge Management},
pages={1433--1442},
year={2018},
organization={ACM}
}