Skip to content

Does Deception Leave a Content Independent Stylistic Trace?

Notifications You must be signed in to change notification settings

ReDASers/CODASPY-2022-Deception

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

12 Commits
 
 
 
 
 
 
 
 

Repository files navigation

Introduction

This repository contains code from our CODASPY 2022 poster Does Deception Leave a Content Independent Stylistic Trace?

Datasets

You can download our cleaned datasets from Zenodo.

We use the following datasets for our experiments:

  • The Amazon Reviews dataset consists of 21,000 English Amazon Reviews labeled as either real or fake. The reviews cover a variety of products with no particular product dominating the dataset. We filter out the non-English reviews and mark the fake and real reviews as deceptive and non-deceptive respectively. The final dataset is evenly balanced with 10,493 deceptive samples and 10,481 non-deceptive samples.

  • The Employment Scam Aegean Dataset, henceforth referred to as the Job Scams dataset consists of 17,880 human-annotated job listings. We use the description field as the sample text and the fraudulent field as our label. We clean the descriptions by removing all HTML tags, empty descriptions, and duplicates. Our final dataset is a heavily unbalanced dataset with 603 deceptive and 14,173 non-deceptive samples.

  • The Email Benchmarking dataset is a phishing dataset consisting of both legitimate and phishing emails. We use the bodies extracted using PhishBench as the text.

  • The Liar dataset consists of political statements made by US speakers assigned a fine-grain truthfullness label by PolitiFact. We use the claim as the text, and label "pants-fire," "false," "barely-true" "half-true," as deceptive and "mostly-true," and "true" claims as non-deceptive.

    • It contains 5669 deceptive and 7167 truthful statements.
    • On 11/27/2021, we noticed that a lot of the sample claims start with the phrase "Says that [claim]" or "Says [claim]". We modified our cleaning script to remove this signature.
  • The WELFake dataset is a more general fake news dataset intended to prevent overfitting. It combines 72,134 news articles from four pre-existing datasets (Kaggle, McIntire, Reuters, and BuzzFeed Political). It is roughly balanced with 35,028 real news articles and 37,106 fake news articles. As with the LIAR dataset, we consider fake news as deceptive and real news as truthful.

Code

In the data folder, we have a set of python scripts. The read_[dataset].py scripts read the raw data and compile it to jsonlines. The clean_[dataset].py files clean the dataset.

Note: Because of langdetect's stochastic nature, the datasets generated are non-deterministic.

Experiments

In the deep learning folder, we have our code for the deep learning experiments.

Citation

If this code helped you, please cite our CODASPY paper.

@inproceedings{
    10.1145/3508398.3519358,
    author = {Zeng, Victor and Liu, Xuting and Verma, Rakesh M.},
    title = {Does Deception Leave a Content Independent Stylistic Trace?},
    year = {2022},
    isbn = {9781450392204},
    publisher = {Association for Computing Machinery},
    address = {New York, NY, USA},
    url = {https://doi.org/10.1145/3508398.3519358},
    doi = {10.1145/3508398.3519358},
    booktitle = {Proceedings of the Twelfth ACM Conference on Data and Application Security and Privacy},
    pages = {349–351},
    numpages = {3},
    keywords = {domain-independent deception detection, dataset quality/cleaning},
    location = {Baltimore, MD, USA},
    series = {CODASPY '22}
}

About

Does Deception Leave a Content Independent Stylistic Trace?

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published