Skip to content

Code base for 'Detecting Fake News using Siamese BERT Network' paper.

License

Notifications You must be signed in to change notification settings

anbhimi/fake_news_detection

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

18 Commits
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Fake News Detection

This GitHub repository is a codebase for the paper 'Detecting Fake News using Siamese BERT Network'. The paper explores the possibilities of classifying news statements. We have experimented using LSTM, Bidirectional LSTM, and BERT models. The Dataset used in this paper is introduced in "Liar, Liar Pants on Fire" paper and is an open-source dataset. The dataset can be found at liar_liar_dataset.

LSTM Model

The LSTM Model uses news-statement which is fed into the network as tokens which are padded to a maximum length. The model is trained to 20 epochs with 'Adam' optimizer and 'BinaryCrossentropy' loss function. The model uses Keras and TensorFlow's framework for training.

Bidirectional LSTM

The architecture of the Bidirectional LSTM Model is similar to the LSTM Model.

Simple BERT Model

The Simple BERT Model uses news-statements which are fed into the BERT network. The news-statements are tokenized, converted into ids, and padded to a maximum length. The resulted tensors are converted into a data loader which is fed into a BERT model for classification. The model is trained for 5 epochs with 'Adam' optimizer and 'CrossEntropyLoss' function in PyTorch.

BERT with Metadata (Siamese BERT Architecture)

BERT with Metadata uses news-statements and metadata (The metadata includes - subjects, speakers, jobs, states, and affiliations). Both the features (news statements and metadata) are subjected to tokenization, conversion to ids, and padding to a maximum length. The resulted tensors are introduced into two separate BERT models. The results from the BERT models are concatenated with credits and used for classification. The model is trained for 5 epochs with 'Adam' optimizer and 'CrossEntropyLoss' function in PyTorch.

BERT with Metadata

Final BERT Model (Siamese BERT Architecture)

The Final BERT Model uses news-statements, metadata, true words, and false words. All the features are subjected to tokenization, conversion to ids, and padding to a maximum length. The resulted tensors are introduced into separate BERT models. The results from the BERT models are concatenated with credits and used for classification. The model is trained for 5 epochs with 'Adam' optimizer and 'CrossEntropyLoss' function in PyTorch.

Final BERT Model

About

Code base for 'Detecting Fake News using Siamese BERT Network' paper.

Topics

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published