Skip to content

suraj-s13/Fake_News_Detection

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

7 Commits
 
 
 
 
 
 

Repository files navigation

CodeClause_Fake_News_Detection

forthebadge made-with-python

What It Does:

Fake news on different platforms is spreading widely and is a matter of serious concern, as it can cause social wars and permanent breakage of the bonds established among people. A lot of research is already going on focused on the classification of fake news.

1️⃣ Data Collection
2️⃣ Data Preprocessing
3️⃣ Exploratory Data Analysis
4️⃣ Dataset Balancing & Scaling
5️⃣ Machine Learning Models Training & Evaluation

Prerequisites:

I would highly recommend that before the hack night, you should have some kind of toolchain and development environment already installed and ready. If you have no idea where to start with this, try a combination like:

1️⃣ Python
2️⃣ scikit-learn / sklearn
3️⃣ Pandas
4️⃣ NumPy
5️⃣ Swaborn
6️⃣ An environment to work in - something like Jupyter or colab

For Linux people, your package manager should be able to handle all of this. If it somehow can't, see if you can at least install Python and pip and then use pip to install the above packages.

Dataset:

You can collect raw dataset from here.

Attribute Information:
1️⃣ Title
2️⃣ Text
3️⃣ Lable

Machine Learning Model:

1️⃣TfidfVectorizer

TF (Term Frequency): The number of times a word appears in a document is its Term Frequency. A higher value means a term appears more often than others, and so, the document is a good match when the term is part of the search terms.

IDF (Inverse Document Frequency): Words that occur many times a document, but also occur many times in many others, may be irrelevant. IDF is a measure of how significant a term is in the entire corpus.

The TfidfVectorizer converts a collection of raw documents into a matrix of TF-IDF features.

2️⃣PassiveAggressiveClassifier

Passive Aggressive algorithms are online learning algorithms. Such an algorithm remains passive for a correct classification outcome, and turns aggressive in the event of a miscalculation, updating and adjusting. Unlike most other algorithms, it does not converge. Its purpose is to make updates that correct the loss, causing very little change in the norm of the weight vector.

Updated on 2025-02-02 09:44:03

About

No description, website, or topics provided.

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published