Page 786 - Emerging Trends and Innovations in Web-Based Applications and Technologies
P. 786

International Journal of Trend in Scientific Research and Development (IJTSRD) @ www.ijtsrd.com eISSN: 2456-6470
             Each model is executed with a fixed random state of 42 to   independence assumption holds reasonably true. The model
             ensure consistent and reproducible results.        has three variations: Multinomial, Bernoulli, and Gaussian.
                                                                The Multinomial Naïve Bayes is particularly suited for text
             Figure 2 Comparison of Most Common Words in Fake vs
                                                                data, as  it  models word occurrences in documents. Input
             Real News Articles
                                                                features are generally derived from BoW or TF-IDF, and the
             4.  Machine Learning Models                        model computes the likelihood of a document being fake or
             4.1.  Random Forest                                real  based  on  word  frequencies.  Naïve  Bayes  is
             Random Forest is a widely used ensemble machine learning   computationally  efficient  and  works  well  for  smaller
             method, particularly effective in text classification tasks like   datasets. However, its independence assumption can limit its
             fake  news  detection.  It  functions  by  building  multiple   ability  to  capture  complex  dependencies  between  words,
             decision trees during training, and the final output is based   especially  in  more  nuanced  tasks  where  contextual
             on the majority vote from individual tree classifications (for   relationships matter [18]. Despite these limitations, Naïve
             classification) or an average prediction (for regression). For   Bayes  is  often  chosen  for  fake  news  detection  due  to  its
             Natural  Language  Processing  (NLP),  Random  Forest  uses   speed, simplicity, and reasonable performance, particularly
             features extracted from text, such as term frequency-inverse   when used as a baseline model or alongside other methods.
             document  frequency  (TFIDF)  or  word  embeddings,  to   4.4.  Neural Networks
             perform classification [16].
                                                                Neural Networks, particularly deep learning architectures,
             A key strength of Random Forest lies in its ability to process   have  significantly  advanced  NLP  by  enabling  models  to
             high-dimensional datasets, a typical characteristic of textual   autonomously learn features from textual data. In fake news
             data.  By  combining  predictions  from  multiple  trees,  it   detection,  neural  networks  use  preprocessed  text  inputs,
             reduces  overfitting  and  enhances  generalization.  In  fake   such as word embeddings or token sequences, to understand
             news detection, Random Forest can detect patterns in word   intricate relationships and contextual details. The simplest
             usage and sentence structure across large datasets, helping   form, Feedforward Neural Networks (FNNs), processes input
             identify  important  features.  Additionally,  it  ranks  feature   features via fully connected layers.
             importance, allowing for the interpretation of which terms or   However, more advanced architectures, such as Recurrent
             phrases  are  most  significant  in  predictions.  However,   Neural  Networks  (RNNs)  and  Long  Short-Term  Memory
             effective  hyperparameter  tuning—such  as  the  number  of   (LSTM) networks, are better suited for sequential data like
             trees and maximum tree depth—is necessary for optimal   text, capturing the flow of information across a document
             results. Despite its reliance on aggregated features, which   and identifying inconsistencies that may indicate fake news.
             might  miss  some  contextual  nuances,  Random  Forest   Transformer-based models, like BERT, further enhance this
             remains a reliable and interpretable method for tasks like
                                                                by  leveraging  self-attention  mechanisms  to  understand
             fake news detection.
                                                                context and relationships across entire documents. Neural
             4.2.  Support Vector Machine (SVM)                 networks  are  particularly  strong  in  processing  large,
             Support  Vector  Machine  (SVM)  is  a  powerful  supervised   unstructured data and uncovering deep patterns that other
             learning technique well-suited for text classification tasks   models  might  miss.  However,  they  require  significant
             due to its robust handling of high-dimensional and sparse   computational resources and vast labeled datasets for high
             data. In the context of NLP tasks like fake news detection,   accuracy. Techniques like dropout regularization are often
             SVM identifies a hyperplane that best separates data points   used to prevent overfitting. Despite these challenges, neural
             from different classes. By maximizing the margin between   networks are currently the state-of-the-art in NLP and excel
             the classes, SVM ensures reliable and precise classification. It   in tasks such as fake news detection.
             employs  kernel  functions—such  as  linear,  polynomial,  or   4.5.  Logistic Regression
             radial  basis  functions  (RBF)—to  map  data  into  higher-  Logistic Regression is a simple but effective linear model
             dimensional spaces where linear separation is achievable.   commonly employed for binary classification tasks, including
             For text classification, features are commonly derived from   fake news detection. It predicts the probability of an instance
             bag-of-words  (BoW),  TF-IDF,  or  word  embeddings.  SVM   belonging to a particular class by applying a logistic function
             excels  in  fake  news  detection  by  analyzing  word   to the weighted sum of input features. In NLP, features for
             distributions and patterns [17].
                                                                Logistic Regression are often derived from techniques like
             One of SVM's major advantages is its ability to manage noisy   bag-of-words,  TF-IDF,  or  word  embeddings.  Despite  its
             and  imbalanced  datasets,  making  it  well-suited  for  real-  simplicity, Logistic Regression performs remarkably well in
             world fake news detection tasks. However, its computational   fake news detection, especially when combined with robust
             cost increases with large datasets, especially when nonlinear   feature engineering. The model assigns weights to features
             kernels are used. Proper tuning of hyperparameters, such as   based  on  their  relevance  to  the  classification,  making  it
             kernel selection and regularization parameters, is critical to   interpretable. For example, it can highlight specific terms or
             achieving optimal  performance. Despite these  challenges,   phrases  strongly  associated  with  fake  news.  However,
             SVM remains a popular tool for NLP due to its scalability and   Logistic  Regression  struggles  to  capture  nonlinear
             efficiency in high-dimensional settings.           relationships and contextual nuances in text, which are vital
             4.3.  Naïve Bayes                                  for  identifying  more  complex  fake  news  patterns.
                                                                Regularization methods, such as L1 and L2 penalties, are
             Naïve Bayes is a probabilistic approach often used in text
                                                                frequently used to avoid overfitting, especially with high-
             classification tasks due to its simplicity and effectiveness.
                                                                dimensional data. While not as sophisticated as advanced
             Based  on  Bayes'  theorem,  it  assumes  that  features  are
                                                                models like neural networks, Logistic Regression remains a
             conditionally independent given the class label. Despite the   solid baseline for fake news detection due to its simplicity,
             "naïve" assumption, Naïve Bayes performs well in NLP tasks,   efficiency, and interpretability.
             such  as  fake  news  detection,  particularly  when  the

             IJTSRD | Special Issue on Emerging Trends and Innovations in Web-Based Applications and Technologies   Page 776
   781   782   783   784   785   786   787   788   789   790   791