Keywords

1 Introduction

Sentiment analysis is the process of communicating and interpreting feelings, views, ideas, and written emotions to enhance the consistency of various ML and DL methods. With the increased use of digital media for users to express themselves and engage with one another, sentiment analysis is an important technique for extracting their opinions, sentiments, and emotions. It has been one of the most active study fields, using numerous DL and ML models to analyze viewpoints and sentiments. Today, Twitter and social networks discussed what is trending in social media, whether locally or internationally.

The Corona Virus Disease 2019 (COVID-19) Disease has spread exponentially all over the world. By May 9, 2021, there will be 22,295,911 confirmed cases across India [1]. This pandemic has sparked widespread public interest in India. Since the coronavirus emerged so quickly in China, learning the public opinion variations will help the government manage and monitor the pandemic’s growth, as well as support more scientific and successful work. For eg, is it feasible and appropriate for people to lock down in India because it is the first time this has occurred in India? Sentiment analysis is a valuable tool for quickly obtaining people’s insights from vast amounts of text data. From opinion analysis, direct input from the public on government policy on coronavirus can be found, which is critical knowledge for government decision-making for pandemic prevention. However, since India already has 1.36 billion inhabitants, it is difficult to ask anyone how they feel about coronavirus. At the moment, the emotion classification methodology allows one to investigate public sentiment heterogeneity.

According to dsayce.com, in May 2020, there would be an average of 6000 tweets in each second, 350,000 tweets in each minute, 500 million tweets in each day, and 200 billion tweets in each year [3].

In the year 2020, blog.twitter.com announced that #COVID 19 is the most famous hashtag this year. It was tweeted nearly 400 million times, and #StayHome was the third most popular hashtag [blog.twitter.com]. They also stated that this year, 7000 tweets about TV and movies are sent every minute [2].

According to the “DSAYCE” website, 6000 tweets are posted every second on average, with 500 million tweets being reported every day. In Japan, Twitter users set the highest world record of “143,199” tweets in each second when watching Hayao Miyazaki’s animation classic “Castle in the Sky” on Saturday third August 2018. In 2016, 6000 tweets are received each second, 350,000 tweets are sent in each minute on average, 500 million tweets are received every day, and 200 billion tweets are in every year [2].

This research paper has organized into five sections in the II-section we have reviewed the research papers that have already been done, in section-III has covered the theoretical background of the research paper, in the section-IV, we have given proposed methods that I have used, and finally, in section-V, we have given the conclusion and future scope of this research paper.

2 Literature Review

In this section, we review the research work that has already been published by the researcher in the area of sentiment predictions and opinion mining, as well as numerous methods and strategies for analyzing sentiment; we have chosen the best research papers in the area of sentiment predictions.

In sentiment analysis, ML-based methods are commonly used; these approaches include training and testing datasets. Supervised classifier learning was used to prepare the dataset [14]. Uni-grams, bi-grams, n-grams, parts of speech (POS) codes, and bag-of-words (BOW), for example, are features of ML-based strategies. It employs three models: NB, SVM, and maximal entropy (ME). Machine learning analyses sentiments using supervised, semi-supervised, and unsupervised methods.

Kumar, Sudhanshu, et al. propose [15] that this paper aims to use the feature extraction method to evaluate the impact of user’s emotions based on their gender and age, as well as to create a dataset with user reviews and gender and age information, as well as to present user’s expression through extensive experiments, and finally to discuss the comparison of dictionary-based and machine learning methods. This paper employs various sentiment analysis strategies on 900 Facebook users with age and gender records, divides the 900 users into four categories, and then applies sentiment analysis to individual groups. With an accuracy of 78%, [15] used Feature Extraction, Dictionary-Based Classifier, Machine Learning-based Classifier, SVM, NB, and Maximum Entropy to evaluate sentiment analysis.

Kabir, Monika, and colleagues suggest [16] In this case, data sets are mined from various forms of feedback. These data sets are of various sizes and domains, and different methods are used to improve precision in the data sets. For sentiment analysis, [16] employs various ML and classification methods, including RF, DT, SVM, Boosting, Bagging, and DT. For sentiment prediction, three popular data sets such as Amazon, Yelp, and IMDb on various fields such as brands, Movies, and restaurants are used [16] employs data sets to categorize and test emotion polarity.

Jagdale et al. suggest [17] that this paper used a dataset from the Amazon website that included six categories of product reviews: laptops, cameras, televisions, mobile phones, tablets, and video surveillance. Analyze and describe each product review using various ML techniques such as SVM and NB. Using NB and SVM, they achieved accuracy on camera 98.17 percent and 93.54 percent, respectively. They have used a mixed approach, which is a mixture of lexicon-based and machine learning techniques. Machine-learning and lexicon-based methods are used to improve precision.

Jain Achin et al. suggest [18], and this paper is mostly concerned with analyzing English tweets. This paper used R studio and the Twitter API to compile tweets with blogging hashtags. This paper evaluates and implements sentiment analysis on “RenewableEnergy” using five different forms of ML techniques on 5000 Twitter tweets (SVM, KNN, NB, AdaBoost, Bagging). Using the CfsSubsetEval function selection process and SVM, they finally achieved 92.96 percent precision.

Saad, Shihab Elbagir, and colleagues suggest [19] Data sets are derived from the Twitter API and categorized as positive or negative tweets [19] performs sentiment analysis in five stages. Step 1: Assume that each tweet is classified as positive or negative. Step 2: Determine each tweet’s polarity. Step 3: Determine each class’s total emotion polarity. Step 4: assign a score of +2 for high positive, +1 for moderate positive, −1 for moderate negative, and − 2 for high negative to each level [5] analyses emotions in the Twitter data collection using ordinal regression and four ML-based algorithms. These algorithms are used for categorization: SVR (support vector regression), SoftMax, RF (Random Forest), and DT (Decision Trees). For sentiment analysis, SVR and RF have almost identical precision but outperform the Multinomial logistic regression classifier. The Decision tree classifier achieved the best precision of 91.81 percent in this case.

Yadav, Nikhil, and colleagues suggest [20]. This paper makes use of a dataset from Twitter as well as the Kaggle repository. Often employs numerous ML-based algorithms such as NB, RF, DT, SVM, and XGBoost, and then compares multiple classifiers to distinguish tweets on a certain product based on positive or negative sentiments [20] assists the corporate company in developing stronger strategic business strategies for their brands. The dataset is evaluated in five measures in this article. The first step is to compile a list of data sets. Step 2 is the preprocessing step. Step 3 is the feature extraction process. Step 4 is the grouping process. Step 5 is the assessment process. This paper relies mostly on English tweets and dissects human thoughts and emotions.

Chaturvedi, Saumya, et al. suggest [21] that this paper uses some ML-based strategies like NB and SVM for exploring business intelligence emotions that are used for business process decisions, as well as Big-data principles. The decision used the following measures to evaluate the emotions for the market process.

  1. (a)

    Collection of Data

  2. (b)

    Preprocessing and cleaning the dataset

  3. (c)

    Prediction of the sentiments

  4. (d)

    Classification of the sentiments

  5. (e)

    Final output

Dataset collected from online services and analyzed using machine learning algorithms for market intelligence and sentiment analysis.

Hemalatha et al. suggest [22] that this paper is used to test emotions using ML tools such as Sentiment140, Twittratr, and Tweet feel, as well as ML techniques such as the Naive Bayes (NB) algorithm, the Maximum Entropy (ME) algorithm, and data-sets gathered from the social networking fields such as Twitter and others. Then, on the data collection, apply maximum Entropy and sentiment analysis methods, and finally, equate the Maximum Entropy algorithm to other sentiment analysis tools. Figure 1 shows the comparative analysis of maximum Entropy and sentiment analysis tools.

Fig. 1
A stacked bar graph plots percentage versus maximum entropy and sentiment analysis tools for positive percentage, negative percentage, and accuracy. Maximum entropy has 88 percent accuracy, tweet feel with 82 percent and sentiment 140 with 70 percent.

Comparision of Maximum Entropy and Sentiment Analysis Tools [22]

The Fig. 1 above explicitly reveals that maximum entropy outperforms sentiment analysis methods in terms of precision.

The [23] use over 4000,00 ratings for sentiment predictions by using three ML algorithms, and this paper aims to improve sentiment analysis accuracy using NB, DT, and SVM. The evaluation was carried out using a ten-fold cross-validation procedure. They performed sentiment analysis on cell phone ratings, sorted them into favorable and negative emotions in almost equal proportions, and then analyzed the dataset using machine learning models. SVM has the highest precision of the three versions, at 81.75 percent.

Many authors use Deep learning models for sentiment analysis, DL is a form of artificial neural network (ANN) that uses several layers of networks to learn tasks. The composition of a neural network works the same as the human physiological brain and its processing units are known as neurons. Neurons can learn to do tasks in the same way as the brain does. Recurrent Neural Networks (RNN) and Feed-forward Neural Networks (FNN) are the two types of neural networks.

Chandra, Yogesh, et al. suggest [24]. A machine learning model, a deep learning model, and polarity-based approaches are used in this article. To determine the percentage of trust in classification systems, tweet data is first obtained from Twitter and then transmitted through a machine learning classifier. The ratio of positive and negative tweets is calculated using polarity-based techniques. Deep learning algorithms are then used to interpret the tweets. The tweets were classified using CNN-RNN, LSTM, and RNN templates. Deep learning models, such as CNN-RNN and LSTM, outperformed machine learning models. This paper has used the Twitter API to retrieve live tweets and identify them using voting-based classification and polarity-based classification. The dataset was obtained from the Twitter API/Kaggle repositories and the Kaggle kernel. Using machine learning classifiers, we achieved an accuracy of 81 percent to 90 percent for training and testing datasets.

Alarifi, Abdulaziz, et al. suggest [25] This paper gathered data from Amazon and employs a greedy approach since it uses the best classifier, a cat swarm optimization-based long short-term memory neural network (CSO-LSTMNN). The CSO-LSTMNN algorithm, the greedy method, and the particle swarm optimization (PSO) algorithm are used to compare results. The CSO-LSTMNN algorithm produces better results in terms of multifaceted nature and proficiency. CSO-LSTMNN employs five polarity levels: negative, relational negative (RN), neutral, positive, and relational positive (RP), with each level’s threshold value balanced.

Mohamed Neha et al. suggest [26] and uses four deep learning methods for processing movie reviews datasets. This paper uses a data collection from IMDB that contains 50 k movie ratings. The data set is also split into half positive and half negative data sets. The dataset is first pre-processed with Word2Vec before applying deep learning algorithms such as LSTM, RNN CNN, Multilayer Perceptron (MLP), and a hybrid model that combines CNN and LSTM. Finally, the Hybrid model had an accuracy of 89.2 percent, while CNN, LSTM, and MLP had an accuracy of 87.7 percent, 86.64 percent, and 86.74 percent, respectively.

Alharbi et al. [27] provides sentiment analysis for the Arabic language. Many Arabic datasets are gathered from various sources and then subjected to sentiment analysis and comparison with other languages’ datasets. The first small dataset is taken from Opinion Corpus for Arabic (OCA), and it has 500 movie comments. The second dataset is taken from Large Scale Arabic Book Reviews (LABR), and it contains book reviews from consumers on a scale of 5 to 1, and it is taken from www.goodreads.com, and it contains 63,000 Arabic reviews [27] considered three major levels: text, statement, and entity aspect. Find the issue and articulate the dataset with positive or negative sentiments at the document level. They have calculated individual sentences at the sentence level.

Ramadhani et al. suggest [28] for sentiment analysis using deep learning models; they retrieved tweets in both Korean and English from the Twitter API and used three stages for text mining from the Twitter API: information acquisition, information extraction, and data mining. Finally, they ran data into two DL models, Multilayer Perceptron (MLP) and Deep Neural Network (DNN), and obtained 67.45 percent and 77.45 percent accuracy on the training dataset, respectively.

Heikal et al. propose [29] for sentiment prediction of the Arabic dataset, taking 10,000 tweets from the Arabic social sentiment prediction dataset (ASTD). They used an ensemble-based CNN approach and the LSTM deep learning model. First, they used an individual variant on the sample and obtained 64.30 percent and 64.75 percent accuracy using CNN and LSTM, respectively. Eventually, they used an ensemble-based approach using a soft voting classifier and obtained 65.05 percent accuracy.

Wang, Jian, et al. recommend [30] for Chinese online shopping product reviews. They collected 13,000 votes, 7000 of which were positive and the rest were negative. They used three models for sentiment analysis: LSTM, LSTM dependent on Nadam with L2 regularisation, and Bi-directional LSTM. After reviewing the performance of both models, they discovered that the LSTM based on Nadam with L2 had greater accuracy than the other two models, which provided about 98 percent accuracy.

3 Theoretical Background

  • ML: For sentiment predictions, we have taken many machine learning techniques SVM, RF, DT, K-Nearest Neighbor (KNN), Logistic Regression (LR), and ensemble voting classifier.

  • SVM: Support vector machines are supervised machine learning algorithms that are both efficient and scalable. They are used for classification and regression. However, they are most often found in classification problems. SVMs were initially developed in the 1960s, but they were refined in 1990. As opposed to other machine learning algorithms, SVMs have a distinct implementation process. Because of their ability to manage several continuous and categorical variables, they are highly common [4].

An SVM model is essentially a representation of various groups in a hyperplane in a multidimensional space. SVM creates a hyperplane iteratively to minimize error. SVM aims to classify datasets to find the maximal marginal hyperplane (MMH) [4].

  • Support Vectors: Support vectors are data points that are in marginal lines which are parallel to the hyperplane. These data points are mainly used for separating the data sets into classes.

  • Hyperplane: As seen in the above Fig. 2, it is a decision plane or space that divides the dataset into different classes.

  • Margin: It can be described as the distance of the hyperplane and marginal plane that separates groups of closet data points. It is computed as the distance between the hyperplane and the marginal line. A maximum margin is regarded as a positive margin, whereas a small margin is regarded as a poor margin.

  • RANDOM FOREST: Tin Kam Ho prepared a random forests decision algorithm in 1995, with the help of the random subspace technique, in which Ho described how to apply Eugene Kleinberg’s “stochastic discrimination” approach to classification.

RF, also known as random decision forests, is an ensemble-based technique for classification, regression, and other tasks that works like training a large number of decision trees and then outputting the class that is the mode of the classes (classification) or mean/average estimation (regression) of the individual trees [5] (Fig. 3).

Fig. 2
A graph of y axis versus x axis plots spherical objects grouped as class A and class B. The distance between 2 objects is called as hyperplane and the object near the hyperplane is called support vectors. A margin line runs perpendicular to the line between two classes.

SVM Working [4]

Fig. 3
A flow diagram begins with training set leads to training samples 1, 2, to n. It further leads to voting and prediction.

Working of RF [6]

Random Forest Algorithm took the following four steps to classify the dataset.

  1. (a)

    Selection of Random sample

  2. (b)

    Construct a decision tree and predict the result

  3. (c)

    Perform ensemble technique

  4. (d)

    Select the best result as the final prediction

  • DECISION TREE: these are statistical modeling techniques used in ML and DT for analytics and data processing. It employs a DT to progress from assumptions about the item (represented by nodes) to judgments about the items and target value (represented in the leaf nodes). Classification trees are tree structures in which the target nodes have a distinct range of values [7] (Fig. 4).

  • Root Node: The decision tree begins at the root node. It contains the whole dataset, which is then split into two or more homogeneous sets.

  • Leaf Node: Leaf nodes are the tree’s last output node, and the tree cannot be further separated until obtaining a leaf node.

The decision tree uses three main methods to make a decision which are the Entropy, Gini index, and Information gain.

Fig. 4
A flow diagram of decision node leads to 2 decision nodes. Decision node 1 has 2 leaf nodes labeled, sub tree. Decision node 2 has leaf node and decision node further leads to 2 leaf nodes.

Working of DT [8]

ENTROPY (E): It is used to assess a dataset’s impurity or randomness.

$$ \textbf{E}\left(\textbf{X}\right)={\textbf{I}}_{\textbf{E}}\left(\textbf{P}\textbf{1},\textbf{P}\textbf{2},\dots \dots, \textbf{P}\textbf{j}\right) $$
$$ \textbf{E}\left(\textbf{X}\right)=-\sum \limits_{\textbf{i}=\textbf{1}}^{\textbf{j}}{\textbf{p}}_{\textbf{i}\kern0.375em {\log}_{\textbf{2}}{\textbf{P}}_{\textbf{i}.}} $$

INFORMATION GAIN (IG): To identify the best function that can be used as a root node in terms of information gain.

$$ \textbf{IG}=\textbf{E}\left(\textbf{X}\right)\hbox{--} \textbf{E}\left(\textbf{X}|\textbf{a}\right) $$
$$ \textbf{IG}=-\sum \limits_{i=\textbf{1}}^j{p}_{i\kern0.375em {\log}_{\textbf{2}}{P}_i}-\sum \limits_{i=\textbf{1}}^j{p}_{\left(i|a\right)\kern0.375em {\log}_{\textbf{2}}{P}_{\left(i|a\right)}} $$

Where E(X) Entropy of X and E(X|a) Sum of all entropy children P is the probability.

GINI INDEX (GI): It is calculated by subtracting the total of every class’s squared percentages from one. It always considers the largest partitions that are simple to enforce, whereas information gain always considers smaller partitions that have distinct values.

KNN: k-NN is a non-parametric classification technique introduced in 1951 by Evelyn Fix and Joseph Hodges and then enhanced by Thomas Cover. It is used in both classification and regression problems. In all circumstances, the response consists of the k closest training instances in the data set [9]. KNN predicts the new data points based on ‘feature similarities,’ and the new data point will predict a value based on how exactly it resembles the points in the training collection [10] (Fig. 5).

Fig. 5
2 graphs illustrate working of K N N. The graphs plot X 2 versus X 1. Before K N N, two categories A and B of spherical objects with new data point at the center. After K N N the new data data point is assigned to category 1.

working of KNN [11]

KNN uses the many steps to classify the datasets which are.

  1. (a)

    ‘K’ number of neighbors selection.

  2. (b)

    Euclidean distance calculation for ‘K’.

  3. (c)

    Taken k as per the Euclidean distance

  4. (d)

    Count the number of data points in each group among these k neighbors.

  5. (e)

    Assign the latest data points to the group with the highest number of neighbors.

LR: it is a classification technique that uses supervised learning to predict the likelihood of a target value. Because the target variable is dichotomous, means there are only two possible classes [12].

An LR model forecasts P(Y = 1) as a function of X mathematically. It is a very basic ML technique that can be used to solve various classification problems such as spam diagnosis, diabetes prediction, cancer detection, Sentiment analysis, and so on [12]. We fit an “S” form logistic function that predicts two maximum values in LR (0 or 1) (Fig. 6).

Fig. 6
A line graph plots fir Y versus X has an S curve with 3 points where y equals 0.3 and y equals 0.8. A horizontal dashed line in the middle at 0.5 represents threshold value.

working of LR [13]

The logistic feature curve shows the probability of anything like whether the cells are cancerous or not, when a mouse is a fat or not depending on its weight, and so on [13].

ENSEMBLE VOTING CLASSIFIER: Ensemble approaches are strategies for developing different models and then combining them to achieve better performance. Ensemble models usually yield more reliable results than a single model [31]. A Voting Classifier is an ML technique that trains on various ML and DL models and predicts an outcome (class) based on the highest likelihood of the taken class as the output [32].

It essentially merges the results of each classifier that passed into Voting Classifier and then predicts the output class depending on the voting majority. Instead of developing individual dedicated models and determining their precision, we can develop a single model that trains on these models and predicts performance based on their cumulative plurality of voting for every output class [32] (Fig. 7).

Fig. 7
A block diagram begins with intrusion data classifies into 4 classifiers 1, 2, 3, and 4. Each has corresponding decision boundaries further leads to combiner and final ensemble.

Architecture of voting classifier [33]

The voting classifier has two methods to ensemble the models.

Hard Voting Classifier: The predicted performance class in hard voting is the class with the maximum majority of votes, i.e. the class with the high probability of being analyzed by any of the classifiers. Assume that we have three classifiers that have their output classes (A, A, B), and the majority predicted A. As a result, A will be the final prediction [32].

Soft Voting Classifier: The performance class of soft voting is the prediction technique that depends on the average of the probabilities assigned to that class. Assume that provided any input to three models, the estimation chances for each class A = (0.30, 0.47, 0.53) and B = (0.30, 0.47, 0.53). (0.20, 0.32, 0.40). So, if the average for class A is 0.4333 and the average for class B is 0.3067, class A is the winner since it has the highest likelihood averaged by each classifier [32].

PIPELINING: Pipelines are used to divide the machine learning workflows into separate, interchangeable, scalable sections that can then be pipelined together to continually increase the model’s accuracy and produce a reliable algorithm. Generally pipelining includes the following steps.

Preprocessing and Vocab building: Unwanted texts (stop words), punctuation, URLs, handles, and so on that have little sentimental meaning are removed. Then adding unique preprocessed words to a vocabulary.

Feature Extraction: Iteratively going through and data illustration to remove features using a frequency dictionary and then creating a function matrix.

4 Proposed Method

In this part, we will go over the methodology we used in this paper. We build a Voting ensemble of ML models for the sentiment prediction of Covid-19 tweets. In the Voting ensemble-based model, we have included various ML classifiers for classifying the dataset into positive as well as negative tweets. It is divided into four major stages. The first stage is data preprocessing, which includes word sorting and tokenization, as well as stop words elimination and cleaning the dataset. The second stage is pipelining, in which we have passed two methods first one is TFIDF vectorizer and the second one is a machine learning classifier. TFIDF vectorizer and classifier model are used in the third stage for making a single model. In the fourth stage, we have applied a hard voting classifier (Fig. 8).

Fig. 8
A flow diagram begins with Covid 19 data sheets leads to pre-processing dataset leads to 5 pipelining. Each leads to T F I D F plus L R, T F I D F plus S V M, T F I D F plus D T, T F I D F plus K N N, and T F I D F plus R F, together leads to hard voting classifier, and final result.

Working of the ensemble-based model

Figure 1 shows the working of the ensemble-based technique, in the preprocessing stage of the technique we have used the following steps.

  • Remove special characters: in this step, we have removed special characters from the tweets like _,? $, etc.

  • Remove Accented characters: In this, we have removed accented characters from the tweets like â, î or ô, etc.

  • Remove Emails: In this step, we have removed emails from the tweets.

  • Remove html tags: In this steps we have removed html tags like <html>, </html>, etc.

  • Remove URLs: In this step, we have removed all URLs from the tweets.

  • Lower case: In this step, we have converted all upper case tweets into lower case.

After preprocessing we have used pipelining in that we have passed TFIDF vectorizer and machine learning classifier. Then prepare a single pipeline model. Finally, we have compared all models with their accuracy.

After preparing all models, we used an ensemble voting classifier to make average accuracy. By using the Hard voting classifier we got 94.7% accuracy on the testing dataset.

5 Experiment and Results

For the experiment, we have used Covid-19 tweets which have been downloaded from the Kaggle repository that has 1.4 million tweets related to covid-19. Dataset has four types of sentiments which are positive, extremely positive, negative, and extremely negative. For the experiment purpose, we have converted extremely positive as positive and extremely negative as negative. Then by using an individual machine learning classifier we got the following results (Table 1) by using Random Forest ML – classifier we got 90% accuracy which is better than other classifiers. Figure 9 shows the comparative analysis of individual models.

Table 1 Individual model accuracy
Fig. 9
A bar graph illustrates the accuracy values of classifiers as follows: D T, 81; K N N, 72; S V M, 81; L R, 76; R F, 90.

Comparative analysis of classifiers

After calculating the individual accuracy we got that some classifiers perform well and some are not performing well so we have used the concept of pipelining with the help of TFIDF vectorizer. In the pipeline, we have passed one ML-classifier with TFIDF vectorizer and prepared a model, After preparing all models we have compared their results using Table 2 which shows the accuracy of each model.

Table 2 Pipeline model accuracy

After preparing models using the concept of pipeline techniques, we got that some models perform well and some are performing poorly so to overcome this problem we have used an ensemble model for the average accuracy. Figure 10 shows the comparative analysis of pipelining models.

Fig. 10
A bar graph illustrates the accuracy values of pipelining models as follows: T F I D F plus L R, 89; T F I D F plus S V M, 95; T F I D F plus D T, 91; T F I D F plus K N N, 77; T F I D F plus R F, 93; Adaboost plus D T, 87.

Comparative analysis of pipelining models

Figure 10 shows that the pipeline model using TFIDF + SVM has the highest accuracy and TFIDF + KNN has the worst accuracy compared to other models.

To overcome the above problem we have used the concept of ensemble model using a Hard Voting Classifier. Finally, we have passed all models in the voting machine and got an accuracy of 94.70% which is better than other classifiers and pipeline models.

6 Conclusion and Future Scope

In this research paper, we introduced an ensemble-based model for sentiment prediction. This paper aims to support the government in this pandemic situation for taking the right decision using public opinions, it will also be helpful for the researcher to research in the area of sentiment predictions. This paper uses the concept of pipelining in which we have passed a machine learning classifier with a TFIDF vectorizer.

This research paper implementation part is divided into four levels, The first level is data preprocessing, which involves word sorting and tokenization, as well as stop words removal and dataset cleaning. The second level is pipelining, by which we have passed two methods: a TFIDF vectorizer and a machine learning classifier. In the third level, a TFIDF vectorizer and classifier model was used to create a single model. We used a hard voting classifier in the fourth level. We have used covid-19 tweets which have been downloaded from the Kaggle. We have also calculated the accuracy of the individual model and then finally by using hard voting we got an accuracy of 94.70% which is better than all models.

For future studies, we have observed that if we train our model using the BERT model and some neural network models, we can improve accuracy. I have also found that no work has used ensemble-based models using BERT and Neural networks.