Sentimental Analysis
Sentimental Analysis
ABSTRACT:
The objective of this project is to conduct a comparative study of
sentimental analysis through text using machine learning techniques,
specifically employing an Early Classification Based Approach for Fault
Classification. The emotions under consideration are sadness, joy, love, anger,
fear, and ego, which are commonly expressed in text data. To achieve this
goal, three machine learning algorithms, namely Stochastic Decision Tree
(DT), Random Forest (RF), and Long Short-Term Memory (LSTM), are
utilized. The study utilizes a data set containing diverse text samples
expressing different emotions, which is processed to remove irrelevant
information and then split into training and testing sets. The DT, RF, and
LSTM models are trained on the training set and evaluated on the testing set
using various evaluation metrics. The experimental results provide insights
into the performance of the three algorithms for emotion classification from
text. The findings reveal that DT and RF achieve comparable accuracy levels,
while LSTM outperforms them in terms of overall classification accuracy.
Key Words: Decision Trees, Random Forest, Long ShortTerm Memory
INPUTS CONSIDERED:
● Dataset of emotions:
This set consists of set of emotions sourced from “kaggle.com”.
● Image Pre-processing:
Techniques for data cleaning, data training.
● Machine Learning Libraries:
Machine learning libraries: NumPy, IO, Keras
Frameworks: Django
Database: SQL
● Hardware Tools: PyCharm, SQL workbench
EXPECTED OUTPUT:
● Emotion Identification:
Identifies the emotions from the input data.
● Emotion Classification:
Predicts the type of emotion.
LITERATURE STUDY:
RAM : 8GB
Software Configuration:
Operating System : windows 7/8/10
IDE : PyCharm
A tree has many analogies in real life and turns out that it has
influenced a wide area of machine learning, covering both classification and
regression. In decision analysis, a decision tree can be used to represent
decisions and decision making visually and explicitly. As the name goes, it
uses a tree-like model of decisions. Though a commonly used tool in data
mining for deriving a strategy to reach a particular goal. A decision tree is
drawn upside down with its root at the top. In the image on the left, the bold
text in black represents a condition/internal node, based on which the tree
splits into branches/ edges. The end of the branch that doesn’t split anymore
is the decision/leaf, in this case, whether the passenger died or survived,
represented as red and green text respectively Although, a real dataset will
have a lot more features and this will just be a branch in a much bigger tree,
but you can’t ignore the simplicity of this algorithm. The feature importance
is clear, and relations can be viewed easily. his methodology is more
commonly known as learning decision tree from data and the above tree is
called Classification tree as the target is to classify passengers as survived or
died. Regression trees are represented in the same manner, just they predict
continuous values like the price of a house. In general, Decision Tree
algorithms are referred to as CART or Classification and Regression Trees.