https://www.gnu.org/licenses/gpl-3.0-standalone.htmlhttps://www.gnu.org/licenses/gpl-3.0-standalone.html
This repository contains data and code for emotion detection in text, developed in the EASIER project on sign language translation. The end goal within the project is to provide emotional cues to an avatar for the translation direction from spoken to sign language. While the EASIER project is multilingual, emotion detection so far is restricted to German.
https://choosealicense.com/licenses/other/https://choosealicense.com/licenses/other/
Dataset Card for "emotion"
Dataset Summary
Emotion is a dataset of English Twitter messages with six basic emotions: anger, fear, joy, love, sadness, and surprise. For more detailed information please refer to the paper.
Supported Tasks and Leaderboards
More Information Needed
Languages
More Information Needed
Dataset Structure
Data Instances
An example looks as follows. { "text": "im feeling quite sad… See the full description on the dataset page: https://huggingface.co/datasets/dair-ai/emotion.
Attribution-NonCommercial-NoDerivs 4.0 (CC BY-NC-ND 4.0)https://creativecommons.org/licenses/by-nc-nd/4.0/
License information was derived automatically
The audio dataset consists of a collection of texts spoken with four distinct emotions. These texts are spoken in English and represent four different emotional states: euphoria, joy, sadness and surprise. Each audio clip captures the tone, intonation, and nuances of speech as individuals convey their emotions through their voice. The dataset includes a diverse range of speakers, ensuring variability in age, gender, and cultural backgrounds*, allowing for a more comprehensive representation of the emotional spectrum. The dataset is labeled and organized based on the emotion expressed in each audio sample, making it a valuable resource for emotion recognition and analysis. Researchers and developers can utilize this dataset to train and evaluate machine learning models and algorithms, aiming to accurately recognize and classify emotions in speech.
Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
In this dataset, there are Bangla sentences manually labeled with three emotions- Happy, Sad, and Angry.
This dataset is used in Multilabel sentiment analysis and emotion detection for YouTube comments in different kinds of Bengali videos.
There are two files in the folder. There are might be multiple comments with same text. Also it may be noted that, the comments collected here contain abusive and vulgar words, slangs and personal attack. Therefore, we ensure that all annotators are adults.
Sentiment.csv
Id - Unique id number for the comment. Text - Text of the data Label - 1 (3 class label) or 2 (5 class label) Score - Denotes the polarity of the comment. In three class labelling : 1(positive), 0 (neutral), -1(negative) In three class labelling : 2 (highly positive), 1(positive), 0 (neutral), -1(negative), -2(highly negative) Lan - Language of the comment. EN (English), BN (Bengali), RN (Romanized Bangla) Domain - Category of the video.
Emotion.csv
Id - Unique id number for the comment. Text - Text of the data emotion - Corresponding emotion of the comment. Anger/Joy/Disgust/Fear/Surprise/Sad/None (no emotion found) Lan - Language of the comment. EN (English), BN (Bengali), RN (Romanized Bangla) Domain - Category of the video.
If you use the dataset in any research work, please cite the following paper as
N. Irtiza Tripto and M. Eunus Ali, "Detecting Multilabel Sentiment and Emotions from Bangla YouTube Comments," 2018 International Conference on Bangla Speech and Language Processing (ICBSLP), Sylhet, 2018, pp. 1-6.
doi: 10.1109/ICBSLP.2018.8554875
It will be helpful for researchers specially in analyzing sentiments from social media in non-English language
Dataset Description:
This dataset comprises two text files, one containing a list of positive words and the other a list of negative words. These files are intended to serve as essential resources for sentiment analysis and natural language processing tasks.
Positive Words File: This file contains a collection of words and terms that typically convey positive sentiment or emotions. These words are often associated with happiness, satisfaction, approval, or positive experiences.
Negative Words File: The second file includes a compilation of words and phrases that commonly express negative sentiments or emotions. These words are often related to displeasure, disappointment, criticism, or negative experiences.
Use Cases:
The dataset can be used in various applications and research areas, including:
Sentiment Analysis: Researchers and developers can employ these word lists to help assess the sentiment of text data and categorize it as positive, negative, or neutral.
Text Classification: These datasets can be utilized in text classification tasks, such as determining the sentiment of product reviews, social media posts, or customer feedback.
Emotion Detection: The words in these lists can aid in emotion detection and understanding the emotional tone of text.
NLP Research: Linguists and researchers in the field of natural language processing (NLP) may use these datasets to analyze language patterns and emotional expression in texts.
Content Analysis: Businesses and organizations can leverage this data to gain insights into public opinion, brand perception, and content analysis.
These lists of positive and negative words serve as valuable references for sentiment-related analyses, enabling automated assessment of the emotional content of textual data in a wide range of applications.
Attribution-ShareAlike 4.0 (CC BY-SA 4.0)https://creativecommons.org/licenses/by-sa/4.0/
License information was derived automatically
Emotions in Literature
Detecting Fine-Grained Emotions in Literature
Please cite:
@Article{app13137502, AUTHOR = {Rei, Luis and Mladenić, Dunja}, TITLE = {Detecting Fine-Grained Emotions in Literature}, JOURNAL = {Applied Sciences}, VOLUME = {13}, YEAR = {2023}, NUMBER = {13}, ARTICLE-NUMBER = {7502}, URL = {https://www.mdpi.com/2076-3417/13/13/7502}, ISSN = {2076-3417}, DOI = {10.3390/app13137502} }
Emotion detection in text is a fundamental aspect of affective computing and is closely linked to natural language processing. Its applications span various domains, from interactive chatbots to marketing and customer service. This research specifically focuses on its significance in literature analysis and understanding. To facilitate this, we present a novel approach that involves creating a multi-label fine-grained emotion detection dataset, derived from literary sources. Our methodology employs a simple yet effective semi-supervised technique. We leverage textual entailment classification to perform emotion-specific weak-labeling, selecting examples with the highest and lowest scores from a large corpus. Utilizing these emotion-specific datasets, we train binary pseudo-labeling classifiers for each individual emotion. By applying this process to the selected examples, we construct a multi-label dataset. Using this dataset, we train models and evaluate their performance within a traditional supervised setting. Our model achieves an F1 score of 0.59 on our labeled gold set, showcasing its ability to effectively detect fine-grained emotions. Furthermore, we conduct evaluations of the model's performance in zero- and few-shot transfer scenarios using benchmark datasets. Notably, our results indicate that the knowledge learned from our dataset exhibits transferability across diverse data domains, demonstrating its potential for broader applications beyond emotion detection in literature. Our contribution thus includes a multi-label fine-grained emotion detection dataset built from literature, the semi-supervised approach used to create it, as well as the models trained on it. This work provides a solid foundation for advancing emotion detection techniques and their utilization in various scenarios, especially within the cultural heritage analysis.
admiration: finds something admirable, impressive or worthy of respect
amusement: finds something funny, entertaining or amusing
anger: is angry, furious, or strongly displeased; displays ire, rage, or wrath
annoyance: is annoyed or irritated
approval: expresses a favorable opinion, approves, endorses or agrees with something or someone
boredom: feels bored, uninterested, monotony, tedium
calmness: is calm, serene, free from agitation or disturbance, experiences emotional tranquility
caring: cares about the well-being of someone else, feels sympathy, compassion, affectionate concern towards someone, displays kindness or generosity
courage: feels courage or the ability to do something that frightens one, displays fearlessness or bravery
curiosity: is interested, curious, or has strong desire to learn something
desire: has a desire or ambition, wants something, wishes for something to happen
despair: feels despair, helpless, powerless, loss or absence of hope, desperation, despondency
disappointment: feels sadness or displeasure caused by the non-fulfillment of hopes or expectations, being or let down, expresses regret due to the unfavorable outcome of a decision
disapproval: expresses an unfavorable opinion, disagrees or disapproves of something or someone
disgust: feels disgust, revulsion, finds something or someone unpleasant, offensive or hateful
doubt: has doubt or is uncertain about something, bewildered, confused, or shows lack of understanding
embarrassment: feels embarrassed, awkward, self-conscious, shame, or humiliation
envy: is covetous, feels envy or jealousy; begrudges or resents someone for their achievements, possessions, or qualities
excitement: feels excitement or great enthusiasm and eagerness
faith: expresses religious faith, has a strong belief in the doctrines of a religion, or trust in god
fear: is afraid or scared due to a threat, danger, or harm
frustration: feels frustrated: upset or annoyed because of inability to change or achieve something
gratitude: is thankful or grateful for something
greed: is greedy, rapacious, avaricious, or has selfish desire to acquire or possess more than what one needs
grief: feels grief or intense sorrow, or grieves for someone who has died
guilt: feels guilt, remorse, or regret to have committed wrong or failed in an obligation
indifference: is uncaring, unsympathetic, uncharitable, or callous, shows indifference, lack of concern, coldness towards someone
joy: is happy, feels joy, great pleasure, elation, satisfaction, contentment, or delight
love: feels love, strong affection, passion, or deep romantic attachment for someone
nervousness: feels nervous, anxious, worried, uneasy, apprehensive, stressed, troubled or tense
nostalgia: feels nostalgia, longing or wistful affection for the past, something lost, or for a period in one's life, feels homesickness, a longing for one's home, city, or country while being away; longing for a familiar place
optimism: feels optimism or hope, is hopeful or confident about the future, that something good may happen, or the success of something - pain: feels physical pain or is experiences physical suffering
pride: is proud, feels pride from one's own achievements, self-fulfillment, or from the achievements of those with whom one is closely associated, or from qualities or possessions that are widely admired
relief: feels relaxed, relief from tension or anxiety
sadness: feels sadness, sorrow, unhappiness, depression, dejection
surprise: is surprised, astonished or shocked by something unexpected
trust: trusts or has confidence in someone, or believes that someone is good, honest, or reliable
The Global Emotion Detection and Recognition Market size was valued at USD 26.15 Billion in 2022 and is expected to surpass USD 140.07 Billion by 2031, expanding at a CAGR of 20.5% during the forecast period, 2023 – 2031. Emotion detection and recognition involves a process of deployment of advanced technology and software that are primarily designed to read or detect human emotions. It usually evaluates emotions through using image processing technology that examines facial expression. Emotions have a significant impact on how we behave and think and understanding emotion of a consumer or patient is a vital aspect for various business industries.
Social communication is an important part of our lives and emotional expressions help us comprehend the mental state of people around us, allowing us to communicate more effectively. Emotion detection and recognition research has gained traction as a result of its wide applications in perceptual and cognitive sciences as well as affective computing and computer animation. This technology is extensively used by businesses to detect consumer opinions that may have an impact on their purchasing and consumption behavior, which have a major impact on the level of production. The law enforcement sector uses several technique for human monitoring and tracking using a lie detector test.
Increasing awareness regarding benefits of emotion recognition systems in the surveillance sector and its wide applications across several industries propel market growth.
Rising introduction of innovative technology in wearable devices and expansion of end-users of the technology for various emotional detection applications are expected to fuel the market expansion.
Growing demand for enhanced client experiences as well as the requirement for a human touch in digital interactions are key factors driving market growth.
High cost of the device systems and issue related to misinterpretation of emotional analysis are major hindrance to the growth of the market.
<spa
Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
video, audio and text collection
Attribution-ShareAlike 4.0 (CC BY-SA 4.0)https://creativecommons.org/licenses/by-sa/4.0/
License information was derived automatically
This dataset advances the "Emotions dataset for NLP" by Praveen (https://www.kaggle.com/datasets/praveengovi/emotions-dataset-for-nlp/data).
This dataset contains an additional neutral emotion label, which was obtained from the "Emotion Detection from Text" by Pashupati Gupta (https://www.kaggle.com/datasets/pashupatigupta/emotion-detection-from-text).
Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
The most popular news portal's Facebook pages such as Prothom Alo, BBC Bangla, BD News 24, Bangla Tribune, Kaler Kantho, Daily Jugantor are picked to build the dataset. Following a manual collection of posts, a total of 130 posts for 11 news topics were obtained and converted into a CSV file. The dataset is annotated in Ekman's seven universal emotions and they are collected using a self-developed scraper algorithm.
https://www.verifiedmarketresearch.com/privacy-policy/https://www.verifiedmarketresearch.com/privacy-policy/
Emotion AI Market size was valued at USD 5.23 Billion in 2024 and is projected to reach USD 28.63 Billion by 2032, growing at a CAGR of 22.5% from 2025 to 2032.
The Emotion AI market is driven by the rising demand for enhanced customer experience across industries like retail, healthcare, and automotive. Businesses are leveraging AI-driven emotion recognition to personalize interactions, improve engagement, and boost sales. The growing adoption of AI in sentiment analysis, chatbots, and virtual assistants further fuels market expansion.
Advancements in deep learning, facial recognition, and natural language processing (NLP) are accelerating the development of Emotion AI. Increasing investments in AI research and the integration of emotion recognition in smart devices, security systems, and entertainment platforms are also key growth drivers.
BanglaEmotion is a manually annotated Bangla Emotion corpus, which incorporates the diversity of fine-grained emotion expressions in social-media text. More fine-grained emotion labels are considered such as Sadness, Happiness, Disgust, Surprise, Fear and Anger - which are, according to Paul Ekman (1999), the six basic emotion categories. For this task, a large amount of raw text data are collected from the user’s comments on two different Facebook groups (Ekattor TV and Airport Magistrates) and from the public post of a popular blogger and activist Dr. Imran H Sarker. These comments are mostly reactions to ongoing socio-political issues and towards the economic success and failure of Bangladesh. A total of 32923 comments are scraped from the three sources aforementioned above. Out of these, a total of 6314 comments were annotated into the six categories. The distribution of the annotated corpus is as follows:
sad = 1341 happy = 1908 disgust = 703 surprise = 562 fear = 384 angry = 1416
A balanced set is also provided from the above data and split the dataset into training and test set of equal ratio. A proportion of 5:1 is used for training and evaluation purposes. More information on the dataset and the experiments on it could be found in our paper (related links below).
https://www.marketreportanalytics.com/privacy-policyhttps://www.marketreportanalytics.com/privacy-policy
The Emotion Recognition and Sentiment Analysis Software Market is experiencing robust growth, projected to reach $849.76 million in 2025 and maintain a Compound Annual Growth Rate (CAGR) of 14.15% from 2025 to 2033. This expansion is fueled by several key drivers. Increasing adoption of AI-powered solutions across diverse sectors, including customer service, market research, and healthcare (patient diagnosis), is a primary factor. Businesses leverage these tools to gain valuable insights into customer preferences, improve product development, and personalize user experiences. The rise of cloud-based deployment models further accelerates market growth, offering scalability, cost-effectiveness, and enhanced accessibility. Furthermore, the growing need for effective brand monitoring and reputation management, particularly on social media, is driving demand for sentiment analysis tools. While data privacy concerns and ethical considerations surrounding emotion recognition technology pose certain restraints, the overall market outlook remains exceptionally positive. The market is segmented by application (customer service/experience, product/market research, patient diagnosis, others) and deployment (on-premises, cloud-based), reflecting the diverse use cases and deployment preferences of different industries. North America currently holds a significant market share, driven by early adoption and technological advancements. However, APAC is expected to exhibit substantial growth in the coming years, fueled by increasing digitalization and a burgeoning tech industry in countries like China and Japan. Leading companies are focusing on strategic partnerships, acquisitions, and the development of innovative solutions to maintain a competitive edge in this rapidly evolving landscape. The competitive landscape is characterized by a mix of established tech giants like Microsoft and IBM alongside specialized emotion AI companies. The market’s success hinges on the continuous improvement of algorithm accuracy, addressing ethical concerns, and ensuring responsible data handling. Future growth will depend on advancements in deep learning and computer vision, enabling more nuanced and accurate emotion recognition across various modalities, including facial expressions, voice tone, and text analysis. Addressing data bias and ensuring compliance with data privacy regulations are crucial for sustainable growth. The market's segmentation reflects its adaptability across various industries, underscoring its potential for widespread application and sustained expansion throughout the forecast period.
Attribution-NonCommercial-ShareAlike 4.0 (CC BY-NC-SA 4.0)https://creativecommons.org/licenses/by-nc-sa/4.0/
License information was derived automatically
Contact Information
If you would like further information about the RAVDESS Facial Landmark Tracking data set, or if you experience any issues downloading files, please contact us at ravdess@gmail.com.
Tracking Examples
Watch a sample of the facial tracking results.
Description
The RAVDESS Facial Landmark Tracking dataset set contains tracked facial landmark movements from the Ryerson Audio-Visual Database of Emotional Speech and Song (RAVDESS) [RAVDESS Zenodo page]. Motion tracking of actors' faces was produced by OpenFace 2.1.0 (Baltrusaitis, T., Zadeh, A., Lim, Y. C., & Morency, L. P., 2018). Tracked information includes: facial landmark detection, head pose estimation, facial action unit recognition, and eye-gaze estimation.
The Facial Landmark Tracking dataset was created in the Affective Data Science Lab.
This data set contains tracking for all 2452 RAVDESS trials. All tracking movement data are contained in "FacialTracking_Actors_01-24.zip", which contains 2452 .CSV files. Each actor has 104 tracked trials (60 speech, 44 song). Note, there are no song files for Actor 18.
Total Tracked Files = (24 Actors x 60 Speech trials) + (23 Actors x 44 Song trials) = 2452 files.
Tracking results for each trial are provided as individual comma separated value files (CSV format). File naming convention of tracked files is identical to that of the RAVDESS. For example, tracked file "01-01-01-01-01-01-01.csv" corresponds to RAVDESS audio-video file "01-01-01-01-01-01-01.mp4". For a complete description of the RAVDESS file naming convention and experimental manipulations, please see the RAVDESS Zenodo page.
Tracking overlay videos for all trials are also provided (720p Xvid, .avi), one zip file per Actor. As the RAVDESS does not contain "ground truth" facial landmark locations, the overlay videos provide a visual 'sanity check' for researchers to confirm the general accuracy of the tracking results. The file naming convention of tracking overlay videos also matches that of the RAVDESS. For example, tracking video "01-01-01-01-01-01-01.avi" corresponds to RAVDESS audio-video file "01-01-01-01-01-01-01.mp4".
Tracking File Output Format
This data set retained OpenFace's data output format, described here in detail. The resolution of all input videos was 1280x720. When tracking output units are in pixels, their range of values is (0,0) (top left corner) to (1280,720) (bottom right corner).
Columns 1-3 = Timing and Detection Confidence
Columns 4-291 = Eye Gaze Detection
Columns 292-297 = Head pose
Columns 298-433 = Facial Landmarks locations in 2D
Columns 434-637 = Facial Landmarks locations in 3D
Columns 638-677 = Rigid and non-rigid shape parameters
Parameters of a point distribution model (PDM) that describe the rigid face shape (location, scale and rotation) and non-rigid face shape (deformation due to expression and identity). For more details, please refer to chapter 4.2 of my Tadas Baltrusaitis's PhD thesis [download link].
Columns 687-712 = Facial Action Units
Facial Action Units (AUs) are a way to describe human facial movements (Ekman, Friesen, and Hager, 2002) [wiki link]. More information on OpenFace's implementation of AUs can be found here.
Note, OpenFace's columns 2 and 5 (face_id and success, respectively) were not included in this data set. These values were redundant as a single face was detected in all frames, in all 2452 trials.
Tracking Overlay Videos
Tracking overlay videos visualize most aspects of the tracking output described above.
Camera Parameters and 3D Calibration Procedure
This data set contains accurate estimates of actors' 3D head poses. To produce these, camera parameters at the time of recording were required (distance from camera to actor, and camera field of view). These values were used with OpenCV's camera calibration procedure, described here, to produce estimates of the camera's focal length and optical center at the time of actor recordings. The four values produced by the calibration procedure (fx,fy,cx,cy) were input to OpenFace as command line arguments during facial tracking, described here, to produce accurate estimates of 3D head pose.
Camera
MIT Licensehttps://opensource.org/licenses/MIT
License information was derived automatically
Text-based afective computing
We collected a dataset of tweets primarily written in Guarani (and Jopara, a code-switching language that combines Guarani and Spanish) and annotated them for three widely-used dimensions in sentiment analysis:
emotion recognition (https://huggingface.co/datasets/mmaguero/gn-emotion-recognition),
humor detection (this repo, https://huggingface.co/datasets/mmaguero/gn-humor-detection), and
offensive language identification… See the full description on the dataset page: https://huggingface.co/datasets/mmaguero/gn-humor-detection.
https://www.marketresearchintellect.com/privacy-policyhttps://www.marketresearchintellect.com/privacy-policy
The market size of the Emotion Recognition And Sentiment Analysis Market is categorized based on Type (Customer Feedback, Social Media Monitoring, Market Research, Product Development) and Application (Sentiment Analysis Software, Emotion Detection Tools, Text Analysis, Voice Sentiment Analysis, Image Sentiment Analysis) and geographical regions (North America, Europe, Asia-Pacific, South America, and Middle-East and Africa).
This report provides insights into the market size and forecasts the value of the market, expressed in USD million, across these defined segments.
Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
Hate speech detection and emotion recognition are emerging research fields for Natural Language Processing (NLP). The text that expresses hate speech also convey some emotion. Hence, hate speech and emotion recognition are related research areas. Therefore, attention is necessary to combine both fields in the same research. With this motivation, a multilevel dataset for natural Bangla text is proposed for Bangla hate speech and emotion recognition. The dataset contains 16407 statements collected from natural comments and categorized into two labels. Firstly, it is categorized as either hate or non-hate. Then, each category is further divided into three emotions: happy, sad, and angry. The dataset is validated with Cohen’s Kappa and Fliess’ Kappa measures that show good scores. The dataset can be applicable to machine learning and natural language processing tasks for Bangla text.
Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
A multilingual fine-grained emotion dataset. The dataset consists of human annotated Finnish (25k) and English sentences (30k). Plutchik’s core emotions are used to annotate the dataset with the addition of neutral to create a multilabel multiclass dataset. The dataset is carefully evaluated using language-specific BERT models and SVMs to show that XED performs on par with other similar datasets and is therefore a useful tool for sentiment analysis and emotion detection.
Attribution-NonCommercial 4.0 (CC BY-NC 4.0)https://creativecommons.org/licenses/by-nc/4.0/
License information was derived automatically
Emotional Video Data,including multiple races, multiple indoor scenes, multiple age groups, multiple languages, multiple emotions (11 types of facial emotions, 15 types of inner emotions). For each sentence in each video, annotated emotion types (including facial emotions and inner emotions), start & end timestamp, text transcription.This dataset can be used for tasks such as emotion recognition and sentiment analysis, enhancing model performance in real and complex tasks.rnQuality tested by various AI companies. We strictly adhere to data protection regulations and privacy standards, ensuring the maintenance of user privacy and legal rights throughout the data collection, storage, and usage processes, our datasets are all GDPR, CCPA, PIPL complied. For more details, please refer to the link: https://www.nexdata.ai/datasets/speechrecog/977?source=Kaggle
The video data format is .mp4, the annotation file format is .json;
Including multiple races, multiple indoor scenes, multiple age groups, multiple languages, multiple emotions (11 types of facial emotions, 15 types of inner emotions);
Indoor;
Camera or smartphone;
1,003 people, each person has one or several videos with multiple emotions; Race distribution: 232 people of Asian, 614 people of Caucasian, 157 people of black race; gender distribution: 410 people of male, 593 people of female; age distribution: 569 young people, 330 middle-aged people and 104 old-aged people;
For each sentence in each video, annotated emotion types (including facial emotions and inner emotions), start & end timestamp, text transcription;
Collecting accuracy: according to the 'collecting requirement', the collecting accuracy is over 97%; Label annotation accuracy: the accuracy of language, race, gender and age group labeling is over 97%; file annotation accuracy: the word accuracy rate of text transcription is over 85%;
Commercial License
https://www.gnu.org/licenses/gpl-3.0-standalone.htmlhttps://www.gnu.org/licenses/gpl-3.0-standalone.html
This repository contains data and code for emotion detection in text, developed in the EASIER project on sign language translation. The end goal within the project is to provide emotional cues to an avatar for the translation direction from spoken to sign language. While the EASIER project is multilingual, emotion detection so far is restricted to German.