Skip to content

Sentimental Analysis from text: A text analysis algorithm based on ETL approach with multiprocessing for parallel processing of multiple text files.

Notifications You must be signed in to change notification settings

suwesh/Multi-core_SentimentEngine

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

13 Commits
 
 
 
 
 
 
 
 

Repository files navigation

Multi-core SentimentEngine

Sentimental Analysis from text: A rule-based text analysis algorithm based on ETL approach with multiprocessing for parallel processing of multiple text files. This algorithm is able to utilize multiple cores of a processor, and is designed to run on a server as an engine and not a end to end web application. The algorithm takes URLs and id as input and calculates the positive, negative, ... scores from the article text and updates the values to output destination, data from here can be used for other applications.

This document provides insights and guidance on the Sentimental analysis algorithm “Multi-core SentimentEngine”. The data extraction and NLP problem is tasked with solving ways to design and implement a sentimental analysis algorithm, which analyzes textual data from web and assess its sentiments, positive, negative, or neutral, along with determining more variables such as subjectivity score, readability. Problem approach: The problem is solved using an extract, transform, load approach with multiprocessing for parallel processing of text files.

  1. Data Extraction: • The function createdf() creates a dataframe containing URL and URL_ID from the Input.xlsx file. • The function extractdf() is a function that leverages requests liberary and a extract_urltext() function to fetch HTML content, then uses BeautifulSoup for efficient parsing of content. The title is extracted from <title> tag and article text is extracted using class td-pb-span8 td-main-content. • The extractdf() function constructs filenames with unique URL_ID and uses “with open” to write the extracted text to respective files in UTF-8 encoding.
  2. Data Transformation: • The read_stopwords() function reads stop words from multiple files and returns a set of stop words. The clean_stopwords() uses NLTK’s word_tokenize function to tokenizes text and returns only those words not in stop words. The clean_stopwords_directory() applies clean_stopwords() function to all text files in the directory rewriting the files. • Create_posneg_dict() is a function that reads the provided files in MasterDirectiry and returns two sets, one with positive words and other with negative words.
  3. Calculate variables: • The runengine() function runs a loop for each file the directory provided and tokenizes the texts into words (word_tokenize) and sentences (sentence_tokenize) and calculating each variable by passing these as parameters to separate functions made incorporated with the formulae provided with the problem. • The appendtodf() function appends a new columns with calculated variables for each row where URL ID matches to a dataframe. • Finally the result dataframe for each process running the runengine() function appends its respective output dataframes to a list which will be concatenated to one single dataframe.
  4. Data Load: • The leadoutput() function stores the concatenated dataframe as Output.xlsx file to specified path.

The implementation of multiprocessing and error handling: The program uses multiprocessing library to create multiple processes equal to the number of logical processors(cores) found in the executing system. The number of files to process are split to each process almost equally. The processes execute the runengine() function parallelly processing separate files list divided for each process.

  1. Significant Performance Gains: Parallel processing effectively leverages multiple CPU cores, distributing workload and substantially reducing processing time, especially for large datasets.
  2. Shared Data Handling: The use of a managed list to store output dataframes by each process ensures efficient communication and data collection between parallel processes.
  3. Error Handling: Error handling is essential to prevent potential issues during multiprocessing, ensuring the integrity of the final output.

How to run the .py file to generate output(engine.py):

  1. Download the SentimentEngine folder from the drive link and extract from zip file. https://drive.google.com/drive/folders/1z8DWG6xZwhDRujmelOlypP_9nXEWjME3?usp=drive_link
  2. Run the command (in terminal opened in same directory as folderpath i.e SentimentEngine) pip install -r requirements.txt
  3. Change the folder path variable in the engine.py python source file: folderpath = r'C:\Users/suwes/SentimentEngine/' for the running system and execute engine.py. (run: python engine.py in terminal opened in same directory as folderpath). Python should be pre-installed in this directory. Also change the "class" variable of the article content in the extract_urltext() function with the class of respective webpage.
  4. After the program completes its execution, the output excel file named “Output.xlsx” is generated with specified structure format in the same directory.

Required dependencies(python liberaries): • multiprocessing • os • pandas • requests • bs4 and BeautifulSoup • string • nltk • time

About

Sentimental Analysis from text: A text analysis algorithm based on ETL approach with multiprocessing for parallel processing of multiple text files.

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages