Skip to content

One of the projects used in the second edition of Ten Days of Code by GNU/Linux Users' Group, NIT Durgapur, It is a search engine with functionalities of crawling the web pages and storing them. It makes use of MongoDB to store the data found in the pages. It also has an web interface to search the crawled pages.

License

Notifications You must be signed in to change notification settings

mayankt18/glugle

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

91 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

forthebadge

Quick Links

Search engine made using Python and flask

Project Setup

  1. Clone the git in a folder
  2. Make a virtual environment in the folder

For Windows

pip install virtualenv
cd project_folder
virtualenv env
.\env\Scripts\activate

For Linux

pip install virtualenv #for version 2 and below 
pip3 install virtualenv #for version 3
cd project file
virtualenv env
source ./env/bin/activate
  1. Install requirements
pip3 install -r requirements.txt
  1. Start the crawler
python ./crawler/spider.py #for version 2 and below
python3 ./crawler/spider.py #for version 3
  1. Start the application
python ./app/app.py #for version 2 and below
python3 ./app/app.py #for version 3

About

One of the projects used in the second edition of Ten Days of Code by GNU/Linux Users' Group, NIT Durgapur, It is a search engine with functionalities of crawling the web pages and storing them. It makes use of MongoDB to store the data found in the pages. It also has an web interface to search the crawled pages.

Topics

Resources

License

Stars

Watchers

Forks