I switched to a new method of installing this program for a better developer experience and better modifiability.
As a result, the PYPI package scan-for-webcams
is not maintained anymore, and
should be considered deprecated.
btw: we have a discord channel ! join
- Note
- Table of contents
- Usage
- Installation
- Experimental Vision-Language Model Support
- Places365: On device footage classification
- Graphical User Interface
- Demo
-
python sfw search MJPG
: for public MJPG streamers -
python sfw search webcamXP
: for public webcamXP streamers -
python sfw search yawCam
: for public yawCam steamers -
python sfw search hipcam
: for public hipcam streamers -
python sfw search rtsp
: DANGER searches for rtsp servers on shodan, and performs enumeration on them to try and find streams -
python sfw play {url}
: plays the camera stream- for
rtsp://
streams, sfw will play it in a GUI viewer (seesfw/rtsp.py
) - for all other streams, sfw will open it in the default browser.
- for
-
python sfw search ... --gui
: displays the scanned cameras in a grid visually. -
python sfw search --help
: for more options and help
The program will output a list of links with the format of ip_address:port
, and descriptions of the image beneath it.
If your terminal supports links, click the link and open it in your browser, otherwise, copy the link and open it in your browser.
-
Clone this repository:
git clone https://github.com/JettChenT/scan-for-webcams
-
install requirements.txt:
pip install -r requirements.txt
-
set up shodan: go to shodan.io, register/log in and grab your API key
-
set up clarifai: go to clarifai.com, register/log in, create an application and grab your API key. Alternatively, use the local places365 model.
-
setup geoip: go to geo.ipify.org, register/log in and grab your API key
-
Add API keys:
- run
python sfw setup
- enter your shodan, clarifai and geoip API keys
- run
And then you can run the program!
We are working on supporting generating descriptions for footages
via vision language models, which can generate descriptions for images in
natural language, instead of just tags, and can run on-device.
Currently, we support a llama.cpp
based llava 7b
model.
To use this, run the following commands:
pip install -U --no-cache-dir llama-cpp-python@git+https://github.com/damian0815/llama-cpp-python/@4ec35390d72faba70942b9605dfcbde2bda0bdad
pip install huggingface_hub wurlitzer
python sfw search webcamXP --vllm=True --parallel=False # or any other search command with the latter two flags
Note: our current VLLM integration is experimental, and may not work as expected. Parallel mode must be disabled for VLLM as of now.
It is now possible to run the Places365 model, a model fined tuned for real world locations, to get information about webcam footage.
To use this model, you need to install the following packages:
pip install -r requirements-places.txt
Then, you can run the program with the --places
flag:
python sfw search MJPG --tag=False --places
The model will take some time to download, and will be cached for future use.
To get a display of the live streams of all the scanned webcams, simply append the --gui
option to your command.