✂️🕷️ Spider-Cut is a Network Mapper Framework (NMAP Framework)

Overview

Build with Python Latest Release Last Comitment Python Version


Spider-Cut is a Network Mapper Framework (NMAP Framework)

Installation    |    Usage    |    Creators    |    Donate

Installation

# Kali Linux | WSL

# clone the repo
$ git clone https://github.com/XFORWORKS/SpiderCut

# change the working directory to SpiderCut
$ cd SpiderCut
  
# install the requirements  
$ python -m pip install -r requirements.txt

# run the bigger installation process
$ python setup.py

Usage

# Kali Linux | WSL

$ python spidercut.py --help
usage: spidercut.py [-h] [-v] [-s] [-m] [-rH] [-sn] [-rh] [-eX]
                              [-ex] [-a]

Spider-Cut is a Network Mapper Framework (NMAP Framework)

optional arguments:
  -h, --help       show this help message and exit
  -v, --version    shows the current Version of Spider-Cut
  -s, --single     scan a single Target
  -m, --multiple   scan multiple Targets
  -rH, --rhost     scan a range of Hosts
  -sn, --subnet    scan an entire Subnet
  -rh, --ranhost   scan random Hosts
  -eX, --exclude   Excluding Targets from a scan
  -ex, --excllist  Excluding Targets using a list
  -a, --agscan     perform an Aggressive scan

Created By XFORWORKS (Retr0 & Ctript0)

To run the script:

# Kali Linux | WSL

$ python spidercut.py

Creators

The Script is designed by Cript0 (like the banners, the colors, etc...)

Everything other is from Retr0 (like the commands, the framework idea, etc...)

Donate

PayPal

Releases

Latest Release

You might also like...
Distributed Crawler Management Framework Based on Scrapy, Scrapyd, Django and Vue.js

Gerapy Distributed Crawler Management Framework Based on Scrapy, Scrapyd, Scrapyd-Client, Scrapyd-API, Django and Vue.js. Documentation Documentation

Async Python 3.6+ web scraping micro-framework based on asyncio
Async Python 3.6+ web scraping micro-framework based on asyncio

Ruia 🕸️ Async Python 3.6+ web scraping micro-framework based on asyncio. ⚡ Write less, run faster. Overview Ruia is an async web scraping micro-frame

A high-level distributed crawling framework.

Cola: high-level distributed crawling framework Overview Cola is a high-level distributed crawling framework, used to crawl pages and extract structur

Transistor, a Python web scraping framework for intelligent use cases.
Transistor, a Python web scraping framework for intelligent use cases.

Web data collection and storage for intelligent use cases. transistor About The web is full of data. Transistor is a web scraping framework for collec

Web crawling framework  based on asyncio.
Web crawling framework based on asyncio.

Web crawling framework for everyone. Written with asyncio, uvloop and aiohttp. Requirements Python3.5+ Installation pip install gain pip install uvloo

PyQuery-based scraping micro-framework.

demiurge PyQuery-based scraping micro-framework. Supports Python 2.x and 3.x. Documentation: http://demiurge.readthedocs.org Installing demiurge $ pip

Crawler do site Fundamentus.com com o uso do framework scrapy, tanto da aba detalhada como a de resumo.

Crawler do site Fundamentus.com com o uso do framework scrapy, tanto da aba detalhada como a de resumo. (Todas as infomações)

A simple django-rest-framework api using web scraping

Apicell You can use this api to search in google, bing, pypi and subscene and get results Method : POST Parameter : query Example import request url =

Python framework to scrape Pastebin pastes and analyze them
Python framework to scrape Pastebin pastes and analyze them

pastepwn - Paste-Scraping Python Framework Pastebin is a very helpful tool to store or rather share ascii encoded data online. In the world of OSINT,

Releases(1.0)
  • 1.0(Dec 25, 2021)

    Build with Python Last Comitment Python Version


    Spider-Cut is a Network Mapper Framework (NMAP Framework)

    Installation

    # Kali Linux | WSL
    
    # clone the repo
    $ git clone https://github.com/XFORWORKS/SpiderCut
    
    # change the working directory to SpiderCut
    $ cd SpiderCut
      
    # install the requirements  
    $ python -m pip install -r requirements.txt
    
    # run the bigger installation process
    $ python setup.py
    

    Usage

    # Kali Linux | WSL
    
    $ python spidercut.py --help
    usage: spidercut.py [-h] [-v] [-s] [-m] [-rH] [-sn] [-rh] [-eX]
                                  [-ex] [-a]
    
    Spider-Cut is a Network Mapper Framework (NMAP Framework)
    
    optional arguments:
      -h, --help       show this help message and exit
      -v, --version    shows the current Version of Spider-Cut
      -s, --single     scan a single Target
      -m, --multiple   scan multiple Targets
      -rH, --rhost     scan a range of Hosts
      -sn, --subnet    scan an entire Subnet
      -rh, --ranhost   scan random Hosts
      -eX, --exclude   Excluding Targets from a scan
      -ex, --excllist  Excluding Targets using a list
      -a, --agscan     perform an Aggressive scan
    
    Created By XFORWORKS (Retr0 & Cript0)
    

    To run the script:

    # Kali Linux | WSL
    
    $ python spidercut.py
    

    Creators

    The Script is designed by Cript0 (like the banners, the colors, etc...)

    Everything other is from Retr0 (like the commands, the framework idea, etc...)

    Donate

    PayPal

    Source code(tar.gz)
    Source code(zip)
    SpiderCut.rar(171.16 KB)
    SpiderCut.zip(226.92 KB)
Owner
XforWorks
Nothing to see here...
XforWorks
Jobinja.ir jobs scraper.

Jobinja.ir Dataset Introduction This project is a simple web scraper that scraps pages of jobinja.ir concurrently and writes and update (if file gets

Iman Kermani 3 Apr 15, 2022
The core packages of security analyzer web crawler

Security Analyzer 🐍 A large scale web crawler (considered also as vulnerability scanner tool) to take an overview about security of Moroccan sites Cu

Security Analyzer 10 Jul 03, 2022
Scrapy-soccer-games - Scraping information about soccer games from a few websites

scrapy-soccer-games Esse projeto tem por finalidade pegar informação de tabela d

Caio Alves 2 Jul 20, 2022
A Python Covid-19 cases tracker that scrapes data off the web and presents the number of Cases, Recovered Cases, and Deaths that occurred because of the pandemic.

A Python Covid-19 cases tracker that scrapes data off the web and presents the number of Cases, Recovered Cases, and Deaths that occurred because of the pandemic.

Alex Papadopoulos 1 Nov 13, 2021
Danbooru scraper with python

Danbooru Version: 0.0.1 License under: MIT License Dependencies Python: = 3.9.7 beautifulsoup4 cloudscraper Example of use Danbooru from danbooru imp

Sugarbell 2 Oct 27, 2022
用python爬取江苏几大高校的就业网站,并提供3种方式通知给用户,分别是通过微信发送、命令行直接输出、windows气泡通知。

crawler_for_university 用python爬取江苏几大高校的就业网站,并提供3种方式通知给用户,分别是通过微信发送、命令行直接输出、windows气泡通知。 环境依赖 wxpy,requests,bs4等库 功能描述 该项目基于python,通过爬虫爬各高校的就业信息网,爬取招聘信

8 Aug 16, 2021
Parse feeds in Python

feedparser - Parse Atom and RSS feeds in Python. Copyright 2010-2020 Kurt McKee Kurt McKee 1.5k Dec 30, 2022

PS5 bot to find a console in france for chrismas 🎄🎅🏻 NOT FOR SCALPERS

Une PS5 pour Noël Python + Chrome --headless = une PS5 pour noël MacOS Installer chrome Tweaker le .yaml pour la listes sites a scrap et les criteres

Olivier Giniaux 3 Feb 13, 2022
Scraping script for stats on covid19 pandemic status in Chiba prefecture, Japan

About 千葉県の地域別の詳細感染者統計(Excelファイル) をCSVに変換し、かつ地域別の日時感染者集計値を出力するスクリプトです。 Requirement POSIX互換なシェル, e.g. GNU Bash (1) curl (1) python = 3.8 pandas = 1.1.

Conv4Japan 1 Nov 29, 2021
Minimal set of tools to conduct stealthy scraping.

Stealthy Scraping Tools Do not use puppeteer and playwright for scraping. Explanation. We only use the CDP to obtain the page source and to get the ab

Nikolai Tschacher 88 Jan 04, 2023
A scrapy pipeline that provides an easy way to store files and images using various folder structures.

scrapy-folder-tree This is a scrapy pipeline that provides an easy way to store files and images using various folder structures. Supported folder str

Panagiotis Simakis 7 Oct 23, 2022
哔哩哔哩爬取器:以个人为中心

Open Bilibili Crawer 哔哩哔哩是一个信息非常丰富的社交平台,我们基于此构造社交网络。在该网络中,节点包括用户(up主),以及视频、专栏等创作产物;关系包括:用户之间,包括关注关系(following/follower),回复关系(评论区),转发关系(对视频or动态转发);用户对创

Boshen Shi 3 Oct 21, 2021
A tool for scraping and organizing data from NewsBank API searches

nbscraper Overview This simple tool automates the process of copying, pasting, and organizing data from NewsBank API searches. Curerntly, nbscrape onl

0 Jun 17, 2021
Automatically download and crop key information from the arxiv daily paper.

Arxiv daily 速览 功能:按关键词筛选arxiv每日最新paper,自动获取摘要,自动截取文中表格和图片。 1 测试环境 Ubuntu 16+ Python3.7 torch 1.9 Colab GPU 2 使用演示 首先下载权重baiduyun 提取码:il87,放置于code/Pars

HeoLis 20 Jul 30, 2022
Complete pipeline for crawling online newspaper article.

Complete pipeline for crawling online newspaper article. The articles are stored to MongoDB. The whole pipeline is dockerized, thus the user does not need to worry about dependencies. Additionally, d

newspipe 4 May 27, 2022
Web crawling framework based on asyncio.

Web crawling framework for everyone. Written with asyncio, uvloop and aiohttp. Requirements Python3.5+ Installation pip install gain pip install uvloo

Jiuli Gao 2k Jan 05, 2023
A python script to extract answers to any question on Quora (Quora+ included)

quora-plus-bypass A python script to extract answers to any question on Quora (Quora+ included) Requirements Python 3.x

Nitin Narayanan 10 Aug 18, 2022
High available distributed ip proxy pool, powerd by Scrapy and Redis

高可用IP代理池 README | 中文文档 本项目所采集的IP资源都来自互联网,愿景是为大型爬虫项目提供一个高可用低延迟的高匿IP代理池。 项目亮点 代理来源丰富 代理抓取提取精准 代理校验严格合理 监控完备,鲁棒性强 架构灵活,便于扩展 各个组件分布式部署 快速开始 注意,代码请在release

SpiderClub 5.2k Jan 03, 2023
OSTA web scraper, for checking the status of school buses in Ottawa

OSTA-La-Vista OSTA web scraper, for checking the status of school buses in Ottawa. Getting Started Using a Raspberry Pi, download Python 3, and option

1 Jan 28, 2022
This tool crawls a list of websites and download all PDF and office documents

This tool crawls a list of websites and download all PDF and office documents. Then it analyses the PDF documents and tries to detect accessibility issues.

AccessibilityLU 7 Sep 30, 2022