✂️🕷️ Spider-Cut is a Network Mapper Framework (NMAP Framework)

Overview

Build with Python Latest Release Last Comitment Python Version


Spider-Cut is a Network Mapper Framework (NMAP Framework)

Installation    |    Usage    |    Creators    |    Donate

Installation

# Kali Linux | WSL

# clone the repo
$ git clone https://github.com/XFORWORKS/SpiderCut

# change the working directory to SpiderCut
$ cd SpiderCut
  
# install the requirements  
$ python -m pip install -r requirements.txt

# run the bigger installation process
$ python setup.py

Usage

# Kali Linux | WSL

$ python spidercut.py --help
usage: spidercut.py [-h] [-v] [-s] [-m] [-rH] [-sn] [-rh] [-eX]
                              [-ex] [-a]

Spider-Cut is a Network Mapper Framework (NMAP Framework)

optional arguments:
  -h, --help       show this help message and exit
  -v, --version    shows the current Version of Spider-Cut
  -s, --single     scan a single Target
  -m, --multiple   scan multiple Targets
  -rH, --rhost     scan a range of Hosts
  -sn, --subnet    scan an entire Subnet
  -rh, --ranhost   scan random Hosts
  -eX, --exclude   Excluding Targets from a scan
  -ex, --excllist  Excluding Targets using a list
  -a, --agscan     perform an Aggressive scan

Created By XFORWORKS (Retr0 & Ctript0)

To run the script:

# Kali Linux | WSL

$ python spidercut.py

Creators

The Script is designed by Cript0 (like the banners, the colors, etc...)

Everything other is from Retr0 (like the commands, the framework idea, etc...)

Donate

PayPal

Releases

Latest Release

You might also like...
Distributed Crawler Management Framework Based on Scrapy, Scrapyd, Django and Vue.js

Gerapy Distributed Crawler Management Framework Based on Scrapy, Scrapyd, Scrapyd-Client, Scrapyd-API, Django and Vue.js. Documentation Documentation

Async Python 3.6+ web scraping micro-framework based on asyncio
Async Python 3.6+ web scraping micro-framework based on asyncio

Ruia 🕸️ Async Python 3.6+ web scraping micro-framework based on asyncio. ⚡ Write less, run faster. Overview Ruia is an async web scraping micro-frame

A high-level distributed crawling framework.

Cola: high-level distributed crawling framework Overview Cola is a high-level distributed crawling framework, used to crawl pages and extract structur

Transistor, a Python web scraping framework for intelligent use cases.
Transistor, a Python web scraping framework for intelligent use cases.

Web data collection and storage for intelligent use cases. transistor About The web is full of data. Transistor is a web scraping framework for collec

Web crawling framework  based on asyncio.
Web crawling framework based on asyncio.

Web crawling framework for everyone. Written with asyncio, uvloop and aiohttp. Requirements Python3.5+ Installation pip install gain pip install uvloo

PyQuery-based scraping micro-framework.

demiurge PyQuery-based scraping micro-framework. Supports Python 2.x and 3.x. Documentation: http://demiurge.readthedocs.org Installing demiurge $ pip

Crawler do site Fundamentus.com com o uso do framework scrapy, tanto da aba detalhada como a de resumo.

Crawler do site Fundamentus.com com o uso do framework scrapy, tanto da aba detalhada como a de resumo. (Todas as infomações)

A simple django-rest-framework api using web scraping

Apicell You can use this api to search in google, bing, pypi and subscene and get results Method : POST Parameter : query Example import request url =

Python framework to scrape Pastebin pastes and analyze them
Python framework to scrape Pastebin pastes and analyze them

pastepwn - Paste-Scraping Python Framework Pastebin is a very helpful tool to store or rather share ascii encoded data online. In the world of OSINT,

Releases(1.0)
  • 1.0(Dec 25, 2021)

    Build with Python Last Comitment Python Version


    Spider-Cut is a Network Mapper Framework (NMAP Framework)

    Installation

    # Kali Linux | WSL
    
    # clone the repo
    $ git clone https://github.com/XFORWORKS/SpiderCut
    
    # change the working directory to SpiderCut
    $ cd SpiderCut
      
    # install the requirements  
    $ python -m pip install -r requirements.txt
    
    # run the bigger installation process
    $ python setup.py
    

    Usage

    # Kali Linux | WSL
    
    $ python spidercut.py --help
    usage: spidercut.py [-h] [-v] [-s] [-m] [-rH] [-sn] [-rh] [-eX]
                                  [-ex] [-a]
    
    Spider-Cut is a Network Mapper Framework (NMAP Framework)
    
    optional arguments:
      -h, --help       show this help message and exit
      -v, --version    shows the current Version of Spider-Cut
      -s, --single     scan a single Target
      -m, --multiple   scan multiple Targets
      -rH, --rhost     scan a range of Hosts
      -sn, --subnet    scan an entire Subnet
      -rh, --ranhost   scan random Hosts
      -eX, --exclude   Excluding Targets from a scan
      -ex, --excllist  Excluding Targets using a list
      -a, --agscan     perform an Aggressive scan
    
    Created By XFORWORKS (Retr0 & Cript0)
    

    To run the script:

    # Kali Linux | WSL
    
    $ python spidercut.py
    

    Creators

    The Script is designed by Cript0 (like the banners, the colors, etc...)

    Everything other is from Retr0 (like the commands, the framework idea, etc...)

    Donate

    PayPal

    Source code(tar.gz)
    Source code(zip)
    SpiderCut.rar(171.16 KB)
    SpiderCut.zip(226.92 KB)
Owner
XforWorks
Nothing to see here...
XforWorks
An helper library to scrape data from TikTok in one line, using the Influencer Hunters APIs.

TikTok Scraper An utility library to scrape data from TikTok hassle-free Go to the website » View Demo · Report Bug · Request Feature About The Projec

6 Jan 08, 2023
Dex-scrapper - Hobby project for scrapping dex data on VeChain

Folders /zumo_abis # abi extracted from zumo repo /zumo_pools # runtime e

3 Jan 20, 2022
A leetcode scraper to compile all questions in leetcode free tier to text file. pdf also available.

A leetcode scraper to compile all questions in leetcode free tier to text file, pdf also available. if new questions get added, run again to get new questions.

3 Dec 07, 2021
Command line program to download documents from web portals.

command line document download made easy Highlights list available documents in json format or download them filter documents using string matching re

16 Dec 26, 2022
A simple, configurable and expandable combined shop scraper to minimize the costs of ordering several items

combined-shop-scraper A simple, configurable and expandable combined shop scraper to minimize the costs of ordering several items. Features Define an

2 Dec 13, 2021
HappyScrapper - Google news web scrapper with python

HappyScrapper ~ Google news web scrapper INSTALLATION ♦ Clone the repository ♦ O

Jhon Aguiar 0 Nov 07, 2022
Scrape Twitter for Tweets

Backers Thank you to all our backers! 🙏 [Become a backer] Sponsors Support this project by becoming a sponsor. Your logo will show up here with a lin

Ahmet Taspinar 2.2k Jan 05, 2023
Telegram group scraper tool

Telegram Group Scrapper

Wahyusaputra 2 Jan 11, 2022
Scraping followers of an instagram account

ScrapInsta A script to scraping data from Instagram Install First of all you can run: pip install scrapinsta After that you need to install these requ

Matheus Kolln 1 Sep 05, 2021
联通手机营业厅自动做任务、签到、领流量、领积分等。

联通手机营业厅自动完成每日任务,领流量、签到获取积分等,月底流量不发愁。 功能 沃之树领流量、浇水(12M日流量) 每日签到(1积分+翻倍4积分+第七天1G流量日包) 天天抽奖,每天三次免费机会(随机奖励) 游戏中心每日打卡(连续打卡,积分递增至最高

2k May 06, 2021
Kusonime scraper using python3

Features Scrap from url Scrap from recommendation Search by query Todo [+] Search by genre Example # Get download url from kusonime import Scrap

MhankBarBar 2 Jan 28, 2022
This is a simple website crawler which asks for a website link from the user to crawl and find specific data from the given website address.

This is a simple website crawler which asks for a website link from the user to crawl and find specific data from the given website address.

Faisal Ahmed 1 Jan 10, 2022
Scraping Top Repositories for Topics on GitHub,

0.-Webscrapping-using-python Scraping Top Repositories for Topics on GitHub, Web scraping is the process of extracting and parsing data from websites

Dev Aravind D Satprem 2 Mar 18, 2022
Script used to download data for stocks.

This script is useful for downloading stock market data for a wide range of companies specified by their respective tickers. The script reads in the d

Carmelo Gonzales 71 Oct 04, 2022
A Python web scraper to scrape latest posts from official Coinbase's Blog.

Coinbase Blog Scraper A Python web scraper to scrape latest posts from official Coinbase's Blog. IDEA It scrapes up latest blog posts from https://blo

Lucas Villela 3 Feb 18, 2022
中国大学生在线 四史自动答题刷分(现仅支持英雄篇)

中国大学生在线 “四史”学习教育竞答 自动答题 刷分 (现仅支持英雄篇,已更新可用) 若对您有所帮助,记得点个Star 🌟 !!! 中国大学生在线 “四史”学习教育竞答 自动答题 刷分 (现仅支持英雄篇,已更新可用) 🥰 🥰 🥰 依赖 本项目依赖的第三方库: requests 在终端执行以下

XWhite 229 Dec 12, 2022
Grab the changelog from releases on Github

release-notes-scraper This simple script can be used to grab the release notes for projects from github that do not keep a CHANGELOG, but publish thei

Dan Čermák 4 Apr 01, 2022
Web and PDF Scraper Refactoring

Web and PDF Scraper Refactoring This repository contains the example code of the Web and PDF scraper code roast. Here are the links to the videos: Par

18 Dec 31, 2022
👨🏼‍⚖️ reddit bot that turns comment chains into ace attorney scenes

Ace Attorney reddit bot 👨🏼‍⚖️ Reddit bot that turns comment chains into ace attorney scenes. You'll need to sign up for streamable and reddit and se

763 Nov 17, 2022
Pyrics is a tool to scrape lyrics, get rhymes, generate relevant lyrics with rhymes.

Pyrics Pyrics is a tool to scrape lyrics, get rhymes, generate relevant lyrics with rhymes. ./test/run.py provides the full function in terminal cmd

MisterDK 1 Feb 12, 2022