Skip to content

judemont/netexplorer

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

23 Commits
 
 
 
 
 
 
 
 

Repository files navigation

NetExplorer

Web analysis tool


Installation :

pip install netexplorer

Examples :

from netexplorer import Crawler, SitesMap

crawler = Crawler("https://futureofthe.tech", maxSites=300, maxDepth=2, threads=5)
crawler.start()

sitesMap = SitesMap()
sitesMap.show()

Example result :

Screenshot From 2025-01-22 09-31-41

Documentation :

Crawler :

A web crawler class to recursively explore and store web pages.

Parameters:

  • entryUrl (str): The initial URL to start crawling from.
  • maxSites (int|None): The maximum number of sites to crawl. If None, there is no limit.
  • maxDepth (int|None): The maximum depth to crawl. If None, there is no limit.
  • threads (int): The maximum number of threads to use for crawling.
  • headers (dict): Headers to use for HTTP requests.
  • dbPath (str): Path to the SQLite database file.
  • resume (bool): Whether to resume from the existing database or start fresh.

Methods:

  • start(): Starts the crawling process.

SitesMap :

A class to visualize the map of sites and their links using a graph.

Attributes:

  • dbPath (str): Path to the SQLite database file.
  • fixedPointsSize (bool): Whether to use a fixed size for points (nodes) in the graph.
  • pointsSize (int): The size of the points (nodes) in the graph.
  • edgesWidth (float): The width of the edges (links) in the graph.
  • layout (str): The layout to use for the graph visualization.

Methods:

  • show(): Displays the graph of sites and their links.