Web Crawling

Libraries that analyze the content of websites.

Newest releases

southwolf Mac Bot A simple crawler sending Telegram notification when Refurbished Macbook Air / Pro in stock. Usage Create a Telegram Bot API account Create a H

s0rg crawley Crawls web pages and prints any link it can find. Scan depth (by default - 0) can be configured. features fast SAX-parser (powered by golang.o

reactor-joy Reactor Crawler Simple CLI content crawler for Joyreactor. He'll find all media content on the page you've provided and save it. If there will be any

amirgamil A Unix-style personal search engine and web crawler for your digital footprint

detectify Fetch web pages using headless Chrome, storing all fetched resources including JavaScript files. Run arbitrary JavaScript on many web pages and see the returned values

Zartenc Collyzar provides a very simple configuration and tools to implement distributed crawling/scraping.

el10savio GoCrawler - A distributed web crawler implemented using Go, Postgres, RabbitMQ and Docker

ianmarmour A simple bot written in go that tracks for availability of stock from nvidia's store and automatically add items to your checkout.

denverquane Discord Bot to scrape Among Us on-screen data, and automatically mute/unmute players during the course of the game!

IAmStoxe A golang utility to spider through a website searching for additional links with support for JavaScript rendering.

rocketlaunchr Quickly scrape Google Search Results.

jaeles-project GoSpider GoSpider - Fast web spider written in Go Installation go get -u github.com/jaeles-project/gospider Features Fast web crawling Brute force and parse sitemap.xml Parse robots.txt

hakluke hakrawler What is it? hakrawler is a Go web crawler designed for easy, quick discovery of endpoints and assets within a web application. It can be used to discover: Forms Endpoints Subdomains Re

alash3al scrapyd-go an drop-in replacement for scrapyd that is more easy to be scalable and distributed on any number of commodity machines with no hassle, each scrapyd-go instance is a stateless microservice, all instances m

geziyor Geziyor Geziyor is a blazing fast web crawling and web scraping framework. It can be used to crawl websites and extract structured data from them. Geziyor is useful for a wide range of purposes such as data mining, moni

alash3al Scraply Scraply a simple dom scraper to fetch information from any html based website using jQuery like syntax and convert that info to JSON APIs How it works? it works by simple define some macros/e

crackcomm About chromedp Package chromedp is a faster, simpler way to drive browsers in Go using the Chrome Debugging Protocol (for Chrome, Edge, Safari, etc) without external dependencies (ie, Selenium, PhantomJS, etc). NOTE:

MontFerret Ferret What is it? ferret is a web scraping system. It aims to simplify data extraction from the web for UI testing, machine learning, analytics and more. ferret allows users to focus o

gocolly Colly Lightning Fast and Elegant Scraping Framework for Gophers Colly provides a clean interface to write any kind of crawler/scraper/spider. With Colly you can easily extract structured data from websites, which can

sensepost A golang, web screenshot utility using Chrome Headless. introduction gowitness is a website screenshot utility written in Golang, that uses Chrome Headless to generate screens

infinitbyte GOPA, A Spider Written in Go. Goal Light weight, low footprint, memory requirement should < 100MB Easy to deploy, no runtime or dependency required Easy to use, no programming or scripts ability needed,

integrii DO NOT USE! This project does not work anymore due to changes in Chrome. Use the Chrome DevTools protocol in chromedp instead! headlessChrome 🤖 Support only for Ubuntu on Docker for now. Mac appears t

gocolly Colly Lightning Fast and Elegant Scraping Framework for Gophers Colly provides a clean interface to write any kind of crawler/scraper/spider. With Colly you can easily extract structured data from websites, which can

kgretzky dcrawl dcrawl is a simple, but smart, multi-threaded web crawler for randomly gathering huge lists of unique domain names. How it works? dcrawl takes one site URL as input and detects all <a href=...&

benbjohnson deprecation warning active phantomjs development has ended, in favor of using Chrome's new headless functionality (reference). Instead of using this library, consider using a go package that uses this new api such as ch

PuerkitoBio gocrawl gocrawl is a polite, slim and concurrent web crawler written in Go. For a simpler yet more flexible web crawler written in a more idiomatic Go style, you may want to take a look at fetchbot, a package that bu

anaskhan96 soup Web Scraper in Go, similar to BeautifulSoup soup is a small web scraper package for Go, with its interface highly similar to that of BeautifulSoup. Exported variables and functions implemented till now : v

wspl About Creeper is a next-generation crawler which fetches web page by creeper script. As a cross-platform embedded crawler, you can use it for your news app, subscribe program, etc. Warning: At present this project

badoux goscraper Golang package to quickly return a preview of a webpage, you can get easily its title, description & images Usage func main() { s, err := goscraper.Scrape("https://www.w3.org/", 5) if err !=

bunsenapp go-selenium Introduction Yes, yet another Selenium Web Driver library has been brought to the table. This one, however, is slightly different. Easy to understand. Full test coverage by unit tests a

wcong ants-go open source, restful, distributed crawler engine gitter comming up Persistence Dynamic Master design of ants-go ants I wrote a crawler engine named ants in

cathalgarvey Sqrape - Simple Query Scraping with CSS and Go Reflection by Cathal Garvey, ©2016, Released under the GNU AGPLv3 What When scraping web content, one usually hopes that the content is laid out logically, an