Tiki Crawl | |
This is the alpha version of a crawler tool for checking links or gathering content from websites. It relies on the Crawler library from Spatie. Kudo to those guys.
|
Introduction | |
Tiki-Crawl is a powerful tool that allows you to crawl and index external websites or content from remote sources. It enables you to create a local index of content from external sites, which can be useful for various purposes such as aggregating content, building a search index, or archiving information. With Tiki-Crawl, you can specify a list of URLs or feeds from external website or remote sources that you want to index, and TikiWiki will periodically crawl these sources to fetch and index the content. This process is often referred to as web scraping and web crawling
|
What is Tiki-Crawl? | |
Tiki-Crawl is a web scraping and web crawling feature that enables users to fetch and index content from external websites or remote data sources. It automates the process of gathering information from these sources and can stores it.
|
Requirements | |
To use Tiki-Crawl effectively, ensure that you have this version installed (but it may very well be that it works on older versions):
|
Installation | |
First you need to clone the tool from https://gitlab.com/tikiwiki/tiki-crawl. This piece of code has been tested with
Copy to clipboard
|
Usage | |
First you need to configure your options. You can override any options from Copy to clipboard
Copy to clipboard
|
Config options | ||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||
This are configuration options that you can override in you new
|
Roadmap | |
|
Notes | |
|
Related Link | |