High level overview
This package will crawl your entire site and will put the content in a search index. This way, the entire content of your site is searchable. Think of it as a private Google search index.
The configuration for each site that needs to be crawled is saved in the
site_search_configs table. You can manually create a row in that table or run this artisan command:
Next, you can fill up a search index by executing the crawl command. Before that commands actually start crawling, it will clean up old indexes whose names start with the
index_base_name specified in the
site_search_configs table. After that, it will create a new empty Meilisearch index. The name of that new index will be saved in the
pending_index_name column of the
A search profile class will determine which pages get crawled and which pages should be put in the Meilisearch index. An indexer class will transform the HTML of a page to something that can be saved in the index.
When crawling your site, multiple concurrent connections are used to speed up the crawling process.
After the site has been crawled, the Meilisearch index whose name is in
site_search_configs will be deleted. The
index_name will now be set to the value of