![]() |
Break the depth limit of 20?
Is the Depth limit of 20 a script limitation? a resource limitation? some sort of loop avoidance?
I ask because I tried to spider a directory where each new page of results is considered a new level, and there are categories with more than 20 pages. Can we break this limit somehow? Thanks! |
Just change it in the config file:
Code:
define('SPIDER_MAX_LIMIT',20); // max (re)index search depth - used for shell and admin panel dropdown |
Thanks-a-bunch Charter!
Off-side, are you the only developer behind PHPDigger? Do u take donations? |
Antoine was the previous developer, releasing the initial version through v.1.6.2, and I have since been the current developer. There have also been contributions posted in the forums and/or listed in the CREDITS, CHANGELOG, and README files. Some history about the change in developers can be found here.
|
Thanks.
I changed the depth limit to 60 and now i try to rerun the spider over the same domain so it will add the rest of links not spidered beyond the initial 20 hops, however it won't spider any link but the very first page and then stop. Ideas? |
Check the values in the update sites table via the admin panel.
|
They match my proposals: depth 60 and links 0 (aka all).
|
Some thoughts...
- Try using the textbox, 60, 0, no. - View the robots.txt file for changes. - Look for meta revisit-after/robots tags. - Enter the site at a different location. |
- Used both text and combo box
- No robots.txt present - No revisits on the code - Thats the only thing i should try now. However, does it make sense to index both www.domain.com and domain.com when they're 99% of the times the same thing? shouldn't this be implemented (even as a switch?) on the code of the digger? |
Set PHPDIG_IN_DOMAIN to true in the config file.
|
All times are GMT -8. The time now is 04:02 PM. |
Powered by vBulletin® Version 3.7.3
Copyright ©2000 - 2025, Jelsoft Enterprises Ltd.
Copyright © 2001 - 2005, ThinkDing LLC. All Rights Reserved.