View Full Version : What about sites on free hosting?
druesome
10-27-2003, 05:08 AM
Hey, I wonder if PHP Dig can accommodate sites such as those hosted by Geocities or Tripod etc. I tried to crawl one site in Geocities but instead it tried to crawl the whole Geocities network!! Is there any way around this?
Rolandks
10-27-2003, 08:22 AM
It works for me. :confused:
SITE : http://dsyn2.tripod.com/
Exclude paths :
- @NONE@
1:http://dsyn2.tripod.com/main.html
links found : 8
http://dsyn2.tripod.com/main.html
http://dsyn2.tripod.com/link.html
http://dsyn2.tripod.com/rev.html
http://dsyn2.tripod.com/
http://dsyn2.tripod.com/fic.html
http://dsyn2.tripod.com/rant.html
http://dsyn2.tripod.com/more.html
http://dsyn2.tripod.com/rmore.html
Optimizing tables...
Indexing complete !
-Roland-
druesome
10-27-2003, 08:41 AM
Hey,
What I meant are URL's like:
www.geocities.com/Area51/Space/1109 (non-existent)
or perhaps like
www.brinkster.com/druesome (non-existent)
because these are how their URL's are formatted. Totally forgot that tripod gave out subdomains, that's why yours worked, roland.:)
Charter
10-28-2003, 07:18 PM
Hi. Does setting PHPDIG_DEFAULT_INDEX to false have any effect?
druesome
10-28-2003, 09:46 PM
Nope. Right now PHPdig cannot accept slashes, it only takes the domain part of a URL.
So, if I try to crawl:
http://www.geocities.com/Area51/Space/
it smokes out Area51/Space and crawls:
http://www.geocities.com/
which practically is a nightmare..;)
BUT, i should make it clear that http://www.geocities.com/Area51/Space/ is also crawled. I think the reason why the rest of geocities is crawled is because the page contains links to other geocities sites or to the geocities homepage. This is a problem because you may be losing control over which sites to include especially if they are hosted on free servers. I have had experience with another search engine software called xavatoria, which is written in Perl, that addresses this problem quite well. Whichever site I crawl, no matter what folder, it only includes pages within the given folder and the subfolders it links to. So:
http://www.geocities.com/Area51/Space/
is treated as the root URL, and pages like
http://www.geocities.com/Area51/Space/index.html
http://www.geocities.com/Area51/Space/aboutme/
http://www.geocities.com/Area51/Space/contactme/
http://www.geocities.com/Area51/Space/favorites/movies.html
http://www.geocities.com/Area51/Space/favorites/books.html
are crawled exclusively.
I hope I'm making some sense here.
:)
Thanks.
Charter
10-29-2003, 03:44 AM
Hi. Yep, that makes sense. As a temporary solution, how about setting the index levels to one when crawling those links via browser or text file?
druesome
11-05-2003, 11:45 AM
Hmm.. can you show me how to do that through a text file?
Charter
11-07-2003, 11:35 AM
Hi. Perhaps try the following.
Set the following in the config file:
define('SPIDER_MAX_LIMIT',1); // default = 20
define('SPIDER_DEFAULT_LIMIT',1); // default = 3
define('RESPIDER_LIMIT',1); // default = 4
Also, set the following in the config file:
define('LIMIT_DAYS',0); // default = 7
Then make a text file with a list of full URLs, one per line, and try indexing from shell.
vBulletin® v3.7.3, Copyright ©2000-2024, Jelsoft Enterprises Ltd.