You are viewing a plain text version of this content. The canonical link for it is here.
Posted to dev@nutch.apache.org by Uroš Gruber <ur...@sir-mag.com> on 2006/09/06 19:43:05 UTC
[Fwd: Re: get CrawlDatum]
A while ago I posted this on dev list but without reply. I wonder if
this is right approach and If I continue to create this feature?
Do you think this idea would help nutch or maybe this is dead end and
you've already talked about this.
regards
Uros
Andrzej Bialecki wrote:
> Uroš Gruber wrote:
>> ParseData.metadata sounds nice, but I think I'm lost again :)
>> If I understand code flow the best place would be in Fetcher [262]
>>
>> but i'm not sure that datum holds info of url being fetched
>
> On the input to the fetcher you get a URL and a CrawlDatum (originally
> coming from the crawldb). Check for example how the segment name is
> passed around in metadata, you can use the same method.
>
Hi,
I made some draft patch. But there is still some problems I see. I know
code needs to be cleaned and test. But right now I don't know what
number set to external urls. For internal linking works great.
What is the whole idea of this changes.
Injected urls always get hop 0. While fetching/updating/generating hop
value is incremented by 1. (still no idea what to do with external
link). Then I can add config value max_hop etc. to limit fetcher and
generator to create more urls.
This way it's possible to limit crawling vertically
Comments are welcome.
Re: [Fwd: Re: get CrawlDatum]
Posted by Uroš Gruber <ur...@sir-mag.com>.
Andrzej Bialecki wrote:
> Uroš Gruber wrote:
>> I made some draft patch. But there is still some problems I see. I
>> know code needs to be cleaned and test. But right now I don't know
>> what number set to external urls. For internal linking works great.
>
> (the patch changes CrawlDatum itself, I think it would be better to
> put the hop counter in CrawlDatum.metaData.)
>
I can try to make with metaData
>>
>> What is the whole idea of this changes.
>>
>> Injected urls always get hop 0. While fetching/updating/generating
>> hop value is incremented by 1. (still no idea what to do with
>> external link). Then I can add config value max_hop etc. to limit
>> fetcher and generator to create more urls.
>>
>> This way it's possible to limit crawling vertically
>>
>> Comments are welcome.
>
> Well, it really depends on what you want to do when you encounter an
> external link. Do you want to restart the counter, i.e. crawl the new
> site at full depth up to max_hop? Then set hop=0. Do you want to
> terminate the crawl at that link? then set hop=max_hop.
>
I talk with my friend about this and here is what we've came up. Let say
URLs manualy injected are good and checked by human and probably you
wan't to start from it. So setting hop to 0 at injection is ok. While
crawling we have some sort of filtering by host (regexp etc.). We need
no worry about urls we don't have in our list so hop can be set whatever
it's, maybe to max_hop.
But here a scenario We add foo.com and bar.com from injection. After
crawling we find on site foo.com link to bar.com/hop/hop/index.html
We can set url hop to 0 or to max because we can update this after we
found this url on bar.com site.
Checking for hop needs to be done while updating I think, so we don't
end up with bunch of urls having hop greater than max_hop.
I will try to make a decent patch for this to check and if there is any
idea by others please make a comment on this.
regards
Uros
Re: [Fwd: Re: get CrawlDatum]
Posted by Andrzej Bialecki <ab...@getopt.org>.
Uroš Gruber wrote:
> I made some draft patch. But there is still some problems I see. I
> know code needs to be cleaned and test. But right now I don't know
> what number set to external urls. For internal linking works great.
(the patch changes CrawlDatum itself, I think it would be better to put
the hop counter in CrawlDatum.metaData.)
>
> What is the whole idea of this changes.
>
> Injected urls always get hop 0. While fetching/updating/generating hop
> value is incremented by 1. (still no idea what to do with external
> link). Then I can add config value max_hop etc. to limit fetcher and
> generator to create more urls.
>
> This way it's possible to limit crawling vertically
>
> Comments are welcome.
Well, it really depends on what you want to do when you encounter an
external link. Do you want to restart the counter, i.e. crawl the new
site at full depth up to max_hop? Then set hop=0. Do you want to
terminate the crawl at that link? then set hop=max_hop.
--
Best regards,
Andrzej Bialecki <><
___. ___ ___ ___ _ _ __________________________________
[__ || __|__/|__||\/| Information Retrieval, Semantic Web
___|||__|| \| || | Embedded Unix, System Integration
http://www.sigram.com Contact: info at sigram dot com