@@ -39,8 +39,8 @@ use crate::{
39
39
} ;
40
40
41
41
use super :: {
42
- encoded_body, robot_client:: RobotClient , wander_prirotiser :: WanderPrioritiser , CrawlDatum ,
43
- DatumStream , Domain , Error , Result , RetrieableUrl , Site , WarcWriter , WeightedUrl , WorkerJob ,
42
+ encoded_body, robot_client:: RobotClient , wander_prioritiser :: WanderPrioritiser , CrawlDatum ,
43
+ DatumSink , Domain , Error , Result , RetrieableUrl , Site , WarcWriter , WeightedUrl , WorkerJob ,
44
44
MAX_CONTENT_LENGTH , MAX_OUTGOING_URLS_PER_PAGE ,
45
45
} ;
46
46
@@ -126,7 +126,8 @@ impl WorkerThread {
126
126
}
127
127
}
128
128
129
- pub struct JobExecutor < S : DatumStream > {
129
+ /// JobExecutor receives a job from the coordinator and crawls the urls in the job.
130
+ pub struct JobExecutor < S : DatumSink > {
130
131
writer : Arc < S > ,
131
132
client : RobotClient ,
132
133
has_gotten_429_response : bool ,
@@ -144,7 +145,7 @@ pub struct JobExecutor<S: DatumStream> {
144
145
job : WorkerJob ,
145
146
}
146
147
147
- impl < S : DatumStream > JobExecutor < S > {
148
+ impl < S : DatumSink > JobExecutor < S > {
148
149
pub fn new (
149
150
job : WorkerJob ,
150
151
config : Arc < CrawlerConfig > ,
0 commit comments