• Home
  • Line#
  • Scopes#
  • Navigate#
  • Raw
  • Download
1 // Copyright (c) 2011 The Chromium Authors. All rights reserved.
2 // Use of this source code is governed by a BSD-style license that can be
3 // found in the LICENSE file.
4 
5 #include "net/url_request/url_request_job.h"
6 
7 #include "base/compiler_specific.h"
8 #include "base/message_loop.h"
9 #include "base/string_number_conversions.h"
10 #include "base/string_util.h"
11 #include "net/base/auth.h"
12 #include "net/base/host_port_pair.h"
13 #include "net/base/io_buffer.h"
14 #include "net/base/load_states.h"
15 #include "net/base/net_errors.h"
16 #include "net/base/network_delegate.h"
17 #include "net/http/http_response_headers.h"
18 #include "net/url_request/url_request.h"
19 #include "net/url_request/url_request_context.h"
20 #include "net/url_request/url_request_job_tracker.h"
21 
22 namespace net {
23 
URLRequestJob(URLRequest * request)24 URLRequestJob::URLRequestJob(URLRequest* request)
25     : request_(request),
26       done_(false),
27       prefilter_bytes_read_(0),
28       postfilter_bytes_read_(0),
29       filter_input_byte_count_(0),
30       filter_needs_more_output_space_(false),
31       filtered_read_buffer_len_(0),
32       has_handled_response_(false),
33       expected_content_size_(-1),
34       deferred_redirect_status_code_(-1),
35       ALLOW_THIS_IN_INITIALIZER_LIST(method_factory_(this)) {
36   g_url_request_job_tracker.AddNewJob(this);
37 }
38 
SetUpload(UploadData * upload)39 void URLRequestJob::SetUpload(UploadData* upload) {
40 }
41 
SetExtraRequestHeaders(const HttpRequestHeaders & headers)42 void URLRequestJob::SetExtraRequestHeaders(
43     const HttpRequestHeaders& headers) {
44 }
45 
Kill()46 void URLRequestJob::Kill() {
47   // Make sure the request is notified that we are done.  We assume that the
48   // request took care of setting its error status before calling Kill.
49   if (request_)
50     NotifyCanceled();
51 }
52 
DetachRequest()53 void URLRequestJob::DetachRequest() {
54   request_ = NULL;
55 }
56 
57 // This function calls ReadData to get stream data. If a filter exists, passes
58 // the data to the attached filter. Then returns the output from filter back to
59 // the caller.
Read(IOBuffer * buf,int buf_size,int * bytes_read)60 bool URLRequestJob::Read(IOBuffer* buf, int buf_size, int *bytes_read) {
61   bool rv = false;
62 
63   DCHECK_LT(buf_size, 1000000);  // sanity check
64   DCHECK(buf);
65   DCHECK(bytes_read);
66   DCHECK(filtered_read_buffer_ == NULL);
67   DCHECK_EQ(0, filtered_read_buffer_len_);
68 
69   *bytes_read = 0;
70 
71   // Skip Filter if not present
72   if (!filter_.get()) {
73     rv = ReadRawDataHelper(buf, buf_size, bytes_read);
74   } else {
75     // Save the caller's buffers while we do IO
76     // in the filter's buffers.
77     filtered_read_buffer_ = buf;
78     filtered_read_buffer_len_ = buf_size;
79 
80     if (ReadFilteredData(bytes_read)) {
81       rv = true;   // we have data to return
82     } else {
83       rv = false;  // error, or a new IO is pending
84     }
85   }
86   if (rv && *bytes_read == 0)
87     NotifyDone(URLRequestStatus());
88   return rv;
89 }
90 
StopCaching()91 void URLRequestJob::StopCaching() {
92   // Nothing to do here.
93 }
94 
GetLoadState() const95 LoadState URLRequestJob::GetLoadState() const {
96   return LOAD_STATE_IDLE;
97 }
98 
GetUploadProgress() const99 uint64 URLRequestJob::GetUploadProgress() const {
100   return 0;
101 }
102 
GetCharset(std::string * charset)103 bool URLRequestJob::GetCharset(std::string* charset) {
104   return false;
105 }
106 
GetResponseInfo(HttpResponseInfo * info)107 void URLRequestJob::GetResponseInfo(HttpResponseInfo* info) {
108 }
109 
GetResponseCookies(std::vector<std::string> * cookies)110 bool URLRequestJob::GetResponseCookies(std::vector<std::string>* cookies) {
111   return false;
112 }
113 
SetupFilter() const114 Filter* URLRequestJob::SetupFilter() const {
115   return NULL;
116 }
117 
IsRedirectResponse(GURL * location,int * http_status_code)118 bool URLRequestJob::IsRedirectResponse(GURL* location,
119                                        int* http_status_code) {
120   // For non-HTTP jobs, headers will be null.
121   HttpResponseHeaders* headers = request_->response_headers();
122   if (!headers)
123     return false;
124 
125   std::string value;
126   if (!headers->IsRedirect(&value))
127     return false;
128 
129   *location = request_->url().Resolve(value);
130   *http_status_code = headers->response_code();
131   return true;
132 }
133 
IsSafeRedirect(const GURL & location)134 bool URLRequestJob::IsSafeRedirect(const GURL& location) {
135   return true;
136 }
137 
NeedsAuth()138 bool URLRequestJob::NeedsAuth() {
139   return false;
140 }
141 
GetAuthChallengeInfo(scoped_refptr<AuthChallengeInfo> * auth_info)142 void URLRequestJob::GetAuthChallengeInfo(
143     scoped_refptr<AuthChallengeInfo>* auth_info) {
144   // This will only be called if NeedsAuth() returns true, in which
145   // case the derived class should implement this!
146   NOTREACHED();
147 }
148 
SetAuth(const string16 & username,const string16 & password)149 void URLRequestJob::SetAuth(const string16& username,
150                             const string16& password) {
151   // This will only be called if NeedsAuth() returns true, in which
152   // case the derived class should implement this!
153   NOTREACHED();
154 }
155 
CancelAuth()156 void URLRequestJob::CancelAuth() {
157   // This will only be called if NeedsAuth() returns true, in which
158   // case the derived class should implement this!
159   NOTREACHED();
160 }
161 
ContinueWithCertificate(X509Certificate * client_cert)162 void URLRequestJob::ContinueWithCertificate(
163     X509Certificate* client_cert) {
164   // The derived class should implement this!
165   NOTREACHED();
166 }
167 
ContinueDespiteLastError()168 void URLRequestJob::ContinueDespiteLastError() {
169   // Implementations should know how to recover from errors they generate.
170   // If this code was reached, we are trying to recover from an error that
171   // we don't know how to recover from.
172   NOTREACHED();
173 }
174 
FollowDeferredRedirect()175 void URLRequestJob::FollowDeferredRedirect() {
176   DCHECK(deferred_redirect_status_code_ != -1);
177 
178   // NOTE: deferred_redirect_url_ may be invalid, and attempting to redirect to
179   // such an URL will fail inside FollowRedirect.  The DCHECK above asserts
180   // that we called OnReceivedRedirect.
181 
182   // It is also possible that FollowRedirect will drop the last reference to
183   // this job, so we need to reset our members before calling it.
184 
185   GURL redirect_url = deferred_redirect_url_;
186   int redirect_status_code = deferred_redirect_status_code_;
187 
188   deferred_redirect_url_ = GURL();
189   deferred_redirect_status_code_ = -1;
190 
191   FollowRedirect(redirect_url, redirect_status_code);
192 }
193 
GetMimeType(std::string * mime_type) const194 bool URLRequestJob::GetMimeType(std::string* mime_type) const {
195   return false;
196 }
197 
GetResponseCode() const198 int URLRequestJob::GetResponseCode() const {
199   return -1;
200 }
201 
GetSocketAddress() const202 HostPortPair URLRequestJob::GetSocketAddress() const {
203   return HostPortPair();
204 }
205 
~URLRequestJob()206 URLRequestJob::~URLRequestJob() {
207   g_url_request_job_tracker.RemoveJob(this);
208 }
209 
NotifyHeadersComplete()210 void URLRequestJob::NotifyHeadersComplete() {
211   if (!request_ || !request_->delegate())
212     return;  // The request was destroyed, so there is no more work to do.
213 
214   if (has_handled_response_)
215     return;
216 
217   DCHECK(!request_->status().is_io_pending());
218 
219   // Initialize to the current time, and let the subclass optionally override
220   // the time stamps if it has that information.  The default request_time is
221   // set by URLRequest before it calls our Start method.
222   request_->response_info_.response_time = base::Time::Now();
223   GetResponseInfo(&request_->response_info_);
224 
225   // When notifying the delegate, the delegate can release the request
226   // (and thus release 'this').  After calling to the delgate, we must
227   // check the request pointer to see if it still exists, and return
228   // immediately if it has been destroyed.  self_preservation ensures our
229   // survival until we can get out of this method.
230   scoped_refptr<URLRequestJob> self_preservation(this);
231 
232   GURL new_location;
233   int http_status_code;
234   if (IsRedirectResponse(&new_location, &http_status_code)) {
235     const GURL& url = request_->url();
236 
237     // Move the reference fragment of the old location to the new one if the
238     // new one has none. This duplicates mozilla's behavior.
239     if (url.is_valid() && url.has_ref() && !new_location.has_ref()) {
240       GURL::Replacements replacements;
241       // Reference the |ref| directly out of the original URL to avoid a
242       // malloc.
243       replacements.SetRef(url.spec().data(),
244                           url.parsed_for_possibly_invalid_spec().ref);
245       new_location = new_location.ReplaceComponents(replacements);
246     }
247 
248     bool defer_redirect = false;
249     request_->ReceivedRedirect(new_location, &defer_redirect);
250 
251     // Ensure that the request wasn't detached or destroyed in ReceivedRedirect
252     if (!request_ || !request_->delegate())
253       return;
254 
255     // If we were not cancelled, then maybe follow the redirect.
256     if (request_->status().is_success()) {
257       if (defer_redirect) {
258         deferred_redirect_url_ = new_location;
259         deferred_redirect_status_code_ = http_status_code;
260       } else {
261         FollowRedirect(new_location, http_status_code);
262       }
263       return;
264     }
265   } else if (NeedsAuth()) {
266     scoped_refptr<AuthChallengeInfo> auth_info;
267     GetAuthChallengeInfo(&auth_info);
268     // Need to check for a NULL auth_info because the server may have failed
269     // to send a challenge with the 401 response.
270     if (auth_info) {
271       request_->delegate()->OnAuthRequired(request_, auth_info);
272       // Wait for SetAuth or CancelAuth to be called.
273       return;
274     }
275   }
276 
277   has_handled_response_ = true;
278   if (request_->status().is_success())
279     filter_.reset(SetupFilter());
280 
281   if (!filter_.get()) {
282     std::string content_length;
283     request_->GetResponseHeaderByName("content-length", &content_length);
284     if (!content_length.empty())
285       base::StringToInt64(content_length, &expected_content_size_);
286   }
287 
288   request_->ResponseStarted();
289 }
290 
NotifyReadComplete(int bytes_read)291 void URLRequestJob::NotifyReadComplete(int bytes_read) {
292   if (!request_ || !request_->delegate())
293     return;  // The request was destroyed, so there is no more work to do.
294 
295   // TODO(darin): Bug 1004233. Re-enable this test once all of the chrome
296   // unit_tests have been fixed to not trip this.
297   //DCHECK(!request_->status().is_io_pending());
298 
299   // The headers should be complete before reads complete
300   DCHECK(has_handled_response_);
301 
302   OnRawReadComplete(bytes_read);
303 
304   // Don't notify if we had an error.
305   if (!request_->status().is_success())
306     return;
307 
308   // When notifying the delegate, the delegate can release the request
309   // (and thus release 'this').  After calling to the delgate, we must
310   // check the request pointer to see if it still exists, and return
311   // immediately if it has been destroyed.  self_preservation ensures our
312   // survival until we can get out of this method.
313   scoped_refptr<URLRequestJob> self_preservation(this);
314 
315   prefilter_bytes_read_ += bytes_read;
316   if (filter_.get()) {
317     // Tell the filter that it has more data
318     FilteredDataRead(bytes_read);
319 
320     // Filter the data.
321     int filter_bytes_read = 0;
322     if (ReadFilteredData(&filter_bytes_read)) {
323       postfilter_bytes_read_ += filter_bytes_read;
324       if (request_->context() && request_->context()->network_delegate()) {
325         request_->context()->network_delegate()->NotifyReadCompleted(
326             request_, filter_bytes_read);
327       }
328       request_->delegate()->OnReadCompleted(request_, filter_bytes_read);
329     }
330   } else {
331     postfilter_bytes_read_ += bytes_read;
332     if (request_->context() && request_->context()->network_delegate()) {
333       request_->context()->network_delegate()->NotifyReadCompleted(
334           request_, bytes_read);
335     }
336     request_->delegate()->OnReadCompleted(request_, bytes_read);
337   }
338 }
339 
NotifyStartError(const URLRequestStatus & status)340 void URLRequestJob::NotifyStartError(const URLRequestStatus &status) {
341   DCHECK(!has_handled_response_);
342   has_handled_response_ = true;
343   if (request_) {
344     request_->set_status(status);
345     request_->ResponseStarted();
346   }
347 }
348 
NotifyDone(const URLRequestStatus & status)349 void URLRequestJob::NotifyDone(const URLRequestStatus &status) {
350   DCHECK(!done_) << "Job sending done notification twice";
351   if (done_)
352     return;
353   done_ = true;
354 
355   // Unless there was an error, we should have at least tried to handle
356   // the response before getting here.
357   DCHECK(has_handled_response_ || !status.is_success());
358 
359   // As with NotifyReadComplete, we need to take care to notice if we were
360   // destroyed during a delegate callback.
361   if (request_) {
362     request_->set_is_pending(false);
363     // With async IO, it's quite possible to have a few outstanding
364     // requests.  We could receive a request to Cancel, followed shortly
365     // by a successful IO.  For tracking the status(), once there is
366     // an error, we do not change the status back to success.  To
367     // enforce this, only set the status if the job is so far
368     // successful.
369     if (request_->status().is_success())
370       request_->set_status(status);
371   }
372 
373   g_url_request_job_tracker.OnJobDone(this, status);
374 
375   // Complete this notification later.  This prevents us from re-entering the
376   // delegate if we're done because of a synchronous call.
377   MessageLoop::current()->PostTask(
378       FROM_HERE,
379       method_factory_.NewRunnableMethod(&URLRequestJob::CompleteNotifyDone));
380 }
381 
CompleteNotifyDone()382 void URLRequestJob::CompleteNotifyDone() {
383   // Check if we should notify the delegate that we're done because of an error.
384   if (request_ &&
385       !request_->status().is_success() &&
386       request_->delegate()) {
387     // We report the error differently depending on whether we've called
388     // OnResponseStarted yet.
389     if (has_handled_response_) {
390       // We signal the error by calling OnReadComplete with a bytes_read of -1.
391       if (request_->context() && request_->context()->network_delegate())
392         request_->context()->network_delegate()->NotifyReadCompleted(
393             request_, -1);
394       request_->delegate()->OnReadCompleted(request_, -1);
395     } else {
396       has_handled_response_ = true;
397       request_->ResponseStarted();
398     }
399   }
400 }
401 
NotifyCanceled()402 void URLRequestJob::NotifyCanceled() {
403   if (!done_) {
404     NotifyDone(URLRequestStatus(URLRequestStatus::CANCELED,
405                                 ERR_ABORTED));
406   }
407 }
408 
NotifyRestartRequired()409 void URLRequestJob::NotifyRestartRequired() {
410   DCHECK(!has_handled_response_);
411   if (GetStatus().status() != URLRequestStatus::CANCELED)
412     request_->Restart();
413 }
414 
ReadRawData(IOBuffer * buf,int buf_size,int * bytes_read)415 bool URLRequestJob::ReadRawData(IOBuffer* buf, int buf_size,
416                                 int *bytes_read) {
417   DCHECK(bytes_read);
418   *bytes_read = 0;
419   NotifyDone(URLRequestStatus());
420   return false;
421 }
422 
FilteredDataRead(int bytes_read)423 void URLRequestJob::FilteredDataRead(int bytes_read) {
424   DCHECK(filter_.get());  // don't add data if there is no filter
425   filter_->FlushStreamBuffer(bytes_read);
426 }
427 
ReadFilteredData(int * bytes_read)428 bool URLRequestJob::ReadFilteredData(int* bytes_read) {
429   DCHECK(filter_.get());  // don't add data if there is no filter
430   DCHECK(filtered_read_buffer_ != NULL);  // we need to have a buffer to fill
431   DCHECK_GT(filtered_read_buffer_len_, 0);  // sanity check
432   DCHECK_LT(filtered_read_buffer_len_, 1000000);  // sanity check
433   DCHECK(raw_read_buffer_ == NULL);  // there should be no raw read buffer yet
434 
435   bool rv = false;
436   *bytes_read = 0;
437 
438   if (is_done())
439     return true;
440 
441   if (!filter_needs_more_output_space_ && !filter_->stream_data_len()) {
442     // We don't have any raw data to work with, so
443     // read from the socket.
444     int filtered_data_read;
445     if (ReadRawDataForFilter(&filtered_data_read)) {
446       if (filtered_data_read > 0) {
447         filter_->FlushStreamBuffer(filtered_data_read);  // Give data to filter.
448       } else {
449         return true;  // EOF
450       }
451     } else {
452       return false;  // IO Pending (or error)
453     }
454   }
455 
456   if ((filter_->stream_data_len() || filter_needs_more_output_space_)
457       && !is_done()) {
458     // Get filtered data.
459     int filtered_data_len = filtered_read_buffer_len_;
460     Filter::FilterStatus status;
461     int output_buffer_size = filtered_data_len;
462     status = filter_->ReadData(filtered_read_buffer_->data(),
463                                &filtered_data_len);
464 
465     if (filter_needs_more_output_space_ && 0 == filtered_data_len) {
466       // filter_needs_more_output_space_ was mistaken... there are no more bytes
467       // and we should have at least tried to fill up the filter's input buffer.
468       // Correct the state, and try again.
469       filter_needs_more_output_space_ = false;
470       return ReadFilteredData(bytes_read);
471     }
472 
473     switch (status) {
474       case Filter::FILTER_DONE: {
475         filter_needs_more_output_space_ = false;
476         *bytes_read = filtered_data_len;
477         rv = true;
478         break;
479       }
480       case Filter::FILTER_NEED_MORE_DATA: {
481         filter_needs_more_output_space_ =
482             (filtered_data_len == output_buffer_size);
483         // We have finished filtering all data currently in the buffer.
484         // There might be some space left in the output buffer. One can
485         // consider reading more data from the stream to feed the filter
486         // and filling up the output buffer. This leads to more complicated
487         // buffer management and data notification mechanisms.
488         // We can revisit this issue if there is a real perf need.
489         if (filtered_data_len > 0) {
490           *bytes_read = filtered_data_len;
491           rv = true;
492         } else {
493           // Read again since we haven't received enough data yet (e.g., we may
494           // not have a complete gzip header yet)
495           rv = ReadFilteredData(bytes_read);
496         }
497         break;
498       }
499       case Filter::FILTER_OK: {
500         filter_needs_more_output_space_ =
501             (filtered_data_len == output_buffer_size);
502         *bytes_read = filtered_data_len;
503         rv = true;
504         break;
505       }
506       case Filter::FILTER_ERROR: {
507         filter_needs_more_output_space_ = false;
508         NotifyDone(URLRequestStatus(URLRequestStatus::FAILED,
509                    ERR_CONTENT_DECODING_FAILED));
510         rv = false;
511         break;
512       }
513       default: {
514         NOTREACHED();
515         filter_needs_more_output_space_ = false;
516         rv = false;
517         break;
518       }
519     }
520   } else {
521     // we are done, or there is no data left.
522     rv = true;
523   }
524 
525   if (rv) {
526     // When we successfully finished a read, we no longer need to
527     // save the caller's buffers. Release our reference.
528     filtered_read_buffer_ = NULL;
529     filtered_read_buffer_len_ = 0;
530   }
531   return rv;
532 }
533 
GetStatus()534 const URLRequestStatus URLRequestJob::GetStatus() {
535   if (request_)
536     return request_->status();
537   // If the request is gone, we must be cancelled.
538   return URLRequestStatus(URLRequestStatus::CANCELED,
539                           ERR_ABORTED);
540 }
541 
SetStatus(const URLRequestStatus & status)542 void URLRequestJob::SetStatus(const URLRequestStatus &status) {
543   if (request_)
544     request_->set_status(status);
545 }
546 
ReadRawDataForFilter(int * bytes_read)547 bool URLRequestJob::ReadRawDataForFilter(int* bytes_read) {
548   bool rv = false;
549 
550   DCHECK(bytes_read);
551   DCHECK(filter_.get());
552 
553   *bytes_read = 0;
554 
555   // Get more pre-filtered data if needed.
556   // TODO(mbelshe): is it possible that the filter needs *MORE* data
557   //    when there is some data already in the buffer?
558   if (!filter_->stream_data_len() && !is_done()) {
559     IOBuffer* stream_buffer = filter_->stream_buffer();
560     int stream_buffer_size = filter_->stream_buffer_size();
561     rv = ReadRawDataHelper(stream_buffer, stream_buffer_size, bytes_read);
562   }
563   return rv;
564 }
565 
ReadRawDataHelper(IOBuffer * buf,int buf_size,int * bytes_read)566 bool URLRequestJob::ReadRawDataHelper(IOBuffer* buf, int buf_size,
567                                       int* bytes_read) {
568   DCHECK(!request_->status().is_io_pending());
569   DCHECK(raw_read_buffer_ == NULL);
570 
571   // Keep a pointer to the read buffer, so we have access to it in the
572   // OnRawReadComplete() callback in the event that the read completes
573   // asynchronously.
574   raw_read_buffer_ = buf;
575   bool rv = ReadRawData(buf, buf_size, bytes_read);
576 
577   if (!request_->status().is_io_pending()) {
578     // If the read completes synchronously, either success or failure,
579     // invoke the OnRawReadComplete callback so we can account for the
580     // completed read.
581     OnRawReadComplete(*bytes_read);
582   }
583   return rv;
584 }
585 
FollowRedirect(const GURL & location,int http_status_code)586 void URLRequestJob::FollowRedirect(const GURL& location, int http_status_code) {
587   g_url_request_job_tracker.OnJobRedirect(this, location, http_status_code);
588 
589   int rv = request_->Redirect(location, http_status_code);
590   if (rv != OK)
591     NotifyDone(URLRequestStatus(URLRequestStatus::FAILED, rv));
592 }
593 
OnRawReadComplete(int bytes_read)594 void URLRequestJob::OnRawReadComplete(int bytes_read) {
595   DCHECK(raw_read_buffer_);
596   if (bytes_read > 0) {
597     RecordBytesRead(bytes_read);
598   }
599   raw_read_buffer_ = NULL;
600 }
601 
RecordBytesRead(int bytes_read)602 void URLRequestJob::RecordBytesRead(int bytes_read) {
603   filter_input_byte_count_ += bytes_read;
604   UpdatePacketReadTimes();  // Facilitate stats recording if it is active.
605   g_url_request_job_tracker.OnBytesRead(this, raw_read_buffer_->data(),
606                                         bytes_read);
607 }
608 
FilterHasData()609 bool URLRequestJob::FilterHasData() {
610     return filter_.get() && filter_->stream_data_len();
611 }
612 
UpdatePacketReadTimes()613 void URLRequestJob::UpdatePacketReadTimes() {
614 }
615 
616 }  // namespace net
617