// choose http GET/method to download func connectByHttp(p *page.Page, req *request.Request) (*http.Response, error) { client := &http.Client{ CheckRedirect: req.GetRedirectFunc(), } httpreq, err := http.NewRequest(req.GetMethod(), req.GetUrl(), strings.NewReader(req.GetPostdata())) if header := req.GetHeader(); header != nil { httpreq.Header = req.GetHeader() } if cookies := req.GetCookies(); cookies != nil { for i := range cookies { httpreq.AddCookie(cookies[i]) } } var resp *http.Response if resp, err = client.Do(httpreq); err != nil { if e, ok := err.(*url.Error); ok && e.Err != nil && e.Err.Error() == "normal" { // normal } else { mlog.LogInst().LogError(err.Error()) p.SetStatus(true, err.Error()) //fmt.Printf("client do error %v \r\n", err) return nil, err } } return resp, nil }
// add Request to Schedule func (this *Spider) AddRequest(req *request.Request) *Spider { if req == nil { mlog.LogInst().LogError("request is nil") return this } else if req.GetUrl() == "" { mlog.LogInst().LogError("request is empty") return this } this.pScheduler.Push(req) return this }
func (this *QueueScheduler) Push(requ *request.Request) { this.locker.Lock() var key [md5.Size]byte if this.rm { key = md5.Sum([]byte(requ.GetUrl())) if _, ok := this.rmKey[key]; ok { this.locker.Unlock() return } } e := this.queue.PushBack(requ) if this.rm { this.rmKey[key] = e } this.locker.Unlock() }
// choose a proxy server to excute http GET/method to download func connectByHttpProxy(p *page.Page, in_req *request.Request) (*http.Response, error) { request, _ := http.NewRequest("GET", in_req.GetUrl(), nil) proxy, err := url.Parse(in_req.GetProxyHost()) if err != nil { return nil, err } client := &http.Client{ Transport: &http.Transport{ Proxy: http.ProxyURL(proxy), }, } resp, err := client.Do(request) if err != nil { return nil, err } return resp, nil }
// Download file and change the charset of page charset. func (this *HttpDownloader) downloadFile(p *page.Page, req *request.Request) (*page.Page, string) { var err error var urlstr string if urlstr = req.GetUrl(); len(urlstr) == 0 { mlog.LogInst().LogError("url is empty") p.SetStatus(true, "url is empty") return p, "" } var resp *http.Response if proxystr := req.GetProxyHost(); len(proxystr) != 0 { //using http proxy //fmt.Print("HttpProxy Enter ",proxystr,"\n") resp, err = connectByHttpProxy(p, req) } else { //normal http download //fmt.Print("Http Normal Enter \n",proxystr,"\n") resp, err = connectByHttp(p, req) } if err != nil { return p, "" } //b, _ := ioutil.ReadAll(resp.Body) //fmt.Printf("Resp body %v \r\n", string(b)) p.SetHeader(resp.Header) p.SetCookies(resp.Cookies()) // get converter to utf-8 var bodyStr string if resp.Header.Get("Content-Encoding") == "gzip" { bodyStr = this.changeCharsetEncodingAutoGzipSupport(resp.Header.Get("Content-Type"), resp.Body) } else { bodyStr = this.changeCharsetEncodingAuto(resp.Header.Get("Content-Type"), resp.Body) } //fmt.Printf("utf-8 body %v \r\n", bodyStr) defer resp.Body.Close() return p, bodyStr }