(澳门威尼斯人官网) 爬虫 页面不存在_百度搜索
澳门威尼斯人官网
澳门威尼斯人官网
当前位置 : 澳门威尼斯人官网 > 澳门威尼斯人官网

(澳门威尼斯人官网) 爬虫 页面不存在_百度搜索

1.学写爬虫,遇到一个问题,加了values={"wd":"test","ie":"utf-8"}就报错了,也就是我向百度页面搜索里面输入test就报错,百度了一圈没找到答案,请教一下各位大神如何处理,如果可以的话能否说明这是为什么,以及怎么处理类似的情况,谢谢!

2.代码如下(版本2.7):
(randHeader用来生成随机的Header,savef保存输出结果,logging.info用来调试,try处理可能的错误,其他部分就是爬虫的主体)

#coding:utf-8
import random
import urllib2
import urllib
import os
import logging
logging.basicConfiglevel=logging.INFO

def savefhtml,tpath:
    #12输入输出目录不存在,就创建
    if not os.path.existstpath:
        os.makedirstpath
    #12打开需要保存的文件(该命令下,如果文件不存在就创建新文件)
    with openos.path.jointpath,2.html,w as f:
        f.writehtml
    with openos.path.jointpath,2.xml,w as f:
        f.writehtml
    with openos.path.jointpath,2.txt,w as f:
        f.writehtml
        
def randHeader:
    
    head_connection = [Keep-Alive,close]
    head_accept = [text/html, application/xhtml+xml, */*]
    head_accept_language = [zh-CN,fr-FR;q=0.5,en-US,en;q=0.8,zh-Hans-CN;q=0.5,zh-Hans;q=0.3]
    head_user_agent = [Mozilla/5.0 Windows NT 6.3; WOW64; Trident/7.0; rv:11.0 like Gecko,
                       Mozilla/5.0 Windows NT 5.1 AppleWebKit/537.36 KHTML, like Gecko Chrome/28.0.1500.95 Safari/537.36,
                       Mozilla/5.0 Windows NT 6.1; WOW64; Trident/7.0; SLCC2; .NET CLR 2.0.50727; .NET CLR 3.5.30729; .NET CLR 3.0.30729; Media Center PC 6.0; .NET4.0C; rv:11.0 like Gecko,
                       Mozilla/5.0 Windows; U; Windows NT 5.2 Gecko/2008070208 Firefox/3.0.1,
                       Mozilla/5.0 Windows; U; Windows NT 5.1 Gecko/20070309 Firefox/2.0.0.3,
                       Mozilla/5.0 Windows; U; Windows NT 5.1 Gecko/20070803 Firefox/1.5.0.12,
                       Opera/9.27 Windows NT 5.2; U; zh-cn,
                       Mozilla/5.0 Macintosh; PPC Mac OS X; U; en Opera 8.0,
                       Opera/8.0 Macintosh; PPC Mac OS X; U; en,
                       Mozilla/5.0 Windows; U; Windows NT 5.1; en-US; rv:1.8.1.12 Gecko/20080219 Firefox/2.0.0.12 Navigator/9.0.0.6,
                       Mozilla/4.0 compatible; MSIE 8.0; Windows NT 6.1; Win64; x64; Trident/4.0,
                       Mozilla/4.0 compatible; MSIE 8.0; Windows NT 6.1; Trident/4.0,
                       Mozilla/5.0 compatible; MSIE 10.0; Windows NT 6.1; WOW64; Trident/6.0; SLCC2; .NET CLR 2.0.50727; .NET CLR 3.5.30729; .NET CLR 3.0.30729; Media Center PC 6.0; InfoPath.2; .NET4.0C; .NET4.0E,
                       Mozilla/5.0 Windows NT 6.1; WOW64 AppleWebKit/537.1 KHTML, like Gecko Maxthon/4.0.6.2000 Chrome/26.0.1410.43 Safari/537.1 ,
                       Mozilla/5.0 compatible; MSIE 10.0; Windows NT 6.1; WOW64; Trident/6.0; SLCC2; .NET CLR 2.0.50727; .NET CLR 3.5.30729; .NET CLR 3.0.30729; Media Center PC 6.0; InfoPath.2; .NET4.0C; .NET4.0E; QQBrowser/7.3.9825.400,
                       Mozilla/5.0 Windows NT 6.1; WOW64; rv:21.0 Gecko/20100101 Firefox/21.0 ,
                       Mozilla/5.0 Windows NT 6.1; WOW64 AppleWebKit/537.1 KHTML, like Gecko Chrome/21.0.1180.92 Safari/537.1 LBBROWSER,
                       Mozilla/5.0 compatible; MSIE 10.0; Windows NT 6.1; WOW64; Trident/6.0; BIDUBrowser 2.x,
                       Mozilla/5.0 Windows NT 6.1; WOW64 AppleWebKit/536.11 KHTML, like Gecko Chrome/20.0.1132.11 TaoBrowser/3.0 Safari/536.11]
    
    
    header = {
        Connection: head_connection[0],
        Accept: head_accept[0],
        Accept-Language: head_accept_language[1],
        User-Agent: head_user_agent[random.randrange0,lenhead_user_agent]
    }
    return header
"""
#输出随机Header
for i in range10:
    printrandHeader
"""
url1="http://www.baidu.com/"
kkk=url1.split"/"[2]
headers=randHeader
logging.info"headers=%s" % headers

values={"wd":"test","ie":"utf-8"}
data=urllib.urlencodevalues
request=urllib2.Requesturl1,data,headers
#HTTP错误处理
try:
    urllib2.urlopenrequest
except URLError,e:
    logging.info"e.code=%s" % e.code
    logging.info"e.read=" % e.read
response=urllib2.urlopenrequest
html=response.read

#根目录
cpath="I:\\"
#拼输出目录
tpath=os.path.joincpath,kkk
savefhtml,tpath

3.输出结果如下:

<!DOCTYPE html>

<!--STATUS OK-->

<html>

<head>

    <meta http-equiv="X-UA-Compatible" content="IE=edge,chrome=1">

    <meta http-equiv="content-type" content="text/html;charset=utf-8">

    <meta content="always" name="referrer">

    <title>页面不存在_百度搜索</title>

    <style data-for="result">
    ...由于太多,后面就省略了

因为你的百度地址错了
实际上还要加一个/s,
http://www.baidu.com/s

完整包括关键词的地址是这样
http://www.baidu.com/s?wd=key...

包括页码的地址是这样的
http://www.baidu.com/s?wd=key...

你如果不加s的话,就会提示页面没找到

====================================

#/usr/bin/python
# -*- coding: Utf-8 -*-
# -*- author: Wd0g  -*-
import requests,gzip

def httpGeturl:
    try:
        res = requests.session.geturl,verify=True
    except Exception as e:
        return False
    try:
        data = res.content.decodeutf-8
    except:
        try:
            data = res.content.decodegbk
        except:
            data = gzip.decompressres.content.decode
    return data

def baidukeywords, pageNum:
    url = http://www.baidu.com/s?wd=%s&pn=%d %keywords, pageNum*10
    return httpGeturl

printbaiduwd0g,0

这是python3.5的代码,我这里是没有出现你说过的问题

一个简单的Python爬虫写这么多代码。。。。。。

import urllib.request
html = urllib.request.urlopen"http://www.baidu.com"
content = html.read
content = content.decode"utf-8"
printcontent

写爬虫的话用requests库更为方便一些,不用设置复杂的Header

import requests
response = requests.gethttp://www.baidu.com
response.status_code # 200
response.text # <!doctype html><html>.....</html>
import urllib
import requests

keyword = test
url = http://www.baidu.com/s?wd=%s % urllib.quotekeyword
res = requests.geturl
html = res.content

广告位

澳门威尼斯人官网