1.使用工具:
| 1 2 | Python 3.5 BeautifulSoup | 
2.抓取网站:
csdn首页文章列表 http://blog.csdn.net/
3、实现代码:
| 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 | __author__ = 'Administrator' import urllib.request import re from bs4 import BeautifulSoup ######################################################## # #              抓取csdn首页文章http://blog.csdn.net/?&page=1 # #             参数baseUrl是要访问的网站地址 # ######################################################## class CsdnUtils(object):     def __init__(self):         user_agent='Mozilla/5.0 (Windows NT 6.1; WOW64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/50.0.2661.102 Safari/537.36'         self.headers ={'Cache-Control':'max-age=0',                         'Connection':'keep-alive',                         'Accept': 'text/html,application/xhtml+xml,application/xml;q=0.9,image/webp,*/*;q=0.8',                         'User-Agent':user_agent,                         }         self.contentAll=set()     def getPage(self,url=None):         request=urllib.request.Request(url,headers=self.headers)         response=urllib.request.urlopen(request)         soup=BeautifulSoup(response.read(),"html.parser")         #print(soup.prettify())         return soup     def parsePage(self,url=None,page_num=None):         soup=self.getPage(url)         itemBlog=soup.find_all('dl','blog_list clearfix')         cnArticle=CsdnUtils         print("========================第",page_num,"页======================================")         for i,itemSingle in enumerate(itemBlog):             cnArticle.num=i             cnArticle.author=itemSingle.find('a','nickname').string             cnArticle.postTime=itemSingle.find('label').string             cnArticle.articleView=itemSingle.find('em').string             cnArticle.title=itemSingle.find('h3',"tracking-ad").string             cnArticle.url=itemSingle.find("h3").find("a").get("href")             print("数据:",cnArticle.num+1,'\t',cnArticle.author,'\t',cnArticle.postTime,'\t',cnArticle.articleView,'\t',cnArticle.title,'\t',cnArticle.url) #######     执行    ######## if __name__ =="__main__":     #要抓取的网页地址'http://blog.csdn.net/?&page={}'.format(i+1),i+1)     url = "http://blog.csdn.net"     cnblog=CsdnUtils()     for i in range(0,2):         cnblog.parsePage(url,i+1) | 
4.执行脚本
| 1 | $ python3 spyder.py | 

