总共四个文件,新手,大佬勿喷,欢迎指导
1,爬取首页所有栏目并入MySQL,这个是单独运行的,因为总共就16个栏目,所以只是练手,并没有在此之后继续爬后续的文章内容
# -*- coding: UTF-8 -*-import refrom lxml import htmlimport pymysqlfrom Tools import ToolsMethodsconnect = pymysql.connect(host='localhost', user='root', passwd='44124985', db='acgcloumn', port=3306, charset='utf8')# 获取游标cursor = connect.cursor()def getCloumnData(tree): first_titleArr = tree.xpath('//div[@class="topbar-bg"]/div/div/ul/li/a/text()') first_urlList = tree.xpath('//div[@class="topbar-bg"]/div/div/ul/li/a/@href') pid = 0 #插入主栏目 cursor.execute("use acgcloumn") for i in range(len(first_titleArr)): cursor.execute("insert into cloumn_tab (name,weburl,pid) values (%s,%s,%s)",(first_titleArr[i],first_urlList[i],0)) connect.commit() sub_countArr = tree.xpath('//div[@class="topbar-bg"]/div/div/ul/li/ul[@class="ui-nav-submain"]/li[contains(@class,"ui-nav-subitem")]/a/text()') sub_titleArr = tree.xpath('//div[@class="topbar-bg"]/div/div/ul/li/ul/li/a/text()') sub_urlList = tree.xpath('//div[@class="topbar-bg"]/div/div/ul/li/ul/li/a/@href') cursor.execute("use acgcloumn") for j in range(len(sub_titleArr)): cursor.execute("insert into cloumn_tab (name,weburl) values (%s,%s)",(sub_titleArr[j],sub_urlList[j])) connect.commit()if __name__ == '__main__': toolsObj = ToolsMethods() mytree = toolsObj.getHtmlTrees('http://acg.178.com/') getCloumnData(mytree) connect.close()复制代码
2.这个是爬栏目进来后的页面,这里运行起来就可以把整个这个栏目爬光光
# -*- coding: UTF-8 -*-import refrom lxml import htmlfrom lxml import etreeimport pymysqlimport osimport threadingfrom Tools import ToolsMethodsfrom ContentDemo import ContentFileMethodsAllTitleImageList = []ALLNextUrlList = []DownLoadPicSwithc = 1finish = FalseTitlePagePid = 1#这里就手填吧,标题页的pid是子栏目的id,总共就16个#图片处理区def getTotalPages(url): global finish #获取一共有多少页 tool_obj = ToolsMethods() tree = tool_obj.getHtmlTrees(url) maxPageList = tree.xpath('//div[@class="splitpage"]/div/a/@href') for e in maxPageList: print e noLoaded = tool_obj.stringIntoMD5(e) if noLoaded == True: getPageTitleUrlAndImage(e,maxPageList) else: print '这个加载过' def getPageTitleUrlAndImage(url,arr): #获取每个页面的跳转路径和图片路径 tool_obj = ToolsMethods() tree = tool_obj.getHtmlTrees(url) allNextUrlList = [] allTitleImgList = [] nextUrlList = tree.xpath('//div[@class="imgtextlist"]/ul/li/div/a/@href') tmpImgList = tree.xpath('//div[@class="imgtextlist"]/ul/li/div/a/img/@src') ALLNextUrlList.extend(nextUrlList) AllTitleImageList.extend(tmpImgList) if url == arr[-1]:#如果传进来的url是当前数组的最后一位就加载这个url看是否还有后续的url getTotalPages(url) def savePicIn(): if (DownLoadPicSwithc == 1): print '开始下载图片' path = 'titleImg' os.makedirs('f:\\acgTitleImg\\'+path+'\\') toolsObj = ToolsMethods() for picUrl in AllTitleImageList: print picUrl mat = re.search(r'([0-9]+\.jpg|[0-9]+\.png|[0-9]+\.gif)',str(picUrl))#生成图片名称 if mat: picName = mat.group(0) toolsObj.downLoadPicture(picUrl,picName,path) print '图片下载完成' #标题跳转处理区def getArticlePageContent(list_url,next_pid): content_obj = ContentFileMethods() content_obj.receivePageUrl(list_url,next_pid)if __name__ == '__main__': toolsObj = ToolsMethods() urllist,idlist= toolsObj.getTitleUrl()#返回ID作为标题页的pid,然后idlist的下标是内容页的pid,注意下标0其实pid是1 #每个子栏目下面的所有文章标题页,首先要获取最大页数,然后获取所有的标题URL getTotalPages(urllist[0]) # for url in urllist: # getTotalPages(url) #拿到了跳转链接和图片,下载图片存入本地 # savePicIn()#先不下载图片以防被屏蔽,另外标题页的图片没有入库,标题页也没有爬内容源码,这行基本不用 print '图片保存完成' getArticlePageContent(ALLNextUrlList,1)#这里1是写死因为urllist[0],到时候从for里取就ok print '全部完成'复制代码
3.接第二步,对文章详情页进行解析并入MySQL
# -*- coding: UTF-8 -*-import refrom lxml import htmlimport pymysqlfrom Tools import ToolsMethodsconnect = pymysql.connect(host='localhost', user='root', passwd='44124985', db='acgcloumn', port=3306, charset='utf8')# 获取游标cursor = connect.cursor()def getCloumnData(tree): first_titleArr = tree.xpath('//div[@class="topbar-bg"]/div/div/ul/li/a/text()') first_urlList = tree.xpath('//div[@class="topbar-bg"]/div/div/ul/li/a/@href') pid = 0 #插入主栏目 cursor.execute("use acgcloumn") for i in range(len(first_titleArr)): cursor.execute("insert into cloumn_tab (name,weburl,pid) values (%s,%s,%s)",(first_titleArr[i],first_urlList[i],0)) connect.commit() sub_countArr = tree.xpath('//div[@class="topbar-bg"]/div/div/ul/li/ul[@class="ui-nav-submain"]/li[contains(@class,"ui-nav-subitem")]/a/text()') sub_titleArr = tree.xpath('//div[@class="topbar-bg"]/div/div/ul/li/ul/li/a/text()') sub_urlList = tree.xpath('//div[@class="topbar-bg"]/div/div/ul/li/ul/li/a/@href') cursor.execute("use acgcloumn") for j in range(len(sub_titleArr)): cursor.execute("insert into cloumn_tab (name,weburl) values (%s,%s)",(sub_titleArr[j],sub_urlList[j])) connect.commit()if __name__ == '__main__': toolsObj = ToolsMethods() mytree = toolsObj.getHtmlTrees('http://acg.178.com/') getCloumnData(mytree) connect.close()复制代码
4.这个是自己用的工具类,很烂,大佬轻喷,可以不用看,欢迎指导
# -*- coding: UTF-8 -*-import randomimport reimport requestsimport pymysqlimport hashlibimport osfrom lxml import htmlHaveLoadList = [] class ToolsMethods(object): def getHeads(self): agentList = ['Mozilla/5.0 (Macintosh; U; Intel Mac OS X 10_6_8; en-us) AppleWebKit/534.50 (KHTML, like Gecko) Version/5.1 Safari/534.50','Mozilla/5.0 (Windows; U; Windows NT 6.1; en-us) AppleWebKit/534.50 (KHTML, like Gecko) Version/5.1 Safari/534.50','Mozilla/5.0 (Windows NT 10.0; WOW64; rv:38.0) Gecko/20100101 Firefox/38.0','Mozilla/5.0 (Windows NT 10.0; WOW64; Trident/7.0; .NET4.0C; .NET4.0E; .NET CLR 2.0.50727; .NET CLR 3.0.30729; .NET CLR 3.5.30729; InfoPath.3; rv:11.0) like Gecko','Mozilla/5.0 (compatible; MSIE 9.0; Windows NT 6.1; Trident/5.0','User-Agent: Mozilla/5.0 (Macintosh; Intel Mac OS X 10.12; rv:56.0) Gecko/20100101 Firefox/56.0','User-Agent: Mozilla/5.0 (Linux; Android 6.0; Nexus 5 Build/MRA58N) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/62.0.3202.89 Mobile Safari/537.36','User-Agent: Mozilla/5.0 (Macintosh; Intel Mac OS X 10_12_6) AppleWebKit/604.3.5 (KHTML, like Gecko) Version/11.0.1 Safari/604.3.5','User-Agent:Mozilla/5.0 (Windows NT 6.3; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/62.0.3202.75 Safari/537.36 Request Payload','User-Agent: Mozilla/5.0 (Windows NT 6.3; Win64; x64; rv:56.0)','User-Agent Mozilla/5.0 (Windows NT 6.3; WOW64; Trident/7.0; rv:11.0) like Gecko','User-Agent:Mozilla/5.0 (Macintosh; U; Intel Mac OS X 10_6_8; en-us) AppleWebKit/534.50 (KHTML, like Gecko) Version/5.1 Safari/534.50','User-Agent:Mozilla/5.0 (Windows; U; Windows NT 6.1; en-us) AppleWebKit/534.50 (KHTML, like Gecko) Version/5.1 Safari/534.50','User-Agent:Mozilla/4.0 (compatible; MSIE 8.0; Windows NT 6.0; Trident/4.0)','User-Agent:Mozilla/4.0 (compatible; MSIE 7.0; Windows NT 6.0)','User-Agent: Mozilla/4.0 (compatible; MSIE 6.0; Windows NT 5.1)','User-Agent: Mozilla/5.0 (Macintosh; Intel Mac OS X 10.6; rv:2.0.1) Gecko/20100101 Firefox/4.0.1','User-Agent:Mozilla/5.0 (Windows NT 6.1; rv:2.0.1) Gecko/20100101 Firefox/4.0.1','User-Agent:Opera/9.80 (Macintosh; Intel Mac OS X 10.6.8; U; en) Presto/2.8.131 Version/11.11','User-Agent:Opera/9.80 (Windows NT 6.1; U; en) Presto/2.8.131 Version/11.11','User-Agent: Mozilla/5.0 (Macintosh; Intel Mac OS X 10_7_0) AppleWebKit/535.11 (KHTML, like Gecko) Chrome/17.0.963.56 Safari/535.11','User-Agent: Mozilla/4.0 (compatible; MSIE 7.0; Windows NT 5.1; Maxthon 2.0)','User-Agent: Mozilla/4.0 (compatible; MSIE 7.0; Windows NT 5.1; TencentTraveler 4.0)','User-Agent: Mozilla/4.0 (compatible; MSIE 7.0; Windows NT 5.1)','User-Agent: Mozilla/4.0 (compatible; MSIE 7.0; Windows NT 5.1; The World)','User-Agent: Mozilla/4.0 (compatible; MSIE 7.0; Windows NT 5.1; Trident/4.0; SE 2.X MetaSr 1.0; SE 2.X MetaSr 1.0; .NET CLR 2.0.50727; SE 2.X MetaSr 1.0)','User-Agent: Mozilla/4.0 (compatible; MSIE 7.0; Windows NT 5.1; 360SE)','User-Agent: Mozilla/4.0 (compatible; MSIE 7.0; Windows NT 5.1)','User-Agent: Mozilla/4.0 (compatible; MSIE 7.0; Windows NT 5.1; Avant Browser)','User-Agent:Mozilla/5.0 (iPhone; U; CPU iPhone OS 4_3_3 like Mac OS X; en-us) AppleWebKit/533.17.9 (KHTML, like Gecko) Version/5.0.2 Mobile/8J2 Safari/6533.18.5','User-Agent:Mozilla/5.0 (iPod; U; CPU iPhone OS 4_3_3 like Mac OS X; en-us) AppleWebKit/533.17.9 (KHTML, like Gecko) Version/5.0.2 Mobile/8J2 Safari/6533.18.5','User-Agent:Mozilla/5.0 (iPad; U; CPU OS 4_3_3 like Mac OS X; en-us) AppleWebKit/533.17.9 (KHTML, like Gecko) Version/5.0.2 Mobile/8J2 Safari/6533.18.5'] random_num = random.randint(1,len(agentList)) user_agent = agentList[random_num-1] # myHeaders = {'User-Agent': user_agent} myHeaders = { 'Host' : 'acg.178.com', 'User-Agent' : user_agent, 'Accept':'text/html,application/xhtml+xml,application/xml;q=0.9,*/*;q=0.8', 'Connection':'keep-alive' } return myHeaders def getTitleUrl(object): connect = pymysql.connect( host='localhost', user='root', passwd='44124985', db='acgcloumn', port=3306, charset='utf8' ) # 获取游标 cursor = connect.cursor() cursor.execute("use acgcloumn") sql = 'select weburl,id from cloumn_tab where id > 8' cursor.execute(sql) result = cursor.fetchall() urlList = [] idlist = [] for a in result: urlList.append(a[0]) idlist.append(a[1]) connect.commit() connect.close() return (urlList,idlist) def getHtmlTrees(self,pageUrl): toolobj = ToolsMethods() headerCls = toolobj.getHeads() response = requests.get(pageUrl,headers=headerCls) statueCode = response.status_code page_content = response.text if statueCode == 404: print 404 return 404 tree = html.fromstring(page_content) return tree def stringIntoMD5(self,str): m = hashlib.md5() m.update(str) md5Str = m.hexdigest() if md5Str not in HaveLoadList: HaveLoadList.append(md5Str) return True else: return False def downLoadPicture(self,picUrl,picName,path): # urllib.urlretrieve(picUrl,'f:\\acgTitleImg\\'+path+'\\'+picName)#urllib一句代码下载 tools_obj = ToolsMethods() headers = tools_obj.getHeads() path_pic = 'f:\\acgTitleImg\\'+path+'\\'+picName picResponse = requests.get(picUrl) if picResponse.status_code == 200: open(path_pic, 'wb').write(picResponse.content)复制代码