import requests
from bs4 import BeautifulSoup
import pymysql
print("连接到mysql服务器...")
db = pymysql.connect("127.0.0.1", "root", "1234", charset="utf8")
print("连接成功!")
cursor = db.cursor()
cursor.execute("use test")
num = 0
headers = {
'User-Agent': 'Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/62.0.3202.89 Safari/537.36 OPR/49.0.2725.47'}
url = "http://maoyan.com/board/4?offset="
for i in range(0, 100, 10):
url = "http://maoyan.com/board/4?offset=" + str(i)
r = requests.get(url, headers=headers)
content = r.text
soup = BeautifulSoup(content, 'lxml')
div_name = soup.find_all(class_='name')
div_star = soup.find_all(class_='star')
div_time = soup.find_all(class_='releasetime')
div_score = soup.find_all(class_='score')
long = len(div_name)
for i in range(0, long):
num = num + 1
number = str(num)
name = div_name[i].get_text()
actor = div_star[i].get_text()
date = div_time[i].get_text()
score = div_score[i].get_text()
insert_name = ("insert into maoyantop100(num,name,actor,date,score)" "values(%s,%s,%s,%s,%s)")
data = (number,name,actor,date,score)
cursor.execute(insert_name, data)
db.commit()
print("第 "+ str(num) +" 条插入完成")