import time
import pydash
import base64
import requests
from lxml import etree
from aip import AipFace
from pathlib import Path
#惟一必須填的信息就這三行
APP_ID = "xxxxxxxx"
API_KEY = "xxxxxxxxxxxxxxxx"
SECRET_KEY = "xxxxxxxxxxxxxxxx"
# 過濾顏值閾值,存儲空間大的請隨意
BEAUTY_THRESHOLD = 55
AUTHORIZATION = "oauth c3cef7c66a1843f8b3a9e6a1e3160e20"
# 若是權限錯誤,瀏覽器中打開知乎,在開發者工具複製一個,無需登陸
# 建議最好換一個,由於不知道知乎的反爬蟲策略,若是太多人用同一個,可能會影響程序運行
# 每次請求知乎的討論列表長度,不建議設定太長,注意節操
LIMIT = 5
# 這是話題『美女』的 ID,其是『顏值』(20013528)的父話題
SOURCE = "19552207"
USER_AGENT = "Mozilla/5.0 (Windows NT 5.1) AppleWebKit/534.55.3 (KHTML, like Gecko) Version/5.1.5 Safari/534.55.3"
REFERER = "https://www.zhihu.com/topic/%s/newest" % SOURCE
# 某話題下討論列表請求 url
BASE_URL = "https://www.zhihu.com/api/v4/topics/%s/feeds/timeline_activity"
# 初始請求 url 附帶的請求參數
URL_QUERY = "?include=data%5B%3F%28target.type%3Dtopic_sticky_module%29%5D.target.data%5B%3F%28target.type%3Danswer%29%5D.target.content%2Crelationship.is_authorized%2Cis_author%2Cvoting%2Cis_thanked%2Cis_nothelp%3Bdata%5B%3F%28target.type%3Dtopic_sticky_module%29%5D.target.data%5B%3F%28target.type%3Danswer%29%5D.target.is_normal%2Ccomment_count%2Cvoteup_count%2Ccontent%2Crelevant_info%2Cexcerpt.author.badge%5B%3F%28type%3Dbest_answerer%29%5D.topics%3Bdata%5B%3F%28target.type%3Dtopic_sticky_module%29%5D.target.data%5B%3F%28target.type%3Darticle%29%5D.target.content%2Cvoteup_count%2Ccomment_count%2Cvoting%2Cauthor.badge%5B%3F%28type%3Dbest_answerer%29%5D.topics%3Bdata%5B%3F%28target.type%3Dtopic_sticky_module%29%5D.target.data%5B%3F%28target.type%3Dpeople%29%5D.target.answer_count%2Carticles_count%2Cgender%2Cfollower_count%2Cis_followed%2Cis_following%2Cbadge%5B%3F%28type%3Dbest_answerer%29%5D.topics%3Bdata%5B%3F%28target.type%3Danswer%29%5D.target.content%2Crelationship.is_authorized%2Cis_author%2Cvoting%2Cis_thanked%2Cis_nothelp%3Bdata%5B%3F%28target.type%3Danswer%29%5D.target.author.badge%5B%3F%28type%3Dbest_answerer%29%5D.topics%3Bdata%5B%3F%28target.type%3Darticle%29%5D.target.content%2Cauthor.badge%5B%3F%28type%3Dbest_answerer%29%5D.topics%3Bdata%5B%3F%28target.type%3Dquestion%29%5D.target.comment_count&limit=" + str(
LIMIT)
HEADERS = {
"User-Agent": USER_AGENT,
"Referer": REFERER,
"authorization": AUTHORIZATION
}
def fetch_image(url):
try:
response = requests.get(url, headers=HEADERS)
except Exception as e:
raise e
return response.content
def fetch_activities(url):
try:
response = requests.get(url, headers=HEADERS)
except Exception as e:
raise e
return response.json()
def parser_activities(datums, face_detective):
for data in datums["data"]:
target = data["target"]
if "content" not in target or "question" not in target or "author" not in target:
continue
html = etree.HTML(target["content"])
seq = 0
title = target["question"]["title"]
author = target["author"]["name"]
images = html.xpath("//img/@src")
for image in images:
if not image.startswith("http"):
continue
image_data = fetch_image(image)
score = face_detective(image_data)
if not score:
continue
name = "{}--{}--{}--{}.jpg".format(score, author, title, seq)
seq = seq + 1
path = Path(__file__).parent.joinpath("image").joinpath(name)
try:
f = open(path, "wb")
f.write(image_data)
f.flush()
f.close()
print(path)
time.sleep(2)
except Exception as e:
continue
if not datums["paging"]["is_end"]:
return datums["paging"]["next"]
else:
return None
def init_detective(app_id, api_key, secret_key):
client = AipFace(app_id, api_key, secret_key)
options = {"face_field": "age,gender,beauty,qualities"}
def detective(image):
image = str(base64.b64encode(image), "utf-8")
response = client.detect(str(image), "BASE64", options)
response = response.get("result")
if not response:
return
if (not response) or (response["face_num"] == 0):
return
face_list = response["face_list"]
if pydash.get(face_list, "0.face_probability") < 0.6:
return
if pydash.get(face_list, "0.beauty") < BEAUTY_THRESHOLD:
return
if pydash.get(face_list, "0.gender.type") != "female":
return
score = pydash.get(face_list, "0.beauty")
return score
return detective
def main():
face_detective = init_detective(APP_ID, API_KEY, SECRET_KEY)
url = BASE_URL % SOURCE + URL_QUERY
while url is not None:
datums = fetch_activities(url)
url = parser_activities(datums, face_detective)
time.sleep(5)
if __name__ == '__main__':
main()
更多詳情請參考文章出處知乎高顏值圖片爬取html