import osimport jsonimport datetimeimport requestsimport bs4 as bsimport numpy as npimport pandas as pdimport urllib.request# from IPython.core.display import HTMLdef generate_path(path):folder = os.path.exists(path)if not folder:os.makedirs(path)def video(video_url,file_name):return urllib.request.urlretrieve(video_url,file_name)def cover(cover_url,file_name):return urllib.request.urlretrieve(cover_url,file_name)def time(timestamp):return str(datetime.datetime.fromtimestamp(timestamp))api = requests.get('https://raw.githubusercontent.com/xjincomm/Douyin/master/Douyin%20Trending%20API.txt').textre=requests.get(api)soup=bs.BeautifulSoup(re.content,'html.parser')last_update = json.loads(soup.text)['data']['active_time']trending_data = json.loads(soup.text)['data']['word_list']trend = pd.DataFrame(trending_data)date=last_update.split(' ')[0]Word_cover=[]for i in trend['word_cover']:if type(i)==dict:Word_cover.append(i['url_list'][0])else:Word_cover.append(None)trend = trend.drop(columns = ['word_cover','challenge_id'])# word means the title of the topicgenerate_path('./trend')#trend['cover']=['<img src="'+ str(i) + '" width="60" >' for i in Word_cover]#trend_visual = HTML(trend.head(3).to_html(escape=False ,formatters=trend['cover']))#trend.to_html('./trend/trend_'+last_update+'.html', escape=False)trend.to_csv('./trend/trend_'+date+'.csv',encoding = 'utf-8-sig', index = False)#trend_visualdef scraper(topic):generate_path('./'+topic)topic_api='https://aweme-hl.snssdk.com/aweme/v1/hot/search/video/list/?hotword='re=requests.get(topic_api+topic)soup=bs.BeautifulSoup(re.content,'html.parser')data = json.loads(soup.text)data = data['aweme_list']desc = [info['desc'] for info in data]time_stamp = [info['create_time'] for info in data]create_time = [time(info['create_time']) for info in data]nickname = [info['author']['nickname'] for info in data]verify = [info['author']['custom_verify'] for info in data]share_count = [info['statistics']['share_count'] for info in data]forward_count = [info['statistics']['forward_count'] for info in data]like_count = [info['statistics']['digg_count'] for info in data]comment_count = [info['statistics']['comment_count'] for info in data]download_count = [info['statistics']['download_count'] for info in data]cover_url = [info['video']['cover']['url_list'][0] for info in data]cover_visual = ['<img src="'+ url + '" width="100" >' for url in cover_url]video_url = []for info in data:try:video_url.append([i for i in info['video']['play_addr']['url_list']][0])except:video_url.append(None)df=pd.DataFrame({'desc':desc,'nickname':nickname,'verify':verify,'time_stamp':time_stamp,'create_time':create_time,'share_count':share_count,'forward_count':forward_count,'like_count':like_count,'comment_count':comment_count,'download_count':download_count,'video_url':video_url,'cover_visual':cover_visual})df.to_csv('./'+topic+'/'+topic+'.csv',encoding='utf-8-sig',index=False)#df.to_html('./'+topic+'/'+topic+'.html',escape=False)#video_visual = HTML(df.to_html(escape=False ,formatters=df['cover_visual']))for num in range(0,len(data)):try:video(df['video_url'][num],'./'+topic+'/'+str(df['time_stamp'][num])+'.mp4')print('topic: '+topic+', video #'+str(num)+': '+str(df['time_stamp'][num])+'......Succeeded')except:print('topic: '+topic+', video #'+str(num)+': '+str(df['time_stamp'][num])+'......Failed')continuedef douyin_trend():for i in trend['word']:scraper(i)def douyin_topic(topic):scraper(topic)
Đăng nhận xét