cost_util.py 20 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371372373374375376377378379380381382383384385386387388389390391392393394395396397398399400401402403404405406407408409410411412413414415416417418419420421422423424425426427428429430431432433434435436437438439440441442443444445446447448449450451452453454455456457458459460461462463464465466467468469470471472473474475476477478479480481482483484485486487488489490491492493494495496497498499500501502503504505506507508509510511512513514515516517518519520521522523524525526527528529530531532533534535536537538539540541542543544545546547548549550551552553554555556557558559560561562563564565566567568569570571572573574575576577578579580581582583584585586587588589590591592593594595596597598599600601602603604605606607608609610611612613614615616617618619620621622623624625626627628629630631632633634635636637638639640641642643644645646647648649650651652653654655656657658659660661662663664665666667668669670671672673674675676
  1. import json
  2. import random
  3. import requests
  4. import time
  5. from datetime import datetime
  6. import pymysql
  7. import logging
  8. from concurrent.futures import ThreadPoolExecutor
  9. from model.DataBaseUtils import MysqlUtils
  10. logging.getLogger().setLevel(logging.WARNING)
  11. from model.ComUtils import *
  12. from model.DateUtils import DateUtils
  13. du = DateUtils()
  14. db = MysqlUtils()
  15. max_workers = 10
  16. count = []
  17. t = du.get_n_days(-10)
  18. def get_campaign(account_id, access_token, flag,campaign_ids,dt):
  19. path = 'campaigns/get'
  20. fields = ('campaign_id', 'campaign_name', 'configured_status', 'campaign_type', 'promoted_object_type',
  21. 'daily_budget','budget_reach_date','created_time','last_modified_time','speed_mode','is_deleted')
  22. url = 'https://api.e.qq.com/v1.3/' + path
  23. li = []
  24. page = 1
  25. while True:
  26. parameters = {
  27. 'access_token': access_token,
  28. 'timestamp': int(time.time()),
  29. 'nonce': str(time.time()) + str(random.randint(0, 999999)),
  30. 'fields': fields,
  31. "filtering": [{
  32. "field": "campaign_id",
  33. "operator": "IN",
  34. "values":
  35. campaign_ids.split(',')
  36. }],
  37. "account_id": account_id,
  38. "page": page,
  39. "page_size": 100,
  40. "is_deleted": False
  41. }
  42. for k in parameters:
  43. if type(parameters[k]) is not str:
  44. parameters[k] = json.dumps(parameters[k])
  45. while True:
  46. r = requests.get(url, params=parameters).json()
  47. code = r['code']
  48. if code == 11017:
  49. time.sleep(61)
  50. else:
  51. break
  52. # print(r)
  53. total_page = r['data']['page_info']['total_page']
  54. if page > total_page:
  55. break
  56. else:
  57. page += 1
  58. if r.get("data"):
  59. for i in r['data']['list']:
  60. li.append((str(i['campaign_id']), i['campaign_name'], i['configured_status'], i['campaign_type'],
  61. i['promoted_object_type'],i['daily_budget'],i.get('budget_reach_date'),
  62. DateUtils.stamp_to_str(i['created_time']),
  63. DateUtils.stamp_to_str(i['last_modified_time']),i.get('speed_mode'),i.get('is_deleted'),
  64. account_id, flag, dt))
  65. # print(li)
  66. """mp 没有 speed_mode,is_deleted,budget_reach_date"""
  67. if li.__len__() > 0:
  68. print(f"{account_id}有计划:", li.__len__())
  69. sql = "replace into campaign_info values(%s,%s,%s,%s,%s,%s,%s,%s,%s,%s,%s,%s,%s,%s)"
  70. db.quchen_text.executeMany(sql, li)
  71. db.close()
  72. def get_adcreatives(account_id,access_token,flag,adc_ids,dt): # 获取创意
  73. url = 'https://api.e.qq.com/v1.1/adcreatives/get'
  74. li =[]
  75. page = 1
  76. while True:
  77. parameters = {
  78. 'access_token': access_token,
  79. 'timestamp': int(time.time()),
  80. 'nonce': str(time.time()) + str(random.randint(0, 999999)),
  81. 'fields': ('campaign_id', 'adcreative_id', 'adcreative_name', 'adcreative_elements', 'promoted_object_type', 'page_type',
  82. 'page_spec', 'link_page_spec', 'universal_link_url', 'promoted_object_id','site_set'),
  83. "filtering": [{
  84. "field": "adcreative_id",
  85. "operator": "IN",
  86. "values": adc_ids.split(',')
  87. }],
  88. "account_id": account_id,
  89. "page": page,
  90. "page_size": 100,
  91. "is_deleted": False
  92. }
  93. for k in parameters:
  94. if type(parameters[k]) is not str:
  95. parameters[k] = json.dumps(parameters[k])
  96. while True:
  97. h = requests.get(url, params=parameters)
  98. # print(h.json())
  99. if h.status_code == 200:
  100. r = h.json()
  101. # print(r)
  102. break
  103. else:
  104. time.sleep(1)
  105. print("爬取失败 等待1s")
  106. if 'data' in r.keys():
  107. is_video=0
  108. for i in r['data']['list']:
  109. # print(i)
  110. if flag=='MP':
  111. if len(i['adcreative_elements'])>0:
  112. d = i['adcreative_elements']
  113. title =d.get('title','')
  114. description = d.get('description', '')
  115. if 'image' in d.keys():
  116. image=d.get('image','')
  117. elif 'image_list' in d.keys():
  118. image =','.join(d.get('image_list'))
  119. elif 'video' in d.keys():
  120. image = d['video']
  121. is_video=1
  122. else:
  123. image = ''
  124. else:
  125. title = image=''
  126. li.append((
  127. i['adcreative_id'],i['adcreative_name'],i['campaign_id'],image,title,
  128. i.get('promoted_object_type',''),i.get('page_type',''),
  129. i['page_spec'].get('page_id',''),i.get('promoted_object_id',''),
  130. '',description,'MP',account_id,dt,is_video
  131. ))
  132. else:
  133. if len(i['adcreative_elements'])>0:
  134. d =i['adcreative_elements']
  135. if 'image' in d.keys():
  136. image =d['image']
  137. elif 'element_story' in d.keys():
  138. image= ','.join([x['image'] for x in d['element_story']])
  139. else:
  140. image=''
  141. title =d.get('title','')
  142. description = d.get('description','')
  143. else:
  144. image=title=description=''
  145. li.append(
  146. (
  147. i['adcreative_id'], i['adcreative_name'], i['campaign_id'],image,title,
  148. i.get('promoted_object_type', ''), i.get('page_type', ''),
  149. i['page_spec'].get('page_id', ''), i.get('promoted_object_id', ''),
  150. ','.join(i['site_set']),description,'GDT',account_id,dt,is_video
  151. )
  152. )
  153. total_page = r['data']['page_info']['total_page']
  154. if total_page > page:
  155. page += 1
  156. else:
  157. break
  158. else:
  159. break
  160. if len(li)>0:
  161. print(f"{account_id}有创意:",len(li))
  162. sql='replace into adcreative_info values(%s,%s,%s,%s,%s,%s,%s,%s,%s,%s,%s,%s,%s,%s,%s) '
  163. db.quchen_text.executeMany(sql,li)
  164. def images_info_get(account_id,access_token,image_ids): # 获取图片信息
  165. fields = ('image_id','width','height','file_size','signature','preview_url')
  166. interface = 'images/get'
  167. url = 'https://api.e.qq.com/v1.3/' + interface
  168. page = 1
  169. li = []
  170. while True:
  171. common_parameters = {
  172. 'access_token': access_token,
  173. 'timestamp': int(time.time()),
  174. 'nonce': str(time.time()) + str(random.randint(0, 999999)),
  175. 'fields': fields
  176. }
  177. parameters = {
  178. "account_id": account_id,
  179. "filtering":[{
  180. "field": "image_id",
  181. "operator": "IN",
  182. "values": image_ids.split(',')
  183. }],
  184. "page": page,
  185. "page_size": 100
  186. }
  187. parameters.update(common_parameters)
  188. for k in parameters:
  189. if type(parameters[k]) is not str:
  190. parameters[k] = json.dumps(parameters[k])
  191. while True:
  192. h = requests.get(url, params=parameters)
  193. # print(h.text)
  194. if h.status_code == 200:
  195. r = h.json()
  196. break
  197. else:
  198. time.sleep(1)
  199. print("请求出错 等待1s..")
  200. if 'data' in r.keys():
  201. li.extend(r['data']['list'])
  202. total_page = r['data']['page_info']['total_page']
  203. if total_page > page:
  204. page += 1
  205. else:
  206. break
  207. # print(li)
  208. data = []
  209. for i in li:
  210. data.append((i['image_id'],i['width'],i['height'],i['signature'],i['preview_url']))
  211. # print(data)
  212. print(f"{account_id} 有图片:", li.__len__())
  213. if li.__len__() > 0:
  214. sql="replace into image_info value (%s,%s,%s,%s,%s)"
  215. db.quchen_text.executeMany(sql, data)
  216. db.close()
  217. def video_info_get(account_id,access_token,image_ids): # 获取视频信息
  218. fields = ('video_id','width','height','file_size','signature','preview_url')
  219. interface = 'videos/get'
  220. url = 'https://api.e.qq.com/v1.3/' + interface
  221. page = 1
  222. li = []
  223. while True:
  224. common_parameters = {
  225. 'access_token': access_token,
  226. 'timestamp': int(time.time()),
  227. 'nonce': str(time.time()) + str(random.randint(0, 999999)),
  228. 'fields': fields
  229. }
  230. parameters = {
  231. "account_id": account_id,
  232. "filtering":[{
  233. "field": "media_id",
  234. "operator": "IN",
  235. "values": image_ids.split(',')
  236. }],
  237. "page": page,
  238. "page_size": 100
  239. }
  240. parameters.update(common_parameters)
  241. for k in parameters:
  242. if type(parameters[k]) is not str:
  243. parameters[k] = json.dumps(parameters[k])
  244. while True:
  245. h = requests.get(url, params=parameters)
  246. # print(h.text)
  247. if h.status_code == 200:
  248. r = h.json()
  249. break
  250. else:
  251. time.sleep(1)
  252. print("请求出错 等待1s..")
  253. if 'data' in r.keys():
  254. li.extend(r['data']['list'])
  255. total_page = r['data']['page_info']['total_page']
  256. if total_page > page:
  257. page += 1
  258. else:
  259. break
  260. # print(li)
  261. data = []
  262. for i in li:
  263. data.append((i['video_id'],i['width'],i['height'],i['signature'],i['preview_url']))
  264. # print(data)
  265. print(f"{account_id} 有视频:", li.__len__())
  266. if li.__len__() > 0:
  267. sql="replace into video_info value (%s,%s,%s,%s,%s)"
  268. db.quchen_text.executeMany(sql, data)
  269. db.close()
  270. def ad_info():
  271. accounts = db.quchen_text.getData("""
  272. select account_id,access_token,name channel,'GDT' type from advertiser_qq where name !='' or name is not null
  273. union
  274. select account_id,access_token,name channel,'MP' type from advertiser_vx where name !='' or name is not null
  275. """)
  276. total_data =[]
  277. executor = ThreadPoolExecutor(max_workers=max_workers)
  278. for i in accounts:
  279. # print(i)
  280. account_id =i[0]
  281. access_token = i[1]
  282. type = i[3]
  283. executor.submit(get_ad_info,account_id, access_token,type,total_data)
  284. executor.shutdown()
  285. print(len(total_data))
  286. if len(total_data)>0:
  287. sql="replace into ad_info values(%s,%s,%s,%s,%s,%s,%s) "
  288. db.quchen_text.executeMany(sql, total_data)
  289. """获取广告基础信息"""
  290. def get_ad_info(account_id, access_token, flag,ad_ids,dt):
  291. path = 'ads/get'
  292. fields = ('ad_id', 'ad_name', 'adcreative_id', 'adgroup_id', 'campaign_id')
  293. url = 'https://api.e.qq.com/v1.3/' + path
  294. li = []
  295. page = 1
  296. while True:
  297. parameters = {
  298. 'access_token': access_token,
  299. 'timestamp': int(time.time()),
  300. 'nonce': str(time.time()) + str(random.randint(0, 999999)),
  301. 'fields': fields,
  302. "filtering": [{
  303. "field": "ad_id",
  304. "operator": "IN",
  305. "values":
  306. ad_ids.split(',')
  307. }],
  308. "account_id": account_id,
  309. "page": page,
  310. "page_size": 100,
  311. "is_deleted": False
  312. }
  313. for k in parameters:
  314. if type(parameters[k]) is not str:
  315. parameters[k] = json.dumps(parameters[k])
  316. while True:
  317. r = requests.get(url, params=parameters).json()
  318. code = r['code']
  319. if code == 11017:
  320. time.sleep(61)
  321. else:
  322. break
  323. # print(r)
  324. total_page = r['data']['page_info']['total_page']
  325. if page > total_page:
  326. break
  327. else:
  328. page += 1
  329. if r.get("data"):
  330. for i in r['data']['list']:
  331. li.append((str(i['ad_id']), i['ad_name'], i['adcreative_id'], i['campaign_id'], i['adgroup_id'],
  332. account_id, flag, dt))
  333. if li.__len__()>0:
  334. print(f"{account_id}有广告:",li.__len__())
  335. sql = "replace into ad_info values(%s,%s,%s,%s,%s,%s,%s,%s) "
  336. db.quchen_text.executeMany(sql, li)
  337. db.close()
  338. def get_ad_cost_day(account_id,access_token,flag,st,et):
  339. if flag == 'MP':
  340. ad_cost_day_mp(account_id,access_token, st, et)
  341. else:
  342. ad_cost_day_gdt(account_id,access_token, st, et)
  343. def ad_cost_day_gdt(account_id,access_token,st,et):
  344. url = 'https://api.e.qq.com/v1.3/daily_reports/get'
  345. fields = ('date', 'ad_id','adgroup_id', 'cost', 'view_count', 'ctr', 'follow_count','web_order_count','order_amount')
  346. li = []
  347. page = 1
  348. while True:
  349. parameters = {
  350. 'access_token': access_token,
  351. 'timestamp': int(time.time()),
  352. 'nonce': str(time.time()) + str(random.randint(0, 999999)),
  353. 'fields': fields,
  354. "account_id": account_id,
  355. "group_by" : ['ad_id','date'],
  356. "level": 'REPORT_LEVEL_AD',
  357. "page": page,
  358. "page_size": 1000,
  359. "date_range": {
  360. "start_date": st,
  361. "end_date": et
  362. }
  363. }
  364. for k in parameters:
  365. if type(parameters[k]) is not str:
  366. parameters[k] = json.dumps(parameters[k])
  367. while True:
  368. r = requests.get(url, params=parameters).json()
  369. # print(r)
  370. code =r['code']
  371. if code==11017:
  372. time.sleep(61)
  373. else:
  374. break
  375. if r.get("data"):
  376. for i in r['data']['list']:
  377. if i['cost']>0:
  378. li.append(
  379. (
  380. i['date'], i['ad_id'],i['adgroup_id'], i['cost']/100, i['view_count'], i['ctr']*i['view_count'],
  381. i['follow_count'],i['web_order_count'],i['order_amount']/100,account_id,'GDT'
  382. )
  383. )
  384. total_page = r['data']['page_info']['total_page']
  385. if page >= total_page:
  386. break
  387. else:
  388. page += 1
  389. # print(li)
  390. if len(li) > 0:
  391. print(f"{account_id} have ad cost :{len(li)} ")
  392. db.quchen_text.executeMany('replace into ad_cost_day values(%s,%s,%s,%s,%s,%s,%s,%s,%s,%s,%s)', li)
  393. db.close()
  394. def ad_cost_day_mp(account_id,access_token,st,et):
  395. url = 'https://api.e.qq.com/v1.3/daily_reports/get'
  396. fields = ('date', 'ad_id','adgroup_id', 'cost', 'view_count', 'valid_click_count', 'official_account_follow_count','order_count','order_amount')
  397. li = []
  398. page = 1
  399. while True:
  400. parameters = {
  401. 'access_token': access_token,
  402. 'timestamp': int(time.time()),
  403. 'nonce': str(time.time()) + str(random.randint(0, 999999)),
  404. 'fields': fields,
  405. "account_id": account_id,
  406. "level": 'REPORT_LEVEL_AD_WECHAT',
  407. "page": page,
  408. "page_size": 1000,
  409. "date_range": {
  410. "start_date": st,
  411. "end_date": et
  412. }
  413. }
  414. for k in parameters:
  415. if type(parameters[k]) is not str:
  416. parameters[k] = json.dumps(parameters[k])
  417. while True:
  418. r = requests.get(url, params=parameters).json()
  419. # print(r['data']['list'])
  420. # import pandas as pd
  421. # print(pd.DataFrame(r['data']['list']))
  422. code = r['code']
  423. if code == 11017:
  424. time.sleep(61)
  425. else:
  426. break
  427. if r.get("data"):
  428. for i in r['data']['list']:
  429. if i['cost']>0:
  430. li.append(
  431. (
  432. i['date'],i['ad_id'],i['adgroup_id'],i['cost']/100,i['view_count'],i['valid_click_count'],
  433. i['official_account_follow_count'],i['order_count'],i['order_amount']/100,account_id,'MP'
  434. )
  435. )
  436. total_page = r['data']['page_info']['total_page']
  437. if page >=total_page:
  438. break
  439. else:
  440. page += 1
  441. # print(li)
  442. # exit()
  443. if len(li) > 0:
  444. print(f"{account_id} have ad cost :{len(li)} ")
  445. db.quchen_text.executeMany('replace into ad_cost_day values(%s,%s,%s,%s,%s,%s,%s,%s,%s,%s,%s)', li)
  446. db.close()
  447. def daily_reports_get(access_token, account_id, level, start_date, end_date, fields): # 获取wx投放计划日报数据
  448. interface = 'daily_reports/get'
  449. url = 'https://api.e.qq.com/v1.3/' + interface
  450. common_parameters = {
  451. 'access_token': access_token,
  452. 'timestamp': int(time.time()),
  453. 'nonce': str(time.time()) + str(random.randint(0, 999999)),
  454. 'fields': fields
  455. }
  456. parameters = {
  457. "account_id": account_id,
  458. "level": level,
  459. "date_range":
  460. {
  461. "start_date": start_date,
  462. "end_date": end_date
  463. },
  464. "page": 1,
  465. "page_size": 1000,
  466. "fields":
  467. [
  468. ]
  469. }
  470. parameters.update(common_parameters)
  471. for k in parameters:
  472. if type(parameters[k]) is not str:
  473. parameters[k] = json.dumps(parameters[k])
  474. while True:
  475. r = requests.get(url, params=parameters)
  476. if r.status_code == 200:
  477. break
  478. else:
  479. time.sleep(1)
  480. print("请求出错 等待1s..")
  481. return r.json()
  482. def daily_qq_reports_get(access_token, account_id, compaign_id, level, start_date, end_date, fields): # 获取gdt投放计划日报数据
  483. interface = 'daily_reports/get'
  484. url = 'https://api.e.qq.com/v1.1/' + interface
  485. common_parameters = {
  486. 'access_token': access_token,
  487. 'timestamp': int(time.time()),
  488. 'nonce': str(time.time()) + str(random.randint(0, 999999)),
  489. 'fields': fields
  490. }
  491. parameters = {
  492. "account_id": account_id,
  493. "filtering":
  494. [
  495. {
  496. "field": "campaign_id",
  497. "operator": "EQUALS",
  498. "values":
  499. [
  500. compaign_id
  501. ]
  502. }
  503. ],
  504. "level": level,
  505. "date_range":
  506. {
  507. "start_date": start_date,
  508. "end_date": end_date
  509. },
  510. "page": 1,
  511. "page_size": 1000,
  512. "fields":
  513. [
  514. ]
  515. }
  516. parameters.update(common_parameters)
  517. for k in parameters:
  518. if type(parameters[k]) is not str:
  519. parameters[k] = json.dumps(parameters[k])
  520. r = requests.get(url, params=parameters)
  521. return r.json()
  522. def mysql_insert_adcreative(data):
  523. db = pymysql.connect('rm-bp1c9cj79872tx3aaro.mysql.rds.aliyuncs.com', 'superc', 'Cc719199895', 'quchen_text')
  524. cursor = db.cursor()
  525. sql = 'replace into adcreative (campaign_id,adcreative_id,adcreative_name,image_id,title,promoted_object_type,page_type,page_id,link_page_id,promoted_object_id) values (%s,%s,%s,%s,%s,%s,%s,%s,%s,%s)'
  526. try:
  527. cursor.executemany(sql, data)
  528. db.commit()
  529. print('insert [adcreative] ', len(data))
  530. except:
  531. db.rollback()
  532. print('insert [adcreative] defeat')
  533. if __name__ == '__main__':
  534. account_id = 19206910
  535. access_token = '89079ccc8db047b078a0108e36a7e276'
  536. #
  537. account_id2 = 14709511
  538. access_token2 = 'e87f7b6f860eaeef086ddcc9c3614678'
  539. get_ad_cost_day(account_id,access_token,'MP','2021-04-09','2021-04-09')
  540. # get_adcreatives(account_id,access_token,'MP','3187867673','2021-04-09')