cost_util.py 16 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371372373374375376377378379380381382383384385386387388389390391392393394395396397398399400401402403404405406407408409410411412413414415416417418419420421422423424425426427428429430431432433434435436437438439440441442443444445446447448449450451452453454455456457458459460461462463464465466467468469470471472473474475476477478479480481482483484485486487488489490491492493494495496497498499500501502503504505506507508509510511512513514515516517518519520521522523524525526527528529530531532533534535536537538539540541542543544545546547548
  1. import json
  2. import random
  3. import requests
  4. import time
  5. from datetime import datetime
  6. import pymysql
  7. import logging
  8. from concurrent.futures import ThreadPoolExecutor
  9. from model.DataBaseUtils import MysqlUtils
  10. logging.getLogger().setLevel(logging.WARNING)
  11. from model.ComUtils import *
  12. from model.DateUtils import DateUtils
  13. du = DateUtils()
  14. db = MysqlUtils()
  15. max_workers = 10
  16. count = []
  17. t = du.get_n_days(-10)
  18. def get_adcreatives(account_id,access_token,flag,adc_ids,dt): # 获取创意
  19. url = 'https://api.e.qq.com/v1.1/adcreatives/get'
  20. li =[]
  21. page = 1
  22. while True:
  23. parameters = {
  24. 'access_token': access_token,
  25. 'timestamp': int(time.time()),
  26. 'nonce': str(time.time()) + str(random.randint(0, 999999)),
  27. 'fields': ('campaign_id', 'adcreative_id', 'adcreative_name', 'adcreative_elements', 'promoted_object_type', 'page_type',
  28. 'page_spec', 'link_page_spec', 'universal_link_url', 'promoted_object_id','site_set'),
  29. "filtering": [{
  30. "field": "adcreative_id",
  31. "operator": "IN",
  32. "values": adc_ids.split(',')
  33. }],
  34. "account_id": account_id,
  35. "page": page,
  36. "page_size": 100,
  37. "is_deleted": False
  38. }
  39. for k in parameters:
  40. if type(parameters[k]) is not str:
  41. parameters[k] = json.dumps(parameters[k])
  42. while True:
  43. h = requests.get(url, params=parameters)
  44. # print(h.json())
  45. if h.status_code == 200:
  46. r = h.json()
  47. # print(r)
  48. break
  49. else:
  50. time.sleep(1)
  51. print("爬取失败 等待1s")
  52. if 'data' in r.keys():
  53. is_video=0
  54. for i in r['data']['list']:
  55. # print(i)
  56. if flag=='MP':
  57. if len(i['adcreative_elements'])>0:
  58. d = i['adcreative_elements']
  59. title =d.get('title','')
  60. description = d.get('description', '')
  61. if 'image' in d.keys():
  62. image=d.get('image','')
  63. elif 'image_list' in d.keys():
  64. image =','.join(d.get('image_list'))
  65. elif 'short_video_struct' in d.keys():
  66. image = d['short_video_struct']["short_video1"]
  67. is_video=1
  68. else:
  69. image = ''
  70. else:
  71. title = image=''
  72. li.append((
  73. i['adcreative_id'],i['adcreative_name'],i['campaign_id'],image,title,
  74. i.get('promoted_object_type',''),i.get('page_type',''),
  75. i['page_spec'].get('page_id',''),i.get('promoted_object_id',''),
  76. '',description,'MP',account_id,dt,is_video
  77. ))
  78. else:
  79. if len(i['adcreative_elements'])>0:
  80. d =i['adcreative_elements']
  81. if 'image' in d.keys():
  82. image =d['image']
  83. elif 'element_story' in d.keys():
  84. image= ','.join([x['image'] for x in d['element_story']])
  85. else:
  86. image=''
  87. title =d.get('title','')
  88. description = d.get('description','')
  89. else:
  90. image=title=description=''
  91. li.append(
  92. (
  93. i['adcreative_id'], i['adcreative_name'], i['campaign_id'],image,title,
  94. i.get('promoted_object_type', ''), i.get('page_type', ''),
  95. i['page_spec'].get('page_id', ''), i.get('promoted_object_id', ''),
  96. ','.join(i['site_set']),description,'GDT',account_id,dt,is_video
  97. )
  98. )
  99. total_page = r['data']['page_info']['total_page']
  100. if total_page > page:
  101. page += 1
  102. else:
  103. break
  104. else:
  105. break
  106. if len(li)>0:
  107. print(f"{account_id}有创意:",len(li))
  108. sql='replace into adcreative_info values(%s,%s,%s,%s,%s,%s,%s,%s,%s,%s,%s,%s,%s,%s,%s) '
  109. db.quchen_text.executeMany(sql,li)
  110. def images_info_get(account_id,access_token,image_ids): # 获取图片信息
  111. fields = ('image_id','width','height','file_size','signature','preview_url')
  112. interface = 'images/get'
  113. url = 'https://api.e.qq.com/v1.3/' + interface
  114. page = 1
  115. li = []
  116. while True:
  117. common_parameters = {
  118. 'access_token': access_token,
  119. 'timestamp': int(time.time()),
  120. 'nonce': str(time.time()) + str(random.randint(0, 999999)),
  121. 'fields': fields
  122. }
  123. parameters = {
  124. "account_id": account_id,
  125. "filtering":[{
  126. "field": "image_id",
  127. "operator": "IN",
  128. "values": image_ids.split(',')
  129. }],
  130. "page": page,
  131. "page_size": 100
  132. }
  133. parameters.update(common_parameters)
  134. for k in parameters:
  135. if type(parameters[k]) is not str:
  136. parameters[k] = json.dumps(parameters[k])
  137. while True:
  138. h = requests.get(url, params=parameters)
  139. # print(h.text)
  140. if h.status_code == 200:
  141. r = h.json()
  142. break
  143. else:
  144. time.sleep(1)
  145. print("请求出错 等待1s..")
  146. if 'data' in r.keys():
  147. li.extend(r['data']['list'])
  148. total_page = r['data']['page_info']['total_page']
  149. if total_page > page:
  150. page += 1
  151. else:
  152. break
  153. # print(li)
  154. data = []
  155. for i in li:
  156. data.append((i['image_id'],i['width'],i['height'],i['signature'],i['preview_url']))
  157. # print(data)
  158. print(f"{account_id} 有图片:", li.__len__())
  159. if li.__len__() > 0:
  160. sql="replace into image_info value (%s,%s,%s,%s,%s)"
  161. db.quchen_text.executeMany(sql, data)
  162. db.close()
  163. def ad_info():
  164. accounts = db.quchen_text.getData("""
  165. select account_id,access_token,name channel,'GDT' type from advertiser_qq where name !='' or name is not null
  166. union
  167. select account_id,access_token,name channel,'MP' type from advertiser_vx where name !='' or name is not null
  168. """)
  169. total_data =[]
  170. executor = ThreadPoolExecutor(max_workers=max_workers)
  171. for i in accounts:
  172. # print(i)
  173. account_id =i[0]
  174. access_token = i[1]
  175. type = i[3]
  176. executor.submit(get_ad_info,account_id, access_token,type,total_data)
  177. executor.shutdown()
  178. print(len(total_data))
  179. if len(total_data)>0:
  180. sql="replace into ad_info values(%s,%s,%s,%s,%s,%s,%s) "
  181. db.quchen_text.executeMany(sql, total_data)
  182. """获取广告基础信息"""
  183. def get_ad_info(account_id, access_token, flag,ad_ids,dt):
  184. path = 'ads/get'
  185. fields = ('ad_id', 'ad_name', 'adcreative_id', 'adgroup_id', 'campaign_id')
  186. url = 'https://api.e.qq.com/v1.3/' + path
  187. li = []
  188. page = 1
  189. while True:
  190. parameters = {
  191. 'access_token': access_token,
  192. 'timestamp': int(time.time()),
  193. 'nonce': str(time.time()) + str(random.randint(0, 999999)),
  194. 'fields': fields,
  195. "filtering": [{
  196. "field": "ad_id",
  197. "operator": "IN",
  198. "values":
  199. ad_ids.split(',')
  200. }],
  201. "account_id": account_id,
  202. "page": page,
  203. "page_size": 100,
  204. "is_deleted": False
  205. }
  206. for k in parameters:
  207. if type(parameters[k]) is not str:
  208. parameters[k] = json.dumps(parameters[k])
  209. while True:
  210. r = requests.get(url, params=parameters).json()
  211. code = r['code']
  212. if code == 11017:
  213. time.sleep(61)
  214. else:
  215. break
  216. # print(r)
  217. total_page = r['data']['page_info']['total_page']
  218. if page > total_page:
  219. break
  220. else:
  221. page += 1
  222. if r.get("data"):
  223. for i in r['data']['list']:
  224. li.append((str(i['ad_id']), i['ad_name'], i['adcreative_id'], i['campaign_id'], i['adgroup_id'],
  225. account_id, flag, dt))
  226. if li.__len__()>0:
  227. print(f"{account_id}有广告:",li.__len__())
  228. sql = "replace into ad_info values(%s,%s,%s,%s,%s,%s,%s,%s) "
  229. db.quchen_text.executeMany(sql, li)
  230. db.close()
  231. def get_ad_cost_day(account_id,access_token,flag,st,et):
  232. if flag == 'MP':
  233. ad_cost_day_mp(account_id,access_token, st, et)
  234. else:
  235. ad_cost_day_gdt(account_id,access_token, st, et)
  236. def ad_cost_day_gdt(account_id,access_token,st,et):
  237. url = 'https://api.e.qq.com/v1.3/daily_reports/get'
  238. fields = ('date', 'ad_id','adgroup_id', 'cost', 'view_count', 'ctr', 'follow_count','web_order_count','order_amount')
  239. li = []
  240. page = 1
  241. while True:
  242. parameters = {
  243. 'access_token': access_token,
  244. 'timestamp': int(time.time()),
  245. 'nonce': str(time.time()) + str(random.randint(0, 999999)),
  246. 'fields': fields,
  247. "account_id": account_id,
  248. "group_by" : ['ad_id','date'],
  249. "level": 'REPORT_LEVEL_AD',
  250. "page": page,
  251. "page_size": 1000,
  252. "date_range": {
  253. "start_date": st,
  254. "end_date": et
  255. }
  256. }
  257. for k in parameters:
  258. if type(parameters[k]) is not str:
  259. parameters[k] = json.dumps(parameters[k])
  260. while True:
  261. r = requests.get(url, params=parameters).json()
  262. # print(r)
  263. code =r['code']
  264. if code==11017:
  265. time.sleep(61)
  266. else:
  267. break
  268. if r.get("data"):
  269. for i in r['data']['list']:
  270. if i['cost']>0:
  271. li.append(
  272. (
  273. i['date'], i['ad_id'],i['adgroup_id'], i['cost']/100, i['view_count'], i['ctr']*i['view_count'],
  274. i['follow_count'],i['web_order_count'],i['order_amount']/100,account_id,'GDT'
  275. )
  276. )
  277. total_page = r['data']['page_info']['total_page']
  278. if page >= total_page:
  279. break
  280. else:
  281. page += 1
  282. # print(li)
  283. if len(li) > 0:
  284. print(f"{account_id} have ad cost :{len(li)} ")
  285. db.quchen_text.executeMany('replace into ad_cost_day values(%s,%s,%s,%s,%s,%s,%s,%s,%s,%s,%s)', li)
  286. db.close()
  287. def ad_cost_day_mp(account_id,access_token,st,et):
  288. url = 'https://api.e.qq.com/v1.3/daily_reports/get'
  289. fields = ('date', 'ad_id','adgroup_id', 'cost', 'view_count', 'valid_click_count', 'official_account_follow_count','order_count','order_amount')
  290. li = []
  291. page = 1
  292. while True:
  293. parameters = {
  294. 'access_token': access_token,
  295. 'timestamp': int(time.time()),
  296. 'nonce': str(time.time()) + str(random.randint(0, 999999)),
  297. 'fields': fields,
  298. "account_id": account_id,
  299. "level": 'REPORT_LEVEL_AD_WECHAT',
  300. "page": page,
  301. "page_size": 1000,
  302. "date_range": {
  303. "start_date": st,
  304. "end_date": et
  305. }
  306. }
  307. for k in parameters:
  308. if type(parameters[k]) is not str:
  309. parameters[k] = json.dumps(parameters[k])
  310. while True:
  311. r = requests.get(url, params=parameters).json()
  312. # print(r['data']['list'])
  313. # import pandas as pd
  314. # print(pd.DataFrame(r['data']['list']))
  315. code = r['code']
  316. if code == 11017:
  317. time.sleep(61)
  318. else:
  319. break
  320. if r.get("data"):
  321. for i in r['data']['list']:
  322. if i['cost']>0:
  323. li.append(
  324. (
  325. i['date'],i['ad_id'],i['adgroup_id'],i['cost']/100,i['view_count'],i['valid_click_count'],
  326. i['official_account_follow_count'],i['order_count'],i['order_amount']/100,account_id,'MP'
  327. )
  328. )
  329. total_page = r['data']['page_info']['total_page']
  330. if page >=total_page:
  331. break
  332. else:
  333. page += 1
  334. # print(li)
  335. # exit()
  336. if len(li) > 0:
  337. print(f"{account_id} have ad cost :{len(li)} ")
  338. db.quchen_text.executeMany('replace into ad_cost_day values(%s,%s,%s,%s,%s,%s,%s,%s,%s,%s,%s)', li)
  339. db.close()
  340. def daily_reports_get(access_token, account_id, level, start_date, end_date, fields): # 获取wx投放计划日报数据
  341. interface = 'daily_reports/get'
  342. url = 'https://api.e.qq.com/v1.3/' + interface
  343. common_parameters = {
  344. 'access_token': access_token,
  345. 'timestamp': int(time.time()),
  346. 'nonce': str(time.time()) + str(random.randint(0, 999999)),
  347. 'fields': fields
  348. }
  349. parameters = {
  350. "account_id": account_id,
  351. "level": level,
  352. "date_range":
  353. {
  354. "start_date": start_date,
  355. "end_date": end_date
  356. },
  357. "page": 1,
  358. "page_size": 1000,
  359. "fields":
  360. [
  361. ]
  362. }
  363. parameters.update(common_parameters)
  364. for k in parameters:
  365. if type(parameters[k]) is not str:
  366. parameters[k] = json.dumps(parameters[k])
  367. while True:
  368. r = requests.get(url, params=parameters)
  369. if r.status_code == 200:
  370. break
  371. else:
  372. time.sleep(1)
  373. print("请求出错 等待1s..")
  374. return r.json()
  375. def daily_qq_reports_get(access_token, account_id, compaign_id, level, start_date, end_date, fields): # 获取gdt投放计划日报数据
  376. interface = 'daily_reports/get'
  377. url = 'https://api.e.qq.com/v1.1/' + interface
  378. common_parameters = {
  379. 'access_token': access_token,
  380. 'timestamp': int(time.time()),
  381. 'nonce': str(time.time()) + str(random.randint(0, 999999)),
  382. 'fields': fields
  383. }
  384. parameters = {
  385. "account_id": account_id,
  386. "filtering":
  387. [
  388. {
  389. "field": "campaign_id",
  390. "operator": "EQUALS",
  391. "values":
  392. [
  393. compaign_id
  394. ]
  395. }
  396. ],
  397. "level": level,
  398. "date_range":
  399. {
  400. "start_date": start_date,
  401. "end_date": end_date
  402. },
  403. "page": 1,
  404. "page_size": 1000,
  405. "fields":
  406. [
  407. ]
  408. }
  409. parameters.update(common_parameters)
  410. for k in parameters:
  411. if type(parameters[k]) is not str:
  412. parameters[k] = json.dumps(parameters[k])
  413. r = requests.get(url, params=parameters)
  414. return r.json()
  415. def mysql_insert_adcreative(data):
  416. db = pymysql.connect('rm-bp1c9cj79872tx3aaro.mysql.rds.aliyuncs.com', 'superc', 'Cc719199895', 'quchen_text')
  417. cursor = db.cursor()
  418. sql = 'replace into adcreative (campaign_id,adcreative_id,adcreative_name,image_id,title,promoted_object_type,page_type,page_id,link_page_id,promoted_object_id) values (%s,%s,%s,%s,%s,%s,%s,%s,%s,%s)'
  419. try:
  420. cursor.executemany(sql, data)
  421. db.commit()
  422. print('insert [adcreative] ', len(data))
  423. except:
  424. db.rollback()
  425. print('insert [adcreative] defeat')
  426. if __name__ == '__main__':
  427. account_id = 19206910
  428. access_token = '89079ccc8db047b078a0108e36a7e276'
  429. #
  430. account_id2 = 14709511
  431. access_token2 = 'e87f7b6f860eaeef086ddcc9c3614678'
  432. get_ad_cost_day(account_id,access_token,'MP','2021-04-09','2021-04-09')
  433. # get_adcreatives(account_id,access_token,'MP','3187867673','2021-04-09')