cost_util.py 16 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371372373374375376377378379380381382383384385386387388389390391392393394395396397398399400401402403404405406407408409410411412413414415416417418419420421422423424425426427428429430431432433434435436437438439440441442443444445446447448449450451452453454455456457458459460461462463464465466467468469470471472473474475476477478479480481482483484485486487488489490491492493494495496497498499500501502503504505506507508509510511512513514515516517518519520521522523524525526527528529530531532533534535536537538539540541542543544545
  1. import json
  2. import random
  3. import requests
  4. import time
  5. from datetime import datetime
  6. import pymysql
  7. import logging
  8. from concurrent.futures import ThreadPoolExecutor
  9. from model.DataBaseUtils import MysqlUtils
  10. logging.getLogger().setLevel(logging.WARNING)
  11. from model.ComUtils import *
  12. from model.DateUtils import DateUtils
  13. du = DateUtils()
  14. db = MysqlUtils()
  15. max_workers = 10
  16. count = []
  17. t = du.get_n_days(-10)
  18. def get_adcreatives(account_id,access_token,flag,adc_ids,dt): # 获取创意
  19. url = 'https://api.e.qq.com/v1.1/adcreatives/get'
  20. li =[]
  21. page = 1
  22. while True:
  23. parameters = {
  24. 'access_token': access_token,
  25. 'timestamp': int(time.time()),
  26. 'nonce': str(time.time()) + str(random.randint(0, 999999)),
  27. 'fields': ('campaign_id', 'adcreative_id', 'adcreative_name', 'adcreative_elements', 'promoted_object_type', 'page_type',
  28. 'page_spec', 'link_page_spec', 'universal_link_url', 'promoted_object_id','site_set'),
  29. "filtering": [{
  30. "field": "adcreative_id",
  31. "operator": "IN",
  32. "values": adc_ids.split(',')
  33. }],
  34. "account_id": account_id,
  35. "page": page,
  36. "page_size": 100,
  37. "is_deleted": False
  38. }
  39. for k in parameters:
  40. if type(parameters[k]) is not str:
  41. parameters[k] = json.dumps(parameters[k])
  42. while True:
  43. h = requests.get(url, params=parameters)
  44. # print(h.json())
  45. if h.status_code == 200:
  46. r = h.json()
  47. # print(r)
  48. break
  49. else:
  50. time.sleep(1)
  51. print("爬取失败 等待1s")
  52. if 'data' in r.keys():
  53. for i in r['data']['list']:
  54. # print(i)
  55. if flag=='MP':
  56. if len(i['adcreative_elements'])>0:
  57. d = i['adcreative_elements']
  58. title =d.get('title','')
  59. description = d.get('description', '')
  60. if 'image' in d.keys():
  61. image=d.get('image','')
  62. elif 'image_list' in d.keys():
  63. image =','.join(d.get('image_list'))
  64. else:
  65. image=''
  66. else:
  67. title = image=''
  68. li.append((
  69. i['adcreative_id'],i['adcreative_name'],i['campaign_id'],image,title,
  70. i.get('promoted_object_type',''),i.get('page_type',''),
  71. i['page_spec'].get('page_id',''),i.get('promoted_object_id',''),
  72. '',description,'MP',account_id,dt
  73. ))
  74. else:
  75. if len(i['adcreative_elements'])>0:
  76. d =i['adcreative_elements']
  77. if 'image' in d.keys():
  78. image =d['image']
  79. elif 'element_story' in d.keys():
  80. image= ','.join([x['image'] for x in d['element_story']])
  81. else:
  82. image=''
  83. title =d.get('title','')
  84. description = d.get('description','')
  85. else:
  86. image=title=description=''
  87. li.append(
  88. (
  89. i['adcreative_id'], i['adcreative_name'], i['campaign_id'],image,title,
  90. i.get('promoted_object_type', ''), i.get('page_type', ''),
  91. i['page_spec'].get('page_id', ''), i.get('promoted_object_id', ''),
  92. ','.join(i['site_set']),description,'GDT',account_id,dt
  93. )
  94. )
  95. total_page = r['data']['page_info']['total_page']
  96. if total_page > page:
  97. page += 1
  98. else:
  99. break
  100. else:
  101. break
  102. if len(li)>0:
  103. print(f"{account_id}有创意:",len(li))
  104. sql='replace into adcreative_info values(%s,%s,%s,%s,%s,%s,%s,%s,%s,%s,%s,%s,%s,%s) '
  105. db.quchen_text.executeMany(sql,li)
  106. def images_info_get(account_id,access_token,image_ids): # 获取图片信息
  107. fields = ('image_id','width','height','file_size','signature','preview_url')
  108. interface = 'images/get'
  109. url = 'https://api.e.qq.com/v1.3/' + interface
  110. page = 1
  111. li = []
  112. while True:
  113. common_parameters = {
  114. 'access_token': access_token,
  115. 'timestamp': int(time.time()),
  116. 'nonce': str(time.time()) + str(random.randint(0, 999999)),
  117. 'fields': fields
  118. }
  119. parameters = {
  120. "account_id": account_id,
  121. "filtering":[{
  122. "field": "image_id",
  123. "operator": "IN",
  124. "values": image_ids.split(',')
  125. }],
  126. "page": page,
  127. "page_size": 100
  128. }
  129. parameters.update(common_parameters)
  130. for k in parameters:
  131. if type(parameters[k]) is not str:
  132. parameters[k] = json.dumps(parameters[k])
  133. while True:
  134. h = requests.get(url, params=parameters)
  135. # print(h.text)
  136. if h.status_code == 200:
  137. r = h.json()
  138. break
  139. else:
  140. time.sleep(1)
  141. print("请求出错 等待1s..")
  142. if 'data' in r.keys():
  143. li.extend(r['data']['list'])
  144. total_page = r['data']['page_info']['total_page']
  145. if total_page > page:
  146. page += 1
  147. else:
  148. break
  149. # print(li)
  150. data = []
  151. for i in li:
  152. data.append((i['image_id'],i['width'],i['height'],i['signature'],i['preview_url']))
  153. # print(data)
  154. print(f"{account_id} 有图片:", li.__len__())
  155. if li.__len__() > 0:
  156. sql="replace into image_info value (%s,%s,%s,%s,%s)"
  157. db.quchen_text.executeMany(sql, data)
  158. db.close()
  159. def ad_info():
  160. accounts = db.quchen_text.getData("""
  161. select account_id,access_token,name channel,'GDT' type from advertiser_qq where name !='' or name is not null
  162. union
  163. select account_id,access_token,name channel,'MP' type from advertiser_vx where name !='' or name is not null
  164. """)
  165. total_data =[]
  166. executor = ThreadPoolExecutor(max_workers=max_workers)
  167. for i in accounts:
  168. # print(i)
  169. account_id =i[0]
  170. access_token = i[1]
  171. type = i[3]
  172. executor.submit(get_ad_info,account_id, access_token,type,total_data)
  173. executor.shutdown()
  174. print(len(total_data))
  175. if len(total_data)>0:
  176. sql="replace into ad_info values(%s,%s,%s,%s,%s,%s,%s) "
  177. db.quchen_text.executeMany(sql, total_data)
  178. """获取广告基础信息"""
  179. def get_ad_info(account_id, access_token, flag,ad_ids,dt):
  180. path = 'ads/get'
  181. fields = ('ad_id', 'ad_name', 'adcreative_id', 'adgroup_id', 'campaign_id')
  182. url = 'https://api.e.qq.com/v1.3/' + path
  183. li = []
  184. page = 1
  185. while True:
  186. parameters = {
  187. 'access_token': access_token,
  188. 'timestamp': int(time.time()),
  189. 'nonce': str(time.time()) + str(random.randint(0, 999999)),
  190. 'fields': fields,
  191. "filtering": [{
  192. "field": "ad_id",
  193. "operator": "IN",
  194. "values":
  195. ad_ids.split(',')
  196. }],
  197. "account_id": account_id,
  198. "page": page,
  199. "page_size": 100,
  200. "is_deleted": False
  201. }
  202. for k in parameters:
  203. if type(parameters[k]) is not str:
  204. parameters[k] = json.dumps(parameters[k])
  205. while True:
  206. r = requests.get(url, params=parameters).json()
  207. code = r['code']
  208. if code == 11017:
  209. time.sleep(61)
  210. else:
  211. break
  212. # print(r)
  213. total_page = r['data']['page_info']['total_page']
  214. if page > total_page:
  215. break
  216. else:
  217. page += 1
  218. if r.get("data"):
  219. for i in r['data']['list']:
  220. li.append((str(i['ad_id']), i['ad_name'], i['adcreative_id'], i['campaign_id'], i['adgroup_id'],
  221. account_id, flag, dt))
  222. if li.__len__()>0:
  223. print(f"{account_id}有广告:",li.__len__())
  224. sql = "replace into ad_info values(%s,%s,%s,%s,%s,%s,%s,%s) "
  225. db.quchen_text.executeMany(sql, li)
  226. db.close()
  227. def get_ad_cost_day(account_id,access_token,flag,st,et):
  228. if flag == 'MP':
  229. ad_cost_day_mp(account_id,access_token, st, et)
  230. else:
  231. ad_cost_day_gdt(account_id,access_token, st, et)
  232. def ad_cost_day_gdt(account_id,access_token,st,et):
  233. url = 'https://api.e.qq.com/v1.3/daily_reports/get'
  234. fields = ('date', 'ad_id','adgroup_id', 'cost', 'view_count', 'ctr', 'follow_count','web_order_count','order_amount')
  235. li = []
  236. page = 1
  237. while True:
  238. parameters = {
  239. 'access_token': access_token,
  240. 'timestamp': int(time.time()),
  241. 'nonce': str(time.time()) + str(random.randint(0, 999999)),
  242. 'fields': fields,
  243. "account_id": account_id,
  244. "group_by" : ['ad_id','date'],
  245. "level": 'REPORT_LEVEL_AD',
  246. "page": page,
  247. "page_size": 1000,
  248. "date_range": {
  249. "start_date": st,
  250. "end_date": et
  251. }
  252. }
  253. for k in parameters:
  254. if type(parameters[k]) is not str:
  255. parameters[k] = json.dumps(parameters[k])
  256. while True:
  257. r = requests.get(url, params=parameters).json()
  258. # print(r)
  259. code =r['code']
  260. if code==11017:
  261. time.sleep(61)
  262. else:
  263. break
  264. if r.get("data"):
  265. for i in r['data']['list']:
  266. if i['cost']>0:
  267. li.append(
  268. (
  269. i['date'], i['ad_id'],i['adgroup_id'], i['cost']/100, i['view_count'], i['ctr']*i['view_count'],
  270. i['follow_count'],i['web_order_count'],i['order_amount']/100,account_id,'GDT'
  271. )
  272. )
  273. total_page = r['data']['page_info']['total_page']
  274. if page >= total_page:
  275. break
  276. else:
  277. page += 1
  278. # print(li)
  279. if len(li) > 0:
  280. print(f"{account_id} have ad cost :{len(li)} ")
  281. db.quchen_text.executeMany('replace into ad_cost_day values(%s,%s,%s,%s,%s,%s,%s,%s,%s,%s,%s)', li)
  282. db.close()
  283. def ad_cost_day_mp(account_id,access_token,st,et):
  284. url = 'https://api.e.qq.com/v1.3/daily_reports/get'
  285. fields = ('date', 'ad_id','adgroup_id', 'cost', 'view_count', 'valid_click_count', 'official_account_follow_count','order_count','order_amount')
  286. li = []
  287. page = 1
  288. while True:
  289. parameters = {
  290. 'access_token': access_token,
  291. 'timestamp': int(time.time()),
  292. 'nonce': str(time.time()) + str(random.randint(0, 999999)),
  293. 'fields': fields,
  294. "account_id": account_id,
  295. "level": 'REPORT_LEVEL_AD_WECHAT',
  296. "page": page,
  297. "page_size": 1000,
  298. "date_range": {
  299. "start_date": st,
  300. "end_date": et
  301. }
  302. }
  303. for k in parameters:
  304. if type(parameters[k]) is not str:
  305. parameters[k] = json.dumps(parameters[k])
  306. while True:
  307. r = requests.get(url, params=parameters).json()
  308. # print(r['data']['list'])
  309. # import pandas as pd
  310. # print(pd.DataFrame(r['data']['list']))
  311. code = r['code']
  312. if code == 11017:
  313. time.sleep(61)
  314. else:
  315. break
  316. if r.get("data"):
  317. for i in r['data']['list']:
  318. if i['cost']>0:
  319. li.append(
  320. (
  321. i['date'],i['ad_id'],i['adgroup_id'],i['cost']/100,i['view_count'],i['valid_click_count'],
  322. i['official_account_follow_count'],i['order_count'],i['order_amount']/100,account_id,'MP'
  323. )
  324. )
  325. total_page = r['data']['page_info']['total_page']
  326. if page >=total_page:
  327. break
  328. else:
  329. page += 1
  330. # print(li)
  331. # exit()
  332. if len(li) > 0:
  333. print(f"{account_id} have ad cost :{len(li)} ")
  334. db.quchen_text.executeMany('replace into ad_cost_day values(%s,%s,%s,%s,%s,%s,%s,%s,%s,%s,%s)', li)
  335. db.close()
  336. def daily_reports_get(access_token, account_id, level, start_date, end_date, fields): # 获取wx投放计划日报数据
  337. interface = 'daily_reports/get'
  338. url = 'https://api.e.qq.com/v1.3/' + interface
  339. common_parameters = {
  340. 'access_token': access_token,
  341. 'timestamp': int(time.time()),
  342. 'nonce': str(time.time()) + str(random.randint(0, 999999)),
  343. 'fields': fields
  344. }
  345. parameters = {
  346. "account_id": account_id,
  347. "level": level,
  348. "date_range":
  349. {
  350. "start_date": start_date,
  351. "end_date": end_date
  352. },
  353. "page": 1,
  354. "page_size": 1000,
  355. "fields":
  356. [
  357. ]
  358. }
  359. parameters.update(common_parameters)
  360. for k in parameters:
  361. if type(parameters[k]) is not str:
  362. parameters[k] = json.dumps(parameters[k])
  363. while True:
  364. r = requests.get(url, params=parameters)
  365. if r.status_code == 200:
  366. break
  367. else:
  368. time.sleep(1)
  369. print("请求出错 等待1s..")
  370. return r.json()
  371. def daily_qq_reports_get(access_token, account_id, compaign_id, level, start_date, end_date, fields): # 获取gdt投放计划日报数据
  372. interface = 'daily_reports/get'
  373. url = 'https://api.e.qq.com/v1.1/' + interface
  374. common_parameters = {
  375. 'access_token': access_token,
  376. 'timestamp': int(time.time()),
  377. 'nonce': str(time.time()) + str(random.randint(0, 999999)),
  378. 'fields': fields
  379. }
  380. parameters = {
  381. "account_id": account_id,
  382. "filtering":
  383. [
  384. {
  385. "field": "campaign_id",
  386. "operator": "EQUALS",
  387. "values":
  388. [
  389. compaign_id
  390. ]
  391. }
  392. ],
  393. "level": level,
  394. "date_range":
  395. {
  396. "start_date": start_date,
  397. "end_date": end_date
  398. },
  399. "page": 1,
  400. "page_size": 1000,
  401. "fields":
  402. [
  403. ]
  404. }
  405. parameters.update(common_parameters)
  406. for k in parameters:
  407. if type(parameters[k]) is not str:
  408. parameters[k] = json.dumps(parameters[k])
  409. r = requests.get(url, params=parameters)
  410. return r.json()
  411. def mysql_insert_adcreative(data):
  412. db = pymysql.connect('rm-bp1c9cj79872tx3aaro.mysql.rds.aliyuncs.com', 'superc', 'Cc719199895', 'quchen_text')
  413. cursor = db.cursor()
  414. sql = 'replace into adcreative (campaign_id,adcreative_id,adcreative_name,image_id,title,promoted_object_type,page_type,page_id,link_page_id,promoted_object_id) values (%s,%s,%s,%s,%s,%s,%s,%s,%s,%s)'
  415. try:
  416. cursor.executemany(sql, data)
  417. db.commit()
  418. print('insert [adcreative] ', len(data))
  419. except:
  420. db.rollback()
  421. print('insert [adcreative] defeat')
  422. if __name__ == '__main__':
  423. account_id = 19206910
  424. access_token = '89079ccc8db047b078a0108e36a7e276'
  425. #
  426. account_id2 = 14709511
  427. access_token2 = 'e87f7b6f860eaeef086ddcc9c3614678'
  428. get_ad_cost_day(account_id,access_token,'MP','2021-04-09','2021-04-09')
  429. # get_adcreatives(account_id,access_token,'MP','3187867673','2021-04-09')