""" @Author : HuangJian @Description : 各站点店铺asin详情表-- 月抓取 @SourceTable : us_asin_detail_product_2023 @SinkTable : ods_asin_detail_product @CreateTime : 2022/05/19 14:55 @UpdateTime : 2022/05/19 14:55 """ import os import sys sys.path.append(os.path.dirname(sys.path[0])) from utils.ssh_util import SSHUtil from utils.db_util import DBUtil from utils.common_util import CommonUtil from utils.common_util import DateTypes from utils.hdfs_utils import HdfsUtils from utils.spark_util import SparkUtil if __name__ == '__main__': site_name = CommonUtil.get_sys_arg(1, None) asin_type = CommonUtil.get_sys_arg(2, None) date_type = CommonUtil.get_sys_arg(3, None) date_info = CommonUtil.get_sys_arg(4, None) assert site_name is not None, "site_name 不能为空!" assert date_type is not None, "date_type 不能为空!" assert date_info is not None, "date_info 不能为空!" hive_table = f"ods_keepa_asin_trend" partition_dict = { "site_name": site_name, "asin_type": asin_type, "date_type": date_type, "date_info": date_info } # 落表路径校验 hdfs_path = CommonUtil.build_hdfs_path(hive_table, partition_dict=partition_dict) print(f"hdfs_path is {hdfs_path}") import_table = f"{site_name}_keepa_trend_{asin_type}" db_type = 'postgresql' print("当前链接的数据库为:", db_type, " 同步的表为:", import_table) sql_query = f""" select id, asin, title, img_url, listed_since, first_bsr_label, last_bsr_label, last_bsr_rank, last_price, last_count_reviews, currency, times_list, price_list, rank_list, first_bsr_list, last_bsr_list, count_reviews_list, variations, data_zoom_start, date_range, color, size, last_bsr_rating, rating_count, rating_count_list, min_list, max_list, current_list, avg_list, avg90_list, avg180_list, bsr_orders_list, bsr_orders_sale_list, created_time, updated_time from {import_table} where 1=1 and \$CONDITIONS """ # 进行schema和数据校验 CommonUtil.check_schema_before_import(db_type=db_type, site_name=site_name, query=sql_query, hive_tb_name=hive_table, msg_usr=['chenyuanjie']) # 生成导出脚本 import_sh = CommonUtil.build_import_sh(site_name=site_name, db_type=db_type, query=sql_query, hdfs_path=hdfs_path, map_num=20, key='asin') # 导入前先删除原始hdfs数据 HdfsUtils.delete_hdfs_file(hdfs_path) # 创建ssh Client对象--用于执行cmd命令 client = SSHUtil.get_ssh_client() SSHUtil.exec_command_async(client, import_sh, ignore_err=False) # 创建lzo索引和修复元数据 CommonUtil.after_import(hdfs_path=hdfs_path, hive_tb=hive_table) # 关闭链接 client.close()