Compare commits

..

No commits in common. "617103e03ac33d5e29b4d492f6725a3417a9c730" and "3b15dae69caa44cc93bca5bdf26fe3f18dc57b3c" have entirely different histories.

2 changed files with 102 additions and 90 deletions

View File

@ -0,0 +1,102 @@
/*******Main Section**************************************************************************/
\set ON_ERROR_STOP on
\set AUTOCOMMIT on
\timing on
delete from p10_sa.S98_S_livechat
;
insert into p10_sa.S98_S_livechat
( conference_id
, create_date
, start_date
, start_url
, referrer
, chat_dura
, queue_dura
, visitor_id
, visit_nick
, visitor_ip
, visitor_email
, operator_id
, group_id
, group_name
, rate
, last_rate_comment
, operator_1_nick
, operator_2_nick
, operator_3_nick
, operator_4_nick
, etl_tx_dt )
select
conference_id
, create_date
, start_date
, start_url
, referrer
, chat_dura
, queue_dura
, visitor_id
, visit_nick
, visitor_ip
, visitor_email
, operator_id
, group_id
, group_name
, rate
, last_rate_comment
, operator_1_nick
, operator_2_nick
, operator_3_nick
, operator_4_nick
, etl_tx_dt
from p00_tal.S98_S_livechat
;
delete from p12_sfull.S98_S_livechat
;
;
insert into p12_sfull.S98_S_livechat
( conference_id
, create_date
, start_date
, start_url
, referrer
, chat_dura
, queue_dura
, visitor_id
, visit_nick
, visitor_ip
, visitor_email
, operator_id
, group_id
, group_name
, rate
, last_rate_comment
, operator_1_nick
, operator_2_nick
, operator_3_nick
, operator_4_nick
, etl_tx_dt )
select
conference_id
, create_date
, start_date
, start_url
, referrer
, chat_dura
, queue_dura
, visitor_id
, visit_nick
, visitor_ip
, visitor_email
, operator_id
, group_id
, group_name
, rate
, last_rate_comment
, operator_1_nick
, operator_2_nick
, operator_3_nick
, operator_4_nick
, etl_tx_dt
from p10_sa.S98_S_livechat
;
\q

View File

@ -1,90 +0,0 @@
#!/usr/bin/python
# -*- encoding=utf-8 -*-
from airflow import DAG
from datetime import datetime, timedelta
from airflow.contrib.hooks.ssh_hook import SSHHook
from airflow.contrib.operators.ssh_operator import SSHOperator
from airflow.sensors.external_task_sensor import ExternalTaskSensor
import json
from airflow.operators.email_operator import EmailOperator
from airflow.utils.trigger_rule import TriggerRule
sshHook = SSHHook(ssh_conn_id ='ssh_air')
default_args = {
'owner': 'info@idgvalue.com',
'email': [''],
'email_on_failure': True,
'email_on_retry':True,
'start_date': datetime(2022, 9, 12),
'depends_on_past': False,
'retries': 6,
'retry_delay': timedelta(minutes=10),
}
dag = DAG('wf_dag_tk_cust', default_args=default_args,
schedule_interval="0 0 * * *",
catchup=False,
dagrun_timeout=timedelta(minutes=160),
max_active_runs=3)
task_failed = EmailOperator (
dag=dag,
trigger_rule=TriggerRule.ONE_FAILED,
task_id="task_failed",
to=["info@idgvalue.com"],
cc=[""],
subject="tk_cust_failed",
html_content='<h3>您好tk_cust作业失败请及时处理" </h3>')
file_Tk01 = SSHOperator(
ssh_hook=sshHook,
task_id='file_Tk01',
command='python /data/airflow/bin/FILELOD.py {{ params.my_param }} {{ ds_nodash }} >>/data/airflow/logs/file_load/file_load_{{ds_nodash}}.log 2>&1 ',
params={'my_param':"Tk01"},
depends_on_past=False,
retries=3,
dag=dag)
file_Update_af0912 = SSHOperator(
ssh_hook=sshHook,
task_id='file_Update_af0912',
command='python /data/airflow/bin/FILELOD.py {{ params.my_param }} {{ ds_nodash }} >>/data/airflow/logs/file_load/file_load_{{ds_nodash}}.log 2>&1 ',
params={'my_param':"Update_af0912"},
depends_on_past=False,
retries=3,
dag=dag)
file_Merge_af0912 = SSHOperator(
ssh_hook=sshHook,
task_id='file_Merge_af0912',
command='python /data/airflow/bin/FILELOD.py {{ params.my_param }} {{ ds_nodash }} >>/data/airflow/logs/file_load/file_load_{{ds_nodash}}.log 2>&1 ',
params={'my_param':"Merge_af0912"},
depends_on_past=False,
retries=3,
dag=dag)
file_SCRM_Contact_af0912 = SSHOperator(
ssh_hook=sshHook,
task_id='file_SCRM_Contact_af0912',
command='python /data/airflow/bin/FILELOD.py {{ params.my_param }} {{ ds_nodash }} >>/data/airflow/logs/file_load/file_load_{{ds_nodash}}.log 2>&1 ',
params={'my_param':"SCRM_Contact_af0912"},
depends_on_past=False,
retries=3,
dag=dag)
file_SCRM_Contact_bf0912 = SSHOperator(
ssh_hook=sshHook,
task_id='file_SCRM_Contact_bf0912',
command='python /data/airflow/bin/FILELOD.py {{ params.my_param }} {{ ds_nodash }} >>/data/airflow/logs/file_load/file_load_{{ds_nodash}}.log 2>&1 ',
params={'my_param':"SCRM_Contact_bf0912"},
depends_on_past=False,
retries=3,
dag=dag)