我有两个表,需要分别在table_name和file_name上应用join。问题是与表2中的file_name相比,table_name有一些额外的字符串。
使用正则表达式,我如何从table_name中删除额外的字符串,使其与表2的file_name兼容?
TABLE 1:
table_name audit_record_count
Immunology_COVID-19_Treatment_202006221630_01.csv 1260124
Immunology_COVID-19_Trial_Design_202006221630_01.csv 2173762
Immunology_COVID-19_Planned_Treatment_202006221630_01.csv 1350135
Immunology_COVID-19_Patient_Characteristic_202006221630_01.csv 2173762
Immunology_COVID-19_Intervention_Type_202006221630_01.csv 2173762
Immunology_COVID-19_Arm_202006221630_01.csv 4
Immunology_COVID-19_Actual_Treatment_202006221630_01.csv 2173762
Immunology_COVID-19_Publication_202006221630_01.csv 2173762
Immunology_COVID-19_Outcome_202006221630_01.csv 2173762
Immunology_COVID-19_Intervention_Type_Factor_202006221630_01.csv 2173762
Immunology_COVID-19_Inclusion_Criteria_202006221630_01.csv 2173762
Immunology_COVID-19_Curation_202006221630_01.csv 2173762
TABLE 2:
file_name csv_record_count
Treatment 1260124
Trial_Design 2173762
Planned_Treatment 1350135
Patient_Characteristic 2173762
Intervention_Type 2173762
Arm 4
Actual_Treatment 2173762
Publication 2173762
Outcome 2173762
Intervention_Type_Factor 2173762
Inclusion_Criteria 2173762
Curation 2173762我尝试过的:
audit_file_df = spark.read.csv(
f"s3://{config['raw_bucket']}/{config['landing_directory']}/{config['audit_file']}/{watermark_timestamp}*.csv",
header=False, inferSchema=True) \
.withColumnRenamed("_c0", "table_name").withColumnRenamed("_c1", "audit_record_count")\
.selectExpr("regexp_extract(table_name, '^(.(?!(\\\\d{12}_\\\\d{2,4}.csv|\\\\d{12}.csv)))*', 0) AS table_name",'audit_record_count')
print("audit_file_df :",audit_file_df)
audit_file_df.show()
validation_df = audit_file_df.join(schema_validation_df, how='inner', on=audit_file_df['table_name'] == schema_validation_df['file_name']).withColumn("count_match",
col=col(
'audit_record_count') == col(
'csv_record_count'))
print("Record validation result")
validation_df.show()我可以从table_name中删除时间戳,但不能提取file_name来使连接条件工作。
加法
免疫学_新冠肺炎未修复可能会更改为其他文件,table_name的格式为:
TA_Indication_data_timestamp_nn.csv发布于 2020-07-01 04:28:38
在表1中创建一个包含data部件的附加列:
df = df.withColumn('data', F.regexp_extract(F.col('table_name'), '.*?_.*?_(.*)_\d{12}_\d{2}\.csv', 1))给出
+----------------------------------------------------------------+---------+------------------------+
|table_name |audit_rec|data |
+----------------------------------------------------------------+---------+------------------------+
|Immunology_COVID-19_Treatment_202006221630_01.csv |1260124 |Treatment |
|Immunology_COVID-19_Trial_Design_202006221630_01.csv |2173762 |Trial_Design |
|Immunology_COVID-19_Planned_Treatment_202006221630_01.csv |1350135 |Planned_Treatment |
|Immunology_COVID-19_Patient_Characteristic_202006221630_01.csv |2173762 |Patient_Characteristic |
|Immunology_COVID-19_Intervention_Type_202006221630_01.csv |2173762 |Intervention_Type |
|Immunology_COVID-19_Arm_202006221630_01.csv |4 |Arm |
|Immunology_COVID-19_Actual_Treatment_202006221630_01.csv |2173762 |Actual_Treatment |
|Immunology_COVID-19_Publication_202006221630_01.csv |2173762 |Publication |
|Immunology_COVID-19_Outcome_202006221630_01.csv |2173762 |Outcome |
|Immunology_COVID-19_Intervention_Type_Factor_202006221630_01.csv|2173762 |Intervention_Type_Factor|
|Immunology_COVID-19_Inclusion_Criteria_202006221630_01.csv |2173762 |Inclusion_Criteria |
|Immunology_COVID-19_Curation_202006221630_01.csv |2173762 |Curation |
+----------------------------------------------------------------+---------+------------------------+然后,您可以使用table1.data和table2.file_name连接这些表,并继续执行问题中已经给出的审计检查。
正则表达式的棘手之处在于使用non-greedy限定符,因为data部分本身可以包含下划线字符。
https://stackoverflow.com/questions/62664307
复制相似问题