Skip to content

Commit

Permalink
post-scan update
Browse files Browse the repository at this point in the history
  • Loading branch information
li000363 committed Sep 23, 2024
1 parent e2f11b5 commit af333e0
Show file tree
Hide file tree
Showing 6 changed files with 21 additions and 46 deletions.
10 changes: 5 additions & 5 deletions mitre/src/phsafe_safetab_reader/cef_config.ini
Original file line number Diff line number Diff line change
@@ -1,8 +1,8 @@
[paths]
per_dir = s3://v-s3-das-ite-sourcedata-252999262699-us-gov-west-1/2010-convert/cef/us/per/
unit_dir = s3://v-s3-das-ite-sourcedata-252999262699-us-gov-west-1/2010-convert/cef/us/unit/
per_dir_pr = s3://v-s3-das-ite-sourcedata-252999262699-us-gov-west-1/2010-convert/cef/pr/per/
unit_dir_pr = s3://v-s3-das-ite-sourcedata-252999262699-us-gov-west-1/2010-convert/cef/pr/unit/
grfc_dir = s3://v-s3-das-ite-sourcedata-252999262699-us-gov-west-1/2010-convert/grfc/
per_dir = <US Person CEF Path>
unit_dir = <US Unit CEF Path>
per_dir_pr = <PR Person CEF Path>
unit_dir_pr = <PR Unit CEF Path>
grfc_dir = <GRFC Path>
per_file_format = CEF20_PER_%%s.txt
unit_file_format = CEF20_UNIT_%%s.txt
3 changes: 0 additions & 3 deletions mitre/src/phsafe_safetab_reader/cef_validator_classes.py
Original file line number Diff line number Diff line change
Expand Up @@ -5,7 +5,6 @@
# This file was automatically generated by ../etl_2020/census_etl/spec_scanner.py on Fri Oct 30 14:06:08 2020
# Command line: ../etl_2020/census_etl/spec_scanner.py ../etl_2020/cef/ADO-DEV-CEF_FOR_CDL_Layouts-2020DRPS-20200611.xlsx --output_parser ../das_decennial/programs/reader/cef_2020/cef_validator_classes.py --tablenames CEF20_UNIT CEF20_PERSON CEF20_CNT

# Automatically generated on Fri Oct 30 14:06:08 2020 by /mnt/users/will0555/new-das-vm/das-vm-config/das_decennial/das_framework/ctools/schema/table.py

def leftpad(x,width):
return ' '*(width-len(str(x)))+str(x)
Expand Down Expand Up @@ -488,7 +487,6 @@ def parse_piped_line(line):



# Automatically generated on Fri Oct 30 14:06:08 2020 by /mnt/users/will0555/new-das-vm/das-vm-config/das_decennial/das_framework/ctools/schema/table.py

def leftpad(x,width):
return ' '*(width-len(str(x)))+str(x)
Expand Down Expand Up @@ -1023,7 +1021,6 @@ def parse_piped_line(line):



# Automatically generated on Fri Oct 30 14:06:08 2020 by /mnt/users/will0555/new-das-vm/das-vm-config/das_decennial/das_framework/ctools/schema/table.py

def leftpad(x,width):
return ' '*(width-len(str(x)))+str(x)
Expand Down
10 changes: 5 additions & 5 deletions mitre/src/phsafe_safetab_reader/safetab_cef_config.ini
Original file line number Diff line number Diff line change
@@ -1,9 +1,9 @@
[paths]
per_dir = s3://v-s3-das-prod-data-412241963457-us-gov-west-1/mft/cdl-to-das/PER/
unit_dir = s3://v-s3-das-prod-data-412241963457-us-gov-west-1/mft/cdl-to-das/UNIT/
per_dir_pr = s3://v-s3-das-prod-data-412241963457-us-gov-west-1/mft/cdl-to-das/PER/
unit_dir_pr = s3://v-s3-das-prod-data-412241963457-us-gov-west-1/mft/cdl-to-das/UNIT/
grfc_dir = s3://v-s3-das-common-drps-412241963457-us-gov-west-1/2020/production/grfc/
per_dir = <US Person CEF Path>
unit_dir = <US Unit CEF Path>
per_dir_pr = <PR Person CEF Path>
unit_dir_pr = <PR Unit CEF Path>
grfc_dir = <GRFC Path>
per_file_format = CEF20_PER_%%s.txt
unit_file_format = CEF20_UNIT_%%s.txt
geo_file_format = grfc_tab20_%%s.txt
Expand Down
10 changes: 5 additions & 5 deletions mitre/src/phsafe_safetab_reader/safetab_cef_config_2010.ini
Original file line number Diff line number Diff line change
@@ -1,9 +1,9 @@
[paths]
per_dir = s3://v-s3-das-ite-data-252999262699-us-gov-west-1/cef/2010_2020fmt_qrace_v2/per/
unit_dir = s3://v-s3-das-ite-data-252999262699-us-gov-west-1/cef/2010_2020fmt_qrace_v2/unit/
per_dir_pr = s3://v-s3-das-ite-data-252999262699-us-gov-west-1/cef/2010_2020fmt_qrace_v2/per/
unit_dir_pr = s3://v-s3-das-ite-data-252999262699-us-gov-west-1/cef/2010_2020fmt_qrace_v2/unit/
grfc_dir = s3://v-s3-das-ite-data-252999262699-us-gov-west-1/2010-convert/grfc/
per_dir = <US Person CEF Path>
unit_dir = <US Unit CEF Path>
per_dir_pr = <PR Person CEF Path>
unit_dir_pr = <PR Unit CEF Path>
grfc_dir = <GRFC Path>
per_file_format = CEF20_PER_%%s.txt
unit_file_format = CEF20_UNIT_%%s.txt
geo_file_format = grfc_tab20_%%s.txt
Expand Down
12 changes: 6 additions & 6 deletions mitre/src/phsafe_safetab_reader/safetab_h_cef_config_2010.ini
Original file line number Diff line number Diff line change
@@ -1,10 +1,10 @@
[paths]
per_dir = s3://v-s3-das-ite-data-252999262699-us-gov-west-1/cef/2010_2020fmt_qrace_v2/per/
unit_dir = s3://v-s3-das-ite-data-252999262699-us-gov-west-1/cef/2010_2020fmt_qrace_v2/unit/
per_dir_pr = s3://v-s3-das-ite-data-252999262699-us-gov-west-1/cef/2010_2020fmt_qrace_v2/per/
unit_dir_pr = s3://v-s3-das-ite-data-252999262699-us-gov-west-1/cef/2010_2020fmt_qrace_v2/unit/
grfc_dir = s3://v-s3-das-ite-data-252999262699-us-gov-west-1/2010-convert/grfc/
pop_dir = s3://v-s3-das-ite-data-252999262699-us-gov-west-1/pop-group-totals/
per_dir = <US Person CEF Path>
unit_dir = <US Unit CEF Path>
per_dir_pr = <PR Person CEF Path>
unit_dir_pr = <PR Unit CEF Path>
grfc_dir = <GRFC Path>
pop_dir = <POP Path>
per_file_format = CEF20_PER_%%s.txt
unit_file_format = CEF20_UNIT_%%s.txt
geo_file_format = grfc_tab20_%%s.txt
Expand Down
22 changes: 0 additions & 22 deletions mitre/src/phsafe_safetab_reader/safetab_p_cef_reader.py
Original file line number Diff line number Diff line change
Expand Up @@ -266,25 +266,3 @@ def get_person_df(self) -> SparkDataFrame:

print("person_df:", p_reader.get_person_df())

'''
import uuid
unique_id = uuid.uuid1()
model_write = f"cef_reader_{unique_id}"
file_name = f"{model_write}.txt"
local_file = f"/tmp/{file_name}"
s3_file = f"s3://uscb-decennial-ite-das/users/zhou0021/debug/reader/{file_name}"
print("TEST123-----------------------------")
for i in range(10):
print(reader.get_person_df().rdd.takeSample(False,1))
print("PERDF-----------------")
print(reader.get_unit_df().rdd.takeSample(False,1))
print("UNIDF_________________")
with open(local_file, "at+") as f:
f.write(f"PER_DF:{reader.get_person_df().rdd.takeSample(False,1)}\n\n")
f.write(f"UNIT_DF:{reader.get_unit_df().rdd.takeSample(False,1)}\n\n")
cmd = ['/usr/bin/aws', 's3', 'cp', local_file, s3_file]
subprocess.Popen(cmd, stderr=subprocess.PIPE)
'''

0 comments on commit af333e0

Please sign in to comment.