-
Notifications
You must be signed in to change notification settings - Fork 106
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[presidential map] decide on synchronization between FECP and Postgres #4172
Comments
Data needed to support presidential map is "specially processed raw data". That is, it had not been through the regular coding process. Data expert (Paul) need to review and give a "go" signal before it can be published. The final set of tables were in FECP database only. The previous ticket had added these tables to our cloud database, and initial load of these tables had been done. Now at each filing deadline, after the final set of required FECP tables had been refreshed, their counterpart in our cloud PostgreSQL database will be refresh. |
This final set of tables can be divided into two categories: summary data and detail schedule data. Due to the different nature of these tables, the refresh process will be different. (the process to refresh 2016 data will be the same described below for 2020 data, IF data changes) The following summary tables will be completely refreshed. The amount of data need to be refreshed is small and a total refresh is clean and simple. The following details tables has much more data. An incremental refresh process (only update the changed data) will be used. |
For the complete total refresh, a python program will read from the source FECP database and re-load data into their postgreSQL counterpart in the cloud databases. |
For the two detail schedule tables, the record of changes are captured in an audit table in the intermediate database. A separate Python program, which use this audit table as the "driver", to grab the changed rows in FECP and insert into ((or delete from) their postgreSQL counterpart in the cloud database, in multi sessions in case the amount of changed rows are large. |
Materialized view in intermediate database: Package in intermediate database: Python program and shell scripts used load_pres_data16.ksh |
Summary
currently data needed to support presidential map is manually loaded into PostgreSQL. We would like to identify a way that data will be routinely loaded/synchronized from FECP --> PostgreSQL
We need to get a better idea about how difficult this task is and if we have the appropriate data for this work to reflect correctly in the endpoint.
Completion criteria
Technical considerations
The text was updated successfully, but these errors were encountered: