Commit
This commit does not belong to any branch on this repository, and may belong to a fork outside of the repository.
* Create the DatabaseConnector * Implement DatabaseConnector for the DB connectors * Add DatabaseConnector to std imports * Flake8 fix * Remove reference to padding in copy() * Add database_discover and fix inheritance * Remove strict_length from copy() * Put strict_length back in original order * Remove strict_length stub from BQ * Fix discover_database export statement * Add return annotation to mysql table_exists * Black formatter pass * create bigquery folder in databases folde * create query parity between bigquery and redshift * mock up copy functionality for bigquery * fix typo * add duplicate function to bigquery * move transaction to helper function * implement upsert * fix imports and packages * add get tables and views methods * add query return flexibility * match bigquery apis with redshift * make s3 to gcs more generic * add transaction support to bigquery * remove logs * add gcs docs * process job config in function * finish todo's (and add one more lol) * [ wip ] AttributeError * add raw download param * drop raw download * copy from GCS docstring * copy s3 docs * copy docs * docstrings * control flow * add source path to aws transfer spec * add Code object to imports * cleaning up slightly * check status code * nice * pass in required param * add pattern handling * add quote character to LoadJobConfig * add schema to copy from gcs * drop dist and sortkeys No longer input params * add delimiter param * use schema definition * write column mapping helper * pass in formatted schema to load_uri fn * rename new file * move file with jason's changes * move new changes back into file to maintain history * remove extraneous fn and move project job config * get back to test parity * fix bad merge conflict * remove extra params from copy sig * clarify transaction guidance * clean up list blobs * clean up storage transfer polling * upgrade cloud storage package * use list of schema mappings * scaffolded big file function 😎 * add to docs * default to compression we can make this more flexible, just scaffolding * add temp logging we can drop this later just trying to get a handle on cycle time * use decompress * add logging * implement unzipping and reuploading cloud file * logging error * Add destination path * Small fix * add todo's * drop max wait time * add kwargs to put blob Potentially useful for metadata (content type, etc.) * add verbosity to description * black formatted * add gcs to/from helpers * write to_bigquery function * update big file logic * allow jagged rows logic * test additional methods * add duplicate table test * test drop flag for duplicate * basic test for upsert * add typing * move non-essential logs to debug * move logs to debug * hey, it works! * add UUID support for bigquery type map * add datetime to bigquery type map * address comments * address comments * drop GCS class function we can pick this up later but it doesn't currently work * move class back to old location with new import * revert to old name * remove transaction error handler * add description conditional block for s3 * change one more conditional to s3 * handle empty source paths * reverting new import path --------- Co-authored-by: Jason Walker <[email protected]> Co-authored-by: Ian <[email protected]> Co-authored-by: Kasia Hinkson <[email protected]>
- Loading branch information