Source code for pudl.convert.ferc_to_sqlite

"""A script for cloning the FERC Form 1 database into SQLite.

This script generates a SQLite database that is a clone/mirror of the original
FERC Form1 database. We use this cloned database as the starting point for the
main PUDL ETL process. The underlying work in the script is being done in
import argparse
import pathlib
import sys
from pathlib import Path

import yaml

import pudl
from pudl.settings import FercToSqliteSettings
from pudl.workspace.datastore import Datastore

# Create a logger to output any messages we might have...
[docs]logger = pudl.logging_helpers.get_logger(__name__)
[docs]def parse_command_line(argv): """Parse command line arguments. See the -h option. Args: argv (str): Command line arguments, including caller filename. Returns: dict: Dictionary of command line arguments and their parsed values. """ parser = argparse.ArgumentParser(description=__doc__) parser.add_argument( "settings_file", type=str, default="", help="path to YAML settings file." ) parser.add_argument( "--logfile", default=None, type=str, help="If specified, write logs to this file.", ) parser.add_argument( "-c", "--clobber", action="store_true", help="""Clobber existing sqlite database if it exists. If clobber is not included but the sqlite databse already exists the _build will fail.""", default=False, ) parser.add_argument( "--sandbox", action="store_true", default=False, help="Use the Zenodo sandbox rather than production", ) parser.add_argument( "-b", "--batch-size", default=50, type=int, help="Specify number of XBRL instances to be processed at a time (defaults to 50)", ) parser.add_argument( "-w", "--workers", default=None, type=int, help="Specify number of worker processes for parsing XBRL filings.", ) parser.add_argument( "--gcs-cache-path", type=str, help="Load datastore resources from Google Cloud Storage. Should be gs://bucket[/path_prefix]", ) parser.add_argument( "--bypass-local-cache", action="store_true", default=False, help="If enabled, the local file cache for datastore will not be used.", ) parser.add_argument( "--loglevel", help="Set logging level (DEBUG, INFO, WARNING, ERROR, or CRITICAL).", default="INFO", ) arguments = parser.parse_args(argv[1:]) return arguments
[docs]def main(): # noqa: C901 """Clone the FERC Form 1 FoxPro database into SQLite.""" args = parse_command_line(sys.argv) # Display logged output from the PUDL package: pudl.logging_helpers.configure_root_logger( logfile=args.logfile, loglevel=args.loglevel ) with pathlib.Path(args.settings_file).open() as f: script_settings = yaml.safe_load(f) defaults = pudl.workspace.setup.get_defaults() pudl_in = script_settings.get("pudl_in", defaults["pudl_in"]) pudl_out = script_settings.get("pudl_out", defaults["pudl_out"]) pudl_settings = pudl.workspace.setup.derive_paths( pudl_in=pudl_in, pudl_out=pudl_out ) parsed_settings = FercToSqliteSettings().parse_obj( script_settings["ferc_to_sqlite_settings"] ) # Configure how we want to obtain raw input data: ds_kwargs = dict( gcs_cache_path=args.gcs_cache_path, sandbox=pudl_settings.get("sandbox", False) ) if not args.bypass_local_cache: ds_kwargs["local_cache_path"] = Path(pudl_settings["pudl_in"]) / "data" pudl_settings["sandbox"] = args.sandbox if parsed_settings.ferc1_dbf_to_sqlite_settings: pudl.extract.ferc1.dbf2sqlite( ferc1_to_sqlite_settings=parsed_settings.ferc1_dbf_to_sqlite_settings, pudl_settings=pudl_settings, clobber=args.clobber, datastore=Datastore(**ds_kwargs), ) pudl.extract.xbrl.xbrl2sqlite( ferc_to_sqlite_settings=parsed_settings, pudl_settings=pudl_settings, clobber=args.clobber, datastore=Datastore(**ds_kwargs), batch_size=args.batch_size, workers=args.workers, )
if __name__ == "__main__": sys.exit(main())