Skip to content

Commit

Permalink
Develop (#81)
Browse files Browse the repository at this point in the history
* created develop branch with python3 Readme

* python3 working through Allocation and Accumulation and MakeFinalTables, numpy file creation (#76)

* working through Allocation and Accumulation

* catch max/min eval in accumulation for empty upstream arrays

* update table clean in by_RPU metrics

* add encoding to state_dict load, f-strings, pathlib

* Update to readme

* Get python working (#80)

* working through Allocation and Accumulation

* catch max/min eval in accumulation for empty upstream arrays

* update table clean in by_RPU metrics

* add encoding to state_dict load, f-strings, pathlib

* reinstate CRS conversion in PointInPoly to retain matching outcomes, alter initial value for MIN/MAX cols in Accumulation

* sort cat table by numpy array comids

* remove RuntimeWarning from processing output

* use f-strings, .query(), and iterrows

* remove duplicate line to instantiate tables

* add check_metrics script

* modify check_metrics script

* order ControlTable by Final_Table_Name to group metrics

* check/make all dirs in MakeFinalTables individually

Co-authored-by: rickD <[email protected]>
  • Loading branch information
mhweber and debboutr authored Feb 12, 2021
1 parent fbee61f commit 23aa972
Show file tree
Hide file tree
Showing 6 changed files with 395 additions and 376 deletions.
256 changes: 128 additions & 128 deletions ControlTable_StreamCat.csv

Large diffs are not rendered by default.

57 changes: 14 additions & 43 deletions MakeFinalTables.py
Original file line number Diff line number Diff line change
Expand Up @@ -15,7 +15,7 @@
import zipfile
import numpy as np
import pandas as pd
from pathlib2 import Path
from pathlib import Path

from stream_cat_config import OUT_DIR, LENGTHS, FINAL_DIR

Expand All @@ -38,9 +38,8 @@ def build_stats(tbl, stats):
FINAL_DIR = Path(FINAL_DIR) # TODO: change this in the config
ctl = pd.read_csv("ControlTable_StreamCat.csv") # TODO move CONTROL_TABLE to config

inputs = np.load("accum_npy/vpu_inputs.npy").item()
inputs = np.load("accum_npy/vpu_inputs.npy", allow_pickle=True).item()

tables = dict()
runners = ctl.query("run == 1").groupby("Final_Table_Name")
tables = runners["FullTableName"].unique().to_dict()
# check that all accumulated files are present
Expand All @@ -55,23 +54,29 @@ def build_stats(tbl, stats):

if len(missing) > 0:
for miss in missing:
print("Missing {}".format(miss.name))
print "Check output from StreamCat.py"
print(f"Missing {miss.name}")
print("Check output from StreamCat.py")
sys.exit()

states_lookup = Path("state_dict.npz")
states_dict = np.load(str(states_lookup))["data"].item()
states_dict = np.load(str(states_lookup),
allow_pickle=True,
encoding="latin1")["data"].item()

STATES_DIR = FINAL_DIR.parents[0] / "States"
if not FINAL_DIR.exists():
FINAL_DIR.mkdir(parents=True)
if not (FINAL_DIR / "zips").exists():
(FINAL_DIR / "zips").mkdir()
if not STATES_DIR.exists():
STATES_DIR.mkdir()
if not (STATES_DIR / "zips").exists():
(STATES_DIR / "zips").mkdir()

region_fn = "{}_Region{}.csv"
for table, metrics in tables.items():

print("Running {} .....into {}".format(table, FINAL_DIR))
print(f"Running {table} .....into {FINAL_DIR}")
# this will print stats for every final table, used for metadata
stats = dict()
# Looop through NHD Hydro-regions
Expand Down Expand Up @@ -243,39 +248,5 @@ def build_stats(tbl, stats):
print(table)

for stat in stats:
print stat + " " + str(stats[stat])
print "All Done....."

###########################
#table = "RoadStreamCrossings"
#aa = []
#for f in os.listdir(str(REDO_DIR)):
# s = f.split("_Region")[0]
# if not s in aa:
# aa.append(f.split("_Region")[0])
#FINAL_DIR = Path("L:/Priv/CORFiles/Geospatial_Library_Projects/StreamCat/FTP_Staging/HydroRegions")
#for table in tables:
#for table in aa[49:-1]:
# print(table)
# for vpu in inputs:
# print(vpu)
# orig = pd.read_csv(FINAL_DIR / region_fn.format(table,vpu))
# new = pd.read_csv(REDO_DIR / region_fn.format(table,vpu))
# if not orig.equals(new):
# print(table, vpu, orig.equals(new))

#for col in orig.columns:
# print(col, (orig[col] == new[col]).all())
# if not (orig[col] == new[col]).all():
# break
#
#qq = pd.merge(orig[["COMID", col]], new[["COMID", col]],
# on="COMID", suffixes=("_orig", "_new"))


#for state in states_dict:
#
# f = fn.format(table, state)
# orig = pd.read_csv(STATES_DIR / f)
# new = pd.read_csv(REDO_STATES / f)
# print(table, state, orig.equals(new))
print (stat + " " + str(stats[stat]))
print("All Done.....")
89 changes: 68 additions & 21 deletions README.md
Original file line number Diff line number Diff line change
@@ -1,55 +1,102 @@
# StreamCat

## Description:
The StreamCat Dataset (http://www2.epa.gov/national-aquatic-resource-surveys/streamcat) provides summaries of natural and anthropogenic landscape features for ~2.65 million streams, and their associated catchments, within the conterminous USA. This repo contains code used in StreamCat to process a suite of landscape rasters to watersheds for streams and their associated catchments (local reach contributing area) within the conterminous USA using the [NHDPlus Version 2](http://www.horizon-systems.com/NHDPlus/NHDPlusV2_data.php) as the geospatial framework. See [Running-StreamCat-Scripts](https://github.com/USEPA/StreamCat/wiki/Running-StreamCat-Scripts) for details on running the scripts to produce StreamCat data.
The StreamCat Dataset (http://www2.epa.gov/national-aquatic-resource-surveys/streamcat) provides summaries of natural and anthropogenic landscape features for ~2.65 million streams, and their associated catchments, within the conterminous USA. This repo contains code used in StreamCat to process a suite of landscape rasters to watersheds for streams and their associated catchments (local reach contributing area) within the conterminous USA using the [NHDPlus Version 2](http://www.horizon-systems.com/NHDPlus/NHDPlusV2_data.php) as the geospatial framework.

## Necessary Python Packages and Installation Tips
The scripts for StreamCat rely on several python modules a user will need to install such as numpy, pandas, gdal, fiona, rasterio, geopandas, shapely, pysal, and ArcPy with an ESRI license (minimal steps still using ArcPy). We highly recommend using a scientific python distribution such as [Anaconda](https://www.continuum.io/downloads) or [Enthought Canopy](https://www.enthought.com/products/canopy/). We used the conda package manager to install necessary python modules. Our essential packages and versions used are listed below (Windows 64 and Python 2.7.11):
The scripts for StreamCat rely on several python modules a user will need to install such as numpy, pandas, gdal, fiona, rasterio, geopandas, shapely, pysal, and ArcPy with an ESRI license (minimal steps still using ArcPy). We highly recommend using a scientific python distribution such as [Anaconda](https://www.continuum.io/downloads) or [Enthought Canopy](https://www.enthought.com/products/canopy/). We used the conda package manager to install necessary python modules. Our essential packages and versions when code was last used are listed below - note that other configurations may work, we simply have verified this particular combination (Windows 64 and Python 3.6.10):

| Package | Version |
| ------------- |--------------:|
| fiona | 1.7.7 |
| gdal | 2.2.0 |
| geopandas | 0.2.1 |
| geos | 3.5.1 |
| libgdal | 2.0.0 |
| numpy | 1.12.1 |
| pandas | 0.20.2 |
| pyproj | 1.9.5.1 |
| pysal | 1.13.0 |
| rasterio | 1.0a9 |
| shapely | 1.5.17 |
| python | 3.6.10 |
| fiona | 1.8.9.post2 |
| gdal | 2.4.4 |
| geopandas | 0.8.1 |
| geos | 3.8.1 |
| libgdal | 2.4.4 |
| numpy | 1.19.1 |
| pandas | 1.1.1 |
| pyproj | 2.6.1 |
| rasterio | 1.1.5 |
| shapely | 1.7.1 |

If you are using Anaconda, creating a new, clean 'StreamCat' environment with these needed packages can be done easily and simply one of several ways:

* In your conda shell, add one necessary channel and then download the streamcat environment from the Anaconda cloud:
+ conda config --add channels conda-forge
+ conda env create mweber36/streamcat
+ conda env create mweber36/StreamCat

* Alternatively, using the streamcat.yml file in this repository, in your conda shell cd to the directory where your streamcat.yml file is located and run:
+ conda env create -f StreamCat.yml
+ conda env create -f streamcat_py3.yml

* To build environment yourself, do:
+ conda env create -n StreamCat rasterio geopandas
+ pip install georasters
+ conda create --name StreamCat -c conda-forge python=3.6 geopandas rasterio=1.1.5=py36h2409764_0

* To activate this new environment and open Spyder, type the following at the conda prompt
* To activate this new environment, you'll need to install Spyder in the environment, and possibly re-install pyqt with specific version (we did). You may even need to uninstall pyqt after installing Spyder (as below) and then specifically re-install:

+ install spyder=4.1.4=py36h9f0ad1d_0 -c conda-forge
+ install pyqt=5.12.3=py36h6538335_1 -c conda-forge

* To open Spyder, type the following at the conda prompt
+ activate Streamcat

Then

+ Spyder

Finally, to use arcpy in this new environment, you will need to copy your Arc .pth file into your new environment. Copy the .pth file for your install of ArcGIS located in a directory like:
Finally, to use arcpy in this new environment, you will need to copy several ArcPro files and folders to your new environment as follows:

+ C:/Program Files/ArcGIS/Pro/bin/Python/envs/arcgispro-py3/Lib/site-packages/ArcGISPro.pth

+ C:/Program Files/ArcGIS/Pro/bin/Python/envs/arcgispro-py3/Lib/site-packages/Arcgisscripting

+ C:\Python27\ArcGISx6410.3\Lib\site-packages\DTBGGP64.pth
+ C:/Program Files/ArcGIS/Pro/bin/Python/envs/arcgispro-py3/Lib/site-packages/arcpy_wmx

+ C:/Program Files/ArcGIS/Pro/bin/Python/envs/arcgispro-py3/Lib/site-packages/Gapy

To your environment directory which should look something like:

+ C:\Anaconda\envs\streamcat\Lib\site-packages\DTBGGP64.pth
+ C:/Users/mweber/AppData/Local/Continuum/anaconda3/envs/StreamCat/Lib/site-packages

You will also need to install dateutils:
+ conda install -c anaconda python-dateutil

Note that the exact paths may vary depending on the version of ArcGIS and Anaconda you have installed and the configuration of your computer

## How to Run Scripts
### The scripts make use of 'control tables' to pass all the particular parameters to the three primary scripts:

+ [StreamCat_PreProcessing.py](https://github.com/USEPA/StreamCat/blob/master/StreamCat_PreProcessing.py)
+ [StreamCat.py](https://github.com/USEPA/StreamCat/blob/master/StreamCat.py)
+ [MakeFinalTables.py](https://github.com/USEPA/StreamCat/blob/master/StreamCat_functions.py).

In turn, these scripts rely on a generic functions in [StreamCat_functions.py](https://github.com/USEPA/StreamCat/blob/master/StreamCat_functions.py).

To generate the riparian buffers we used in [StreamCat](ftp://newftp.epa.gov/EPADataCommons/ORD/NHDPlusLandscapeAttributes/StreamCat/Documentation/ReadMe.html) we used the code in [RiparianBuffers.py](https://github.com/USEPA/StreamCat/blob/master/RiparianBuffer.py)

To generate percent full for catchments on the US border for point features, we used the code in [border.py](https://github.com/USEPA/StreamCat/blob/master/border.py)

Examples of control tables used in scripts are:
+ [RasterControlTable](https://github.com/USEPA/StreamCat/blob/master/RasterControlTable.csv)
+ [ReclassTable](https://github.com/USEPA/StreamCat/blob/master/ReclassTable.csv)
+ [FieldCalcTable.](https://github.com/USEPA/StreamCat/blob/master/FieldCalcTable.csv)
+ [Lithology_lookup](https://github.com/USEPA/StreamCat/blob/master/Lithology_lookup.csv)
+ [NLCD2006_lookup](https://github.com/USEPA/StreamCat/blob/master/NLCD2006_lookup.csv)
+ [ControlTable_StreamCat](https://github.com/USEPA/StreamCat/blob/master/ControlTable_StreamCat.csv)
+ [MakeFinalTables](https://github.com/USEPA/StreamCat/blob/master/MakeFinalTables.csv)

### Running StreamCat.py to generate new StreamCat metrics

After editing the control tables to provide necessary information, such as directory paths, the following stesps will excecute processes to generate new watershed metrics for the conterminous US. All examples in the control table are for layers (e.g., STATSGO % clay content of soils) that were processed as part of the StreamCat Dataset. This example assumes run in Anaconda within Conda shell.

1. Edit [ControlTable_StreamCat](https://github.com/USEPA/StreamCat/blob/master/ControlTable_StreamCat.csv) and set desired layer's "run" column to 1. All other columns should be set to 0
2. Open a Conda shell and type "activate StreamCat"
3. At the Conda shell type: "Python<space>"
4. Drag and drop "StreamCat.py" to the Conda shell from a file manager followed by another space
5. Drag and drop the control table to the Conda shell

Final text in Conda shell should resemble this: python C:\some_path\StreamCat.py C:\some_other_path\ControlTable.csv


## EPA Disclaimer
The United States Environmental Protection Agency (EPA) GitHub project code is provided on an "as is" basis and the user assumes responsibility for its use. EPA has relinquished control of the information and no longer has responsibility to protect the integrity , confidentiality, or availability of the information. Any reference to specific commercial products, processes, or services by service mark, trademark, manufacturer, or otherwise, does not constitute or imply their endorsement, recommendation or favoring by EPA. The EPA seal and logo shall not be used in any manner to imply endorsement of any commercial product or activity by EPA or the United States Government.
Expand Down
Loading

0 comments on commit 23aa972

Please sign in to comment.