mirror of
https://github.com/DOI-DO/j40-cejst-2.git
synced 2025-07-29 20:41:17 -07:00
Added tract grandfathering language to UI
This commit is contained in:
parent
d22c348504
commit
a58edbc724
10 changed files with 49 additions and 23 deletions
|
@ -282,6 +282,7 @@ TILES_SCORE_COLUMNS = {
|
|||
# The NEW final score value INCLUDES the adjacency index.
|
||||
field_names.FINAL_SCORE_N_BOOLEAN: "SN_C",
|
||||
field_names.FINAL_SCORE_N_BOOLEAN_V1_0: "SN_C_V10",
|
||||
field_names.GRANDFATHERED_N_COMMUNITIES_V1_0: "SN_GRAND",
|
||||
field_names.IS_TRIBAL_DAC: "SN_T",
|
||||
field_names.DIABETES_LOW_INCOME_FIELD: "DLI",
|
||||
field_names.ASTHMA_LOW_INCOME_FIELD: "ALI",
|
||||
|
|
|
@ -1,23 +1,25 @@
|
|||
These files are used as inputs to unit tests. Some notes in their creation is below.
|
||||
# How to generate the sample data in this folder
|
||||
|
||||
The sample data in this folder can be easily generated by debugging the `data_pipeline/etl/score/etl_score_post.py` file
|
||||
and exporting data using the debugger console. Examples of this exporting are below.
|
||||
|
||||
## Why in pickle format?
|
||||
|
||||
Exporting as a Pickle file keeps all the metadata about the columns including the data types. If we were to export as CSV then we will need
|
||||
to code the data types in the test fixtures for all the columns for the comparison to be correct.
|
||||
|
||||
## Exporting the test data
|
||||
|
||||
First, verify the code works as expected before exporting the data. You will not be able to inspect the data exports as they are in binary.
|
||||
You will be using the debugger to export the data. Note that it is best to export a small subset of the data for faster test execution.
|
||||
|
||||
### create_tile_data test
|
||||
1. Place a breakpoint in `data_pipeline/etl/score/etl_score_post.py` in the `transform` method right after the call to
|
||||
`_create_tile_data` and start the debugger running the Generate Post Score command (`generate-score-post`).
|
||||
1. Partially export the `output_score_county_state_merged_df` and `self.output_score_tiles_df` data to a pickle file once the debugger pauses
|
||||
at the breakpoint. Use these sample commands in the debugger console. Note that we are using head and tail to have territories in the sample data.
|
||||
|
||||
### create_tile_data_expected.pkl
|
||||
1. Set a breakpoint in the `test_create_tile_data` method in `data_pipeline/etl/score/tests/test_score_post.py`
|
||||
after the call to `_create_tile_data` and debug the test.
|
||||
2. Extract a subset of the `output_tiles_df_actual` dataframe. Do not extract the whole score as the file
|
||||
will be too big and the test will run slow. Also, you need to extract the same tracts that are in
|
||||
the `create_tile_score_data_input.pkl` input data. For example, use the following command once the breakpoint is reached
|
||||
to extract a few rows at the top and bottom of the score. This will some capture states and territories.
|
||||
```python
|
||||
import pandas as pd
|
||||
pd.concat([output_tiles_df_actual.head(3), output_tiles_df_actual.tail(3)], ignore_index=True).to_pickle('data_pipeline/etl/score/tests/snapshots/create_tile_data_expected.pkl')
|
||||
pd.concat([output_score_county_state_merged_df.head(3), output_score_county_state_merged_df.tail(4)], ignore_index=True).to_pickle('data_pipeline/etl/score/tests/snapshots/create_tile_score_data_input.pkl')
|
||||
pd.concat([self.output_score_tiles_df.head(3), self.output_score_tiles_df.tail(4)], ignore_index=True).to_pickle('data_pipeline/etl/score/tests/snapshots/create_tile_data_expected.pkl')
|
||||
```
|
||||
|
||||
### create_tile_score_data_input.pkl
|
||||
1. Set a breakpoint in the transform method in `data_pipeline/etl/score/etl_score_post.py` before the call to
|
||||
`_create_tile_data` and run the post scoring.
|
||||
2. Extract a subset of the `output_score_county_state_merged_df` dataframe. Do not extract the whole score as the file
|
||||
will be too big and the test will run slow. For example, use the following command once the breakpoint is reached
|
||||
to extract a few rows at the top and bottom of the score. This will some capture states and territories.
|
||||
```python
|
||||
pd.concat([output_score_county_state_merged_df.head(3), output_score_county_state_merged_df.tail(3)], ignore_index=True).to_pickle('data_pipeline/etl/score/tests/snapshots/create_tile_score_data_input.pkl')
|
||||
```
|
Binary file not shown.
Binary file not shown.
Loading…
Add table
Add a link
Reference in a new issue