|
| 1 | +# Sidewalk Dataset Preprocessing |
| 2 | + |
| 3 | +In this document we'll describe how we preprocess dataset we're given, and derive what we needed for querying Google's [street view API](https://developers.google.com/maps/documentation/streetview/intro) for sidewalk images. |
| 4 | + |
| 5 | +## About Dataset and Resource |
| 6 | + |
| 7 | +We have our sidewalk and street data from Boston Hack 2018's dataset. The whole dataset can be obtained from project description [here](https://docs.google.com/document/d/1jS3QsgjQLZyYoZzs0WbrA_SrOWAhEUv6Cc_a8X0oHJA/). |
| 8 | + |
| 9 | +The street dataset consists of line segments of streets in Boston; and sidewalk contains shape polygons of planned sidewalk blocks. Both datasets use coordinates under *EPSG:6492*'s ([description](https://epsg.io/6492)) projection settings. |
| 10 | + |
| 11 | +### Coordinate Conversion |
| 12 | + |
| 13 | +Because Google's street view API only accept *WGS84* (worldwide coordinate system, [ref](https://en.wikipedia.org/wiki/World_Geodetic_System)), while the dataset uses another coordinate system, we need to transform location coordinates beforehand. There are several Python libraries supporting coordinate conversion, such as `pyproj`. However, we found that transfomation result from local machine suffers from significant numeral error, which can't be directly applied in any further task. |
| 14 | + |
| 15 | +To address this issue, we use [epsg.io](http://epsg.io/)'s free API to obtain coordinates with higher precision ([API description](https://github.com/klokantech/epsg.io)). We wrote a function that converts a series of locations from one projection system to the other for this task. |
| 16 | + |
| 17 | +### Sidewalk Block Partition |
| 18 | + |
| 19 | +Most of the sidewalk blocks are too long to be included in single image, hence we may want to separate single block into multiple partitions of similar length, then retrieve images separately. So the next thing to do is to know about how many photos to be taken along the sidewalk, along with location of each partition block. We desinged an algorithm to approximate a simpler shape of the sidewalk, then do partition on the simplified shape to obtain requried partitions. |
| 20 | + |
| 21 | +The approximation algorithm is based on the assumpiton that most of the blocks are neary rectangular-shaped. Firstable, we approxmiate the block as a quadrilateral defined by points which obtain extreme value (maximum and minimum) on both coordinate directions. Then, take the longest side of the quadrilateral to decide number of partitions and main heading direction to the block. Last we partitioned the longest edge along with the edge across, and take the midpoint of evenly segmented points on both sides as partition's center. |
| 22 | + |
| 23 | +### Street Matching and Camera Settings |
| 24 | + |
| 25 | +The main problem of the dataset is that the sidewalk dataset doesn't provide any information about which road it belongs to. Also, we need to know relative position between street and sidewalk to make sure the camera is set up on the correct side of target sidewalk in order to obtain valid result. In spatial database applications, the most basic way for various kinds of searching operations (such as intersection and nearest-neighbor search) is to build up index for regtangualr bounding boxes of all itmes, then work on relationship between bounding boxes before any analysis on detailed shapes inside. Another useful clue that can be used to match street segment and sidewalk is using angle difference between their heading directions, since idealy they're nearly parallel to each other. |
| 26 | + |
| 27 | +So this is how we deal with this case. First, we build up an R-tree index for all line segments in the street dataset, then use the center of target sidewalk partition to do nearest-neighbor query (on bounding boxes) for possible candidates, then use angle differenceto find out segment that best fits. |
| 28 | + |
| 29 | +### Obtaining Sidewalk Images from Google's Street View API |
| 30 | + |
| 31 | +Other than solving camera heading, we need to decide other factors such as camera position, pitch rotation, and zooming for querying street view images, while guarantee the area we want appears is large enough on result image. However, obtaining the right camera settings is another researchable problem far beyond our main goal, so we just set all camera positions to be some point away from partition centner at a fixed distance, and use default value in Google's API for other settings. |
| 32 | + |
| 33 | +### Future Work |
| 34 | + |
| 35 | +We're told that the original datset for sidewalks [Link](https://data.boston.gov/dataset/sidewalk-inventory) has more attributes that are crucial to refine all above tasks, such as street it belongs to and sidewalk width. Future updates will combine this dataset so that we can obtain more accurate camera position and settings for the target. |
| 36 | + |
| 37 | +Also, the assumptions we used for generating camera settings can fail in some cases, and there are still lots of possible improvement can be done in the future, such as approximating parameters other than position and heading for the query. |
| 38 | + |
| 39 | +### Dependencies |
| 40 | + |
| 41 | +All the programs are developed under Python3.7, and requre these external libraries other than built-ins: |
| 42 | + |
| 43 | +* `requests`, library for HTTP requests |
| 44 | +* `pyshp`, library for reading Shapefile |
| 45 | +* `rtree`, wrapper library for [libspatialindex](https://libspatialindex.org/) to build up R-tree index. |
| 46 | +* `nvector`, library that provides tools for solving common geographical questions |
| 47 | +* `geopy`, library for geocoding and distance computation |
| 48 | +* `PyGeodesy`, library for geodesy operations |
| 49 | + |
| 50 | +First, run this command to install all dependencies: |
| 51 | + |
| 52 | +``` |
| 53 | +python3.7 -m pip install -r dependencies.txt |
| 54 | +``` |
| 55 | + |
| 56 | +Then follow [this](http://toblerity.org/rtree/install.html) instruction to have `libspatialindex` installed. For macOS users (which is not covered from official website), run `brew install spatialindex` through `homebrew` to install this package. |
| 57 | + |
| 58 | +### How to run |
| 59 | +#### Coordination Conversion |
| 60 | + |
| 61 | +Say you have your Shapefiles for street and sidewalk in directories `ST_DB` and `SW_DB`, and want to output the converted result to `ST_OUT_PATH`, `SW_OUT_PATH`, respectively. Then you can run this command to convert both datasets: |
| 62 | +``` |
| 63 | +python3 dataset_convert.py "ST_DB" "SW_DB" "ST_OUT_PATH" "SW_OUT_PATH" |
| 64 | +``` |
| 65 | + |
| 66 | +#### Partitioning, Query Generation and Obtain Images |
| 67 | + |
| 68 | +To partition converted sidewalk datset and generate query parameters, use `query_generation.py` and run: |
| 69 | + |
| 70 | +``` |
| 71 | +python3.7 query_generation.py [YOUR_PARAMETERS...] |
| 72 | +``` |
| 73 | + |
| 74 | +For details about parameters, check them out by running: |
| 75 | + |
| 76 | +``` |
| 77 | +python3.7 query_generation.py -h |
| 78 | +``` |
| 79 | + |
| 80 | +Running `query_generation` will output three text files: |
| 81 | +- `sidewalk_info.txt`: Information of sidewalk blocks, mainly about how many partitions are made for this block for querying. |
| 82 | +- `metadata.txt`: Metadata for partition queries, includeing their center position, heading direction and the matched street segment. |
| 83 | +- `queries.txt`: Query parameters for Google street view API (in JSON string format) |
| 84 | + |
| 85 | +As for obtaining sidewalk from queries generated, right now we haven't solution to obtain. Since Google's street view API is chraged service, user may consider not requesting all images at once. |
| 86 | + |
| 87 | +However, if you'd like to sample from the set of all queries, then you can try `sample_retrive.py` and `query_sampling.py`, which samples from output result above, filter out ones with valid result from Google's API, and query street view images from the samples. |
| 88 | + |
| 89 | +Google's API require product key and signature secret for forming queries, so you need to store your authetication credentials in a JSON file with this format: |
| 90 | + |
| 91 | +```json |
| 92 | +{ |
| 93 | + "key": "[Your Google API key here]", |
| 94 | + "secret": "[And your API signature secret here]" |
| 95 | +} |
| 96 | +``` |
| 97 | + |
| 98 | +Suppose that you store all outputs from `query_generation` under `partitions` directory, and store the keys in `credentials.json`. Then running the following commands will randomly select 100 valid queries (out of 1000 samples) and store images obtained to `sample_output` directory (with *test_sample_* as filename prefix): |
| 99 | + |
| 100 | +``` |
| 101 | +python3.7 query_sampling.py partitions/queries.txt partitions/metadata.txt 100 credentials.json partitions/samples.txt 1000 |
| 102 | +python3.7 sample_retrive.py partitions/samples.txt credentials.json sample_output test_sample_ [-v] |
| 103 | +``` |
| 104 | + |
0 commit comments