Crunchy Data reposted this
When was the last time you truly tested the product your team is building? Last week, last month, last year? At Crunchy Data, we embrace “dogfooding” – using our own products as our customers would. While working on the new Spatial Analytics in Crunchy Bridge, I decided to put it to test on a personal project: mapping Greece’s 15,000 km coastline to identify the best spots for water sports based on 10 years of hourly weather data and precise coastal geometry. As a result, I now have a list of hidden gems to visit with my surf for next year based on the period of the year and time of the day :). More immediately useful, though, are the insights I gained into geospatial workload challenges—where our product excels and where it can improve. Combined with feedback from other dogfooding users, like Paul Ramsey's vehicle routing with Overture data example, and real customer input, these insights guide us to solve real-world problems, unlike many teams that start with solutions and later search for problems. Here are some highlights from my experience. Let me know if these resonate: Easily Accessing Geospatial Data: Support for various geospatial file formats made it easy to load things like coastal geometries and water quality data from various sources (e.g. shapefile, geoparquet) Integrating Cloud Native Data: By creating foreign tables that pointed to online weather data, I bypassed bulky downloads and storage – selecting only the data I needed, joining it with local geospatial fact tables and writing out the relevant data directly to S3 in compressed parquet format. Query performance: Query pushdowns to DuckDB improved processing speed of some PostGIS functions by 100%, though I had to keep certain considerations in mind to get the best speed up (e.g., avoiding coordinate projections within the "push-downable" queries). Seamless compatibility: I worked with both local PostgreSQL tables and foreign tables pointing to parquet files in S3. This allowed me to get best of both worlds - convenience of keeping compressed data in object storage and querying it with a powerful vectorized engine; making full use of PostGIS capabilities such as spatial indexes. I was also able to visualize my data in Qgis from both types of tables with its built-in Postgres integration. Honestly, without Crunchy Bridge for Analytics (CBA), I don’t think I could take my fun little project to this scale. I would have had to spend a lot more time accessing, loading and transforming data, spend much more money in storing large data files and keeping computational resources running. Thankfully, the benefits I’ve seen in my tests have applied to real-world workloads for our customers as well, who say that CBA helped them to streamline complex data pipelines, make full use of their existing Postgres knowhow / application logic and avoid complex integrations. https://lnkd.in/dxGXytVp #postgres, #postgis, #crunchydata, #geospatial, #duckdb