<?xml version="1.0" encoding="UTF-8"?><rss xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:atom="http://www.w3.org/2005/Atom" version="2.0" xmlns:cc="http://cyber.law.harvard.edu/rss/creativeCommonsRssModule.html">
    <channel>
        <title><![CDATA[DataSD - Medium]]></title>
        <description><![CDATA[Helping San Diegans get and use data - Medium]]></description>
        <link>https://medium.com/datasd?source=rss----cd6fdd92a319---4</link>
        <image>
            <url>https://cdn-images-1.medium.com/proxy/1*TGH72Nnw24QL3iV9IOm4VA.png</url>
            <title>DataSD - Medium</title>
            <link>https://medium.com/datasd?source=rss----cd6fdd92a319---4</link>
        </image>
        <generator>Medium</generator>
        <lastBuildDate>Fri, 10 Apr 2026 21:19:50 GMT</lastBuildDate>
        <atom:link href="https://medium.com/feed/datasd" rel="self" type="application/rss+xml"/>
        <webMaster><![CDATA[yourfriends@medium.com]]></webMaster>
        <atom:link href="http://medium.superfeedr.com" rel="hub"/>
        <item>
            <title><![CDATA[Netlify Site of the Week]]></title>
            <link>https://medium.com/datasd/netlify-site-of-the-week-2e9f9580a8c3?source=rss----cd6fdd92a319---4</link>
            <guid isPermaLink="false">https://medium.com/p/2e9f9580a8c3</guid>
            <category><![CDATA[open-data]]></category>
            <category><![CDATA[san-diego]]></category>
            <dc:creator><![CDATA[Maksim Pecherskiy]]></dc:creator>
            <pubDate>Tue, 13 Mar 2018 17:26:05 GMT</pubDate>
            <atom:updated>2018-03-13T17:25:56.031Z</atom:updated>
            <content:encoded><![CDATA[<p>By: Sydney Yockey, on Aug 23, 2017</p><p>As you know by now, our Open Data Portal is hard at work keeping San Diego’s data accessible and up-to-date. As we strive to keep San Diego at the forefront of SmartCity technology, more citizens and city officials continue to recognize the portal’s benefits.</p><p>Our site and team are even turning heads beyond city boundaries! Netlify, an online platform that easily allows users to deploy websites, recently crowned our Open Data Portal the “Netlify Site of the Week.” By using the hassle-free services Netlify offers, we saved the city from spending money on expensive website builders, and saved ourselves from hours of tedious work. We are proud to epitomize “innovation, creativity, and utility,” and join the other sites in Netlify’s hall of fame.</p><p>Check out Netlify’s <a href="https://www.netlify.com/site-of-the-week/san-diego-open-data-portal/?utm_content=buffer41a7d&amp;utm_medium=social&amp;utm_source=twitter.com&amp;utm_campaign=buffer">article</a> and the <a href="https://data.sandiego.gov/">Open Data Portal</a> to read more about how our team aims to make San Diego more efficient and informed.</p><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/0*mwHP1xdDpRD7HhdD.png" /></figure><img src="https://medium.com/_/stat?event=post.clientViewed&referrerSource=full_rss&postId=2e9f9580a8c3" width="1" height="1" alt=""><hr><p><a href="https://medium.com/datasd/netlify-site-of-the-week-2e9f9580a8c3">Netlify Site of the Week</a> was originally published in <a href="https://medium.com/datasd">DataSD</a> on Medium, where people are continuing the conversation by highlighting and responding to this story.</p>]]></content:encoded>
        </item>
        <item>
            <title><![CDATA[A few updates to StreetsSD]]></title>
            <link>https://medium.com/datasd/a-few-updates-to-streetssd-a6b3e5a50610?source=rss----cd6fdd92a319---4</link>
            <guid isPermaLink="false">https://medium.com/p/a6b3e5a50610</guid>
            <category><![CDATA[software-development]]></category>
            <category><![CDATA[open-data]]></category>
            <category><![CDATA[data]]></category>
            <category><![CDATA[street]]></category>
            <dc:creator><![CDATA[Maksim Pecherskiy]]></dc:creator>
            <pubDate>Tue, 13 Mar 2018 17:23:40 GMT</pubDate>
            <atom:updated>2018-03-13T17:23:43.756Z</atom:updated>
            <content:encoded><![CDATA[<p>Hey there everyone! For all the fans of <a href="http://streets.sandiego.gov/">StreetsSD</a> we have some great news!</p><p>We recently completed a small sprint and made a few fixes and improvements.</p><p>We love listening to our users, and one of the most requested features has been a way to type in an address and see what streets have been or will be repaired around where you live or work.</p><p>We implemented an autocompleting search using Mapzen’s awesome <a href="https://mapzen.com/products/search/">search api</a>. Now you can search for a place (like Balboa Park), or an address and it will zoom right in for you:</p><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/0*huoKM8MflBxNOg33.jpg" /></figure><p>We also made a few performance tweaks by moving hosting from Github Pages to Amazon S3 and CloudFront to align with our overall infrastructure and deployment strategy. Hopefully sometime soon we’ll be able to make a few tweaks <a href="https://data.sandiego.gov/stories/portal-speedup/">like we did on data.sandiego.gov</a> to significantly increase the loading speed.</p><p>Finally, there is one more tweak we made, but we’re not ready to talk about it yet, because it needs a bigger piece before it goes into production. However, you can probably find it by <a href="https://github.com/cityofsandiego/streetsSD">looking at the open source code</a> driving StreetsSD.</p><p>As always, we learn, iterate and continuously improve our products. <a href="https://github.com/cityofsandiego/streetsSD/issues">We Love To Hear and Implement your suggestions!</a>!</p><p>But we love pull requests even more!</p><img src="https://medium.com/_/stat?event=post.clientViewed&referrerSource=full_rss&postId=a6b3e5a50610" width="1" height="1" alt=""><hr><p><a href="https://medium.com/datasd/a-few-updates-to-streetssd-a6b3e5a50610">A few updates to StreetsSD</a> was originally published in <a href="https://medium.com/datasd">DataSD</a> on Medium, where people are continuing the conversation by highlighting and responding to this story.</p>]]></content:encoded>
        </item>
        <item>
            <title><![CDATA[City Streetlights Get Smart]]></title>
            <link>https://medium.com/datasd/city-streetlights-get-smart-670e6a30730e?source=rss----cd6fdd92a319---4</link>
            <guid isPermaLink="false">https://medium.com/p/670e6a30730e</guid>
            <category><![CDATA[government]]></category>
            <category><![CDATA[tech]]></category>
            <category><![CDATA[street]]></category>
            <category><![CDATA[san-diego]]></category>
            <dc:creator><![CDATA[Maksim Pecherskiy]]></dc:creator>
            <pubDate>Tue, 13 Mar 2018 17:14:15 GMT</pubDate>
            <atom:updated>2018-03-13T17:14:14.727Z</atom:updated>
            <content:encoded><![CDATA[<p>by: Lorie Cosio Azar</p><p>The Energy &amp; Sustainability Division within the Environmental Services Department has been upgrading City streetlights to SMART adaptive control LED lighting systems.</p><p>Our systems include adaptive control meters, which provide the City with real-time data, energy savings, and asset management and improve lighting throughout the City.</p><p>The map below displays each retrofitted streetlight in the City. Orange dots represent retrofitted induction lights; green dots represent LED lights with adaptive controls; and blue dots represent all other LED lights.</p><p>See the map below!</p><p><a href="https://cityofsandiego.carto.com/u/andrellbower/viz/b0c3ff3f-4a61-4da7-84e3-0e7a134ef067/embed_map&#39;">https://cityofsandiego.carto.com/u/andrellbower/viz/b0c3ff3f-4a61-4da7-84e3-0e7a134ef067/embed_map</a></p><img src="https://medium.com/_/stat?event=post.clientViewed&referrerSource=full_rss&postId=670e6a30730e" width="1" height="1" alt=""><hr><p><a href="https://medium.com/datasd/city-streetlights-get-smart-670e6a30730e">City Streetlights Get Smart</a> was originally published in <a href="https://medium.com/datasd">DataSD</a> on Medium, where people are continuing the conversation by highlighting and responding to this story.</p>]]></content:encoded>
        </item>
        <item>
            <title><![CDATA[San Diego’s Open Data Portal Video]]></title>
            <link>https://medium.com/datasd/san-diegos-open-data-portal-video-86679e8d2698?source=rss----cd6fdd92a319---4</link>
            <guid isPermaLink="false">https://medium.com/p/86679e8d2698</guid>
            <category><![CDATA[datasd]]></category>
            <category><![CDATA[open-data]]></category>
            <category><![CDATA[san-diego]]></category>
            <dc:creator><![CDATA[Maksim Pecherskiy]]></dc:creator>
            <pubDate>Tue, 13 Mar 2018 17:07:34 GMT</pubDate>
            <atom:updated>2018-03-13T17:07:25.778Z</atom:updated>
            <content:encoded><![CDATA[<p>by: Sydney Yockey</p><p>The Data &amp; Analytics program aims to empower both officials and residents with easy access to San Diego’s data. But if people don’t know about the portal, they won’t be able to build apps, hold the City accountable, or make decisions based on data.</p><p>That’s where I came in. As a summer intern focused on community engagement, I took on the task of bringing more attention to the portal. I’m excited to share my last project: a video championing the Open Data Portal. Using concrete examples and minimal technical jargon, I hope to make the portal both comprehensible and intriguing.</p><iframe src="https://cdn.embedly.com/widgets/media.html?src=https%3A%2F%2Fwww.youtube.com%2Fembed%2FvUbNO2PeNeA%3Ffeature%3Doembed&amp;url=http%3A%2F%2Fwww.youtube.com%2Fwatch%3Fv%3DvUbNO2PeNeA&amp;image=https%3A%2F%2Fi.ytimg.com%2Fvi%2FvUbNO2PeNeA%2Fhqdefault.jpg&amp;key=a19fcc184b9711e1b4764040d3dc5c07&amp;type=text%2Fhtml&amp;schema=youtube" width="854" height="480" frameborder="0" scrolling="no"><a href="https://medium.com/media/964b8e26dcb01dd89cdbf66c16f134ae/href">https://medium.com/media/964b8e26dcb01dd89cdbf66c16f134ae/href</a></iframe><img src="https://medium.com/_/stat?event=post.clientViewed&referrerSource=full_rss&postId=86679e8d2698" width="1" height="1" alt=""><hr><p><a href="https://medium.com/datasd/san-diegos-open-data-portal-video-86679e8d2698">San Diego’s Open Data Portal Video</a> was originally published in <a href="https://medium.com/datasd">DataSD</a> on Medium, where people are continuing the conversation by highlighting and responding to this story.</p>]]></content:encoded>
        </item>
        <item>
            <title><![CDATA[Solar Permits Prediction]]></title>
            <link>https://medium.com/datasd/solar-permits-prediction-city-of-san-diego-open-data-portal-131e40c531db?source=rss----cd6fdd92a319---4</link>
            <guid isPermaLink="false">https://medium.com/p/131e40c531db</guid>
            <category><![CDATA[solar]]></category>
            <category><![CDATA[solar-energy]]></category>
            <category><![CDATA[permits]]></category>
            <category><![CDATA[solar-data]]></category>
            <dc:creator><![CDATA[Maksim Pecherskiy]]></dc:creator>
            <pubDate>Tue, 13 Mar 2018 17:05:22 GMT</pubDate>
            <atom:updated>2018-03-13T17:29:14.216Z</atom:updated>
            <content:encoded><![CDATA[<p>by: Randy Askin</p><p>The <a href="https://www.ge.com/digital/minds-machines">2017 GE Minds &amp; Machines</a> conference featured an “Appathon” where teams were challenged to come up with a working application in a day and a half. The challenge theme was all about decarbonization, decentralization, and digitization in the global energy sector. Each team was provided with a Predix Kit that included an Intel Nook and various sensors, sample data from power grid feeders in Ireland, and access to the <a href="http://developer.currentbyge.com/cityiq">CityIQ API</a> for accessing the intelligent street lamps in the city of San Diego.</p><p>Our team took inspiration from <a href="https://www.google.com/get/sunroof/data-explorer/">Google’s “Project Sunroof,”</a> which analyzes satellite imagery and 3D renderings of rooftops in various cities to calculate potential savings from solar energy. Our idea was to take this data for solar potential, subtract the existing installations extrapolated from the <a href="https://data.sandiego.gov/datasets/solar-permits/">permit dataset on data.sandiego.gov</a>, and use the city budget and other financial information to provide city planners with strategic information around solar subsidies.</p><p>We also built a regression model using <a href="http://predix.io/">Predix</a> analytics and services to calculate the best way to balance residential, commercial, and industrial solar installations to arrive at the city’s goal of carbon neutrality by 2035. Our basic finding was that, even though commercial and industrial buildings generally have larger rooftop footprints, the most economical and scalable solution would be to target widespread residential installations. We then used income data for households in the area to estimate the relative affordability and required subsidies for various income brackets in San Diego.</p><h3>Check it out!</h3><ul><li><a href="https://devpost.com/software/mm17">Devpost</a></li><li><a href="https://solar-sd.run.aws-usw02-pr.ice.predix.io/">Running Application</a></li><li><a href="https://github.com/randyaskin/mm17">Github</a></li></ul><p><em>Originally published at </em><a href="https://data.sandiego.gov/stories/solar-permits-prediction/"><em>https://data.sandiego.gov/stories/solar-permits-prediction/</em></a><em>.</em></p><img src="https://medium.com/_/stat?event=post.clientViewed&referrerSource=full_rss&postId=131e40c531db" width="1" height="1" alt=""><hr><p><a href="https://medium.com/datasd/solar-permits-prediction-city-of-san-diego-open-data-portal-131e40c531db">Solar Permits Prediction</a> was originally published in <a href="https://medium.com/datasd">DataSD</a> on Medium, where people are continuing the conversation by highlighting and responding to this story.</p>]]></content:encoded>
        </item>
        <item>
            <title><![CDATA[Remembering Arnaud Vedy]]></title>
            <link>https://medium.com/datasd/remembering-arnaud-vedy-e09e687f800?source=rss----cd6fdd92a319---4</link>
            <guid isPermaLink="false">https://medium.com/p/e09e687f800</guid>
            <category><![CDATA[data-science]]></category>
            <dc:creator><![CDATA[Andrell Bower]]></dc:creator>
            <pubDate>Wed, 17 Jan 2018 21:27:43 GMT</pubDate>
            <atom:updated>2018-01-17T21:30:00.141Z</atom:updated>
            <content:encoded><![CDATA[<p>If you’ve paid much attention to smart city or open data or data science in San Diego local government, you probably ran across Arnuad Vedy.</p><p>Arnaud was the City of San Diego’s first data scientist. He joined the Data &amp; Analytics team in October 2016, and right away he was speaking at events, teaching data science and Python at meetups, and spending his free time with community nerds fiddling around with tech gadgets like <a href="https://www.thethingsnetwork.org/docs/">LoRa</a>. And, obviously, solving problems in the City using data and technology.</p><p>One time, he did this:</p><h3>Arnaud Vedy on Twitter</h3><p>NYC on my desk using #Google #ARCore and #pixelphone. #lunchtimeprojects #ar https://t.co/q3ud8TVTxf</p><p>We lost Arnaud in a motorcycle accident in early November. I was not at all prepared for how devastating it is to lose a co-worker, especially one you work closely with and is also a friend. He was smart and talented, but not condescending. I could, and did, ask many dumb questions about Python, GIS, gadgets, France, artisan pizza, and coffee. He answered my questions patiently and never made me feel stupid, or like I was wasting his time. We ate many lunches together, along with our boss Maksim, and chatted a lot about code, and life.</p><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/1*-AJFpzMPeU8qgmaly8V2CA.jpeg" /><figcaption>Here is the Data &amp; Analytics team checking out Mission Trails Park. We plan to publish trails and species monitoring data, and we wanted to know a little bit more about the City’s Open Space parks. (L-R: Maksim Pecherskiy, Andrell Bower, Arnaud Vedy)</figcaption></figure><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/1*rHgy9RUGr0ftWs4MvOXbbg.jpeg" /><figcaption>Maksim, left, and Arnaud, right, liked to embark on random, zany adventures together. Here they are about to start a 56-mile bicycle race. Arnaud was already an experienced cyclist, but Maksim hadn’t even really trained. Typical Maksim.</figcaption></figure><p>Data &amp; Analytics is a 3-person team within a larger department of only 13, that we affectionately refer to as PandA (Performance &amp; Analytics). Arnaud’s death was a blow to the entire department, and to people in various other departments who had worked with him, not just the two of us left in Data &amp; Analytics. The City Council even adjourned a meeting in Arnaud’s memory.</p><p>Council member Chris Ward read about Arnaud’s life and his accomplishments with the City. At this point, I was done crying about it all the time, and it made me really happy to hear him remembered and have him memorialized in the official record of the City of San Diego. Well, I was almost done crying at that point.</p><p>Anyway, here’s the clip:</p><iframe src="https://cdn.embedly.com/widgets/media.html?src=https%3A%2F%2Fwww.youtube.com%2Fembed%2F-DNO71PiS1g%3Ffeature%3Doembed&amp;url=http%3A%2F%2Fwww.youtube.com%2Fwatch%3Fv%3D-DNO71PiS1g&amp;image=https%3A%2F%2Fi.ytimg.com%2Fvi%2F-DNO71PiS1g%2Fhqdefault.jpg&amp;key=a19fcc184b9711e1b4764040d3dc5c07&amp;type=text%2Fhtml&amp;schema=youtube" width="854" height="480" frameborder="0" scrolling="no"><a href="https://medium.com/media/6318031417b5bd406caea52cf6e982da/href">https://medium.com/media/6318031417b5bd406caea52cf6e982da/href</a></iframe><p>We also have this certificate hanging in the hall leading to our wing of the 8th floor of the City Administration Building.</p><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/1*uGVh8ttn0b5vWEA9YcrMzw.jpeg" /><figcaption>San Diego City Council members signed this certificate after they adjourned a meeting in Arnaud’s memory. We have it hanging among various press clippings in the hallway that leads to our area. I’m happy to walk by it every workday and think about my friend for moment. Thanks, Council.</figcaption></figure><p>You’ll see some posts on this blog that I published under my account after his death. He absolutely wanted you guys to learn all kinds of geeky stuff about GIS and data science, so I had to get these out into the world and share a bit of his passion and knowledge.</p><p>We talked all the time about how both of us started off in other disciplines (myself in journalism, him in Geography) and found our way to programming, data analysis, and data visualization. Our code was a little sloppy sometimes, but we were both excited to be doing something that we discovered such a passion for that we had spent our free time teaching ourselves the skills.</p><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/1*qci8hbDe_paoDGezkj9_-A.jpeg" /><figcaption>Arnaud shows maps to kids. He was really good at teaching people how to work with data and maps. I learned a ton from him, and I’ve been posting more stuff he wrote on this blog under my account. Maybe one of these kids will grow up to be a map whiz because of this cool map they saw.</figcaption></figure><p>So, if you really want to channel Arnaud, <a href="https://www.teepublic.com/t-shirt/508381-i-just-want-to-drink-coffee-create-stuff-and-sleep">wear this shirt</a>. I proudly display a magnet version of that shirt in my not-at-all-dull cubicle.</p><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/1*DQU-HPlCMQDd4z-IiI3H3g.jpeg" /><figcaption>One of my other coworkers bought these magnets because he liked Arnaud’s shirt. I have it here to channel Arnaud when I’m working in my excellent and otherwise inspiring cubicle.</figcaption></figure><p>Even better, build yourself a crazy powerful Linux computer that can handle Lidar data, spend your Saturdays coding, and always make maps with dark backgrounds (just kidding, channel me and put some thought into what your color choices communicate).</p><img src="https://medium.com/_/stat?event=post.clientViewed&referrerSource=full_rss&postId=e09e687f800" width="1" height="1" alt=""><hr><p><a href="https://medium.com/datasd/remembering-arnaud-vedy-e09e687f800">Remembering Arnaud Vedy</a> was originally published in <a href="https://medium.com/datasd">DataSD</a> on Medium, where people are continuing the conversation by highlighting and responding to this story.</p>]]></content:encoded>
        </item>
        <item>
            <title><![CDATA[Using spatial data and Mapshaper to answer questions]]></title>
            <link>https://medium.com/datasd/using-spatial-data-and-mapshaper-to-answer-questions-6176b4cab1f6?source=rss----cd6fdd92a319---4</link>
            <guid isPermaLink="false">https://medium.com/p/6176b4cab1f6</guid>
            <category><![CDATA[data-science]]></category>
            <dc:creator><![CDATA[Andrell Bower]]></dc:creator>
            <pubDate>Mon, 08 Jan 2018 23:00:33 GMT</pubDate>
            <atom:updated>2018-01-17T21:32:34.445Z</atom:updated>
            <content:encoded><![CDATA[<p>Sure, you could simply display spatial data on a map. But you’d be limiting yourself. Spatial data can also help you answer questions about all sorts of topics. To demonstrate it, let’s answer a question with spatial data.</p><h3>The question</h3><p>How many times did someone report a pothole to the City through the Get It Done applications in 2017 for the Downtown area?</p><h3>The tool</h3><p>Although we could use various GIS tools to answer this question, here we are using <a href="http://mapshaper.org/">Mapshaper</a>. Why? Because you will be able to reproduce the workflow without having to install anything, and you’ll be surprised that an online tool can handle this sort of analysis.</p><h3>The steps</h3><ol><li>Download data</li></ol><p>To start with, download the following two datasets from our <a href="https://data.sandiego.gov/">open data portal</a>:</p><ul><li><a href="https://seshat.datasd.org/get_it_done_311/get_it_done_311_requests_datasd.csv">Get It Done requests</a></li><li><a href="http://seshat.datasd.org/sde/cmty_plan/cmty_plan_datasd.zip">Community Planning District boundaries</a></li></ul><p>2. Import data into Mapshaper</p><p>Go to Mapshaper and select/import the Get It Done dataset. First thing’s first, let’s rename the dataset to ‘gid’ for easier processing by opening the console (top-right button) and typing the following command: -rename-layers gid</p><p>At this point, this dataset is not interpreted as spatial data, and all you can see is a grid of points. To solve this issue, we are going to tell Mapshaper which columns of the .csv file contain spatial data by typing the following command into the console: -points x=long y=lat</p><p>If you’re curious about all the commands you can use in Mapshaper, check out this <a href="https://github.com/mbloch/mapshaper/wiki/Command-Reference#-calc">reference</a>.</p><p>3. Filter data</p><p>Two elements of the question we are attempting to answer can be addressed by filtering the Get It Done data:</p><ul><li>‘In 2017’ with the following command: -filter ‘requested_datetime &gt; “2016–12–31” &amp;&amp; requested_datetime &lt; “2018–01–01”’</li><li>‘Requests for potholes’ with the following command: -filter “service_name == ‘Pothole’”</li></ul><p>Now your screen should look like this:</p><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/1*1PmqmkiRLk83h8t1sW_Yxw.png" /><figcaption>Mapshaper now shows just pothole service requests made in 2017.</figcaption></figure><p>4. Import more data into Mapshaper</p><p>To answer the last element of the question, ‘in the Downtown area’, we need to perform a spatial join between the neighborhoods from the Community Planning shapefile and the filtered Get It Done data.</p><p>To do so, first import the Community shapefile by clicking on ‘gid’ (top-center), then ‘add a file’, then ‘import’. Rename the layer to ‘neighborhoods’: -rename-layers neighborhoods target=cmty_plan_datasd</p><p>One last thing — reprojection. If you type -info in the console, you will notice that both spatial layers are using different coordinates systems. All you need to do here is to put both layers in the same system with the following command: -proj wgs84</p><p>5. Perform a spatial join</p><p>You are now all set to operate a spatial join between the two layers. With a spatial join, we will end up with a column that contains the neighborhood name from the ‘neighborhoods’ layer for each row of the ‘gid’ layer.</p><p>To perform the join, select the ‘gid’ layer from the layer switcher (top-center), and type the following line: -join neighborhoods fields=‘cpname’</p><p>If you now explore the gid data, you will notice an extra field at the bottom with the neighborhood information for each pothole:</p><figure><img alt="" src="https://cdn-images-1.medium.com/max/969/1*dXUa9kTbWtOAF-v5De9eWw.png" /><figcaption>Each row, which represents one pothole service request, now has an associated neighborhood name.</figcaption></figure><p>6. Answer the question</p><p>You can now answer the initial question with a last command: -calc ‘count()’ where=“cpname == ‘Downtown’”</p><h3>The findings</h3><p>At the time of this writing, 814 potholes were reported with the Get It Done applications for the downtown area in 2017. Put that in your report or slide deck!</p><p>Do you have more questions? I did.</p><p>What about Uptown (1059), or Golden Hill (194), or La Jolla (725)? What is Uptown, anyway (Mission Hills + Hillcrest?)? Have the number of potholes reported per month citywide gone down (yes! See the table below).</p><iframe src="https://cdn.embedly.com/widgets/media.html?src=https%3A%2F%2Fairtable.com%2Fembed%2Fshre8xjSeRhcIw8Fa&amp;url=https%3A%2F%2Fairtable.com%2Fshre8xjSeRhcIw8Fa&amp;image=https%3A%2F%2Fstatic.airtable.com%2Fimages%2Foembed%2Fairtable.png&amp;key=a19fcc184b9711e1b4764040d3dc5c07&amp;type=text%2Fhtml&amp;schema=airtable" width="800" height="533" frameborder="0" scrolling="no"><a href="https://medium.com/media/3bda8c54ae88fbcb1e7a6a3ea88d0c2a/href">https://medium.com/media/3bda8c54ae88fbcb1e7a6a3ea88d0c2a/href</a></iframe><p>For that last one, I had to run multiple calculations and use requested_datetime in the where clause. There might be a better way to do that, but I don’t know it.</p><p>What I do know, is that spatial data has all the answers.</p><p><a href="https://medium.com/datasd/remembering-arnaud-vedy-e09e687f800"><em>Arnaud Vedy</em></a><em> contributed to this article.</em></p><img src="https://medium.com/_/stat?event=post.clientViewed&referrerSource=full_rss&postId=6176b4cab1f6" width="1" height="1" alt=""><hr><p><a href="https://medium.com/datasd/using-spatial-data-and-mapshaper-to-answer-questions-6176b4cab1f6">Using spatial data and Mapshaper to answer questions</a> was originally published in <a href="https://medium.com/datasd">DataSD</a> on Medium, where people are continuing the conversation by highlighting and responding to this story.</p>]]></content:encoded>
        </item>
        <item>
            <title><![CDATA[What to do with GIS files]]></title>
            <link>https://medium.com/datasd/what-to-do-with-gis-files-f974b1b6d498?source=rss----cd6fdd92a319---4</link>
            <guid isPermaLink="false">https://medium.com/p/f974b1b6d498</guid>
            <category><![CDATA[gis]]></category>
            <dc:creator><![CDATA[Andrell Bower]]></dc:creator>
            <pubDate>Wed, 03 Jan 2018 00:37:57 GMT</pubDate>
            <atom:updated>2018-01-17T21:32:07.888Z</atom:updated>
            <content:encoded><![CDATA[<p>Most of the spatial data on our <a href="https://data.sandiego.gov/">open data portal</a> is available as a shapefile (.shp) — a GIS vector data format developed by Esri. Unlike tabular data (.csv, .xls, etc.), GIS vector data cannot be opened and analyzed with Excel or any traditional spreadsheet software. So, how should you open these files?</p><h3>The online way</h3><p>The quickest way to open and visualize the content of a shapefile is to use a tool called <a href="http://mapshaper.org/">Mapshaper</a>.</p><p>Simply drag the freshly downloaded .zip file from our portal into the web interface, click ‘Import’ and visualize your data! By clicking on the ‘i’ at the top-right of the interface, you will be able to view the attributes of each feature as you hover it.</p><p>More adventurous / advanced users are encouraged to read the project’s wiki. Indeed, Mapshaper can do way more than open shapefiles. You can join, dissolve, reproject, simplify, and export as well as perform other essential GIS vector operations.</p><h3>The desktop way</h3><p>If you do not mind installing software on your computer, <a href="https://qgis.org/en/site/forusers/download.html">QGIS</a> is the way to go. Thanks to its numerous plugins and toolboxes, this free and open source software (available for Windows / Mac / Linux) will cover most of your GIS needs. Opening one of our shapefiles is as easy as selecting it in the file explorer (left bar). Wondering what to do next, now that you have opened a file? Check out the <a href="https://docs.qgis.org/2.18/en/docs/training_manual/index.html">QGIS training manual</a> and <a href="https://docs.qgis.org/2.18/en/docs/user_manual/">QGIS user guide</a>!</p><h3>The Python way</h3><p>If you are a developer or a data scientist, you are probably dealing with tabular data using Python. When it comes to spatial data, you can take advantage of plenty of Python packages to open, edit, and process shapefiles.</p><p>Here is a non-exhaustive list of packages:</p><ul><li><a href="https://pypi.python.org/pypi/pyshp">pyshp</a></li><li><a href="http://www.gdal.org/">GDAL</a></li><li><a href="https://pypi.python.org/pypi/Shapely">Shapely</a></li><li><a href="https://pypi.python.org/pypi/Fiona">Fiona</a></li><li><a href="http://geopandas.org/">GeoPandas</a></li></ul><p><em>This article was originally written by </em><a href="https://medium.com/datasd/remembering-arnaud-vedy-e09e687f800"><em>Arnaud Vedy</em></a></p><img src="https://medium.com/_/stat?event=post.clientViewed&referrerSource=full_rss&postId=f974b1b6d498" width="1" height="1" alt=""><hr><p><a href="https://medium.com/datasd/what-to-do-with-gis-files-f974b1b6d498">What to do with GIS files</a> was originally published in <a href="https://medium.com/datasd">DataSD</a> on Medium, where people are continuing the conversation by highlighting and responding to this story.</p>]]></content:encoded>
        </item>
        <item>
            <title><![CDATA[Geo data: up-to-date and in multiple formats]]></title>
            <link>https://medium.com/datasd/our-geo-data-has-become-more-relevant-city-of-san-diego-open-data-portal-8cf8f9d2d65e?source=rss----cd6fdd92a319---4</link>
            <guid isPermaLink="false">https://medium.com/p/8cf8f9d2d65e</guid>
            <category><![CDATA[open-data]]></category>
            <dc:creator><![CDATA[Andrell Bower]]></dc:creator>
            <pubDate>Thu, 30 Nov 2017 18:19:53 GMT</pubDate>
            <atom:updated>2018-01-17T21:33:07.714Z</atom:updated>
            <content:encoded><![CDATA[<p>Dear readers / followers, we have some exciting news! Our geographic data is now integrated with <a href="https://data.sandiego.gov/stories/why-data-automation-matters-data-portals/">Poseidon</a>, our data automation system.</p><h3>Current geo data</h3><p>It means you now have access to the most current geographic data at all times. Some data sources are updated every month. For others, this process occurs on a daily basis. No matter the schedule of the dataset, our portal will now automatically update along with it. This ensures that as a user of the portal, you will be retrieving the most up-to-date data available.</p><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/0*clF9o_oConb0KbwU.jpg" /></figure><h3>New formats</h3><p>Until recently, we were sharing geo data exclusively as shapefiles — one of the most commonly used format for geo data. Nevertheless, there are a few reasons why it should not be the only format of geo data offered on an open data portal:</p><ul><li>It is not an open standard format.</li><li>It is actually multiple files. These files need to be zipped, making it inconvenient to share.</li><li>It is not a practical format for web developers. Few web mapping javascript libraries support shapefiles natively, which usually leads to data conversion.</li><li>It is not the most practical format for non-technical users. It requires software installation to be visualized / used for data analysis.</li></ul><p>For these reasons, we have decided to provide you with additional geo data formats. For now (and we will be adding more) they are: GeoJSON, TopoJSON, and Geobuf.</p><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/0*BkVoQPWEMyZR9GwQ.png" /></figure><ul><li><strong>GeoJSON</strong>: This is the most common alternative to shapefiles. <a href="http://geojson.org/">GeoJSON</a> is an open standard format based on the JSON specification. It is supported by most desktop and web based mapping/GIS solutions. Geo data in this format is also human-readable. This means it can be opened with a text editor, which make it very easy to use. The only drawback with GeoJSON when compared to shapefile is size. This is a consequence of GeoJSON’s lack of topology support.</li><li><strong>TopoJSON</strong>: This <a href="https://github.com/topojson/topojson">format</a> is an extension of GeoJSON that encodes topology. As a result, TopoJSON files that contain line and polygon geometries are significantly smaller than both GeoJSON and shapefiles.</li><li><strong>Geobuf</strong>: <a href="https://github.com/mapbox/geobuf">Geobuf</a> is a relatively new format introduced by Mapbox. Geobuf provides lossless compression of GeoJSON data into <a href="https://developers.google.com/protocol-buffers/">protocol buffers</a>. It is fast to decode and very light, making it attractive for developers who want to use large geo datasets in their web applications. Here is a comparative table to demonstrate the compression power of the Geobuf specification on our data:</li></ul><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/0*jjhwBGEdYgewFZCX.png" /></figure><p>So, you now have access to the most current geo data in multiple formats and we are pretty excited about it. What will you do with it? Let us know at <a href="https://data.sandiego.gov/stories/gis-update/data@sandiego.gov">data@sandiego.gov</a> and stay tuned for more announcements!</p><p><em>Originally written by </em><a href="https://medium.com/datasd/remembering-arnaud-vedy-e09e687f800"><em>Arnaud Vedy</em></a><em> and published at </em><a href="https://data.sandiego.gov/stories/gis-update/"><em>https://data.sandiego.gov</em></a></p><img src="https://medium.com/_/stat?event=post.clientViewed&referrerSource=full_rss&postId=8cf8f9d2d65e" width="1" height="1" alt=""><hr><p><a href="https://medium.com/datasd/our-geo-data-has-become-more-relevant-city-of-san-diego-open-data-portal-8cf8f9d2d65e">Geo data: up-to-date and in multiple formats</a> was originally published in <a href="https://medium.com/datasd">DataSD</a> on Medium, where people are continuing the conversation by highlighting and responding to this story.</p>]]></content:encoded>
        </item>
        <item>
            <title><![CDATA[A better open data portal]]></title>
            <link>https://medium.com/datasd/a-better-open-data-portal-city-of-san-diego-open-data-portal-ccc450e76872?source=rss----cd6fdd92a319---4</link>
            <guid isPermaLink="false">https://medium.com/p/ccc450e76872</guid>
            <category><![CDATA[open-data]]></category>
            <dc:creator><![CDATA[Andrell Bower]]></dc:creator>
            <pubDate>Thu, 30 Nov 2017 18:15:46 GMT</pubDate>
            <atom:updated>2018-03-13T17:33:46.461Z</atom:updated>
            <content:encoded><![CDATA[<p>We are excited — and somewhat exhausted — to present to San Diegans a new and improved Open Data portal today.</p><p>The portal has a fresh look, but more importantly, we rebuilt the technology behind the portal and upgraded our workflow for keeping data up-to-date.</p><h3>Just yesterday</h3><figure><img alt="" src="https://cdn-images-1.medium.com/max/623/0*xi--XxpZtHxT5y4-.jpg" /></figure><p>We initially launched our data portal by buying a ready-made portal product from a company. As our team was newer, this made a lot of sense, and the portal the vendor made for us was exactly what we needed at that time.</p><p>But we’ve always had a long-term vision for open data that relied on high-tech strategies. We never wanted to publish data that was manually updated. We want City employees to spend their time using data to serve residents, not trying to remember to update data on an open data portal. We also want our data to be put to use helping the City perform better. This means the data needs to be accessible not only for simple downloads, but also for alerting, prediction, and other advanced analytics.</p><p>Ready-made portal products are more geared toward a manual update process, so when we started working to automate data updates late last year, our automation solution didn’t work as well with our portal product as we had hoped.</p><p>Normally, a client uploads a dataset to its cloud, and then that dataset is available for download, for preview in the browser, and through its data API.</p><p>But we were not hosting our data in the cloud service our portal provider offered. We uploaded to a different cloud service and then linked the files to the portal. Not using a direct upload limited our ability to show a preview in the browser or make the data API available without manual intervention. We also couldn’t trigger a change in the extremely important metadata field of “Date Updated” without a manual upload.</p><p>Although a ready-made portal product is convenient, it is also expensive. Adding functionality and features as we sought continuous improvement in our program would end up costing taxpayers money that could be better spent.</p><p>Taking all of these factors together, we decided to take San Diego’s open data program to another level and build a portal ourselves. We have a team with the technical expertise to run our own data portal, so we went for it!</p><h3>Today</h3><figure><img alt="" src="https://cdn-images-1.medium.com/max/595/0*g_Kf4kFiYoDYYeTQ.jpg" /></figure><p>At a minimum, a data portal is just a website. All it really needs is a page with a list of datasets (the catalog) and individual pages for each dataset that contain links to download the data. A ready-made portal product is a website with an admin interface that makes it easy for just about anybody to manually update data.</p><p>The reality is that some open data programs do depend on employees to upload data manually. This is because open data became a thing long after government and technology were invented. In fact, doing open data well is very, very, very complex (did we say very?). Not all departments and divisions of a government use the same database technology. Government information is housed in hundreds of different systems or even on spreadsheets in shared drives. Some data even comes from third-parties that are contracted to provide services on behalf of a government, and therefore maintain data on their own servers.</p><p>Our team spent a significant amount of time writing code to build our automation process that pulls in data from our City’s various systems, transforms it, and puts it into the our cloud. We could have probably scripted a solution that would copy files from our cloud to the portal cloud. But it would have taken some work, and the effort would be hampered by the fact that we don’t have control over a product made by an outside company. As they upgraded their product, our script might have broken over, and over again.</p><p>A portal we created that worked perfectly with our automation process was the obvious solution. However, we can’t take ALL the credit for this new portal. We based our new portal on an open-source project by <a href="http://github.com/timwis">Tim Wisniewski</a>, Chief Data Officer for the City of Philadelphia. The open-source portal he created, called <a href="https://github.com/timwis/jkan">JKAN</a>, was instrumental in getting us up and running with our own version. (Thanks, Tim!!!)</p><h3>Tomorrow+</h3><p>Moving forward, this data portal is the new hub for all open data and analytics projects for the Performance &amp; Analytics Department. We used to maintain a separate informational site, datasd.org, that we no longer need since we have full control over this one. Keep checking back for the latest news, how-tos, data stories, visualizations, and of course, the most up-to-date City data.</p><p>If you are intrigued by the work we’ve done, our portal is also open-source! <a href="https://github.com/cityofsandiego/seaboard">Check out our code</a>, copy it, and make your own data portal, aka website with a catalog. We use <a href="https://jekyllrb.com/docs/home/">Jekyll</a>, a static site generator, to manage all the files and pages of our portal.</p><p>You also need to check out our <a href="https://github.com/mrmaksimize/docker-airflow">automation process</a>. We are using <a href="https://github.com/apache/incubator-airflow">Airflow</a>, another open-source tool, originally developed to improve workflow for employees of companies that crunch data.</p><p>Basically, each little task that we need to complete when we extract and transform data is handled with bits of Python code. Examples of these tasks include connecting to a database, pulling a file from one of the City’s shared drives, detecting the current date and using that to grab the correct time interval of data, etc.</p><p>Airflow keeps all these bits organized into a hierarchy of jobs, and we get alerts if anything fails. We can also use Airflow to schedule automatic kick-off of jobs, which is one of the major headaches of automation. Keep an eye on this blog for more in-depth and technical explanations of how we are using Airflow.</p><p>If you are a person who was using the data API from our previous portal, you will instead have to work with the link to the csv. We also had to change how we stored data in our cloud, so please plan on updating any previous links to csv files by the end of March. You can get updated links by checking the dataset pages on this new portal.</p><p>We hope you love this new portal, and we especially hope you love having automatically updated data! Let us know what you think.</p><img src="https://medium.com/_/stat?event=post.clientViewed&referrerSource=full_rss&postId=ccc450e76872" width="1" height="1" alt=""><hr><p><a href="https://medium.com/datasd/a-better-open-data-portal-city-of-san-diego-open-data-portal-ccc450e76872">A better open data portal</a> was originally published in <a href="https://medium.com/datasd">DataSD</a> on Medium, where people are continuing the conversation by highlighting and responding to this story.</p>]]></content:encoded>
        </item>
    </channel>
</rss>