Show HN: What's in My Location History?
The article explores Google Maps location history and Google Fit data analysis using Python scripts. It reveals discrepancies, accuracy limitations, and personal reflections on walking patterns and lifestyle changes.
Read original articleThe article discusses the author's exploration of their Google Maps location history data, obtained through Google Takeout in JSON format. The data includes raw location packets from the author's phone, processed locally and by Google servers. The article also mentions the use of Google Fit to track physical activity, with data downloaded in TCX format. A Python script is used to analyze and visualize the data, revealing insights such as walking speed averages and distances over time. Discrepancies between Maps and Fit data are highlighted, showing variations in recorded activities. The author reflects on the data's accuracy and limitations, noting instances where the data may not fully represent their activities. Additionally, the author shares observations on their walking patterns over different periods, correlating them with personal experiences like school years and lifestyle changes. The article provides a detailed look into the analysis of location and activity data obtained from Google services.
Related
The unbearable sorrow of Apple Weather
The author criticizes Apple Weather iOS app for inconsistent temperature data presentation, focusing on alignment and color issues in temperature bars. Adjusting text size improved alignment, but concerns remain about decision-making process.
Google: Stop Burning Counterterrorism Operations
Google's Project Zero and TAG exposed a U.S.-allied government's counterterrorism operation, sparking debate on the impact of revealing such information. Cyber operations play a crucial role in counterterrorism efforts, balancing intelligence gathering with minimizing risks.
Leaking URLs to the Clown
The author describes leaking URLs during Mac app testing, with a unique URL receiving requests from a random "cloud" service every three hours. This raises privacy concerns and highlights potential risks for users.
Google has been lying about their search results [video]
A leak from Google's GitHub shows the search algorithm tracks user clicks and time on pages, raising concerns about search result accuracy, treatment of smaller websites, and SEO strategies.
How I scraped 6 years of Reddit posts in JSON
The article covers scraping 6 years of Reddit posts for self-promotion data, highlighting challenges like post limits and cutoffs. Pushshift is suggested for Reddit archives. Extracting URLs and checking website status are explained. Findings reveal 40% of sites inactive. Trends in online startups are discussed.
Very impressive system that had a big budget in the mid 2010s (mostly built by zurich, iirc). They even hired a bunch of tvcs to walk around movie theaters and scan the wifi ssids, so timeline could show you what movie you saw. It had a photos integration as well that would show the pics you took that day. All sorts of plans for more delightful features like that. I think the value prop was that deeply integrating people's memories made maps a stickier product.
The investment and headcount started getting cut post-pandemic, like everything else at Google. Lots of team churn, not just on Timeline but on hulk and the semantic location service which undergirded it. When I was last there SLS was literally 1 guy who either could not or would not leave. Those services became abandonware, along with all the flumes and postwrite processors responsible for cleaning up the data and enforcing heuristics. Exactly 0 people on the web UI - some of the directories were literally un-reviewable (the code owners had left and no one in geo had MPA-approval). That decay led to a noticeable decline in the accuracy of reported trips. Users weren't happy, angry reports started piling up about inaccurate/missing trips. It was embarassing. Timeline was moved to the back burner and the idea of being a cute time capsule for users no longer aligned with the AI maximalists.
By '22 the investment and headcount was slashed. IMO the ODLH death march was as much about throwing in the towel for Timeline as a product as it was about getting location data off of Google's servers.
They frame it positively, of course ("now your data lives on your device!") but AFAICT it's all downside. I can't browse my location history on a nice big screen, and (very annoyingly) the app does not let you view your aggregate history over the span of a month or year -- only a single day. Plus, if you lose your phone, you lose all that precious data, unless you configure the app to automatically sync your history to Google's cloud...wait what? Wasn't not doing that the whole point? Just baffling.
Hundreds of thousands of these per person, stored forever in a data centre somewhere.
Imagine how much carbon could have been saved by just fixing this to 1-2 decimal places.
The nice part is that it's not some data silo but that it supports open formats and you can import / export everything very easily.
I wonder if the author actually tests the site with Netscape!
Related
The unbearable sorrow of Apple Weather
The author criticizes Apple Weather iOS app for inconsistent temperature data presentation, focusing on alignment and color issues in temperature bars. Adjusting text size improved alignment, but concerns remain about decision-making process.
Google: Stop Burning Counterterrorism Operations
Google's Project Zero and TAG exposed a U.S.-allied government's counterterrorism operation, sparking debate on the impact of revealing such information. Cyber operations play a crucial role in counterterrorism efforts, balancing intelligence gathering with minimizing risks.
Leaking URLs to the Clown
The author describes leaking URLs during Mac app testing, with a unique URL receiving requests from a random "cloud" service every three hours. This raises privacy concerns and highlights potential risks for users.
Google has been lying about their search results [video]
A leak from Google's GitHub shows the search algorithm tracks user clicks and time on pages, raising concerns about search result accuracy, treatment of smaller websites, and SEO strategies.
How I scraped 6 years of Reddit posts in JSON
The article covers scraping 6 years of Reddit posts for self-promotion data, highlighting challenges like post limits and cutoffs. Pushshift is suggested for Reddit archives. Extracting URLs and checking website status are explained. Findings reveal 40% of sites inactive. Trends in online startups are discussed.