What I would really like is a tool that takes a video or series of photos, and automatic catalogues the contents. This would be nice from a ‘document all the things’ perspective, but really, really convenient in case of a fire or theft claim.
If you haven’t already, I strongly urge everyone reading this comment to stand up and do a video walkthrough of your house, today. Do all of your book spines, jewellery, DVDs, games, clothes, tools, cutlery, etc. Take photos of your bike’s serial numbers (usually under the hub).
Store the videos and photos in a cloud album or even a free tier somewhere. Email it to yourself, whatever, just don’t forget to do it.
It might take half an hour or, but this evidence is priceless (in terms of time, but it does actually have an monetary value) if you ever need to claim insurance in case of a fire or buglary.
It simply isn’t possible to remember everything you own.
Some insurers demand photographic evidence of recent ownership - I found this out the hard way (who here has a photo of themselves with their bike?! I even had the receipt!)
Same. I've imagined 'spines.com' for books, where the work of linking book spines to ISBN's, etc, has been crowdsourced and you can point your phone camera at a shelf and look up reviews, etc.
When I visit spines.com it does a redirect to a self-publishing tool called booxAI.com . I don't think that's what you were referring to, or did the spines tool sell their domain name?
Apple iOS does this, kindasorta, but it's not real time. It does text recognition on text on images in your photostream, and you can search for text in your photos using the search bar.
Apple lidar is absolutely suited to this, and I’ve always thought there is an opportunity to integrate it into the dollhouse addon in Homeassistant.
Or, create a digital twin of your garden, and simulate light shadows throughout the day after adding or removing a tree. Add pruning schedules to a fruit tree.
It’s trivial to do a scan, but it hasn’t really taken off in a practical sense.
I wish when people published things like this they would take the time to notice that their requirements.txt and READMEs don't actually work; e.g., try it out in a fresh VM or container divorced from their working environment. Not even their arguments match what's in the README.
It blows my mind how much effort people, particularly python ml people, put into creating, training, blogging, promoting their github repo and then give so little thought to whether anybody will be able to use it. You're lucky if you get an incomplete and unversioned requirements.txt and no mention of what python version was used. If it requires conda, just give up. You might get an injunction to "install pytorch" and you're on your own.
While watching an interview recently, where the interviewee was sitting in front of their bookshelf, I was trying to discern the book titles to add to my reading list. I screenshotted the person/bookshelf and tried asking ChatGPT+/GPT4 to list all the books. It could only identity a tiny fraction.
This is a nice time saving tool for "minimalist" information hoarding.
A pet project of mine is to thin out my bookshelf only to books I regularly reach for and store "never going to read" books out of sight.
The idea is, if I can save the details of the "never going to read books" and acquire a digital copy of them, it may be easier for me to psychologically let go of the physical copy and gain the storage space again.
I was going to take a photo of my crowded bookshelves and manually put the ISBN and titles into a spreadsheet. Keeping the photos simply for extra reference. Your project making the photo clickable is a great bridge between the data and artifact.
This is interesting. For me it seems like me not reaching out for books is a consequence of lazy "rotation", i.e., these are the same books I'd eventually find time to read if I could see them around often enough, so when I have some leisure I don't forget about them.
Anyway this is our (my wife and mine) hypothesis - so we are currently working on rotating books, let's see how that works out : - )
HTML has image maps. It's been there since longer than many of the people who might read this post have even been alive. You don't need SVG and JS for this.
Was expecting to find a project where you add switch behind a book and "clicking" it opens up secret entrance behind the bookshelf, but this is also very cool.
Would it make sense to have the user click and then use that point as a SAM prompt? It might let you find a book even if the initial SAM query doesn't find it.
Post author here. I like this idea. I plan to explore it and make a more generic solution. I'd love to have a point-and-click interface for annotating scenes.
For example, I'd like to be able to click on pieces of coffee equipment in a photo of my coffee setup so I can add sticky note annotations when you hover over each item.
For the bookshelves idea specifically, I would love to have a correction system in place. The problem isn't so much SAM as it is Grounding DINO, the model I'm using for object identification. I then pass each identified region to SAM and map the segmentation mask to the box.
Grounding DINO detects a lot of book spines, but often misses 1-2. I am planning to try out YOLO-World (https://github.com/AILab-CVC/YOLO-World), which, in my limited testing, performs better for this task.
If you haven’t already, I strongly urge everyone reading this comment to stand up and do a video walkthrough of your house, today. Do all of your book spines, jewellery, DVDs, games, clothes, tools, cutlery, etc. Take photos of your bike’s serial numbers (usually under the hub).
Store the videos and photos in a cloud album or even a free tier somewhere. Email it to yourself, whatever, just don’t forget to do it.
It might take half an hour or, but this evidence is priceless (in terms of time, but it does actually have an monetary value) if you ever need to claim insurance in case of a fire or buglary.
It simply isn’t possible to remember everything you own.
Some insurers demand photographic evidence of recent ownership - I found this out the hard way (who here has a photo of themselves with their bike?! I even had the receipt!)