I can’t wait for the technology to get basic enough where I can roll my own self hosted instance of it without it taking months. Because I can see a way it’s doable without a centralized service to get around that. But for mass consumer level, I can see that becoming true. But this can be applied to every bit of software currently. All of it can be ran by you, if you have time. Hell I’ve got my own cloud (hosted at my home ) music streaming service.
A lot of that is doable now - like, how many grocery stores are even nearby to someone, so writing a custom bit of code to check the website of each, one by one, and looking for previously manually-identified items could be automated.
One major downside is prioritization of large chain stores at the expense of smaller mom & pop ones that don’t maintain a constant inventory system accessible via the web. Someone could even volunteer their time to build them a database backend, but still they’d have to see the value in actually scanning the items every time or else it would quickly fall behind.
That’s precisely what I was thinking, but reflecting more on it, I don’t know how well it would handle the webpages, so maybe some other languages mixed in too (I’m out of date, maybe PHP?). If AI writing code worked it would lower the barrier, but I’m not certain we’re quite there yet to trust anything it would create.
Python web scraping is just fine, with the llms you.have the option of either extracting the html and having the LLM read.over that, or having a vision ai OCR the page and make its own decision of what to extract.
I can’t wait for the technology to get basic enough where I can roll my own self hosted instance of it without it taking months. Because I can see a way it’s doable without a centralized service to get around that. But for mass consumer level, I can see that becoming true. But this can be applied to every bit of software currently. All of it can be ran by you, if you have time. Hell I’ve got my own cloud (hosted at my home ) music streaming service.
A lot of that is doable now - like, how many grocery stores are even nearby to someone, so writing a custom bit of code to check the website of each, one by one, and looking for previously manually-identified items could be automated.
One major downside is prioritization of large chain stores at the expense of smaller mom & pop ones that don’t maintain a constant inventory system accessible via the web. Someone could even volunteer their time to build them a database backend, but still they’d have to see the value in actually scanning the items every time or else it would quickly fall behind.
Yeah, it wouldn’t be a huge lift if you’re familiar with python.
Done!
That’s precisely what I was thinking, but reflecting more on it, I don’t know how well it would handle the webpages, so maybe some other languages mixed in too (I’m out of date, maybe PHP?). If AI writing code worked it would lower the barrier, but I’m not certain we’re quite there yet to trust anything it would create.
Python web scraping is just fine, with the llms you.have the option of either extracting the html and having the LLM read.over that, or having a vision ai OCR the page and make its own decision of what to extract.
Yeah I was going to say, I’ve done similar for clients with regards to competitor pricing