Dobrica Pavlinušić's random unstructured stuff
Grep: Revision 7

Grep is RSS feed searcher and cacher. Actually, more correct is to think about Grep as your search console.
If you are into buzzword land, you can even call that information worker workbench but, enough of that.

It's mostly useful with web sites that provide search results as RSS feed. If they include full content in feeds even better. One example of such web service is SocialText wiki for which Grep was originally written.

I know that I wrote it in some wiki workspace, but where ?!

Following that idea, Grep gained powerful plugin mechanism which enable users (err, developers who can write 10-line perl module) to scrape any site which has form or rest API and produce search results as links. While doing that, it will also fetch result pages and cache them locally. Have in mind that this is a slow process which puts much load on remote server, so use it sparsely. However, once remote results are fetched, they are always available in local cache for quick reference, even when offline!

For now, here is a feature list:

  • cache all results locally, great for off-line use
  • supports credential spoofing using Cookies (useful to login into protected areas)
  • comes with easy bookmarklet subscribe helper
  • written using Jifty and KinoSearch
  • scrapers for other wikis and CMS engines which can be source of items for Grep
  • de-duplicate local results (based on near duplicates detection which is great for sites which change just slightly like wikis)

Source code is in development Subversion repository

fetchrss: http://svn.rot13.org/index.cgi/Grep/rss
  • There was an error: 404 Not Found