skimmer is a lightweight feed reader inspired by newsboat and yarnc from yarn.social. skimmer is very minimal and deliberately lacks features. That is to say skimmer’s best feature is what it doesn’t do. skimmer tries to do two things well.
That’s it. That is skimmer secret power. It does only two things. There is no elaborate user interface beyond standard input, standard output and standard error found on POSIX type operating systems. Even if you invoke it in “interactive” mode your choices are limited, press enter and go to next item, press “n” and mark the item read, press “s” and save the item, press “q” and quit interactive mode.
By storing the item information in an SQLite3 database (like newsboat’s cache.db file) I can re-purpose the feed content as needed. An example would be generating a personal news aggregation page. Another might be to convert the entries to BibTeX and manage them as reference. Lots of options are possible.
As mentioned skimmer was very much inspired by newsboat. In fact it uses newsboat’s urls list format. That’s because skimmer isn’t trying to replace newsboat as a reader of all feeds but instead gives me more options for how I read the feeds I’ve collected.
The newsboat urls file boils down to a list of urls, one per line with an optional “label” added after the url using the notation of space, double quote, tilde, label content followed by a double quote and end of line. That’s really easy to parse. You can add comments using the hash mark with hash mark and anything to the right ignored when the urls are read in to skimmer.
UPDATE: 2023-10-31, In using the experimental skimmer app in practive I have found some feed sources still white list access based on user agent strings (not something that is concidered a “best practice” today). Unfortunately it is highly in conconsistant to know which string is accepted. As a result maintaining a list of feeds is really challenging unless you can specific a user agent string per feed source for those that need it. As a result I’ve add an additional column of content to the newsboat url file format. A user agent can be included after a feed’s label by adding a space and the user agent string value.
skimmer uses SQLite 3 database with two tables for managing feeds and their content. It doesn’t use newsboat’s cache.db. The name of the skimmer database ends in “.skim” and pairs with the name of the urls file. Example if I have a urls list named “my_news.txt” skimmer will use a database file (and create it if it doesn’t exist) called “my_news.skim”. Each time skimmer reads the urls file it will replace the content in the skimmer database file except for any notations about a given item having been read or saved.
Presently skimmer is focused on reading RSS 2, Atom and jsonfeeds as that is provided by the Go package skimmer uses (i.e. goread). Someday, maybe, I hope to include support for Gopher or Gemini feeds.
skimmer [OPTIONS] URL_LIST_FILENAME
skimmer [OPTIONS] SKIMMER_DB_FILENAME [TIME_RANGE]
skimmer have two ways to invoke it. You can fetch the contents from list of URLs in newsboat urls file format. You can read the items from the related skimmer database.
Fetch and read my newsboat feeds from .newsboat/urls
.
This will create a .newsboat/urls.skim
if it doesn’t exist.
Remember invoking skimmer with a URLs file will retrieve feeds and their
contents and invoking skimmer with the skimmer database file will let
you read them.
skimmer .newsboat/urls
skimmer .newsboat/urls.skim
This will fetch and read the feeds frommy-news.urls
.
This will create a my-news.skim
file. When the skimmer
database is read a simplistic interactive mode is presented.
skimmer my-news.urls
skimmer -i my-news.skim
The same method is used to update your my-news.skim
file
and read it.
Export the current state of the skimmer database channels to a urls file. Feeds that failed to be retrieved will not be in the database channels table channels table. This is an easy way to get rid of the cruft and dead feeds.
skimmer -urls my-news.skim >my-news.urls
Prune the items in the database older than today.
skimmer -prune my-news.skim today
Prune the items older than September 30, 2023.
skimmer -prune my-news.skim \
"2023-09-30 23:59:59"
skimmer is an experiment. The compiled binaries are not necessarily tested. To compile from source you need to have git, make, Pandoc, SQLite3 and Go.
Installation process I used to setup skimmer on a new machine.
git clone https://github.com/rsdoiel/skimmer
cd skimmer
make
make install
This experiment would not be possible with the authors of newsboat, SQLite3, Pandoc and the gofeed package for Go.