~/Blog

Brandon Rozek

Photo of Brandon Rozek

PhD Student @ RPI studying Automated Reasoning in AI and Linux Enthusiast.

Blogroll From Subscriptions

Published on

Updated on

While I was browsing around personal websites, I found a fun little piece of code from Jake Bauer’s links page.

grep "xmlUrl" static/subscriptions.opml |\
sed 's/.*text=\"\(.*\)\" xmlUrl=\"\(https\?:\/\/[^\/]*\/\)\(.*\)\" .*/<li><a href=\"\2\">\1<\/a> (<a href=\"\2\3\">feed<\/a>)<\/li>/g'

This takes the subscriptions exported from yarr and generates a HTML list which you can then include in a blogroll page.

Running this script on my export from Feedbin yielded some extra metadata being shown in the HTML. For example: Joke Bauer type="rss" (Feed). So let’s edit the code snippet above so that it works for my subscription export.

From my subscriptions.xml here’s an example entry:

<outline text="Jake Bauer" title="Jake Bauer" type="rss" xmlUrl="https://www.paritybit.ca/feed.xml" htmlUrl="https://www.paritybit.ca/"/>

It looks like I need to extract the title, xmlUrl, and htmlUrl attributes in that specific order. I’ll use the same technique from a previous post on capturing quoated strings.

grep "xmlUrl" subscriptions.xml |\
sed 's/.*title=\"\([^\"]*\)\".*xmlUrl=\"\([^\"]*\)\".*htmlUrl=\"\([^\"]*\)\".*/<li><a href=\"\3\">\1<\/a> (<a href=\"\2\">feed<\/a>)<\/li>/g'

We can then clean this up into a create_blogroll script saved within ~/.local/bin.

#!/bin/sh

set -o errexit
set -o nounset

show_usage() {
    echo "Usage: create_blogroll [subscriptions.xml]"
    exit 1
}

# Check argument count
if [ "$#" -ne 1 ]; then
    show_usage
fi

QUOTED_STR="\"\([^\"]*\)\""
XML_EXPR=".*title=$QUOTED_STR.*xmlUrl=$QUOTED_STR.*htmlUrl=$QUOTED_STR.*"
HTML_EXPR="<li><a href=\"\3\">\1<\/a> (<a href=\"\2\">feed<\/a>)<\/li>"
REPLACE_EXPR="s/$XML_EXPR/$HTML_EXPR/g"

grep "xmlUrl" "$1" | sed "$REPLACE_EXPR"
Reply via Email Buy me a Coffee
Was this useful? Feel free to share: Hacker News Reddit Twitter