huffduff-video has been decommissioned. So long, and thanks for all the fish!
For alternatives, check out Breadcrum, PodQueue, Podify, and youtube-dl-api-server.
Extracts the audio from videos on YouTube, Vimeo, and many more sites and sends it to Huffduffer.
See huffduff-video.snarfed.org for bookmarklet and usage details.
Uses yt-dlp to download the video and extract its audio track. Stores the resulting MP3 file in Backblaze B2.
License: this project is placed in the public domain. Alternatively, you may use it under the CC0 license.
- Podify is a self-hosted app that also serves feeds of the generated MP3 files. Backed by youtube-dl.
- youtube-dl-api-server is a web front-end that uses youtube-dl to extract and return a video's metadata.
- Flask webapp and Chrome extension for using youtube-dl to download a video to local disk.
- iOS workflow that does the same thing as huffduff-video, except all client side: downloads a YouTube video, converts it to MP3, uploads the MP3 to Dropbox, and passes it to Huffduffer.
huffduff-video has a few specific requirements that make it a bit harder than usual to find a host, so right now it's on a full VM, on AWS EC2. I'd love to switch to a serverless/containerized host instead, but I haven't found one that satisfies all of the requirements yet:
- Python 3 WSGI application server
- able to install and use ffmpeg, generally as a system package
- long-running HTTP requests, often over 60s
- streaming HTTP responses aka "hanging GETs"
-
= 1G memory
-
= 2G disk (largest output file in Dec 2019 was 1.7G)
- Lots of egress bandwidth, often >200G/mo
Many of the major serverless PaaS hosts didn't/don't support all of these, especially streaming HTTP responses, since they often have a frontend in front of the application server that buffers entire HTTP responses before returning them.
- AWS Lambda: only 512MB disk but can use EFS for more;
only Javasupports streaming, other languages unclear; expensive-ish egress bandwidth ($.09/G) - Google Cloud Run:
no streamingnow supports streaming! In memory file system though, which gets expensive; expensive-ish egress bandwidth ($.085/G) - App Engine Standard: no streaming or system packages
- App Engine Flexible: pricing is a bit prohibitive, ~$40/mo minimum
- Azure Functions and App Service: seems like no streaming or system packages, but hard to tell for sure
Most other smaller serverless hosts (eg Heroku, Zeit, Serverless) don't allow installing system packages like ffmpeg or support streaming HTTP responses either.
I track monthly costs here. They come from this B2 billing page, and before that, this AWS billing page. The B2 bucket web UI shows the current total number of files and total bytes stored in the huffduff-video
bucket.
I've configured the bucket's lifecycle to hide files after 31 days, and delete them 1 day after that. I also configured the bucket settings to send the Cache-Control: max-age=210240
HTTP header to let clients cache files for up to a year.
I originally used AWS S3 instead of B2, but S3 eventually got too expensive. As of 11/21/2019, huffduff-video was storing ~200GB steady state, and downloads were using well over 2T/month of bandwidth, so my S3 bill alone was >$200/month.
Currently on an AWS EC2 t2.micro instance on Ubuntu 20. unattended-upgrades
is on, with the default configuration; logs are in /var/log/unattended-upgrades/
.
I started it originally on a t2.micro. I migrated it to a t2.nano on 2016-03-24, but usage outgrew the nano's CPU quota, so I migrated back to a t2.micro on 2016-05-25.
I did both migrations by making an snapshot of the t2.micro's EBS volume, making an AMI from the snapshot, then launching a new t2.nano instance using that AMI. Details.
Here's how I set it up:
# set up swap
sudo dd if=/dev/zero of=/var/swapfile bs=1M count=4096
sudo chmod 600 /var/swapfile
sudo mkswap /var/swapfile
sudo swapon /var/swapfile
# add my dotfiles
mkdir src
cd src
git clone [email protected]:snarfed/dotfiles.git
cd
ln -s src/dotfiles/.cshrc
ln -s src/dotfiles/.gitconfig
ln -s src/dotfiles/.git_excludes
ln -s src/dotfiles/.python
# install core system packages and config
sudo apt-get update
sudo apt-get install apache2 libapache2-mod-wsgi-py3 tcsh python3 python3-pip ffmpeg
sudo pip3 install -U pip
sudo chsh ubuntu
# enter /bin/tcsh
# install and set up huffduff-video
cd ~/src
git clone https://github.com/snarfed/huffduff-video.git
cd huffduff-video
sudo pip3 install -r requirements.txt
# add these lines to /etc/httpd/conf/httpd.conf
#
# # rest is for huffduff-video!
# Options FollowSymLinks
# WSGIScriptAlias /get /var/www/cgi-bin/app.py
# LogLevel info
#
# # tune number of prefork server processes
# StartServers 8
# ServerLimit 12
# MaxClients 12
# MaxRequestsPerChild 4000
# start apache
sudo service apache2 start
systemctl status apache2.service
sudo systemctl enable apache2.service
sudo chmod a+rx /var/log/apache2
sudo chmod -R a+r /var/log/apache2
# on local laptop
cd ~/src/huffduff-video/
scp b2_* aws_* ubuntu@[IP]:src/huffduff-video/
# back on EC2
cd /var/www/
sudo mkdir cgi-bin
cd cgi-bin
sudo ln -s ~/src/huffduff-video/app.py
cd /var/www/html
sudo ln -s ~/src/huffduff-video/static/index.html
sudo ln -s ~/src/huffduff-video/static/robots.txt
sudo ln -s ~/src/huffduff-video/static/util.js
# install cron jobs
cd
cat > ~/crontab << EOF
# clean up /tmp every hour
0 * * * * find /tmp/ -user www-data -not -newermt yesterday | xargs rm
# auto upgrade yt-dlp daily
10 10 * * * sudo pip3 install -U yt-dlp; sudo service apache2 restart
# recopy robots.txt to S3 since our bucket expiration policy deletes it monthly
1 2 3 * * aws s3 cp --acl=public-read ~/src/huffduff-video/s3_robots.txt s3://huffduff-video/robots.txt
EOF
crontab crontab
It's possible to set Apache up on macOS to run Python like the production Linux setup, eg with Homebrew Apache and UWSGI, but it's a bit complicated. The simpler approach is to make a virtualenv, install requirements.txt
and gunicorn
in it, and then run appp.py
under gunicorn
with eg :
gunicorn --workers 1 --threads 10 -b :8080 app
The app will serve on localhost:8080. Run with eg http://localhost:8080/?url=...
huffduff-video is pretty small and simple, it doesn't have many unusual dependencies or needs, so I've generally had good luck using Ubuntu's do-release-upgrade
tool to upgrade from one Ubuntu LTS version to the next (more, even more):
sudo apt-get update
sudo apt-get upgrade
sudo do-release-upgrade
Python installed packages may disappear, make sure to reinstall those with sudo! Otherwise Apache's mod_wsgi won't see them, or will see older versions.
sudo pip3 install -r requirements.txt
I followed the Certbot Apache instructions to mint an SSL certificate, install it, and set up a cron job to renew it every 3 months:
sudo snap install core; sudo snap refresh core
sudo snap install --classic certbot
sudo certbot --apache
# answer questions; domain is huffduff-video.snarfed.org
I use Honeycomb to monitor huffduff-video with black box HTTP probes to its home page. If enough of them fail in a given time window, it emails me.
I use CloudWatch to monitor and alert on EC2 instance system checks and CPU quota. When alarms fire, it emails me.
To get system-level custom metrics for memory, swap, and disk space, set up Amazon's custom monitoring scripts.
sudo yum install perl-DateTime perl-Sys-Syslog perl-LWP-Protocol-https
wget http://aws-cloudwatch.s3.amazonaws.com/downloads/CloudWatchMonitoringScripts-1.2.1.zip
unzip CloudWatchMonitoringScripts-1.2.1.zip
rm CloudWatchMonitoringScripts-1.2.1.zip
cd aws-scripts-mon
cp awscreds.template awscreds.conf
# fill in awscreds.conf
./mon-put-instance-data.pl --aws-credential-file ~/aws-scripts-mon/awscreds.conf --mem-util --swap-util --disk-space-util --disk-path=/ --verify
crontab -e
# add this line:
# * * * * * ./mon-put-instance-data.pl --aws-credential-file ~/aws-scripts-mon/awscreds.conf --mem-util --swap-util --disk-space-util --disk-path=/ --from-cron
To set up HTTP and application level monitoring, I had to:
- add an IAM policy
- install the logs agent with
sudo yum install awslogs
- add my IAM credentials to
/etc/awslogs/awscli.conf
and set region to us-west-2 - add these lines to
/etc/awslogs/awslogs.conf
:
[/var/log/httpd/access_log]
file = /var/log/httpd/access_log*
log_group_name = /var/log/httpd/access_log
log_stream_name = {instance_id}
datetime_format = %d/%b/%Y:%H:%M:%S %z
[/var/log/httpd/error_log]
file = /var/log/httpd/error_log*
log_group_name = /var/log/httpd/error_log
log_stream_name = {instance_id}
datetime_format = %b %d %H:%M:%S %Y
# WSGI writes Python exception stack traces to this log file across multiple
# lines, and I'd love to collect them multi_line_start_pattern or something
# similar, but each line is prefixed with the same timestamp + severity + etc
# prefix as other lines, so I can't.
- start the agent and restart it on boot:
sudo service awslogs start
sudo service awslogs status
sudo chkconfig awslogs on
- wait a while, then check that the logs are flowing:
aws --region us-west-2 logs describe-log-groups
aws --region us-west-2 logs describe-log-streams --log-group-name /var/log/httpd/access_log
aws --region us-west-2 logs describe-log-streams --log-group-name /var/log/httpd/error_log
- define a few metric filters so we can graph and query HTTP status codes, error messages, etc:
aws logs put-metric-filter --region us-west-2 \
--log-group-name /var/log/httpd/access_log \
--filter-name HTTPRequests \
--filter-pattern '[ip, id, user, timestamp, request, status, bytes]' \
--metric-transformations metricName=count,metricNamespace=huffduff-video,metricValue=1
aws logs put-metric-filter --region us-west-2 \
--log-group-name /var/log/httpd/error_log \
--filter-name PythonErrors \
--filter-pattern '[timestamp, error_label, prefix = "ERROR:root:ERROR:", ...]' \
--metric-transformations metricName=errors,metricNamespace=huffduff-video,metricValue=1
aws --region us-west-2 logs describe-metric-filters --log-group-name /var/log/httpd/access_log
aws --region us-west-2 logs describe-metric-filters --log-group-name /var/log/httpd/error_log
Back in April 2015, I did a bit of research to understand who was downloading huffduff-video files, to see if I could optimize its bandwidth usage by blocking non-human users.
As always, measure first, then optimize. To learn a bit more about who's downloading these files, I turned on S3 access logging, waited 24h, then ran these commands to collect and aggregate the logs:
aws --profile personal s3 sync s3://huffduff-video/logs .
grep -R REST.GET.OBJECT . | grep ' 200 ' | grep -vE 'robots.txt|logs/20' \
| sed -E 's/[A-Za-z0-9\/+=_-]{32,76}/X/g' | cut -d' ' -f8,20- | sort | uniq -c | sort -n -r > user_agents
grep -R REST.GET.OBJECT . | grep ' 200 ' | grep -vE 'robots.txt|logs/20' \
| cut -d' ' -f5 | sort | uniq -c | sort -n -r > ips
This gave me some useful baseline numbers. Over a 24h period, there were 482 downloads, 318 of which came from bots. (That's 2/3!) Out of the six top user agents by downloads, five were bots. The one exception was the Overcast podcast app.
- Flipboard Proxy (142 downloads)
- Googlebot (67)
- Twitterbot (39)
- Overcast (47)
- Yahoo! Slurp (36)
- Googlebot-Video (34)
(Side note: Googlebot-Video is polite and includes Etag
or If-Modified-Since
when it refetches files. It sent 68 requests, but exactly half of those resulted in an empty 304
response. Thanks Googlebot-Video!)
I switched huffduff-video to use S3 URLs on the huffduff-video.s3.amazonaws.com
virtual host, added a robots.txt
file that blocks all bots, waited 24h, and then measured again. The vast majority of huffduff-video links on Huffduffer are still on the s3.amazonaws.com
domain, which doesn't serve my robots.txt
, so I didn't expect a big difference...but I was wrong. Twitterbot had roughly the same number, but the rest were way down:
- Overcast (76)
- Twitterbot (36)
- FlipboardProxy (33)
- iTunes (OS X) (21)
- Yahoo! Slurp (20)
- libwww-perl (18)
- Googlebot (14)
(Googlebot-Video was way farther down the chart with just 4 downloads.)
This may have been due to the fact that my first measurement was Wed-Thurs, and the second was Fri-Sat, which are slower social media and link sharing days. Still, I'm hoping some of it was due to robots.txt
. Fingers crossed the bots will eventually go away altogether!
To update the robots.txt
file:
aws --profile personal s3 cp --acl=public-read ~/src/huffduff-video/s3_robots.txt s3://huffduff-video/robots.txt
I put this in a cron job to run every 30d. I had to run aws configure
first and give it the key id and secret.
To find a specific bot's IPs:
$ grep -R FlipboardProxy . | cut -d' ' -f5 |sort |uniq
34.207.219.235
34.229.167.12
34.229.216.231
52.201.0.135
52.207.240.171
54.152.58.154
54.210.190.43
54.210.24.16
...and then to block them, add them to the bucket policy:
{
"Version": "2012-10-17",
"Id": "Block IPs",
"Statement": [
{
"Sid": "Block FlipboardProxy (IPs collected 1/25-26/2017)",
"Effect": "Deny",
"Principal": "*",
"Action": "s3:*",
"Resource": "arn:aws:s3:::huffduff-video/*",
"Condition": {
"IpAddress": {
"aws:SourceIp": [
"34.207.219.235/32",
"34.229.167.12/32",
"34.229.216.231/32",
"52.201.0.135/32",
"52.207.240.171/32",
"54.152.58.154/32",
"54.210.190.43/32",
"54.210.24.16/32"
]
}
}
}
]
}
While doing this, I discovered something a bit interesting: Huffduffer itself seems to download a copy of every podcast that gets huffduffed, ie the full MP3 file. It does this with no user agent, from 146.185.159.94, which reverse DNS resolves to huffduffer.com.
I can't tell that any Huffduffer feature is based on the actual audio from each podcast, so I wonder why they download them. I doubt they keep them all. Jeremy probably knows why!
Something also downloads a lot from 54.154.42.3 (on Amazon EC2) with user agent Ruby
. No reverse DNS there though.