First setup an unprivileged Ubuntu container with Plex Media Server installed. Inside the container take note of the id of the plex group.
# Your Plex group's ID may be different
$ getent group plex | cut -d : -f3
998UPDATED 22.11.2022
It's been two years since the last update, so here's the updated working script as per the comments below.
Thanks to BryanHaley for this.
setInterval(function () {
video = document.getElementsByTagName('ytd-playlist-video-renderer')[0];
video.querySelector('#primary button[aria-label="Action menu"]').click();| 'use strict'; | |
| exports.handler = (event, context, callback) => { | |
| // Get request and request headers | |
| const request = event.Records[0].cf.request; | |
| const headers = request.headers; | |
| // Configure authentication | |
| const authUser = 'user'; | |
| const authPass = 'pass'; |
| // find old workspace folders on any node (master/slave) | |
| // currently only look for deleted branches of multibranch (pipeline?) jobs, and remove them | |
| // also finds builds older than 90 days. Set flags below to true to actually do deletes | |
| import hudson.model.*; | |
| import hudson.util.*; | |
| import jenkins.model.*; | |
| import hudson.FilePath; | |
| import hudson.FilePath.FileCallable; | |
| import hudson.slaves.OfflineCause; |
There are two modes when you don't want Certbot to edit your configuration:
Webroot is better because it doesn't need to replace Nginx (to bind to port 80) to renew certificates.
In the following, we're setting up mydomain.com to be served from /var/www/mydomain, and challenges will be served from /var/www/letsencrypt.
Product: Sagitta Brutalis 1080 (PN S3480-GTX-1080-2697-128)
Software: Hashcat v3.00-beta-145-g069634a, Nvidia driver 367.18
Accelerator: 8x Nvidia GTX 1080 Founders Edition
| #!/bin/bash | |
| touch /tmp/Recent.xyz | |
| while true | |
| do | |
| echo "Checking again..." | |
| wget -q https://apt.dockerproject.org/repo/dists/ubuntu-xenial/main/binary-amd64/Release -O /tmp/Release.xyz | |
| if test /tmp/Release.xyz -nt Recent.xyz; then | |
| mplayer /usr/share/sounds/ubuntu/stereo/phone-incoming-call.ogg |
| # copied from http://gbayer.com/development/moving-files-from-one-git-repository-to-another-preserving-history/ | |
| git clone <git repository A url> # clone source repository | |
| cd <git repository A directory> | |
| git remote rm origin # to make sure it doesn't affect the original repository | |
| git filter-branch --subdirectory-filter <directory 1> -- --all # remove all files other than the ones needed | |
| mkdir <directory 1> # move them into another directory where they will be stored in the destination repository (if needed) | |
| mv * <directory 1> | |
| git add . | |
| git commit |
Scalability - You can load balance multiple instances of your application behind front end server. This will allow you to handle more volume, and increase stability in the event one of your instances goes down.
Security - Apache, Tomcat, and Glassfish all support SSL, but if you decide to use Apache, most likely thats where you should configure it. If you want additional protection against attacks (DoS, XSS, SQL injection, etc.) you can install the mod_security web application firewall.
Additional Features - Apache has a bunch of nice modules available for URL rewriting, interfacing with other programming languages, authentication, and a ton of other stuff.
Clustering - By using Apache HTTP as a front end you can let Apache HTTP act as a front door to your content to multiple Apache Tomcat instances. If one of your Apache Tomcats fails, Apache HTTP ignores it and your Sysadmin can sleep through the nigh