Compare commits
166 commits
1.40.3.855
...
master
| Author | SHA1 | Date | |
|---|---|---|---|
| 492da39dea | |||
|
|
252e2b2b5c | ||
|
|
2d16c92cb3 | ||
|
|
3faea15cf3 | ||
|
|
0a43190ac9 | ||
|
|
68fa685a3c | ||
|
|
7378fb237d | ||
|
|
1a7cf74b6e | ||
|
|
e0048bb650 | ||
|
|
be4e12409d | ||
|
|
302c6e5702 | ||
|
|
ecb66b1ca3 | ||
|
|
e3bbf927d8 | ||
|
|
ee772de585 | ||
|
|
468a4995e0 | ||
|
|
a56f4d3e85 | ||
|
|
094d2071e2 | ||
|
|
2186587d37 | ||
|
|
c1ab38b66e | ||
|
|
8551fb88e9 | ||
|
|
e64b2579d7 | ||
|
|
cb3568d81a | ||
|
|
4a4dc88e6c | ||
|
|
9f0882a70e | ||
|
|
82cbb863bc | ||
|
|
6374e18803 | ||
|
|
944fc560c6 | ||
|
|
c9ffa88637 | ||
|
|
d4262cc4e9 | ||
|
|
2d9cdff163 | ||
|
|
163c200dcb | ||
|
|
77cca7947b | ||
|
|
d63927007f | ||
|
|
8056a354f5 | ||
|
|
f32896aed4 | ||
|
|
144b87d05b | ||
|
|
64aed46150 | ||
|
|
a1bac2c988 | ||
|
|
215506d3cd | ||
|
|
7adf1b75ca | ||
|
|
a5ad980f28 | ||
|
|
2e4d6ad903 | ||
|
|
806b88c37a | ||
|
|
dbc5f5c85b | ||
|
|
c2b3335d4f | ||
|
|
c0b54e8512 | ||
|
|
b36fa22efe | ||
|
|
62d7a9ecfe | ||
|
|
71640b3d41 | ||
|
|
65664b148a | ||
|
|
3644491d23 | ||
|
|
86f90113e5 | ||
|
|
0f27fcacc2 | ||
|
|
5d98680464 | ||
|
|
b5d4cc7e71 | ||
|
|
727499b357 | ||
|
|
f0f381c0db | ||
|
|
54bac1c3cf | ||
|
|
c87f1010a7 | ||
|
|
3975f3f4ba | ||
|
|
635f9ddf03 | ||
|
|
6987e66f3c | ||
|
|
0900661b6f | ||
|
|
c14dd5a47c | ||
|
|
b65ef27e65 | ||
|
|
70c49ef2d4 | ||
|
|
c27d92587e | ||
|
|
0465b8095e | ||
|
|
3044e621bf | ||
|
|
303d122c01 | ||
|
|
3833657a3a | ||
|
|
c35f69d4fa | ||
|
|
bf5b90a8e2 | ||
|
|
20c539447b | ||
|
|
0c7513ed90 | ||
|
|
05ec76e7e9 | ||
|
|
5d77e8d8ba | ||
|
|
12308d98c6 | ||
|
|
3ee08e388d | ||
|
|
803ec8f8b6 | ||
|
|
97c9c78c0c | ||
|
|
60c1ec04cf | ||
|
|
5be1cb4c1c | ||
|
|
d5127aff6d | ||
|
|
a2e1bebb34 | ||
|
|
989f857351 | ||
|
|
3fbfd24ffa | ||
|
|
dbabe2bee4 | ||
|
|
0a30a9391e | ||
|
|
2d5bbae0dd | ||
|
|
8a22d79bee | ||
|
|
fa3f35a0f0 | ||
|
|
1faeb23abb | ||
|
|
23a7477bd6 | ||
|
|
b9716badd9 | ||
|
|
7577e696ce | ||
|
|
8217b62d9a | ||
|
|
54de1404ca | ||
|
|
aa406f1496 | ||
|
|
56d67a38c4 | ||
|
|
c8c020c491 | ||
|
|
875eb3dc41 | ||
|
|
bd368dfbf3 | ||
|
|
b29a1b8320 | ||
|
|
5521df85df | ||
|
|
af6901e341 | ||
|
|
025e5da7cb | ||
|
|
25ff849152 | ||
|
|
ed26e879a2 | ||
|
|
57555832cd | ||
|
|
1705b4cee7 | ||
|
|
3297c9b082 | ||
|
|
74f663120a | ||
|
|
49fe7202e9 | ||
|
|
7c019750ce | ||
|
|
081ace2aee | ||
|
|
0102ec4fa6 | ||
|
|
86c5c83041 | ||
|
|
a54fdfbce9 | ||
|
|
88b49cd0be | ||
|
|
4dd0b7f2d2 | ||
|
|
8e9b62b828 | ||
|
|
bb423a4d3b | ||
|
|
877a9bd1a3 | ||
|
|
3b7c8dd08b | ||
|
|
9e313afccf | ||
|
|
1a6eae12a2 | ||
|
|
4d1b0e164e | ||
|
|
b3fae00aee | ||
|
|
7e207423f3 | ||
|
|
fedcb7ae7a | ||
|
|
a420bdda82 | ||
|
|
42af1591d5 | ||
|
|
15a4ebbcfb | ||
|
|
8d628cc4ff | ||
|
|
5be889429b | ||
|
|
885401d6e4 | ||
|
|
603e86d095 | ||
|
|
5c3b47763c | ||
|
|
4e96653f26 | ||
|
|
521b7d06d8 | ||
|
|
d5716af54b | ||
|
|
a42349565e | ||
|
|
cc4cae358a | ||
|
|
c4de83bb82 | ||
|
|
2ca56eb1ce | ||
|
|
b646ebaabe | ||
|
|
409420813e | ||
|
|
7d865543ea | ||
|
|
c7717810ae | ||
|
|
9fdb4a4894 | ||
|
|
29400873e3 | ||
|
|
aa902dd60b | ||
|
|
a2cb9a9595 | ||
|
|
64197c0cf8 | ||
|
|
259c06aaa9 | ||
|
|
c6e9645aab | ||
|
|
a3a21e9824 | ||
|
|
80fc8c4237 | ||
|
|
34130ffa25 | ||
|
|
c25a8e637e | ||
|
|
658c63714c | ||
|
|
d177f3a00e | ||
|
|
603a79335f | ||
|
|
fa08911915 | ||
|
|
3cced1d334 |
33 changed files with 1073 additions and 608 deletions
49
.forgejo/workflows/build.yml
Normal file
49
.forgejo/workflows/build.yml
Normal file
|
|
@ -0,0 +1,49 @@
|
||||||
|
name: Build and push image
|
||||||
|
|
||||||
|
on:
|
||||||
|
push:
|
||||||
|
branches: [master]
|
||||||
|
schedule:
|
||||||
|
- cron: '0 4 * * *' # daily, picks up LSIO base image updates
|
||||||
|
workflow_dispatch:
|
||||||
|
|
||||||
|
jobs:
|
||||||
|
build:
|
||||||
|
runs-on: docker
|
||||||
|
steps:
|
||||||
|
- name: Checkout
|
||||||
|
uses: actions/checkout@v4
|
||||||
|
with:
|
||||||
|
ref: master
|
||||||
|
fetch-depth: 0
|
||||||
|
token: ${{ secrets.PUSH_TOKEN }}
|
||||||
|
|
||||||
|
- name: Merge upstream
|
||||||
|
run: |
|
||||||
|
git config user.name "forgejo-runner"
|
||||||
|
git config user.email "runner@forge.jannis.wtf"
|
||||||
|
git fetch origin master
|
||||||
|
git reset --hard origin/master
|
||||||
|
git remote add upstream https://github.com/linuxserver/docker-plex
|
||||||
|
git fetch upstream
|
||||||
|
git merge upstream/master --no-edit
|
||||||
|
git push origin master
|
||||||
|
|
||||||
|
- name: Set up Buildx
|
||||||
|
uses: docker/setup-buildx-action@v3
|
||||||
|
|
||||||
|
- name: Login to Forgejo registry
|
||||||
|
uses: docker/login-action@v3
|
||||||
|
with:
|
||||||
|
registry: forge.jannis.wtf
|
||||||
|
username: ${{ secrets.REGISTRY_USER }}
|
||||||
|
password: ${{ secrets.REGISTRY_TOKEN }}
|
||||||
|
|
||||||
|
- name: Build and push
|
||||||
|
uses: docker/build-push-action@v5
|
||||||
|
with:
|
||||||
|
context: .
|
||||||
|
push: true
|
||||||
|
tags: |
|
||||||
|
forge.jannis.wtf/fej/plex-wireguard:latest
|
||||||
|
forge.jannis.wtf/fej/plex-wireguard:${{ github.sha }}
|
||||||
6
.github/CONTRIBUTING.md
vendored
6
.github/CONTRIBUTING.md
vendored
|
|
@ -6,7 +6,7 @@
|
||||||
* Read, and fill the Pull Request template
|
* Read, and fill the Pull Request template
|
||||||
* If this is a fix for a typo (in code, documentation, or the README) please file an issue and let us sort it out. We do not need a PR
|
* If this is a fix for a typo (in code, documentation, or the README) please file an issue and let us sort it out. We do not need a PR
|
||||||
* If the PR is addressing an existing issue include, closes #\<issue number>, in the body of the PR commit message
|
* If the PR is addressing an existing issue include, closes #\<issue number>, in the body of the PR commit message
|
||||||
* If you want to discuss changes, you can also bring it up in [#dev-talk](https://discordapp.com/channels/354974912613449730/757585807061155840) in our [Discord server](https://discord.gg/YWrKVTn)
|
* If you want to discuss changes, you can also bring it up in [#dev-talk](https://discordapp.com/channels/354974912613449730/757585807061155840) in our [Discord server](https://linuxserver.io/discord)
|
||||||
|
|
||||||
## Common files
|
## Common files
|
||||||
|
|
||||||
|
|
@ -105,10 +105,10 @@ docker build \
|
||||||
-t linuxserver/plex:latest .
|
-t linuxserver/plex:latest .
|
||||||
```
|
```
|
||||||
|
|
||||||
The ARM variants can be built on x86_64 hardware using `multiarch/qemu-user-static`
|
The ARM variants can be built on x86_64 hardware and vice versa using `lscr.io/linuxserver/qemu-static`
|
||||||
|
|
||||||
```bash
|
```bash
|
||||||
docker run --rm --privileged multiarch/qemu-user-static:register --reset
|
docker run --rm --privileged lscr.io/linuxserver/qemu-static --reset
|
||||||
```
|
```
|
||||||
|
|
||||||
Once registered you can define the dockerfile to use with `-f Dockerfile.aarch64`.
|
Once registered you can define the dockerfile to use with `-f Dockerfile.aarch64`.
|
||||||
|
|
|
||||||
2
.github/ISSUE_TEMPLATE/config.yml
vendored
2
.github/ISSUE_TEMPLATE/config.yml
vendored
|
|
@ -1,7 +1,7 @@
|
||||||
blank_issues_enabled: false
|
blank_issues_enabled: false
|
||||||
contact_links:
|
contact_links:
|
||||||
- name: Discord chat support
|
- name: Discord chat support
|
||||||
url: https://discord.gg/YWrKVTn
|
url: https://linuxserver.io/discord
|
||||||
about: Realtime support / chat with the community and the team.
|
about: Realtime support / chat with the community and the team.
|
||||||
|
|
||||||
- name: Discourse discussion forum
|
- name: Discourse discussion forum
|
||||||
|
|
|
||||||
3
.github/workflows/call_issue_pr_tracker.yml
vendored
3
.github/workflows/call_issue_pr_tracker.yml
vendored
|
|
@ -8,6 +8,9 @@ on:
|
||||||
pull_request_review:
|
pull_request_review:
|
||||||
types: [submitted,edited,dismissed]
|
types: [submitted,edited,dismissed]
|
||||||
|
|
||||||
|
permissions:
|
||||||
|
contents: read
|
||||||
|
|
||||||
jobs:
|
jobs:
|
||||||
manage-project:
|
manage-project:
|
||||||
permissions:
|
permissions:
|
||||||
|
|
|
||||||
3
.github/workflows/call_issues_cron.yml
vendored
3
.github/workflows/call_issues_cron.yml
vendored
|
|
@ -4,6 +4,9 @@ on:
|
||||||
- cron: '33 12 * * *'
|
- cron: '33 12 * * *'
|
||||||
workflow_dispatch:
|
workflow_dispatch:
|
||||||
|
|
||||||
|
permissions:
|
||||||
|
contents: read
|
||||||
|
|
||||||
jobs:
|
jobs:
|
||||||
stale:
|
stale:
|
||||||
permissions:
|
permissions:
|
||||||
|
|
|
||||||
143
.github/workflows/external_trigger.yml
vendored
143
.github/workflows/external_trigger.yml
vendored
|
|
@ -3,6 +3,9 @@ name: External Trigger Main
|
||||||
on:
|
on:
|
||||||
workflow_dispatch:
|
workflow_dispatch:
|
||||||
|
|
||||||
|
permissions:
|
||||||
|
contents: read
|
||||||
|
|
||||||
jobs:
|
jobs:
|
||||||
external-trigger-master:
|
external-trigger-master:
|
||||||
runs-on: ubuntu-latest
|
runs-on: ubuntu-latest
|
||||||
|
|
@ -11,18 +14,31 @@ jobs:
|
||||||
|
|
||||||
- name: External Trigger
|
- name: External Trigger
|
||||||
if: github.ref == 'refs/heads/master'
|
if: github.ref == 'refs/heads/master'
|
||||||
|
env:
|
||||||
|
SKIP_EXTERNAL_TRIGGER: ${{ vars.SKIP_EXTERNAL_TRIGGER }}
|
||||||
run: |
|
run: |
|
||||||
if [ -n "${{ secrets.PAUSE_EXTERNAL_TRIGGER_PLEX_MASTER }}" ]; then
|
printf "# External trigger for docker-plex\n\n" >> $GITHUB_STEP_SUMMARY
|
||||||
echo "**** Github secret PAUSE_EXTERNAL_TRIGGER_PLEX_MASTER is set; skipping trigger. ****"
|
if grep -q "^plex_master_" <<< "${SKIP_EXTERNAL_TRIGGER}"; then
|
||||||
echo "Github secret \`PAUSE_EXTERNAL_TRIGGER_PLEX_MASTER\` is set; skipping trigger." >> $GITHUB_STEP_SUMMARY
|
echo "> [!NOTE]" >> $GITHUB_STEP_SUMMARY
|
||||||
|
echo "> Github organizational variable \`SKIP_EXTERNAL_TRIGGER\` contains \`plex_master_\`; will skip trigger if version matches." >> $GITHUB_STEP_SUMMARY
|
||||||
|
elif grep -q "^plex_master" <<< "${SKIP_EXTERNAL_TRIGGER}"; then
|
||||||
|
echo "> [!WARNING]" >> $GITHUB_STEP_SUMMARY
|
||||||
|
echo "> Github organizational variable \`SKIP_EXTERNAL_TRIGGER\` contains \`plex_master\`; skipping trigger." >> $GITHUB_STEP_SUMMARY
|
||||||
exit 0
|
exit 0
|
||||||
fi
|
fi
|
||||||
echo "**** External trigger running off of master branch. To disable this trigger, set a Github secret named \"PAUSE_EXTERNAL_TRIGGER_PLEX_MASTER\". ****"
|
echo "> [!NOTE]" >> $GITHUB_STEP_SUMMARY
|
||||||
echo "External trigger running off of master branch. To disable this trigger, set a Github secret named \`PAUSE_EXTERNAL_TRIGGER_PLEX_MASTER\`" >> $GITHUB_STEP_SUMMARY
|
echo "> External trigger running off of master branch. To disable this trigger, add \`plex_master\` into the Github organizational variable \`SKIP_EXTERNAL_TRIGGER\`." >> $GITHUB_STEP_SUMMARY
|
||||||
echo "**** Retrieving external version ****"
|
printf "\n## Retrieving external version\n\n" >> $GITHUB_STEP_SUMMARY
|
||||||
EXT_RELEASE=$(curl -s "https://plex.tv/api/downloads/5.json" | jq -r '. | .computer.Linux.version')
|
EXT_RELEASE=$(curl -s "https://plex.tv/api/downloads/5.json" | jq -r '. | .computer.Linux.version')
|
||||||
|
echo "Type is \`custom_json\`" >> $GITHUB_STEP_SUMMARY
|
||||||
|
if grep -q "^plex_master_${EXT_RELEASE}" <<< "${SKIP_EXTERNAL_TRIGGER}"; then
|
||||||
|
echo "> [!WARNING]" >> $GITHUB_STEP_SUMMARY
|
||||||
|
echo "> Github organizational variable \`SKIP_EXTERNAL_TRIGGER\` matches current external release; skipping trigger." >> $GITHUB_STEP_SUMMARY
|
||||||
|
exit 0
|
||||||
|
fi
|
||||||
if [ -z "${EXT_RELEASE}" ] || [ "${EXT_RELEASE}" == "null" ]; then
|
if [ -z "${EXT_RELEASE}" ] || [ "${EXT_RELEASE}" == "null" ]; then
|
||||||
echo "**** Can't retrieve external version, exiting ****"
|
echo "> [!WARNING]" >> $GITHUB_STEP_SUMMARY
|
||||||
|
echo "> Can't retrieve external version, exiting" >> $GITHUB_STEP_SUMMARY
|
||||||
FAILURE_REASON="Can't retrieve external version for plex branch master"
|
FAILURE_REASON="Can't retrieve external version for plex branch master"
|
||||||
GHA_TRIGGER_URL="https://github.com/linuxserver/docker-plex/actions/runs/${{ github.run_id }}"
|
GHA_TRIGGER_URL="https://github.com/linuxserver/docker-plex/actions/runs/${{ github.run_id }}"
|
||||||
curl -X POST -H "Content-Type: application/json" --data '{"avatar_url": "https://cdn.discordapp.com/avatars/354986384542662657/df91181b3f1cf0ef1592fbe18e0962d7.png","embeds": [{"color": 16711680,
|
curl -X POST -H "Content-Type: application/json" --data '{"avatar_url": "https://cdn.discordapp.com/avatars/354986384542662657/df91181b3f1cf0ef1592fbe18e0962d7.png","embeds": [{"color": 16711680,
|
||||||
|
|
@ -30,25 +46,43 @@ jobs:
|
||||||
"username": "Github Actions"}' ${{ secrets.DISCORD_WEBHOOK }}
|
"username": "Github Actions"}' ${{ secrets.DISCORD_WEBHOOK }}
|
||||||
exit 1
|
exit 1
|
||||||
fi
|
fi
|
||||||
EXT_RELEASE=$(echo ${EXT_RELEASE} | sed 's/[~,%@+;:/]//g')
|
EXT_RELEASE_SANITIZED=$(echo ${EXT_RELEASE} | sed 's/[~,%@+;:/]//g')
|
||||||
echo "**** External version: ${EXT_RELEASE} ****"
|
echo "Sanitized external version: \`${EXT_RELEASE_SANITIZED}\`" >> $GITHUB_STEP_SUMMARY
|
||||||
echo "External version: ${EXT_RELEASE}" >> $GITHUB_STEP_SUMMARY
|
echo "Retrieving last pushed version" >> $GITHUB_STEP_SUMMARY
|
||||||
echo "**** Retrieving last pushed version ****"
|
|
||||||
image="linuxserver/plex"
|
image="linuxserver/plex"
|
||||||
tag="latest"
|
tag="latest"
|
||||||
token=$(curl -sX GET \
|
token=$(curl -sX GET \
|
||||||
"https://ghcr.io/token?scope=repository%3Alinuxserver%2Fplex%3Apull" \
|
"https://ghcr.io/token?scope=repository%3Alinuxserver%2Fplex%3Apull" \
|
||||||
| jq -r '.token')
|
| jq -r '.token')
|
||||||
multidigest=$(curl -s \
|
multidigest=$(curl -s \
|
||||||
|
--header "Accept: application/vnd.docker.distribution.manifest.v2+json" \
|
||||||
|
--header "Accept: application/vnd.oci.image.index.v1+json" \
|
||||||
|
--header "Authorization: Bearer ${token}" \
|
||||||
|
"https://ghcr.io/v2/${image}/manifests/${tag}")
|
||||||
|
if jq -e '.layers // empty' <<< "${multidigest}" >/dev/null 2>&1; then
|
||||||
|
# If there's a layer element it's a single-arch manifest so just get that digest
|
||||||
|
digest=$(jq -r '.config.digest' <<< "${multidigest}")
|
||||||
|
else
|
||||||
|
# Otherwise it's multi-arch or has manifest annotations
|
||||||
|
if jq -e '.manifests[]?.annotations // empty' <<< "${multidigest}" >/dev/null 2>&1; then
|
||||||
|
# Check for manifest annotations and delete if found
|
||||||
|
multidigest=$(jq 'del(.manifests[] | select(.annotations))' <<< "${multidigest}")
|
||||||
|
fi
|
||||||
|
if [[ $(jq '.manifests | length' <<< "${multidigest}") -gt 1 ]]; then
|
||||||
|
# If there's still more than one digest, it's multi-arch
|
||||||
|
multidigest=$(jq -r ".manifests[] | select(.platform.architecture == \"amd64\").digest?" <<< "${multidigest}")
|
||||||
|
else
|
||||||
|
# Otherwise it's single arch
|
||||||
|
multidigest=$(jq -r ".manifests[].digest?" <<< "${multidigest}")
|
||||||
|
fi
|
||||||
|
if digest=$(curl -s \
|
||||||
--header "Accept: application/vnd.docker.distribution.manifest.v2+json" \
|
--header "Accept: application/vnd.docker.distribution.manifest.v2+json" \
|
||||||
|
--header "Accept: application/vnd.oci.image.manifest.v1+json" \
|
||||||
--header "Authorization: Bearer ${token}" \
|
--header "Authorization: Bearer ${token}" \
|
||||||
"https://ghcr.io/v2/${image}/manifests/${tag}" \
|
"https://ghcr.io/v2/${image}/manifests/${multidigest}"); then
|
||||||
| jq -r 'first(.manifests[].digest)')
|
digest=$(jq -r '.config.digest' <<< "${digest}");
|
||||||
digest=$(curl -s \
|
fi
|
||||||
--header "Accept: application/vnd.docker.distribution.manifest.v2+json" \
|
fi
|
||||||
--header "Authorization: Bearer ${token}" \
|
|
||||||
"https://ghcr.io/v2/${image}/manifests/${multidigest}" \
|
|
||||||
| jq -r '.config.digest')
|
|
||||||
image_info=$(curl -sL \
|
image_info=$(curl -sL \
|
||||||
--header "Authorization: Bearer ${token}" \
|
--header "Authorization: Bearer ${token}" \
|
||||||
"https://ghcr.io/v2/${image}/blobs/${digest}")
|
"https://ghcr.io/v2/${image}/blobs/${digest}")
|
||||||
|
|
@ -60,45 +94,54 @@ jobs:
|
||||||
IMAGE_RELEASE=$(echo ${image_info} | jq -r '.Labels.build_version' | awk '{print $3}')
|
IMAGE_RELEASE=$(echo ${image_info} | jq -r '.Labels.build_version' | awk '{print $3}')
|
||||||
IMAGE_VERSION=$(echo ${IMAGE_RELEASE} | awk -F'-ls' '{print $1}')
|
IMAGE_VERSION=$(echo ${IMAGE_RELEASE} | awk -F'-ls' '{print $1}')
|
||||||
if [ -z "${IMAGE_VERSION}" ]; then
|
if [ -z "${IMAGE_VERSION}" ]; then
|
||||||
echo "**** Can't retrieve last pushed version, exiting ****"
|
echo "> [!WARNING]" >> $GITHUB_STEP_SUMMARY
|
||||||
|
echo "Can't retrieve last pushed version, exiting" >> $GITHUB_STEP_SUMMARY
|
||||||
FAILURE_REASON="Can't retrieve last pushed version for plex tag latest"
|
FAILURE_REASON="Can't retrieve last pushed version for plex tag latest"
|
||||||
curl -X POST -H "Content-Type: application/json" --data '{"avatar_url": "https://cdn.discordapp.com/avatars/354986384542662657/df91181b3f1cf0ef1592fbe18e0962d7.png","embeds": [{"color": 16711680,
|
curl -X POST -H "Content-Type: application/json" --data '{"avatar_url": "https://cdn.discordapp.com/avatars/354986384542662657/df91181b3f1cf0ef1592fbe18e0962d7.png","embeds": [{"color": 16711680,
|
||||||
"description": "**Trigger Failed** \n**Reason:** '"${FAILURE_REASON}"' \n"}],
|
"description": "**Trigger Failed** \n**Reason:** '"${FAILURE_REASON}"' \n"}],
|
||||||
"username": "Github Actions"}' ${{ secrets.DISCORD_WEBHOOK }}
|
"username": "Github Actions"}' ${{ secrets.DISCORD_WEBHOOK }}
|
||||||
exit 1
|
exit 1
|
||||||
fi
|
fi
|
||||||
echo "**** Last pushed version: ${IMAGE_VERSION} ****"
|
echo "Last pushed version: \`${IMAGE_VERSION}\`" >> $GITHUB_STEP_SUMMARY
|
||||||
echo "Last pushed version: ${IMAGE_VERSION}" >> $GITHUB_STEP_SUMMARY
|
if [ "${EXT_RELEASE_SANITIZED}" == "${IMAGE_VERSION}" ]; then
|
||||||
if [ "${EXT_RELEASE}" == "${IMAGE_VERSION}" ]; then
|
echo "Sanitized version \`${EXT_RELEASE_SANITIZED}\` already pushed, exiting" >> $GITHUB_STEP_SUMMARY
|
||||||
echo "**** Version ${EXT_RELEASE} already pushed, exiting ****"
|
|
||||||
echo "Version ${EXT_RELEASE} already pushed, exiting" >> $GITHUB_STEP_SUMMARY
|
|
||||||
exit 0
|
exit 0
|
||||||
elif [ $(curl -s https://ci.linuxserver.io/job/Docker-Pipeline-Builders/job/docker-plex/job/master/lastBuild/api/json | jq -r '.building') == "true" ]; then
|
elif [ $(curl -s https://ci.linuxserver.io/job/Docker-Pipeline-Builders/job/docker-plex/job/master/lastBuild/api/json | jq -r '.building') == "true" ]; then
|
||||||
echo "**** New version ${EXT_RELEASE} found; but there already seems to be an active build on Jenkins; exiting ****"
|
echo "New version \`${EXT_RELEASE}\` found; but there already seems to be an active build on Jenkins; exiting" >> $GITHUB_STEP_SUMMARY
|
||||||
echo "New version ${EXT_RELEASE} found; but there already seems to be an active build on Jenkins; exiting" >> $GITHUB_STEP_SUMMARY
|
|
||||||
exit 0
|
exit 0
|
||||||
else
|
else
|
||||||
echo "**** New version ${EXT_RELEASE} found; old version was ${IMAGE_VERSION}. Triggering new build ****"
|
if [[ "${artifacts_found}" == "false" ]]; then
|
||||||
echo "New version ${EXT_RELEASE} found; old version was ${IMAGE_VERSION}. Triggering new build" >> $GITHUB_STEP_SUMMARY
|
echo "> [!WARNING]" >> $GITHUB_STEP_SUMMARY
|
||||||
response=$(curl -iX POST \
|
echo "> New version detected, but not all artifacts are published yet; skipping trigger" >> $GITHUB_STEP_SUMMARY
|
||||||
https://ci.linuxserver.io/job/Docker-Pipeline-Builders/job/docker-plex/job/master/buildWithParameters?PACKAGE_CHECK=false \
|
FAILURE_REASON="New version ${EXT_RELEASE} for plex tag latest is detected, however not all artifacts are uploaded to upstream release yet. Will try again later."
|
||||||
--user ${{ secrets.JENKINS_USER }}:${{ secrets.JENKINS_TOKEN }} | grep -i location | sed "s|^[L|l]ocation: \(.*\)|\1|")
|
curl -X POST -H "Content-Type: application/json" --data '{"avatar_url": "https://cdn.discordapp.com/avatars/354986384542662657/df91181b3f1cf0ef1592fbe18e0962d7.png","embeds": [{"color": 9802903,
|
||||||
echo "**** Jenkins job queue url: ${response%$'\r'} ****"
|
"description": "**Trigger Failed** \n**Reason:** '"${FAILURE_REASON}"' \n"}],
|
||||||
echo "**** Sleeping 10 seconds until job starts ****"
|
"username": "Github Actions"}' ${{ secrets.DISCORD_WEBHOOK }}
|
||||||
sleep 10
|
else
|
||||||
buildurl=$(curl -s "${response%$'\r'}api/json" | jq -r '.executable.url')
|
printf "\n## Trigger new build\n\n" >> $GITHUB_STEP_SUMMARY
|
||||||
buildurl="${buildurl%$'\r'}"
|
echo "New sanitized version \`${EXT_RELEASE_SANITIZED}\` found; old version was \`${IMAGE_VERSION}\`. Triggering new build" >> $GITHUB_STEP_SUMMARY
|
||||||
echo "**** Jenkins job build url: ${buildurl} ****"
|
if [[ "${artifacts_found}" == "true" ]]; then
|
||||||
echo "Jenkins job build url: ${buildurl}" >> $GITHUB_STEP_SUMMARY
|
echo "All artifacts seem to be uploaded." >> $GITHUB_STEP_SUMMARY
|
||||||
echo "**** Attempting to change the Jenkins job description ****"
|
fi
|
||||||
curl -iX POST \
|
response=$(curl -iX POST \
|
||||||
"${buildurl}submitDescription" \
|
https://ci.linuxserver.io/job/Docker-Pipeline-Builders/job/docker-plex/job/master/buildWithParameters?PACKAGE_CHECK=false \
|
||||||
--user ${{ secrets.JENKINS_USER }}:${{ secrets.JENKINS_TOKEN }} \
|
--user ${{ secrets.JENKINS_USER }}:${{ secrets.JENKINS_TOKEN }} | grep -i location | sed "s|^[L|l]ocation: \(.*\)|\1|")
|
||||||
--data-urlencode "description=GHA external trigger https://github.com/${{ github.repository }}/actions/runs/${{ github.run_id }}" \
|
echo "Jenkins [job queue url](${response%$'\r'})" >> $GITHUB_STEP_SUMMARY
|
||||||
--data-urlencode "Submit=Submit"
|
echo "Sleeping 10 seconds until job starts" >> $GITHUB_STEP_SUMMARY
|
||||||
echo "**** Notifying Discord ****"
|
sleep 10
|
||||||
TRIGGER_REASON="A version change was detected for plex tag latest. Old version:${IMAGE_VERSION} New version:${EXT_RELEASE}"
|
buildurl=$(curl -s "${response%$'\r'}api/json" | jq -r '.executable.url')
|
||||||
curl -X POST -H "Content-Type: application/json" --data '{"avatar_url": "https://cdn.discordapp.com/avatars/354986384542662657/df91181b3f1cf0ef1592fbe18e0962d7.png","embeds": [{"color": 9802903,
|
buildurl="${buildurl%$'\r'}"
|
||||||
"description": "**Build Triggered** \n**Reason:** '"${TRIGGER_REASON}"' \n**Build URL:** '"${buildurl}display/redirect"' \n"}],
|
echo "Jenkins job [build url](${buildurl})" >> $GITHUB_STEP_SUMMARY
|
||||||
"username": "Github Actions"}' ${{ secrets.DISCORD_WEBHOOK }}
|
echo "Attempting to change the Jenkins job description" >> $GITHUB_STEP_SUMMARY
|
||||||
|
curl -iX POST \
|
||||||
|
"${buildurl}submitDescription" \
|
||||||
|
--user ${{ secrets.JENKINS_USER }}:${{ secrets.JENKINS_TOKEN }} \
|
||||||
|
--data-urlencode "description=GHA external trigger https://github.com/${{ github.repository }}/actions/runs/${{ github.run_id }}" \
|
||||||
|
--data-urlencode "Submit=Submit"
|
||||||
|
echo "**** Notifying Discord ****"
|
||||||
|
TRIGGER_REASON="A version change was detected for plex tag latest. Old version:${IMAGE_VERSION} New version:${EXT_RELEASE_SANITIZED}"
|
||||||
|
curl -X POST -H "Content-Type: application/json" --data '{"avatar_url": "https://cdn.discordapp.com/avatars/354986384542662657/df91181b3f1cf0ef1592fbe18e0962d7.png","embeds": [{"color": 9802903,
|
||||||
|
"description": "**Build Triggered** \n**Reason:** '"${TRIGGER_REASON}"' \n**Build URL:** '"${buildurl}display/redirect"' \n"}],
|
||||||
|
"username": "Github Actions"}' ${{ secrets.DISCORD_WEBHOOK }}
|
||||||
|
fi
|
||||||
fi
|
fi
|
||||||
|
|
|
||||||
27
.github/workflows/external_trigger_scheduler.yml
vendored
27
.github/workflows/external_trigger_scheduler.yml
vendored
|
|
@ -5,6 +5,9 @@ on:
|
||||||
- cron: '14 * * * *'
|
- cron: '14 * * * *'
|
||||||
workflow_dispatch:
|
workflow_dispatch:
|
||||||
|
|
||||||
|
permissions:
|
||||||
|
contents: read
|
||||||
|
|
||||||
jobs:
|
jobs:
|
||||||
external-trigger-scheduler:
|
external-trigger-scheduler:
|
||||||
runs-on: ubuntu-latest
|
runs-on: ubuntu-latest
|
||||||
|
|
@ -15,31 +18,31 @@ jobs:
|
||||||
|
|
||||||
- name: External Trigger Scheduler
|
- name: External Trigger Scheduler
|
||||||
run: |
|
run: |
|
||||||
echo "**** Branches found: ****"
|
printf "# External trigger scheduler for docker-plex\n\n" >> $GITHUB_STEP_SUMMARY
|
||||||
git for-each-ref --format='%(refname:short)' refs/remotes
|
printf "Found the branches:\n\n%s\n" "$(git for-each-ref --format='- %(refname:lstrip=3)' refs/remotes)" >> $GITHUB_STEP_SUMMARY
|
||||||
for br in $(git for-each-ref --format='%(refname:short)' refs/remotes)
|
for br in $(git for-each-ref --format='%(refname:lstrip=3)' refs/remotes)
|
||||||
do
|
do
|
||||||
br=$(echo "$br" | sed 's|origin/||g')
|
if [[ "${br}" == "HEAD" ]]; then
|
||||||
echo "**** Evaluating branch ${br} ****"
|
printf "\nSkipping %s.\n" ${br} >> $GITHUB_STEP_SUMMARY
|
||||||
|
continue
|
||||||
|
fi
|
||||||
|
printf "\n## Evaluating \`%s\`\n\n" ${br} >> $GITHUB_STEP_SUMMARY
|
||||||
ls_jenkins_vars=$(curl -sX GET https://raw.githubusercontent.com/linuxserver/docker-plex/${br}/jenkins-vars.yml)
|
ls_jenkins_vars=$(curl -sX GET https://raw.githubusercontent.com/linuxserver/docker-plex/${br}/jenkins-vars.yml)
|
||||||
ls_branch=$(echo "${ls_jenkins_vars}" | yq -r '.ls_branch')
|
ls_branch=$(echo "${ls_jenkins_vars}" | yq -r '.ls_branch')
|
||||||
ls_trigger=$(echo "${ls_jenkins_vars}" | yq -r '.external_type')
|
ls_trigger=$(echo "${ls_jenkins_vars}" | yq -r '.external_type')
|
||||||
if [[ "${br}" == "${ls_branch}" ]] && [[ "${ls_trigger}" != "os" ]]; then
|
if [[ "${br}" == "${ls_branch}" ]] && [[ "${ls_trigger}" != "os" ]]; then
|
||||||
echo "**** Branch ${br} appears to be live and trigger is not os; checking workflow. ****"
|
echo "Branch appears to be live and trigger is not os; checking workflow." >> $GITHUB_STEP_SUMMARY
|
||||||
if curl -sfX GET https://raw.githubusercontent.com/linuxserver/docker-plex/${br}/.github/workflows/external_trigger.yml > /dev/null 2>&1; then
|
if curl -sfX GET https://raw.githubusercontent.com/linuxserver/docker-plex/${br}/.github/workflows/external_trigger.yml > /dev/null 2>&1; then
|
||||||
echo "**** Workflow exists. Triggering external trigger workflow for branch ${br} ****."
|
echo "Triggering external trigger workflow for branch." >> $GITHUB_STEP_SUMMARY
|
||||||
echo "Triggering external trigger workflow for branch ${br}" >> $GITHUB_STEP_SUMMARY
|
|
||||||
curl -iX POST \
|
curl -iX POST \
|
||||||
-H "Authorization: token ${{ secrets.CR_PAT }}" \
|
-H "Authorization: token ${{ secrets.CR_PAT }}" \
|
||||||
-H "Accept: application/vnd.github.v3+json" \
|
-H "Accept: application/vnd.github.v3+json" \
|
||||||
-d "{\"ref\":\"refs/heads/${br}\"}" \
|
-d "{\"ref\":\"refs/heads/${br}\"}" \
|
||||||
https://api.github.com/repos/linuxserver/docker-plex/actions/workflows/external_trigger.yml/dispatches
|
https://api.github.com/repos/linuxserver/docker-plex/actions/workflows/external_trigger.yml/dispatches
|
||||||
else
|
else
|
||||||
echo "**** Workflow doesn't exist; skipping trigger. ****"
|
echo "Skipping branch due to no external trigger workflow present." >> $GITHUB_STEP_SUMMARY
|
||||||
echo "Skipping branch ${br} due to no external trigger workflow present." >> $GITHUB_STEP_SUMMARY
|
|
||||||
fi
|
fi
|
||||||
else
|
else
|
||||||
echo "**** ${br} is either a dev branch, or has no external version; skipping trigger. ****"
|
echo "Skipping branch due to being detected as dev branch or having no external version." >> $GITHUB_STEP_SUMMARY
|
||||||
echo "Skipping branch ${br} due to being detected as dev branch or having no external version." >> $GITHUB_STEP_SUMMARY
|
|
||||||
fi
|
fi
|
||||||
done
|
done
|
||||||
|
|
|
||||||
6
.github/workflows/greetings.yml
vendored
6
.github/workflows/greetings.yml
vendored
|
|
@ -2,8 +2,14 @@ name: Greetings
|
||||||
|
|
||||||
on: [pull_request_target, issues]
|
on: [pull_request_target, issues]
|
||||||
|
|
||||||
|
permissions:
|
||||||
|
contents: read
|
||||||
|
|
||||||
jobs:
|
jobs:
|
||||||
greeting:
|
greeting:
|
||||||
|
permissions:
|
||||||
|
issues: write
|
||||||
|
pull-requests: write
|
||||||
runs-on: ubuntu-latest
|
runs-on: ubuntu-latest
|
||||||
steps:
|
steps:
|
||||||
- uses: actions/first-interaction@v1
|
- uses: actions/first-interaction@v1
|
||||||
|
|
|
||||||
42
.github/workflows/package_trigger.yml
vendored
42
.github/workflows/package_trigger.yml
vendored
|
|
@ -1,42 +0,0 @@
|
||||||
name: Package Trigger Main
|
|
||||||
|
|
||||||
on:
|
|
||||||
workflow_dispatch:
|
|
||||||
|
|
||||||
jobs:
|
|
||||||
package-trigger-master:
|
|
||||||
runs-on: ubuntu-latest
|
|
||||||
steps:
|
|
||||||
- uses: actions/checkout@v4.1.1
|
|
||||||
|
|
||||||
- name: Package Trigger
|
|
||||||
if: github.ref == 'refs/heads/master'
|
|
||||||
run: |
|
|
||||||
if [ -n "${{ secrets.PAUSE_PACKAGE_TRIGGER_PLEX_MASTER }}" ]; then
|
|
||||||
echo "**** Github secret PAUSE_PACKAGE_TRIGGER_PLEX_MASTER is set; skipping trigger. ****"
|
|
||||||
echo "Github secret \`PAUSE_PACKAGE_TRIGGER_PLEX_MASTER\` is set; skipping trigger." >> $GITHUB_STEP_SUMMARY
|
|
||||||
exit 0
|
|
||||||
fi
|
|
||||||
if [ $(curl -s https://ci.linuxserver.io/job/Docker-Pipeline-Builders/job/docker-plex/job/master/lastBuild/api/json | jq -r '.building') == "true" ]; then
|
|
||||||
echo "**** There already seems to be an active build on Jenkins; skipping package trigger ****"
|
|
||||||
echo "There already seems to be an active build on Jenkins; skipping package trigger" >> $GITHUB_STEP_SUMMARY
|
|
||||||
exit 0
|
|
||||||
fi
|
|
||||||
echo "**** Package trigger running off of master branch. To disable, set a Github secret named \"PAUSE_PACKAGE_TRIGGER_PLEX_MASTER\". ****"
|
|
||||||
echo "Package trigger running off of master branch. To disable, set a Github secret named \`PAUSE_PACKAGE_TRIGGER_PLEX_MASTER\`" >> $GITHUB_STEP_SUMMARY
|
|
||||||
response=$(curl -iX POST \
|
|
||||||
https://ci.linuxserver.io/job/Docker-Pipeline-Builders/job/docker-plex/job/master/buildWithParameters?PACKAGE_CHECK=true \
|
|
||||||
--user ${{ secrets.JENKINS_USER }}:${{ secrets.JENKINS_TOKEN }} | grep -i location | sed "s|^[L|l]ocation: \(.*\)|\1|")
|
|
||||||
echo "**** Jenkins job queue url: ${response%$'\r'} ****"
|
|
||||||
echo "**** Sleeping 10 seconds until job starts ****"
|
|
||||||
sleep 10
|
|
||||||
buildurl=$(curl -s "${response%$'\r'}api/json" | jq -r '.executable.url')
|
|
||||||
buildurl="${buildurl%$'\r'}"
|
|
||||||
echo "**** Jenkins job build url: ${buildurl} ****"
|
|
||||||
echo "Jenkins job build url: ${buildurl}" >> $GITHUB_STEP_SUMMARY
|
|
||||||
echo "**** Attempting to change the Jenkins job description ****"
|
|
||||||
curl -iX POST \
|
|
||||||
"${buildurl}submitDescription" \
|
|
||||||
--user ${{ secrets.JENKINS_USER }}:${{ secrets.JENKINS_TOKEN }} \
|
|
||||||
--data-urlencode "description=GHA package trigger https://github.com/${{ github.repository }}/actions/runs/${{ github.run_id }}" \
|
|
||||||
--data-urlencode "Submit=Submit"
|
|
||||||
105
.github/workflows/package_trigger_scheduler.yml
vendored
105
.github/workflows/package_trigger_scheduler.yml
vendored
|
|
@ -5,6 +5,9 @@ on:
|
||||||
- cron: '0 9 * * 1'
|
- cron: '0 9 * * 1'
|
||||||
workflow_dispatch:
|
workflow_dispatch:
|
||||||
|
|
||||||
|
permissions:
|
||||||
|
contents: read
|
||||||
|
|
||||||
jobs:
|
jobs:
|
||||||
package-trigger-scheduler:
|
package-trigger-scheduler:
|
||||||
runs-on: ubuntu-latest
|
runs-on: ubuntu-latest
|
||||||
|
|
@ -14,37 +17,87 @@ jobs:
|
||||||
fetch-depth: '0'
|
fetch-depth: '0'
|
||||||
|
|
||||||
- name: Package Trigger Scheduler
|
- name: Package Trigger Scheduler
|
||||||
|
env:
|
||||||
|
SKIP_PACKAGE_TRIGGER: ${{ vars.SKIP_PACKAGE_TRIGGER }}
|
||||||
run: |
|
run: |
|
||||||
echo "**** Branches found: ****"
|
printf "# Package trigger scheduler for docker-plex\n\n" >> $GITHUB_STEP_SUMMARY
|
||||||
git for-each-ref --format='%(refname:short)' refs/remotes
|
printf "Found the branches:\n\n%s\n" "$(git for-each-ref --format='- %(refname:lstrip=3)' refs/remotes)" >> $GITHUB_STEP_SUMMARY
|
||||||
for br in $(git for-each-ref --format='%(refname:short)' refs/remotes)
|
for br in $(git for-each-ref --format='%(refname:lstrip=3)' refs/remotes)
|
||||||
do
|
do
|
||||||
br=$(echo "$br" | sed 's|origin/||g')
|
if [[ "${br}" == "HEAD" ]]; then
|
||||||
echo "**** Evaluating branch ${br} ****"
|
printf "\nSkipping %s.\n" ${br} >> $GITHUB_STEP_SUMMARY
|
||||||
ls_branch=$(curl -sX GET https://raw.githubusercontent.com/linuxserver/docker-plex/${br}/jenkins-vars.yml | yq -r '.ls_branch')
|
continue
|
||||||
if [ "${br}" == "${ls_branch}" ]; then
|
fi
|
||||||
echo "**** Branch ${br} appears to be live; checking workflow. ****"
|
printf "\n## Evaluating \`%s\`\n\n" ${br} >> $GITHUB_STEP_SUMMARY
|
||||||
if curl -sfX GET https://raw.githubusercontent.com/linuxserver/docker-plex/${br}/.github/workflows/package_trigger.yml > /dev/null 2>&1; then
|
JENKINS_VARS=$(curl -sX GET https://raw.githubusercontent.com/linuxserver/docker-plex/${br}/jenkins-vars.yml)
|
||||||
echo "**** Workflow exists. Triggering package trigger workflow for branch ${br}. ****"
|
if ! curl -sfX GET https://raw.githubusercontent.com/linuxserver/docker-plex/${br}/Jenkinsfile >/dev/null 2>&1; then
|
||||||
echo "Triggering package trigger workflow for branch ${br}" >> $GITHUB_STEP_SUMMARY
|
echo "> [!WARNING]" >> $GITHUB_STEP_SUMMARY
|
||||||
triggered_branches="${triggered_branches}${br} "
|
echo "> No Jenkinsfile found. Branch is either deprecated or is an early dev branch." >> $GITHUB_STEP_SUMMARY
|
||||||
curl -iX POST \
|
skipped_branches="${skipped_branches}${br} "
|
||||||
-H "Authorization: token ${{ secrets.CR_PAT }}" \
|
elif [[ "${br}" == $(yq -r '.ls_branch' <<< "${JENKINS_VARS}") ]]; then
|
||||||
-H "Accept: application/vnd.github.v3+json" \
|
echo "Branch appears to be live; checking workflow." >> $GITHUB_STEP_SUMMARY
|
||||||
-d "{\"ref\":\"refs/heads/${br}\"}" \
|
README_VARS=$(curl -sX GET https://raw.githubusercontent.com/linuxserver/docker-plex/${br}/readme-vars.yml)
|
||||||
https://api.github.com/repos/linuxserver/docker-plex/actions/workflows/package_trigger.yml/dispatches
|
if [[ $(yq -r '.project_deprecation_status' <<< "${README_VARS}") == "true" ]]; then
|
||||||
sleep 30
|
echo "> [!WARNING]" >> $GITHUB_STEP_SUMMARY
|
||||||
|
echo "> Branch appears to be deprecated; skipping trigger." >> $GITHUB_STEP_SUMMARY
|
||||||
|
skipped_branches="${skipped_branches}${br} "
|
||||||
|
elif [[ $(yq -r '.skip_package_check' <<< "${JENKINS_VARS}") == "true" ]]; then
|
||||||
|
echo "> [!WARNING]" >> $GITHUB_STEP_SUMMARY
|
||||||
|
echo "> Skipping branch ${br} due to \`skip_package_check\` being set in \`jenkins-vars.yml\`." >> $GITHUB_STEP_SUMMARY
|
||||||
|
skipped_branches="${skipped_branches}${br} "
|
||||||
|
elif grep -q "^plex_${br}" <<< "${SKIP_PACKAGE_TRIGGER}"; then
|
||||||
|
echo "> [!WARNING]" >> $GITHUB_STEP_SUMMARY
|
||||||
|
echo "> Github organizational variable \`SKIP_PACKAGE_TRIGGER\` contains \`plex_${br}\`; skipping trigger." >> $GITHUB_STEP_SUMMARY
|
||||||
|
skipped_branches="${skipped_branches}${br} "
|
||||||
|
elif [ $(curl -s https://ci.linuxserver.io/job/Docker-Pipeline-Builders/job/docker-plex/job/${br}/lastBuild/api/json | jq -r '.building' 2>/dev/null) == "true" ]; then
|
||||||
|
echo "> [!WARNING]" >> $GITHUB_STEP_SUMMARY
|
||||||
|
echo "> There already seems to be an active build on Jenkins; skipping package trigger for ${br}" >> $GITHUB_STEP_SUMMARY
|
||||||
|
skipped_branches="${skipped_branches}${br} "
|
||||||
else
|
else
|
||||||
echo "**** Workflow doesn't exist; skipping trigger. ****"
|
echo "> [!NOTE]" >> $GITHUB_STEP_SUMMARY
|
||||||
echo "Skipping branch ${br} due to no package trigger workflow present." >> $GITHUB_STEP_SUMMARY
|
echo "> Triggering package trigger for branch ${br}" >> $GITHUB_STEP_SUMMARY
|
||||||
|
printf "> To disable, add \`plex_%s\` into the Github organizational variable \`SKIP_PACKAGE_TRIGGER\`.\n\n" "${br}" >> $GITHUB_STEP_SUMMARY
|
||||||
|
triggered_branches="${triggered_branches}${br} "
|
||||||
|
response=$(curl -iX POST \
|
||||||
|
https://ci.linuxserver.io/job/Docker-Pipeline-Builders/job/docker-plex/job/${br}/buildWithParameters?PACKAGE_CHECK=true \
|
||||||
|
--user ${{ secrets.JENKINS_USER }}:${{ secrets.JENKINS_TOKEN }} | grep -i location | sed "s|^[L|l]ocation: \(.*\)|\1|")
|
||||||
|
if [[ -z "${response}" ]]; then
|
||||||
|
echo "> [!WARNING]" >> $GITHUB_STEP_SUMMARY
|
||||||
|
echo "> Jenkins build could not be triggered. Skipping branch."
|
||||||
|
continue
|
||||||
|
fi
|
||||||
|
echo "Jenkins [job queue url](${response%$'\r'})" >> $GITHUB_STEP_SUMMARY
|
||||||
|
echo "Sleeping 10 seconds until job starts" >> $GITHUB_STEP_SUMMARY
|
||||||
|
sleep 10
|
||||||
|
buildurl=$(curl -s "${response%$'\r'}api/json" | jq -r '.executable.url')
|
||||||
|
buildurl="${buildurl%$'\r'}"
|
||||||
|
echo "Jenkins job [build url](${buildurl})" >> $GITHUB_STEP_SUMMARY
|
||||||
|
echo "Attempting to change the Jenkins job description" >> $GITHUB_STEP_SUMMARY
|
||||||
|
if ! curl -ifX POST \
|
||||||
|
"${buildurl}submitDescription" \
|
||||||
|
--user ${{ secrets.JENKINS_USER }}:${{ secrets.JENKINS_TOKEN }} \
|
||||||
|
--data-urlencode "description=GHA package trigger https://github.com/${{ github.repository }}/actions/runs/${{ github.run_id }}" \
|
||||||
|
--data-urlencode "Submit=Submit"; then
|
||||||
|
echo "> [!WARNING]" >> $GITHUB_STEP_SUMMARY
|
||||||
|
echo "> Unable to change the Jenkins job description."
|
||||||
|
fi
|
||||||
|
sleep 20
|
||||||
fi
|
fi
|
||||||
else
|
else
|
||||||
echo "**** ${br} appears to be a dev branch; skipping trigger. ****"
|
|
||||||
echo "Skipping branch ${br} due to being detected as dev branch." >> $GITHUB_STEP_SUMMARY
|
echo "Skipping branch ${br} due to being detected as dev branch." >> $GITHUB_STEP_SUMMARY
|
||||||
fi
|
fi
|
||||||
done
|
done
|
||||||
echo "**** Package check build(s) triggered for branch(es): ${triggered_branches} ****"
|
if [[ -n "${triggered_branches}" ]] || [[ -n "${skipped_branches}" ]]; then
|
||||||
echo "**** Notifying Discord ****"
|
if [[ -n "${triggered_branches}" ]]; then
|
||||||
curl -X POST -H "Content-Type: application/json" --data '{"avatar_url": "https://cdn.discordapp.com/avatars/354986384542662657/df91181b3f1cf0ef1592fbe18e0962d7.png","embeds": [{"color": 9802903,
|
NOTIFY_BRANCHES="**Triggered:** ${triggered_branches} \n"
|
||||||
"description": "**Package Check Build(s) Triggered for plex** \n**Branch(es):** '"${triggered_branches}"' \n**Build URL:** '"https://ci.linuxserver.io/blue/organizations/jenkins/Docker-Pipeline-Builders%2Fdocker-plex/activity/"' \n"}],
|
NOTIFY_BUILD_URL="**Build URL:** https://ci.linuxserver.io/blue/organizations/jenkins/Docker-Pipeline-Builders%2Fdocker-plex/activity/ \n"
|
||||||
"username": "Github Actions"}' ${{ secrets.DISCORD_WEBHOOK }}
|
echo "**** Package check build(s) triggered for branch(es): ${triggered_branches} ****"
|
||||||
|
fi
|
||||||
|
if [[ -n "${skipped_branches}" ]]; then
|
||||||
|
NOTIFY_BRANCHES="${NOTIFY_BRANCHES}**Skipped:** ${skipped_branches} \n"
|
||||||
|
fi
|
||||||
|
echo "**** Notifying Discord ****"
|
||||||
|
curl -X POST -H "Content-Type: application/json" --data '{"avatar_url": "https://cdn.discordapp.com/avatars/354986384542662657/df91181b3f1cf0ef1592fbe18e0962d7.png","embeds": [{"color": 9802903,
|
||||||
|
"description": "**Package Check Build(s) for plex** \n'"${NOTIFY_BRANCHES}"''"${NOTIFY_BUILD_URL}"'"}],
|
||||||
|
"username": "Github Actions"}' ${{ secrets.DISCORD_WEBHOOK }}
|
||||||
|
fi
|
||||||
|
|
|
||||||
2
.github/workflows/permissions.yml
vendored
2
.github/workflows/permissions.yml
vendored
|
|
@ -5,6 +5,8 @@ on:
|
||||||
- '**/run'
|
- '**/run'
|
||||||
- '**/finish'
|
- '**/finish'
|
||||||
- '**/check'
|
- '**/check'
|
||||||
|
- 'root/migrations/*'
|
||||||
|
|
||||||
jobs:
|
jobs:
|
||||||
permission_check:
|
permission_check:
|
||||||
uses: linuxserver/github-workflows/.github/workflows/init-svc-executable-permissions.yml@v1
|
uses: linuxserver/github-workflows/.github/workflows/init-svc-executable-permissions.yml@v1
|
||||||
|
|
|
||||||
2
.gitignore
vendored
2
.gitignore
vendored
|
|
@ -1,3 +1,5 @@
|
||||||
|
# Wireguard configs, you never know
|
||||||
|
wg*.conf
|
||||||
# Windows image file caches
|
# Windows image file caches
|
||||||
Thumbs.db
|
Thumbs.db
|
||||||
ehthumbs.db
|
ehthumbs.db
|
||||||
|
|
|
||||||
21
Dockerfile
21
Dockerfile
|
|
@ -1,8 +1,8 @@
|
||||||
# syntax=docker/dockerfile:1
|
# syntax=docker/dockerfile:1
|
||||||
|
|
||||||
FROM ghcr.io/linuxserver/unrar:latest as unrar
|
FROM ghcr.io/linuxserver/unrar:latest AS unrar
|
||||||
|
|
||||||
FROM ghcr.io/linuxserver/baseimage-ubuntu:jammy
|
FROM ghcr.io/linuxserver/baseimage-ubuntu:noble
|
||||||
|
|
||||||
# set version label
|
# set version label
|
||||||
ARG BUILD_DATE
|
ARG BUILD_DATE
|
||||||
|
|
@ -12,7 +12,7 @@ LABEL build_version="Linuxserver.io version:- ${VERSION} Build-date:- ${BUILD_DA
|
||||||
LABEL maintainer="thelamer"
|
LABEL maintainer="thelamer"
|
||||||
|
|
||||||
#Add needed nvidia environment variables for https://github.com/NVIDIA/nvidia-docker
|
#Add needed nvidia environment variables for https://github.com/NVIDIA/nvidia-docker
|
||||||
ENV NVIDIA_DRIVER_CAPABILITIES="compute,video,utility"
|
ENV NVIDIA_DRIVER_CAPABILITIES="compute,video,utility,graphics"
|
||||||
|
|
||||||
# global environment settings
|
# global environment settings
|
||||||
ENV DEBIAN_FRONTEND="noninteractive" \
|
ENV DEBIAN_FRONTEND="noninteractive" \
|
||||||
|
|
@ -23,19 +23,22 @@ ENV DEBIAN_FRONTEND="noninteractive" \
|
||||||
PLEX_MEDIA_SERVER_MAX_PLUGIN_PROCS="6" \
|
PLEX_MEDIA_SERVER_MAX_PLUGIN_PROCS="6" \
|
||||||
PLEX_MEDIA_SERVER_USER="abc" \
|
PLEX_MEDIA_SERVER_USER="abc" \
|
||||||
PLEX_MEDIA_SERVER_INFO_VENDOR="Docker" \
|
PLEX_MEDIA_SERVER_INFO_VENDOR="Docker" \
|
||||||
PLEX_MEDIA_SERVER_INFO_DEVICE="Docker Container (LinuxServer.io)"
|
PLEX_MEDIA_SERVER_INFO_DEVICE="Docker Container (LinuxServer.io)" \
|
||||||
|
TMPDIR=/run/plex-temp \
|
||||||
|
ATTACHED_DEVICES_PERMS="/dev/dri /dev/dvb -type c"
|
||||||
|
|
||||||
RUN \
|
RUN \
|
||||||
echo "**** install runtime packages ****" && \
|
echo "**** install runtime packages & wireguard ****" && \
|
||||||
apt-get update && \
|
apt-get update && \
|
||||||
apt-get install -y \
|
apt-get install -y \
|
||||||
jq \
|
|
||||||
udev \
|
udev \
|
||||||
wget && \
|
wget \
|
||||||
|
iproute2 \
|
||||||
|
wireguard-tools &&\
|
||||||
echo "**** install plex ****" && \
|
echo "**** install plex ****" && \
|
||||||
if [ -z ${PLEX_RELEASE+x} ]; then \
|
if [ -z ${PLEX_RELEASE+x} ]; then \
|
||||||
PLEX_RELEASE=$(curl -sX GET 'https://plex.tv/api/downloads/5.json' \
|
PLEX_RELEASE=$(curl -sX GET 'https://plex.tv/api/downloads/5.json' \
|
||||||
| jq -r '.computer.Linux.version'); \
|
| jq -r '.computer.Linux.version'); \
|
||||||
fi && \
|
fi && \
|
||||||
curl -o \
|
curl -o \
|
||||||
/tmp/plexmediaserver.deb -L \
|
/tmp/plexmediaserver.deb -L \
|
||||||
|
|
@ -43,6 +46,7 @@ RUN \
|
||||||
dpkg -i /tmp/plexmediaserver.deb && \
|
dpkg -i /tmp/plexmediaserver.deb && \
|
||||||
echo "**** ensure abc user's home folder is /app ****" && \
|
echo "**** ensure abc user's home folder is /app ****" && \
|
||||||
usermod -d /app abc && \
|
usermod -d /app abc && \
|
||||||
|
printf "Linuxserver.io version: ${VERSION}\nBuild-date: ${BUILD_DATE}" > /build_version && \
|
||||||
echo "**** cleanup ****" && \
|
echo "**** cleanup ****" && \
|
||||||
apt-get clean && \
|
apt-get clean && \
|
||||||
rm -rf \
|
rm -rf \
|
||||||
|
|
@ -54,6 +58,7 @@ RUN \
|
||||||
# add local files
|
# add local files
|
||||||
COPY root/ /
|
COPY root/ /
|
||||||
|
|
||||||
|
|
||||||
# add unrar
|
# add unrar
|
||||||
COPY --from=unrar /usr/bin/unrar-ubuntu /usr/bin/unrar
|
COPY --from=unrar /usr/bin/unrar-ubuntu /usr/bin/unrar
|
||||||
|
|
||||||
|
|
|
||||||
|
|
@ -1,8 +1,8 @@
|
||||||
# syntax=docker/dockerfile:1
|
# syntax=docker/dockerfile:1
|
||||||
|
|
||||||
FROM ghcr.io/linuxserver/unrar:arm64v8-latest as unrar
|
FROM ghcr.io/linuxserver/unrar:arm64v8-latest AS unrar
|
||||||
|
|
||||||
FROM ghcr.io/linuxserver/baseimage-ubuntu:arm64v8-jammy
|
FROM ghcr.io/linuxserver/baseimage-ubuntu:arm64v8-noble
|
||||||
|
|
||||||
# set version label
|
# set version label
|
||||||
ARG BUILD_DATE
|
ARG BUILD_DATE
|
||||||
|
|
@ -20,20 +20,20 @@ ENV DEBIAN_FRONTEND="noninteractive" \
|
||||||
PLEX_MEDIA_SERVER_MAX_PLUGIN_PROCS="6" \
|
PLEX_MEDIA_SERVER_MAX_PLUGIN_PROCS="6" \
|
||||||
PLEX_MEDIA_SERVER_USER="abc" \
|
PLEX_MEDIA_SERVER_USER="abc" \
|
||||||
PLEX_MEDIA_SERVER_INFO_VENDOR="Docker" \
|
PLEX_MEDIA_SERVER_INFO_VENDOR="Docker" \
|
||||||
PLEX_MEDIA_SERVER_INFO_DEVICE="Docker Container (LinuxServer.io)"
|
PLEX_MEDIA_SERVER_INFO_DEVICE="Docker Container (LinuxServer.io)" \
|
||||||
|
TMPDIR=/run/plex-temp \
|
||||||
|
ATTACHED_DEVICES_PERMS="/dev/dri /dev/dvb -type c"
|
||||||
|
|
||||||
RUN \
|
RUN \
|
||||||
echo "**** install runtime packages ****" && \
|
echo "**** install runtime packages ****" && \
|
||||||
apt-get update && \
|
apt-get update && \
|
||||||
apt-get install -y \
|
apt-get install -y \
|
||||||
jq \
|
|
||||||
udev \
|
udev \
|
||||||
unrar \
|
|
||||||
wget && \
|
wget && \
|
||||||
echo "**** install plex ****" && \
|
echo "**** install plex ****" && \
|
||||||
if [ -z ${PLEX_RELEASE+x} ]; then \
|
if [ -z ${PLEX_RELEASE+x} ]; then \
|
||||||
PLEX_RELEASE=$(curl -sX GET 'https://plex.tv/api/downloads/5.json' \
|
PLEX_RELEASE=$(curl -sX GET 'https://plex.tv/api/downloads/5.json' \
|
||||||
| jq -r '.computer.Linux.version'); \
|
| jq -r '.computer.Linux.version'); \
|
||||||
fi && \
|
fi && \
|
||||||
curl -o \
|
curl -o \
|
||||||
/tmp/plexmediaserver.deb -L \
|
/tmp/plexmediaserver.deb -L \
|
||||||
|
|
@ -41,6 +41,7 @@ RUN \
|
||||||
dpkg -i /tmp/plexmediaserver.deb && \
|
dpkg -i /tmp/plexmediaserver.deb && \
|
||||||
echo "**** ensure abc user's home folder is /app ****" && \
|
echo "**** ensure abc user's home folder is /app ****" && \
|
||||||
usermod -d /app abc && \
|
usermod -d /app abc && \
|
||||||
|
printf "Linuxserver.io version: ${VERSION}\nBuild-date: ${BUILD_DATE}" > /build_version && \
|
||||||
echo "**** cleanup ****" && \
|
echo "**** cleanup ****" && \
|
||||||
apt-get clean && \
|
apt-get clean && \
|
||||||
rm -rf \
|
rm -rf \
|
||||||
|
|
|
||||||
676
Jenkinsfile
vendored
676
Jenkinsfile
vendored
|
|
@ -8,7 +8,7 @@ pipeline {
|
||||||
}
|
}
|
||||||
// Input to determine if this is a package check
|
// Input to determine if this is a package check
|
||||||
parameters {
|
parameters {
|
||||||
string(defaultValue: 'false', description: 'package check run', name: 'PACKAGE_CHECK')
|
string(defaultValue: 'false', description: 'package check run', name: 'PACKAGE_CHECK')
|
||||||
}
|
}
|
||||||
// Configuration for the variables used for this specific repo
|
// Configuration for the variables used for this specific repo
|
||||||
environment {
|
environment {
|
||||||
|
|
@ -17,6 +17,8 @@ pipeline {
|
||||||
GITLAB_TOKEN=credentials('b6f0f1dd-6952-4cf6-95d1-9c06380283f0')
|
GITLAB_TOKEN=credentials('b6f0f1dd-6952-4cf6-95d1-9c06380283f0')
|
||||||
GITLAB_NAMESPACE=credentials('gitlab-namespace-id')
|
GITLAB_NAMESPACE=credentials('gitlab-namespace-id')
|
||||||
DOCKERHUB_TOKEN=credentials('docker-hub-ci-pat')
|
DOCKERHUB_TOKEN=credentials('docker-hub-ci-pat')
|
||||||
|
QUAYIO_API_TOKEN=credentials('quayio-repo-api-token')
|
||||||
|
GIT_SIGNING_KEY=credentials('484fbca6-9a4f-455e-b9e3-97ac98785f5f')
|
||||||
JSON_URL = 'https://plex.tv/api/downloads/5.json'
|
JSON_URL = 'https://plex.tv/api/downloads/5.json'
|
||||||
JSON_PATH = '.computer.Linux.version'
|
JSON_PATH = '.computer.Linux.version'
|
||||||
BUILD_VERSION_ARG = 'PLEX_RELEASE'
|
BUILD_VERSION_ARG = 'PLEX_RELEASE'
|
||||||
|
|
@ -33,22 +35,49 @@ pipeline {
|
||||||
CI_PORT='32400'
|
CI_PORT='32400'
|
||||||
CI_SSL='false'
|
CI_SSL='false'
|
||||||
CI_DELAY='120'
|
CI_DELAY='120'
|
||||||
CI_DOCKERENV='TZ=US/Pacific'
|
CI_DOCKERENV=''
|
||||||
CI_AUTH='user:password'
|
CI_AUTH=''
|
||||||
CI_WEBPATH='/web/index.html'
|
CI_WEBPATH='/web/index.html'
|
||||||
}
|
}
|
||||||
stages {
|
stages {
|
||||||
|
stage("Set git config"){
|
||||||
|
steps{
|
||||||
|
sh '''#!/bin/bash
|
||||||
|
cat ${GIT_SIGNING_KEY} > /config/.ssh/id_sign
|
||||||
|
chmod 600 /config/.ssh/id_sign
|
||||||
|
ssh-keygen -y -f /config/.ssh/id_sign > /config/.ssh/id_sign.pub
|
||||||
|
echo "Using $(ssh-keygen -lf /config/.ssh/id_sign) to sign commits"
|
||||||
|
git config --global gpg.format ssh
|
||||||
|
git config --global user.signingkey /config/.ssh/id_sign
|
||||||
|
git config --global commit.gpgsign true
|
||||||
|
'''
|
||||||
|
}
|
||||||
|
}
|
||||||
// Setup all the basic environment variables needed for the build
|
// Setup all the basic environment variables needed for the build
|
||||||
stage("Set ENV Variables base"){
|
stage("Set ENV Variables base"){
|
||||||
steps{
|
steps{
|
||||||
|
echo "Running on node: ${NODE_NAME}"
|
||||||
sh '''#! /bin/bash
|
sh '''#! /bin/bash
|
||||||
containers=$(docker ps -aq)
|
echo "Pruning builder"
|
||||||
|
docker builder prune -f --builder container || :
|
||||||
|
containers=$(docker ps -q)
|
||||||
if [[ -n "${containers}" ]]; then
|
if [[ -n "${containers}" ]]; then
|
||||||
docker stop ${containers}
|
BUILDX_CONTAINER_ID=$(docker ps -qf 'name=buildx_buildkit')
|
||||||
|
for container in ${containers}; do
|
||||||
|
if [[ "${container}" == "${BUILDX_CONTAINER_ID}" ]]; then
|
||||||
|
echo "skipping buildx container in docker stop"
|
||||||
|
else
|
||||||
|
echo "Stopping container ${container}"
|
||||||
|
docker stop ${container}
|
||||||
|
fi
|
||||||
|
done
|
||||||
fi
|
fi
|
||||||
docker system prune -af --volumes || : '''
|
docker system prune -f --volumes || :
|
||||||
|
docker image prune -af || :
|
||||||
|
'''
|
||||||
script{
|
script{
|
||||||
env.EXIT_STATUS = ''
|
env.EXIT_STATUS = ''
|
||||||
|
env.CI_TEST_ATTEMPTED = ''
|
||||||
env.LS_RELEASE = sh(
|
env.LS_RELEASE = sh(
|
||||||
script: '''docker run --rm quay.io/skopeo/stable:v1 inspect docker://ghcr.io/${LS_USER}/${CONTAINER_NAME}:latest 2>/dev/null | jq -r '.Labels.build_version' | awk '{print $3}' | grep '\\-ls' || : ''',
|
script: '''docker run --rm quay.io/skopeo/stable:v1 inspect docker://ghcr.io/${LS_USER}/${CONTAINER_NAME}:latest 2>/dev/null | jq -r '.Labels.build_version' | awk '{print $3}' | grep '\\-ls' || : ''',
|
||||||
returnStdout: true).trim()
|
returnStdout: true).trim()
|
||||||
|
|
@ -67,8 +96,12 @@ pipeline {
|
||||||
env.CODE_URL = 'https://github.com/' + env.LS_USER + '/' + env.LS_REPO + '/commit/' + env.GIT_COMMIT
|
env.CODE_URL = 'https://github.com/' + env.LS_USER + '/' + env.LS_REPO + '/commit/' + env.GIT_COMMIT
|
||||||
env.DOCKERHUB_LINK = 'https://hub.docker.com/r/' + env.DOCKERHUB_IMAGE + '/tags/'
|
env.DOCKERHUB_LINK = 'https://hub.docker.com/r/' + env.DOCKERHUB_IMAGE + '/tags/'
|
||||||
env.PULL_REQUEST = env.CHANGE_ID
|
env.PULL_REQUEST = env.CHANGE_ID
|
||||||
env.TEMPLATED_FILES = 'Jenkinsfile README.md LICENSE .editorconfig ./.github/CONTRIBUTING.md ./.github/FUNDING.yml ./.github/ISSUE_TEMPLATE/config.yml ./.github/ISSUE_TEMPLATE/issue.bug.yml ./.github/ISSUE_TEMPLATE/issue.feature.yml ./.github/PULL_REQUEST_TEMPLATE.md ./.github/workflows/external_trigger_scheduler.yml ./.github/workflows/greetings.yml ./.github/workflows/package_trigger_scheduler.yml ./.github/workflows/call_issue_pr_tracker.yml ./.github/workflows/call_issues_cron.yml ./.github/workflows/permissions.yml ./.github/workflows/external_trigger.yml ./.github/workflows/package_trigger.yml'
|
env.TEMPLATED_FILES = 'Jenkinsfile README.md LICENSE .editorconfig ./.github/CONTRIBUTING.md ./.github/FUNDING.yml ./.github/ISSUE_TEMPLATE/config.yml ./.github/ISSUE_TEMPLATE/issue.bug.yml ./.github/ISSUE_TEMPLATE/issue.feature.yml ./.github/PULL_REQUEST_TEMPLATE.md ./.github/workflows/external_trigger_scheduler.yml ./.github/workflows/greetings.yml ./.github/workflows/package_trigger_scheduler.yml ./.github/workflows/call_issue_pr_tracker.yml ./.github/workflows/call_issues_cron.yml ./.github/workflows/permissions.yml ./.github/workflows/external_trigger.yml'
|
||||||
|
if ( env.SYFT_IMAGE_TAG == null ) {
|
||||||
|
env.SYFT_IMAGE_TAG = 'latest'
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
echo "Using syft image tag ${SYFT_IMAGE_TAG}"
|
||||||
sh '''#! /bin/bash
|
sh '''#! /bin/bash
|
||||||
echo "The default github branch detected as ${GH_DEFAULT_BRANCH}" '''
|
echo "The default github branch detected as ${GH_DEFAULT_BRANCH}" '''
|
||||||
script{
|
script{
|
||||||
|
|
@ -177,6 +210,8 @@ pipeline {
|
||||||
env.VERSION_TAG = env.EXT_RELEASE_CLEAN + '-ls' + env.LS_TAG_NUMBER
|
env.VERSION_TAG = env.EXT_RELEASE_CLEAN + '-ls' + env.LS_TAG_NUMBER
|
||||||
env.META_TAG = env.EXT_RELEASE_CLEAN + '-ls' + env.LS_TAG_NUMBER
|
env.META_TAG = env.EXT_RELEASE_CLEAN + '-ls' + env.LS_TAG_NUMBER
|
||||||
env.EXT_RELEASE_TAG = 'version-' + env.EXT_RELEASE_CLEAN
|
env.EXT_RELEASE_TAG = 'version-' + env.EXT_RELEASE_CLEAN
|
||||||
|
env.BUILDCACHE = 'docker.io/lsiodev/buildcache,registry.gitlab.com/linuxserver.io/docker-jenkins-builder/lsiodev-buildcache,ghcr.io/linuxserver/lsiodev-buildcache,quay.io/linuxserver.io/lsiodev-buildcache'
|
||||||
|
env.CITEST_IMAGETAG = 'latest'
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
@ -201,6 +236,8 @@ pipeline {
|
||||||
env.META_TAG = env.EXT_RELEASE_CLEAN + '-pkg-' + env.PACKAGE_TAG + '-dev-' + env.COMMIT_SHA
|
env.META_TAG = env.EXT_RELEASE_CLEAN + '-pkg-' + env.PACKAGE_TAG + '-dev-' + env.COMMIT_SHA
|
||||||
env.EXT_RELEASE_TAG = 'version-' + env.EXT_RELEASE_CLEAN
|
env.EXT_RELEASE_TAG = 'version-' + env.EXT_RELEASE_CLEAN
|
||||||
env.DOCKERHUB_LINK = 'https://hub.docker.com/r/' + env.DEV_DOCKERHUB_IMAGE + '/tags/'
|
env.DOCKERHUB_LINK = 'https://hub.docker.com/r/' + env.DEV_DOCKERHUB_IMAGE + '/tags/'
|
||||||
|
env.BUILDCACHE = 'docker.io/lsiodev/buildcache,registry.gitlab.com/linuxserver.io/docker-jenkins-builder/lsiodev-buildcache,ghcr.io/linuxserver/lsiodev-buildcache,quay.io/linuxserver.io/lsiodev-buildcache'
|
||||||
|
env.CITEST_IMAGETAG = 'develop'
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
@ -225,6 +262,8 @@ pipeline {
|
||||||
env.EXT_RELEASE_TAG = 'version-' + env.EXT_RELEASE_CLEAN
|
env.EXT_RELEASE_TAG = 'version-' + env.EXT_RELEASE_CLEAN
|
||||||
env.CODE_URL = 'https://github.com/' + env.LS_USER + '/' + env.LS_REPO + '/pull/' + env.PULL_REQUEST
|
env.CODE_URL = 'https://github.com/' + env.LS_USER + '/' + env.LS_REPO + '/pull/' + env.PULL_REQUEST
|
||||||
env.DOCKERHUB_LINK = 'https://hub.docker.com/r/' + env.PR_DOCKERHUB_IMAGE + '/tags/'
|
env.DOCKERHUB_LINK = 'https://hub.docker.com/r/' + env.PR_DOCKERHUB_IMAGE + '/tags/'
|
||||||
|
env.BUILDCACHE = 'docker.io/lsiodev/buildcache,registry.gitlab.com/linuxserver.io/docker-jenkins-builder/lsiodev-buildcache,ghcr.io/linuxserver/lsiodev-buildcache,quay.io/linuxserver.io/lsiodev-buildcache'
|
||||||
|
env.CITEST_IMAGETAG = 'develop'
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
@ -247,7 +286,7 @@ pipeline {
|
||||||
-v ${WORKSPACE}:/mnt \
|
-v ${WORKSPACE}:/mnt \
|
||||||
-e AWS_ACCESS_KEY_ID=\"${S3_KEY}\" \
|
-e AWS_ACCESS_KEY_ID=\"${S3_KEY}\" \
|
||||||
-e AWS_SECRET_ACCESS_KEY=\"${S3_SECRET}\" \
|
-e AWS_SECRET_ACCESS_KEY=\"${S3_SECRET}\" \
|
||||||
ghcr.io/linuxserver/baseimage-alpine:3.19 s6-envdir -fn -- /var/run/s6/container_environment /bin/bash -c "\
|
ghcr.io/linuxserver/baseimage-alpine:3.23 s6-envdir -fn -- /var/run/s6/container_environment /bin/bash -c "\
|
||||||
apk add --no-cache python3 && \
|
apk add --no-cache python3 && \
|
||||||
python3 -m venv /lsiopy && \
|
python3 -m venv /lsiopy && \
|
||||||
pip install --no-cache-dir -U pip && \
|
pip install --no-cache-dir -U pip && \
|
||||||
|
|
@ -297,7 +336,7 @@ pipeline {
|
||||||
echo "Jenkinsfile is up to date."
|
echo "Jenkinsfile is up to date."
|
||||||
fi
|
fi
|
||||||
echo "Starting Stage 2 - Delete old templates"
|
echo "Starting Stage 2 - Delete old templates"
|
||||||
OLD_TEMPLATES=".github/ISSUE_TEMPLATE.md .github/ISSUE_TEMPLATE/issue.bug.md .github/ISSUE_TEMPLATE/issue.feature.md .github/workflows/call_invalid_helper.yml .github/workflows/stale.yml"
|
OLD_TEMPLATES=".github/ISSUE_TEMPLATE.md .github/ISSUE_TEMPLATE/issue.bug.md .github/ISSUE_TEMPLATE/issue.feature.md .github/workflows/call_invalid_helper.yml .github/workflows/stale.yml .github/workflows/package_trigger.yml"
|
||||||
for i in ${OLD_TEMPLATES}; do
|
for i in ${OLD_TEMPLATES}; do
|
||||||
if [[ -f "${i}" ]]; then
|
if [[ -f "${i}" ]]; then
|
||||||
TEMPLATES_TO_DELETE="${i} ${TEMPLATES_TO_DELETE}"
|
TEMPLATES_TO_DELETE="${i} ${TEMPLATES_TO_DELETE}"
|
||||||
|
|
@ -321,6 +360,35 @@ pipeline {
|
||||||
else
|
else
|
||||||
echo "No templates to delete"
|
echo "No templates to delete"
|
||||||
fi
|
fi
|
||||||
|
echo "Starting Stage 2.5 - Update init diagram"
|
||||||
|
if ! grep -q 'init_diagram:' readme-vars.yml; then
|
||||||
|
echo "Adding the key 'init_diagram' to readme-vars.yml"
|
||||||
|
sed -i '\\|^#.*changelog.*$|d' readme-vars.yml
|
||||||
|
sed -i 's|^changelogs:|# init diagram\\ninit_diagram:\\n\\n# changelog\\nchangelogs:|' readme-vars.yml
|
||||||
|
fi
|
||||||
|
mkdir -p ${TEMPDIR}/d2
|
||||||
|
docker run --rm -v ${TEMPDIR}/d2:/output -e PUID=$(id -u) -e PGID=$(id -g) -e RAW="true" ghcr.io/linuxserver/d2-builder:latest ${CONTAINER_NAME}:latest
|
||||||
|
ls -al ${TEMPDIR}/d2
|
||||||
|
yq -ei ".init_diagram |= load_str(\\"${TEMPDIR}/d2/${CONTAINER_NAME}-latest.d2\\")" readme-vars.yml
|
||||||
|
if [[ $(md5sum readme-vars.yml | cut -c1-8) != $(md5sum ${TEMPDIR}/docker-${CONTAINER_NAME}/readme-vars.yml | cut -c1-8) ]]; then
|
||||||
|
echo "'init_diagram' has been updated. Updating repo and exiting build, new one will trigger based on commit."
|
||||||
|
mkdir -p ${TEMPDIR}/repo
|
||||||
|
git clone https://github.com/${LS_USER}/${LS_REPO}.git ${TEMPDIR}/repo/${LS_REPO}
|
||||||
|
cd ${TEMPDIR}/repo/${LS_REPO}
|
||||||
|
git checkout -f master
|
||||||
|
cp ${WORKSPACE}/readme-vars.yml ${TEMPDIR}/repo/${LS_REPO}/readme-vars.yml
|
||||||
|
git add readme-vars.yml
|
||||||
|
git commit -m 'Bot Updating Templated Files'
|
||||||
|
git pull https://LinuxServer-CI:${GITHUB_TOKEN}@github.com/${LS_USER}/${LS_REPO}.git master
|
||||||
|
git push https://LinuxServer-CI:${GITHUB_TOKEN}@github.com/${LS_USER}/${LS_REPO}.git master
|
||||||
|
echo "true" > /tmp/${COMMIT_SHA}-${BUILD_NUMBER}
|
||||||
|
echo "Updating templates and exiting build, new one will trigger based on commit"
|
||||||
|
rm -Rf ${TEMPDIR}
|
||||||
|
exit 0
|
||||||
|
else
|
||||||
|
echo "false" > /tmp/${COMMIT_SHA}-${BUILD_NUMBER}
|
||||||
|
echo "Init diagram is unchanged"
|
||||||
|
fi
|
||||||
echo "Starting Stage 3 - Update templates"
|
echo "Starting Stage 3 - Update templates"
|
||||||
CURRENTHASH=$(grep -hs ^ ${TEMPLATED_FILES} | md5sum | cut -c1-8)
|
CURRENTHASH=$(grep -hs ^ ${TEMPLATED_FILES} | md5sum | cut -c1-8)
|
||||||
cd ${TEMPDIR}/docker-${CONTAINER_NAME}
|
cd ${TEMPDIR}/docker-${CONTAINER_NAME}
|
||||||
|
|
@ -354,7 +422,7 @@ pipeline {
|
||||||
fi
|
fi
|
||||||
echo "Starting Stage 4 - External repo updates: Docs, Unraid Template and Readme Sync to Docker Hub"
|
echo "Starting Stage 4 - External repo updates: Docs, Unraid Template and Readme Sync to Docker Hub"
|
||||||
mkdir -p ${TEMPDIR}/docs
|
mkdir -p ${TEMPDIR}/docs
|
||||||
git clone https://github.com/linuxserver/docker-documentation.git ${TEMPDIR}/docs/docker-documentation
|
git clone --depth=1 https://github.com/linuxserver/docker-documentation.git ${TEMPDIR}/docs/docker-documentation
|
||||||
if [[ "${BRANCH_NAME}" == "${GH_DEFAULT_BRANCH}" ]] && [[ (! -f ${TEMPDIR}/docs/docker-documentation/docs/images/docker-${CONTAINER_NAME}.md) || ("$(md5sum ${TEMPDIR}/docs/docker-documentation/docs/images/docker-${CONTAINER_NAME}.md | awk '{ print $1 }')" != "$(md5sum ${TEMPDIR}/docker-${CONTAINER_NAME}/.jenkins-external/docker-${CONTAINER_NAME}.md | awk '{ print $1 }')") ]]; then
|
if [[ "${BRANCH_NAME}" == "${GH_DEFAULT_BRANCH}" ]] && [[ (! -f ${TEMPDIR}/docs/docker-documentation/docs/images/docker-${CONTAINER_NAME}.md) || ("$(md5sum ${TEMPDIR}/docs/docker-documentation/docs/images/docker-${CONTAINER_NAME}.md | awk '{ print $1 }')" != "$(md5sum ${TEMPDIR}/docker-${CONTAINER_NAME}/.jenkins-external/docker-${CONTAINER_NAME}.md | awk '{ print $1 }')") ]]; then
|
||||||
cp ${TEMPDIR}/docker-${CONTAINER_NAME}/.jenkins-external/docker-${CONTAINER_NAME}.md ${TEMPDIR}/docs/docker-documentation/docs/images/
|
cp ${TEMPDIR}/docker-${CONTAINER_NAME}/.jenkins-external/docker-${CONTAINER_NAME}.md ${TEMPDIR}/docs/docker-documentation/docs/images/
|
||||||
cd ${TEMPDIR}/docs/docker-documentation
|
cd ${TEMPDIR}/docs/docker-documentation
|
||||||
|
|
@ -372,8 +440,8 @@ pipeline {
|
||||||
echo "Docs update not needed, skipping"
|
echo "Docs update not needed, skipping"
|
||||||
fi
|
fi
|
||||||
mkdir -p ${TEMPDIR}/unraid
|
mkdir -p ${TEMPDIR}/unraid
|
||||||
git clone https://github.com/linuxserver/docker-templates.git ${TEMPDIR}/unraid/docker-templates
|
git clone --depth=1 https://github.com/linuxserver/docker-templates.git ${TEMPDIR}/unraid/docker-templates
|
||||||
git clone https://github.com/linuxserver/templates.git ${TEMPDIR}/unraid/templates
|
git clone --depth=1 https://github.com/linuxserver/templates.git ${TEMPDIR}/unraid/templates
|
||||||
if [[ -f ${TEMPDIR}/unraid/docker-templates/linuxserver.io/img/${CONTAINER_NAME}-logo.png ]]; then
|
if [[ -f ${TEMPDIR}/unraid/docker-templates/linuxserver.io/img/${CONTAINER_NAME}-logo.png ]]; then
|
||||||
sed -i "s|master/linuxserver.io/img/linuxserver-ls-logo.png|master/linuxserver.io/img/${CONTAINER_NAME}-logo.png|" ${TEMPDIR}/docker-${CONTAINER_NAME}/.jenkins-external/${CONTAINER_NAME}.xml
|
sed -i "s|master/linuxserver.io/img/linuxserver-ls-logo.png|master/linuxserver.io/img/${CONTAINER_NAME}-logo.png|" ${TEMPDIR}/docker-${CONTAINER_NAME}/.jenkins-external/${CONTAINER_NAME}.xml
|
||||||
elif [[ -f ${TEMPDIR}/unraid/docker-templates/linuxserver.io/img/${CONTAINER_NAME}-icon.png ]]; then
|
elif [[ -f ${TEMPDIR}/unraid/docker-templates/linuxserver.io/img/${CONTAINER_NAME}-icon.png ]]; then
|
||||||
|
|
@ -383,9 +451,9 @@ pipeline {
|
||||||
echo "Updating Unraid template"
|
echo "Updating Unraid template"
|
||||||
cd ${TEMPDIR}/unraid/templates/
|
cd ${TEMPDIR}/unraid/templates/
|
||||||
GH_TEMPLATES_DEFAULT_BRANCH=$(git remote show origin | grep "HEAD branch:" | sed 's|.*HEAD branch: ||')
|
GH_TEMPLATES_DEFAULT_BRANCH=$(git remote show origin | grep "HEAD branch:" | sed 's|.*HEAD branch: ||')
|
||||||
if grep -wq "${CONTAINER_NAME}" ${TEMPDIR}/unraid/templates/unraid/ignore.list && [[ -f ${TEMPDIR}/unraid/templates/unraid/deprecated/${CONTAINER_NAME}.xml ]]; then
|
if grep -wq "^${CONTAINER_NAME}$" ${TEMPDIR}/unraid/templates/unraid/ignore.list && [[ -f ${TEMPDIR}/unraid/templates/unraid/deprecated/${CONTAINER_NAME}.xml ]]; then
|
||||||
echo "Image is on the ignore list, and already in the deprecation folder."
|
echo "Image is on the ignore list, and already in the deprecation folder."
|
||||||
elif grep -wq "${CONTAINER_NAME}" ${TEMPDIR}/unraid/templates/unraid/ignore.list; then
|
elif grep -wq "^${CONTAINER_NAME}$" ${TEMPDIR}/unraid/templates/unraid/ignore.list; then
|
||||||
echo "Image is on the ignore list, marking Unraid template as deprecated"
|
echo "Image is on the ignore list, marking Unraid template as deprecated"
|
||||||
cp ${TEMPDIR}/docker-${CONTAINER_NAME}/.jenkins-external/${CONTAINER_NAME}.xml ${TEMPDIR}/unraid/templates/unraid/
|
cp ${TEMPDIR}/docker-${CONTAINER_NAME}/.jenkins-external/${CONTAINER_NAME}.xml ${TEMPDIR}/unraid/templates/unraid/
|
||||||
git add -u unraid/${CONTAINER_NAME}.xml
|
git add -u unraid/${CONTAINER_NAME}.xml
|
||||||
|
|
@ -478,10 +546,10 @@ pipeline {
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
/* #######################
|
/* #######################
|
||||||
GitLab Mirroring
|
GitLab Mirroring and Quay.io Repo Visibility
|
||||||
####################### */
|
####################### */
|
||||||
// Ping into Gitlab to mirror this repo and have a registry endpoint
|
// Ping into Gitlab to mirror this repo and have a registry endpoint & mark this repo on Quay.io as public
|
||||||
stage("GitLab Mirror"){
|
stage("GitLab Mirror and Quay.io Visibility"){
|
||||||
when {
|
when {
|
||||||
environment name: 'EXIT_STATUS', value: ''
|
environment name: 'EXIT_STATUS', value: ''
|
||||||
}
|
}
|
||||||
|
|
@ -497,6 +565,8 @@ pipeline {
|
||||||
"visibility":"public"}' '''
|
"visibility":"public"}' '''
|
||||||
sh '''curl -H "Private-Token: ${GITLAB_TOKEN}" -X PUT "https://gitlab.com/api/v4/projects/Linuxserver.io%2F${LS_REPO}" \
|
sh '''curl -H "Private-Token: ${GITLAB_TOKEN}" -X PUT "https://gitlab.com/api/v4/projects/Linuxserver.io%2F${LS_REPO}" \
|
||||||
-d "mirror=true&import_url=https://github.com/linuxserver/${LS_REPO}.git" '''
|
-d "mirror=true&import_url=https://github.com/linuxserver/${LS_REPO}.git" '''
|
||||||
|
sh '''curl -H "Content-Type: application/json" -H "Authorization: Bearer ${QUAYIO_API_TOKEN}" -X POST "https://quay.io/api/v1/repository${QUAYIMAGE/quay.io/}/changevisibility" \
|
||||||
|
-d '{"visibility":"public"}' ||: '''
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
/* ###############
|
/* ###############
|
||||||
|
|
@ -527,8 +597,45 @@ pipeline {
|
||||||
--label \"org.opencontainers.image.title=Plex\" \
|
--label \"org.opencontainers.image.title=Plex\" \
|
||||||
--label \"org.opencontainers.image.description=[Plex](https://plex.tv) organizes video, music and photos from personal media libraries and streams them to smart TVs, streaming boxes and mobile devices. This container is packaged as a standalone Plex Media Server. Straightforward design and bulk actions mean getting things done faster.\" \
|
--label \"org.opencontainers.image.description=[Plex](https://plex.tv) organizes video, music and photos from personal media libraries and streams them to smart TVs, streaming boxes and mobile devices. This container is packaged as a standalone Plex Media Server. Straightforward design and bulk actions mean getting things done faster.\" \
|
||||||
--no-cache --pull -t ${IMAGE}:${META_TAG} --platform=linux/amd64 \
|
--no-cache --pull -t ${IMAGE}:${META_TAG} --platform=linux/amd64 \
|
||||||
--provenance=false --sbom=false \
|
--provenance=true --sbom=true --builder=container --load \
|
||||||
--build-arg ${BUILD_VERSION_ARG}=${EXT_RELEASE} --build-arg VERSION=\"${VERSION_TAG}\" --build-arg BUILD_DATE=${GITHUB_DATE} ."
|
--build-arg ${BUILD_VERSION_ARG}=${EXT_RELEASE} --build-arg VERSION=\"${VERSION_TAG}\" --build-arg BUILD_DATE=${GITHUB_DATE} ."
|
||||||
|
sh '''#! /bin/bash
|
||||||
|
set -e
|
||||||
|
IFS=',' read -ra CACHE <<< "$BUILDCACHE"
|
||||||
|
for i in "${CACHE[@]}"; do
|
||||||
|
docker tag ${IMAGE}:${META_TAG} ${i}:amd64-${COMMIT_SHA}-${BUILD_NUMBER}
|
||||||
|
done
|
||||||
|
'''
|
||||||
|
withCredentials([
|
||||||
|
[
|
||||||
|
$class: 'UsernamePasswordMultiBinding',
|
||||||
|
credentialsId: 'Quay.io-Robot',
|
||||||
|
usernameVariable: 'QUAYUSER',
|
||||||
|
passwordVariable: 'QUAYPASS'
|
||||||
|
]
|
||||||
|
]) {
|
||||||
|
retry_backoff(5,5) {
|
||||||
|
sh '''#! /bin/bash
|
||||||
|
set -e
|
||||||
|
echo $DOCKERHUB_TOKEN | docker login -u linuxserverci --password-stdin
|
||||||
|
echo $GITHUB_TOKEN | docker login ghcr.io -u LinuxServer-CI --password-stdin
|
||||||
|
echo $GITLAB_TOKEN | docker login registry.gitlab.com -u LinuxServer.io --password-stdin
|
||||||
|
echo $QUAYPASS | docker login quay.io -u $QUAYUSER --password-stdin
|
||||||
|
|
||||||
|
if [[ "${PACKAGE_CHECK}" != "true" ]]; then
|
||||||
|
declare -A pids
|
||||||
|
IFS=',' read -ra CACHE <<< "$BUILDCACHE"
|
||||||
|
for i in "${CACHE[@]}"; do
|
||||||
|
docker push ${i}:amd64-${COMMIT_SHA}-${BUILD_NUMBER} &
|
||||||
|
pids[$!]="$i"
|
||||||
|
done
|
||||||
|
for p in "${!pids[@]}"; do
|
||||||
|
wait "$p" || { [[ "${pids[$p]}" != *"quay.io"* ]] && exit 1; }
|
||||||
|
done
|
||||||
|
fi
|
||||||
|
'''
|
||||||
|
}
|
||||||
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
// Build MultiArch Docker containers for push to LS Repo
|
// Build MultiArch Docker containers for push to LS Repo
|
||||||
|
|
@ -559,8 +666,45 @@ pipeline {
|
||||||
--label \"org.opencontainers.image.title=Plex\" \
|
--label \"org.opencontainers.image.title=Plex\" \
|
||||||
--label \"org.opencontainers.image.description=[Plex](https://plex.tv) organizes video, music and photos from personal media libraries and streams them to smart TVs, streaming boxes and mobile devices. This container is packaged as a standalone Plex Media Server. Straightforward design and bulk actions mean getting things done faster.\" \
|
--label \"org.opencontainers.image.description=[Plex](https://plex.tv) organizes video, music and photos from personal media libraries and streams them to smart TVs, streaming boxes and mobile devices. This container is packaged as a standalone Plex Media Server. Straightforward design and bulk actions mean getting things done faster.\" \
|
||||||
--no-cache --pull -t ${IMAGE}:amd64-${META_TAG} --platform=linux/amd64 \
|
--no-cache --pull -t ${IMAGE}:amd64-${META_TAG} --platform=linux/amd64 \
|
||||||
--provenance=false --sbom=false \
|
--provenance=true --sbom=true --builder=container --load \
|
||||||
--build-arg ${BUILD_VERSION_ARG}=${EXT_RELEASE} --build-arg VERSION=\"${VERSION_TAG}\" --build-arg BUILD_DATE=${GITHUB_DATE} ."
|
--build-arg ${BUILD_VERSION_ARG}=${EXT_RELEASE} --build-arg VERSION=\"${VERSION_TAG}\" --build-arg BUILD_DATE=${GITHUB_DATE} ."
|
||||||
|
sh '''#! /bin/bash
|
||||||
|
set -e
|
||||||
|
IFS=',' read -ra CACHE <<< "$BUILDCACHE"
|
||||||
|
for i in "${CACHE[@]}"; do
|
||||||
|
docker tag ${IMAGE}:amd64-${META_TAG} ${i}:amd64-${COMMIT_SHA}-${BUILD_NUMBER}
|
||||||
|
done
|
||||||
|
'''
|
||||||
|
withCredentials([
|
||||||
|
[
|
||||||
|
$class: 'UsernamePasswordMultiBinding',
|
||||||
|
credentialsId: 'Quay.io-Robot',
|
||||||
|
usernameVariable: 'QUAYUSER',
|
||||||
|
passwordVariable: 'QUAYPASS'
|
||||||
|
]
|
||||||
|
]) {
|
||||||
|
retry_backoff(5,5) {
|
||||||
|
sh '''#! /bin/bash
|
||||||
|
set -e
|
||||||
|
echo $DOCKERHUB_TOKEN | docker login -u linuxserverci --password-stdin
|
||||||
|
echo $GITHUB_TOKEN | docker login ghcr.io -u LinuxServer-CI --password-stdin
|
||||||
|
echo $GITLAB_TOKEN | docker login registry.gitlab.com -u LinuxServer.io --password-stdin
|
||||||
|
echo $QUAYPASS | docker login quay.io -u $QUAYUSER --password-stdin
|
||||||
|
|
||||||
|
if [[ "${PACKAGE_CHECK}" != "true" ]]; then
|
||||||
|
declare -A pids
|
||||||
|
IFS=',' read -ra CACHE <<< "$BUILDCACHE"
|
||||||
|
for i in "${CACHE[@]}"; do
|
||||||
|
docker push ${i}:amd64-${COMMIT_SHA}-${BUILD_NUMBER} &
|
||||||
|
pids[$!]="$i"
|
||||||
|
done
|
||||||
|
for p in "${!pids[@]}"; do
|
||||||
|
wait "$p" || { [[ "${pids[$p]}" != *"quay.io"* ]] && exit 1; }
|
||||||
|
done
|
||||||
|
fi
|
||||||
|
'''
|
||||||
|
}
|
||||||
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
stage('Build ARM64') {
|
stage('Build ARM64') {
|
||||||
|
|
@ -569,10 +713,6 @@ pipeline {
|
||||||
}
|
}
|
||||||
steps {
|
steps {
|
||||||
echo "Running on node: ${NODE_NAME}"
|
echo "Running on node: ${NODE_NAME}"
|
||||||
echo 'Logging into Github'
|
|
||||||
sh '''#! /bin/bash
|
|
||||||
echo $GITHUB_TOKEN | docker login ghcr.io -u LinuxServer-CI --password-stdin
|
|
||||||
'''
|
|
||||||
sh "sed -r -i 's|(^FROM .*)|\\1\\n\\nENV LSIO_FIRST_PARTY=true|g' Dockerfile.aarch64"
|
sh "sed -r -i 's|(^FROM .*)|\\1\\n\\nENV LSIO_FIRST_PARTY=true|g' Dockerfile.aarch64"
|
||||||
sh "docker buildx build \
|
sh "docker buildx build \
|
||||||
--label \"org.opencontainers.image.created=${GITHUB_DATE}\" \
|
--label \"org.opencontainers.image.created=${GITHUB_DATE}\" \
|
||||||
|
|
@ -588,18 +728,52 @@ pipeline {
|
||||||
--label \"org.opencontainers.image.title=Plex\" \
|
--label \"org.opencontainers.image.title=Plex\" \
|
||||||
--label \"org.opencontainers.image.description=[Plex](https://plex.tv) organizes video, music and photos from personal media libraries and streams them to smart TVs, streaming boxes and mobile devices. This container is packaged as a standalone Plex Media Server. Straightforward design and bulk actions mean getting things done faster.\" \
|
--label \"org.opencontainers.image.description=[Plex](https://plex.tv) organizes video, music and photos from personal media libraries and streams them to smart TVs, streaming boxes and mobile devices. This container is packaged as a standalone Plex Media Server. Straightforward design and bulk actions mean getting things done faster.\" \
|
||||||
--no-cache --pull -f Dockerfile.aarch64 -t ${IMAGE}:arm64v8-${META_TAG} --platform=linux/arm64 \
|
--no-cache --pull -f Dockerfile.aarch64 -t ${IMAGE}:arm64v8-${META_TAG} --platform=linux/arm64 \
|
||||||
--provenance=false --sbom=false \
|
--provenance=true --sbom=true --builder=container --load \
|
||||||
--build-arg ${BUILD_VERSION_ARG}=${EXT_RELEASE} --build-arg VERSION=\"${VERSION_TAG}\" --build-arg BUILD_DATE=${GITHUB_DATE} ."
|
--build-arg ${BUILD_VERSION_ARG}=${EXT_RELEASE} --build-arg VERSION=\"${VERSION_TAG}\" --build-arg BUILD_DATE=${GITHUB_DATE} ."
|
||||||
sh "docker tag ${IMAGE}:arm64v8-${META_TAG} ghcr.io/linuxserver/lsiodev-buildcache:arm64v8-${COMMIT_SHA}-${BUILD_NUMBER}"
|
sh '''#! /bin/bash
|
||||||
retry(5) {
|
set -e
|
||||||
sh "docker push ghcr.io/linuxserver/lsiodev-buildcache:arm64v8-${COMMIT_SHA}-${BUILD_NUMBER}"
|
IFS=',' read -ra CACHE <<< "$BUILDCACHE"
|
||||||
|
for i in "${CACHE[@]}"; do
|
||||||
|
docker tag ${IMAGE}:arm64v8-${META_TAG} ${i}:arm64v8-${COMMIT_SHA}-${BUILD_NUMBER}
|
||||||
|
done
|
||||||
|
'''
|
||||||
|
withCredentials([
|
||||||
|
[
|
||||||
|
$class: 'UsernamePasswordMultiBinding',
|
||||||
|
credentialsId: 'Quay.io-Robot',
|
||||||
|
usernameVariable: 'QUAYUSER',
|
||||||
|
passwordVariable: 'QUAYPASS'
|
||||||
|
]
|
||||||
|
]) {
|
||||||
|
retry_backoff(5,5) {
|
||||||
|
sh '''#! /bin/bash
|
||||||
|
set -e
|
||||||
|
echo $DOCKERHUB_TOKEN | docker login -u linuxserverci --password-stdin
|
||||||
|
echo $GITHUB_TOKEN | docker login ghcr.io -u LinuxServer-CI --password-stdin
|
||||||
|
echo $GITLAB_TOKEN | docker login registry.gitlab.com -u LinuxServer.io --password-stdin
|
||||||
|
echo $QUAYPASS | docker login quay.io -u $QUAYUSER --password-stdin
|
||||||
|
if [[ "${PACKAGE_CHECK}" != "true" ]]; then
|
||||||
|
declare -A pids
|
||||||
|
IFS=',' read -ra CACHE <<< "$BUILDCACHE"
|
||||||
|
for i in "${CACHE[@]}"; do
|
||||||
|
docker push ${i}:arm64v8-${COMMIT_SHA}-${BUILD_NUMBER} &
|
||||||
|
pids[$!]="$i"
|
||||||
|
done
|
||||||
|
for p in "${!pids[@]}"; do
|
||||||
|
wait "$p" || { [[ "${pids[$p]}" != *"quay.io"* ]] && exit 1; }
|
||||||
|
done
|
||||||
|
fi
|
||||||
|
'''
|
||||||
|
}
|
||||||
}
|
}
|
||||||
sh '''#! /bin/bash
|
sh '''#! /bin/bash
|
||||||
containers=$(docker ps -aq)
|
containers=$(docker ps -aq)
|
||||||
if [[ -n "${containers}" ]]; then
|
if [[ -n "${containers}" ]]; then
|
||||||
docker stop ${containers}
|
docker stop ${containers}
|
||||||
fi
|
fi
|
||||||
docker system prune -af --volumes || : '''
|
docker system prune -f --volumes || :
|
||||||
|
docker image prune -af || :
|
||||||
|
'''
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
@ -624,7 +798,7 @@ pipeline {
|
||||||
docker run --rm \
|
docker run --rm \
|
||||||
-v /var/run/docker.sock:/var/run/docker.sock:ro \
|
-v /var/run/docker.sock:/var/run/docker.sock:ro \
|
||||||
-v ${TEMPDIR}:/tmp \
|
-v ${TEMPDIR}:/tmp \
|
||||||
ghcr.io/anchore/syft:latest \
|
ghcr.io/anchore/syft:${SYFT_IMAGE_TAG} \
|
||||||
${LOCAL_CONTAINER} -o table=/tmp/package_versions.txt
|
${LOCAL_CONTAINER} -o table=/tmp/package_versions.txt
|
||||||
NEW_PACKAGE_TAG=$(md5sum ${TEMPDIR}/package_versions.txt | cut -c1-8 )
|
NEW_PACKAGE_TAG=$(md5sum ${TEMPDIR}/package_versions.txt | cut -c1-8 )
|
||||||
echo "Package tag sha from current packages in buit container is ${NEW_PACKAGE_TAG} comparing to old ${PACKAGE_TAG} from github"
|
echo "Package tag sha from current packages in buit container is ${NEW_PACKAGE_TAG} comparing to old ${PACKAGE_TAG} from github"
|
||||||
|
|
@ -700,12 +874,21 @@ pipeline {
|
||||||
script{
|
script{
|
||||||
env.CI_URL = 'https://ci-tests.linuxserver.io/' + env.IMAGE + '/' + env.META_TAG + '/index.html'
|
env.CI_URL = 'https://ci-tests.linuxserver.io/' + env.IMAGE + '/' + env.META_TAG + '/index.html'
|
||||||
env.CI_JSON_URL = 'https://ci-tests.linuxserver.io/' + env.IMAGE + '/' + env.META_TAG + '/report.json'
|
env.CI_JSON_URL = 'https://ci-tests.linuxserver.io/' + env.IMAGE + '/' + env.META_TAG + '/report.json'
|
||||||
|
env.CI_TEST_ATTEMPTED = 'true'
|
||||||
}
|
}
|
||||||
sh '''#! /bin/bash
|
sh '''#! /bin/bash
|
||||||
set -e
|
set -e
|
||||||
docker pull ghcr.io/linuxserver/ci:latest
|
if grep -q 'docker-baseimage' <<< "${LS_REPO}"; then
|
||||||
|
echo "Detected baseimage, setting LSIO_FIRST_PARTY=true"
|
||||||
|
if [ -n "${CI_DOCKERENV}" ]; then
|
||||||
|
CI_DOCKERENV="LSIO_FIRST_PARTY=true|${CI_DOCKERENV}"
|
||||||
|
else
|
||||||
|
CI_DOCKERENV="LSIO_FIRST_PARTY=true"
|
||||||
|
fi
|
||||||
|
fi
|
||||||
|
docker pull ghcr.io/linuxserver/ci:${CITEST_IMAGETAG}
|
||||||
if [ "${MULTIARCH}" == "true" ]; then
|
if [ "${MULTIARCH}" == "true" ]; then
|
||||||
docker pull ghcr.io/linuxserver/lsiodev-buildcache:arm64v8-${COMMIT_SHA}-${BUILD_NUMBER}
|
docker pull ghcr.io/linuxserver/lsiodev-buildcache:arm64v8-${COMMIT_SHA}-${BUILD_NUMBER} --platform=arm64
|
||||||
docker tag ghcr.io/linuxserver/lsiodev-buildcache:arm64v8-${COMMIT_SHA}-${BUILD_NUMBER} ${IMAGE}:arm64v8-${META_TAG}
|
docker tag ghcr.io/linuxserver/lsiodev-buildcache:arm64v8-${COMMIT_SHA}-${BUILD_NUMBER} ${IMAGE}:arm64v8-${META_TAG}
|
||||||
fi
|
fi
|
||||||
docker run --rm \
|
docker run --rm \
|
||||||
|
|
@ -715,6 +898,7 @@ pipeline {
|
||||||
-e DOCKER_LOGS_TIMEOUT=\"${CI_DELAY}\" \
|
-e DOCKER_LOGS_TIMEOUT=\"${CI_DELAY}\" \
|
||||||
-e TAGS=\"${CI_TAGS}\" \
|
-e TAGS=\"${CI_TAGS}\" \
|
||||||
-e META_TAG=\"${META_TAG}\" \
|
-e META_TAG=\"${META_TAG}\" \
|
||||||
|
-e RELEASE_TAG=\"latest\" \
|
||||||
-e PORT=\"${CI_PORT}\" \
|
-e PORT=\"${CI_PORT}\" \
|
||||||
-e SSL=\"${CI_SSL}\" \
|
-e SSL=\"${CI_SSL}\" \
|
||||||
-e BASE=\"${DIST_IMAGE}\" \
|
-e BASE=\"${DIST_IMAGE}\" \
|
||||||
|
|
@ -724,7 +908,11 @@ pipeline {
|
||||||
-e WEB_SCREENSHOT=\"${CI_WEB}\" \
|
-e WEB_SCREENSHOT=\"${CI_WEB}\" \
|
||||||
-e WEB_AUTH=\"${CI_AUTH}\" \
|
-e WEB_AUTH=\"${CI_AUTH}\" \
|
||||||
-e WEB_PATH=\"${CI_WEBPATH}\" \
|
-e WEB_PATH=\"${CI_WEBPATH}\" \
|
||||||
-t ghcr.io/linuxserver/ci:latest \
|
-e NODE_NAME=\"${NODE_NAME}\" \
|
||||||
|
-e SYFT_IMAGE_TAG=\"${CI_SYFT_IMAGE_TAG:-${SYFT_IMAGE_TAG}}\" \
|
||||||
|
-e COMMIT_SHA=\"${COMMIT_SHA}\" \
|
||||||
|
-e BUILD_NUMBER=\"${BUILD_NUMBER}\" \
|
||||||
|
-t ghcr.io/linuxserver/ci:${CITEST_IMAGETAG} \
|
||||||
python3 test_build.py'''
|
python3 test_build.py'''
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
@ -739,37 +927,25 @@ pipeline {
|
||||||
environment name: 'EXIT_STATUS', value: ''
|
environment name: 'EXIT_STATUS', value: ''
|
||||||
}
|
}
|
||||||
steps {
|
steps {
|
||||||
withCredentials([
|
retry_backoff(5,5) {
|
||||||
[
|
sh '''#! /bin/bash
|
||||||
$class: 'UsernamePasswordMultiBinding',
|
set -e
|
||||||
credentialsId: 'Quay.io-Robot',
|
for PUSHIMAGE in "${IMAGE}" "${GITLABIMAGE}" "${GITHUBIMAGE}" "${QUAYIMAGE}"; do
|
||||||
usernameVariable: 'QUAYUSER',
|
[[ ${PUSHIMAGE%%/*} =~ \\. ]] && PUSHIMAGEPLUS="${PUSHIMAGE}" || PUSHIMAGEPLUS="docker.io/${PUSHIMAGE}"
|
||||||
passwordVariable: 'QUAYPASS'
|
IFS=',' read -ra CACHE <<< "$BUILDCACHE"
|
||||||
]
|
for i in "${CACHE[@]}"; do
|
||||||
]) {
|
if [[ "${PUSHIMAGEPLUS}" == "$(cut -d "/" -f1 <<< ${i})"* ]]; then
|
||||||
retry(5) {
|
CACHEIMAGE=${i}
|
||||||
sh '''#! /bin/bash
|
fi
|
||||||
set -e
|
|
||||||
echo $DOCKERHUB_TOKEN | docker login -u linuxserverci --password-stdin
|
|
||||||
echo $GITHUB_TOKEN | docker login ghcr.io -u LinuxServer-CI --password-stdin
|
|
||||||
echo $GITLAB_TOKEN | docker login registry.gitlab.com -u LinuxServer.io --password-stdin
|
|
||||||
echo $QUAYPASS | docker login quay.io -u $QUAYUSER --password-stdin
|
|
||||||
for PUSHIMAGE in "${GITHUBIMAGE}" "${GITLABIMAGE}" "${QUAYIMAGE}" "${IMAGE}"; do
|
|
||||||
docker tag ${IMAGE}:${META_TAG} ${PUSHIMAGE}:${META_TAG}
|
|
||||||
docker tag ${PUSHIMAGE}:${META_TAG} ${PUSHIMAGE}:latest
|
|
||||||
docker tag ${PUSHIMAGE}:${META_TAG} ${PUSHIMAGE}:${EXT_RELEASE_TAG}
|
|
||||||
if [ -n "${SEMVER}" ]; then
|
|
||||||
docker tag ${PUSHIMAGE}:${META_TAG} ${PUSHIMAGE}:${SEMVER}
|
|
||||||
fi
|
|
||||||
docker push ${PUSHIMAGE}:latest
|
|
||||||
docker push ${PUSHIMAGE}:${META_TAG}
|
|
||||||
docker push ${PUSHIMAGE}:${EXT_RELEASE_TAG}
|
|
||||||
if [ -n "${SEMVER}" ]; then
|
|
||||||
docker push ${PUSHIMAGE}:${SEMVER}
|
|
||||||
fi
|
|
||||||
done
|
done
|
||||||
'''
|
docker buildx imagetools create --prefer-index=false -t ${PUSHIMAGE}:${META_TAG} -t ${PUSHIMAGE}:latest -t ${PUSHIMAGE}:${EXT_RELEASE_TAG} ${CACHEIMAGE}:amd64-${COMMIT_SHA}-${BUILD_NUMBER} || \
|
||||||
}
|
{ if [[ "${PUSHIMAGE}" != "${QUAYIMAGE}" ]]; then exit 1; fi; }
|
||||||
|
if [ -n "${SEMVER}" ]; then
|
||||||
|
docker buildx imagetools create --prefer-index=false -t ${PUSHIMAGE}:${SEMVER} ${CACHEIMAGE}:amd64-${COMMIT_SHA}-${BUILD_NUMBER} || \
|
||||||
|
{ if [[ "${PUSHIMAGE}" != "${QUAYIMAGE}" ]]; then exit 1; fi; }
|
||||||
|
fi
|
||||||
|
done
|
||||||
|
'''
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
@ -780,57 +956,41 @@ pipeline {
|
||||||
environment name: 'EXIT_STATUS', value: ''
|
environment name: 'EXIT_STATUS', value: ''
|
||||||
}
|
}
|
||||||
steps {
|
steps {
|
||||||
withCredentials([
|
retry_backoff(5,5) {
|
||||||
[
|
sh '''#! /bin/bash
|
||||||
$class: 'UsernamePasswordMultiBinding',
|
set -e
|
||||||
credentialsId: 'Quay.io-Robot',
|
for MANIFESTIMAGE in "${IMAGE}" "${GITLABIMAGE}" "${GITHUBIMAGE}" "${QUAYIMAGE}"; do
|
||||||
usernameVariable: 'QUAYUSER',
|
[[ ${MANIFESTIMAGE%%/*} =~ \\. ]] && MANIFESTIMAGEPLUS="${MANIFESTIMAGE}" || MANIFESTIMAGEPLUS="docker.io/${MANIFESTIMAGE}"
|
||||||
passwordVariable: 'QUAYPASS'
|
IFS=',' read -ra CACHE <<< "$BUILDCACHE"
|
||||||
]
|
for i in "${CACHE[@]}"; do
|
||||||
]) {
|
if [[ "${MANIFESTIMAGEPLUS}" == "$(cut -d "/" -f1 <<< ${i})"* ]]; then
|
||||||
retry(5) {
|
CACHEIMAGE=${i}
|
||||||
sh '''#! /bin/bash
|
fi
|
||||||
set -e
|
done
|
||||||
echo $DOCKERHUB_TOKEN | docker login -u linuxserverci --password-stdin
|
docker buildx imagetools create --prefer-index=false -t ${MANIFESTIMAGE}:amd64-${META_TAG} -t ${MANIFESTIMAGE}:amd64-latest -t ${MANIFESTIMAGE}:amd64-${EXT_RELEASE_TAG} ${CACHEIMAGE}:amd64-${COMMIT_SHA}-${BUILD_NUMBER} || \
|
||||||
echo $GITHUB_TOKEN | docker login ghcr.io -u LinuxServer-CI --password-stdin
|
{ if [[ "${MANIFESTIMAGE}" != "${QUAYIMAGE}" ]]; then exit 1; fi; }
|
||||||
echo $GITLAB_TOKEN | docker login registry.gitlab.com -u LinuxServer.io --password-stdin
|
docker buildx imagetools create --prefer-index=false -t ${MANIFESTIMAGE}:arm64v8-${META_TAG} -t ${MANIFESTIMAGE}:arm64v8-latest -t ${MANIFESTIMAGE}:arm64v8-${EXT_RELEASE_TAG} ${CACHEIMAGE}:arm64v8-${COMMIT_SHA}-${BUILD_NUMBER} || \
|
||||||
echo $QUAYPASS | docker login quay.io -u $QUAYUSER --password-stdin
|
{ if [[ "${MANIFESTIMAGE}" != "${QUAYIMAGE}" ]]; then exit 1; fi; }
|
||||||
if [ "${CI}" == "false" ]; then
|
if [ -n "${SEMVER}" ]; then
|
||||||
docker pull ghcr.io/linuxserver/lsiodev-buildcache:arm64v8-${COMMIT_SHA}-${BUILD_NUMBER}
|
docker buildx imagetools create --prefer-index=false -t ${MANIFESTIMAGE}:amd64-${SEMVER} ${CACHEIMAGE}:amd64-${COMMIT_SHA}-${BUILD_NUMBER} || \
|
||||||
docker tag ghcr.io/linuxserver/lsiodev-buildcache:arm64v8-${COMMIT_SHA}-${BUILD_NUMBER} ${IMAGE}:arm64v8-${META_TAG}
|
{ if [[ "${MANIFESTIMAGE}" != "${QUAYIMAGE}" ]]; then exit 1; fi; }
|
||||||
|
docker buildx imagetools create --prefer-index=false -t ${MANIFESTIMAGE}:arm64v8-${SEMVER} ${CACHEIMAGE}:arm64v8-${COMMIT_SHA}-${BUILD_NUMBER} || \
|
||||||
|
{ if [[ "${MANIFESTIMAGE}" != "${QUAYIMAGE}" ]]; then exit 1; fi; }
|
||||||
fi
|
fi
|
||||||
for MANIFESTIMAGE in "${IMAGE}" "${GITLABIMAGE}" "${GITHUBIMAGE}" "${QUAYIMAGE}"; do
|
done
|
||||||
docker tag ${IMAGE}:amd64-${META_TAG} ${MANIFESTIMAGE}:amd64-${META_TAG}
|
for MANIFESTIMAGE in "${IMAGE}" "${GITLABIMAGE}" "${GITHUBIMAGE}" "${QUAYIMAGE}"; do
|
||||||
docker tag ${MANIFESTIMAGE}:amd64-${META_TAG} ${MANIFESTIMAGE}:amd64-latest
|
docker buildx imagetools create -t ${MANIFESTIMAGE}:latest ${MANIFESTIMAGE}:amd64-latest ${MANIFESTIMAGE}:arm64v8-latest || \
|
||||||
docker tag ${MANIFESTIMAGE}:amd64-${META_TAG} ${MANIFESTIMAGE}:amd64-${EXT_RELEASE_TAG}
|
{ if [[ "${MANIFESTIMAGE}" != "${QUAYIMAGE}" ]]; then exit 1; fi; }
|
||||||
docker tag ${IMAGE}:arm64v8-${META_TAG} ${MANIFESTIMAGE}:arm64v8-${META_TAG}
|
docker buildx imagetools create -t ${MANIFESTIMAGE}:${META_TAG} ${MANIFESTIMAGE}:amd64-${META_TAG} ${MANIFESTIMAGE}:arm64v8-${META_TAG} || \
|
||||||
docker tag ${MANIFESTIMAGE}:arm64v8-${META_TAG} ${MANIFESTIMAGE}:arm64v8-latest
|
{ if [[ "${MANIFESTIMAGE}" != "${QUAYIMAGE}" ]]; then exit 1; fi; }
|
||||||
docker tag ${MANIFESTIMAGE}:arm64v8-${META_TAG} ${MANIFESTIMAGE}:arm64v8-${EXT_RELEASE_TAG}
|
docker buildx imagetools create -t ${MANIFESTIMAGE}:${EXT_RELEASE_TAG} ${MANIFESTIMAGE}:amd64-${EXT_RELEASE_TAG} ${MANIFESTIMAGE}:arm64v8-${EXT_RELEASE_TAG} || \
|
||||||
if [ -n "${SEMVER}" ]; then
|
{ if [[ "${MANIFESTIMAGE}" != "${QUAYIMAGE}" ]]; then exit 1; fi; }
|
||||||
docker tag ${MANIFESTIMAGE}:amd64-${META_TAG} ${MANIFESTIMAGE}:amd64-${SEMVER}
|
if [ -n "${SEMVER}" ]; then
|
||||||
docker tag ${MANIFESTIMAGE}:arm64v8-${META_TAG} ${MANIFESTIMAGE}:arm64v8-${SEMVER}
|
docker buildx imagetools create -t ${MANIFESTIMAGE}:${SEMVER} ${MANIFESTIMAGE}:amd64-${SEMVER} ${MANIFESTIMAGE}:arm64v8-${SEMVER} || \
|
||||||
fi
|
{ if [[ "${MANIFESTIMAGE}" != "${QUAYIMAGE}" ]]; then exit 1; fi; }
|
||||||
docker push ${MANIFESTIMAGE}:amd64-${META_TAG}
|
fi
|
||||||
docker push ${MANIFESTIMAGE}:amd64-${EXT_RELEASE_TAG}
|
done
|
||||||
docker push ${MANIFESTIMAGE}:amd64-latest
|
'''
|
||||||
docker push ${MANIFESTIMAGE}:arm64v8-${META_TAG}
|
|
||||||
docker push ${MANIFESTIMAGE}:arm64v8-latest
|
|
||||||
docker push ${MANIFESTIMAGE}:arm64v8-${EXT_RELEASE_TAG}
|
|
||||||
if [ -n "${SEMVER}" ]; then
|
|
||||||
docker push ${MANIFESTIMAGE}:amd64-${SEMVER}
|
|
||||||
docker push ${MANIFESTIMAGE}:arm64v8-${SEMVER}
|
|
||||||
fi
|
|
||||||
done
|
|
||||||
for MANIFESTIMAGE in "${IMAGE}" "${GITLABIMAGE}" "${GITHUBIMAGE}" "${QUAYIMAGE}"; do
|
|
||||||
docker buildx imagetools create -t ${MANIFESTIMAGE}:latest ${MANIFESTIMAGE}:amd64-latest ${MANIFESTIMAGE}:arm64v8-latest
|
|
||||||
docker buildx imagetools create -t ${MANIFESTIMAGE}:${META_TAG} ${MANIFESTIMAGE}:amd64-${META_TAG} ${MANIFESTIMAGE}:arm64v8-${META_TAG}
|
|
||||||
docker buildx imagetools create -t ${MANIFESTIMAGE}:${EXT_RELEASE_TAG} ${MANIFESTIMAGE}:amd64-${EXT_RELEASE_TAG} ${MANIFESTIMAGE}:arm64v8-${EXT_RELEASE_TAG}
|
|
||||||
if [ -n "${SEMVER}" ]; then
|
|
||||||
docker buildx imagetools create -t ${MANIFESTIMAGE}:${SEMVER} ${MANIFESTIMAGE}:amd64-${SEMVER} ${MANIFESTIMAGE}:arm64v8-${SEMVER}
|
|
||||||
fi
|
|
||||||
done
|
|
||||||
'''
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
@ -845,23 +1005,41 @@ pipeline {
|
||||||
environment name: 'EXIT_STATUS', value: ''
|
environment name: 'EXIT_STATUS', value: ''
|
||||||
}
|
}
|
||||||
steps {
|
steps {
|
||||||
echo "Pushing New tag for current commit ${META_TAG}"
|
|
||||||
sh '''curl -H "Authorization: token ${GITHUB_TOKEN}" -X POST https://api.github.com/repos/${LS_USER}/${LS_REPO}/git/tags \
|
|
||||||
-d '{"tag":"'${META_TAG}'",\
|
|
||||||
"object": "'${COMMIT_SHA}'",\
|
|
||||||
"message": "Tagging Release '${EXT_RELEASE_CLEAN}'-ls'${LS_TAG_NUMBER}' to master",\
|
|
||||||
"type": "commit",\
|
|
||||||
"tagger": {"name": "LinuxServer Jenkins","email": "jenkins@linuxserver.io","date": "'${GITHUB_DATE}'"}}' '''
|
|
||||||
echo "Pushing New release for Tag"
|
|
||||||
sh '''#! /bin/bash
|
sh '''#! /bin/bash
|
||||||
|
echo "Auto-generating release notes"
|
||||||
|
if [ "$(git tag --points-at HEAD)" != "" ]; then
|
||||||
|
echo "Existing tag points to current commit, suggesting no new LS changes"
|
||||||
|
AUTO_RELEASE_NOTES="No changes"
|
||||||
|
else
|
||||||
|
AUTO_RELEASE_NOTES=$(curl -fsL -H "Authorization: token ${GITHUB_TOKEN}" -H "Accept: application/vnd.github+json" -X POST https://api.github.com/repos/${LS_USER}/${LS_REPO}/releases/generate-notes \
|
||||||
|
-d '{"tag_name":"'${META_TAG}'",\
|
||||||
|
"target_commitish": "master"}' \
|
||||||
|
| jq -r '.body' | sed 's|## What.s Changed||')
|
||||||
|
fi
|
||||||
|
echo "Pushing New tag for current commit ${META_TAG}"
|
||||||
|
curl -H "Authorization: token ${GITHUB_TOKEN}" -X POST https://api.github.com/repos/${LS_USER}/${LS_REPO}/git/tags \
|
||||||
|
-d '{"tag":"'${META_TAG}'",\
|
||||||
|
"object": "'${COMMIT_SHA}'",\
|
||||||
|
"message": "Tagging Release '${EXT_RELEASE_CLEAN}'-ls'${LS_TAG_NUMBER}' to master",\
|
||||||
|
"type": "commit",\
|
||||||
|
"tagger": {"name": "LinuxServer-CI","email": "ci@linuxserver.io","date": "'${GITHUB_DATE}'"}}'
|
||||||
|
echo "Pushing New release for Tag"
|
||||||
echo "Data change at JSON endpoint ${JSON_URL}" > releasebody.json
|
echo "Data change at JSON endpoint ${JSON_URL}" > releasebody.json
|
||||||
echo '{"tag_name":"'${META_TAG}'",\
|
jq -n \
|
||||||
"target_commitish": "master",\
|
--arg tag_name "$META_TAG" \
|
||||||
"name": "'${META_TAG}'",\
|
--arg target_commitish "master" \
|
||||||
"body": "**LinuxServer Changes:**\\n\\n'${LS_RELEASE_NOTES}'\\n\\n**Remote Changes:**\\n\\n' > start
|
--arg ci_url "${CI_URL:-N/A}" \
|
||||||
printf '","draft": false,"prerelease": false}' >> releasebody.json
|
--arg ls_notes "$AUTO_RELEASE_NOTES" \
|
||||||
paste -d'\\0' start releasebody.json > releasebody.json.done
|
--arg remote_notes "$(cat releasebody.json)" \
|
||||||
curl -H "Authorization: token ${GITHUB_TOKEN}" -X POST https://api.github.com/repos/${LS_USER}/${LS_REPO}/releases -d @releasebody.json.done'''
|
'{
|
||||||
|
"tag_name": $tag_name,
|
||||||
|
"target_commitish": $target_commitish,
|
||||||
|
"name": $tag_name,
|
||||||
|
"body": ("**CI Report:**\\n\\n" + $ci_url + "\\n\\n**LinuxServer Changes:**\\n\\n" + $ls_notes + "\\n\\n**Remote Changes:**\\n\\n" + $remote_notes),
|
||||||
|
"draft": false,
|
||||||
|
"prerelease": false }' > releasebody.json.done
|
||||||
|
curl -H "Authorization: token ${GITHUB_TOKEN}" -X POST https://api.github.com/repos/${LS_USER}/${LS_REPO}/releases -d @releasebody.json.done
|
||||||
|
'''
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
// Add protection to the release branch
|
// Add protection to the release branch
|
||||||
|
|
@ -899,116 +1077,174 @@ EOF
|
||||||
) '''
|
) '''
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
// If this is a Pull request send the CI link as a comment on it
|
|
||||||
stage('Pull Request Comment') {
|
|
||||||
when {
|
|
||||||
not {environment name: 'CHANGE_ID', value: ''}
|
|
||||||
environment name: 'EXIT_STATUS', value: ''
|
|
||||||
}
|
|
||||||
steps {
|
|
||||||
sh '''#! /bin/bash
|
|
||||||
# Function to retrieve JSON data from URL
|
|
||||||
get_json() {
|
|
||||||
local url="$1"
|
|
||||||
local response=$(curl -s "$url")
|
|
||||||
if [ $? -ne 0 ]; then
|
|
||||||
echo "Failed to retrieve JSON data from $url"
|
|
||||||
return 1
|
|
||||||
fi
|
|
||||||
local json=$(echo "$response" | jq .)
|
|
||||||
if [ $? -ne 0 ]; then
|
|
||||||
echo "Failed to parse JSON data from $url"
|
|
||||||
return 1
|
|
||||||
fi
|
|
||||||
echo "$json"
|
|
||||||
}
|
|
||||||
|
|
||||||
build_table() {
|
|
||||||
local data="$1"
|
|
||||||
|
|
||||||
# Get the keys in the JSON data
|
|
||||||
local keys=$(echo "$data" | jq -r 'to_entries | map(.key) | .[]')
|
|
||||||
|
|
||||||
# Check if keys are empty
|
|
||||||
if [ -z "$keys" ]; then
|
|
||||||
echo "JSON report data does not contain any keys or the report does not exist."
|
|
||||||
return 1
|
|
||||||
fi
|
|
||||||
|
|
||||||
# Build table header
|
|
||||||
local header="| Tag | Passed |\\n| --- | --- |\\n"
|
|
||||||
|
|
||||||
# Loop through the JSON data to build the table rows
|
|
||||||
local rows=""
|
|
||||||
for build in $keys; do
|
|
||||||
local status=$(echo "$data" | jq -r ".[\\"$build\\"].test_success")
|
|
||||||
if [ "$status" = "true" ]; then
|
|
||||||
status="✅"
|
|
||||||
else
|
|
||||||
status="❌"
|
|
||||||
fi
|
|
||||||
local row="| "$build" | "$status" |\\n"
|
|
||||||
rows="${rows}${row}"
|
|
||||||
done
|
|
||||||
|
|
||||||
local table="${header}${rows}"
|
|
||||||
local escaped_table=$(echo "$table" | sed 's/\"/\\\\"/g')
|
|
||||||
echo "$escaped_table"
|
|
||||||
}
|
|
||||||
|
|
||||||
if [[ "${CI}" = "true" ]]; then
|
|
||||||
# Retrieve JSON data from URL
|
|
||||||
data=$(get_json "$CI_JSON_URL")
|
|
||||||
# Create table from JSON data
|
|
||||||
table=$(build_table "$data")
|
|
||||||
echo -e "$table"
|
|
||||||
|
|
||||||
curl -X POST -H "Authorization: token $GITHUB_TOKEN" \
|
|
||||||
-H "Accept: application/vnd.github.v3+json" \
|
|
||||||
"https://api.github.com/repos/$LS_USER/$LS_REPO/issues/$PULL_REQUEST/comments" \
|
|
||||||
-d "{\\"body\\": \\"I am a bot, here are the test results for this PR: \\n${CI_URL}\\n${SHELLCHECK_URL}\\n${table}\\"}"
|
|
||||||
else
|
|
||||||
curl -X POST -H "Authorization: token $GITHUB_TOKEN" \
|
|
||||||
-H "Accept: application/vnd.github.v3+json" \
|
|
||||||
"https://api.github.com/repos/$LS_USER/$LS_REPO/issues/$PULL_REQUEST/comments" \
|
|
||||||
-d "{\\"body\\": \\"I am a bot, here is the pushed image/manifest for this PR: \\n\\n\\`${GITHUBIMAGE}:${META_TAG}\\`\\"}"
|
|
||||||
fi
|
|
||||||
'''
|
|
||||||
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
/* ######################
|
/* ######################
|
||||||
Send status to Discord
|
Comment on PR and Send status to Discord
|
||||||
###################### */
|
###################### */
|
||||||
post {
|
post {
|
||||||
always {
|
always {
|
||||||
script{
|
script {
|
||||||
|
env.JOB_DATE = sh(
|
||||||
|
script: '''date '+%Y-%m-%dT%H:%M:%S%:z' ''',
|
||||||
|
returnStdout: true).trim()
|
||||||
if (env.EXIT_STATUS == "ABORTED"){
|
if (env.EXIT_STATUS == "ABORTED"){
|
||||||
sh 'echo "build aborted"'
|
sh 'echo "build aborted"'
|
||||||
}
|
}else{
|
||||||
else if (currentBuild.currentResult == "SUCCESS"){
|
if (currentBuild.currentResult == "SUCCESS"){
|
||||||
sh ''' curl -X POST -H "Content-Type: application/json" --data '{"avatar_url": "https://raw.githubusercontent.com/linuxserver/docker-templates/master/linuxserver.io/img/jenkins-avatar.png","embeds": [{"color": 1681177,\
|
if (env.GITHUBIMAGE =~ /lspipepr/){
|
||||||
"description": "**Build:** '${BUILD_NUMBER}'\\n**CI Results:** '${CI_URL}'\\n**ShellCheck Results:** '${SHELLCHECK_URL}'\\n**Status:** Success\\n**Job:** '${RUN_DISPLAY_URL}'\\n**Change:** '${CODE_URL}'\\n**External Release:**: '${RELEASE_LINK}'\\n**DockerHub:** '${DOCKERHUB_LINK}'\\n"}],\
|
env.JOB_WEBHOOK_STATUS='Success'
|
||||||
"username": "Jenkins"}' ${BUILDS_DISCORD} '''
|
env.JOB_WEBHOOK_COLOUR=3957028
|
||||||
}
|
env.JOB_WEBHOOK_FOOTER='PR Build'
|
||||||
else {
|
}else if (env.GITHUBIMAGE =~ /lsiodev/){
|
||||||
sh ''' curl -X POST -H "Content-Type: application/json" --data '{"avatar_url": "https://raw.githubusercontent.com/linuxserver/docker-templates/master/linuxserver.io/img/jenkins-avatar.png","embeds": [{"color": 16711680,\
|
env.JOB_WEBHOOK_STATUS='Success'
|
||||||
"description": "**Build:** '${BUILD_NUMBER}'\\n**CI Results:** '${CI_URL}'\\n**ShellCheck Results:** '${SHELLCHECK_URL}'\\n**Status:** failure\\n**Job:** '${RUN_DISPLAY_URL}'\\n**Change:** '${CODE_URL}'\\n**External Release:**: '${RELEASE_LINK}'\\n**DockerHub:** '${DOCKERHUB_LINK}'\\n"}],\
|
env.JOB_WEBHOOK_COLOUR=3957028
|
||||||
|
env.JOB_WEBHOOK_FOOTER='Dev Build'
|
||||||
|
}else{
|
||||||
|
env.JOB_WEBHOOK_STATUS='Success'
|
||||||
|
env.JOB_WEBHOOK_COLOUR=1681177
|
||||||
|
env.JOB_WEBHOOK_FOOTER='Live Build'
|
||||||
|
}
|
||||||
|
}else{
|
||||||
|
if (env.GITHUBIMAGE =~ /lspipepr/){
|
||||||
|
env.JOB_WEBHOOK_STATUS='Failure'
|
||||||
|
env.JOB_WEBHOOK_COLOUR=12669523
|
||||||
|
env.JOB_WEBHOOK_FOOTER='PR Build'
|
||||||
|
}else if (env.GITHUBIMAGE =~ /lsiodev/){
|
||||||
|
env.JOB_WEBHOOK_STATUS='Failure'
|
||||||
|
env.JOB_WEBHOOK_COLOUR=12669523
|
||||||
|
env.JOB_WEBHOOK_FOOTER='Dev Build'
|
||||||
|
}else{
|
||||||
|
env.JOB_WEBHOOK_STATUS='Failure'
|
||||||
|
env.JOB_WEBHOOK_COLOUR=16711680
|
||||||
|
env.JOB_WEBHOOK_FOOTER='Live Build'
|
||||||
|
}
|
||||||
|
}
|
||||||
|
sh ''' curl -X POST -H "Content-Type: application/json" --data '{"avatar_url": "https://raw.githubusercontent.com/linuxserver/docker-templates/master/linuxserver.io/img/jenkins-avatar.png","embeds": [{"'color'": '${JOB_WEBHOOK_COLOUR}',\
|
||||||
|
"footer": {"text" : "'"${JOB_WEBHOOK_FOOTER}"'"},\
|
||||||
|
"timestamp": "'${JOB_DATE}'",\
|
||||||
|
"description": "**Build:** '${BUILD_NUMBER}'\\n**CI Results:** '${CI_URL}'\\n**ShellCheck Results:** '${SHELLCHECK_URL}'\\n**Status:** '${JOB_WEBHOOK_STATUS}'\\n**Job:** '${RUN_DISPLAY_URL}'\\n**Change:** '${CODE_URL}'\\n**External Release:**: '${RELEASE_LINK}'\\n**DockerHub:** '${DOCKERHUB_LINK}'\\n"}],\
|
||||||
"username": "Jenkins"}' ${BUILDS_DISCORD} '''
|
"username": "Jenkins"}' ${BUILDS_DISCORD} '''
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
script {
|
||||||
|
if (env.GITHUBIMAGE =~ /lspipepr/){
|
||||||
|
if (env.CI_TEST_ATTEMPTED == "true"){
|
||||||
|
sh '''#! /bin/bash
|
||||||
|
# Function to retrieve JSON data from URL
|
||||||
|
get_json() {
|
||||||
|
local url="$1"
|
||||||
|
local response=$(curl -s "$url")
|
||||||
|
if [ $? -ne 0 ]; then
|
||||||
|
echo "Failed to retrieve JSON data from $url"
|
||||||
|
return 1
|
||||||
|
fi
|
||||||
|
local json=$(echo "$response" | jq .)
|
||||||
|
if [ $? -ne 0 ]; then
|
||||||
|
echo "Failed to parse JSON data from $url"
|
||||||
|
return 1
|
||||||
|
fi
|
||||||
|
echo "$json"
|
||||||
|
}
|
||||||
|
|
||||||
|
build_table() {
|
||||||
|
local data="$1"
|
||||||
|
|
||||||
|
# Get the keys in the JSON data
|
||||||
|
local keys=$(echo "$data" | jq -r 'to_entries | map(.key) | .[]')
|
||||||
|
|
||||||
|
# Check if keys are empty
|
||||||
|
if [ -z "$keys" ]; then
|
||||||
|
echo "JSON report data does not contain any keys or the report does not exist."
|
||||||
|
return 1
|
||||||
|
fi
|
||||||
|
|
||||||
|
# Build table header
|
||||||
|
local header="| Tag | Passed |\\n| --- | --- |\\n"
|
||||||
|
|
||||||
|
# Loop through the JSON data to build the table rows
|
||||||
|
local rows=""
|
||||||
|
for build in $keys; do
|
||||||
|
local status=$(echo "$data" | jq -r ".[\\"$build\\"].test_success")
|
||||||
|
if [ "$status" = "true" ]; then
|
||||||
|
status="✅"
|
||||||
|
else
|
||||||
|
status="❌"
|
||||||
|
fi
|
||||||
|
local row="| "$build" | "$status" |\\n"
|
||||||
|
rows="${rows}${row}"
|
||||||
|
done
|
||||||
|
|
||||||
|
local table="${header}${rows}"
|
||||||
|
local escaped_table=$(echo "$table" | sed 's/\"/\\\\"/g')
|
||||||
|
echo "$escaped_table"
|
||||||
|
}
|
||||||
|
|
||||||
|
if [[ "${CI}" = "true" ]]; then
|
||||||
|
# Retrieve JSON data from URL
|
||||||
|
data=$(get_json "$CI_JSON_URL")
|
||||||
|
# Create table from JSON data
|
||||||
|
table=$(build_table "$data")
|
||||||
|
echo -e "$table"
|
||||||
|
|
||||||
|
curl -X POST -H "Authorization: token $GITHUB_TOKEN" \
|
||||||
|
-H "Accept: application/vnd.github.v3+json" \
|
||||||
|
"https://api.github.com/repos/$LS_USER/$LS_REPO/issues/$PULL_REQUEST/comments" \
|
||||||
|
-d "{\\"body\\": \\"I am a bot, here are the test results for this PR: \\n${CI_URL}\\n${SHELLCHECK_URL}\\n${table}\\"}"
|
||||||
|
else
|
||||||
|
curl -X POST -H "Authorization: token $GITHUB_TOKEN" \
|
||||||
|
-H "Accept: application/vnd.github.v3+json" \
|
||||||
|
"https://api.github.com/repos/$LS_USER/$LS_REPO/issues/$PULL_REQUEST/comments" \
|
||||||
|
-d "{\\"body\\": \\"I am a bot, here is the pushed image/manifest for this PR: \\n\\n\\`${GITHUBIMAGE}:${META_TAG}\\`\\"}"
|
||||||
|
fi
|
||||||
|
'''
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
sh '''#!/bin/bash
|
||||||
|
rm -rf /config/.ssh/id_sign
|
||||||
|
rm -rf /config/.ssh/id_sign.pub
|
||||||
|
git config --global --unset gpg.format
|
||||||
|
git config --global --unset user.signingkey
|
||||||
|
git config --global --unset commit.gpgsign
|
||||||
|
'''
|
||||||
}
|
}
|
||||||
cleanup {
|
cleanup {
|
||||||
sh '''#! /bin/bash
|
sh '''#! /bin/bash
|
||||||
echo "Performing docker system prune!!"
|
echo "Pruning builder!!"
|
||||||
containers=$(docker ps -aq)
|
docker builder prune -f --builder container || :
|
||||||
|
containers=$(docker ps -q)
|
||||||
if [[ -n "${containers}" ]]; then
|
if [[ -n "${containers}" ]]; then
|
||||||
docker stop ${containers}
|
BUILDX_CONTAINER_ID=$(docker ps -qf 'name=buildx_buildkit')
|
||||||
|
for container in ${containers}; do
|
||||||
|
if [[ "${container}" == "${BUILDX_CONTAINER_ID}" ]]; then
|
||||||
|
echo "skipping buildx container in docker stop"
|
||||||
|
else
|
||||||
|
echo "Stopping container ${container}"
|
||||||
|
docker stop ${container}
|
||||||
|
fi
|
||||||
|
done
|
||||||
fi
|
fi
|
||||||
docker system prune -af --volumes || :
|
docker system prune -f --volumes || :
|
||||||
|
docker image prune -af || :
|
||||||
'''
|
'''
|
||||||
cleanWs()
|
cleanWs()
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
def retry_backoff(int max_attempts, int power_base, Closure c) {
|
||||||
|
int n = 0
|
||||||
|
while (n < max_attempts) {
|
||||||
|
try {
|
||||||
|
c()
|
||||||
|
return
|
||||||
|
} catch (err) {
|
||||||
|
if ((n + 1) >= max_attempts) {
|
||||||
|
throw err
|
||||||
|
}
|
||||||
|
sleep(power_base ** n)
|
||||||
|
n++
|
||||||
|
}
|
||||||
|
}
|
||||||
|
return
|
||||||
|
}
|
||||||
|
|
|
||||||
70
README.md
70
README.md
|
|
@ -1,11 +1,24 @@
|
||||||
|
<<<<<<< HEAD
|
||||||
<!-- DO NOT EDIT THIS FILE MANUALLY -->
|
<!-- DO NOT EDIT THIS FILE MANUALLY -->
|
||||||
<!-- Please read https://github.com/linuxserver/docker-plex/blob/master/.github/CONTRIBUTING.md -->
|
<!-- Please read https://github.com/linuxserver/docker-plex/blob/master/.github/CONTRIBUTING.md -->
|
||||||
|
=======
|
||||||
|
due to the recent purge of Hosters by plex, this fork adds wireguard functionallity to the plex container as a quick workaround. all traffic gets tunneled.
|
||||||
|
|
||||||
|
wireguard onfig file has to be mounted in /etc/wireguard/wg0.conf. e.g.:
|
||||||
|
```
|
||||||
|
volumes:
|
||||||
|
- ./wg0.conf:/etc/wireguard/wg0.conf
|
||||||
|
```
|
||||||
|
|
||||||
|
<!-- DO NOT EDIT THIS FILE MANUALLY -->
|
||||||
|
<!-- Please read the https://github.com/linuxserver/docker-plex/blob/master/.github/CONTRIBUTING.md -->
|
||||||
|
|
||||||
|
>>>>>>> 72bdfaa (Update README.md)
|
||||||
[](https://linuxserver.io)
|
[](https://linuxserver.io)
|
||||||
|
|
||||||
[](https://blog.linuxserver.io "all the things you can do with our containers including How-To guides, opinions and much more!")
|
[](https://blog.linuxserver.io "all the things you can do with our containers including How-To guides, opinions and much more!")
|
||||||
[](https://discord.gg/YWrKVTn "realtime support / chat with the community and the team.")
|
[](https://linuxserver.io/discord "realtime support / chat with the community and the team.")
|
||||||
[](https://discourse.linuxserver.io "post on our community forum.")
|
[](https://discourse.linuxserver.io "post on our community forum.")
|
||||||
[](https://fleet.linuxserver.io "an online web interface which displays all of our maintained images.")
|
|
||||||
[](https://github.com/linuxserver "view the source for all of our repositories.")
|
[](https://github.com/linuxserver "view the source for all of our repositories.")
|
||||||
[](https://opencollective.com/linuxserver "please consider helping us by either donating or contributing to our budget")
|
[](https://opencollective.com/linuxserver "please consider helping us by either donating or contributing to our budget")
|
||||||
|
|
||||||
|
|
@ -20,9 +33,8 @@ The [LinuxServer.io](https://linuxserver.io) team brings you another container r
|
||||||
Find us at:
|
Find us at:
|
||||||
|
|
||||||
* [Blog](https://blog.linuxserver.io) - all the things you can do with our containers including How-To guides, opinions and much more!
|
* [Blog](https://blog.linuxserver.io) - all the things you can do with our containers including How-To guides, opinions and much more!
|
||||||
* [Discord](https://discord.gg/YWrKVTn) - realtime support / chat with the community and the team.
|
* [Discord](https://linuxserver.io/discord) - realtime support / chat with the community and the team.
|
||||||
* [Discourse](https://discourse.linuxserver.io) - post on our community forum.
|
* [Discourse](https://discourse.linuxserver.io) - post on our community forum.
|
||||||
* [Fleet](https://fleet.linuxserver.io) - an online web interface which displays all of our maintained images.
|
|
||||||
* [GitHub](https://github.com/linuxserver) - view the source for all of our repositories.
|
* [GitHub](https://github.com/linuxserver) - view the source for all of our repositories.
|
||||||
* [Open Collective](https://opencollective.com/linuxserver) - please consider helping us by either donating or contributing to our budget
|
* [Open Collective](https://opencollective.com/linuxserver) - please consider helping us by either donating or contributing to our budget
|
||||||
|
|
||||||
|
|
@ -41,7 +53,7 @@ Find us at:
|
||||||
|
|
||||||
[Plex](https://plex.tv) organizes video, music and photos from personal media libraries and streams them to smart TVs, streaming boxes and mobile devices. This container is packaged as a standalone Plex Media Server. Straightforward design and bulk actions mean getting things done faster.
|
[Plex](https://plex.tv) organizes video, music and photos from personal media libraries and streams them to smart TVs, streaming boxes and mobile devices. This container is packaged as a standalone Plex Media Server. Straightforward design and bulk actions mean getting things done faster.
|
||||||
|
|
||||||
[](https://plex.tv)
|
[](https://plex.tv)
|
||||||
|
|
||||||
## Supported Architectures
|
## Supported Architectures
|
||||||
|
|
||||||
|
|
@ -55,19 +67,21 @@ The architectures supported by this image are:
|
||||||
| :----: | :----: | ---- |
|
| :----: | :----: | ---- |
|
||||||
| x86-64 | ✅ | amd64-\<version tag\> |
|
| x86-64 | ✅ | amd64-\<version tag\> |
|
||||||
| arm64 | ✅ | arm64v8-\<version tag\> |
|
| arm64 | ✅ | arm64v8-\<version tag\> |
|
||||||
| armhf | ❌ | |
|
|
||||||
|
|
||||||
## Application Setup
|
## Application Setup
|
||||||
|
|
||||||
Webui can be found at `<your-ip>:32400/web`
|
Webui can be found at `<your-ip>:32400/web`
|
||||||
|
|
||||||
**note**: If there is no value set for the VERSION variable, then no updates will take place.
|
>[!NOTE]
|
||||||
|
>If there is no value set for the VERSION variable, then no updates will take place.
|
||||||
|
|
||||||
**note**: For new users, no updates will take place on the first run of the container as there is no preferences file to read your token from, to update restart the Docker container after logging in through the webui.
|
>[!NOTE]
|
||||||
|
>For new users, no updates will take place on the first run of the container as there is no preferences file to read your token from, to update restart the Docker container after logging in through the webui.
|
||||||
|
|
||||||
Valid settings for VERSION are:-
|
Valid settings for VERSION are:-
|
||||||
|
|
||||||
**info**: YOU CANNOT UPDATE TO A PLEXPASS ONLY (BETA) VERSION IF YOU ARE NOT LOGGED IN WITH A PLEXPASS ACCOUNT.
|
>[!NOTE]
|
||||||
|
>YOU CANNOT UPDATE TO A PLEXPASS ONLY (BETA) VERSION IF YOU ARE NOT LOGGED IN WITH A PLEXPASS ACCOUNT.
|
||||||
|
|
||||||
+ **`docker`**: Let Docker handle the Plex Version, we keep our Dockerhub Endpoint up to date with the latest public builds. This is the same as leaving this setting out of your create command.
|
+ **`docker`**: Let Docker handle the Plex Version, we keep our Dockerhub Endpoint up to date with the latest public builds. This is the same as leaving this setting out of your create command.
|
||||||
+ **`latest`**: will update plex to the latest version available that you are entitled to.
|
+ **`latest`**: will update plex to the latest version available that you are entitled to.
|
||||||
|
|
@ -99,10 +113,31 @@ We automatically add the necessary environment variable that will utilise all th
|
||||||
|
|
||||||
Best effort is made to install tools to allow mounting in /dev/dri on Arm devices. In most cases if /dev/dri exists on the host it should just work. If running a Raspberry Pi 4 be sure to enable `dtoverlay=vc4-fkms-v3d` in your usercfg.txt.
|
Best effort is made to install tools to allow mounting in /dev/dri on Arm devices. In most cases if /dev/dri exists on the host it should just work. If running a Raspberry Pi 4 be sure to enable `dtoverlay=vc4-fkms-v3d` in your usercfg.txt.
|
||||||
|
|
||||||
|
## Read-Only Operation
|
||||||
|
|
||||||
|
This image can be run with a read-only container filesystem. For details please [read the docs](https://docs.linuxserver.io/misc/read-only/).
|
||||||
|
|
||||||
|
### Caveats
|
||||||
|
|
||||||
|
* Runtime update of Plex (and thus Plexpass builds) is not supported.
|
||||||
|
* Transcode directory must be mounted to a host path or tmpfs.
|
||||||
|
|
||||||
|
## Non-Root Operation
|
||||||
|
|
||||||
|
This image can be run with a non-root user. For details please [read the docs](https://docs.linuxserver.io/misc/non-root/).
|
||||||
|
|
||||||
|
### Caveats
|
||||||
|
|
||||||
|
* Runtime update of Plex (and thus Plexpass builds) is not supported.
|
||||||
|
* Transcode directory must be mounted to a host path or tmpfs.
|
||||||
|
|
||||||
## Usage
|
## Usage
|
||||||
|
|
||||||
To help you get started creating a container from this image you can either use docker-compose or the docker cli.
|
To help you get started creating a container from this image you can either use docker-compose or the docker cli.
|
||||||
|
|
||||||
|
>[!NOTE]
|
||||||
|
>Unless a parameter is flagged as 'optional', it is *mandatory* and a value must be provided.
|
||||||
|
|
||||||
### docker-compose (recommended, [click here for more info](https://docs.linuxserver.io/general/docker-compose))
|
### docker-compose (recommended, [click here for more info](https://docs.linuxserver.io/general/docker-compose))
|
||||||
|
|
||||||
```yaml
|
```yaml
|
||||||
|
|
@ -119,7 +154,7 @@ services:
|
||||||
- VERSION=docker
|
- VERSION=docker
|
||||||
- PLEX_CLAIM= #optional
|
- PLEX_CLAIM= #optional
|
||||||
volumes:
|
volumes:
|
||||||
- /path/to/library:/config
|
- /path/to/plex/library:/config
|
||||||
- /path/to/tvseries:/tv
|
- /path/to/tvseries:/tv
|
||||||
- /path/to/movies:/movies
|
- /path/to/movies:/movies
|
||||||
restart: unless-stopped
|
restart: unless-stopped
|
||||||
|
|
@ -136,7 +171,7 @@ docker run -d \
|
||||||
-e TZ=Etc/UTC \
|
-e TZ=Etc/UTC \
|
||||||
-e VERSION=docker \
|
-e VERSION=docker \
|
||||||
-e PLEX_CLAIM= `#optional` \
|
-e PLEX_CLAIM= `#optional` \
|
||||||
-v /path/to/library:/config \
|
-v /path/to/plex/library:/config \
|
||||||
-v /path/to/tvseries:/tv \
|
-v /path/to/tvseries:/tv \
|
||||||
-v /path/to/movies:/movies \
|
-v /path/to/movies:/movies \
|
||||||
--restart unless-stopped \
|
--restart unless-stopped \
|
||||||
|
|
@ -158,6 +193,8 @@ Containers are configured using parameters passed at runtime (such as those abov
|
||||||
| `-v /config` | Plex library location. *This can grow very large, 50gb+ is likely for a large collection.* |
|
| `-v /config` | Plex library location. *This can grow very large, 50gb+ is likely for a large collection.* |
|
||||||
| `-v /tv` | Media goes here. Add as many as needed e.g. `/movies`, `/tv`, etc. |
|
| `-v /tv` | Media goes here. Add as many as needed e.g. `/movies`, `/tv`, etc. |
|
||||||
| `-v /movies` | Media goes here. Add as many as needed e.g. `/movies`, `/tv`, etc. |
|
| `-v /movies` | Media goes here. Add as many as needed e.g. `/movies`, `/tv`, etc. |
|
||||||
|
| `--read-only=true` | Run container with a read-only filesystem. Please [read the docs](https://docs.linuxserver.io/misc/read-only/). |
|
||||||
|
| `--user=1000:1000` | Run container with a non-root user. Please [read the docs](https://docs.linuxserver.io/misc/non-root/). |
|
||||||
|
|
||||||
## Environment variables from files (Docker secrets)
|
## Environment variables from files (Docker secrets)
|
||||||
|
|
||||||
|
|
@ -322,7 +359,8 @@ Below are the instructions for updating containers:
|
||||||
|
|
||||||
### Image Update Notifications - Diun (Docker Image Update Notifier)
|
### Image Update Notifications - Diun (Docker Image Update Notifier)
|
||||||
|
|
||||||
**tip**: We recommend [Diun](https://crazymax.dev/diun/) for update notifications. Other tools that automatically update containers unattended are not recommended or supported.
|
>[!TIP]
|
||||||
|
>We recommend [Diun](https://crazymax.dev/diun/) for update notifications. Other tools that automatically update containers unattended are not recommended or supported.
|
||||||
|
|
||||||
## Building locally
|
## Building locally
|
||||||
|
|
||||||
|
|
@ -337,16 +375,20 @@ docker build \
|
||||||
-t lscr.io/linuxserver/plex:latest .
|
-t lscr.io/linuxserver/plex:latest .
|
||||||
```
|
```
|
||||||
|
|
||||||
The ARM variants can be built on x86_64 hardware using `multiarch/qemu-user-static`
|
The ARM variants can be built on x86_64 hardware and vice versa using `lscr.io/linuxserver/qemu-static`
|
||||||
|
|
||||||
```bash
|
```bash
|
||||||
docker run --rm --privileged multiarch/qemu-user-static:register --reset
|
docker run --rm --privileged lscr.io/linuxserver/qemu-static --reset
|
||||||
```
|
```
|
||||||
|
|
||||||
Once registered you can define the dockerfile to use with `-f Dockerfile.aarch64`.
|
Once registered you can define the dockerfile to use with `-f Dockerfile.aarch64`.
|
||||||
|
|
||||||
## Versions
|
## Versions
|
||||||
|
|
||||||
|
* **15.03.26:** - Allow TMPDIR to be changed to better support read-only containers
|
||||||
|
* **15.03.26:** - Fix initial claim setup on non-root containers
|
||||||
|
* **04.11.24:** - Add Nvidia capability needed for h265
|
||||||
|
* **18.07.24:** - Rebase to Ubuntu Noble.
|
||||||
* **12.02.24:** - Use universal hardware acceleration blurb
|
* **12.02.24:** - Use universal hardware acceleration blurb
|
||||||
* **09.01.24:** - Set ownership on TranscoderTempDirectory when it's been saved in Preferences.
|
* **09.01.24:** - Set ownership on TranscoderTempDirectory when it's been saved in Preferences.
|
||||||
* **16.08.23:** - Install unrar from [linuxserver repo](https://github.com/linuxserver/docker-unrar).
|
* **16.08.23:** - Install unrar from [linuxserver repo](https://github.com/linuxserver/docker-unrar).
|
||||||
|
|
|
||||||
|
|
@ -23,6 +23,6 @@ repo_vars:
|
||||||
- CI_PORT='32400'
|
- CI_PORT='32400'
|
||||||
- CI_SSL='false'
|
- CI_SSL='false'
|
||||||
- CI_DELAY='120'
|
- CI_DELAY='120'
|
||||||
- CI_DOCKERENV='TZ=US/Pacific'
|
- CI_DOCKERENV=''
|
||||||
- CI_AUTH='user:password'
|
- CI_AUTH=''
|
||||||
- CI_WEBPATH='/web/index.html'
|
- CI_WEBPATH='/web/index.html'
|
||||||
|
|
|
||||||
|
|
@ -1,151 +1,148 @@
|
||||||
NAME VERSION TYPE
|
NAME VERSION TYPE
|
||||||
adduser 3.118ubuntu5 deb
|
adduser 3.137ubuntu1 deb
|
||||||
apt 2.4.12 deb
|
apt 2.8.3 deb
|
||||||
apt-utils 2.4.12 deb
|
apt-utils 2.8.3 deb
|
||||||
base-files 12ubuntu4.6 deb
|
base-files 13ubuntu10.4 deb
|
||||||
base-passwd 3.5.52build1 deb
|
base-passwd 3.6.3build1 deb
|
||||||
bash 5.1-6ubuntu1.1 deb
|
bash 5.2.21-2ubuntu4 deb
|
||||||
bsdutils 1:2.37.2-4ubuntu3.4 deb
|
bsdutils 1:2.39.3-9ubuntu6.5 deb
|
||||||
ca-certificates 20230311ubuntu0.22.04.1 deb
|
ca-certificates 20240203 deb
|
||||||
catatonit 0.1.7-1 deb
|
catatonit 0.1.7-1 deb
|
||||||
coreutils 8.32-4.1ubuntu1.2 deb
|
coreutils 9.4-3ubuntu6.2 deb
|
||||||
cron 3.0pl1-137ubuntu3 deb
|
cron 3.0pl1-184ubuntu2 deb
|
||||||
curl 7.81.0-1ubuntu1.16 deb
|
cron-daemon-common 3.0pl1-184ubuntu2 deb
|
||||||
dash 0.5.11+git20210903+057cd650a4ed-3build1 deb
|
curl 8.5.0-2ubuntu10.8 deb
|
||||||
debconf 1.5.79ubuntu1 deb
|
dash 0.5.12-6ubuntu5 deb
|
||||||
debianutils 5.5-1ubuntu2 deb
|
debconf 1.5.86ubuntu1 deb
|
||||||
diffutils 1:3.8-0ubuntu2 deb
|
debianutils 5.17build1 deb
|
||||||
dirmngr 2.2.27-3ubuntu2.1 deb
|
diffutils 1:3.10-1build1 deb
|
||||||
dpkg 1.21.1ubuntu2.3 deb
|
dirmngr 2.4.4-2ubuntu17.4 deb
|
||||||
e2fsprogs 1.46.5-2ubuntu1.1 deb
|
dpkg 1.22.6ubuntu6.5 deb
|
||||||
findutils 4.8.0-1ubuntu3 deb
|
e2fsprogs 1.47.0-2.4~exp1ubuntu4.1 deb
|
||||||
gcc-12-base 12.3.0-1ubuntu1~22.04 deb
|
findutils 4.9.0-5build1 deb
|
||||||
gnupg 2.2.27-3ubuntu2.1 deb
|
gcc-14-base 14.2.0-4ubuntu2~24.04.1 deb
|
||||||
gnupg-l10n 2.2.27-3ubuntu2.1 deb
|
gnupg 2.4.4-2ubuntu17.4 deb
|
||||||
gnupg-utils 2.2.27-3ubuntu2.1 deb
|
gnupg-l10n 2.4.4-2ubuntu17.4 deb
|
||||||
gpg 2.2.27-3ubuntu2.1 deb
|
gnupg-utils 2.4.4-2ubuntu17.4 deb
|
||||||
gpg-agent 2.2.27-3ubuntu2.1 deb
|
gpg 2.4.4-2ubuntu17.4 deb
|
||||||
gpg-wks-client 2.2.27-3ubuntu2.1 deb
|
gpg-agent 2.4.4-2ubuntu17.4 deb
|
||||||
gpg-wks-server 2.2.27-3ubuntu2.1 deb
|
gpg-wks-client 2.4.4-2ubuntu17.4 deb
|
||||||
gpgconf 2.2.27-3ubuntu2.1 deb
|
gpgconf 2.4.4-2ubuntu17.4 deb
|
||||||
gpgsm 2.2.27-3ubuntu2.1 deb
|
gpgsm 2.4.4-2ubuntu17.4 deb
|
||||||
gpgv 2.2.27-3ubuntu2.1 deb
|
gpgv 2.4.4-2ubuntu17.4 deb
|
||||||
grep 3.7-1build1 deb
|
grep 3.11-4build1 deb
|
||||||
gzip 1.10-4ubuntu4.1 deb
|
gzip 1.12-1ubuntu3.1 deb
|
||||||
hostname 3.23ubuntu2 deb
|
hostname 3.23+nmu2ubuntu2 deb
|
||||||
init-system-helpers 1.62 deb
|
init-system-helpers 1.66ubuntu1 deb
|
||||||
jq 1.6-2.1ubuntu3 deb
|
jq 1.7.1-3ubuntu0.24.04.2 deb
|
||||||
libacl1 2.3.1-1 deb
|
keyboxd 2.4.4-2ubuntu17.4 deb
|
||||||
libapt-pkg6.0 2.4.12 deb
|
krb5-locales 1.20.1-6ubuntu2.6 deb
|
||||||
libassuan0 2.5.5-1build1 deb
|
libacl1 2.3.2-1build1.1 deb
|
||||||
libattr1 1:2.5.1-1build1 deb
|
libapt-pkg6.0t64 2.8.3 deb
|
||||||
libaudit-common 1:3.0.7-1build1 deb
|
libassuan0 2.5.6-1build1 deb
|
||||||
libaudit1 1:3.0.7-1build1 deb
|
libattr1 1:2.5.2-1build1.1 deb
|
||||||
libblkid1 2.37.2-4ubuntu3.4 deb
|
libaudit-common 1:3.1.2-2.1build1.1 deb
|
||||||
libbrotli1 1.0.9-2build6 deb
|
libaudit1 1:3.1.2-2.1build1.1 deb
|
||||||
libbsd0 0.11.5-1 deb
|
libblkid1 2.39.3-9ubuntu6.5 deb
|
||||||
libbz2-1.0 1.0.8-5build1 deb
|
libbrotli1 1.1.0-2build2 deb
|
||||||
libc-bin 2.35-0ubuntu3.8 deb
|
libbsd0 0.12.1-1build1.1 deb
|
||||||
libc6 2.35-0ubuntu3.8 deb
|
libbz2-1.0 1.0.8-5.1build0.1 deb
|
||||||
libcap-ng0 0.7.9-2.2build3 deb
|
libc-bin 2.39-0ubuntu8.7 deb
|
||||||
libcap2 1:2.44-1ubuntu0.22.04.1 deb
|
libc6 2.39-0ubuntu8.7 deb
|
||||||
libcom-err2 1.46.5-2ubuntu1.1 deb
|
libcap-ng0 0.8.4-2build2 deb
|
||||||
libcrypt1 1:4.4.27-1 deb
|
libcap2 1:2.66-5ubuntu2.4 deb
|
||||||
libcurl4 7.81.0-1ubuntu1.16 deb
|
libcom-err2 1.47.0-2.4~exp1ubuntu4.1 deb
|
||||||
libdb5.3 5.3.28+dfsg1-0.8ubuntu3 deb
|
libcrypt1 1:4.4.36-4build1 deb
|
||||||
libdebconfclient0 0.261ubuntu1 deb
|
libcurl4t64 8.5.0-2ubuntu10.8 deb
|
||||||
libext2fs2 1.46.5-2ubuntu1.1 deb
|
libdb5.3t64 5.3.28+dfsg2-7 deb
|
||||||
libffi8 3.4.2-4 deb
|
libdebconfclient0 0.271ubuntu3 deb
|
||||||
libgcc-s1 12.3.0-1ubuntu1~22.04 deb
|
libext2fs2t64 1.47.0-2.4~exp1ubuntu4.1 deb
|
||||||
libgcrypt20 1.9.4-3ubuntu3 deb
|
libffi8 3.4.6-1build1 deb
|
||||||
libgmp10 2:6.2.1+dfsg-3ubuntu1 deb
|
libgcc-s1 14.2.0-4ubuntu2~24.04.1 deb
|
||||||
libgnutls30 3.7.3-4ubuntu1.5 deb
|
libgcrypt20 1.10.3-2build1 deb
|
||||||
libgpg-error0 1.43-3 deb
|
libgmp10 2:6.3.0+dfsg-2ubuntu6.1 deb
|
||||||
libgssapi-krb5-2 1.19.2-2ubuntu0.3 deb
|
libgnutls30t64 3.8.3-1.1ubuntu3.5 deb
|
||||||
libhogweed6 3.7.3-1build2 deb
|
libgpg-error0 1.47-3build2.1 deb
|
||||||
libidn2-0 2.3.2-2build1 deb
|
libgssapi-krb5-2 1.20.1-6ubuntu2.6 deb
|
||||||
libjq1 1.6-2.1ubuntu3 deb
|
libhogweed6t64 3.9.1-2.2build1.1 deb
|
||||||
libk5crypto3 1.19.2-2ubuntu0.3 deb
|
libidn2-0 2.3.7-2build1.1 deb
|
||||||
libkeyutils1 1.6.1-2ubuntu3 deb
|
libjq1 1.7.1-3ubuntu0.24.04.2 deb
|
||||||
libkmod2 29-1ubuntu1 deb
|
libk5crypto3 1.20.1-6ubuntu2.6 deb
|
||||||
libkrb5-3 1.19.2-2ubuntu0.3 deb
|
libkeyutils1 1.6.3-3build1 deb
|
||||||
libkrb5support0 1.19.2-2ubuntu0.3 deb
|
libkmod2 31+20240202-2ubuntu7.2 deb
|
||||||
libksba8 1.6.0-2ubuntu0.2 deb
|
libkrb5-3 1.20.1-6ubuntu2.6 deb
|
||||||
libldap-2.5-0 2.5.17+dfsg-0ubuntu0.22.04.1 deb
|
libkrb5support0 1.20.1-6ubuntu2.6 deb
|
||||||
libldap-common 2.5.17+dfsg-0ubuntu0.22.04.1 deb
|
libksba8 1.6.6-1build1 deb
|
||||||
liblz4-1 1.9.3-2build2 deb
|
libldap-common 2.6.10+dfsg-0ubuntu0.24.04.1 deb
|
||||||
liblzma5 5.2.5-2ubuntu1 deb
|
libldap2 2.6.10+dfsg-0ubuntu0.24.04.1 deb
|
||||||
libmd0 1.0.4-1build1 deb
|
liblz4-1 1.9.4-1build1.1 deb
|
||||||
libmount1 2.37.2-4ubuntu3.4 deb
|
liblzma5 5.6.1+really5.4.5-1ubuntu0.2 deb
|
||||||
libncurses6 6.3-2ubuntu0.1 deb
|
libmd0 1.1.0-2build1.1 deb
|
||||||
libncursesw6 6.3-2ubuntu0.1 deb
|
libmount1 2.39.3-9ubuntu6.5 deb
|
||||||
libnettle8 3.7.3-1build2 deb
|
libncursesw6 6.4+20240113-1ubuntu2 deb
|
||||||
libnghttp2-14 1.43.0-1ubuntu0.2 deb
|
libnettle8t64 3.9.1-2.2build1.1 deb
|
||||||
libnpth0 1.6-3build2 deb
|
libnghttp2-14 1.59.0-1ubuntu0.2 deb
|
||||||
libnsl2 1.3.0-2build2 deb
|
libnpth0t64 1.6-3.1build1 deb
|
||||||
libonig5 6.9.7.1-2build1 deb
|
libonig5 6.9.9-1build1 deb
|
||||||
libp11-kit0 0.24.0-6build1 deb
|
libp11-kit0 0.25.3-4ubuntu2.1 deb
|
||||||
libpam-modules 1.4.0-11ubuntu2.4 deb
|
libpam-modules 1.5.3-5ubuntu5.5 deb
|
||||||
libpam-modules-bin 1.4.0-11ubuntu2.4 deb
|
libpam-modules-bin 1.5.3-5ubuntu5.5 deb
|
||||||
libpam-runtime 1.4.0-11ubuntu2.4 deb
|
libpam-runtime 1.5.3-5ubuntu5.5 deb
|
||||||
libpam0g 1.4.0-11ubuntu2.4 deb
|
libpam0g 1.5.3-5ubuntu5.5 deb
|
||||||
libpcre2-8-0 10.39-3ubuntu0.1 deb
|
libpcre2-8-0 10.42-4ubuntu2.1 deb
|
||||||
libpcre3 2:8.39-13ubuntu0.22.04.1 deb
|
libproc2-0 2:4.0.4-4ubuntu3.2 deb
|
||||||
libprocps8 2:3.3.17-6ubuntu2.1 deb
|
libpsl5t64 0.21.2-1.1build1 deb
|
||||||
libpsl5 0.21.0-1.2build2 deb
|
libreadline8t64 8.2-4build1 deb
|
||||||
libreadline8 8.1.2-1 deb
|
librtmp1 2.4+20151223.gitfa8646d.1-2build7 deb
|
||||||
librtmp1 2.4+20151223.gitfa8646d.1-2build4 deb
|
libsasl2-2 2.1.28+dfsg1-5ubuntu3.1 deb
|
||||||
libsasl2-2 2.1.27+dfsg2-3ubuntu1.2 deb
|
libsasl2-modules 2.1.28+dfsg1-5ubuntu3.1 deb
|
||||||
libsasl2-modules 2.1.27+dfsg2-3ubuntu1.2 deb
|
libsasl2-modules-db 2.1.28+dfsg1-5ubuntu3.1 deb
|
||||||
libsasl2-modules-db 2.1.27+dfsg2-3ubuntu1.2 deb
|
libseccomp2 2.5.5-1ubuntu3.1 deb
|
||||||
libseccomp2 2.5.3-2ubuntu2 deb
|
libselinux1 3.5-2ubuntu2.1 deb
|
||||||
libselinux1 3.3-1build2 deb
|
libsemanage-common 3.5-1build5 deb
|
||||||
libsemanage-common 3.3-1build2 deb
|
libsemanage2 3.5-1build5 deb
|
||||||
libsemanage2 3.3-1build2 deb
|
libsepol2 3.5-2build1 deb
|
||||||
libsepol2 3.3-1build1 deb
|
libsmartcols1 2.39.3-9ubuntu6.5 deb
|
||||||
libsmartcols1 2.37.2-4ubuntu3.4 deb
|
libsqlite3-0 3.45.1-1ubuntu2.5 deb
|
||||||
libsqlite3-0 3.37.2-2ubuntu0.3 deb
|
libss2 1.47.0-2.4~exp1ubuntu4.1 deb
|
||||||
libss2 1.46.5-2ubuntu1.1 deb
|
libssh-4 0.10.6-2ubuntu0.4 deb
|
||||||
libssh-4 0.9.6-2ubuntu0.22.04.3 deb
|
libssl3t64 3.0.13-0ubuntu3.9 deb
|
||||||
libssl3 3.0.2-0ubuntu1.16 deb
|
libstdc++6 14.2.0-4ubuntu2~24.04.1 deb
|
||||||
libstdc++6 12.3.0-1ubuntu1~22.04 deb
|
libsystemd0 255.4-1ubuntu8.15 deb
|
||||||
libsystemd0 249.11-0ubuntu3.12 deb
|
libtasn1-6 4.19.0-3ubuntu0.24.04.2 deb
|
||||||
libtasn1-6 4.18.0-4build1 deb
|
libtinfo6 6.4+20240113-1ubuntu2 deb
|
||||||
libtinfo6 6.3-2ubuntu0.1 deb
|
libudev1 255.4-1ubuntu8.15 deb
|
||||||
libtirpc-common 1.3.2-2ubuntu0.1 deb
|
libunistring5 1.1-2build1.1 deb
|
||||||
libtirpc3 1.3.2-2ubuntu0.1 deb
|
libuuid1 2.39.3-9ubuntu6.5 deb
|
||||||
libudev1 249.11-0ubuntu3.12 deb
|
libxxhash0 0.8.2-2build1 deb
|
||||||
libunistring2 1.0-1 deb
|
libzstd1 1.5.5+dfsg2-2build1.1 deb
|
||||||
libuuid1 2.37.2-4ubuntu3.4 deb
|
locales 2.39-0ubuntu8.7 deb
|
||||||
libxxhash0 0.8.1-1 deb
|
login 1:4.13+dfsg1-4ubuntu3.2 deb
|
||||||
libzstd1 1.4.8+dfsg-3build1 deb
|
logsave 1.47.0-2.4~exp1ubuntu4.1 deb
|
||||||
locales 2.35-0ubuntu3.8 deb
|
mawk 1.3.4.20240123-1build1 deb
|
||||||
login 1:4.8.1-2ubuntu2.2 deb
|
mount 2.39.3-9ubuntu6.5 deb
|
||||||
logsave 1.46.5-2ubuntu1.1 deb
|
ncurses-base 6.4+20240113-1ubuntu2 deb
|
||||||
lsb-base 11.1.0ubuntu4 deb
|
ncurses-bin 6.4+20240113-1ubuntu2 deb
|
||||||
mawk 1.3.4.20200120-3 deb
|
netcat-openbsd 1.226-1ubuntu2 deb
|
||||||
mount 2.37.2-4ubuntu3.4 deb
|
openssl 3.0.13-0ubuntu3.9 deb
|
||||||
ncurses-base 6.3-2ubuntu0.1 deb
|
passwd 1:4.13+dfsg1-4ubuntu3.2 deb
|
||||||
ncurses-bin 6.3-2ubuntu0.1 deb
|
perl-base 5.38.2-3.2ubuntu0.2 deb
|
||||||
netcat 1.218-4ubuntu1 deb
|
pinentry-curses 1.2.1-3ubuntu5 deb
|
||||||
netcat-openbsd 1.218-4ubuntu1 deb
|
plexmediaserver 1.43.1.10611-1e34174b1 deb
|
||||||
openssl 3.0.2-0ubuntu1.16 deb
|
procps 2:4.0.4-4ubuntu3.2 deb
|
||||||
passwd 1:4.8.1-2ubuntu2.2 deb
|
publicsuffix 20231001.0357-0.1 deb
|
||||||
perl-base 5.34.0-3ubuntu1.3 deb
|
readline-common 8.2-4build1 deb
|
||||||
pinentry-curses 1.1.1-1build2 deb
|
sed 4.9-2build1 deb
|
||||||
plexmediaserver 1.40.3.8555-fef15d30c deb
|
sensible-utils 0.0.22 deb
|
||||||
procps 2:3.3.17-6ubuntu2.1 deb
|
systemd-dev 255.4-1ubuntu8.15 deb
|
||||||
publicsuffix 20211207.1025-1 deb
|
systemd-hwe-hwdb 255.1.7 deb
|
||||||
readline-common 8.1.2-1 deb
|
systemd-standalone-sysusers 255.4-1ubuntu8.15 deb
|
||||||
sed 4.8-1ubuntu2 deb
|
sysvinit-utils 3.08-6ubuntu3 deb
|
||||||
sensible-utils 0.0.17 deb
|
tar 1.35+dfsg-3build1 deb
|
||||||
systemd-hwe-hwdb 249.11.5 deb
|
tzdata 2026a-0ubuntu0.24.04.1 deb
|
||||||
sysvinit-utils 3.01-1ubuntu1 deb
|
ubuntu-keyring 2023.11.28.1 deb
|
||||||
tar 1.34+dfsg-1ubuntu0.1.22.04.2 deb
|
udev 255.4-1ubuntu8.15 deb
|
||||||
tzdata 2024a-0ubuntu0.22.04 deb
|
unminimize 0.2.1 deb
|
||||||
ubuntu-keyring 2021.03.26 deb
|
util-linux 2.39.3-9ubuntu6.5 deb
|
||||||
udev 249.11-0ubuntu3.12 deb
|
wget 1.21.4-1ubuntu4.1 deb
|
||||||
usrmerge 25ubuntu2 deb
|
zlib1g 1:1.3.dfsg-3.1ubuntu2.1 deb
|
||||||
util-linux 2.37.2-4ubuntu3.4 deb
|
|
||||||
wget 1.21.2-2ubuntu1.1 deb
|
|
||||||
zlib1g 1:1.2.11.dfsg-2ubuntu9.2 deb
|
|
||||||
|
|
|
||||||
|
|
@ -3,24 +3,22 @@
|
||||||
# project information
|
# project information
|
||||||
project_name: plex
|
project_name: plex
|
||||||
project_url: "https://plex.tv"
|
project_url: "https://plex.tv"
|
||||||
project_logo: "http://the-gadgeteer.com/wp-content/uploads/2015/10/plex-logo-e1446990678679.png"
|
project_logo: "https://raw.githubusercontent.com/linuxserver/docker-templates/master/linuxserver.io/img/plex-logo.png"
|
||||||
project_blurb: "[{{ project_name|capitalize }}]({{ project_url }}) organizes video, music and photos from personal media libraries and streams them to smart TVs, streaming boxes and mobile devices. This container is packaged as a standalone Plex Media Server. Straightforward design and bulk actions mean getting things done faster."
|
project_blurb: "[{{ project_name|capitalize }}]({{ project_url }}) organizes video, music and photos from personal media libraries and streams them to smart TVs, streaming boxes and mobile devices. This container is packaged as a standalone Plex Media Server. Straightforward design and bulk actions mean getting things done faster."
|
||||||
project_lsio_github_repo_url: "https://github.com/linuxserver/docker-{{ project_name }}"
|
project_lsio_github_repo_url: "https://github.com/linuxserver/docker-{{ project_name }}"
|
||||||
|
project_categories: "Media Servers,Music,Audiobooks"
|
||||||
# supported architectures
|
# supported architectures
|
||||||
available_architectures:
|
available_architectures:
|
||||||
- {arch: "{{ arch_x86_64 }}", tag: "amd64-latest"}
|
- {arch: "{{ arch_x86_64 }}", tag: "amd64-latest"}
|
||||||
- {arch: "{{ arch_arm64 }}", tag: "arm64v8-latest"}
|
- {arch: "{{ arch_arm64 }}", tag: "arm64v8-latest"}
|
||||||
# development version
|
|
||||||
development_versions: false
|
|
||||||
# container parameters
|
# container parameters
|
||||||
common_param_env_vars_enabled: true #PGID, PUID, etc
|
common_param_env_vars_enabled: true #PGID, PUID, etc
|
||||||
param_container_name: "{{ project_name }}"
|
param_container_name: "{{ project_name }}"
|
||||||
param_usage_include_vols: true
|
param_usage_include_vols: true
|
||||||
param_volumes:
|
param_volumes:
|
||||||
- {vol_path: "/config", vol_host_path: "/path/to/library", desc: "Plex library location. *This can grow very large, 50gb+ is likely for a large collection.*"}
|
- {vol_path: "/config", vol_host_path: "/path/to/{{ project_name }}/library", desc: "Plex library location. *This can grow very large, 50gb+ is likely for a large collection.*"}
|
||||||
- {vol_path: "/tv", vol_host_path: "/path/to/tvseries", desc: "Media goes here. Add as many as needed e.g. `/movies`, `/tv`, etc."}
|
- {vol_path: "/tv", vol_host_path: "/path/to/tvseries", desc: "Media goes here. Add as many as needed e.g. `/movies`, `/tv`, etc."}
|
||||||
- {vol_path: "/movies", vol_host_path: "/path/to/movies", desc: "Media goes here. Add as many as needed e.g. `/movies`, `/tv`, etc."}
|
- {vol_path: "/movies", vol_host_path: "/path/to/movies", desc: "Media goes here. Add as many as needed e.g. `/movies`, `/tv`, etc."}
|
||||||
param_usage_include_ports: false
|
|
||||||
param_usage_include_net: true
|
param_usage_include_net: true
|
||||||
param_net: "host"
|
param_net: "host"
|
||||||
param_net_desc: "Use Host Networking"
|
param_net_desc: "Use Host Networking"
|
||||||
|
|
@ -73,10 +71,64 @@ app_setup_block: |
|
||||||
+ **`latest`**: will update plex to the latest version available that you are entitled to.
|
+ **`latest`**: will update plex to the latest version available that you are entitled to.
|
||||||
+ **`public`**: will update plexpass users to the latest public version, useful for plexpass users that don't want to be on the bleeding edge but still want the latest public updates.
|
+ **`public`**: will update plexpass users to the latest public version, useful for plexpass users that don't want to be on the bleeding edge but still want the latest public updates.
|
||||||
+ **`<specific-version>`**: will select a specific version (eg 0.9.12.4.1192-9a47d21) of plex to install, note you cannot use this to access plexpass versions if you do not have plexpass.
|
+ **`<specific-version>`**: will select a specific version (eg 0.9.12.4.1192-9a47d21) of plex to install, note you cannot use this to access plexpass versions if you do not have plexpass.
|
||||||
|
|
||||||
readme_hwaccel: true
|
readme_hwaccel: true
|
||||||
|
readonly_supported: true
|
||||||
|
readonly_message: |
|
||||||
|
* Runtime update of Plex (and thus Plexpass builds) is not supported.
|
||||||
|
* Transcode directory must be mounted to a host path or tmpfs.
|
||||||
|
nonroot_supported: true
|
||||||
|
nonroot_message: |
|
||||||
|
* Runtime update of Plex (and thus Plexpass builds) is not supported.
|
||||||
|
* Transcode directory must be mounted to a host path or tmpfs.
|
||||||
|
# init diagram
|
||||||
|
init_diagram: |
|
||||||
|
"plex:latest": {
|
||||||
|
docker-mods
|
||||||
|
base {
|
||||||
|
fix-attr +\nlegacy cont-init
|
||||||
|
}
|
||||||
|
docker-mods -> base
|
||||||
|
legacy-services
|
||||||
|
custom services
|
||||||
|
init-services -> legacy-services
|
||||||
|
init-services -> custom services
|
||||||
|
custom services -> legacy-services
|
||||||
|
legacy-services -> ci-service-check
|
||||||
|
init-migrations -> init-adduser
|
||||||
|
init-os-end -> init-config
|
||||||
|
init-config -> init-config-end
|
||||||
|
init-crontab-config -> init-config-end
|
||||||
|
init-plex-update -> init-config-end
|
||||||
|
init-config -> init-crontab-config
|
||||||
|
init-mods-end -> init-custom-files
|
||||||
|
init-adduser -> init-device-perms
|
||||||
|
base -> init-envfile
|
||||||
|
base -> init-migrations
|
||||||
|
init-config-end -> init-mods
|
||||||
|
init-mods-package-install -> init-mods-end
|
||||||
|
init-mods -> init-mods-package-install
|
||||||
|
init-adduser -> init-os-end
|
||||||
|
init-device-perms -> init-os-end
|
||||||
|
init-envfile -> init-os-end
|
||||||
|
init-config -> init-plex-chown
|
||||||
|
init-plex-chown -> init-plex-claim
|
||||||
|
init-plex-claim -> init-plex-update
|
||||||
|
init-custom-files -> init-services
|
||||||
|
init-services -> svc-cron
|
||||||
|
svc-cron -> legacy-services
|
||||||
|
init-services -> svc-plex
|
||||||
|
svc-plex -> legacy-services
|
||||||
|
}
|
||||||
|
Base Images: {
|
||||||
|
"baseimage-ubuntu:noble"
|
||||||
|
}
|
||||||
|
"plex:latest" <- Base Images
|
||||||
# changelog
|
# changelog
|
||||||
changelogs:
|
changelogs:
|
||||||
|
- {date: "15.03.26:", desc: "Allow TMPDIR to be changed to better support read-only containers"}
|
||||||
|
- {date: "15.03.26:", desc: "Fix initial claim setup on non-root containers"}
|
||||||
|
- {date: "04.11.24:", desc: "Add Nvidia capability needed for h265"}
|
||||||
|
- {date: "18.07.24:", desc: "Rebase to Ubuntu Noble."}
|
||||||
- {date: "12.02.24:", desc: "Use universal hardware acceleration blurb"}
|
- {date: "12.02.24:", desc: "Use universal hardware acceleration blurb"}
|
||||||
- {date: "09.01.24:", desc: "Set ownership on TranscoderTempDirectory when it's been saved in Preferences."}
|
- {date: "09.01.24:", desc: "Set ownership on TranscoderTempDirectory when it's been saved in Preferences."}
|
||||||
- {date: "16.08.23:", desc: "Install unrar from [linuxserver repo](https://github.com/linuxserver/docker-unrar)."}
|
- {date: "16.08.23:", desc: "Install unrar from [linuxserver repo](https://github.com/linuxserver/docker-unrar)."}
|
||||||
|
|
|
||||||
|
|
@ -1,19 +1,18 @@
|
||||||
#!/usr/bin/with-contenv bash
|
#!/usr/bin/with-contenv bash
|
||||||
# shellcheck shell=bash
|
# shellcheck shell=bash
|
||||||
|
|
||||||
|
# remove plex temporary directory after unclean stop
|
||||||
|
if [[ -d "${TMPDIR}" ]]; then
|
||||||
|
rm -rf "${TMPDIR}"
|
||||||
|
fi
|
||||||
|
mkdir -p ${TMPDIR}
|
||||||
|
|
||||||
# create folders
|
# create folders
|
||||||
if [[ ! -d "${PLEX_MEDIA_SERVER_APPLICATION_SUPPORT_DIR}" ]]; then
|
if [[ ! -d "${PLEX_MEDIA_SERVER_APPLICATION_SUPPORT_DIR}" ]]; then
|
||||||
mkdir -p "${PLEX_MEDIA_SERVER_APPLICATION_SUPPORT_DIR}"
|
mkdir -p "${PLEX_MEDIA_SERVER_APPLICATION_SUPPORT_DIR}"
|
||||||
lsiown -R abc:abc /config
|
if [[ -z ${LSIO_NON_ROOT_USER} ]]; then
|
||||||
fi
|
lsiown -R abc:abc /config
|
||||||
|
fi
|
||||||
# check Library permissions
|
|
||||||
PUID=${PUID:-911}
|
|
||||||
if [[ ! "$(stat -c %u /config/Library)" == "${PUID}" ]]; then
|
|
||||||
echo "Change in ownership detected, please be patient while we chown existing files"
|
|
||||||
echo "This could take some time"
|
|
||||||
lsiown abc:abc -R \
|
|
||||||
/config/Library
|
|
||||||
fi
|
fi
|
||||||
|
|
||||||
# remove plex pid after unclean stop
|
# remove plex pid after unclean stop
|
||||||
|
|
@ -21,19 +20,31 @@ if [[ -f "/config/Library/Application Support/Plex Media Server/plexmediaserver.
|
||||||
rm -f "/config/Library/Application Support/Plex Media Server/plexmediaserver.pid"
|
rm -f "/config/Library/Application Support/Plex Media Server/plexmediaserver.pid"
|
||||||
fi
|
fi
|
||||||
|
|
||||||
# set permissions on Plex Transcoder Temp Directory
|
if [[ -z ${LSIO_NON_ROOT_USER} ]]; then
|
||||||
PLEX_MEDIA_SERVER_PREFERENCES="${PLEX_MEDIA_SERVER_APPLICATION_SUPPORT_DIR}/Plex Media Server/Preferences.xml"
|
# check Library permissions
|
||||||
if [[ -f "${PLEX_MEDIA_SERVER_PREFERENCES}" ]]; then
|
PUID=${PUID:-911}
|
||||||
TranscoderTempDirectory='\bTranscoderTempDirectory="([^"]+)"'
|
if [[ ! "$(stat -c %u /config/Library)" == "${PUID}" ]]; then
|
||||||
while IFS= read -r line; do
|
echo "Change in ownership detected, please be patient while we chown existing files"
|
||||||
if [[ ${line} =~ ${TranscoderTempDirectory} ]] && [[ -d "${BASH_REMATCH[1]}" ]]; then
|
echo "This could take some time"
|
||||||
echo "Setting permissions on ${BASH_REMATCH[1]}"
|
lsiown -R abc:abc \
|
||||||
lsiown -R abc:abc "${BASH_REMATCH[1]}"
|
/config/Library
|
||||||
fi
|
fi
|
||||||
done <"${PLEX_MEDIA_SERVER_PREFERENCES}"
|
|
||||||
fi
|
|
||||||
|
|
||||||
# permissions (non-recursive) on config root and folders
|
# set permissions on Plex Transcoder Temp Directory
|
||||||
lsiown abc:abc \
|
PLEX_MEDIA_SERVER_PREFERENCES="${PLEX_MEDIA_SERVER_APPLICATION_SUPPORT_DIR}/Plex Media Server/Preferences.xml"
|
||||||
/config \
|
if [[ -f "${PLEX_MEDIA_SERVER_PREFERENCES}" ]]; then
|
||||||
/config/*
|
TranscoderTempDirectory='\bTranscoderTempDirectory="([^"]+)"'
|
||||||
|
while IFS= read -r line; do
|
||||||
|
if [[ ${line} =~ ${TranscoderTempDirectory} ]] && [[ -d "${BASH_REMATCH[1]}" ]]; then
|
||||||
|
echo "Setting permissions on ${BASH_REMATCH[1]}"
|
||||||
|
lsiown -R abc:abc "${BASH_REMATCH[1]}"
|
||||||
|
fi
|
||||||
|
done <"${PLEX_MEDIA_SERVER_PREFERENCES}"
|
||||||
|
fi
|
||||||
|
|
||||||
|
# permissions (non-recursive) on config root and folders
|
||||||
|
lsiown abc:abc \
|
||||||
|
"${TMPDIR}" \
|
||||||
|
/config \
|
||||||
|
/config/*
|
||||||
|
fi
|
||||||
|
|
|
||||||
|
|
@ -19,8 +19,15 @@ if [[ ! -f "${PLEX_MEDIA_SERVER_PREFERENCES}" ]]; then
|
||||||
export PLEX_MEDIA_SERVER_INFO_MODEL
|
export PLEX_MEDIA_SERVER_INFO_MODEL
|
||||||
PLEX_MEDIA_SERVER_INFO_PLATFORM_VERSION=$(uname -r)
|
PLEX_MEDIA_SERVER_INFO_PLATFORM_VERSION=$(uname -r)
|
||||||
export PLEX_MEDIA_SERVER_INFO_PLATFORM_VERSION
|
export PLEX_MEDIA_SERVER_INFO_PLATFORM_VERSION
|
||||||
s6-setuidgid abc /bin/bash -c \
|
|
||||||
'LD_LIBRARY_PATH=/usr/lib/plexmediaserver:/usr/lib/plexmediaserver/lib /usr/lib/plexmediaserver/Plex\ Media\ Server' &
|
if [[ -z ${LSIO_NON_ROOT_USER} ]]; then
|
||||||
|
s6-setuidgid abc /bin/bash -c \
|
||||||
|
'LD_LIBRARY_PATH=/usr/lib/plexmediaserver:/usr/lib/plexmediaserver/lib /usr/lib/plexmediaserver/Plex\ Media\ Server' &
|
||||||
|
else
|
||||||
|
/bin/bash -c \
|
||||||
|
'LD_LIBRARY_PATH=/usr/lib/plexmediaserver:/usr/lib/plexmediaserver/lib /usr/lib/plexmediaserver/Plex\ Media\ Server' &
|
||||||
|
fi
|
||||||
|
|
||||||
PID=$!
|
PID=$!
|
||||||
echo "Waiting for Plex to generate its config"
|
echo "Waiting for Plex to generate its config"
|
||||||
DBNAME="/config/Library/Application Support/Plex Media Server/Plug-in Support/Databases/com.plexapp.plugins.library.db-wal"
|
DBNAME="/config/Library/Application Support/Plex Media Server/Plug-in Support/Databases/com.plexapp.plugins.library.db-wal"
|
||||||
|
|
|
||||||
|
|
@ -1,35 +0,0 @@
|
||||||
#!/usr/bin/with-contenv bash
|
|
||||||
# shellcheck shell=bash
|
|
||||||
|
|
||||||
FILES=$(find /dev/dri /dev/dvb -type c -print 2>/dev/null)
|
|
||||||
|
|
||||||
for i in ${FILES}; do
|
|
||||||
VIDEO_GID=$(stat -c '%g' "${i}")
|
|
||||||
VIDEO_UID=$(stat -c '%u' "${i}")
|
|
||||||
# check if user matches device
|
|
||||||
if id -u abc | grep -qw "${VIDEO_UID}"; then
|
|
||||||
echo "**** permissions for ${i} are good ****"
|
|
||||||
else
|
|
||||||
# check if group matches and that device has group rw
|
|
||||||
if id -G abc | grep -qw "${VIDEO_GID}" && [[ $(stat -c '%A' "${i}" | cut -b 5,6) == "rw" ]]; then
|
|
||||||
echo "**** permissions for ${i} are good ****"
|
|
||||||
# check if device needs to be added to video group
|
|
||||||
elif ! id -G abc | grep -qw "${VIDEO_GID}"; then
|
|
||||||
# check if video group needs to be created
|
|
||||||
VIDEO_NAME=$(getent group "${VIDEO_GID}" | awk -F: '{print $1}')
|
|
||||||
if [[ -z "${VIDEO_NAME}" ]]; then
|
|
||||||
VIDEO_NAME="video$(head /dev/urandom | tr -dc 'a-z0-9' | head -c4)"
|
|
||||||
groupadd "${VIDEO_NAME}"
|
|
||||||
groupmod -g "${VIDEO_GID}" "${VIDEO_NAME}"
|
|
||||||
echo "**** creating video group ${VIDEO_NAME} with id ${VIDEO_GID} ****"
|
|
||||||
fi
|
|
||||||
echo "**** adding ${i} to video group ${VIDEO_NAME} with id ${VIDEO_GID} ****"
|
|
||||||
usermod -a -G "${VIDEO_NAME}" abc
|
|
||||||
fi
|
|
||||||
# check if device has group rw
|
|
||||||
if [[ $(stat -c '%A' "${i}" | cut -b 5,6) != "rw" ]]; then
|
|
||||||
echo -e "**** The device ${i} does not have group read/write permissions, attempting to fix inside the container. ****"
|
|
||||||
chmod g+rw "${i}"
|
|
||||||
fi
|
|
||||||
fi
|
|
||||||
done
|
|
||||||
|
|
@ -1 +0,0 @@
|
||||||
oneshot
|
|
||||||
|
|
@ -1 +0,0 @@
|
||||||
/etc/s6-overlay/s6-rc.d/init-plex-gid-video/run
|
|
||||||
|
|
@ -1,6 +1,11 @@
|
||||||
#!/usr/bin/with-contenv bash
|
#!/usr/bin/with-contenv bash
|
||||||
# shellcheck shell=bash
|
# shellcheck shell=bash
|
||||||
|
|
||||||
|
if [[ -n ${LSIO_READ_ONLY_FS} ]] || [[ -n ${LSIO_NON_ROOT_USER} ]]; then
|
||||||
|
echo "Runtime update not supported with read-only or non-root operation, skipping."
|
||||||
|
exit 0
|
||||||
|
fi
|
||||||
|
|
||||||
# If docker manages versioning exit
|
# If docker manages versioning exit
|
||||||
if [[ -n "${VERSION}" ]] && [[ "${VERSION}" == "docker" ]]; then
|
if [[ -n "${VERSION}" ]] && [[ "${VERSION}" == "docker" ]]; then
|
||||||
echo "Docker is used for versioning skip update check"
|
echo "Docker is used for versioning skip update check"
|
||||||
|
|
|
||||||
14
root/etc/s6-overlay/s6-rc.d/init-wireguard/run
Executable file
14
root/etc/s6-overlay/s6-rc.d/init-wireguard/run
Executable file
|
|
@ -0,0 +1,14 @@
|
||||||
|
#!/usr/bin/with-contenv bash
|
||||||
|
|
||||||
|
# Bring up WireGuard if not already up
|
||||||
|
if ! ip link show dev wg0 > /dev/null 2>&1; then
|
||||||
|
wg-quick up wg0
|
||||||
|
fi
|
||||||
|
|
||||||
|
# Monitor wireguard status
|
||||||
|
while true; do
|
||||||
|
if ! ip link show dev wg0 | grep -q "UP"; then
|
||||||
|
wg-quick up wg0
|
||||||
|
fi
|
||||||
|
sleep 60
|
||||||
|
done
|
||||||
1
root/etc/s6-overlay/s6-rc.d/init-wireguard/type
Normal file
1
root/etc/s6-overlay/s6-rc.d/init-wireguard/type
Normal file
|
|
@ -0,0 +1 @@
|
||||||
|
longrun
|
||||||
|
|
@ -1,11 +1,20 @@
|
||||||
#!/usr/bin/with-contenv bash
|
#!/usr/bin/with-contenv bash
|
||||||
# shellcheck shell=bash
|
# shellcheck shell=bash
|
||||||
|
|
||||||
|
# Wait for WireGuard to be up
|
||||||
|
|
||||||
echo "Starting Plex Media Server. . . (you can ignore the libusb_init error)"
|
echo "Starting Plex Media Server. . . (you can ignore the libusb_init error)"
|
||||||
PLEX_MEDIA_SERVER_INFO_MODEL=$(uname -m)
|
PLEX_MEDIA_SERVER_INFO_MODEL=$(uname -m)
|
||||||
export PLEX_MEDIA_SERVER_INFO_MODEL
|
export PLEX_MEDIA_SERVER_INFO_MODEL
|
||||||
PLEX_MEDIA_SERVER_INFO_PLATFORM_VERSION=$(uname -r)
|
PLEX_MEDIA_SERVER_INFO_PLATFORM_VERSION=$(uname -r)
|
||||||
export PLEX_MEDIA_SERVER_INFO_PLATFORM_VERSION
|
export PLEX_MEDIA_SERVER_INFO_PLATFORM_VERSION
|
||||||
exec \
|
|
||||||
s6-notifyoncheck -d -n 300 -w 1000 -c "nc -z localhost 32400" \
|
if [[ -z ${LSIO_NON_ROOT_USER} ]]; then
|
||||||
s6-setuidgid abc "/usr/lib/plexmediaserver/Plex Media Server"
|
exec \
|
||||||
|
s6-notifyoncheck -d -n 300 -w 1000 -c "nc -z localhost 32400" \
|
||||||
|
s6-setuidgid abc "/usr/lib/plexmediaserver/Plex Media Server"
|
||||||
|
else
|
||||||
|
exec \
|
||||||
|
s6-notifyoncheck -d -n 300 -w 1000 -c "nc -z localhost 32400" \
|
||||||
|
"/usr/lib/plexmediaserver/Plex Media Server"
|
||||||
|
fi
|
||||||
|
|
|
||||||
Loading…
Add table
Add a link
Reference in a new issue