Compare commits
No commits in common. "release" and "2.5.1" have entirely different histories.
|
|
@ -1,10 +1,10 @@
|
||||||
.git*
|
.git
|
||||||
|
.github
|
||||||
|
.travis.yml
|
||||||
*.md
|
*.md
|
||||||
build*
|
|
||||||
docker-compose*
|
|
||||||
env
|
env
|
||||||
test-configuration
|
build*
|
||||||
|
docker-compose.override.yml
|
||||||
.netbox/.git*
|
.netbox/.git*
|
||||||
.netbox/contrib
|
.netbox/.travis.yml
|
||||||
.netbox/scripts
|
.netbox/scripts
|
||||||
.netbox/upgrade.sh
|
|
||||||
|
|
|
||||||
|
|
@ -6,7 +6,7 @@ contact_links:
|
||||||
|
|
||||||
- name: Chat
|
- name: Chat
|
||||||
url: https://join.slack.com/t/netdev-community/shared_invite/zt-mtts8g0n-Sm6Wutn62q_M4OdsaIycrQ
|
url: https://join.slack.com/t/netdev-community/shared_invite/zt-mtts8g0n-Sm6Wutn62q_M4OdsaIycrQ
|
||||||
about: "Usually the quickest way to seek help with small issues is to join our #netbox-docker Slack channel."
|
about: 'Usually the quickest way to seek help with small issues is to join our #netbox-docker Slack channel.'
|
||||||
|
|
||||||
- name: Community Wiki
|
- name: Community Wiki
|
||||||
url: https://github.com/netbox-community/netbox-docker/wiki
|
url: https://github.com/netbox-community/netbox-docker/wiki
|
||||||
|
|
|
||||||
|
|
@ -57,7 +57,7 @@ body:
|
||||||
- type: textarea
|
- type: textarea
|
||||||
id: discussion
|
id: discussion
|
||||||
attributes:
|
attributes:
|
||||||
label: "Discussion: Benefits and Drawbacks"
|
label: 'Discussion: Benefits and Drawbacks'
|
||||||
description: |
|
description: |
|
||||||
Please make your case here:
|
Please make your case here:
|
||||||
- Why do you think this project and the community will benefit from your suggestion?
|
- Why do you think this project and the community will benefit from your suggestion?
|
||||||
|
|
|
||||||
|
|
@ -80,6 +80,6 @@ into the release notes.
|
||||||
Please put an x into the brackets (like `[x]`) if you've completed that task.
|
Please put an x into the brackets (like `[x]`) if you've completed that task.
|
||||||
-->
|
-->
|
||||||
|
|
||||||
- [ ] I have read the comments and followed the PR template.
|
* [ ] I have read the comments and followed the PR template.
|
||||||
- [ ] I have explained my PR according to the information in the comments.
|
* [ ] I have explained my PR according to the information in the comments.
|
||||||
- [ ] My PR targets the `develop` branch.
|
* [ ] My PR targets the `develop` branch.
|
||||||
|
|
|
||||||
|
|
@ -5,40 +5,33 @@ on:
|
||||||
push:
|
push:
|
||||||
branches-ignore:
|
branches-ignore:
|
||||||
- release
|
- release
|
||||||
- renovate/**
|
|
||||||
pull_request:
|
pull_request:
|
||||||
branches-ignore:
|
branches-ignore:
|
||||||
- release
|
- release
|
||||||
|
|
||||||
concurrency:
|
|
||||||
group: ${{ github.workflow }}-${{ github.event.pull_request.number || github.ref }}
|
|
||||||
cancel-in-progress: true
|
|
||||||
|
|
||||||
jobs:
|
jobs:
|
||||||
lint:
|
lint:
|
||||||
runs-on: ubuntu-latest
|
runs-on: ubuntu-latest
|
||||||
name: Checks syntax of our code
|
name: Checks syntax of our code
|
||||||
steps:
|
steps:
|
||||||
- uses: actions/checkout@v4
|
- uses: actions/checkout@v3
|
||||||
with:
|
with:
|
||||||
# Full git history is needed to get a proper
|
# Full git history is needed to get a proper
|
||||||
# list of changed files within `super-linter`
|
# list of changed files within `super-linter`
|
||||||
fetch-depth: 0
|
fetch-depth: 0
|
||||||
- uses: actions/setup-python@v5
|
- uses: actions/setup-python@v4
|
||||||
with:
|
with:
|
||||||
python-version: "3.9"
|
python-version: '3.9'
|
||||||
- name: Lint Code Base
|
- name: Lint Code Base
|
||||||
uses: github/super-linter@v7
|
uses: github/super-linter@v4
|
||||||
env:
|
env:
|
||||||
DEFAULT_BRANCH: develop
|
DEFAULT_BRANCH: develop
|
||||||
GITHUB_TOKEN: ${{ secrets.GITHUB_TOKEN }}
|
GITHUB_TOKEN: ${{ secrets.GITHUB_TOKEN }}
|
||||||
SUPPRESS_POSSUM: true
|
SUPPRESS_POSSUM: true
|
||||||
LINTER_RULES_PATH: /
|
LINTER_RULES_PATH: /
|
||||||
VALIDATE_ALL_CODEBASE: false
|
VALIDATE_ALL_CODEBASE: false
|
||||||
VALIDATE_CHECKOV: false
|
|
||||||
VALIDATE_DOCKERFILE: false
|
VALIDATE_DOCKERFILE: false
|
||||||
VALIDATE_GITLEAKS: false
|
VALIDATE_GITLEAKS: false
|
||||||
VALIDATE_JSCPD: false
|
|
||||||
FILTER_REGEX_EXCLUDE: (.*/)?(LICENSE|configuration/.*)
|
FILTER_REGEX_EXCLUDE: (.*/)?(LICENSE|configuration/.*)
|
||||||
EDITORCONFIG_FILE_NAME: .ecrc
|
EDITORCONFIG_FILE_NAME: .ecrc
|
||||||
DOCKERFILE_HADOLINT_FILE_NAME: .hadolint.yaml
|
DOCKERFILE_HADOLINT_FILE_NAME: .hadolint.yaml
|
||||||
|
|
@ -56,38 +49,32 @@ jobs:
|
||||||
- PRERELEASE=true ./build-latest.sh
|
- PRERELEASE=true ./build-latest.sh
|
||||||
- ./build.sh feature
|
- ./build.sh feature
|
||||||
- ./build.sh develop
|
- ./build.sh develop
|
||||||
os:
|
platform:
|
||||||
- ubuntu-latest
|
- linux/amd64
|
||||||
- self-hosted
|
- linux/arm64
|
||||||
fail-fast: false
|
fail-fast: false
|
||||||
env:
|
env:
|
||||||
GH_ACTION: enable
|
GH_ACTION: enable
|
||||||
GITHUB_TOKEN: ${{ secrets.GITHUB_TOKEN }}
|
GITHUB_TOKEN: ${{ secrets.GITHUB_TOKEN }}
|
||||||
IMAGE_NAMES: docker.io/netboxcommunity/netbox
|
IMAGE_NAMES: docker.io/netboxcommunity/netbox
|
||||||
runs-on: ${{ matrix.os }}
|
runs-on: ubuntu-latest
|
||||||
name: Builds new NetBox Docker Images
|
name: Builds new NetBox Docker Images
|
||||||
steps:
|
steps:
|
||||||
- id: git-checkout
|
- id: git-checkout
|
||||||
name: Checkout
|
name: Checkout
|
||||||
uses: actions/checkout@v4
|
uses: actions/checkout@v3
|
||||||
|
- id: qemu-setup
|
||||||
|
name: Set up QEMU
|
||||||
|
uses: docker/setup-qemu-action@v2
|
||||||
- id: buildx-setup
|
- id: buildx-setup
|
||||||
name: Set up Docker Buildx
|
name: Set up Docker Buildx
|
||||||
uses: docker/setup-buildx-action@v3
|
uses: docker/setup-buildx-action@v2
|
||||||
- id: arm-buildx-platform
|
|
||||||
name: Set BUILDX_PLATFORM to ARM64
|
|
||||||
if: matrix.os == 'self-hosted'
|
|
||||||
run: |
|
|
||||||
echo "BUILDX_PLATFORM=linux/arm64" >>"${GITHUB_ENV}"
|
|
||||||
- id: docker-build
|
- id: docker-build
|
||||||
name: Build the image for '${{ matrix.os }}' with '${{ matrix.build_cmd }}'
|
name: Build the image for '${{ matrix.platform }}' with '${{ matrix.build_cmd }}'
|
||||||
run: ${{ matrix.build_cmd }}
|
run: ${{ matrix.build_cmd }}
|
||||||
env:
|
env:
|
||||||
|
BUILDX_PLATFORM: ${{ matrix.platform }}
|
||||||
BUILDX_BUILDER_NAME: ${{ steps.buildx-setup.outputs.name }}
|
BUILDX_BUILDER_NAME: ${{ steps.buildx-setup.outputs.name }}
|
||||||
- id: arm-time-limit
|
|
||||||
name: Set Netbox container start_period higher on ARM64
|
|
||||||
if: matrix.os == 'self-hosted'
|
|
||||||
run: |
|
|
||||||
echo "NETBOX_START_PERIOD=240s" >>"${GITHUB_ENV}"
|
|
||||||
- id: docker-test
|
- id: docker-test
|
||||||
name: Test the image
|
name: Test the image
|
||||||
run: IMAGE="${FINAL_DOCKER_TAG}" ./test.sh
|
run: IMAGE="${FINAL_DOCKER_TAG}" ./test.sh
|
||||||
|
|
|
||||||
|
|
@ -6,7 +6,7 @@ on:
|
||||||
types:
|
types:
|
||||||
- published
|
- published
|
||||||
schedule:
|
schedule:
|
||||||
- cron: "45 5 * * *"
|
- cron: '45 5 * * *'
|
||||||
workflow_dispatch:
|
workflow_dispatch:
|
||||||
|
|
||||||
jobs:
|
jobs:
|
||||||
|
|
@ -30,55 +30,55 @@ jobs:
|
||||||
steps:
|
steps:
|
||||||
- id: source-checkout
|
- id: source-checkout
|
||||||
name: Checkout
|
name: Checkout
|
||||||
uses: actions/checkout@v4
|
uses: actions/checkout@v3
|
||||||
- id: set-netbox-docker-version
|
- id: set-netbox-docker-version
|
||||||
name: Get Version of NetBox Docker
|
name: Get Version of NetBox Docker
|
||||||
run: echo "version=$(cat VERSION)" >>"$GITHUB_OUTPUT"
|
run: echo "version=$(cat VERSION)" >>"$GITHUB_OUTPUT"
|
||||||
shell: bash
|
shell: bash
|
||||||
- id: check-build-needed
|
- id: qemu-setup
|
||||||
name: Check if the build is needed for '${{ matrix.build_cmd }}'
|
name: Set up QEMU
|
||||||
env:
|
uses: docker/setup-qemu-action@v2
|
||||||
CHECK_ONLY: "true"
|
- id: buildx-setup
|
||||||
|
name: Set up Docker Buildx
|
||||||
|
uses: docker/setup-buildx-action@v2
|
||||||
|
- id: docker-build
|
||||||
|
name: Build the image with '${{ matrix.build_cmd }}'
|
||||||
run: ${{ matrix.build_cmd }}
|
run: ${{ matrix.build_cmd }}
|
||||||
|
- id: test-image
|
||||||
|
name: Test the image
|
||||||
|
run: IMAGE="${FINAL_DOCKER_TAG}" ./test.sh
|
||||||
|
if: steps.docker-build.outputs.skipped != 'true'
|
||||||
# docker.io
|
# docker.io
|
||||||
- id: docker-io-login
|
- id: docker-io-login
|
||||||
name: Login to docker.io
|
name: Login to docker.io
|
||||||
uses: docker/login-action@v3
|
uses: docker/login-action@v2
|
||||||
with:
|
with:
|
||||||
registry: docker.io
|
registry: docker.io
|
||||||
username: ${{ secrets.dockerhub_username }}
|
username: ${{ secrets.dockerhub_username }}
|
||||||
password: ${{ secrets.dockerhub_password }}
|
password: ${{ secrets.dockerhub_password }}
|
||||||
if: steps.check-build-needed.outputs.skipped != 'true'
|
if: steps.docker-build.outputs.skipped != 'true'
|
||||||
- id: buildx-setup
|
|
||||||
name: Set up Docker Buildx
|
|
||||||
uses: docker/setup-buildx-action@v3
|
|
||||||
with:
|
|
||||||
version: "lab:latest"
|
|
||||||
driver: cloud
|
|
||||||
endpoint: "netboxcommunity/netbox-default"
|
|
||||||
if: steps.check-build-needed.outputs.skipped != 'true'
|
|
||||||
# quay.io
|
# quay.io
|
||||||
- id: quay-io-login
|
- id: quay-io-login
|
||||||
name: Login to Quay.io
|
name: Login to Quay.io
|
||||||
uses: docker/login-action@v3
|
uses: docker/login-action@v2
|
||||||
with:
|
with:
|
||||||
registry: quay.io
|
registry: quay.io
|
||||||
username: ${{ secrets.quayio_username }}
|
username: ${{ secrets.quayio_username }}
|
||||||
password: ${{ secrets.quayio_password }}
|
password: ${{ secrets.quayio_password }}
|
||||||
if: steps.check-build-needed.outputs.skipped != 'true'
|
if: steps.docker-build.outputs.skipped != 'true'
|
||||||
# ghcr.io
|
# ghcr.io
|
||||||
- id: ghcr-io-login
|
- id: ghcr-io-login
|
||||||
name: Login to GitHub Container Registry
|
name: Login to GitHub Container Registry
|
||||||
uses: docker/login-action@v3
|
uses: docker/login-action@v2
|
||||||
with:
|
with:
|
||||||
registry: ghcr.io
|
registry: ghcr.io
|
||||||
username: ${{ github.repository_owner }}
|
username: ${{ github.repository_owner }}
|
||||||
password: ${{ secrets.GITHUB_TOKEN }}
|
password: ${{ secrets.GITHUB_TOKEN }}
|
||||||
if: steps.check-build-needed.outputs.skipped != 'true'
|
if: steps.docker-build.outputs.skipped != 'true'
|
||||||
- id: build-and-push
|
- id: build-and-push
|
||||||
name: Push the image
|
name: Push the image
|
||||||
run: ${{ matrix.build_cmd }} --push
|
run: ${{ matrix.build_cmd }} --push
|
||||||
if: steps.check-build-needed.outputs.skipped != 'true'
|
if: steps.docker-build.outputs.skipped != 'true'
|
||||||
env:
|
env:
|
||||||
BUILDX_PLATFORM: ${{ matrix.platform }}
|
BUILDX_PLATFORM: ${{ matrix.platform }}
|
||||||
BUILDX_BUILDER_NAME: ${{ steps.buildx-setup.outputs.name }}
|
BUILDX_BUILDER_NAME: ${{ steps.buildx-setup.outputs.name }}
|
||||||
|
|
|
||||||
|
|
@ -1,6 +1,7 @@
|
||||||
*.sql.gz
|
*.sql.gz
|
||||||
.netbox
|
.netbox
|
||||||
.python-version
|
.initializers
|
||||||
|
docker-compose.override.yml
|
||||||
*.pem
|
*.pem
|
||||||
configuration/*
|
configuration/*
|
||||||
!configuration/configuration.py
|
!configuration/configuration.py
|
||||||
|
|
@ -10,4 +11,5 @@ configuration/ldap/*
|
||||||
!configuration/ldap/ldap_config.py
|
!configuration/ldap/ldap_config.py
|
||||||
!configuration/logging.py
|
!configuration/logging.py
|
||||||
!configuration/plugins.py
|
!configuration/plugins.py
|
||||||
|
prometheus.yml
|
||||||
super-linter.log
|
super-linter.log
|
||||||
|
|
|
||||||
|
|
@ -1,4 +1,5 @@
|
||||||
---
|
---
|
||||||
|
|
||||||
rules:
|
rules:
|
||||||
line-length:
|
line-length:
|
||||||
max: 160
|
max: 120
|
||||||
|
|
|
||||||
35
Dockerfile
35
Dockerfile
|
|
@ -1,5 +1,5 @@
|
||||||
ARG FROM
|
ARG FROM
|
||||||
FROM ${FROM} AS builder
|
FROM ${FROM} as builder
|
||||||
|
|
||||||
RUN export DEBIAN_FRONTEND=noninteractive \
|
RUN export DEBIAN_FRONTEND=noninteractive \
|
||||||
&& apt-get update -qq \
|
&& apt-get update -qq \
|
||||||
|
|
@ -14,6 +14,7 @@ RUN export DEBIAN_FRONTEND=noninteractive \
|
||||||
libsasl2-dev \
|
libsasl2-dev \
|
||||||
libssl-dev \
|
libssl-dev \
|
||||||
libxml2-dev \
|
libxml2-dev \
|
||||||
|
libxml2-dev \
|
||||||
libxmlsec1 \
|
libxmlsec1 \
|
||||||
libxmlsec1-dev \
|
libxmlsec1-dev \
|
||||||
libxmlsec1-openssl \
|
libxmlsec1-openssl \
|
||||||
|
|
@ -30,13 +31,7 @@ RUN export DEBIAN_FRONTEND=noninteractive \
|
||||||
|
|
||||||
ARG NETBOX_PATH
|
ARG NETBOX_PATH
|
||||||
COPY ${NETBOX_PATH}/requirements.txt requirements-container.txt /
|
COPY ${NETBOX_PATH}/requirements.txt requirements-container.txt /
|
||||||
RUN \
|
RUN sed -i -e '/psycopg2-binary/d' requirements.txt && \
|
||||||
# Gunicorn is not needed because we use Nginx Unit
|
|
||||||
sed -i -e '/gunicorn/d' /requirements.txt && \
|
|
||||||
# We need 'social-auth-core[all]' in the Docker image. But if we put it in our own requirements-container.txt
|
|
||||||
# we have potential version conflicts and the build will fail.
|
|
||||||
# That's why we just replace it in the original requirements.txt.
|
|
||||||
sed -i -e 's/social-auth-core/social-auth-core\[all\]/g' /requirements.txt && \
|
|
||||||
/opt/netbox/venv/bin/pip install \
|
/opt/netbox/venv/bin/pip install \
|
||||||
-r /requirements.txt \
|
-r /requirements.txt \
|
||||||
-r /requirements-container.txt
|
-r /requirements-container.txt
|
||||||
|
|
@ -46,7 +41,7 @@ RUN \
|
||||||
###
|
###
|
||||||
|
|
||||||
ARG FROM
|
ARG FROM
|
||||||
FROM ${FROM} AS main
|
FROM ${FROM} as main
|
||||||
|
|
||||||
RUN export DEBIAN_FRONTEND=noninteractive \
|
RUN export DEBIAN_FRONTEND=noninteractive \
|
||||||
&& apt-get update -qq \
|
&& apt-get update -qq \
|
||||||
|
|
@ -60,27 +55,25 @@ RUN export DEBIAN_FRONTEND=noninteractive \
|
||||||
libldap-common \
|
libldap-common \
|
||||||
libpq5 \
|
libpq5 \
|
||||||
libxmlsec1-openssl \
|
libxmlsec1-openssl \
|
||||||
openssh-client \
|
|
||||||
openssl \
|
openssl \
|
||||||
python3 \
|
python3 \
|
||||||
|
python3-distutils \
|
||||||
tini \
|
tini \
|
||||||
&& curl --silent --output /usr/share/keyrings/nginx-keyring.gpg \
|
&& curl -sL https://nginx.org/keys/nginx_signing.key \
|
||||||
https://unit.nginx.org/keys/nginx-keyring.gpg \
|
> /etc/apt/trusted.gpg.d/nginx.asc && \
|
||||||
&& echo "deb [signed-by=/usr/share/keyrings/nginx-keyring.gpg] https://packages.nginx.org/unit/ubuntu/ noble unit" \
|
echo "deb https://packages.nginx.org/unit/ubuntu/ jammy unit" \
|
||||||
> /etc/apt/sources.list.d/unit.list \
|
> /etc/apt/sources.list.d/unit.list \
|
||||||
&& apt-get update -qq \
|
&& apt-get update -qq \
|
||||||
&& apt-get install \
|
&& apt-get install \
|
||||||
--yes -qq --no-install-recommends \
|
--yes -qq --no-install-recommends \
|
||||||
unit=1.33.0-1~noble \
|
unit=1.29.1-1~jammy \
|
||||||
unit-python3.12=1.33.0-1~noble \
|
unit-python3.10=1.29.1-1~jammy \
|
||||||
&& rm -rf /var/lib/apt/lists/*
|
&& rm -rf /var/lib/apt/lists/*
|
||||||
|
|
||||||
COPY --from=builder /opt/netbox/venv /opt/netbox/venv
|
COPY --from=builder /opt/netbox/venv /opt/netbox/venv
|
||||||
|
|
||||||
ARG NETBOX_PATH
|
ARG NETBOX_PATH
|
||||||
COPY ${NETBOX_PATH} /opt/netbox
|
COPY ${NETBOX_PATH} /opt/netbox
|
||||||
# Copy the modified 'requirements*.txt' files, to have the files actually used during installation
|
|
||||||
COPY --from=builder /requirements.txt /requirements-container.txt /opt/netbox/
|
|
||||||
|
|
||||||
COPY docker/configuration.docker.py /opt/netbox/netbox/netbox/configuration.py
|
COPY docker/configuration.docker.py /opt/netbox/netbox/netbox/configuration.py
|
||||||
COPY docker/ldap_config.docker.py /opt/netbox/netbox/netbox/ldap_config.py
|
COPY docker/ldap_config.docker.py /opt/netbox/netbox/netbox/ldap_config.py
|
||||||
|
|
@ -95,11 +88,11 @@ WORKDIR /opt/netbox/netbox
|
||||||
# Must set permissions for '/opt/netbox/netbox/media' directory
|
# Must set permissions for '/opt/netbox/netbox/media' directory
|
||||||
# to g+w so that pictures can be uploaded to netbox.
|
# to g+w so that pictures can be uploaded to netbox.
|
||||||
RUN mkdir -p static /opt/unit/state/ /opt/unit/tmp/ \
|
RUN mkdir -p static /opt/unit/state/ /opt/unit/tmp/ \
|
||||||
&& chown -R unit:root /opt/unit/ media reports scripts \
|
&& chown -R unit:root media /opt/unit/ \
|
||||||
&& chmod -R g+w /opt/unit/ media reports scripts \
|
&& chmod -R g+w media /opt/unit/ \
|
||||||
&& cd /opt/netbox/ && SECRET_KEY="dummyKeyWithMinimumLength-------------------------" /opt/netbox/venv/bin/python -m mkdocs build \
|
&& cd /opt/netbox/ && SECRET_KEY="dummy" /opt/netbox/venv/bin/python -m mkdocs build \
|
||||||
--config-file /opt/netbox/mkdocs.yml --site-dir /opt/netbox/netbox/project-static/docs/ \
|
--config-file /opt/netbox/mkdocs.yml --site-dir /opt/netbox/netbox/project-static/docs/ \
|
||||||
&& SECRET_KEY="dummyKeyWithMinimumLength-------------------------" /opt/netbox/venv/bin/python /opt/netbox/netbox/manage.py collectstatic --no-input
|
&& SECRET_KEY="dummy" /opt/netbox/venv/bin/python /opt/netbox/netbox/manage.py collectstatic --no-input
|
||||||
|
|
||||||
ENV LANG=C.utf8 PATH=/opt/netbox/venv/bin:$PATH
|
ENV LANG=C.utf8 PATH=/opt/netbox/venv/bin:$PATH
|
||||||
ENTRYPOINT [ "/usr/bin/tini", "--" ]
|
ENTRYPOINT [ "/usr/bin/tini", "--" ]
|
||||||
|
|
|
||||||
|
|
@ -34,6 +34,7 @@ There is a more complete [_Getting Started_ guide on our wiki][wiki-getting-star
|
||||||
git clone -b release https://github.com/netbox-community/netbox-docker.git
|
git clone -b release https://github.com/netbox-community/netbox-docker.git
|
||||||
cd netbox-docker
|
cd netbox-docker
|
||||||
tee docker-compose.override.yml <<EOF
|
tee docker-compose.override.yml <<EOF
|
||||||
|
version: '3.4'
|
||||||
services:
|
services:
|
||||||
netbox:
|
netbox:
|
||||||
ports:
|
ports:
|
||||||
|
|
@ -98,7 +99,7 @@ For each of the above tag, there is an extra tag:
|
||||||
## Documentation
|
## Documentation
|
||||||
|
|
||||||
Please refer [to our wiki on GitHub][netbox-docker-wiki] for further information on how to use the NetBox Docker image properly.
|
Please refer [to our wiki on GitHub][netbox-docker-wiki] for further information on how to use the NetBox Docker image properly.
|
||||||
The wiki covers advanced topics such as using files for secrets, configuring TLS, deployment to Kubernetes, monitoring and configuring LDAP.
|
The wiki covers advanced topics such as using files for secrets, configuring TLS, deployment to Kubernetes, monitoring and configuring NAPALM and LDAP.
|
||||||
|
|
||||||
Our wiki is a community effort.
|
Our wiki is a community effort.
|
||||||
Feel free to correct errors, update outdated information or provide additional guides and insights.
|
Feel free to correct errors, update outdated information or provide additional guides and insights.
|
||||||
|
|
|
||||||
|
|
@ -1,11 +1,5 @@
|
||||||
#!/bin/bash
|
#!/bin/bash
|
||||||
|
|
||||||
check_if_tags_exists() {
|
|
||||||
local image=$1
|
|
||||||
local tag=$2
|
|
||||||
skopeo list-tags "docker://$image" | jq -r ".Tags | contains([\"$tag\"])"
|
|
||||||
}
|
|
||||||
|
|
||||||
get_image_label() {
|
get_image_label() {
|
||||||
local label=$1
|
local label=$1
|
||||||
local image=$2
|
local image=$2
|
||||||
|
|
|
||||||
21
build.sh
21
build.sh
|
|
@ -61,7 +61,7 @@ DOCKERFILE The name of Dockerfile to use.
|
||||||
${_GREEN}Default:${_CLEAR} Dockerfile
|
${_GREEN}Default:${_CLEAR} Dockerfile
|
||||||
|
|
||||||
DOCKER_FROM The base image to use.
|
DOCKER_FROM The base image to use.
|
||||||
${_GREEN}Default:${_CLEAR} 'ubuntu:24.04'
|
${_GREEN}Default:${_CLEAR} 'ubuntu:22.04'
|
||||||
|
|
||||||
BUILDX_PLATFORMS
|
BUILDX_PLATFORMS
|
||||||
Specifies the platform(s) to build the image for.
|
Specifies the platform(s) to build the image for.
|
||||||
|
|
@ -103,8 +103,6 @@ GH_ACTION If defined, special 'echo' statements are enabled that set the
|
||||||
- FINAL_DOCKER_TAG: The final value of the DOCKER_TAG env variable
|
- FINAL_DOCKER_TAG: The final value of the DOCKER_TAG env variable
|
||||||
${_GREEN}Default:${_CLEAR} undefined
|
${_GREEN}Default:${_CLEAR} undefined
|
||||||
|
|
||||||
CHECK_ONLY Only checks if the build is needed and sets the GH Action output.
|
|
||||||
|
|
||||||
${_BOLD}Examples:${_CLEAR}
|
${_BOLD}Examples:${_CLEAR}
|
||||||
|
|
||||||
${0} master
|
${0} master
|
||||||
|
|
@ -141,8 +139,7 @@ fi
|
||||||
|
|
||||||
# Check if we have everything needed for the build
|
# Check if we have everything needed for the build
|
||||||
source ./build-functions/check-commands.sh
|
source ./build-functions/check-commands.sh
|
||||||
# Load all build functions
|
|
||||||
source ./build-functions/get-public-image-config.sh
|
|
||||||
source ./build-functions/gh-functions.sh
|
source ./build-functions/gh-functions.sh
|
||||||
|
|
||||||
IMAGE_NAMES="${IMAGE_NAMES-docker.io/netboxcommunity/netbox}"
|
IMAGE_NAMES="${IMAGE_NAMES-docker.io/netboxcommunity/netbox}"
|
||||||
|
|
@ -221,7 +218,7 @@ fi
|
||||||
# Determining the value for DOCKER_FROM
|
# Determining the value for DOCKER_FROM
|
||||||
###
|
###
|
||||||
if [ -z "$DOCKER_FROM" ]; then
|
if [ -z "$DOCKER_FROM" ]; then
|
||||||
DOCKER_FROM="docker.io/ubuntu:24.04"
|
DOCKER_FROM="docker.io/ubuntu:22.04"
|
||||||
fi
|
fi
|
||||||
|
|
||||||
###
|
###
|
||||||
|
|
@ -312,22 +309,19 @@ gh_env "FINAL_DOCKER_TAG=${IMAGE_NAME_TAGS[0]}"
|
||||||
###
|
###
|
||||||
# Checking if the build is necessary,
|
# Checking if the build is necessary,
|
||||||
# meaning build only if one of those values changed:
|
# meaning build only if one of those values changed:
|
||||||
# - a new tag is beeing created
|
|
||||||
# - base image digest
|
# - base image digest
|
||||||
# - netbox git ref (Label: netbox.git-ref)
|
# - netbox git ref (Label: netbox.git-ref)
|
||||||
# - netbox-docker git ref (Label: org.opencontainers.image.revision)
|
# - netbox-docker git ref (Label: org.opencontainers.image.revision)
|
||||||
###
|
###
|
||||||
# Load information from registry (only for first registry in "IMAGE_NAMES")
|
# Load information from registry (only for docker.io)
|
||||||
SHOULD_BUILD="false"
|
SHOULD_BUILD="false"
|
||||||
BUILD_REASON=""
|
BUILD_REASON=""
|
||||||
if [ -z "${GH_ACTION}" ]; then
|
if [ -z "${GH_ACTION}" ]; then
|
||||||
# Asuming non Github builds should always proceed
|
# Asuming non Github builds should always proceed
|
||||||
SHOULD_BUILD="true"
|
SHOULD_BUILD="true"
|
||||||
BUILD_REASON="${BUILD_REASON} interactive"
|
BUILD_REASON="${BUILD_REASON} interactive"
|
||||||
elif [ "false" == "$(check_if_tags_exists "${IMAGE_NAMES[0]}" "$TARGET_DOCKER_TAG")" ]; then
|
|
||||||
SHOULD_BUILD="true"
|
|
||||||
BUILD_REASON="${BUILD_REASON} newtag"
|
|
||||||
else
|
else
|
||||||
|
source ./build-functions/get-public-image-config.sh
|
||||||
echo "Checking labels for '${FINAL_DOCKER_TAG}'"
|
echo "Checking labels for '${FINAL_DOCKER_TAG}'"
|
||||||
BASE_LAST_LAYER=$(get_image_last_layer "${DOCKER_FROM}")
|
BASE_LAST_LAYER=$(get_image_last_layer "${DOCKER_FROM}")
|
||||||
OLD_BASE_LAST_LAYER=$(get_image_label netbox.last-base-image-layer "${FINAL_DOCKER_TAG}")
|
OLD_BASE_LAST_LAYER=$(get_image_label netbox.last-base-image-layer "${FINAL_DOCKER_TAG}")
|
||||||
|
|
@ -357,11 +351,6 @@ else
|
||||||
fi
|
fi
|
||||||
gh_echo "::endgroup::"
|
gh_echo "::endgroup::"
|
||||||
|
|
||||||
if [ "${CHECK_ONLY}" = "true" ]; then
|
|
||||||
echo "Only check if build needed was requested. Exiting"
|
|
||||||
exit 0
|
|
||||||
fi
|
|
||||||
|
|
||||||
###
|
###
|
||||||
# Build the image
|
# Build the image
|
||||||
###
|
###
|
||||||
|
|
|
||||||
|
|
@ -86,9 +86,6 @@ REDIS = {
|
||||||
'tasks': {
|
'tasks': {
|
||||||
'HOST': environ.get('REDIS_HOST', 'localhost'),
|
'HOST': environ.get('REDIS_HOST', 'localhost'),
|
||||||
'PORT': _environ_get_and_map('REDIS_PORT', 6379, _AS_INT),
|
'PORT': _environ_get_and_map('REDIS_PORT', 6379, _AS_INT),
|
||||||
'SENTINELS': [tuple(uri.split(':')) for uri in _environ_get_and_map('REDIS_SENTINELS', '', _AS_LIST) if uri != ''],
|
|
||||||
'SENTINEL_SERVICE': environ.get('REDIS_SENTINEL_SERVICE', 'default'),
|
|
||||||
'SENTINEL_TIMEOUT': _environ_get_and_map('REDIS_SENTINEL_TIMEOUT', 10, _AS_INT),
|
|
||||||
'USERNAME': environ.get('REDIS_USERNAME', ''),
|
'USERNAME': environ.get('REDIS_USERNAME', ''),
|
||||||
'PASSWORD': _read_secret('redis_password', environ.get('REDIS_PASSWORD', '')),
|
'PASSWORD': _read_secret('redis_password', environ.get('REDIS_PASSWORD', '')),
|
||||||
'DATABASE': _environ_get_and_map('REDIS_DATABASE', 0, _AS_INT),
|
'DATABASE': _environ_get_and_map('REDIS_DATABASE', 0, _AS_INT),
|
||||||
|
|
@ -98,8 +95,6 @@ REDIS = {
|
||||||
'caching': {
|
'caching': {
|
||||||
'HOST': environ.get('REDIS_CACHE_HOST', environ.get('REDIS_HOST', 'localhost')),
|
'HOST': environ.get('REDIS_CACHE_HOST', environ.get('REDIS_HOST', 'localhost')),
|
||||||
'PORT': _environ_get_and_map('REDIS_CACHE_PORT', environ.get('REDIS_PORT', '6379'), _AS_INT),
|
'PORT': _environ_get_and_map('REDIS_CACHE_PORT', environ.get('REDIS_PORT', '6379'), _AS_INT),
|
||||||
'SENTINELS': [tuple(uri.split(':')) for uri in _environ_get_and_map('REDIS_CACHE_SENTINELS', '', _AS_LIST) if uri != ''],
|
|
||||||
'SENTINEL_SERVICE': environ.get('REDIS_CACHE_SENTINEL_SERVICE', environ.get('REDIS_SENTINEL_SERVICE', 'default')),
|
|
||||||
'USERNAME': environ.get('REDIS_CACHE_USERNAME', environ.get('REDIS_USERNAME', '')),
|
'USERNAME': environ.get('REDIS_CACHE_USERNAME', environ.get('REDIS_USERNAME', '')),
|
||||||
'PASSWORD': _read_secret('redis_cache_password', environ.get('REDIS_CACHE_PASSWORD', environ.get('REDIS_PASSWORD', ''))),
|
'PASSWORD': _read_secret('redis_cache_password', environ.get('REDIS_CACHE_PASSWORD', environ.get('REDIS_PASSWORD', ''))),
|
||||||
'DATABASE': _environ_get_and_map('REDIS_CACHE_DATABASE', '1', _AS_INT),
|
'DATABASE': _environ_get_and_map('REDIS_CACHE_DATABASE', '1', _AS_INT),
|
||||||
|
|
@ -141,16 +136,17 @@ if 'BANNER_BOTTOM' in environ:
|
||||||
if 'BANNER_LOGIN' in environ:
|
if 'BANNER_LOGIN' in environ:
|
||||||
BANNER_LOGIN = environ.get('BANNER_LOGIN', None)
|
BANNER_LOGIN = environ.get('BANNER_LOGIN', None)
|
||||||
|
|
||||||
|
# Base URL path if accessing NetBox within a directory. For example, if installed at http://example.com/netbox/, set:
|
||||||
|
# BASE_PATH = 'netbox/'
|
||||||
|
BASE_PATH = environ.get('BASE_PATH', '')
|
||||||
|
|
||||||
# Maximum number of days to retain logged changes. Set to 0 to retain changes indefinitely. (Default: 90)
|
# Maximum number of days to retain logged changes. Set to 0 to retain changes indefinitely. (Default: 90)
|
||||||
if 'CHANGELOG_RETENTION' in environ:
|
if 'CHANGELOG_RETENTION' in environ:
|
||||||
CHANGELOG_RETENTION = _environ_get_and_map('CHANGELOG_RETENTION', None, _AS_INT)
|
CHANGELOG_RETENTION = _environ_get_and_map('CHANGELOG_RETENTION', None, _AS_INT)
|
||||||
|
|
||||||
# Maximum number of days to retain job results (scripts and reports). Set to 0 to retain job results in the database indefinitely. (Default: 90)
|
# Maximum number of days to retain job results (scripts and reports). Set to 0 to retain job results in the database indefinitely. (Default: 90)
|
||||||
if 'JOB_RETENTION' in environ:
|
if 'JOBRESULT_RETENTION' in environ:
|
||||||
JOB_RETENTION = _environ_get_and_map('JOB_RETENTION', None, _AS_INT)
|
JOBRESULT_RETENTION = _environ_get_and_map('JOBRESULT_RETENTION', None, _AS_INT)
|
||||||
# JOBRESULT_RETENTION was renamed to JOB_RETENTION in the v3.5.0 release of NetBox. For backwards compatibility, map JOBRESULT_RETENTION to JOB_RETENTION
|
|
||||||
elif 'JOBRESULT_RETENTION' in environ:
|
|
||||||
JOB_RETENTION = _environ_get_and_map('JOBRESULT_RETENTION', None, _AS_INT)
|
|
||||||
|
|
||||||
# API Cross-Origin Resource Sharing (CORS) settings. If CORS_ORIGIN_ALLOW_ALL is set to True, all origins will be
|
# API Cross-Origin Resource Sharing (CORS) settings. If CORS_ORIGIN_ALLOW_ALL is set to True, all origins will be
|
||||||
# allowed. Otherwise, define a list of allowed origins using either CORS_ORIGIN_WHITELIST or
|
# allowed. Otherwise, define a list of allowed origins using either CORS_ORIGIN_WHITELIST or
|
||||||
|
|
@ -188,22 +184,15 @@ EMAIL = {
|
||||||
if 'ENFORCE_GLOBAL_UNIQUE' in environ:
|
if 'ENFORCE_GLOBAL_UNIQUE' in environ:
|
||||||
ENFORCE_GLOBAL_UNIQUE = _environ_get_and_map('ENFORCE_GLOBAL_UNIQUE', None, _AS_BOOL)
|
ENFORCE_GLOBAL_UNIQUE = _environ_get_and_map('ENFORCE_GLOBAL_UNIQUE', None, _AS_BOOL)
|
||||||
|
|
||||||
# By default, netbox sends census reporting data using a single HTTP request each time a worker starts.
|
|
||||||
# This data enables the project maintainers to estimate how many NetBox deployments exist and track the adoption of new versions over time.
|
|
||||||
# The only data reported by this function are the NetBox version, Python version, and a pseudorandom unique identifier.
|
|
||||||
# To opt out of census reporting, set CENSUS_REPORTING_ENABLED to False.
|
|
||||||
if 'CENSUS_REPORTING_ENABLED' in environ:
|
|
||||||
CENSUS_REPORTING_ENABLED = _environ_get_and_map('CENSUS_REPORTING_ENABLED', None, _AS_BOOL)
|
|
||||||
|
|
||||||
# Exempt certain models from the enforcement of view permissions. Models listed here will be viewable by all users and
|
# Exempt certain models from the enforcement of view permissions. Models listed here will be viewable by all users and
|
||||||
# by anonymous users. List models in the form `<app>.<model>`. Add '*' to this list to exempt all models.
|
# by anonymous users. List models in the form `<app>.<model>`. Add '*' to this list to exempt all models.
|
||||||
EXEMPT_VIEW_PERMISSIONS = _environ_get_and_map('EXEMPT_VIEW_PERMISSIONS', '', _AS_LIST)
|
EXEMPT_VIEW_PERMISSIONS = _environ_get_and_map('EXEMPT_VIEW_PERMISSIONS', '', _AS_LIST)
|
||||||
|
|
||||||
# HTTP proxies NetBox should use when sending outbound HTTP requests (e.g. for webhooks).
|
# HTTP proxies NetBox should use when sending outbound HTTP requests (e.g. for webhooks).
|
||||||
HTTP_PROXIES = {
|
# HTTP_PROXIES = {
|
||||||
'http': environ.get('HTTP_PROXY', None),
|
# 'http': 'http://10.10.1.10:3128',
|
||||||
'https': environ.get('HTTPS_PROXY', None),
|
# 'https': 'http://10.10.1.10:1080',
|
||||||
}
|
# }
|
||||||
|
|
||||||
# IP addresses recognized as internal to the system. The debugging toolbar will be available only to clients accessing
|
# IP addresses recognized as internal to the system. The debugging toolbar will be available only to clients accessing
|
||||||
# NetBox from an internal IP.
|
# NetBox from an internal IP.
|
||||||
|
|
@ -221,9 +210,9 @@ if 'GRAPHQL_ENABLED' in environ:
|
||||||
# authenticated to NetBox indefinitely.
|
# authenticated to NetBox indefinitely.
|
||||||
LOGIN_PERSISTENCE = _environ_get_and_map('LOGIN_PERSISTENCE', 'False', _AS_BOOL)
|
LOGIN_PERSISTENCE = _environ_get_and_map('LOGIN_PERSISTENCE', 'False', _AS_BOOL)
|
||||||
|
|
||||||
# When enabled, only authenticated users are permitted to access any part of NetBox.
|
# Setting this to True will permit only authenticated users to access any part of NetBox. By default, anonymous users
|
||||||
# Disabling this will allow unauthenticated users to access most areas of NetBox (but not make any changes).
|
# are permitted to access most data in NetBox (excluding secrets) but not make any changes.
|
||||||
LOGIN_REQUIRED = _environ_get_and_map('LOGIN_REQUIRED', 'True', _AS_BOOL)
|
LOGIN_REQUIRED = _environ_get_and_map('LOGIN_REQUIRED', 'False', _AS_BOOL)
|
||||||
|
|
||||||
# The length of time (in seconds) for which a user will remain logged into the web UI before being prompted to
|
# The length of time (in seconds) for which a user will remain logged into the web UI before being prompted to
|
||||||
# re-authenticate. (Default: 1209600 [14 days])
|
# re-authenticate. (Default: 1209600 [14 days])
|
||||||
|
|
@ -250,6 +239,20 @@ MEDIA_ROOT = environ.get('MEDIA_ROOT', join(_BASE_DIR, 'media'))
|
||||||
# Expose Prometheus monitoring metrics at the HTTP endpoint '/metrics'
|
# Expose Prometheus monitoring metrics at the HTTP endpoint '/metrics'
|
||||||
METRICS_ENABLED = _environ_get_and_map('METRICS_ENABLED', 'False', _AS_BOOL)
|
METRICS_ENABLED = _environ_get_and_map('METRICS_ENABLED', 'False', _AS_BOOL)
|
||||||
|
|
||||||
|
# Credentials that NetBox will uses to authenticate to devices when connecting via NAPALM.
|
||||||
|
if 'NAPALM_USERNAME' in environ:
|
||||||
|
NAPALM_USERNAME = environ.get('NAPALM_USERNAME', None)
|
||||||
|
if 'NAPALM_PASSWORD' in environ:
|
||||||
|
NAPALM_PASSWORD = _read_secret('napalm_password', environ.get('NAPALM_PASSWORD', None))
|
||||||
|
|
||||||
|
# NAPALM timeout (in seconds). (Default: 30)
|
||||||
|
if 'NAPALM_TIMEOUT' in environ:
|
||||||
|
NAPALM_TIMEOUT = _environ_get_and_map('NAPALM_TIMEOUT', None, _AS_INT)
|
||||||
|
|
||||||
|
# # NAPALM optional arguments (see http://napalm.readthedocs.io/en/latest/support/#optional-arguments). Arguments must
|
||||||
|
# # be provided as a dictionary.
|
||||||
|
# NAPALM_ARGS = None
|
||||||
|
|
||||||
# Determine how many objects to display per page within a list. (Default: 50)
|
# Determine how many objects to display per page within a list. (Default: 50)
|
||||||
if 'PAGINATE_COUNT' in environ:
|
if 'PAGINATE_COUNT' in environ:
|
||||||
PAGINATE_COUNT = _environ_get_and_map('PAGINATE_COUNT', None, _AS_INT)
|
PAGINATE_COUNT = _environ_get_and_map('PAGINATE_COUNT', None, _AS_INT)
|
||||||
|
|
@ -286,32 +289,29 @@ if 'RACK_ELEVATION_DEFAULT_UNIT_WIDTH' in environ:
|
||||||
RACK_ELEVATION_DEFAULT_UNIT_WIDTH = _environ_get_and_map('RACK_ELEVATION_DEFAULT_UNIT_WIDTH', None, _AS_INT)
|
RACK_ELEVATION_DEFAULT_UNIT_WIDTH = _environ_get_and_map('RACK_ELEVATION_DEFAULT_UNIT_WIDTH', None, _AS_INT)
|
||||||
|
|
||||||
# Remote authentication support
|
# Remote authentication support
|
||||||
REMOTE_AUTH_AUTO_CREATE_GROUPS = _environ_get_and_map('REMOTE_AUTH_AUTO_CREATE_GROUPS', 'False', _AS_BOOL)
|
|
||||||
REMOTE_AUTH_AUTO_CREATE_USER = _environ_get_and_map('REMOTE_AUTH_AUTO_CREATE_USER', 'False', _AS_BOOL)
|
|
||||||
REMOTE_AUTH_BACKEND = _environ_get_and_map('REMOTE_AUTH_BACKEND', 'netbox.authentication.RemoteUserBackend', _AS_LIST)
|
|
||||||
REMOTE_AUTH_DEFAULT_GROUPS = _environ_get_and_map('REMOTE_AUTH_DEFAULT_GROUPS', '', _AS_LIST)
|
|
||||||
# REMOTE_AUTH_DEFAULT_PERMISSIONS = {} # dicts can't be configured via environment variables. See extra.py instead.
|
|
||||||
REMOTE_AUTH_ENABLED = _environ_get_and_map('REMOTE_AUTH_ENABLED', 'False', _AS_BOOL)
|
REMOTE_AUTH_ENABLED = _environ_get_and_map('REMOTE_AUTH_ENABLED', 'False', _AS_BOOL)
|
||||||
REMOTE_AUTH_GROUP_HEADER = _environ_get_and_map('REMOTE_AUTH_GROUP_HEADER', 'HTTP_REMOTE_USER_GROUP')
|
REMOTE_AUTH_BACKEND = environ.get('REMOTE_AUTH_BACKEND', 'netbox.authentication.RemoteUserBackend')
|
||||||
REMOTE_AUTH_GROUP_SEPARATOR = _environ_get_and_map('REMOTE_AUTH_GROUP_SEPARATOR', '|')
|
|
||||||
REMOTE_AUTH_GROUP_SYNC_ENABLED = _environ_get_and_map('REMOTE_AUTH_GROUP_SYNC_ENABLED', 'False', _AS_BOOL)
|
|
||||||
REMOTE_AUTH_HEADER = environ.get('REMOTE_AUTH_HEADER', 'HTTP_REMOTE_USER')
|
REMOTE_AUTH_HEADER = environ.get('REMOTE_AUTH_HEADER', 'HTTP_REMOTE_USER')
|
||||||
REMOTE_AUTH_USER_EMAIL = environ.get('REMOTE_AUTH_USER_EMAIL', 'HTTP_REMOTE_USER_EMAIL')
|
REMOTE_AUTH_AUTO_CREATE_USER = _environ_get_and_map('REMOTE_AUTH_AUTO_CREATE_USER', 'True', _AS_BOOL)
|
||||||
REMOTE_AUTH_USER_FIRST_NAME = environ.get('REMOTE_AUTH_USER_FIRST_NAME', 'HTTP_REMOTE_USER_FIRST_NAME')
|
REMOTE_AUTH_DEFAULT_GROUPS = _environ_get_and_map('REMOTE_AUTH_DEFAULT_GROUPS', '', _AS_LIST)
|
||||||
REMOTE_AUTH_USER_LAST_NAME = environ.get('REMOTE_AUTH_USER_LAST_NAME', 'HTTP_REMOTE_USER_LAST_NAME')
|
# REMOTE_AUTH_DEFAULT_PERMISSIONS = {}
|
||||||
REMOTE_AUTH_SUPERUSER_GROUPS = _environ_get_and_map('REMOTE_AUTH_SUPERUSER_GROUPS', '', _AS_LIST)
|
|
||||||
REMOTE_AUTH_SUPERUSERS = _environ_get_and_map('REMOTE_AUTH_SUPERUSERS', '', _AS_LIST)
|
|
||||||
REMOTE_AUTH_STAFF_GROUPS = _environ_get_and_map('REMOTE_AUTH_STAFF_GROUPS', '', _AS_LIST)
|
|
||||||
REMOTE_AUTH_STAFF_USERS = _environ_get_and_map('REMOTE_AUTH_STAFF_USERS', '', _AS_LIST)
|
|
||||||
|
|
||||||
# This repository is used to check whether there is a new release of NetBox available. Set to None to disable the
|
# This repository is used to check whether there is a new release of NetBox available. Set to None to disable the
|
||||||
# version check or use the URL below to check for release in the official NetBox repository.
|
# version check or use the URL below to check for release in the official NetBox repository.
|
||||||
RELEASE_CHECK_URL = environ.get('RELEASE_CHECK_URL', None)
|
RELEASE_CHECK_URL = environ.get('RELEASE_CHECK_URL', None)
|
||||||
# RELEASE_CHECK_URL = 'https://api.github.com/repos/netbox-community/netbox/releases'
|
# RELEASE_CHECK_URL = 'https://api.github.com/repos/netbox-community/netbox/releases'
|
||||||
|
|
||||||
|
# The file path where custom reports will be stored. A trailing slash is not needed. Note that the default value of
|
||||||
|
# this setting is derived from the installed location.
|
||||||
|
REPORTS_ROOT = environ.get('REPORTS_ROOT', '/etc/netbox/reports')
|
||||||
|
|
||||||
# Maximum execution time for background tasks, in seconds.
|
# Maximum execution time for background tasks, in seconds.
|
||||||
RQ_DEFAULT_TIMEOUT = _environ_get_and_map('RQ_DEFAULT_TIMEOUT', 300, _AS_INT)
|
RQ_DEFAULT_TIMEOUT = _environ_get_and_map('RQ_DEFAULT_TIMEOUT', 300, _AS_INT)
|
||||||
|
|
||||||
|
# The file path where custom scripts will be stored. A trailing slash is not needed. Note that the default value of
|
||||||
|
# this setting is derived from the installed location.
|
||||||
|
SCRIPTS_ROOT = environ.get('SCRIPTS_ROOT', '/etc/netbox/scripts')
|
||||||
|
|
||||||
# The name to use for the csrf token cookie.
|
# The name to use for the csrf token cookie.
|
||||||
CSRF_COOKIE_NAME = environ.get('CSRF_COOKIE_NAME', 'csrftoken')
|
CSRF_COOKIE_NAME = environ.get('CSRF_COOKIE_NAME', 'csrftoken')
|
||||||
|
|
||||||
|
|
@ -323,23 +323,6 @@ CSRF_TRUSTED_ORIGINS = _environ_get_and_map('CSRF_TRUSTED_ORIGINS', '', _AS_LIST
|
||||||
# The name to use for the session cookie.
|
# The name to use for the session cookie.
|
||||||
SESSION_COOKIE_NAME = environ.get('SESSION_COOKIE_NAME', 'sessionid')
|
SESSION_COOKIE_NAME = environ.get('SESSION_COOKIE_NAME', 'sessionid')
|
||||||
|
|
||||||
# If true, the `includeSubDomains` directive will be included in the HTTP Strict Transport Security (HSTS) header.
|
|
||||||
# This directive instructs the browser to apply the HSTS policy to all subdomains of the current domain.
|
|
||||||
SECURE_HSTS_INCLUDE_SUBDOMAINS = _environ_get_and_map('SECURE_HSTS_INCLUDE_SUBDOMAINS', 'False', _AS_BOOL)
|
|
||||||
|
|
||||||
# If true, the `preload` directive will be included in the HTTP Strict Transport Security (HSTS) header.
|
|
||||||
# This directive instructs the browser to preload the site in HTTPS. Browsers that use the HSTS preload list will force the
|
|
||||||
# site to be accessed via HTTPS even if the user types HTTP in the address bar.
|
|
||||||
SECURE_HSTS_PRELOAD = _environ_get_and_map('SECURE_HSTS_PRELOAD', 'False', _AS_BOOL)
|
|
||||||
|
|
||||||
# If set to a non-zero integer value, the SecurityMiddleware sets the HTTP Strict Transport Security (HSTS) header on all
|
|
||||||
# responses that do not already have it. This will instruct the browser that the website must be accessed via HTTPS,
|
|
||||||
# blocking any HTTP request.
|
|
||||||
SECURE_HSTS_SECONDS = _environ_get_and_map('SECURE_HSTS_SECONDS', 0, _AS_INT)
|
|
||||||
|
|
||||||
# If true, all non-HTTPS requests will be automatically redirected to use HTTPS.
|
|
||||||
SECURE_SSL_REDIRECT = _environ_get_and_map('SECURE_SSL_REDIRECT', 'False', _AS_BOOL)
|
|
||||||
|
|
||||||
# By default, NetBox will store session data in the database. Alternatively, a file path can be specified here to use
|
# By default, NetBox will store session data in the database. Alternatively, a file path can be specified here to use
|
||||||
# local file storage instead. (This can be useful for enabling authentication on a standby instance with read-only
|
# local file storage instead. (This can be useful for enabling authentication on a standby instance with read-only
|
||||||
# database access.) Note that the user as which NetBox runs must have read and write permissions to this path.
|
# database access.) Note that the user as which NetBox runs must have read and write permissions to this path.
|
||||||
|
|
@ -348,3 +331,11 @@ SESSION_FILE_PATH = environ.get('SESSION_FILE_PATH', environ.get('SESSIONS_ROOT'
|
||||||
# Time zone (default: UTC)
|
# Time zone (default: UTC)
|
||||||
TIME_ZONE = environ.get('TIME_ZONE', 'UTC')
|
TIME_ZONE = environ.get('TIME_ZONE', 'UTC')
|
||||||
|
|
||||||
|
# Date/time formatting. See the following link for supported formats:
|
||||||
|
# https://docs.djangoproject.com/en/stable/ref/templates/builtins/#date
|
||||||
|
DATE_FORMAT = environ.get('DATE_FORMAT', 'N j, Y')
|
||||||
|
SHORT_DATE_FORMAT = environ.get('SHORT_DATE_FORMAT', 'Y-m-d')
|
||||||
|
TIME_FORMAT = environ.get('TIME_FORMAT', 'g:i a')
|
||||||
|
SHORT_TIME_FORMAT = environ.get('SHORT_TIME_FORMAT', 'H:i:s')
|
||||||
|
DATETIME_FORMAT = environ.get('DATETIME_FORMAT', 'N j, Y g:i a')
|
||||||
|
SHORT_DATETIME_FORMAT = environ.get('SHORT_DATETIME_FORMAT', 'Y-m-d H:i')
|
||||||
|
|
|
||||||
|
|
@ -15,6 +15,12 @@
|
||||||
# 'file', 'ftp', 'ftps', 'http', 'https', 'irc', 'mailto', 'sftp', 'ssh', 'tel', 'telnet', 'tftp', 'vnc', 'xmpp',
|
# 'file', 'ftp', 'ftps', 'http', 'https', 'irc', 'mailto', 'sftp', 'ssh', 'tel', 'telnet', 'tftp', 'vnc', 'xmpp',
|
||||||
# )
|
# )
|
||||||
|
|
||||||
|
|
||||||
|
## NAPALM optional arguments (see http://napalm.readthedocs.io/en/latest/support/#optional-arguments). Arguments must
|
||||||
|
## be provided as a dictionary.
|
||||||
|
# NAPALM_ARGS = {}
|
||||||
|
|
||||||
|
|
||||||
## Enable installed plugins. Add the name of each plugin to the list.
|
## Enable installed plugins. Add the name of each plugin to the list.
|
||||||
# from netbox.configuration.configuration import PLUGINS
|
# from netbox.configuration.configuration import PLUGINS
|
||||||
# PLUGINS.append('my_plugin')
|
# PLUGINS.append('my_plugin')
|
||||||
|
|
|
||||||
|
|
@ -1,21 +0,0 @@
|
||||||
services:
|
|
||||||
netbox:
|
|
||||||
ports:
|
|
||||||
- "8000:8080"
|
|
||||||
# If you want the Nginx unit status page visible from the
|
|
||||||
# outside of the container add the following port mapping:
|
|
||||||
# - "8001:8081"
|
|
||||||
# healthcheck:
|
|
||||||
# Time for which the health check can fail after the container is started.
|
|
||||||
# This depends mostly on the performance of your database. On the first start,
|
|
||||||
# when all tables need to be created the start_period should be higher than on
|
|
||||||
# subsequent starts. For the first start after major version upgrades of NetBox
|
|
||||||
# the start_period might also need to be set higher.
|
|
||||||
# Default value in our docker-compose.yml is 60s
|
|
||||||
# start_period: 90s
|
|
||||||
# environment:
|
|
||||||
# SKIP_SUPERUSER: "false"
|
|
||||||
# SUPERUSER_API_TOKEN: ""
|
|
||||||
# SUPERUSER_EMAIL: ""
|
|
||||||
# SUPERUSER_NAME: ""
|
|
||||||
# SUPERUSER_PASSWORD: ""
|
|
||||||
|
|
@ -1,3 +1,4 @@
|
||||||
|
version: '3.4'
|
||||||
services:
|
services:
|
||||||
netbox:
|
netbox:
|
||||||
ports:
|
ports:
|
||||||
|
|
|
||||||
|
|
@ -1,5 +0,0 @@
|
||||||
services:
|
|
||||||
netbox:
|
|
||||||
ports:
|
|
||||||
- "127.0.0.1:8000:8080"
|
|
||||||
|
|
||||||
|
|
@ -1,70 +1,44 @@
|
||||||
|
version: '3.4'
|
||||||
services:
|
services:
|
||||||
netbox: &netbox
|
netbox:
|
||||||
image: ${IMAGE-docker.io/netboxcommunity/netbox:latest}
|
image: ${IMAGE-netboxcommunity/netbox:latest}
|
||||||
depends_on:
|
depends_on:
|
||||||
postgres:
|
postgres:
|
||||||
condition: service_healthy
|
condition: service_healthy
|
||||||
redis:
|
redis:
|
||||||
condition: service_healthy
|
condition: service_started
|
||||||
redis-cache:
|
redis-cache:
|
||||||
condition: service_healthy
|
condition: service_started
|
||||||
env_file: env/netbox.env
|
env_file: env/netbox.env
|
||||||
user: 'unit:root'
|
user: 'unit:root'
|
||||||
volumes:
|
volumes:
|
||||||
- ./test-configuration/test_config.py:/etc/netbox/config/test_config.py:z,ro
|
- ./configuration:/etc/netbox/config:z,ro
|
||||||
healthcheck:
|
- ./test-configuration/logging.py:/etc/netbox/config/logging.py:z,ro
|
||||||
test: curl -f http://localhost:8080/login/ || exit 1
|
- ./reports:/etc/netbox/reports:z,ro
|
||||||
start_period: ${NETBOX_START_PERIOD-120s}
|
- ./scripts:/etc/netbox/scripts:z,ro
|
||||||
timeout: 3s
|
- netbox-media-files:/opt/netbox/netbox/media:z
|
||||||
interval: 15s
|
|
||||||
netbox-worker:
|
|
||||||
<<: *netbox
|
|
||||||
command:
|
|
||||||
- /opt/netbox/venv/bin/python
|
|
||||||
- /opt/netbox/netbox/manage.py
|
|
||||||
- rqworker
|
|
||||||
healthcheck:
|
|
||||||
test: ps -aux | grep -v grep | grep -q rqworker || exit 1
|
|
||||||
start_period: 40s
|
|
||||||
timeout: 3s
|
|
||||||
interval: 15s
|
|
||||||
netbox-housekeeping:
|
|
||||||
<<: *netbox
|
|
||||||
command:
|
|
||||||
- /opt/netbox/housekeeping.sh
|
|
||||||
healthcheck:
|
|
||||||
test: ps -aux | grep -v grep | grep -q housekeeping || exit 1
|
|
||||||
start_period: 40s
|
|
||||||
timeout: 3s
|
|
||||||
interval: 15s
|
|
||||||
|
|
||||||
postgres:
|
postgres:
|
||||||
image: docker.io/postgres:16-alpine
|
image: postgres:15-alpine
|
||||||
env_file: env/postgres.env
|
env_file: env/postgres.env
|
||||||
healthcheck:
|
healthcheck:
|
||||||
test: pg_isready -q -t 2 -d $$POSTGRES_DB -U $$POSTGRES_USER ## $$ because of docker-compose
|
test: ["CMD-SHELL", "pg_isready"]
|
||||||
start_period: 20s
|
interval: 10s
|
||||||
interval: 1s
|
|
||||||
timeout: 5s
|
timeout: 5s
|
||||||
retries: 5
|
retries: 5
|
||||||
|
redis:
|
||||||
redis: &redis
|
image: redis:7-alpine
|
||||||
image: docker.io/valkey/valkey:8.0-alpine
|
|
||||||
command:
|
command:
|
||||||
- sh
|
- sh
|
||||||
- -c # this is to evaluate the $REDIS_PASSWORD from the env
|
- -c # this is to evaluate the $REDIS_PASSWORD from the env
|
||||||
- valkey-server --save "" --appendonly no --requirepass $$REDIS_PASSWORD ## $$ because of docker-compose
|
- redis-server --appendonly yes --requirepass $$REDIS_PASSWORD ## $$ because of docker-compose
|
||||||
env_file: env/redis.env
|
env_file: env/redis.env
|
||||||
healthcheck:
|
|
||||||
test: "[ $$(valkey-cli --pass \"$${REDIS_PASSWORD}\" ping) = 'PONG' ]"
|
|
||||||
start_period: 5s
|
|
||||||
timeout: 3s
|
|
||||||
interval: 1s
|
|
||||||
retries: 5
|
|
||||||
redis-cache:
|
redis-cache:
|
||||||
<<: *redis
|
image: redis:7-alpine
|
||||||
|
command:
|
||||||
|
- sh
|
||||||
|
- -c # this is to evaluate the $REDIS_PASSWORD from the env
|
||||||
|
- redis-server --requirepass $$REDIS_PASSWORD ## $$ because of docker-compose
|
||||||
env_file: env/redis-cache.env
|
env_file: env/redis-cache.env
|
||||||
|
|
||||||
volumes:
|
volumes:
|
||||||
netbox-media-files:
|
netbox-media-files:
|
||||||
driver: local
|
driver: local
|
||||||
|
|
|
||||||
|
|
@ -1,22 +1,23 @@
|
||||||
|
version: '3.4'
|
||||||
services:
|
services:
|
||||||
netbox: &netbox
|
netbox: &netbox
|
||||||
image: docker.io/netboxcommunity/netbox:${VERSION-v4.1-3.0.2}
|
image: docker.io/netboxcommunity/netbox:${VERSION-v3.4-2.5.1}
|
||||||
depends_on:
|
depends_on:
|
||||||
- postgres
|
- postgres
|
||||||
- redis
|
- redis
|
||||||
- redis-cache
|
- redis-cache
|
||||||
env_file: env/netbox.env
|
env_file: env/netbox.env
|
||||||
user: "unit:root"
|
user: 'unit:root'
|
||||||
healthcheck:
|
healthcheck:
|
||||||
test: curl -f http://localhost:8080/login/ || exit 1
|
start_period: 60s
|
||||||
start_period: 90s
|
|
||||||
timeout: 3s
|
timeout: 3s
|
||||||
interval: 15s
|
interval: 15s
|
||||||
|
test: "curl -f http://localhost:8080/api/ || exit 1"
|
||||||
volumes:
|
volumes:
|
||||||
- ./configuration:/etc/netbox/config:z,ro
|
- ./configuration:/etc/netbox/config:z,ro
|
||||||
- netbox-media-files:/opt/netbox/netbox/media:rw
|
- ./reports:/etc/netbox/reports:z,ro
|
||||||
- netbox-reports-files:/opt/netbox/netbox/reports:rw
|
- ./scripts:/etc/netbox/scripts:z,ro
|
||||||
- netbox-scripts-files:/opt/netbox/netbox/scripts:rw
|
- netbox-media-files:/opt/netbox/netbox/media:z
|
||||||
netbox-worker:
|
netbox-worker:
|
||||||
<<: *netbox
|
<<: *netbox
|
||||||
depends_on:
|
depends_on:
|
||||||
|
|
@ -27,10 +28,10 @@ services:
|
||||||
- /opt/netbox/netbox/manage.py
|
- /opt/netbox/netbox/manage.py
|
||||||
- rqworker
|
- rqworker
|
||||||
healthcheck:
|
healthcheck:
|
||||||
test: ps -aux | grep -v grep | grep -q rqworker || exit 1
|
|
||||||
start_period: 20s
|
start_period: 20s
|
||||||
timeout: 3s
|
timeout: 3s
|
||||||
interval: 15s
|
interval: 15s
|
||||||
|
test: "ps -aux | grep -v grep | grep -q rqworker || exit 1"
|
||||||
netbox-housekeeping:
|
netbox-housekeeping:
|
||||||
<<: *netbox
|
<<: *netbox
|
||||||
depends_on:
|
depends_on:
|
||||||
|
|
@ -39,47 +40,34 @@ services:
|
||||||
command:
|
command:
|
||||||
- /opt/netbox/housekeeping.sh
|
- /opt/netbox/housekeeping.sh
|
||||||
healthcheck:
|
healthcheck:
|
||||||
test: ps -aux | grep -v grep | grep -q housekeeping || exit 1
|
|
||||||
start_period: 20s
|
start_period: 20s
|
||||||
timeout: 3s
|
timeout: 3s
|
||||||
interval: 15s
|
interval: 15s
|
||||||
|
test: "ps -aux | grep -v grep | grep -q housekeeping || exit 1"
|
||||||
|
|
||||||
# postgres
|
# postgres
|
||||||
postgres:
|
postgres:
|
||||||
image: docker.io/postgres:16-alpine
|
image: docker.io/postgres:15-alpine
|
||||||
healthcheck:
|
|
||||||
test: pg_isready -q -t 2 -d $$POSTGRES_DB -U $$POSTGRES_USER
|
|
||||||
start_period: 20s
|
|
||||||
timeout: 30s
|
|
||||||
interval: 10s
|
|
||||||
retries: 5
|
|
||||||
env_file: env/postgres.env
|
env_file: env/postgres.env
|
||||||
volumes:
|
volumes:
|
||||||
- netbox-postgres-data:/var/lib/postgresql/data
|
- netbox-postgres-data:/var/lib/postgresql/data
|
||||||
|
|
||||||
# redis
|
# redis
|
||||||
redis:
|
redis:
|
||||||
image: docker.io/valkey/valkey:8.0-alpine
|
image: docker.io/redis:7-alpine
|
||||||
command:
|
command:
|
||||||
- sh
|
- sh
|
||||||
- -c # this is to evaluate the $REDIS_PASSWORD from the env
|
- -c # this is to evaluate the $REDIS_PASSWORD from the env
|
||||||
- valkey-server --appendonly yes --requirepass $$REDIS_PASSWORD ## $$ because of docker-compose
|
- redis-server --appendonly yes --requirepass $$REDIS_PASSWORD ## $$ because of docker-compose
|
||||||
healthcheck: &redis-healthcheck
|
|
||||||
test: '[ $$(valkey-cli --pass "$${REDIS_PASSWORD}" ping) = ''PONG'' ]'
|
|
||||||
start_period: 5s
|
|
||||||
timeout: 3s
|
|
||||||
interval: 1s
|
|
||||||
retries: 5
|
|
||||||
env_file: env/redis.env
|
env_file: env/redis.env
|
||||||
volumes:
|
volumes:
|
||||||
- netbox-redis-data:/data
|
- netbox-redis-data:/data
|
||||||
redis-cache:
|
redis-cache:
|
||||||
image: docker.io/valkey/valkey:8.0-alpine
|
image: redis:7-alpine
|
||||||
command:
|
command:
|
||||||
- sh
|
- sh
|
||||||
- -c # this is to evaluate the $REDIS_PASSWORD from the env
|
- -c # this is to evaluate the $REDIS_PASSWORD from the env
|
||||||
- valkey-server --requirepass $$REDIS_PASSWORD ## $$ because of docker-compose
|
- redis-server --requirepass $$REDIS_PASSWORD ## $$ because of docker-compose
|
||||||
healthcheck: *redis-healthcheck
|
|
||||||
env_file: env/redis-cache.env
|
env_file: env/redis-cache.env
|
||||||
volumes:
|
volumes:
|
||||||
- netbox-redis-cache-data:/data
|
- netbox-redis-cache-data:/data
|
||||||
|
|
@ -89,11 +77,7 @@ volumes:
|
||||||
driver: local
|
driver: local
|
||||||
netbox-postgres-data:
|
netbox-postgres-data:
|
||||||
driver: local
|
driver: local
|
||||||
netbox-redis-cache-data:
|
|
||||||
driver: local
|
|
||||||
netbox-redis-data:
|
netbox-redis-data:
|
||||||
driver: local
|
driver: local
|
||||||
netbox-reports-files:
|
netbox-redis-cache-data:
|
||||||
driver: local
|
|
||||||
netbox-scripts-files:
|
|
||||||
driver: local
|
driver: local
|
||||||
|
|
|
||||||
|
|
@ -72,7 +72,8 @@ else
|
||||||
fi
|
fi
|
||||||
|
|
||||||
./manage.py shell --interface python <<END
|
./manage.py shell --interface python <<END
|
||||||
from users.models import Token, User
|
from django.contrib.auth.models import User
|
||||||
|
from users.models import Token
|
||||||
if not User.objects.filter(username='${SUPERUSER_NAME}'):
|
if not User.objects.filter(username='${SUPERUSER_NAME}'):
|
||||||
u=User.objects.create_superuser('${SUPERUSER_NAME}', '${SUPERUSER_EMAIL}', '${SUPERUSER_PASSWORD}')
|
u=User.objects.create_superuser('${SUPERUSER_NAME}', '${SUPERUSER_EMAIL}', '${SUPERUSER_PASSWORD}')
|
||||||
Token.objects.create(user=u, key='${SUPERUSER_API_TOKEN}')
|
Token.objects.create(user=u, key='${SUPERUSER_API_TOKEN}')
|
||||||
|
|
@ -86,7 +87,7 @@ from users.models import Token
|
||||||
try:
|
try:
|
||||||
old_default_token = Token.objects.get(key="0123456789abcdef0123456789abcdef01234567")
|
old_default_token = Token.objects.get(key="0123456789abcdef0123456789abcdef01234567")
|
||||||
if old_default_token:
|
if old_default_token:
|
||||||
print("⚠️ Warning: You have the old default admin API token in your database. This token is widely known; please remove it. Log in as your superuser and check API Tokens in your user menu.")
|
print("⚠️ Warning: You have the old default admin token in your database. This token is widely known; please remove it.")
|
||||||
except Token.DoesNotExist:
|
except Token.DoesNotExist:
|
||||||
pass
|
pass
|
||||||
END
|
END
|
||||||
|
|
|
||||||
|
|
@ -51,7 +51,7 @@ exec unitd \
|
||||||
--control unix:$UNIT_SOCKET \
|
--control unix:$UNIT_SOCKET \
|
||||||
--pid /opt/unit/unit.pid \
|
--pid /opt/unit/unit.pid \
|
||||||
--log /dev/stdout \
|
--log /dev/stdout \
|
||||||
--statedir /opt/unit/state/ \
|
--state /opt/unit/state/ \
|
||||||
--tmpdir /opt/unit/tmp/ \
|
--tmp /opt/unit/tmp/ \
|
||||||
--user unit \
|
--user unit \
|
||||||
--group root
|
--group root
|
||||||
|
|
|
||||||
|
|
@ -29,6 +29,6 @@ REDIS_INSECURE_SKIP_TLS_VERIFY=false
|
||||||
REDIS_PASSWORD=H733Kdjndks81
|
REDIS_PASSWORD=H733Kdjndks81
|
||||||
REDIS_SSL=false
|
REDIS_SSL=false
|
||||||
RELEASE_CHECK_URL=https://api.github.com/repos/netbox-community/netbox/releases
|
RELEASE_CHECK_URL=https://api.github.com/repos/netbox-community/netbox/releases
|
||||||
SECRET_KEY='r(m)9nLGnz$(_q3N4z1k(EFsMCjjjzx08x9VhNVcfd%6RF#r!6DE@+V5Zk2X'
|
SECRET_KEY=r8OwDznj!!dci#P9ghmRfdu1Ysxm0AiPeDCQhKE+N_rClfWNj
|
||||||
SKIP_SUPERUSER=true
|
SKIP_SUPERUSER=true
|
||||||
WEBHOOKS_ENABLED=true
|
WEBHOOKS_ENABLED=true
|
||||||
|
|
|
||||||
|
|
@ -0,0 +1,46 @@
|
||||||
|
from dcim.choices import DeviceStatusChoices
|
||||||
|
from dcim.models import ConsolePort, Device, PowerPort
|
||||||
|
from extras.reports import Report
|
||||||
|
|
||||||
|
|
||||||
|
class DeviceConnectionsReport(Report):
|
||||||
|
description = "Validate the minimum physical connections for each device"
|
||||||
|
|
||||||
|
def test_console_connection(self):
|
||||||
|
|
||||||
|
# Check that every console port for every active device has a connection defined.
|
||||||
|
active = DeviceStatusChoices.STATUS_ACTIVE
|
||||||
|
for console_port in ConsolePort.objects.prefetch_related('device').filter(device__status=active):
|
||||||
|
if console_port.connected_endpoint is None:
|
||||||
|
self.log_failure(
|
||||||
|
console_port.device,
|
||||||
|
"No console connection defined for {}".format(console_port.name)
|
||||||
|
)
|
||||||
|
elif not console_port.connection_status:
|
||||||
|
self.log_warning(
|
||||||
|
console_port.device,
|
||||||
|
"Console connection for {} marked as planned".format(console_port.name)
|
||||||
|
)
|
||||||
|
else:
|
||||||
|
self.log_success(console_port.device)
|
||||||
|
|
||||||
|
def test_power_connections(self):
|
||||||
|
|
||||||
|
# Check that every active device has at least two connected power supplies.
|
||||||
|
for device in Device.objects.filter(status=DeviceStatusChoices.STATUS_ACTIVE):
|
||||||
|
connected_ports = 0
|
||||||
|
for power_port in PowerPort.objects.filter(device=device):
|
||||||
|
if power_port.connected_endpoint is not None:
|
||||||
|
connected_ports += 1
|
||||||
|
if not power_port.connection_status:
|
||||||
|
self.log_warning(
|
||||||
|
device,
|
||||||
|
"Power connection for {} marked as planned".format(power_port.name)
|
||||||
|
)
|
||||||
|
if connected_ports < 2:
|
||||||
|
self.log_failure(
|
||||||
|
device,
|
||||||
|
"{} connected power supplies found (2 needed)".format(connected_ports)
|
||||||
|
)
|
||||||
|
else:
|
||||||
|
self.log_success(device)
|
||||||
|
|
@ -1,5 +1,6 @@
|
||||||
django-auth-ldap==4.8.0
|
django-auth-ldap==4.1.0
|
||||||
django-storages[azure,boto3,dropbox,google,libcloud,sftp]==1.14.4
|
django-storages[azure,boto3,dropbox,google,libcloud,sftp]==1.13.2
|
||||||
dulwich==0.22.1
|
napalm==4.0.0
|
||||||
python3-saml==1.16.0 --no-binary lxml,xmlsec
|
psycopg2==2.9.5
|
||||||
sentry-sdk[django]==2.14.0
|
python3-saml==1.15.0
|
||||||
|
social-auth-core[all]==4.3.0
|
||||||
|
|
|
||||||
|
|
@ -2,6 +2,3 @@ LOGGING = {
|
||||||
'version': 1,
|
'version': 1,
|
||||||
'disable_existing_loggers': True
|
'disable_existing_loggers': True
|
||||||
}
|
}
|
||||||
|
|
||||||
DEFAULT_PERMISSIONS = {}
|
|
||||||
LOGIN_REQUIRED = False
|
|
||||||
49
test.sh
49
test.sh
|
|
@ -14,8 +14,6 @@
|
||||||
# exit when a command exits with an exit code != 0
|
# exit when a command exits with an exit code != 0
|
||||||
set -e
|
set -e
|
||||||
|
|
||||||
source ./build-functions/gh-functions.sh
|
|
||||||
|
|
||||||
# IMAGE is used by `docker-compose.yml` do determine the tag
|
# IMAGE is used by `docker-compose.yml` do determine the tag
|
||||||
# of the Docker Image that is to be used
|
# of the Docker Image that is to be used
|
||||||
if [ "${1}x" != "x" ]; then
|
if [ "${1}x" != "x" ]; then
|
||||||
|
|
@ -37,72 +35,29 @@ if [ -z "${IMAGE}" ]; then
|
||||||
fi
|
fi
|
||||||
|
|
||||||
# The docker compose command to use
|
# The docker compose command to use
|
||||||
doco="docker compose --file docker-compose.test.yml --file docker-compose.test.override.yml --project-name netbox_docker_test"
|
doco="docker compose --file docker-compose.test.yml --project-name netbox_docker_test"
|
||||||
|
|
||||||
test_setup() {
|
test_setup() {
|
||||||
gh_echo "::group:: Test setup"
|
|
||||||
echo "🏗 Setup up test environment"
|
echo "🏗 Setup up test environment"
|
||||||
$doco up --detach --quiet-pull --wait --force-recreate --renew-anon-volumes --no-start
|
$doco up --detach --quiet-pull --wait --force-recreate --renew-anon-volumes --no-start
|
||||||
$doco start postgres
|
$doco start postgres
|
||||||
$doco start redis
|
$doco start redis
|
||||||
$doco start redis-cache
|
$doco start redis-cache
|
||||||
gh_echo "::endgroup::"
|
|
||||||
}
|
}
|
||||||
|
|
||||||
test_netbox_unit_tests() {
|
test_netbox_unit_tests() {
|
||||||
gh_echo "::group:: Netbox unit tests"
|
|
||||||
echo "⏱ Running NetBox Unit Tests"
|
echo "⏱ Running NetBox Unit Tests"
|
||||||
$doco run --rm netbox /opt/netbox/venv/bin/python /opt/netbox/netbox/manage.py test
|
$doco run --rm netbox /opt/netbox/venv/bin/python /opt/netbox/netbox/manage.py test
|
||||||
gh_echo "::endgroup::"
|
|
||||||
}
|
}
|
||||||
|
|
||||||
test_compose_db_setup() {
|
test_compose_db_setup() {
|
||||||
gh_echo "::group:: Netbox DB migrations"
|
|
||||||
echo "⏱ Running NetBox DB migrations"
|
echo "⏱ Running NetBox DB migrations"
|
||||||
$doco run --rm netbox /opt/netbox/venv/bin/python /opt/netbox/netbox/manage.py migrate
|
$doco run --rm netbox /opt/netbox/venv/bin/python /opt/netbox/netbox/manage.py migrate
|
||||||
gh_echo "::endgroup::"
|
|
||||||
}
|
|
||||||
|
|
||||||
test_netbox_start() {
|
|
||||||
gh_echo "::group:: Start Netbox service"
|
|
||||||
echo "⏱ Starting NetBox services"
|
|
||||||
$doco up --detach --wait
|
|
||||||
gh_echo "::endgroup::"
|
|
||||||
}
|
|
||||||
|
|
||||||
test_netbox_web() {
|
|
||||||
gh_echo "::group:: Web service test"
|
|
||||||
echo "⏱ Starting web service test"
|
|
||||||
RESP_CODE=$(
|
|
||||||
curl \
|
|
||||||
--silent \
|
|
||||||
--output /dev/null \
|
|
||||||
--write-out '%{http_code}' \
|
|
||||||
--request GET \
|
|
||||||
--connect-timeout 5 \
|
|
||||||
--max-time 10 \
|
|
||||||
--retry 5 \
|
|
||||||
--retry-delay 0 \
|
|
||||||
--retry-max-time 40 \
|
|
||||||
http://127.0.0.1:8000/login/
|
|
||||||
)
|
|
||||||
if [ "$RESP_CODE" == "200" ]; then
|
|
||||||
echo "Webservice running"
|
|
||||||
else
|
|
||||||
echo "⚠️ Got response code '$RESP_CODE' but expected '200'"
|
|
||||||
exit 1
|
|
||||||
fi
|
|
||||||
gh_echo "::endgroup::"
|
|
||||||
}
|
}
|
||||||
|
|
||||||
test_cleanup() {
|
test_cleanup() {
|
||||||
echo "💣 Cleaning Up"
|
echo "💣 Cleaning Up"
|
||||||
gh_echo "::group:: Docker compose logs"
|
|
||||||
$doco logs --no-color
|
|
||||||
gh_echo "::endgroup::"
|
|
||||||
gh_echo "::group:: Docker compose down"
|
|
||||||
$doco down --volumes
|
$doco down --volumes
|
||||||
gh_echo "::endgroup::"
|
|
||||||
}
|
}
|
||||||
|
|
||||||
echo "🐳🐳🐳 Start testing '${IMAGE}'"
|
echo "🐳🐳🐳 Start testing '${IMAGE}'"
|
||||||
|
|
@ -113,7 +68,5 @@ test_setup
|
||||||
|
|
||||||
test_netbox_unit_tests
|
test_netbox_unit_tests
|
||||||
test_compose_db_setup
|
test_compose_db_setup
|
||||||
test_netbox_start
|
|
||||||
test_netbox_web
|
|
||||||
|
|
||||||
echo "🐳🐳🐳 Done testing '${IMAGE}'"
|
echo "🐳🐳🐳 Done testing '${IMAGE}'"
|
||||||
|
|
|
||||||
Loading…
Reference in New Issue