mirror of
https://github.com/dgtlmoon/changedetection.io.git
synced 2025-11-12 04:27:27 +00:00
Compare commits
227 Commits
skip-chang
...
bugfix-han
| Author | SHA1 | Date | |
|---|---|---|---|
|
|
982b431118 | ||
|
|
b043d477dc | ||
|
|
06bcfb28e5 | ||
|
|
ca3b351bae | ||
|
|
b7e0f0a5e4 | ||
|
|
61f0ac2937 | ||
|
|
fca66eb558 | ||
|
|
359fc48fb4 | ||
|
|
d0efeb9770 | ||
|
|
3416532cd6 | ||
|
|
defc7a340e | ||
|
|
c197c062e1 | ||
|
|
77b59809ca | ||
|
|
f90b170e68 | ||
|
|
c93ca1841c | ||
|
|
57f604dff1 | ||
|
|
8499468749 | ||
|
|
7f6a13ea6c | ||
|
|
9874f0cbc7 | ||
|
|
72834a42fd | ||
|
|
724cb17224 | ||
|
|
4eb4b401a1 | ||
|
|
5d40e16c73 | ||
|
|
492bbce6b6 | ||
|
|
0394a56be5 | ||
|
|
7839551d6b | ||
|
|
9c5588c791 | ||
|
|
5a43a350de | ||
|
|
3c31f023ce | ||
|
|
4cbcc59461 | ||
|
|
4be0260381 | ||
|
|
957a3c1c16 | ||
|
|
85897e0bf9 | ||
|
|
63095f70ea | ||
|
|
8d5b0b5576 | ||
|
|
1b077abd93 | ||
|
|
32ea1a8721 | ||
|
|
fff32cef0d | ||
|
|
8fb146f3e4 | ||
|
|
770b0faa45 | ||
|
|
f6faa90340 | ||
|
|
669fd3ae0b | ||
|
|
17d37fb626 | ||
|
|
dfa7fc3a81 | ||
|
|
cd467df97a | ||
|
|
71bc2fed82 | ||
|
|
738fcfe01c | ||
|
|
3ebb2ab9ba | ||
|
|
ac98bc9144 | ||
|
|
3705ce6681 | ||
|
|
f7ea99412f | ||
|
|
d4715e2bc8 | ||
|
|
8567a83c47 | ||
|
|
77fdf59ae3 | ||
|
|
0e194aa4b4 | ||
|
|
2ba55bb477 | ||
|
|
4c759490da | ||
|
|
58a52c1f60 | ||
|
|
22638399c1 | ||
|
|
e3381776f2 | ||
|
|
26e2f21a80 | ||
|
|
b6009ae9ff | ||
|
|
b046d6ef32 | ||
|
|
e154a3cb7a | ||
|
|
1262700263 | ||
|
|
434c5813b9 | ||
|
|
0a3dc7d77b | ||
|
|
a7e296de65 | ||
|
|
bd0fbaaf27 | ||
|
|
0c111bd9ae | ||
|
|
ed9ac0b7fb | ||
|
|
743a3069bb | ||
|
|
fefc39427b | ||
|
|
2c6faa7c4e | ||
|
|
6168cd2899 | ||
|
|
f3c7c969d8 | ||
|
|
1355c2a245 | ||
|
|
96cf1a06df | ||
|
|
019a4a0375 | ||
|
|
db2f7b80ea | ||
|
|
bfabd7b094 | ||
|
|
d92dbfe765 | ||
|
|
67d2441334 | ||
|
|
3c30bc02d5 | ||
|
|
dcb54117d5 | ||
|
|
b1e32275dc | ||
|
|
e2a6865932 | ||
|
|
f04adb7202 | ||
|
|
1193a7f22c | ||
|
|
0b976827bb | ||
|
|
280e916033 | ||
|
|
5494e61a05 | ||
|
|
e461c0b819 | ||
|
|
d67c654f37 | ||
|
|
06ab34b6af | ||
|
|
ba8676c4ba | ||
|
|
4899c1a4f9 | ||
|
|
9bff1582f7 | ||
|
|
269e3bb7c5 | ||
|
|
9976f3f969 | ||
|
|
1f250aa868 | ||
|
|
1c08d9f150 | ||
|
|
9942107016 | ||
|
|
1eb5726cbf | ||
|
|
b3271ff7bb | ||
|
|
f82d3b648a | ||
|
|
034b1330d4 | ||
|
|
a7d005109f | ||
|
|
048c355e04 | ||
|
|
4026575b0b | ||
|
|
8c466b4826 | ||
|
|
6f072b42e8 | ||
|
|
e318253f31 | ||
|
|
f0f2fe94ce | ||
|
|
26f5c56ba4 | ||
|
|
a1c3107cd6 | ||
|
|
8fef3ff4ab | ||
|
|
baa25c9f9e | ||
|
|
488699b7d4 | ||
|
|
cf3a1ee3e3 | ||
|
|
daae43e9f9 | ||
|
|
cdeedaa65c | ||
|
|
3c9d2ded38 | ||
|
|
9f4364a130 | ||
|
|
5bd9eaf99d | ||
|
|
b1c51c0a65 | ||
|
|
232bd92389 | ||
|
|
e6173357a9 | ||
|
|
f2b8888aff | ||
|
|
9c46f175f9 | ||
|
|
1f27865fdf | ||
|
|
faa42d75e0 | ||
|
|
3b6e6d85bb | ||
|
|
30d6a272ce | ||
|
|
291700554e | ||
|
|
a82fad7059 | ||
|
|
c2fe5ae0d1 | ||
|
|
5beefdb7cc | ||
|
|
872bbba71c | ||
|
|
d578de1a35 | ||
|
|
cdc104be10 | ||
|
|
dd0eeca056 | ||
|
|
a95468be08 | ||
|
|
ace44d0e00 | ||
|
|
ebb8b88621 | ||
|
|
12fc2200de | ||
|
|
52d3d375ba | ||
|
|
08117089e6 | ||
|
|
2ba3a6d53f | ||
|
|
2f636553a9 | ||
|
|
0bde48b282 | ||
|
|
fae1164c0b | ||
|
|
169c293143 | ||
|
|
46cb5cff66 | ||
|
|
05584ea886 | ||
|
|
176a591357 | ||
|
|
15569f9592 | ||
|
|
5f9e475fe0 | ||
|
|
34b8784f50 | ||
|
|
2b054ced8c | ||
|
|
6553980cd5 | ||
|
|
7c12c47204 | ||
|
|
dbd9b470d7 | ||
|
|
83555a9991 | ||
|
|
5bfdb28bd2 | ||
|
|
31a6a6717b | ||
|
|
7da32f9ac3 | ||
|
|
bb732d3d2e | ||
|
|
485e55f9ed | ||
|
|
601a20ea49 | ||
|
|
76996b9eb8 | ||
|
|
fba2b1a39d | ||
|
|
4a91505af5 | ||
|
|
4841c79b4c | ||
|
|
2ba00d2e1d | ||
|
|
19c96f4bdd | ||
|
|
82b900fbf4 | ||
|
|
358a365303 | ||
|
|
a07ca4b136 | ||
|
|
ba8cf2c8cf | ||
|
|
3106b6688e | ||
|
|
2c83845dac | ||
|
|
111266d6fa | ||
|
|
ead610151f | ||
|
|
7e1e763989 | ||
|
|
327cc4af34 | ||
|
|
6008ff516e | ||
|
|
cdcf4b353f | ||
|
|
1ab70f8e86 | ||
|
|
8227c012a7 | ||
|
|
c113d5fb24 | ||
|
|
8c8d4066d7 | ||
|
|
277dc9e1c1 | ||
|
|
fc0fd1ce9d | ||
|
|
bd6127728a | ||
|
|
4101ae00c6 | ||
|
|
62f14df3cb | ||
|
|
560d465c59 | ||
|
|
7929aeddfc | ||
|
|
8294519f43 | ||
|
|
8ba8a220b6 | ||
|
|
aa3c8a9370 | ||
|
|
dbb5468cdc | ||
|
|
329c7620fb | ||
|
|
1f974bfbb0 | ||
|
|
437c8525af | ||
|
|
a2a1d5ae90 | ||
|
|
2566de2aae | ||
|
|
dfec8dbb39 | ||
|
|
5cefb16e52 | ||
|
|
341ae24b73 | ||
|
|
f47c2fb7f6 | ||
|
|
9d742446ab | ||
|
|
e3e022b0f4 | ||
|
|
6de4027c27 | ||
|
|
cda3837355 | ||
|
|
7983675325 | ||
|
|
eef56e52c6 | ||
|
|
8e3195f394 | ||
|
|
e17c2121f7 | ||
|
|
07e279b38d | ||
|
|
2c834cfe37 | ||
|
|
dbb5c666f0 | ||
|
|
70b3493866 | ||
|
|
3b11c474d1 | ||
|
|
890e1e6dcd | ||
|
|
6734fb91a2 |
25
.github/ISSUE_TEMPLATE/bug_report.md
vendored
25
.github/ISSUE_TEMPLATE/bug_report.md
vendored
@@ -1,25 +1,42 @@
|
|||||||
---
|
---
|
||||||
name: Bug report
|
name: Bug report
|
||||||
about: Create a report to help us improve
|
about: Create a bug report, if you don't follow this template, your report will be DELETED
|
||||||
title: ''
|
title: ''
|
||||||
labels: ''
|
labels: 'triage'
|
||||||
assignees: ''
|
assignees: 'dgtlmoon'
|
||||||
|
|
||||||
---
|
---
|
||||||
|
|
||||||
|
**DO NOT USE THIS FORM TO REPORT THAT A PARTICULAR WEBSITE IS NOT SCRAPING/WATCHING AS EXPECTED**
|
||||||
|
|
||||||
|
This form is only for direct bugs and feature requests todo directly with the software.
|
||||||
|
|
||||||
|
Please report watched websites (full URL and _any_ settings) that do not work with changedetection.io as expected [**IN THE DISCUSSION FORUMS**](https://github.com/dgtlmoon/changedetection.io/discussions) or your report will be deleted
|
||||||
|
|
||||||
|
CONSIDER TAKING OUT A SUBSCRIPTION FOR A SMALL PRICE PER MONTH, YOU GET THE BENEFIT OF USING OUR PAID PROXIES AND FURTHERING THE DEVELOPMENT OF CHANGEDETECTION.IO
|
||||||
|
|
||||||
|
THANK YOU
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
**Describe the bug**
|
**Describe the bug**
|
||||||
A clear and concise description of what the bug is.
|
A clear and concise description of what the bug is.
|
||||||
|
|
||||||
**Version**
|
**Version**
|
||||||
In the top right area: 0....
|
*Exact version* in the top right area: 0....
|
||||||
|
|
||||||
**To Reproduce**
|
**To Reproduce**
|
||||||
|
|
||||||
Steps to reproduce the behavior:
|
Steps to reproduce the behavior:
|
||||||
1. Go to '...'
|
1. Go to '...'
|
||||||
2. Click on '....'
|
2. Click on '....'
|
||||||
3. Scroll down to '....'
|
3. Scroll down to '....'
|
||||||
4. See error
|
4. See error
|
||||||
|
|
||||||
|
! ALWAYS INCLUDE AN EXAMPLE URL WHERE IT IS POSSIBLE TO RE-CREATE THE ISSUE - USE THE 'SHARE WATCH' FEATURE AND PASTE IN THE SHARE-LINK!
|
||||||
|
|
||||||
**Expected behavior**
|
**Expected behavior**
|
||||||
A clear and concise description of what you expected to happen.
|
A clear and concise description of what you expected to happen.
|
||||||
|
|
||||||
|
|||||||
4
.github/ISSUE_TEMPLATE/feature_request.md
vendored
4
.github/ISSUE_TEMPLATE/feature_request.md
vendored
@@ -1,8 +1,8 @@
|
|||||||
---
|
---
|
||||||
name: Feature request
|
name: Feature request
|
||||||
about: Suggest an idea for this project
|
about: Suggest an idea for this project
|
||||||
title: ''
|
title: '[feature]'
|
||||||
labels: ''
|
labels: 'enhancement'
|
||||||
assignees: ''
|
assignees: ''
|
||||||
|
|
||||||
---
|
---
|
||||||
|
|||||||
31
.github/test/Dockerfile-alpine
vendored
Normal file
31
.github/test/Dockerfile-alpine
vendored
Normal file
@@ -0,0 +1,31 @@
|
|||||||
|
# Taken from https://github.com/linuxserver/docker-changedetection.io/blob/main/Dockerfile
|
||||||
|
# Test that we can still build on Alpine (musl modified libc https://musl.libc.org/)
|
||||||
|
# Some packages wont install via pypi because they dont have a wheel available under this architecture.
|
||||||
|
|
||||||
|
FROM ghcr.io/linuxserver/baseimage-alpine:3.16
|
||||||
|
ENV PYTHONUNBUFFERED=1
|
||||||
|
|
||||||
|
COPY requirements.txt /requirements.txt
|
||||||
|
|
||||||
|
RUN \
|
||||||
|
apk add --update --no-cache --virtual=build-dependencies \
|
||||||
|
cargo \
|
||||||
|
g++ \
|
||||||
|
gcc \
|
||||||
|
libc-dev \
|
||||||
|
libffi-dev \
|
||||||
|
libxslt-dev \
|
||||||
|
make \
|
||||||
|
openssl-dev \
|
||||||
|
py3-wheel \
|
||||||
|
python3-dev \
|
||||||
|
zlib-dev && \
|
||||||
|
apk add --update --no-cache \
|
||||||
|
libxslt \
|
||||||
|
python3 \
|
||||||
|
py3-pip && \
|
||||||
|
echo "**** pip3 install test of changedetection.io ****" && \
|
||||||
|
pip3 install -U pip wheel setuptools && \
|
||||||
|
pip3 install -U --no-cache-dir --find-links https://wheel-index.linuxserver.io/alpine-3.16/ -r /requirements.txt && \
|
||||||
|
apk del --purge \
|
||||||
|
build-dependencies
|
||||||
15
.github/workflows/containers.yml
vendored
15
.github/workflows/containers.yml
vendored
@@ -85,8 +85,8 @@ jobs:
|
|||||||
version: latest
|
version: latest
|
||||||
driver-opts: image=moby/buildkit:master
|
driver-opts: image=moby/buildkit:master
|
||||||
|
|
||||||
# master always builds :latest
|
# master branch -> :dev container tag
|
||||||
- name: Build and push :latest
|
- name: Build and push :dev
|
||||||
id: docker_build
|
id: docker_build
|
||||||
if: ${{ github.ref }} == "refs/heads/master"
|
if: ${{ github.ref }} == "refs/heads/master"
|
||||||
uses: docker/build-push-action@v2
|
uses: docker/build-push-action@v2
|
||||||
@@ -95,12 +95,12 @@ jobs:
|
|||||||
file: ./Dockerfile
|
file: ./Dockerfile
|
||||||
push: true
|
push: true
|
||||||
tags: |
|
tags: |
|
||||||
${{ secrets.DOCKER_HUB_USERNAME }}/changedetection.io:latest,ghcr.io/${{ github.repository }}:latest
|
${{ secrets.DOCKER_HUB_USERNAME }}/changedetection.io:dev,ghcr.io/${{ github.repository }}:dev
|
||||||
platforms: linux/amd64,linux/arm64,linux/arm/v6,linux/arm/v7
|
platforms: linux/amd64,linux/arm64,linux/arm/v6,linux/arm/v7
|
||||||
cache-from: type=local,src=/tmp/.buildx-cache
|
cache-from: type=local,src=/tmp/.buildx-cache
|
||||||
cache-to: type=local,dest=/tmp/.buildx-cache
|
cache-to: type=local,dest=/tmp/.buildx-cache
|
||||||
|
|
||||||
# A new tagged release is required, which builds :tag
|
# A new tagged release is required, which builds :tag and :latest
|
||||||
- name: Build and push :tag
|
- name: Build and push :tag
|
||||||
id: docker_build_tag_release
|
id: docker_build_tag_release
|
||||||
if: github.event_name == 'release' && startsWith(github.event.release.tag_name, '0.')
|
if: github.event_name == 'release' && startsWith(github.event.release.tag_name, '0.')
|
||||||
@@ -110,7 +110,10 @@ jobs:
|
|||||||
file: ./Dockerfile
|
file: ./Dockerfile
|
||||||
push: true
|
push: true
|
||||||
tags: |
|
tags: |
|
||||||
${{ secrets.DOCKER_HUB_USERNAME }}/changedetection.io:${{ github.event.release.tag_name }},ghcr.io/dgtlmoon/changedetection.io:${{ github.event.release.tag_name }}
|
${{ secrets.DOCKER_HUB_USERNAME }}/changedetection.io:${{ github.event.release.tag_name }}
|
||||||
|
ghcr.io/dgtlmoon/changedetection.io:${{ github.event.release.tag_name }}
|
||||||
|
${{ secrets.DOCKER_HUB_USERNAME }}/changedetection.io:latest
|
||||||
|
ghcr.io/dgtlmoon/changedetection.io:latest
|
||||||
platforms: linux/amd64,linux/arm64,linux/arm/v6,linux/arm/v7
|
platforms: linux/amd64,linux/arm64,linux/arm/v6,linux/arm/v7
|
||||||
cache-from: type=local,src=/tmp/.buildx-cache
|
cache-from: type=local,src=/tmp/.buildx-cache
|
||||||
cache-to: type=local,dest=/tmp/.buildx-cache
|
cache-to: type=local,dest=/tmp/.buildx-cache
|
||||||
@@ -125,5 +128,3 @@ jobs:
|
|||||||
key: ${{ runner.os }}-buildx-${{ github.sha }}
|
key: ${{ runner.os }}-buildx-${{ github.sha }}
|
||||||
restore-keys: |
|
restore-keys: |
|
||||||
${{ runner.os }}-buildx-
|
${{ runner.os }}-buildx-
|
||||||
|
|
||||||
|
|
||||||
|
|||||||
66
.github/workflows/test-container-build.yml
vendored
Normal file
66
.github/workflows/test-container-build.yml
vendored
Normal file
@@ -0,0 +1,66 @@
|
|||||||
|
name: ChangeDetection.io Container Build Test
|
||||||
|
|
||||||
|
# Triggers the workflow on push or pull request events
|
||||||
|
|
||||||
|
# This line doesnt work, even tho it is the documented one
|
||||||
|
#on: [push, pull_request]
|
||||||
|
|
||||||
|
on:
|
||||||
|
push:
|
||||||
|
paths:
|
||||||
|
- requirements.txt
|
||||||
|
- Dockerfile
|
||||||
|
|
||||||
|
pull_request:
|
||||||
|
paths:
|
||||||
|
- requirements.txt
|
||||||
|
- Dockerfile
|
||||||
|
|
||||||
|
# Changes to requirements.txt packages and Dockerfile may or may not always be compatible with arm etc, so worth testing
|
||||||
|
# @todo: some kind of path filter for requirements.txt and Dockerfile
|
||||||
|
jobs:
|
||||||
|
test-container-build:
|
||||||
|
runs-on: ubuntu-latest
|
||||||
|
steps:
|
||||||
|
- uses: actions/checkout@v2
|
||||||
|
- name: Set up Python 3.9
|
||||||
|
uses: actions/setup-python@v2
|
||||||
|
with:
|
||||||
|
python-version: 3.9
|
||||||
|
|
||||||
|
# Just test that the build works, some libraries won't compile on ARM/rPi etc
|
||||||
|
- name: Set up QEMU
|
||||||
|
uses: docker/setup-qemu-action@v1
|
||||||
|
with:
|
||||||
|
image: tonistiigi/binfmt:latest
|
||||||
|
platforms: all
|
||||||
|
|
||||||
|
- name: Set up Docker Buildx
|
||||||
|
id: buildx
|
||||||
|
uses: docker/setup-buildx-action@v1
|
||||||
|
with:
|
||||||
|
install: true
|
||||||
|
version: latest
|
||||||
|
driver-opts: image=moby/buildkit:master
|
||||||
|
|
||||||
|
# https://github.com/dgtlmoon/changedetection.io/pull/1067
|
||||||
|
# Check we can still build under alpine/musl
|
||||||
|
- name: Test that the docker containers can build (musl via alpine check)
|
||||||
|
id: docker_build_musl
|
||||||
|
uses: docker/build-push-action@v2
|
||||||
|
with:
|
||||||
|
context: ./
|
||||||
|
file: ./.github/test/Dockerfile-alpine
|
||||||
|
platforms: linux/amd64,linux/arm64
|
||||||
|
|
||||||
|
- name: Test that the docker containers can build
|
||||||
|
id: docker_build
|
||||||
|
uses: docker/build-push-action@v2
|
||||||
|
# https://github.com/docker/build-push-action#customizing
|
||||||
|
with:
|
||||||
|
context: ./
|
||||||
|
file: ./Dockerfile
|
||||||
|
platforms: linux/arm/v7,linux/arm/v6,linux/amd64,linux/arm64,
|
||||||
|
cache-from: type=local,src=/tmp/.buildx-cache
|
||||||
|
cache-to: type=local,dest=/tmp/.buildx-cache
|
||||||
|
|
||||||
12
.github/workflows/test-only.yml
vendored
12
.github/workflows/test-only.yml
vendored
@@ -1,28 +1,25 @@
|
|||||||
name: ChangeDetection.io Test
|
name: ChangeDetection.io App Test
|
||||||
|
|
||||||
# Triggers the workflow on push or pull request events
|
# Triggers the workflow on push or pull request events
|
||||||
on: [push, pull_request]
|
on: [push, pull_request]
|
||||||
|
|
||||||
jobs:
|
jobs:
|
||||||
test-build:
|
test-application:
|
||||||
runs-on: ubuntu-latest
|
runs-on: ubuntu-latest
|
||||||
steps:
|
steps:
|
||||||
|
|
||||||
- uses: actions/checkout@v2
|
- uses: actions/checkout@v2
|
||||||
- name: Set up Python 3.9
|
- name: Set up Python 3.9
|
||||||
uses: actions/setup-python@v2
|
uses: actions/setup-python@v2
|
||||||
with:
|
with:
|
||||||
python-version: 3.9
|
python-version: 3.9
|
||||||
|
|
||||||
- name: Show env vars
|
|
||||||
run: set
|
|
||||||
|
|
||||||
- name: Install dependencies
|
- name: Install dependencies
|
||||||
run: |
|
run: |
|
||||||
python -m pip install --upgrade pip
|
python -m pip install --upgrade pip
|
||||||
pip install flake8 pytest
|
pip install flake8 pytest
|
||||||
if [ -f requirements.txt ]; then pip install -r requirements.txt; fi
|
if [ -f requirements.txt ]; then pip install -r requirements.txt; fi
|
||||||
if [ -f requirements-dev.txt ]; then pip install -r requirements-dev.txt; fi
|
if [ -f requirements-dev.txt ]; then pip install -r requirements-dev.txt; fi
|
||||||
|
|
||||||
- name: Lint with flake8
|
- name: Lint with flake8
|
||||||
run: |
|
run: |
|
||||||
# stop the build if there are Python syntax errors or undefined names
|
# stop the build if there are Python syntax errors or undefined names
|
||||||
@@ -39,7 +36,4 @@ jobs:
|
|||||||
# Each test is totally isolated and performs its own cleanup/reset
|
# Each test is totally isolated and performs its own cleanup/reset
|
||||||
cd changedetectionio; ./run_all_tests.sh
|
cd changedetectionio; ./run_all_tests.sh
|
||||||
|
|
||||||
# https://github.com/docker/build-push-action/blob/master/docs/advanced/test-before-push.md ?
|
|
||||||
# https://github.com/docker/buildx/issues/59 ? Needs to be one platform?
|
|
||||||
|
|
||||||
# https://github.com/docker/buildx/issues/495#issuecomment-918925854
|
|
||||||
|
|||||||
1
.gitignore
vendored
1
.gitignore
vendored
@@ -8,5 +8,6 @@ __pycache__
|
|||||||
build
|
build
|
||||||
dist
|
dist
|
||||||
venv
|
venv
|
||||||
|
test-datastore
|
||||||
*.egg-info*
|
*.egg-info*
|
||||||
.vscode/settings.json
|
.vscode/settings.json
|
||||||
|
|||||||
@@ -6,7 +6,7 @@ Otherwise, it's always best to PR into the `dev` branch.
|
|||||||
|
|
||||||
Please be sure that all new functionality has a matching test!
|
Please be sure that all new functionality has a matching test!
|
||||||
|
|
||||||
Use `pytest` to validate/test, you can run the existing tests as `pytest tests/test_notifications.py` for example
|
Use `pytest` to validate/test, you can run the existing tests as `pytest tests/test_notification.py` for example
|
||||||
|
|
||||||
```
|
```
|
||||||
pip3 install -r requirements-dev
|
pip3 install -r requirements-dev
|
||||||
|
|||||||
13
Dockerfile
13
Dockerfile
@@ -5,13 +5,14 @@ FROM python:3.8-slim as builder
|
|||||||
ARG CRYPTOGRAPHY_DONT_BUILD_RUST=1
|
ARG CRYPTOGRAPHY_DONT_BUILD_RUST=1
|
||||||
|
|
||||||
RUN apt-get update && apt-get install -y --no-install-recommends \
|
RUN apt-get update && apt-get install -y --no-install-recommends \
|
||||||
libssl-dev \
|
g++ \
|
||||||
libffi-dev \
|
|
||||||
gcc \
|
gcc \
|
||||||
libc-dev \
|
libc-dev \
|
||||||
|
libffi-dev \
|
||||||
|
libssl-dev \
|
||||||
libxslt-dev \
|
libxslt-dev \
|
||||||
zlib1g-dev \
|
make \
|
||||||
g++
|
zlib1g-dev
|
||||||
|
|
||||||
RUN mkdir /install
|
RUN mkdir /install
|
||||||
WORKDIR /install
|
WORKDIR /install
|
||||||
@@ -22,7 +23,8 @@ RUN pip install --target=/dependencies -r /requirements.txt
|
|||||||
|
|
||||||
# Playwright is an alternative to Selenium
|
# Playwright is an alternative to Selenium
|
||||||
# Excluded this package from requirements.txt to prevent arm/v6 and arm/v7 builds from failing
|
# Excluded this package from requirements.txt to prevent arm/v6 and arm/v7 builds from failing
|
||||||
RUN pip install --target=/dependencies playwright~=1.20 \
|
# https://github.com/dgtlmoon/changedetection.io/pull/1067 also musl/alpine (not supported)
|
||||||
|
RUN pip install --target=/dependencies playwright~=1.26 \
|
||||||
|| echo "WARN: Failed to install Playwright. The application can still run, but the Playwright option will be disabled."
|
|| echo "WARN: Failed to install Playwright. The application can still run, but the Playwright option will be disabled."
|
||||||
|
|
||||||
# Final image stage
|
# Final image stage
|
||||||
@@ -58,6 +60,7 @@ EXPOSE 5000
|
|||||||
|
|
||||||
# The actual flask app
|
# The actual flask app
|
||||||
COPY changedetectionio /app/changedetectionio
|
COPY changedetectionio /app/changedetectionio
|
||||||
|
|
||||||
# The eventlet server wrapper
|
# The eventlet server wrapper
|
||||||
COPY changedetection.py /app/changedetection.py
|
COPY changedetection.py /app/changedetection.py
|
||||||
|
|
||||||
|
|||||||
@@ -1,6 +1,8 @@
|
|||||||
|
recursive-include changedetectionio/api *
|
||||||
recursive-include changedetectionio/templates *
|
recursive-include changedetectionio/templates *
|
||||||
recursive-include changedetectionio/static *
|
recursive-include changedetectionio/static *
|
||||||
recursive-include changedetectionio/model *
|
recursive-include changedetectionio/model *
|
||||||
|
recursive-include changedetectionio/tests *
|
||||||
include changedetection.py
|
include changedetection.py
|
||||||
global-exclude *.pyc
|
global-exclude *.pyc
|
||||||
global-exclude node_modules
|
global-exclude node_modules
|
||||||
|
|||||||
@@ -1,42 +1,45 @@
|
|||||||
# changedetection.io
|
## Web Site Change Detection, Monitoring and Notification.
|
||||||

|
|
||||||
<a href="https://hub.docker.com/r/dgtlmoon/changedetection.io" target="_blank" title="Change detection docker hub">
|
|
||||||
<img src="https://img.shields.io/docker/pulls/dgtlmoon/changedetection.io" alt="Docker Pulls"/>
|
|
||||||
</a>
|
|
||||||
<a href="https://hub.docker.com/r/dgtlmoon/changedetection.io" target="_blank" title="Change detection docker hub">
|
|
||||||
<img src="https://img.shields.io/github/v/release/dgtlmoon/changedetection.io" alt="Change detection latest tag version"/>
|
|
||||||
</a>
|
|
||||||
|
|
||||||
## Self-hosted open source change monitoring of web pages.
|
Live your data-life pro-actively, track website content changes and receive notifications via Discord, Email, Slack, Telegram and 70+ more
|
||||||
|
|
||||||
_Know when web pages change! Stay ontop of new information!_
|
[<img src="https://raw.githubusercontent.com/dgtlmoon/changedetection.io/master/docs/screenshot.png" style="max-width:100%;" alt="Self-hosted web page change monitoring" title="Self-hosted web page change monitoring" />](https://lemonade.changedetection.io/start?src=pip)
|
||||||
|
|
||||||
Live your data-life *pro-actively* instead of *re-actively*, do not rely on manipulative social media for consuming important information.
|
|
||||||
|
|
||||||
|
|
||||||
<img src="https://raw.githubusercontent.com/dgtlmoon/changedetection.io/master/screenshot.png" style="max-width:100%;" alt="Self-hosted web page change monitoring" title="Self-hosted web page change monitoring" />
|
[**Don't have time? Let us host it for you! try our extremely affordable subscription use our proxies and support!**](https://lemonade.changedetection.io/start)
|
||||||
|
|
||||||
|
|
||||||
**Get your own private instance now! Let us host it for you!**
|
|
||||||
|
|
||||||
[**Try our $6.99/month subscription - unlimited checks, watches and notifications!**](https://lemonade.changedetection.io/start), choose from different geographical locations, let us handle everything for you.
|
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
#### Example use cases
|
#### Example use cases
|
||||||
|
|
||||||
Know when ...
|
- Products and services have a change in pricing
|
||||||
|
- _Out of stock notification_ and _Back In stock notification_
|
||||||
- Government department updates (changes are often only on their websites)
|
- Governmental department updates (changes are often only on their websites)
|
||||||
- Local government news (changes are often only on their websites)
|
|
||||||
- New software releases, security advisories when you're not on their mailing list.
|
- New software releases, security advisories when you're not on their mailing list.
|
||||||
- Festivals with changes
|
- Festivals with changes
|
||||||
- Realestate listing changes
|
- Realestate listing changes
|
||||||
|
- Know when your favourite whiskey is on sale, or other special deals are announced before anyone else
|
||||||
- COVID related news from government websites
|
- COVID related news from government websites
|
||||||
|
- University/organisation news from their website
|
||||||
- Detect and monitor changes in JSON API responses
|
- Detect and monitor changes in JSON API responses
|
||||||
- API monitoring and alerting
|
- JSON API monitoring and alerting
|
||||||
|
- Changes in legal and other documents
|
||||||
|
- Trigger API calls via notifications when text appears on a website
|
||||||
|
- Glue together APIs using the JSON filter and JSON notifications
|
||||||
|
- Create RSS feeds based on changes in web content
|
||||||
|
- Monitor HTML source code for unexpected changes, strengthen your PCI compliance
|
||||||
|
- You have a very sensitive list of URLs to watch and you do _not_ want to use the paid alternatives. (Remember, _you_ are the product)
|
||||||
|
|
||||||
|
_Need an actual Chrome runner with Javascript support? We support fetching via WebDriver and Playwright!</a>_
|
||||||
|
|
||||||
|
#### Key Features
|
||||||
|
|
||||||
|
- Lots of trigger filters, such as "Trigger on text", "Remove text by selector", "Ignore text", "Extract text", also using regular-expressions!
|
||||||
|
- Target elements with xPath and CSS Selectors, Easily monitor complex JSON with JSONPath or jq
|
||||||
|
- Switch between fast non-JS and Chrome JS based "fetchers"
|
||||||
|
- Easily specify how often a site should be checked
|
||||||
|
- Execute JS before extracting text (Good for logging in, see examples in the UI!)
|
||||||
|
- Override Request Headers, Specify `POST` or `GET` and other methods
|
||||||
|
- Use the "Visual Selector" to help target specific elements
|
||||||
|
|
||||||
**Get monitoring now!**
|
|
||||||
|
|
||||||
```bash
|
```bash
|
||||||
$ pip3 install changedetection.io
|
$ pip3 install changedetection.io
|
||||||
@@ -51,17 +54,5 @@ $ changedetection.io -d /path/to/empty/data/dir -p 5000
|
|||||||
|
|
||||||
Then visit http://127.0.0.1:5000 , You should now be able to access the UI.
|
Then visit http://127.0.0.1:5000 , You should now be able to access the UI.
|
||||||
|
|
||||||
### Features
|
|
||||||
- Website monitoring
|
|
||||||
- Change detection of content and analyses
|
|
||||||
- Filters on change (Select by CSS or JSON)
|
|
||||||
- Triggers (Wait for text, wait for regex)
|
|
||||||
- Notification support
|
|
||||||
- JSON API Monitoring
|
|
||||||
- Parse JSON embedded in HTML
|
|
||||||
- (Reverse) Proxy support
|
|
||||||
- Javascript support via WebDriver
|
|
||||||
- RaspberriPi (arm v6/v7/64 support)
|
|
||||||
|
|
||||||
See https://github.com/dgtlmoon/changedetection.io for more information.
|
See https://github.com/dgtlmoon/changedetection.io for more information.
|
||||||
|
|
||||||
|
|||||||
91
README.md
91
README.md
@@ -1,38 +1,35 @@
|
|||||||
# changedetection.io
|
## Web Site Change Detection, Monitoring and Notification.
|
||||||
|
|
||||||
|
_Live your data-life pro-actively, Detect website changes and perform meaningful actions, trigger notifications via Discord, Email, Slack, Telegram, API calls and many more._
|
||||||
|
|
||||||
|
|
||||||
|
[<img src="https://raw.githubusercontent.com/dgtlmoon/changedetection.io/master/docs/screenshot.png" style="max-width:100%;" alt="Self-hosted web page change monitoring" title="Self-hosted web page change monitoring" />](https://lemonade.changedetection.io/start?src=github)
|
||||||
|
|
||||||
[![Release Version][release-shield]][release-link] [![Docker Pulls][docker-pulls]][docker-link] [![License][license-shield]](LICENSE.md)
|
[![Release Version][release-shield]][release-link] [![Docker Pulls][docker-pulls]][docker-link] [![License][license-shield]](LICENSE.md)
|
||||||
|
|
||||||

|

|
||||||
|
|
||||||
## Self-Hosted, Open Source, Change Monitoring of Web Pages
|
[**Don't have time? Let us host it for you! try our $6.99/month subscription - use our proxies and support!**](https://lemonade.changedetection.io/start) , _half the price of other website change monitoring services and comes with unlimited watches & checks!_
|
||||||
|
|
||||||
_Know when web pages change! Stay ontop of new information!_
|
- Chrome browser included.
|
||||||
|
- Super fast, no registration needed setup.
|
||||||
Live your data-life *pro-actively* instead of *re-actively*.
|
- Start watching and receiving change notifications instantly.
|
||||||
|
|
||||||
Free, Open-source web page monitoring, notification and change detection. Don't have time? [**Try our $6.99/month subscription - unlimited checks and watches!**](https://lemonade.changedetection.io/start)
|
|
||||||
|
|
||||||
|
|
||||||
[<img src="https://raw.githubusercontent.com/dgtlmoon/changedetection.io/master/screenshot.png" style="max-width:100%;" alt="Self-hosted web page change monitoring" title="Self-hosted web page change monitoring" />](https://lemonade.changedetection.io/start)
|
Easily see what changed, examine by word, line, or individual character.
|
||||||
|
|
||||||
|
<img src="https://raw.githubusercontent.com/dgtlmoon/changedetection.io/master/docs/screenshot-diff.png" style="max-width:100%;" alt="Self-hosted web page change monitoring context difference " title="Self-hosted web page change monitoring context difference " />
|
||||||
**Get your own private instance now! Let us host it for you!**
|
|
||||||
|
|
||||||
[**Try our $6.99/month subscription - unlimited checks and watches!**](https://lemonade.changedetection.io/start) , _half the price of other website change monitoring services and comes with unlimited watches & checks!_
|
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
- Automatic Updates, Automatic Backups, No Heroku "paused application", don't miss a change!
|
|
||||||
- Javascript browser included
|
|
||||||
- Unlimited checks and watches!
|
|
||||||
|
|
||||||
|
|
||||||
#### Example use cases
|
#### Example use cases
|
||||||
|
|
||||||
- Products and services have a change in pricing
|
- Products and services have a change in pricing
|
||||||
|
- _Out of stock notification_ and _Back In stock notification_
|
||||||
- Governmental department updates (changes are often only on their websites)
|
- Governmental department updates (changes are often only on their websites)
|
||||||
- New software releases, security advisories when you're not on their mailing list.
|
- New software releases, security advisories when you're not on their mailing list.
|
||||||
- Festivals with changes
|
- Festivals with changes
|
||||||
- Realestate listing changes
|
- Realestate listing changes
|
||||||
|
- Know when your favourite whiskey is on sale, or other special deals are announced before anyone else
|
||||||
- COVID related news from government websites
|
- COVID related news from government websites
|
||||||
- University/organisation news from their website
|
- University/organisation news from their website
|
||||||
- Detect and monitor changes in JSON API responses
|
- Detect and monitor changes in JSON API responses
|
||||||
@@ -44,30 +41,48 @@ Free, Open-source web page monitoring, notification and change detection. Don't
|
|||||||
- Monitor HTML source code for unexpected changes, strengthen your PCI compliance
|
- Monitor HTML source code for unexpected changes, strengthen your PCI compliance
|
||||||
- You have a very sensitive list of URLs to watch and you do _not_ want to use the paid alternatives. (Remember, _you_ are the product)
|
- You have a very sensitive list of URLs to watch and you do _not_ want to use the paid alternatives. (Remember, _you_ are the product)
|
||||||
|
|
||||||
_Need an actual Chrome runner with Javascript support? We support fetching via WebDriver!</a>_
|
_Need an actual Chrome runner with Javascript support? We support fetching via WebDriver and Playwright!</a>_
|
||||||
|
|
||||||
|
#### Key Features
|
||||||
|
|
||||||
|
- Lots of trigger filters, such as "Trigger on text", "Remove text by selector", "Ignore text", "Extract text", also using regular-expressions!
|
||||||
|
- Target elements with xPath and CSS Selectors, Easily monitor complex JSON with JSONPath or jq
|
||||||
|
- Switch between fast non-JS and Chrome JS based "fetchers"
|
||||||
|
- Easily specify how often a site should be checked
|
||||||
|
- Execute JS before extracting text (Good for logging in, see examples in the UI!)
|
||||||
|
- Override Request Headers, Specify `POST` or `GET` and other methods
|
||||||
|
- Use the "Visual Selector" to help target specific elements
|
||||||
|
- Configurable [proxy per watch](https://github.com/dgtlmoon/changedetection.io/wiki/Proxy-configuration)
|
||||||
|
|
||||||
|
We [recommend and use Bright Data](https://brightdata.grsm.io/n0r16zf7eivq) global proxy services, Bright Data will match any first deposit up to $100 using our signup link.
|
||||||
|
|
||||||
## Screenshots
|
## Screenshots
|
||||||
|
|
||||||
Examining differences in content.
|
|
||||||
|
|
||||||
<img src="https://raw.githubusercontent.com/dgtlmoon/changedetection.io/master/screenshot-diff.png" style="max-width:100%;" alt="Self-hosted web page change monitoring context difference " title="Self-hosted web page change monitoring context difference " />
|
|
||||||
|
|
||||||
Please :star: star :star: this project and help it grow! https://github.com/dgtlmoon/changedetection.io/
|
Please :star: star :star: this project and help it grow! https://github.com/dgtlmoon/changedetection.io/
|
||||||
|
|
||||||
|
### Filter by elements using the Visual Selector tool.
|
||||||
|
|
||||||
|
Available when connected to a <a href="https://github.com/dgtlmoon/changedetection.io/wiki/Playwright-content-fetcher">playwright content fetcher</a> (included as part of our subscription service)
|
||||||
|
|
||||||
|
<img src="https://raw.githubusercontent.com/dgtlmoon/changedetection.io/master/docs/visualselector-anim.gif" style="max-width:100%;" alt="Self-hosted web page change monitoring context difference " title="Self-hosted web page change monitoring context difference " />
|
||||||
|
|
||||||
## Installation
|
## Installation
|
||||||
|
|
||||||
### Docker
|
### Docker
|
||||||
|
|
||||||
With Docker composer, just clone this repository and..
|
With Docker composer, just clone this repository and..
|
||||||
|
|
||||||
```bash
|
```bash
|
||||||
$ docker-compose up -d
|
$ docker-compose up -d
|
||||||
```
|
```
|
||||||
|
|
||||||
Docker standalone
|
Docker standalone
|
||||||
```bash
|
```bash
|
||||||
$ docker run -d --restart always -p "127.0.0.1:5000:5000" -v datastore-volume:/datastore --name changedetection.io dgtlmoon/changedetection.io
|
$ docker run -d --restart always -p "127.0.0.1:5000:5000" -v datastore-volume:/datastore --name changedetection.io dgtlmoon/changedetection.io
|
||||||
```
|
```
|
||||||
|
|
||||||
|
`:latest` tag is our latest stable release, `:dev` tag is our bleeding edge `master` branch.
|
||||||
|
|
||||||
### Windows
|
### Windows
|
||||||
|
|
||||||
See the install instructions at the wiki https://github.com/dgtlmoon/changedetection.io/wiki/Microsoft-Windows
|
See the install instructions at the wiki https://github.com/dgtlmoon/changedetection.io/wiki/Microsoft-Windows
|
||||||
@@ -105,9 +120,9 @@ See the wiki for more information https://github.com/dgtlmoon/changedetection.io
|
|||||||
|
|
||||||
|
|
||||||
## Filters
|
## Filters
|
||||||
XPath, JSONPath and CSS support comes baked in! You can be as specific as you need, use XPath exported from various XPath element query creation tools.
|
|
||||||
|
|
||||||
(We support LXML re:test, re:math and re:replace.)
|
XPath, JSONPath, jq, and CSS support comes baked in! You can be as specific as you need, use XPath exported from various XPath element query creation tools.
|
||||||
|
(We support LXML `re:test`, `re:math` and `re:replace`.)
|
||||||
|
|
||||||
## Notifications
|
## Notifications
|
||||||
|
|
||||||
@@ -129,23 +144,31 @@ Just some examples
|
|||||||
|
|
||||||
<a href="https://github.com/caronc/apprise#popular-notification-services">And everything else in this list!</a>
|
<a href="https://github.com/caronc/apprise#popular-notification-services">And everything else in this list!</a>
|
||||||
|
|
||||||
<img src="https://raw.githubusercontent.com/dgtlmoon/changedetection.io/master/screenshot-notifications.png" style="max-width:100%;" alt="Self-hosted web page change monitoring notifications" title="Self-hosted web page change monitoring notifications" />
|
<img src="https://raw.githubusercontent.com/dgtlmoon/changedetection.io/master/docs/screenshot-notifications.png" style="max-width:100%;" alt="Self-hosted web page change monitoring notifications" title="Self-hosted web page change monitoring notifications" />
|
||||||
|
|
||||||
Now you can also customise your notification content!
|
Now you can also customise your notification content!
|
||||||
|
|
||||||
## JSON API Monitoring
|
## JSON API Monitoring
|
||||||
|
|
||||||
Detect changes and monitor data in JSON API's by using the built-in JSONPath selectors as a filter / selector.
|
Detect changes and monitor data in JSON API's by using either JSONPath or jq to filter, parse, and restructure JSON as needed.
|
||||||
|
|
||||||

|

|
||||||
|
|
||||||
This will re-parse the JSON and apply formatting to the text, making it super easy to monitor and detect changes in JSON API results
|
This will re-parse the JSON and apply formatting to the text, making it super easy to monitor and detect changes in JSON API results
|
||||||
|
|
||||||

|

|
||||||
|
|
||||||
|
### JSONPath or jq?
|
||||||
|
|
||||||
|
For more complex parsing, filtering, and modifying of JSON data, jq is recommended due to the built-in operators and functions. Refer to the [documentation](https://stedolan.github.io/jq/manual/) for more specifc information on jq.
|
||||||
|
|
||||||
|
One big advantage of `jq` is that you can use logic in your JSON filter, such as filters to only show items that have a value greater than/less than etc.
|
||||||
|
|
||||||
|
See the wiki https://github.com/dgtlmoon/changedetection.io/wiki/JSON-Selector-Filter-help for more information and examples
|
||||||
|
|
||||||
### Parse JSON embedded in HTML!
|
### Parse JSON embedded in HTML!
|
||||||
|
|
||||||
When you enable a `json:` filter, you can even automatically extract and parse embedded JSON inside a HTML page! Amazingly handy for sites that build content based on JSON, such as many e-commerce websites.
|
When you enable a `json:` or `jq:` filter, you can even automatically extract and parse embedded JSON inside a HTML page! Amazingly handy for sites that build content based on JSON, such as many e-commerce websites.
|
||||||
|
|
||||||
```
|
```
|
||||||
<html>
|
<html>
|
||||||
@@ -155,11 +178,11 @@ When you enable a `json:` filter, you can even automatically extract and parse e
|
|||||||
</script>
|
</script>
|
||||||
```
|
```
|
||||||
|
|
||||||
`json:$.price` would give `23.50`, or you can extract the whole structure
|
`json:$.price` or `jq:.price` would give `23.50`, or you can extract the whole structure
|
||||||
|
|
||||||
## Proxy configuration
|
## Proxy Configuration
|
||||||
|
|
||||||
See the wiki https://github.com/dgtlmoon/changedetection.io/wiki/Proxy-configuration
|
See the wiki https://github.com/dgtlmoon/changedetection.io/wiki/Proxy-configuration , we also support using [BrightData proxy services where possible]( https://github.com/dgtlmoon/changedetection.io/wiki/Proxy-configuration#brightdata-proxy-support)
|
||||||
|
|
||||||
## Raspberry Pi support?
|
## Raspberry Pi support?
|
||||||
|
|
||||||
@@ -177,7 +200,7 @@ Or directly donate an amount PayPal [:
|
||||||
|
import sys
|
||||||
|
print('Shutdown: Got SIGCHLD')
|
||||||
|
# https://stackoverflow.com/questions/40453496/python-multiprocessing-capturing-signals-to-restart-child-processes-or-shut-do
|
||||||
|
pid, status = os.waitpid(-1, os.WNOHANG | os.WUNTRACED | os.WCONTINUED)
|
||||||
|
|
||||||
|
print('Sub-process: pid %d status %d' % (pid, status))
|
||||||
|
if status != 0:
|
||||||
|
sys.exit(1)
|
||||||
|
|
||||||
|
raise SystemExit
|
||||||
|
|
||||||
if __name__ == '__main__':
|
if __name__ == '__main__':
|
||||||
changedetection.main()
|
|
||||||
|
#signal.signal(signal.SIGCHLD, sigchld_handler)
|
||||||
|
|
||||||
|
# The only way I could find to get Flask to shutdown, is to wrap it and then rely on the subsystem issuing SIGTERM/SIGKILL
|
||||||
|
parse_process = multiprocessing.Process(target=changedetection.main)
|
||||||
|
parse_process.daemon = True
|
||||||
|
parse_process.start()
|
||||||
|
import time
|
||||||
|
|
||||||
|
try:
|
||||||
|
while True:
|
||||||
|
time.sleep(1)
|
||||||
|
|
||||||
|
except KeyboardInterrupt:
|
||||||
|
#parse_process.terminate() not needed, because this process will issue it to the sub-process anyway
|
||||||
|
print ("Exited - CTRL+C")
|
||||||
|
|||||||
1
changedetectionio/.gitignore
vendored
1
changedetectionio/.gitignore
vendored
@@ -1 +1,2 @@
|
|||||||
test-datastore
|
test-datastore
|
||||||
|
package-lock.json
|
||||||
|
|||||||
File diff suppressed because it is too large
Load Diff
0
changedetectionio/api/__init__.py
Normal file
0
changedetectionio/api/__init__.py
Normal file
158
changedetectionio/api/api_v1.py
Normal file
158
changedetectionio/api/api_v1.py
Normal file
@@ -0,0 +1,158 @@
|
|||||||
|
from flask_restful import abort, Resource
|
||||||
|
from flask import request, make_response
|
||||||
|
import validators
|
||||||
|
from . import auth
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
|
# https://www.w3.org/Protocols/rfc2616/rfc2616-sec10.html
|
||||||
|
|
||||||
|
class Watch(Resource):
|
||||||
|
def __init__(self, **kwargs):
|
||||||
|
# datastore is a black box dependency
|
||||||
|
self.datastore = kwargs['datastore']
|
||||||
|
self.update_q = kwargs['update_q']
|
||||||
|
|
||||||
|
# Get information about a single watch, excluding the history list (can be large)
|
||||||
|
# curl http://localhost:4000/api/v1/watch/<string:uuid>
|
||||||
|
# ?recheck=true
|
||||||
|
@auth.check_token
|
||||||
|
def get(self, uuid):
|
||||||
|
from copy import deepcopy
|
||||||
|
watch = deepcopy(self.datastore.data['watching'].get(uuid))
|
||||||
|
if not watch:
|
||||||
|
abort(404, message='No watch exists with the UUID of {}'.format(uuid))
|
||||||
|
|
||||||
|
if request.args.get('recheck'):
|
||||||
|
self.update_q.put((1, uuid))
|
||||||
|
return "OK", 200
|
||||||
|
|
||||||
|
# Return without history, get that via another API call
|
||||||
|
watch['history_n'] = watch.history_n
|
||||||
|
return watch
|
||||||
|
|
||||||
|
@auth.check_token
|
||||||
|
def delete(self, uuid):
|
||||||
|
if not self.datastore.data['watching'].get(uuid):
|
||||||
|
abort(400, message='No watch exists with the UUID of {}'.format(uuid))
|
||||||
|
|
||||||
|
self.datastore.delete(uuid)
|
||||||
|
return 'OK', 204
|
||||||
|
|
||||||
|
|
||||||
|
class WatchHistory(Resource):
|
||||||
|
def __init__(self, **kwargs):
|
||||||
|
# datastore is a black box dependency
|
||||||
|
self.datastore = kwargs['datastore']
|
||||||
|
|
||||||
|
# Get a list of available history for a watch by UUID
|
||||||
|
# curl http://localhost:4000/api/v1/watch/<string:uuid>/history
|
||||||
|
def get(self, uuid):
|
||||||
|
watch = self.datastore.data['watching'].get(uuid)
|
||||||
|
if not watch:
|
||||||
|
abort(404, message='No watch exists with the UUID of {}'.format(uuid))
|
||||||
|
return watch.history, 200
|
||||||
|
|
||||||
|
|
||||||
|
class WatchSingleHistory(Resource):
|
||||||
|
def __init__(self, **kwargs):
|
||||||
|
# datastore is a black box dependency
|
||||||
|
self.datastore = kwargs['datastore']
|
||||||
|
|
||||||
|
# Read a given history snapshot and return its content
|
||||||
|
# <string:timestamp> or "latest"
|
||||||
|
# curl http://localhost:4000/api/v1/watch/<string:uuid>/history/<int:timestamp>
|
||||||
|
@auth.check_token
|
||||||
|
def get(self, uuid, timestamp):
|
||||||
|
watch = self.datastore.data['watching'].get(uuid)
|
||||||
|
if not watch:
|
||||||
|
abort(404, message='No watch exists with the UUID of {}'.format(uuid))
|
||||||
|
|
||||||
|
if not len(watch.history):
|
||||||
|
abort(404, message='Watch found but no history exists for the UUID {}'.format(uuid))
|
||||||
|
|
||||||
|
if timestamp == 'latest':
|
||||||
|
timestamp = list(watch.history.keys())[-1]
|
||||||
|
|
||||||
|
with open(watch.history[timestamp], 'r') as f:
|
||||||
|
content = f.read()
|
||||||
|
|
||||||
|
response = make_response(content, 200)
|
||||||
|
response.mimetype = "text/plain"
|
||||||
|
return response
|
||||||
|
|
||||||
|
|
||||||
|
class CreateWatch(Resource):
|
||||||
|
def __init__(self, **kwargs):
|
||||||
|
# datastore is a black box dependency
|
||||||
|
self.datastore = kwargs['datastore']
|
||||||
|
self.update_q = kwargs['update_q']
|
||||||
|
|
||||||
|
@auth.check_token
|
||||||
|
def post(self):
|
||||||
|
# curl http://localhost:4000/api/v1/watch -H "Content-Type: application/json" -d '{"url": "https://my-nice.com", "tag": "one, two" }'
|
||||||
|
json_data = request.get_json()
|
||||||
|
tag = json_data['tag'].strip() if json_data.get('tag') else ''
|
||||||
|
|
||||||
|
if not validators.url(json_data['url'].strip()):
|
||||||
|
return "Invalid or unsupported URL", 400
|
||||||
|
|
||||||
|
extras = {'title': json_data['title'].strip()} if json_data.get('title') else {}
|
||||||
|
|
||||||
|
new_uuid = self.datastore.add_watch(url=json_data['url'].strip(), tag=tag, extras=extras)
|
||||||
|
self.update_q.put((1, new_uuid))
|
||||||
|
return {'uuid': new_uuid}, 201
|
||||||
|
|
||||||
|
# Return concise list of available watches and some very basic info
|
||||||
|
# curl http://localhost:4000/api/v1/watch|python -mjson.tool
|
||||||
|
# ?recheck_all=1 to recheck all
|
||||||
|
@auth.check_token
|
||||||
|
def get(self):
|
||||||
|
list = {}
|
||||||
|
for k, v in self.datastore.data['watching'].items():
|
||||||
|
list[k] = {'url': v['url'],
|
||||||
|
'title': v['title'],
|
||||||
|
'last_checked': v['last_checked'],
|
||||||
|
'last_changed': v.last_changed,
|
||||||
|
'last_error': v['last_error']}
|
||||||
|
|
||||||
|
if request.args.get('recheck_all'):
|
||||||
|
for uuid in self.datastore.data['watching'].keys():
|
||||||
|
self.update_q.put((1, uuid))
|
||||||
|
return {'status': "OK"}, 200
|
||||||
|
|
||||||
|
return list, 200
|
||||||
|
|
||||||
|
class SystemInfo(Resource):
|
||||||
|
def __init__(self, **kwargs):
|
||||||
|
# datastore is a black box dependency
|
||||||
|
self.datastore = kwargs['datastore']
|
||||||
|
self.update_q = kwargs['update_q']
|
||||||
|
|
||||||
|
@auth.check_token
|
||||||
|
def get(self):
|
||||||
|
import time
|
||||||
|
overdue_watches = []
|
||||||
|
|
||||||
|
# Check all watches and report which have not been checked but should have been
|
||||||
|
|
||||||
|
for uuid, watch in self.datastore.data.get('watching', {}).items():
|
||||||
|
# see if now - last_checked is greater than the time that should have been
|
||||||
|
# this is not super accurate (maybe they just edited it) but better than nothing
|
||||||
|
t = watch.threshold_seconds()
|
||||||
|
if not t:
|
||||||
|
# Use the system wide default
|
||||||
|
t = self.datastore.threshold_seconds
|
||||||
|
|
||||||
|
time_since_check = time.time() - watch.get('last_checked')
|
||||||
|
|
||||||
|
# Allow 5 minutes of grace time before we decide it's overdue
|
||||||
|
if time_since_check - (5 * 60) > t:
|
||||||
|
overdue_watches.append(uuid)
|
||||||
|
|
||||||
|
return {
|
||||||
|
'queue_size': self.update_q.qsize(),
|
||||||
|
'overdue_watches': overdue_watches,
|
||||||
|
'uptime': round(time.time() - self.datastore.start_time, 2),
|
||||||
|
'watch_count': len(self.datastore.data.get('watching', {}))
|
||||||
|
}, 200
|
||||||
33
changedetectionio/api/auth.py
Normal file
33
changedetectionio/api/auth.py
Normal file
@@ -0,0 +1,33 @@
|
|||||||
|
from flask import request, make_response, jsonify
|
||||||
|
from functools import wraps
|
||||||
|
|
||||||
|
|
||||||
|
# Simple API auth key comparison
|
||||||
|
# @todo - Maybe short lived token in the future?
|
||||||
|
|
||||||
|
def check_token(f):
|
||||||
|
@wraps(f)
|
||||||
|
def decorated(*args, **kwargs):
|
||||||
|
datastore = args[0].datastore
|
||||||
|
|
||||||
|
config_api_token_enabled = datastore.data['settings']['application'].get('api_access_token_enabled')
|
||||||
|
if not config_api_token_enabled:
|
||||||
|
return
|
||||||
|
|
||||||
|
try:
|
||||||
|
api_key_header = request.headers['x-api-key']
|
||||||
|
except KeyError:
|
||||||
|
return make_response(
|
||||||
|
jsonify("No authorization x-api-key header."), 403
|
||||||
|
)
|
||||||
|
|
||||||
|
config_api_token = datastore.data['settings']['application'].get('api_access_token')
|
||||||
|
|
||||||
|
if api_key_header != config_api_token:
|
||||||
|
return make_response(
|
||||||
|
jsonify("Invalid access - API key invalid."), 403
|
||||||
|
)
|
||||||
|
|
||||||
|
return f(*args, **kwargs)
|
||||||
|
|
||||||
|
return decorated
|
||||||
11
changedetectionio/apprise_asset.py
Normal file
11
changedetectionio/apprise_asset.py
Normal file
@@ -0,0 +1,11 @@
|
|||||||
|
import apprise
|
||||||
|
|
||||||
|
# Create our AppriseAsset and populate it with some of our new values:
|
||||||
|
# https://github.com/caronc/apprise/wiki/Development_API#the-apprise-asset-object
|
||||||
|
asset = apprise.AppriseAsset(
|
||||||
|
image_url_logo='https://raw.githubusercontent.com/dgtlmoon/changedetection.io/master/changedetectionio/static/images/avatar-256x256.png'
|
||||||
|
)
|
||||||
|
|
||||||
|
asset.app_id = "changedetection.io"
|
||||||
|
asset.app_desc = "ChangeDetection.io best and simplest website monitoring and change detection"
|
||||||
|
asset.app_url = "https://changedetection.io"
|
||||||
@@ -4,6 +4,7 @@
|
|||||||
|
|
||||||
import getopt
|
import getopt
|
||||||
import os
|
import os
|
||||||
|
import signal
|
||||||
import sys
|
import sys
|
||||||
|
|
||||||
import eventlet
|
import eventlet
|
||||||
@@ -11,7 +12,21 @@ import eventlet.wsgi
|
|||||||
from . import store, changedetection_app, content_fetcher
|
from . import store, changedetection_app, content_fetcher
|
||||||
from . import __version__
|
from . import __version__
|
||||||
|
|
||||||
|
# Only global so we can access it in the signal handler
|
||||||
|
datastore = None
|
||||||
|
app = None
|
||||||
|
|
||||||
|
def sigterm_handler(_signo, _stack_frame):
|
||||||
|
global app
|
||||||
|
global datastore
|
||||||
|
# app.config.exit.set()
|
||||||
|
print('Shutdown: Got SIGTERM, DB saved to disk')
|
||||||
|
datastore.sync_to_json()
|
||||||
|
# raise SystemExit
|
||||||
|
|
||||||
def main():
|
def main():
|
||||||
|
global datastore
|
||||||
|
global app
|
||||||
ssl_mode = False
|
ssl_mode = False
|
||||||
host = ''
|
host = ''
|
||||||
port = os.environ.get('PORT') or 5000
|
port = os.environ.get('PORT') or 5000
|
||||||
@@ -35,11 +50,6 @@ def main():
|
|||||||
create_datastore_dir = False
|
create_datastore_dir = False
|
||||||
|
|
||||||
for opt, arg in opts:
|
for opt, arg in opts:
|
||||||
# if opt == '--purge':
|
|
||||||
# Remove history, the actual files you need to delete manually.
|
|
||||||
# for uuid, watch in datastore.data['watching'].items():
|
|
||||||
# watch.update({'history': {}, 'last_checked': 0, 'last_changed': 0, 'previous_md5': None})
|
|
||||||
|
|
||||||
if opt == '-s':
|
if opt == '-s':
|
||||||
ssl_mode = True
|
ssl_mode = True
|
||||||
|
|
||||||
@@ -72,9 +82,12 @@ def main():
|
|||||||
"Or use the -C parameter to create the directory.".format(app_config['datastore_path']), file=sys.stderr)
|
"Or use the -C parameter to create the directory.".format(app_config['datastore_path']), file=sys.stderr)
|
||||||
sys.exit(2)
|
sys.exit(2)
|
||||||
|
|
||||||
|
|
||||||
datastore = store.ChangeDetectionStore(datastore_path=app_config['datastore_path'], version_tag=__version__)
|
datastore = store.ChangeDetectionStore(datastore_path=app_config['datastore_path'], version_tag=__version__)
|
||||||
app = changedetection_app(app_config, datastore)
|
app = changedetection_app(app_config, datastore)
|
||||||
|
|
||||||
|
signal.signal(signal.SIGTERM, sigterm_handler)
|
||||||
|
|
||||||
# Go into cleanup mode
|
# Go into cleanup mode
|
||||||
if do_cleanup:
|
if do_cleanup:
|
||||||
datastore.remove_unused_snapshots()
|
datastore.remove_unused_snapshots()
|
||||||
@@ -89,6 +102,14 @@ def main():
|
|||||||
has_password=datastore.data['settings']['application']['password'] != False
|
has_password=datastore.data['settings']['application']['password'] != False
|
||||||
)
|
)
|
||||||
|
|
||||||
|
# Monitored websites will not receive a Referer header when a user clicks on an outgoing link.
|
||||||
|
# @Note: Incompatible with password login (and maybe other features) for now, submit a PR!
|
||||||
|
@app.after_request
|
||||||
|
def hide_referrer(response):
|
||||||
|
if os.getenv("HIDE_REFERER", False):
|
||||||
|
response.headers["Referrer-Policy"] = "no-referrer"
|
||||||
|
return response
|
||||||
|
|
||||||
# Proxy sub-directory support
|
# Proxy sub-directory support
|
||||||
# Set environment var USE_X_SETTINGS=1 on this script
|
# Set environment var USE_X_SETTINGS=1 on this script
|
||||||
# And then in your proxy_pass settings
|
# And then in your proxy_pass settings
|
||||||
@@ -111,4 +132,3 @@ def main():
|
|||||||
else:
|
else:
|
||||||
eventlet.wsgi.server(eventlet.listen((host, int(port))), app)
|
eventlet.wsgi.server(eventlet.listen((host, int(port))), app)
|
||||||
|
|
||||||
|
|
||||||
|
|||||||
@@ -1,35 +1,212 @@
|
|||||||
from abc import ABC, abstractmethod
|
from abc import ABC, abstractmethod
|
||||||
import chardet
|
import chardet
|
||||||
|
import json
|
||||||
import os
|
import os
|
||||||
import requests
|
import requests
|
||||||
import time
|
import time
|
||||||
import urllib3.exceptions
|
|
||||||
import sys
|
import sys
|
||||||
|
|
||||||
|
|
||||||
class EmptyReply(Exception):
|
class Non200ErrorCodeReceived(Exception):
|
||||||
def __init__(self, status_code, url):
|
def __init__(self, status_code, url, screenshot=None, xpath_data=None, page_html=None):
|
||||||
# Set this so we can use it in other parts of the app
|
# Set this so we can use it in other parts of the app
|
||||||
self.status_code = status_code
|
self.status_code = status_code
|
||||||
self.url = url
|
self.url = url
|
||||||
|
self.screenshot = screenshot
|
||||||
|
self.xpath_data = xpath_data
|
||||||
|
self.page_text = None
|
||||||
|
|
||||||
|
if page_html:
|
||||||
|
from changedetectionio import html_tools
|
||||||
|
self.page_text = html_tools.html_to_text(page_html)
|
||||||
return
|
return
|
||||||
|
|
||||||
pass
|
|
||||||
|
|
||||||
|
class JSActionExceptions(Exception):
|
||||||
|
def __init__(self, status_code, url, screenshot, message=''):
|
||||||
|
self.status_code = status_code
|
||||||
|
self.url = url
|
||||||
|
self.screenshot = screenshot
|
||||||
|
self.message = message
|
||||||
|
return
|
||||||
|
|
||||||
|
class PageUnloadable(Exception):
|
||||||
|
def __init__(self, status_code, url, screenshot=False, message=False):
|
||||||
|
# Set this so we can use it in other parts of the app
|
||||||
|
self.status_code = status_code
|
||||||
|
self.url = url
|
||||||
|
self.screenshot = screenshot
|
||||||
|
self.message = message
|
||||||
|
return
|
||||||
|
|
||||||
|
class EmptyReply(Exception):
|
||||||
|
def __init__(self, status_code, url, screenshot=None):
|
||||||
|
# Set this so we can use it in other parts of the app
|
||||||
|
self.status_code = status_code
|
||||||
|
self.url = url
|
||||||
|
self.screenshot = screenshot
|
||||||
|
return
|
||||||
|
|
||||||
|
class ScreenshotUnavailable(Exception):
|
||||||
|
def __init__(self, status_code, url, page_html=None):
|
||||||
|
# Set this so we can use it in other parts of the app
|
||||||
|
self.status_code = status_code
|
||||||
|
self.url = url
|
||||||
|
if page_html:
|
||||||
|
from html_tools import html_to_text
|
||||||
|
self.page_text = html_to_text(page_html)
|
||||||
|
return
|
||||||
|
|
||||||
|
class ReplyWithContentButNoText(Exception):
|
||||||
|
def __init__(self, status_code, url, screenshot=None):
|
||||||
|
# Set this so we can use it in other parts of the app
|
||||||
|
self.status_code = status_code
|
||||||
|
self.url = url
|
||||||
|
self.screenshot = screenshot
|
||||||
|
return
|
||||||
|
|
||||||
class Fetcher():
|
class Fetcher():
|
||||||
error = None
|
error = None
|
||||||
status_code = None
|
status_code = None
|
||||||
content = None
|
content = None
|
||||||
headers = None
|
headers = None
|
||||||
|
|
||||||
|
fetcher_description = "No description"
|
||||||
|
webdriver_js_execute_code = None
|
||||||
|
xpath_element_js = """
|
||||||
|
// Include the getXpath script directly, easier than fetching
|
||||||
|
!function(e,n){"object"==typeof exports&&"undefined"!=typeof module?module.exports=n():"function"==typeof define&&define.amd?define(n):(e=e||self).getXPath=n()}(this,function(){return function(e){var n=e;if(n&&n.id)return'//*[@id="'+n.id+'"]';for(var o=[];n&&Node.ELEMENT_NODE===n.nodeType;){for(var i=0,r=!1,d=n.previousSibling;d;)d.nodeType!==Node.DOCUMENT_TYPE_NODE&&d.nodeName===n.nodeName&&i++,d=d.previousSibling;for(d=n.nextSibling;d;){if(d.nodeName===n.nodeName){r=!0;break}d=d.nextSibling}o.push((n.prefix?n.prefix+":":"")+n.localName+(i||r?"["+(i+1)+"]":"")),n=n.parentNode}return o.length?"/"+o.reverse().join("/"):""}});
|
||||||
|
|
||||||
|
|
||||||
|
const findUpTag = (el) => {
|
||||||
|
let r = el
|
||||||
|
chained_css = [];
|
||||||
|
depth=0;
|
||||||
|
|
||||||
|
// Strategy 1: Keep going up until we hit an ID tag, imagine it's like #list-widget div h4
|
||||||
|
while (r.parentNode) {
|
||||||
|
if(depth==5) {
|
||||||
|
break;
|
||||||
|
}
|
||||||
|
if('' !==r.id) {
|
||||||
|
chained_css.unshift("#"+CSS.escape(r.id));
|
||||||
|
final_selector= chained_css.join(' > ');
|
||||||
|
// Be sure theres only one, some sites have multiples of the same ID tag :-(
|
||||||
|
if (window.document.querySelectorAll(final_selector).length ==1 ) {
|
||||||
|
return final_selector;
|
||||||
|
}
|
||||||
|
return null;
|
||||||
|
} else {
|
||||||
|
chained_css.unshift(r.tagName.toLowerCase());
|
||||||
|
}
|
||||||
|
r=r.parentNode;
|
||||||
|
depth+=1;
|
||||||
|
}
|
||||||
|
return null;
|
||||||
|
}
|
||||||
|
|
||||||
|
|
||||||
|
// @todo - if it's SVG or IMG, go into image diff mode
|
||||||
|
var elements = window.document.querySelectorAll("div,span,form,table,tbody,tr,td,a,p,ul,li,h1,h2,h3,h4, header, footer, section, article, aside, details, main, nav, section, summary");
|
||||||
|
var size_pos=[];
|
||||||
|
// after page fetch, inject this JS
|
||||||
|
// build a map of all elements and their positions (maybe that only include text?)
|
||||||
|
var bbox;
|
||||||
|
for (var i = 0; i < elements.length; i++) {
|
||||||
|
bbox = elements[i].getBoundingClientRect();
|
||||||
|
|
||||||
|
// forget really small ones
|
||||||
|
if (bbox['width'] <20 && bbox['height'] < 20 ) {
|
||||||
|
continue;
|
||||||
|
}
|
||||||
|
|
||||||
|
// @todo the getXpath kind of sucks, it doesnt know when there is for example just one ID sometimes
|
||||||
|
// it should not traverse when we know we can anchor off just an ID one level up etc..
|
||||||
|
// maybe, get current class or id, keep traversing up looking for only class or id until there is just one match
|
||||||
|
|
||||||
|
// 1st primitive - if it has class, try joining it all and select, if theres only one.. well thats us.
|
||||||
|
xpath_result=false;
|
||||||
|
|
||||||
|
try {
|
||||||
|
var d= findUpTag(elements[i]);
|
||||||
|
if (d) {
|
||||||
|
xpath_result =d;
|
||||||
|
}
|
||||||
|
} catch (e) {
|
||||||
|
console.log(e);
|
||||||
|
}
|
||||||
|
|
||||||
|
// You could swap it and default to getXpath and then try the smarter one
|
||||||
|
// default back to the less intelligent one
|
||||||
|
if (!xpath_result) {
|
||||||
|
try {
|
||||||
|
// I've seen on FB and eBay that this doesnt work
|
||||||
|
// ReferenceError: getXPath is not defined at eval (eval at evaluate (:152:29), <anonymous>:67:20) at UtilityScript.evaluate (<anonymous>:159:18) at UtilityScript.<anonymous> (<anonymous>:1:44)
|
||||||
|
xpath_result = getXPath(elements[i]);
|
||||||
|
} catch (e) {
|
||||||
|
console.log(e);
|
||||||
|
continue;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
if(window.getComputedStyle(elements[i]).visibility === "hidden") {
|
||||||
|
continue;
|
||||||
|
}
|
||||||
|
|
||||||
|
size_pos.push({
|
||||||
|
xpath: xpath_result,
|
||||||
|
width: Math.round(bbox['width']),
|
||||||
|
height: Math.round(bbox['height']),
|
||||||
|
left: Math.floor(bbox['left']),
|
||||||
|
top: Math.floor(bbox['top']),
|
||||||
|
childCount: elements[i].childElementCount
|
||||||
|
});
|
||||||
|
}
|
||||||
|
|
||||||
|
|
||||||
|
// inject the current one set in the include_filters, which may be a CSS rule
|
||||||
|
// used for displaying the current one in VisualSelector, where its not one we generated.
|
||||||
|
if (include_filters.length) {
|
||||||
|
q=false;
|
||||||
|
try {
|
||||||
|
// is it xpath?
|
||||||
|
if (include_filters.startsWith('/') || include_filters.startsWith('xpath:')) {
|
||||||
|
q=document.evaluate(include_filters.replace('xpath:',''), document, null, XPathResult.FIRST_ORDERED_NODE_TYPE, null).singleNodeValue;
|
||||||
|
} else {
|
||||||
|
q=document.querySelector(include_filters);
|
||||||
|
}
|
||||||
|
} catch (e) {
|
||||||
|
// Maybe catch DOMException and alert?
|
||||||
|
console.log(e);
|
||||||
|
}
|
||||||
|
bbox=false;
|
||||||
|
if(q) {
|
||||||
|
bbox = q.getBoundingClientRect();
|
||||||
|
}
|
||||||
|
|
||||||
|
if (bbox && bbox['width'] >0 && bbox['height']>0) {
|
||||||
|
size_pos.push({
|
||||||
|
xpath: include_filters,
|
||||||
|
width: bbox['width'],
|
||||||
|
height: bbox['height'],
|
||||||
|
left: bbox['left'],
|
||||||
|
top: bbox['top'],
|
||||||
|
childCount: q.childElementCount
|
||||||
|
});
|
||||||
|
}
|
||||||
|
}
|
||||||
|
// Window.width required for proper scaling in the frontend
|
||||||
|
return {'size_pos':size_pos, 'browser_width': window.innerWidth};
|
||||||
|
"""
|
||||||
|
xpath_data = None
|
||||||
|
|
||||||
# Will be needed in the future by the VisualSelector, always get this where possible.
|
# Will be needed in the future by the VisualSelector, always get this where possible.
|
||||||
screenshot = False
|
screenshot = False
|
||||||
fetcher_description = "No description"
|
|
||||||
system_http_proxy = os.getenv('HTTP_PROXY')
|
system_http_proxy = os.getenv('HTTP_PROXY')
|
||||||
system_https_proxy = os.getenv('HTTPS_PROXY')
|
system_https_proxy = os.getenv('HTTPS_PROXY')
|
||||||
|
|
||||||
# Time ONTOP of the system defined env minimum time
|
# Time ONTOP of the system defined env minimum time
|
||||||
render_extract_delay=0
|
render_extract_delay = 0
|
||||||
|
|
||||||
@abstractmethod
|
@abstractmethod
|
||||||
def get_error(self):
|
def get_error(self):
|
||||||
@@ -42,7 +219,8 @@ class Fetcher():
|
|||||||
request_headers,
|
request_headers,
|
||||||
request_body,
|
request_body,
|
||||||
request_method,
|
request_method,
|
||||||
ignore_status_codes=False):
|
ignore_status_codes=False,
|
||||||
|
current_include_filters=None):
|
||||||
# Should set self.error, self.status_code and self.content
|
# Should set self.error, self.status_code and self.content
|
||||||
pass
|
pass
|
||||||
|
|
||||||
@@ -115,7 +293,15 @@ class base_html_playwright(Fetcher):
|
|||||||
|
|
||||||
# allow per-watch proxy selection override
|
# allow per-watch proxy selection override
|
||||||
if proxy_override:
|
if proxy_override:
|
||||||
self.proxy = {'server': proxy_override}
|
# https://playwright.dev/docs/network#http-proxy
|
||||||
|
from urllib.parse import urlparse
|
||||||
|
parsed = urlparse(proxy_override)
|
||||||
|
proxy_url = "{}://{}:{}".format(parsed.scheme, parsed.hostname, parsed.port)
|
||||||
|
self.proxy = {'server': proxy_url}
|
||||||
|
if parsed.username:
|
||||||
|
self.proxy['username'] = parsed.username
|
||||||
|
if parsed.password:
|
||||||
|
self.proxy['password'] = parsed.password
|
||||||
|
|
||||||
def run(self,
|
def run(self,
|
||||||
url,
|
url,
|
||||||
@@ -123,55 +309,145 @@ class base_html_playwright(Fetcher):
|
|||||||
request_headers,
|
request_headers,
|
||||||
request_body,
|
request_body,
|
||||||
request_method,
|
request_method,
|
||||||
ignore_status_codes=False):
|
ignore_status_codes=False,
|
||||||
|
current_include_filters=None):
|
||||||
|
|
||||||
from playwright.sync_api import sync_playwright
|
from playwright.sync_api import sync_playwright
|
||||||
import playwright._impl._api_types
|
import playwright._impl._api_types
|
||||||
from playwright._impl._api_types import Error, TimeoutError
|
from playwright._impl._api_types import Error, TimeoutError
|
||||||
|
response = None
|
||||||
|
|
||||||
with sync_playwright() as p:
|
with sync_playwright() as p:
|
||||||
browser_type = getattr(p, self.browser_type)
|
browser_type = getattr(p, self.browser_type)
|
||||||
|
|
||||||
# Seemed to cause a connection Exception even tho I can see it connect
|
# Seemed to cause a connection Exception even tho I can see it connect
|
||||||
# self.browser = browser_type.connect(self.command_executor, timeout=timeout*1000)
|
# self.browser = browser_type.connect(self.command_executor, timeout=timeout*1000)
|
||||||
browser = browser_type.connect_over_cdp(self.command_executor, timeout=timeout * 1000)
|
# 60,000 connection timeout only
|
||||||
|
browser = browser_type.connect_over_cdp(self.command_executor, timeout=60000)
|
||||||
|
|
||||||
# Set user agent to prevent Cloudflare from blocking the browser
|
# Set user agent to prevent Cloudflare from blocking the browser
|
||||||
# Use the default one configured in the App.py model that's passed from fetch_site_status.py
|
# Use the default one configured in the App.py model that's passed from fetch_site_status.py
|
||||||
context = browser.new_context(
|
context = browser.new_context(
|
||||||
user_agent=request_headers['User-Agent'] if request_headers.get('User-Agent') else 'Mozilla/5.0',
|
user_agent=request_headers['User-Agent'] if request_headers.get('User-Agent') else 'Mozilla/5.0',
|
||||||
proxy=self.proxy
|
proxy=self.proxy,
|
||||||
|
# This is needed to enable JavaScript execution on GitHub and others
|
||||||
|
bypass_csp=True,
|
||||||
|
# Should never be needed
|
||||||
|
accept_downloads=False
|
||||||
)
|
)
|
||||||
|
|
||||||
|
if len(request_headers):
|
||||||
|
context.set_extra_http_headers(request_headers)
|
||||||
|
|
||||||
page = context.new_page()
|
page = context.new_page()
|
||||||
page.set_viewport_size({"width": 1280, "height": 1024})
|
|
||||||
try:
|
try:
|
||||||
response = page.goto(url, timeout=timeout * 1000, wait_until='commit')
|
page.set_default_navigation_timeout(90000)
|
||||||
# Wait_until = commit
|
page.set_default_timeout(90000)
|
||||||
# - `'commit'` - consider operation to be finished when network response is received and the document started loading.
|
|
||||||
# Better to not use any smarts from Playwright and just wait an arbitrary number of seconds
|
# Listen for all console events and handle errors
|
||||||
# This seemed to solve nearly all 'TimeoutErrors'
|
page.on("console", lambda msg: print(f"Playwright console: Watch URL: {url} {msg.type}: {msg.text} {msg.args}"))
|
||||||
extra_wait = int(os.getenv("WEBDRIVER_DELAY_BEFORE_CONTENT_READY", 5)) + self.render_extract_delay
|
|
||||||
page.wait_for_timeout(extra_wait * 1000)
|
# Bug - never set viewport size BEFORE page.goto
|
||||||
|
|
||||||
|
# Waits for the next navigation. Using Python context manager
|
||||||
|
# prevents a race condition between clicking and waiting for a navigation.
|
||||||
|
with page.expect_navigation():
|
||||||
|
response = page.goto(url, wait_until='load')
|
||||||
|
|
||||||
|
|
||||||
except playwright._impl._api_types.TimeoutError as e:
|
except playwright._impl._api_types.TimeoutError as e:
|
||||||
raise EmptyReply(url=url, status_code=None)
|
context.close()
|
||||||
|
browser.close()
|
||||||
|
# This can be ok, we will try to grab what we could retrieve
|
||||||
|
pass
|
||||||
|
|
||||||
|
except Exception as e:
|
||||||
|
print("other exception when page.goto")
|
||||||
|
print(str(e))
|
||||||
|
context.close()
|
||||||
|
browser.close()
|
||||||
|
raise PageUnloadable(url=url, status_code=None, message=e.message)
|
||||||
|
|
||||||
if response is None:
|
if response is None:
|
||||||
|
context.close()
|
||||||
|
browser.close()
|
||||||
|
print("response object was none")
|
||||||
raise EmptyReply(url=url, status_code=None)
|
raise EmptyReply(url=url, status_code=None)
|
||||||
|
|
||||||
if len(page.content().strip()) == 0:
|
|
||||||
raise EmptyReply(url=url, status_code=None)
|
|
||||||
|
|
||||||
self.status_code = response.status
|
# Removed browser-set-size, seemed to be needed to make screenshots work reliably in older playwright versions
|
||||||
|
# Was causing exceptions like 'waiting for page but content is changing' etc
|
||||||
|
# https://www.browserstack.com/docs/automate/playwright/change-browser-window-size 1280x720 should be the default
|
||||||
|
|
||||||
|
extra_wait = int(os.getenv("WEBDRIVER_DELAY_BEFORE_CONTENT_READY", 5)) + self.render_extract_delay
|
||||||
|
time.sleep(extra_wait)
|
||||||
|
|
||||||
|
if self.webdriver_js_execute_code is not None:
|
||||||
|
try:
|
||||||
|
page.evaluate(self.webdriver_js_execute_code)
|
||||||
|
except Exception as e:
|
||||||
|
# Is it possible to get a screenshot?
|
||||||
|
error_screenshot = False
|
||||||
|
try:
|
||||||
|
page.screenshot(type='jpeg',
|
||||||
|
clip={'x': 1.0, 'y': 1.0, 'width': 1280, 'height': 1024},
|
||||||
|
quality=1)
|
||||||
|
|
||||||
|
# The actual screenshot
|
||||||
|
error_screenshot = page.screenshot(type='jpeg',
|
||||||
|
full_page=True,
|
||||||
|
quality=int(os.getenv("PLAYWRIGHT_SCREENSHOT_QUALITY", 72)))
|
||||||
|
except Exception as s:
|
||||||
|
pass
|
||||||
|
|
||||||
|
raise JSActionExceptions(status_code=response.status, screenshot=error_screenshot, message=str(e), url=url)
|
||||||
|
|
||||||
|
else:
|
||||||
|
# JS eval was run, now we also wait some time if possible to let the page settle
|
||||||
|
if self.render_extract_delay:
|
||||||
|
page.wait_for_timeout(self.render_extract_delay * 1000)
|
||||||
|
|
||||||
|
page.wait_for_timeout(500)
|
||||||
|
|
||||||
self.content = page.content()
|
self.content = page.content()
|
||||||
|
self.status_code = response.status
|
||||||
self.headers = response.all_headers()
|
self.headers = response.all_headers()
|
||||||
|
|
||||||
|
if current_include_filters is not None:
|
||||||
|
page.evaluate("var include_filters={}".format(json.dumps(current_include_filters)))
|
||||||
|
else:
|
||||||
|
page.evaluate("var include_filters=''")
|
||||||
|
|
||||||
|
self.xpath_data = page.evaluate("async () => {" + self.xpath_element_js + "}")
|
||||||
|
|
||||||
|
# Bug 3 in Playwright screenshot handling
|
||||||
# Some bug where it gives the wrong screenshot size, but making a request with the clip set first seems to solve it
|
# Some bug where it gives the wrong screenshot size, but making a request with the clip set first seems to solve it
|
||||||
# JPEG is better here because the screenshots can be very very large
|
# JPEG is better here because the screenshots can be very very large
|
||||||
page.screenshot(type='jpeg', clip={'x': 1.0, 'y': 1.0, 'width': 1280, 'height': 1024})
|
|
||||||
self.screenshot = page.screenshot(type='jpeg', full_page=True, quality=90)
|
# Screenshots also travel via the ws:// (websocket) meaning that the binary data is base64 encoded
|
||||||
|
# which will significantly increase the IO size between the server and client, it's recommended to use the lowest
|
||||||
|
# acceptable screenshot quality here
|
||||||
|
try:
|
||||||
|
# Quality set to 1 because it's not used, just used as a work-around for a bug, no need to change this.
|
||||||
|
page.screenshot(type='jpeg', clip={'x': 1.0, 'y': 1.0, 'width': 1280, 'height': 1024}, quality=1)
|
||||||
|
# The actual screenshot
|
||||||
|
self.screenshot = page.screenshot(type='jpeg', full_page=True, quality=int(os.getenv("PLAYWRIGHT_SCREENSHOT_QUALITY", 72)))
|
||||||
|
except Exception as e:
|
||||||
|
context.close()
|
||||||
|
browser.close()
|
||||||
|
raise ScreenshotUnavailable(url=url, status_code=None)
|
||||||
|
|
||||||
|
if len(self.content.strip()) == 0:
|
||||||
|
context.close()
|
||||||
|
browser.close()
|
||||||
|
print("Content was empty")
|
||||||
|
raise EmptyReply(url=url, status_code=None, screenshot=self.screenshot)
|
||||||
|
|
||||||
context.close()
|
context.close()
|
||||||
browser.close()
|
browser.close()
|
||||||
|
|
||||||
|
if not ignore_status_codes and self.status_code!=200:
|
||||||
|
raise Non200ErrorCodeReceived(url=url, status_code=self.status_code, page_html=self.content, screenshot=self.screenshot)
|
||||||
|
|
||||||
class base_html_webdriver(Fetcher):
|
class base_html_webdriver(Fetcher):
|
||||||
if os.getenv("WEBDRIVER_URL"):
|
if os.getenv("WEBDRIVER_URL"):
|
||||||
@@ -220,7 +496,8 @@ class base_html_webdriver(Fetcher):
|
|||||||
request_headers,
|
request_headers,
|
||||||
request_body,
|
request_body,
|
||||||
request_method,
|
request_method,
|
||||||
ignore_status_codes=False):
|
ignore_status_codes=False,
|
||||||
|
current_include_filters=None):
|
||||||
|
|
||||||
from selenium import webdriver
|
from selenium import webdriver
|
||||||
from selenium.webdriver.common.desired_capabilities import DesiredCapabilities
|
from selenium.webdriver.common.desired_capabilities import DesiredCapabilities
|
||||||
@@ -240,6 +517,14 @@ class base_html_webdriver(Fetcher):
|
|||||||
self.quit()
|
self.quit()
|
||||||
raise
|
raise
|
||||||
|
|
||||||
|
self.driver.set_window_size(1280, 1024)
|
||||||
|
self.driver.implicitly_wait(int(os.getenv("WEBDRIVER_DELAY_BEFORE_CONTENT_READY", 5)))
|
||||||
|
|
||||||
|
if self.webdriver_js_execute_code is not None:
|
||||||
|
self.driver.execute_script(self.webdriver_js_execute_code)
|
||||||
|
# Selenium doesn't automatically wait for actions as good as Playwright, so wait again
|
||||||
|
self.driver.implicitly_wait(int(os.getenv("WEBDRIVER_DELAY_BEFORE_CONTENT_READY", 5)))
|
||||||
|
|
||||||
# @todo - how to check this? is it possible?
|
# @todo - how to check this? is it possible?
|
||||||
self.status_code = 200
|
self.status_code = 200
|
||||||
# @todo somehow we should try to get this working for WebDriver
|
# @todo somehow we should try to get this working for WebDriver
|
||||||
@@ -249,8 +534,8 @@ class base_html_webdriver(Fetcher):
|
|||||||
time.sleep(int(os.getenv("WEBDRIVER_DELAY_BEFORE_CONTENT_READY", 5)) + self.render_extract_delay)
|
time.sleep(int(os.getenv("WEBDRIVER_DELAY_BEFORE_CONTENT_READY", 5)) + self.render_extract_delay)
|
||||||
self.content = self.driver.page_source
|
self.content = self.driver.page_source
|
||||||
self.headers = {}
|
self.headers = {}
|
||||||
|
|
||||||
self.screenshot = self.driver.get_screenshot_as_png()
|
self.screenshot = self.driver.get_screenshot_as_png()
|
||||||
self.quit()
|
|
||||||
|
|
||||||
# Does the connection to the webdriver work? run a test connection.
|
# Does the connection to the webdriver work? run a test connection.
|
||||||
def is_ready(self):
|
def is_ready(self):
|
||||||
@@ -287,9 +572,15 @@ class html_requests(Fetcher):
|
|||||||
request_headers,
|
request_headers,
|
||||||
request_body,
|
request_body,
|
||||||
request_method,
|
request_method,
|
||||||
ignore_status_codes=False):
|
ignore_status_codes=False,
|
||||||
|
current_include_filters=None):
|
||||||
|
|
||||||
proxies={}
|
# Make requests use a more modern looking user-agent
|
||||||
|
if not 'User-Agent' in request_headers:
|
||||||
|
request_headers['User-Agent'] = os.getenv("DEFAULT_SETTINGS_HEADERS_USERAGENT",
|
||||||
|
'Mozilla/5.0 (X11; Linux x86_64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/87.0.4280.66 Safari/537.36')
|
||||||
|
|
||||||
|
proxies = {}
|
||||||
|
|
||||||
# Allows override the proxy on a per-request basis
|
# Allows override the proxy on a per-request basis
|
||||||
if self.proxy_override:
|
if self.proxy_override:
|
||||||
@@ -317,10 +608,14 @@ class html_requests(Fetcher):
|
|||||||
if encoding:
|
if encoding:
|
||||||
r.encoding = encoding
|
r.encoding = encoding
|
||||||
|
|
||||||
|
if not r.content or not len(r.content):
|
||||||
|
raise EmptyReply(url=url, status_code=r.status_code)
|
||||||
|
|
||||||
# @todo test this
|
# @todo test this
|
||||||
# @todo maybe you really want to test zero-byte return pages?
|
# @todo maybe you really want to test zero-byte return pages?
|
||||||
if (not ignore_status_codes and not r) or not r.content or not len(r.content):
|
if r.status_code != 200 and not ignore_status_codes:
|
||||||
raise EmptyReply(url=url, status_code=r.status_code)
|
# maybe check with content works?
|
||||||
|
raise Non200ErrorCodeReceived(url=url, status_code=r.status_code, page_html=r.text)
|
||||||
|
|
||||||
self.status_code = r.status_code
|
self.status_code = r.status_code
|
||||||
self.content = r.text
|
self.content = r.text
|
||||||
|
|||||||
@@ -1,4 +1,5 @@
|
|||||||
import hashlib
|
import hashlib
|
||||||
|
import logging
|
||||||
import os
|
import os
|
||||||
import re
|
import re
|
||||||
import time
|
import time
|
||||||
@@ -9,52 +10,48 @@ from changedetectionio import content_fetcher, html_tools
|
|||||||
urllib3.disable_warnings(urllib3.exceptions.InsecureRequestWarning)
|
urllib3.disable_warnings(urllib3.exceptions.InsecureRequestWarning)
|
||||||
|
|
||||||
|
|
||||||
|
class FilterNotFoundInResponse(ValueError):
|
||||||
|
def __init__(self, msg):
|
||||||
|
ValueError.__init__(self, msg)
|
||||||
|
|
||||||
|
|
||||||
# Some common stuff here that can be moved to a base class
|
# Some common stuff here that can be moved to a base class
|
||||||
|
# (set_proxy_from_list)
|
||||||
class perform_site_check():
|
class perform_site_check():
|
||||||
|
screenshot = None
|
||||||
|
xpath_data = None
|
||||||
|
|
||||||
def __init__(self, *args, datastore, **kwargs):
|
def __init__(self, *args, datastore, **kwargs):
|
||||||
super().__init__(*args, **kwargs)
|
super().__init__(*args, **kwargs)
|
||||||
self.datastore = datastore
|
self.datastore = datastore
|
||||||
|
|
||||||
# If there was a proxy list enabled, figure out what proxy_args/which proxy to use
|
# Doesn't look like python supports forward slash auto enclosure in re.findall
|
||||||
# if watch.proxy use that
|
# So convert it to inline flag "foobar(?i)" type configuration
|
||||||
# fetcher.proxy_override = watch.proxy or main config proxy
|
def forward_slash_enclosed_regex_to_options(self, regex):
|
||||||
# Allows override the proxy on a per-request basis
|
res = re.search(r'^/(.*?)/(\w+)$', regex, re.IGNORECASE)
|
||||||
# ALWAYS use the first one is nothing selected
|
|
||||||
|
|
||||||
def set_proxy_from_list(self, watch):
|
if res:
|
||||||
proxy_args = None
|
regex = res.group(1)
|
||||||
if self.datastore.proxy_list is None:
|
regex += '(?{})'.format(res.group(2))
|
||||||
return None
|
|
||||||
|
|
||||||
# If its a valid one
|
|
||||||
if any([watch['proxy'] in p for p in self.datastore.proxy_list]):
|
|
||||||
proxy_args = watch['proxy']
|
|
||||||
|
|
||||||
# not valid (including None), try the system one
|
|
||||||
else:
|
else:
|
||||||
system_proxy = self.datastore.data['settings']['requests']['proxy']
|
regex += '(?{})'.format('i')
|
||||||
# Is not None and exists
|
|
||||||
if any([system_proxy in p for p in self.datastore.proxy_list]):
|
|
||||||
proxy_args = system_proxy
|
|
||||||
|
|
||||||
# Fallback - Did not resolve anything, use the first available
|
return regex
|
||||||
if proxy_args is None:
|
|
||||||
proxy_args = self.datastore.proxy_list[0][0]
|
|
||||||
|
|
||||||
return proxy_args
|
|
||||||
|
|
||||||
def run(self, uuid):
|
def run(self, uuid):
|
||||||
timestamp = int(time.time()) # used for storage etc too
|
from copy import deepcopy
|
||||||
|
|
||||||
changed_detected = False
|
changed_detected = False
|
||||||
screenshot = False # as bytes
|
screenshot = False # as bytes
|
||||||
stripped_text_from_html = ""
|
stripped_text_from_html = ""
|
||||||
|
|
||||||
watch = self.datastore.data['watching'][uuid]
|
# DeepCopy so we can be sure we don't accidently change anything by reference
|
||||||
|
watch = deepcopy(self.datastore.data['watching'].get(uuid))
|
||||||
|
|
||||||
|
if not watch:
|
||||||
|
return
|
||||||
|
|
||||||
# Protect against file:// access
|
# Protect against file:// access
|
||||||
if re.search(r'^file', watch['url'], re.IGNORECASE) and not os.getenv('ALLOW_FILE_URI', False):
|
if re.search(r'^file', watch.get('url', ''), re.IGNORECASE) and not os.getenv('ALLOW_FILE_URI', False):
|
||||||
raise Exception(
|
raise Exception(
|
||||||
"file:// type access is denied for security reasons."
|
"file:// type access is denied for security reasons."
|
||||||
)
|
)
|
||||||
@@ -62,10 +59,10 @@ class perform_site_check():
|
|||||||
# Unset any existing notification error
|
# Unset any existing notification error
|
||||||
update_obj = {'last_notification_error': False, 'last_error': False}
|
update_obj = {'last_notification_error': False, 'last_error': False}
|
||||||
|
|
||||||
extra_headers = self.datastore.get_val(uuid, 'headers')
|
extra_headers = watch.get('headers', [])
|
||||||
|
|
||||||
# Tweak the base config with the per-watch ones
|
# Tweak the base config with the per-watch ones
|
||||||
request_headers = self.datastore.data['settings']['headers'].copy()
|
request_headers = deepcopy(self.datastore.data['settings']['headers'])
|
||||||
request_headers.update(extra_headers)
|
request_headers.update(extra_headers)
|
||||||
|
|
||||||
# https://github.com/psf/requests/issues/4525
|
# https://github.com/psf/requests/issues/4525
|
||||||
@@ -74,11 +71,13 @@ class perform_site_check():
|
|||||||
if 'Accept-Encoding' in request_headers and "br" in request_headers['Accept-Encoding']:
|
if 'Accept-Encoding' in request_headers and "br" in request_headers['Accept-Encoding']:
|
||||||
request_headers['Accept-Encoding'] = request_headers['Accept-Encoding'].replace(', br', '')
|
request_headers['Accept-Encoding'] = request_headers['Accept-Encoding'].replace(', br', '')
|
||||||
|
|
||||||
timeout = self.datastore.data['settings']['requests']['timeout']
|
timeout = self.datastore.data['settings']['requests'].get('timeout')
|
||||||
url = self.datastore.get_val(uuid, 'url')
|
|
||||||
request_body = self.datastore.get_val(uuid, 'body')
|
url = watch.link
|
||||||
request_method = self.datastore.get_val(uuid, 'method')
|
|
||||||
ignore_status_code = self.datastore.get_val(uuid, 'ignore_status_codes')
|
request_body = self.datastore.data['watching'][uuid].get('body')
|
||||||
|
request_method = self.datastore.data['watching'][uuid].get('method')
|
||||||
|
ignore_status_codes = self.datastore.data['watching'][uuid].get('ignore_status_codes', False)
|
||||||
|
|
||||||
# source: support
|
# source: support
|
||||||
is_source = False
|
is_source = False
|
||||||
@@ -87,24 +86,36 @@ class perform_site_check():
|
|||||||
is_source = True
|
is_source = True
|
||||||
|
|
||||||
# Pluggable content fetcher
|
# Pluggable content fetcher
|
||||||
prefer_backend = watch['fetch_backend']
|
prefer_backend = watch.get('fetch_backend')
|
||||||
if hasattr(content_fetcher, prefer_backend):
|
if hasattr(content_fetcher, prefer_backend):
|
||||||
klass = getattr(content_fetcher, prefer_backend)
|
klass = getattr(content_fetcher, prefer_backend)
|
||||||
else:
|
else:
|
||||||
# If the klass doesnt exist, just use a default
|
# If the klass doesnt exist, just use a default
|
||||||
klass = getattr(content_fetcher, "html_requests")
|
klass = getattr(content_fetcher, "html_requests")
|
||||||
|
|
||||||
proxy_args = self.set_proxy_from_list(watch)
|
proxy_id = self.datastore.get_preferred_proxy_for_watch(uuid=uuid)
|
||||||
fetcher = klass(proxy_override=proxy_args)
|
proxy_url = None
|
||||||
|
if proxy_id:
|
||||||
|
proxy_url = self.datastore.proxy_list.get(proxy_id).get('url')
|
||||||
|
print("UUID {} Using proxy {}".format(uuid, proxy_url))
|
||||||
|
|
||||||
|
fetcher = klass(proxy_override=proxy_url)
|
||||||
|
|
||||||
# Configurable per-watch or global extra delay before extracting text (for webDriver types)
|
# Configurable per-watch or global extra delay before extracting text (for webDriver types)
|
||||||
system_webdriver_delay = self.datastore.data['settings']['application'].get('webdriver_delay', None)
|
system_webdriver_delay = self.datastore.data['settings']['application'].get('webdriver_delay', None)
|
||||||
if watch['webdriver_delay'] is not None:
|
if watch['webdriver_delay'] is not None:
|
||||||
fetcher.render_extract_delay = watch['webdriver_delay']
|
fetcher.render_extract_delay = watch.get('webdriver_delay')
|
||||||
elif system_webdriver_delay is not None:
|
elif system_webdriver_delay is not None:
|
||||||
fetcher.render_extract_delay = system_webdriver_delay
|
fetcher.render_extract_delay = system_webdriver_delay
|
||||||
|
|
||||||
fetcher.run(url, timeout, request_headers, request_body, request_method, ignore_status_code)
|
if watch.get('webdriver_js_execute_code') is not None and watch.get('webdriver_js_execute_code').strip():
|
||||||
|
fetcher.webdriver_js_execute_code = watch.get('webdriver_js_execute_code')
|
||||||
|
|
||||||
|
fetcher.run(url, timeout, request_headers, request_body, request_method, ignore_status_codes, watch.get('include_filters'))
|
||||||
|
fetcher.quit()
|
||||||
|
|
||||||
|
self.screenshot = fetcher.screenshot
|
||||||
|
self.xpath_data = fetcher.xpath_data
|
||||||
|
|
||||||
# Fetching complete, now filters
|
# Fetching complete, now filters
|
||||||
# @todo move to class / maybe inside of fetcher abstract base?
|
# @todo move to class / maybe inside of fetcher abstract base?
|
||||||
@@ -124,27 +135,32 @@ class perform_site_check():
|
|||||||
is_html = False
|
is_html = False
|
||||||
is_json = False
|
is_json = False
|
||||||
|
|
||||||
css_filter_rule = watch['css_filter']
|
include_filters_rule = watch.get('include_filters', [])
|
||||||
|
# include_filters_rule = watch['include_filters']
|
||||||
subtractive_selectors = watch.get(
|
subtractive_selectors = watch.get(
|
||||||
"subtractive_selectors", []
|
"subtractive_selectors", []
|
||||||
) + self.datastore.data["settings"]["application"].get(
|
) + self.datastore.data["settings"]["application"].get(
|
||||||
"global_subtractive_selectors", []
|
"global_subtractive_selectors", []
|
||||||
)
|
)
|
||||||
|
|
||||||
has_filter_rule = css_filter_rule and len(css_filter_rule.strip())
|
has_filter_rule = include_filters_rule and len("".join(include_filters_rule).strip())
|
||||||
has_subtractive_selectors = subtractive_selectors and len(subtractive_selectors[0].strip())
|
has_subtractive_selectors = subtractive_selectors and len(subtractive_selectors[0].strip())
|
||||||
|
|
||||||
if is_json and not has_filter_rule:
|
if is_json and not has_filter_rule:
|
||||||
css_filter_rule = "json:$"
|
include_filters_rule.append("json:$")
|
||||||
has_filter_rule = True
|
has_filter_rule = True
|
||||||
|
|
||||||
if has_filter_rule:
|
if has_filter_rule:
|
||||||
if 'json:' in css_filter_rule:
|
json_filter_prefixes = ['json:', 'jq:']
|
||||||
stripped_text_from_html = html_tools.extract_json_as_string(content=fetcher.content, jsonpath_filter=css_filter_rule)
|
for filter in include_filters_rule:
|
||||||
|
if any(prefix in filter for prefix in json_filter_prefixes):
|
||||||
|
stripped_text_from_html += html_tools.extract_json_as_string(content=fetcher.content, json_filter=filter)
|
||||||
is_html = False
|
is_html = False
|
||||||
|
|
||||||
if is_html or is_source:
|
if is_html or is_source:
|
||||||
|
|
||||||
# CSS Filter, extract the HTML that matches and feed that into the existing inscriptis::get_text
|
# CSS Filter, extract the HTML that matches and feed that into the existing inscriptis::get_text
|
||||||
|
fetcher.content = html_tools.workarounds_for_obfuscations(fetcher.content)
|
||||||
html_content = fetcher.content
|
html_content = fetcher.content
|
||||||
|
|
||||||
# If not JSON, and if it's not text/plain..
|
# If not JSON, and if it's not text/plain..
|
||||||
@@ -154,35 +170,43 @@ class perform_site_check():
|
|||||||
else:
|
else:
|
||||||
# Then we assume HTML
|
# Then we assume HTML
|
||||||
if has_filter_rule:
|
if has_filter_rule:
|
||||||
|
html_content = ""
|
||||||
|
for filter_rule in include_filters_rule:
|
||||||
# For HTML/XML we offer xpath as an option, just start a regular xPath "/.."
|
# For HTML/XML we offer xpath as an option, just start a regular xPath "/.."
|
||||||
if css_filter_rule[0] == '/' or css_filter_rule.startswith('xpath:'):
|
if filter_rule[0] == '/' or filter_rule.startswith('xpath:'):
|
||||||
html_content = html_tools.xpath_filter(xpath_filter=css_filter_rule.replace('xpath:', ''),
|
html_content += html_tools.xpath_filter(xpath_filter=filter_rule.replace('xpath:', ''),
|
||||||
html_content=fetcher.content)
|
html_content=fetcher.content,
|
||||||
|
append_pretty_line_formatting=not is_source)
|
||||||
else:
|
else:
|
||||||
# CSS Filter, extract the HTML that matches and feed that into the existing inscriptis::get_text
|
# CSS Filter, extract the HTML that matches and feed that into the existing inscriptis::get_text
|
||||||
html_content = html_tools.css_filter(css_filter=css_filter_rule, html_content=fetcher.content)
|
html_content += html_tools.include_filters(include_filters=filter_rule,
|
||||||
|
html_content=fetcher.content,
|
||||||
|
append_pretty_line_formatting=not is_source)
|
||||||
|
|
||||||
|
if not html_content.strip():
|
||||||
|
raise FilterNotFoundInResponse(include_filters_rule)
|
||||||
|
|
||||||
if has_subtractive_selectors:
|
if has_subtractive_selectors:
|
||||||
html_content = html_tools.element_removal(subtractive_selectors, html_content)
|
html_content = html_tools.element_removal(subtractive_selectors, html_content)
|
||||||
|
|
||||||
if not is_source:
|
if is_source:
|
||||||
|
stripped_text_from_html = html_content
|
||||||
|
else:
|
||||||
# extract text
|
# extract text
|
||||||
|
do_anchor = self.datastore.data["settings"]["application"].get("render_anchor_tag_content", False)
|
||||||
stripped_text_from_html = \
|
stripped_text_from_html = \
|
||||||
html_tools.html_to_text(
|
html_tools.html_to_text(
|
||||||
html_content,
|
html_content,
|
||||||
render_anchor_tag_content=self.datastore.data["settings"][
|
render_anchor_tag_content=do_anchor
|
||||||
"application"].get(
|
|
||||||
"render_anchor_tag_content", False)
|
|
||||||
)
|
)
|
||||||
|
|
||||||
elif is_source:
|
|
||||||
stripped_text_from_html = html_content
|
|
||||||
|
|
||||||
# Re #340 - return the content before the 'ignore text' was applied
|
# Re #340 - return the content before the 'ignore text' was applied
|
||||||
text_content_before_ignored_filter = stripped_text_from_html.encode('utf-8')
|
text_content_before_ignored_filter = stripped_text_from_html.encode('utf-8')
|
||||||
|
|
||||||
# Re #340 - return the content before the 'ignore text' was applied
|
# Treat pages with no renderable text content as a change? No by default
|
||||||
text_content_before_ignored_filter = stripped_text_from_html.encode('utf-8')
|
empty_pages_are_a_change = self.datastore.data['settings']['application'].get('empty_pages_are_a_change', False)
|
||||||
|
if not is_json and not empty_pages_are_a_change and len(stripped_text_from_html.strip()) == 0:
|
||||||
|
raise content_fetcher.ReplyWithContentButNoText(url=url, status_code=fetcher.get_last_status_code(), screenshot=screenshot)
|
||||||
|
|
||||||
# We rely on the actual text in the html output.. many sites have random script vars etc,
|
# We rely on the actual text in the html output.. many sites have random script vars etc,
|
||||||
# in the future we'll implement other mechanisms.
|
# in the future we'll implement other mechanisms.
|
||||||
@@ -197,38 +221,91 @@ class perform_site_check():
|
|||||||
else:
|
else:
|
||||||
stripped_text_from_html = stripped_text_from_html.encode('utf8')
|
stripped_text_from_html = stripped_text_from_html.encode('utf8')
|
||||||
|
|
||||||
|
# 615 Extract text by regex
|
||||||
|
extract_text = watch.get('extract_text', [])
|
||||||
|
if len(extract_text) > 0:
|
||||||
|
regex_matched_output = []
|
||||||
|
for s_re in extract_text:
|
||||||
|
# incase they specified something in '/.../x'
|
||||||
|
regex = self.forward_slash_enclosed_regex_to_options(s_re)
|
||||||
|
result = re.findall(regex.encode('utf-8'), stripped_text_from_html)
|
||||||
|
|
||||||
|
for l in result:
|
||||||
|
if type(l) is tuple:
|
||||||
|
# @todo - some formatter option default (between groups)
|
||||||
|
regex_matched_output += list(l) + [b'\n']
|
||||||
|
else:
|
||||||
|
# @todo - some formatter option default (between each ungrouped result)
|
||||||
|
regex_matched_output += [l] + [b'\n']
|
||||||
|
|
||||||
|
# Now we will only show what the regex matched
|
||||||
|
stripped_text_from_html = b''
|
||||||
|
text_content_before_ignored_filter = b''
|
||||||
|
if regex_matched_output:
|
||||||
|
# @todo some formatter for presentation?
|
||||||
|
stripped_text_from_html = b''.join(regex_matched_output)
|
||||||
|
text_content_before_ignored_filter = stripped_text_from_html
|
||||||
|
|
||||||
# Re #133 - if we should strip whitespaces from triggering the change detected comparison
|
# Re #133 - if we should strip whitespaces from triggering the change detected comparison
|
||||||
if self.datastore.data['settings']['application'].get('ignore_whitespace', False):
|
if self.datastore.data['settings']['application'].get('ignore_whitespace', False):
|
||||||
fetched_md5 = hashlib.md5(stripped_text_from_html.translate(None, b'\r\n\t ')).hexdigest()
|
fetched_md5 = hashlib.md5(stripped_text_from_html.translate(None, b'\r\n\t ')).hexdigest()
|
||||||
else:
|
else:
|
||||||
fetched_md5 = hashlib.md5(stripped_text_from_html).hexdigest()
|
fetched_md5 = hashlib.md5(stripped_text_from_html).hexdigest()
|
||||||
|
|
||||||
# On the first run of a site, watch['previous_md5'] will be None, set it the current one.
|
############ Blocking rules, after checksum #################
|
||||||
if not watch.get('previous_md5'):
|
blocked = False
|
||||||
watch['previous_md5'] = fetched_md5
|
|
||||||
update_obj["previous_md5"] = fetched_md5
|
|
||||||
|
|
||||||
blocked_by_not_found_trigger_text = False
|
trigger_text = watch.get('trigger_text', [])
|
||||||
|
if len(trigger_text):
|
||||||
if len(watch['trigger_text']):
|
# Assume blocked
|
||||||
# Yeah, lets block first until something matches
|
blocked = True
|
||||||
blocked_by_not_found_trigger_text = True
|
|
||||||
# Filter and trigger works the same, so reuse it
|
# Filter and trigger works the same, so reuse it
|
||||||
|
# It should return the line numbers that match
|
||||||
result = html_tools.strip_ignore_text(content=str(stripped_text_from_html),
|
result = html_tools.strip_ignore_text(content=str(stripped_text_from_html),
|
||||||
wordlist=watch['trigger_text'],
|
wordlist=trigger_text,
|
||||||
|
mode="line numbers")
|
||||||
|
# Unblock if the trigger was found
|
||||||
|
if result:
|
||||||
|
blocked = False
|
||||||
|
|
||||||
|
text_should_not_be_present = watch.get('text_should_not_be_present', [])
|
||||||
|
if len(text_should_not_be_present):
|
||||||
|
# If anything matched, then we should block a change from happening
|
||||||
|
result = html_tools.strip_ignore_text(content=str(stripped_text_from_html),
|
||||||
|
wordlist=text_should_not_be_present,
|
||||||
mode="line numbers")
|
mode="line numbers")
|
||||||
if result:
|
if result:
|
||||||
blocked_by_not_found_trigger_text = False
|
blocked = True
|
||||||
|
|
||||||
if not blocked_by_not_found_trigger_text and watch['previous_md5'] != fetched_md5:
|
# The main thing that all this at the moment comes down to :)
|
||||||
|
if watch.get('previous_md5') != fetched_md5:
|
||||||
changed_detected = True
|
changed_detected = True
|
||||||
update_obj["previous_md5"] = fetched_md5
|
|
||||||
update_obj["last_changed"] = timestamp
|
# Looks like something changed, but did it match all the rules?
|
||||||
|
if blocked:
|
||||||
|
changed_detected = False
|
||||||
|
|
||||||
# Extract title as title
|
# Extract title as title
|
||||||
if is_html:
|
if is_html:
|
||||||
if self.datastore.data['settings']['application']['extract_title_as_title'] or watch['extract_title_as_title']:
|
if self.datastore.data['settings']['application'].get('extract_title_as_title') or watch['extract_title_as_title']:
|
||||||
if not watch['title'] or not len(watch['title']):
|
if not watch['title'] or not len(watch['title']):
|
||||||
update_obj['title'] = html_tools.extract_element(find='title', html_content=fetcher.content)
|
update_obj['title'] = html_tools.extract_element(find='title', html_content=fetcher.content)
|
||||||
|
|
||||||
return changed_detected, update_obj, text_content_before_ignored_filter, fetcher.screenshot
|
if changed_detected:
|
||||||
|
if watch.get('check_unique_lines', False):
|
||||||
|
has_unique_lines = watch.lines_contain_something_unique_compared_to_history(lines=stripped_text_from_html.splitlines())
|
||||||
|
# One or more lines? unsure?
|
||||||
|
if not has_unique_lines:
|
||||||
|
logging.debug("check_unique_lines: UUID {} didnt have anything new setting change_detected=False".format(uuid))
|
||||||
|
changed_detected = False
|
||||||
|
else:
|
||||||
|
logging.debug("check_unique_lines: UUID {} had unique content".format(uuid))
|
||||||
|
|
||||||
|
# Always record the new checksum
|
||||||
|
update_obj["previous_md5"] = fetched_md5
|
||||||
|
|
||||||
|
# On the first run of a site, watch['previous_md5'] will be None, set it the current one.
|
||||||
|
if not watch.get('previous_md5'):
|
||||||
|
watch['previous_md5'] = fetched_md5
|
||||||
|
|
||||||
|
return changed_detected, update_obj, text_content_before_ignored_filter
|
||||||
|
|||||||
@@ -303,33 +303,58 @@ class ValidateCSSJSONXPATHInput(object):
|
|||||||
|
|
||||||
# Re #265 - maybe in the future fetch the page and offer a
|
# Re #265 - maybe in the future fetch the page and offer a
|
||||||
# warning/notice that its possible the rule doesnt yet match anything?
|
# warning/notice that its possible the rule doesnt yet match anything?
|
||||||
|
if not self.allow_json:
|
||||||
|
raise ValidationError("jq not permitted in this field!")
|
||||||
|
|
||||||
|
if 'jq:' in line:
|
||||||
|
try:
|
||||||
|
import jq
|
||||||
|
except ModuleNotFoundError:
|
||||||
|
# `jq` requires full compilation in windows and so isn't generally available
|
||||||
|
raise ValidationError("jq not support not found")
|
||||||
|
|
||||||
|
input = line.replace('jq:', '')
|
||||||
|
|
||||||
|
try:
|
||||||
|
jq.compile(input)
|
||||||
|
except (ValueError) as e:
|
||||||
|
message = field.gettext('\'%s\' is not a valid jq expression. (%s)')
|
||||||
|
raise ValidationError(message % (input, str(e)))
|
||||||
|
except:
|
||||||
|
raise ValidationError("A system-error occurred when validating your jq expression")
|
||||||
|
|
||||||
|
|
||||||
class quickWatchForm(Form):
|
class quickWatchForm(Form):
|
||||||
url = fields.URLField('URL', validators=[validateURL()])
|
url = fields.URLField('URL', validators=[validateURL()])
|
||||||
tag = StringField('Group tag', [validators.Optional(), validators.Length(max=35)])
|
tag = StringField('Group tag', [validators.Optional()])
|
||||||
|
watch_submit_button = SubmitField('Watch', render_kw={"class": "pure-button pure-button-primary"})
|
||||||
|
edit_and_watch_submit_button = SubmitField('Edit > Watch', render_kw={"class": "pure-button pure-button-primary"})
|
||||||
|
|
||||||
|
|
||||||
# Common to a single watch and the global settings
|
# Common to a single watch and the global settings
|
||||||
class commonSettingsForm(Form):
|
class commonSettingsForm(Form):
|
||||||
|
notification_urls = StringListField('Notification URL list', validators=[validators.Optional(), ValidateAppRiseServers()])
|
||||||
notification_urls = StringListField('Notification URL list', validators=[validators.Optional(), ValidateNotificationBodyAndTitleWhenURLisSet(), ValidateAppRiseServers()])
|
notification_title = StringField('Notification title', validators=[validators.Optional(), ValidateTokensList()])
|
||||||
notification_title = StringField('Notification title', default=default_notification_title, validators=[validators.Optional(), ValidateTokensList()])
|
notification_body = TextAreaField('Notification body', validators=[validators.Optional(), ValidateTokensList()])
|
||||||
notification_body = TextAreaField('Notification body', default=default_notification_body, validators=[validators.Optional(), ValidateTokensList()])
|
notification_format = SelectField('Notification format', choices=valid_notification_formats.keys())
|
||||||
notification_format = SelectField('Notification format', choices=valid_notification_formats.keys(), default=default_notification_format)
|
|
||||||
fetch_backend = RadioField(u'Fetch method', choices=content_fetcher.available_fetchers(), validators=[ValidateContentFetcherIsReady()])
|
fetch_backend = RadioField(u'Fetch method', choices=content_fetcher.available_fetchers(), validators=[ValidateContentFetcherIsReady()])
|
||||||
extract_title_as_title = BooleanField('Extract <title> from document and use as watch title', default=False)
|
extract_title_as_title = BooleanField('Extract <title> from document and use as watch title', default=False)
|
||||||
webdriver_delay = IntegerField('Wait seconds before extracting text', validators=[validators.Optional(), validators.NumberRange(min=1, message="Should contain one or more seconds")] )
|
webdriver_delay = IntegerField('Wait seconds before extracting text', validators=[validators.Optional(), validators.NumberRange(min=1,
|
||||||
|
message="Should contain one or more seconds")])
|
||||||
|
|
||||||
class watchForm(commonSettingsForm):
|
class watchForm(commonSettingsForm):
|
||||||
|
|
||||||
url = fields.URLField('URL', validators=[validateURL()])
|
url = fields.URLField('URL', validators=[validateURL()])
|
||||||
tag = StringField('Group tag', [validators.Optional(), validators.Length(max=35)], default='')
|
tag = StringField('Group tag', [validators.Optional()], default='')
|
||||||
|
|
||||||
time_between_check = FormField(TimeBetweenCheckForm)
|
time_between_check = FormField(TimeBetweenCheckForm)
|
||||||
|
|
||||||
css_filter = StringField('CSS/JSON/XPATH Filter', [ValidateCSSJSONXPATHInput()], default='')
|
include_filters = StringListField('CSS/JSONPath/JQ/XPath Filters', [ValidateCSSJSONXPATHInput()], default='')
|
||||||
|
|
||||||
subtractive_selectors = StringListField('Remove elements', [ValidateCSSJSONXPATHInput(allow_xpath=False, allow_json=False)])
|
subtractive_selectors = StringListField('Remove elements', [ValidateCSSJSONXPATHInput(allow_xpath=False, allow_json=False)])
|
||||||
|
|
||||||
|
extract_text = StringListField('Extract text', [ValidateListRegex()])
|
||||||
|
|
||||||
title = StringField('Title', default='')
|
title = StringField('Title', default='')
|
||||||
|
|
||||||
ignore_text = StringListField('Ignore text', [ValidateListRegex()])
|
ignore_text = StringListField('Ignore text', [ValidateListRegex()])
|
||||||
@@ -337,10 +362,19 @@ class watchForm(commonSettingsForm):
|
|||||||
body = TextAreaField('Request body', [validators.Optional()])
|
body = TextAreaField('Request body', [validators.Optional()])
|
||||||
method = SelectField('Request method', choices=valid_method, default=default_method)
|
method = SelectField('Request method', choices=valid_method, default=default_method)
|
||||||
ignore_status_codes = BooleanField('Ignore status codes (process non-2xx status codes as normal)', default=False)
|
ignore_status_codes = BooleanField('Ignore status codes (process non-2xx status codes as normal)', default=False)
|
||||||
|
check_unique_lines = BooleanField('Only trigger when new lines appear', default=False)
|
||||||
trigger_text = StringListField('Trigger/wait for text', [validators.Optional(), ValidateListRegex()])
|
trigger_text = StringListField('Trigger/wait for text', [validators.Optional(), ValidateListRegex()])
|
||||||
|
text_should_not_be_present = StringListField('Block change-detection if text matches', [validators.Optional(), ValidateListRegex()])
|
||||||
|
|
||||||
|
webdriver_js_execute_code = TextAreaField('Execute JavaScript before change detection', render_kw={"rows": "5"}, validators=[validators.Optional()])
|
||||||
|
|
||||||
save_button = SubmitField('Save', render_kw={"class": "pure-button pure-button-primary"})
|
save_button = SubmitField('Save', render_kw={"class": "pure-button pure-button-primary"})
|
||||||
save_and_preview_button = SubmitField('Save & Preview', render_kw={"class": "pure-button pure-button-primary"})
|
|
||||||
proxy = RadioField('Proxy')
|
proxy = RadioField('Proxy')
|
||||||
|
filter_failure_notification_send = BooleanField(
|
||||||
|
'Send a notification when the filter can no longer be found on the page', default=False)
|
||||||
|
|
||||||
|
notification_muted = BooleanField('Notifications Muted / Off', default=False)
|
||||||
|
|
||||||
def validate(self, **kwargs):
|
def validate(self, **kwargs):
|
||||||
if not super().validate():
|
if not super().validate():
|
||||||
@@ -360,7 +394,9 @@ class watchForm(commonSettingsForm):
|
|||||||
class globalSettingsRequestForm(Form):
|
class globalSettingsRequestForm(Form):
|
||||||
time_between_check = FormField(TimeBetweenCheckForm)
|
time_between_check = FormField(TimeBetweenCheckForm)
|
||||||
proxy = RadioField('Proxy')
|
proxy = RadioField('Proxy')
|
||||||
|
jitter_seconds = IntegerField('Random jitter seconds ± check',
|
||||||
|
render_kw={"style": "width: 5em;"},
|
||||||
|
validators=[validators.NumberRange(min=0, message="Should contain zero or more seconds")])
|
||||||
|
|
||||||
# datastore.data['settings']['application']..
|
# datastore.data['settings']['application']..
|
||||||
class globalSettingsApplicationForm(commonSettingsForm):
|
class globalSettingsApplicationForm(commonSettingsForm):
|
||||||
@@ -369,12 +405,18 @@ class globalSettingsApplicationForm(commonSettingsForm):
|
|||||||
global_subtractive_selectors = StringListField('Remove elements', [ValidateCSSJSONXPATHInput(allow_xpath=False, allow_json=False)])
|
global_subtractive_selectors = StringListField('Remove elements', [ValidateCSSJSONXPATHInput(allow_xpath=False, allow_json=False)])
|
||||||
global_ignore_text = StringListField('Ignore Text', [ValidateListRegex()])
|
global_ignore_text = StringListField('Ignore Text', [ValidateListRegex()])
|
||||||
ignore_whitespace = BooleanField('Ignore whitespace')
|
ignore_whitespace = BooleanField('Ignore whitespace')
|
||||||
real_browser_save_screenshot = BooleanField('Save last screenshot when using Chrome?')
|
|
||||||
removepassword_button = SubmitField('Remove password', render_kw={"class": "pure-button pure-button-primary"})
|
removepassword_button = SubmitField('Remove password', render_kw={"class": "pure-button pure-button-primary"})
|
||||||
|
empty_pages_are_a_change = BooleanField('Treat empty pages as a change?', default=False)
|
||||||
render_anchor_tag_content = BooleanField('Render anchor tag content', default=False)
|
render_anchor_tag_content = BooleanField('Render anchor tag content', default=False)
|
||||||
fetch_backend = RadioField('Fetch Method', default="html_requests", choices=content_fetcher.available_fetchers(), validators=[ValidateContentFetcherIsReady()])
|
fetch_backend = RadioField('Fetch Method', default="html_requests", choices=content_fetcher.available_fetchers(), validators=[ValidateContentFetcherIsReady()])
|
||||||
|
api_access_token_enabled = BooleanField('API access token security check enabled', default=True, validators=[validators.Optional()])
|
||||||
password = SaltyPasswordField()
|
password = SaltyPasswordField()
|
||||||
|
|
||||||
|
filter_failure_notification_threshold_attempts = IntegerField('Number of times the filter can be missing before sending a notification',
|
||||||
|
render_kw={"style": "width: 5em;"},
|
||||||
|
validators=[validators.NumberRange(min=0,
|
||||||
|
message="Should contain zero or more attempts")])
|
||||||
|
|
||||||
|
|
||||||
class globalSettingsForm(Form):
|
class globalSettingsForm(Form):
|
||||||
# Define these as FormFields/"sub forms", this way it matches the JSON storage
|
# Define these as FormFields/"sub forms", this way it matches the JSON storage
|
||||||
|
|||||||
@@ -1,26 +1,36 @@
|
|||||||
import json
|
|
||||||
import re
|
|
||||||
from typing import List
|
|
||||||
|
|
||||||
from bs4 import BeautifulSoup
|
from bs4 import BeautifulSoup
|
||||||
from jsonpath_ng.ext import parse
|
|
||||||
import re
|
|
||||||
from inscriptis import get_text
|
from inscriptis import get_text
|
||||||
from inscriptis.model.config import ParserConfig
|
from inscriptis.model.config import ParserConfig
|
||||||
|
from jsonpath_ng.ext import parse
|
||||||
|
from typing import List
|
||||||
|
import json
|
||||||
|
import re
|
||||||
|
|
||||||
|
# HTML added to be sure each result matching a filter (.example) gets converted to a new line by Inscriptis
|
||||||
|
TEXT_FILTER_LIST_LINE_SUFFIX = "<br/>"
|
||||||
|
|
||||||
class JSONNotFound(ValueError):
|
class JSONNotFound(ValueError):
|
||||||
def __init__(self, msg):
|
def __init__(self, msg):
|
||||||
ValueError.__init__(self, msg)
|
ValueError.__init__(self, msg)
|
||||||
|
|
||||||
# Given a CSS Rule, and a blob of HTML, return the blob of HTML that matches
|
# Given a CSS Rule, and a blob of HTML, return the blob of HTML that matches
|
||||||
def css_filter(css_filter, html_content):
|
def include_filters(include_filters, html_content, append_pretty_line_formatting=False):
|
||||||
soup = BeautifulSoup(html_content, "html.parser")
|
soup = BeautifulSoup(html_content, "html.parser")
|
||||||
html_block = ""
|
html_block = ""
|
||||||
for item in soup.select(css_filter, separator=""):
|
r = soup.select(include_filters, separator="")
|
||||||
html_block += str(item)
|
|
||||||
|
|
||||||
return html_block + "\n"
|
for element in r:
|
||||||
|
# When there's more than 1 match, then add the suffix to separate each line
|
||||||
|
# And where the matched result doesn't include something that will cause Inscriptis to add a newline
|
||||||
|
# (This way each 'match' reliably has a new-line in the diff)
|
||||||
|
# Divs are converted to 4 whitespaces by inscriptis
|
||||||
|
if append_pretty_line_formatting and len(html_block) and not element.name in (['br', 'hr', 'div', 'p']):
|
||||||
|
html_block += TEXT_FILTER_LIST_LINE_SUFFIX
|
||||||
|
|
||||||
|
html_block += str(element)
|
||||||
|
|
||||||
|
return html_block
|
||||||
|
|
||||||
def subtractive_css_selector(css_selector, html_content):
|
def subtractive_css_selector(css_selector, html_content):
|
||||||
soup = BeautifulSoup(html_content, "html.parser")
|
soup = BeautifulSoup(html_content, "html.parser")
|
||||||
@@ -36,14 +46,29 @@ def element_removal(selectors: List[str], html_content):
|
|||||||
|
|
||||||
|
|
||||||
# Return str Utf-8 of matched rules
|
# Return str Utf-8 of matched rules
|
||||||
def xpath_filter(xpath_filter, html_content):
|
def xpath_filter(xpath_filter, html_content, append_pretty_line_formatting=False):
|
||||||
from lxml import etree, html
|
from lxml import etree, html
|
||||||
|
|
||||||
tree = html.fromstring(html_content)
|
tree = html.fromstring(bytes(html_content, encoding='utf-8'))
|
||||||
html_block = ""
|
html_block = ""
|
||||||
|
|
||||||
for item in tree.xpath(xpath_filter.strip(), namespaces={'re':'http://exslt.org/regular-expressions'}):
|
r = tree.xpath(xpath_filter.strip(), namespaces={'re': 'http://exslt.org/regular-expressions'})
|
||||||
html_block+= etree.tostring(item, pretty_print=True).decode('utf-8')+"<br/>"
|
#@note: //title/text() wont work where <title>CDATA..
|
||||||
|
|
||||||
|
for element in r:
|
||||||
|
# When there's more than 1 match, then add the suffix to separate each line
|
||||||
|
# And where the matched result doesn't include something that will cause Inscriptis to add a newline
|
||||||
|
# (This way each 'match' reliably has a new-line in the diff)
|
||||||
|
# Divs are converted to 4 whitespaces by inscriptis
|
||||||
|
if append_pretty_line_formatting and len(html_block) and (not hasattr( element, 'tag' ) or not element.tag in (['br', 'hr', 'div', 'p'])):
|
||||||
|
html_block += TEXT_FILTER_LIST_LINE_SUFFIX
|
||||||
|
|
||||||
|
if type(element) == etree._ElementStringResult:
|
||||||
|
html_block += str(element)
|
||||||
|
elif type(element) == etree._ElementUnicodeResult:
|
||||||
|
html_block += str(element)
|
||||||
|
else:
|
||||||
|
html_block += etree.tostring(element, pretty_print=True).decode('utf-8')
|
||||||
|
|
||||||
return html_block
|
return html_block
|
||||||
|
|
||||||
@@ -62,19 +87,35 @@ def extract_element(find='title', html_content=''):
|
|||||||
return element_text
|
return element_text
|
||||||
|
|
||||||
#
|
#
|
||||||
def _parse_json(json_data, jsonpath_filter):
|
def _parse_json(json_data, json_filter):
|
||||||
s=[]
|
if 'json:' in json_filter:
|
||||||
jsonpath_expression = parse(jsonpath_filter.replace('json:', ''))
|
jsonpath_expression = parse(json_filter.replace('json:', ''))
|
||||||
match = jsonpath_expression.find(json_data)
|
match = jsonpath_expression.find(json_data)
|
||||||
|
return _get_stripped_text_from_json_match(match)
|
||||||
|
|
||||||
|
if 'jq:' in json_filter:
|
||||||
|
|
||||||
|
try:
|
||||||
|
import jq
|
||||||
|
except ModuleNotFoundError:
|
||||||
|
# `jq` requires full compilation in windows and so isn't generally available
|
||||||
|
raise Exception("jq not support not found")
|
||||||
|
|
||||||
|
jq_expression = jq.compile(json_filter.replace('jq:', ''))
|
||||||
|
match = jq_expression.input(json_data).all()
|
||||||
|
|
||||||
|
return _get_stripped_text_from_json_match(match)
|
||||||
|
|
||||||
|
def _get_stripped_text_from_json_match(match):
|
||||||
|
s = []
|
||||||
# More than one result, we will return it as a JSON list.
|
# More than one result, we will return it as a JSON list.
|
||||||
if len(match) > 1:
|
if len(match) > 1:
|
||||||
for i in match:
|
for i in match:
|
||||||
s.append(i.value)
|
s.append(i.value if hasattr(i, 'value') else i)
|
||||||
|
|
||||||
# Single value, use just the value, as it could be later used in a token in notifications.
|
# Single value, use just the value, as it could be later used in a token in notifications.
|
||||||
if len(match) == 1:
|
if len(match) == 1:
|
||||||
s = match[0].value
|
s = match[0].value if hasattr(match[0], 'value') else match[0]
|
||||||
|
|
||||||
# Re #257 - Better handling where it does not exist, in the case the original 's' value was False..
|
# Re #257 - Better handling where it does not exist, in the case the original 's' value was False..
|
||||||
if not match:
|
if not match:
|
||||||
@@ -86,16 +127,16 @@ def _parse_json(json_data, jsonpath_filter):
|
|||||||
|
|
||||||
return stripped_text_from_html
|
return stripped_text_from_html
|
||||||
|
|
||||||
def extract_json_as_string(content, jsonpath_filter):
|
def extract_json_as_string(content, json_filter):
|
||||||
|
|
||||||
stripped_text_from_html = False
|
stripped_text_from_html = False
|
||||||
|
|
||||||
# Try to parse/filter out the JSON, if we get some parser error, then maybe it's embedded <script type=ldjson>
|
# Try to parse/filter out the JSON, if we get some parser error, then maybe it's embedded <script type=ldjson>
|
||||||
try:
|
try:
|
||||||
stripped_text_from_html = _parse_json(json.loads(content), jsonpath_filter)
|
stripped_text_from_html = _parse_json(json.loads(content), json_filter)
|
||||||
except json.JSONDecodeError:
|
except json.JSONDecodeError:
|
||||||
|
|
||||||
# Foreach <script json></script> blob.. just return the first that matches jsonpath_filter
|
# Foreach <script json></script> blob.. just return the first that matches json_filter
|
||||||
s = []
|
s = []
|
||||||
soup = BeautifulSoup(content, 'html.parser')
|
soup = BeautifulSoup(content, 'html.parser')
|
||||||
bs_result = soup.findAll('script')
|
bs_result = soup.findAll('script')
|
||||||
@@ -114,7 +155,7 @@ def extract_json_as_string(content, jsonpath_filter):
|
|||||||
# Just skip it
|
# Just skip it
|
||||||
continue
|
continue
|
||||||
else:
|
else:
|
||||||
stripped_text_from_html = _parse_json(json_data, jsonpath_filter)
|
stripped_text_from_html = _parse_json(json_data, json_filter)
|
||||||
if stripped_text_from_html:
|
if stripped_text_from_html:
|
||||||
break
|
break
|
||||||
|
|
||||||
@@ -202,3 +243,17 @@ def html_to_text(html_content: str, render_anchor_tag_content=False) -> str:
|
|||||||
|
|
||||||
return text_content
|
return text_content
|
||||||
|
|
||||||
|
def workarounds_for_obfuscations(content):
|
||||||
|
"""
|
||||||
|
Some sites are using sneaky tactics to make prices and other information un-renderable by Inscriptis
|
||||||
|
This could go into its own Pip package in the future, for faster updates
|
||||||
|
"""
|
||||||
|
|
||||||
|
# HomeDepot.com style <span>$<!-- -->90<!-- -->.<!-- -->74</span>
|
||||||
|
# https://github.com/weblyzard/inscriptis/issues/45
|
||||||
|
if not content:
|
||||||
|
return content
|
||||||
|
|
||||||
|
content = re.sub('<!--\s+-->', '', content)
|
||||||
|
|
||||||
|
return content
|
||||||
|
|||||||
@@ -92,7 +92,7 @@ class import_distill_io_json(Importer):
|
|||||||
|
|
||||||
for d in data.get('data'):
|
for d in data.get('data'):
|
||||||
d_config = json.loads(d['config'])
|
d_config = json.loads(d['config'])
|
||||||
extras = {'title': d['name']}
|
extras = {'title': d.get('name', None)}
|
||||||
|
|
||||||
if len(d['uri']) and good < 5000:
|
if len(d['uri']) and good < 5000:
|
||||||
try:
|
try:
|
||||||
@@ -103,23 +103,20 @@ class import_distill_io_json(Importer):
|
|||||||
pass
|
pass
|
||||||
except IndexError:
|
except IndexError:
|
||||||
pass
|
pass
|
||||||
|
extras['include_filters'] = []
|
||||||
try:
|
try:
|
||||||
extras['css_filter'] = d_config['selections'][0]['frames'][0]['includes'][0]['expr']
|
|
||||||
if d_config['selections'][0]['frames'][0]['includes'][0]['type'] == 'xpath':
|
if d_config['selections'][0]['frames'][0]['includes'][0]['type'] == 'xpath':
|
||||||
extras['css_filter'] = 'xpath:' + extras['css_filter']
|
extras['include_filters'].append('xpath:' + d_config['selections'][0]['frames'][0]['includes'][0]['expr'])
|
||||||
|
else:
|
||||||
|
extras['include_filters'].append(d_config['selections'][0]['frames'][0]['includes'][0]['expr'])
|
||||||
except KeyError:
|
except KeyError:
|
||||||
pass
|
pass
|
||||||
except IndexError:
|
except IndexError:
|
||||||
pass
|
pass
|
||||||
|
|
||||||
try:
|
|
||||||
|
if d.get('tags', False):
|
||||||
extras['tag'] = ", ".join(d['tags'])
|
extras['tag'] = ", ".join(d['tags'])
|
||||||
except KeyError:
|
|
||||||
pass
|
|
||||||
except IndexError:
|
|
||||||
pass
|
|
||||||
|
|
||||||
new_uuid = datastore.add_watch(url=d['uri'].strip(),
|
new_uuid = datastore.add_watch(url=d['uri'].strip(),
|
||||||
extras=extras,
|
extras=extras,
|
||||||
|
|||||||
@@ -1,46 +1,43 @@
|
|||||||
import collections
|
from os import getenv
|
||||||
import os
|
|
||||||
|
|
||||||
import uuid as uuid_builder
|
|
||||||
|
|
||||||
from changedetectionio.notification import (
|
from changedetectionio.notification import (
|
||||||
default_notification_body,
|
default_notification_body,
|
||||||
default_notification_format,
|
default_notification_format,
|
||||||
default_notification_title,
|
default_notification_title,
|
||||||
)
|
)
|
||||||
|
|
||||||
|
_FILTER_FAILURE_THRESHOLD_ATTEMPTS_DEFAULT = 6
|
||||||
|
|
||||||
class model(dict):
|
class model(dict):
|
||||||
base_config = {
|
base_config = {
|
||||||
'note': "Hello! If you change this file manually, please be sure to restart your changedetection.io instance!",
|
'note': "Hello! If you change this file manually, please be sure to restart your changedetection.io instance!",
|
||||||
'watching': {},
|
'watching': {},
|
||||||
'settings': {
|
'settings': {
|
||||||
'headers': {
|
'headers': {
|
||||||
'User-Agent': 'Mozilla/5.0 (X11; Linux x86_64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/87.0.4280.66 Safari/537.36',
|
|
||||||
'Accept': 'text/html,application/xhtml+xml,application/xml;q=0.9,image/avif,image/webp,image/apng,*/*;q=0.8,application/signed-exchange;v=b3;q=0.9',
|
|
||||||
'Accept-Encoding': 'gzip, deflate', # No support for brolti in python requests yet.
|
|
||||||
'Accept-Language': 'en-GB,en-US;q=0.9,en;'
|
|
||||||
},
|
},
|
||||||
'requests': {
|
'requests': {
|
||||||
'timeout': 15, # Default 15 seconds
|
'timeout': int(getenv("DEFAULT_SETTINGS_REQUESTS_TIMEOUT", "45")), # Default 45 seconds
|
||||||
'time_between_check': {'weeks': None, 'days': None, 'hours': 3, 'minutes': None, 'seconds': None},
|
'time_between_check': {'weeks': None, 'days': None, 'hours': 3, 'minutes': None, 'seconds': None},
|
||||||
'workers': 10, # Number of threads, lower is better for slow connections
|
'jitter_seconds': 0,
|
||||||
|
'workers': int(getenv("DEFAULT_SETTINGS_REQUESTS_WORKERS", "10")), # Number of threads, lower is better for slow connections
|
||||||
'proxy': None # Preferred proxy connection
|
'proxy': None # Preferred proxy connection
|
||||||
},
|
},
|
||||||
'application': {
|
'application': {
|
||||||
|
'api_access_token_enabled': True,
|
||||||
'password': False,
|
'password': False,
|
||||||
'base_url' : None,
|
'base_url' : None,
|
||||||
'extract_title_as_title': False,
|
'extract_title_as_title': False,
|
||||||
'fetch_backend': os.getenv("DEFAULT_FETCH_BACKEND", "html_requests"),
|
'empty_pages_are_a_change': False,
|
||||||
|
'fetch_backend': getenv("DEFAULT_FETCH_BACKEND", "html_requests"),
|
||||||
|
'filter_failure_notification_threshold_attempts': _FILTER_FAILURE_THRESHOLD_ATTEMPTS_DEFAULT,
|
||||||
'global_ignore_text': [], # List of text to ignore when calculating the comparison checksum
|
'global_ignore_text': [], # List of text to ignore when calculating the comparison checksum
|
||||||
'global_subtractive_selectors': [],
|
'global_subtractive_selectors': [],
|
||||||
'ignore_whitespace': False,
|
'ignore_whitespace': True,
|
||||||
'render_anchor_tag_content': False,
|
'render_anchor_tag_content': False,
|
||||||
'notification_urls': [], # Apprise URL list
|
'notification_urls': [], # Apprise URL list
|
||||||
# Custom notification content
|
# Custom notification content
|
||||||
'notification_title': default_notification_title,
|
'notification_title': default_notification_title,
|
||||||
'notification_body': default_notification_body,
|
'notification_body': default_notification_body,
|
||||||
'notification_format': default_notification_format,
|
'notification_format': default_notification_format,
|
||||||
'real_browser_save_screenshot': True,
|
|
||||||
'schema_version' : 0,
|
'schema_version' : 0,
|
||||||
'webdriver_delay': None # Extra delay in seconds before extracting text
|
'webdriver_delay': None # Extra delay in seconds before extracting text
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -1,57 +1,217 @@
|
|||||||
|
from distutils.util import strtobool
|
||||||
|
import logging
|
||||||
import os
|
import os
|
||||||
|
import time
|
||||||
import uuid as uuid_builder
|
import uuid
|
||||||
|
|
||||||
minimum_seconds_recheck_time = int(os.getenv('MINIMUM_SECONDS_RECHECK_TIME', 60))
|
minimum_seconds_recheck_time = int(os.getenv('MINIMUM_SECONDS_RECHECK_TIME', 60))
|
||||||
|
mtable = {'seconds': 1, 'minutes': 60, 'hours': 3600, 'days': 86400, 'weeks': 86400 * 7}
|
||||||
|
|
||||||
from changedetectionio.notification import (
|
from changedetectionio.notification import (
|
||||||
default_notification_body,
|
default_notification_format_for_watch
|
||||||
default_notification_format,
|
|
||||||
default_notification_title,
|
|
||||||
)
|
)
|
||||||
|
|
||||||
|
|
||||||
class model(dict):
|
class model(dict):
|
||||||
base_config = {
|
__newest_history_key = None
|
||||||
'url': None,
|
__history_n=0
|
||||||
'tag': None,
|
__base_config = {
|
||||||
'last_checked': 0,
|
#'history': {}, # Dict of timestamp and output stripped filename (removed)
|
||||||
'last_changed': 0,
|
#'newest_history_key': 0, (removed, taken from history.txt index)
|
||||||
'paused': False,
|
|
||||||
'last_viewed': 0, # history key value of the last viewed via the [diff] link
|
|
||||||
'newest_history_key': 0,
|
|
||||||
'title': None,
|
|
||||||
'previous_md5': False,
|
|
||||||
# UUID not needed, should be generated only as a key
|
|
||||||
# 'uuid':
|
|
||||||
'headers': {}, # Extra headers to send
|
|
||||||
'body': None,
|
'body': None,
|
||||||
'method': 'GET',
|
'check_unique_lines': False, # On change-detected, compare against all history if its something new
|
||||||
'history': {}, # Dict of timestamp and output stripped filename
|
'check_count': 0,
|
||||||
'ignore_text': [], # List of text to ignore when calculating the comparison checksum
|
'consecutive_filter_failures': 0, # Every time the CSS/xPath filter cannot be located, reset when all is fine.
|
||||||
# Custom notification content
|
'extract_text': [], # Extract text by regex after filters
|
||||||
'notification_urls': [], # List of URLs to add to the notification Queue (Usually AppRise)
|
|
||||||
'notification_title': default_notification_title,
|
|
||||||
'notification_body': default_notification_body,
|
|
||||||
'notification_format': default_notification_format,
|
|
||||||
'css_filter': "",
|
|
||||||
'subtractive_selectors': [],
|
|
||||||
'trigger_text': [], # List of text or regex to wait for until a change is detected
|
|
||||||
'fetch_backend': None,
|
|
||||||
'extract_title_as_title': False,
|
'extract_title_as_title': False,
|
||||||
|
'fetch_backend': None,
|
||||||
|
'filter_failure_notification_send': strtobool(os.getenv('FILTER_FAILURE_NOTIFICATION_SEND_DEFAULT', 'True')),
|
||||||
|
'headers': {}, # Extra headers to send
|
||||||
|
'ignore_text': [], # List of text to ignore when calculating the comparison checksum
|
||||||
|
'include_filters': [],
|
||||||
|
'last_checked': 0,
|
||||||
|
'last_error': False,
|
||||||
|
'last_viewed': 0, # history key value of the last viewed via the [diff] link
|
||||||
|
'method': 'GET',
|
||||||
|
# Custom notification content
|
||||||
|
'notification_body': None,
|
||||||
|
'notification_format': default_notification_format_for_watch,
|
||||||
|
'notification_muted': False,
|
||||||
|
'notification_title': None,
|
||||||
|
'notification_urls': [], # List of URLs to add to the notification Queue (Usually AppRise)
|
||||||
|
'paused': False,
|
||||||
|
'previous_md5': False,
|
||||||
'proxy': None, # Preferred proxy connection
|
'proxy': None, # Preferred proxy connection
|
||||||
|
'subtractive_selectors': [],
|
||||||
|
'tag': None,
|
||||||
|
'text_should_not_be_present': [], # Text that should not present
|
||||||
# Re #110, so then if this is set to None, we know to use the default value instead
|
# Re #110, so then if this is set to None, we know to use the default value instead
|
||||||
# Requires setting to None on submit if it's the same as the default
|
# Requires setting to None on submit if it's the same as the default
|
||||||
# Should be all None by default, so we use the system default in this case.
|
# Should be all None by default, so we use the system default in this case.
|
||||||
'time_between_check': {'weeks': None, 'days': None, 'hours': None, 'minutes': None, 'seconds': None},
|
'time_between_check': {'weeks': None, 'days': None, 'hours': None, 'minutes': None, 'seconds': None},
|
||||||
'webdriver_delay': None
|
'title': None,
|
||||||
|
'trigger_text': [], # List of text or regex to wait for until a change is detected
|
||||||
|
'url': None,
|
||||||
|
'uuid': str(uuid.uuid4()),
|
||||||
|
'webdriver_delay': None,
|
||||||
|
'webdriver_js_execute_code': None, # Run before change-detection
|
||||||
}
|
}
|
||||||
|
jitter_seconds = 0
|
||||||
|
|
||||||
def __init__(self, *arg, **kw):
|
def __init__(self, *arg, **kw):
|
||||||
self.update(self.base_config)
|
|
||||||
# goes at the end so we update the default object with the initialiser
|
self.update(self.__base_config)
|
||||||
|
self.__datastore_path = kw['datastore_path']
|
||||||
|
|
||||||
|
self['uuid'] = str(uuid.uuid4())
|
||||||
|
|
||||||
|
del kw['datastore_path']
|
||||||
|
|
||||||
|
if kw.get('default'):
|
||||||
|
self.update(kw['default'])
|
||||||
|
del kw['default']
|
||||||
|
|
||||||
|
# Be sure the cached timestamp is ready
|
||||||
|
bump = self.history
|
||||||
|
|
||||||
|
# Goes at the end so we update the default object with the initialiser
|
||||||
super(model, self).__init__(*arg, **kw)
|
super(model, self).__init__(*arg, **kw)
|
||||||
|
|
||||||
|
@property
|
||||||
|
def viewed(self):
|
||||||
|
if int(self['last_viewed']) >= int(self.newest_history_key) :
|
||||||
|
return True
|
||||||
|
|
||||||
|
return False
|
||||||
|
|
||||||
|
def ensure_data_dir_exists(self):
|
||||||
|
if not os.path.isdir(self.watch_data_dir):
|
||||||
|
print ("> Creating data dir {}".format(self.watch_data_dir))
|
||||||
|
os.mkdir(self.watch_data_dir)
|
||||||
|
|
||||||
|
@property
|
||||||
|
def link(self):
|
||||||
|
url = self.get('url', '')
|
||||||
|
if '{%' in url or '{{' in url:
|
||||||
|
from jinja2 import Environment
|
||||||
|
# Jinja2 available in URLs along with https://pypi.org/project/jinja2-time/
|
||||||
|
jinja2_env = Environment(extensions=['jinja2_time.TimeExtension'])
|
||||||
|
return str(jinja2_env.from_string(url).render())
|
||||||
|
return url
|
||||||
|
|
||||||
|
@property
|
||||||
|
def label(self):
|
||||||
|
# Used for sorting
|
||||||
|
if self['title']:
|
||||||
|
return self['title']
|
||||||
|
return self['url']
|
||||||
|
|
||||||
|
@property
|
||||||
|
def last_changed(self):
|
||||||
|
# last_changed will be the newest snapshot, but when we have just one snapshot, it should be 0
|
||||||
|
if self.__history_n <= 1:
|
||||||
|
return 0
|
||||||
|
if self.__newest_history_key:
|
||||||
|
return int(self.__newest_history_key)
|
||||||
|
return 0
|
||||||
|
|
||||||
|
@property
|
||||||
|
def history_n(self):
|
||||||
|
return self.__history_n
|
||||||
|
|
||||||
|
@property
|
||||||
|
def history(self):
|
||||||
|
"""History index is just a text file as a list
|
||||||
|
{watch-uuid}/history.txt
|
||||||
|
|
||||||
|
contains a list like
|
||||||
|
|
||||||
|
{epoch-time},{filename}\n
|
||||||
|
|
||||||
|
We read in this list as the history information
|
||||||
|
|
||||||
|
"""
|
||||||
|
tmp_history = {}
|
||||||
|
|
||||||
|
# Read the history file as a dict
|
||||||
|
fname = os.path.join(self.watch_data_dir, "history.txt")
|
||||||
|
if os.path.isfile(fname):
|
||||||
|
logging.debug("Reading history index " + str(time.time()))
|
||||||
|
with open(fname, "r") as f:
|
||||||
|
for i in f.readlines():
|
||||||
|
if ',' in i:
|
||||||
|
k, v = i.strip().split(',', 2)
|
||||||
|
|
||||||
|
# The index history could contain a relative path, so we need to make the fullpath
|
||||||
|
# so that python can read it
|
||||||
|
if not '/' in v and not '\'' in v:
|
||||||
|
v = os.path.join(self.watch_data_dir, v)
|
||||||
|
else:
|
||||||
|
# It's possible that they moved the datadir on older versions
|
||||||
|
# So the snapshot exists but is in a different path
|
||||||
|
snapshot_fname = v.split('/')[-1]
|
||||||
|
proposed_new_path = os.path.join(self.watch_data_dir, snapshot_fname)
|
||||||
|
if not os.path.exists(v) and os.path.exists(proposed_new_path):
|
||||||
|
v = proposed_new_path
|
||||||
|
|
||||||
|
tmp_history[k] = v
|
||||||
|
|
||||||
|
if len(tmp_history):
|
||||||
|
self.__newest_history_key = list(tmp_history.keys())[-1]
|
||||||
|
|
||||||
|
self.__history_n = len(tmp_history)
|
||||||
|
|
||||||
|
return tmp_history
|
||||||
|
|
||||||
|
@property
|
||||||
|
def has_history(self):
|
||||||
|
fname = os.path.join(self.watch_data_dir, "history.txt")
|
||||||
|
return os.path.isfile(fname)
|
||||||
|
|
||||||
|
# Returns the newest key, but if theres only 1 record, then it's counted as not being new, so return 0.
|
||||||
|
@property
|
||||||
|
def newest_history_key(self):
|
||||||
|
if self.__newest_history_key is not None:
|
||||||
|
return self.__newest_history_key
|
||||||
|
|
||||||
|
if len(self.history) <= 1:
|
||||||
|
return 0
|
||||||
|
|
||||||
|
|
||||||
|
bump = self.history
|
||||||
|
return self.__newest_history_key
|
||||||
|
|
||||||
|
# Save some text file to the appropriate path and bump the history
|
||||||
|
# result_obj from fetch_site_status.run()
|
||||||
|
def save_history_text(self, contents, timestamp):
|
||||||
|
|
||||||
|
self.ensure_data_dir_exists()
|
||||||
|
|
||||||
|
# Small hack so that we sleep just enough to allow 1 second between history snapshots
|
||||||
|
# this is because history.txt indexes/keys snapshots by epoch seconds and we dont want dupe keys
|
||||||
|
if self.__newest_history_key and int(timestamp) == int(self.__newest_history_key):
|
||||||
|
time.sleep(timestamp - self.__newest_history_key)
|
||||||
|
|
||||||
|
snapshot_fname = "{}.txt".format(str(uuid.uuid4()))
|
||||||
|
|
||||||
|
# in /diff/ and /preview/ we are going to assume for now that it's UTF-8 when reading
|
||||||
|
# most sites are utf-8 and some are even broken utf-8
|
||||||
|
with open(os.path.join(self.watch_data_dir, snapshot_fname), 'wb') as f:
|
||||||
|
f.write(contents)
|
||||||
|
f.close()
|
||||||
|
|
||||||
|
# Append to index
|
||||||
|
# @todo check last char was \n
|
||||||
|
index_fname = os.path.join(self.watch_data_dir, "history.txt")
|
||||||
|
with open(index_fname, 'a') as f:
|
||||||
|
f.write("{},{}\n".format(timestamp, snapshot_fname))
|
||||||
|
f.close()
|
||||||
|
|
||||||
|
self.__newest_history_key = timestamp
|
||||||
|
self.__history_n += 1
|
||||||
|
|
||||||
|
# @todo bump static cache of the last timestamp so we dont need to examine the file to set a proper ''viewed'' status
|
||||||
|
return snapshot_fname
|
||||||
|
|
||||||
@property
|
@property
|
||||||
def has_empty_checktime(self):
|
def has_empty_checktime(self):
|
||||||
@@ -62,9 +222,75 @@ class model(dict):
|
|||||||
|
|
||||||
def threshold_seconds(self):
|
def threshold_seconds(self):
|
||||||
seconds = 0
|
seconds = 0
|
||||||
mtable = {'seconds': 1, 'minutes': 60, 'hours': 3600, 'days': 86400, 'weeks': 86400 * 7}
|
|
||||||
for m, n in mtable.items():
|
for m, n in mtable.items():
|
||||||
x = self.get('time_between_check', {}).get(m, None)
|
x = self.get('time_between_check', {}).get(m, None)
|
||||||
if x:
|
if x:
|
||||||
seconds += x * n
|
seconds += x * n
|
||||||
return seconds
|
return seconds
|
||||||
|
|
||||||
|
# Iterate over all history texts and see if something new exists
|
||||||
|
def lines_contain_something_unique_compared_to_history(self, lines: list):
|
||||||
|
local_lines = set([l.decode('utf-8').strip().lower() for l in lines])
|
||||||
|
|
||||||
|
# Compare each lines (set) against each history text file (set) looking for something new..
|
||||||
|
existing_history = set({})
|
||||||
|
for k, v in self.history.items():
|
||||||
|
alist = set([line.decode('utf-8').strip().lower() for line in open(v, 'rb')])
|
||||||
|
existing_history = existing_history.union(alist)
|
||||||
|
|
||||||
|
# Check that everything in local_lines(new stuff) already exists in existing_history - it should
|
||||||
|
# if not, something new happened
|
||||||
|
return not local_lines.issubset(existing_history)
|
||||||
|
|
||||||
|
def get_screenshot(self):
|
||||||
|
fname = os.path.join(self.watch_data_dir, "last-screenshot.png")
|
||||||
|
if os.path.isfile(fname):
|
||||||
|
return fname
|
||||||
|
|
||||||
|
return False
|
||||||
|
|
||||||
|
def __get_file_ctime(self, filename):
|
||||||
|
fname = os.path.join(self.watch_data_dir, filename)
|
||||||
|
if os.path.isfile(fname):
|
||||||
|
return int(os.path.getmtime(fname))
|
||||||
|
return False
|
||||||
|
|
||||||
|
@property
|
||||||
|
def error_text_ctime(self):
|
||||||
|
return self.__get_file_ctime('last-error.txt')
|
||||||
|
|
||||||
|
@property
|
||||||
|
def snapshot_text_ctime(self):
|
||||||
|
if self.history_n==0:
|
||||||
|
return False
|
||||||
|
|
||||||
|
timestamp = list(self.history.keys())[-1]
|
||||||
|
return int(timestamp)
|
||||||
|
|
||||||
|
@property
|
||||||
|
def snapshot_screenshot_ctime(self):
|
||||||
|
return self.__get_file_ctime('last-screenshot.png')
|
||||||
|
|
||||||
|
@property
|
||||||
|
def snapshot_error_screenshot_ctime(self):
|
||||||
|
return self.__get_file_ctime('last-error-screenshot.png')
|
||||||
|
|
||||||
|
@property
|
||||||
|
def watch_data_dir(self):
|
||||||
|
# The base dir of the watch data
|
||||||
|
return os.path.join(self.__datastore_path, self['uuid'])
|
||||||
|
|
||||||
|
def get_error_text(self):
|
||||||
|
"""Return the text saved from a previous request that resulted in a non-200 error"""
|
||||||
|
fname = os.path.join(self.watch_data_dir, "last-error.txt")
|
||||||
|
if os.path.isfile(fname):
|
||||||
|
with open(fname, 'r') as f:
|
||||||
|
return f.read()
|
||||||
|
return False
|
||||||
|
|
||||||
|
def get_error_snapshot(self):
|
||||||
|
"""Return path to the screenshot that resulted in a non-200 error"""
|
||||||
|
fname = os.path.join(self.watch_data_dir, "last-error-screenshot.png")
|
||||||
|
if os.path.isfile(fname):
|
||||||
|
return fname
|
||||||
|
return False
|
||||||
|
|||||||
@@ -14,16 +14,19 @@ valid_tokens = {
|
|||||||
'current_snapshot': ''
|
'current_snapshot': ''
|
||||||
}
|
}
|
||||||
|
|
||||||
|
default_notification_format_for_watch = 'System default'
|
||||||
|
default_notification_format = 'Text'
|
||||||
|
default_notification_body = '{watch_url} had a change.\n---\n{diff}\n---\n'
|
||||||
|
default_notification_title = 'ChangeDetection.io Notification - {watch_url}'
|
||||||
|
|
||||||
valid_notification_formats = {
|
valid_notification_formats = {
|
||||||
'Text': NotifyFormat.TEXT,
|
'Text': NotifyFormat.TEXT,
|
||||||
'Markdown': NotifyFormat.MARKDOWN,
|
'Markdown': NotifyFormat.MARKDOWN,
|
||||||
'HTML': NotifyFormat.HTML,
|
'HTML': NotifyFormat.HTML,
|
||||||
|
# Used only for editing a watch (not for global)
|
||||||
|
default_notification_format_for_watch: default_notification_format_for_watch
|
||||||
}
|
}
|
||||||
|
|
||||||
default_notification_format = 'Text'
|
|
||||||
default_notification_body = '{watch_url} had a change.\n---\n{diff}\n---\n'
|
|
||||||
default_notification_title = 'ChangeDetection.io Notification - {watch_url}'
|
|
||||||
|
|
||||||
def process_notification(n_object, datastore):
|
def process_notification(n_object, datastore):
|
||||||
|
|
||||||
# Get the notification body from datastore
|
# Get the notification body from datastore
|
||||||
@@ -34,7 +37,6 @@ def process_notification(n_object, datastore):
|
|||||||
valid_notification_formats[default_notification_format],
|
valid_notification_formats[default_notification_format],
|
||||||
)
|
)
|
||||||
|
|
||||||
|
|
||||||
# Insert variables into the notification content
|
# Insert variables into the notification content
|
||||||
notification_parameters = create_notification_parameters(n_object, datastore)
|
notification_parameters = create_notification_parameters(n_object, datastore)
|
||||||
|
|
||||||
@@ -48,9 +50,10 @@ def process_notification(n_object, datastore):
|
|||||||
# Anything higher than or equal to WARNING (which covers things like Connection errors)
|
# Anything higher than or equal to WARNING (which covers things like Connection errors)
|
||||||
# raise it as an exception
|
# raise it as an exception
|
||||||
apobjs=[]
|
apobjs=[]
|
||||||
|
sent_objs=[]
|
||||||
|
from .apprise_asset import asset
|
||||||
for url in n_object['notification_urls']:
|
for url in n_object['notification_urls']:
|
||||||
|
apobj = apprise.Apprise(debug=True, asset=asset)
|
||||||
apobj = apprise.Apprise(debug=True)
|
|
||||||
url = url.strip()
|
url = url.strip()
|
||||||
if len(url):
|
if len(url):
|
||||||
print(">> Process Notification: AppRise notifying {}".format(url))
|
print(">> Process Notification: AppRise notifying {}".format(url))
|
||||||
@@ -63,23 +66,36 @@ def process_notification(n_object, datastore):
|
|||||||
|
|
||||||
# So if no avatar_url is specified, add one so it can be correctly calculated into the total payload
|
# So if no avatar_url is specified, add one so it can be correctly calculated into the total payload
|
||||||
k = '?' if not '?' in url else '&'
|
k = '?' if not '?' in url else '&'
|
||||||
if not 'avatar_url' in url:
|
if not 'avatar_url' in url and not url.startswith('mail'):
|
||||||
url += k + 'avatar_url=https://raw.githubusercontent.com/dgtlmoon/changedetection.io/master/changedetectionio/static/images/avatar-256x256.png'
|
url += k + 'avatar_url=https://raw.githubusercontent.com/dgtlmoon/changedetection.io/master/changedetectionio/static/images/avatar-256x256.png'
|
||||||
|
|
||||||
if url.startswith('tgram://'):
|
if url.startswith('tgram://'):
|
||||||
|
# Telegram only supports a limit subset of HTML, remove the '<br/>' we place in.
|
||||||
|
# re https://github.com/dgtlmoon/changedetection.io/issues/555
|
||||||
|
# @todo re-use an existing library we have already imported to strip all non-allowed tags
|
||||||
|
n_body = n_body.replace('<br/>', '\n')
|
||||||
|
n_body = n_body.replace('</br>', '\n')
|
||||||
# real limit is 4096, but minus some for extra metadata
|
# real limit is 4096, but minus some for extra metadata
|
||||||
payload_max_size = 3600
|
payload_max_size = 3600
|
||||||
body_limit = max(0, payload_max_size - len(n_title))
|
body_limit = max(0, payload_max_size - len(n_title))
|
||||||
n_title = n_title[0:payload_max_size]
|
n_title = n_title[0:payload_max_size]
|
||||||
n_body = n_body[0:body_limit]
|
n_body = n_body[0:body_limit]
|
||||||
|
|
||||||
elif url.startswith('discord://'):
|
elif url.startswith('discord://') or url.startswith('https://discordapp.com/api/webhooks') or url.startswith('https://discord.com/api'):
|
||||||
# real limit is 2000, but minus some for extra metadata
|
# real limit is 2000, but minus some for extra metadata
|
||||||
payload_max_size = 1700
|
payload_max_size = 1700
|
||||||
body_limit = max(0, payload_max_size - len(n_title))
|
body_limit = max(0, payload_max_size - len(n_title))
|
||||||
n_title = n_title[0:payload_max_size]
|
n_title = n_title[0:payload_max_size]
|
||||||
n_body = n_body[0:body_limit]
|
n_body = n_body[0:body_limit]
|
||||||
|
|
||||||
|
elif url.startswith('mailto'):
|
||||||
|
# Apprise will default to HTML, so we need to override it
|
||||||
|
# So that whats' generated in n_body is in line with what is going to be sent.
|
||||||
|
# https://github.com/caronc/apprise/issues/633#issuecomment-1191449321
|
||||||
|
if not 'format=' in url and (n_format == 'text' or n_format == 'markdown'):
|
||||||
|
prefix = '?' if not '?' in url else '&'
|
||||||
|
url = "{}{}format={}".format(url, prefix, n_format)
|
||||||
|
|
||||||
apobj.add(url)
|
apobj.add(url)
|
||||||
|
|
||||||
apobj.notify(
|
apobj.notify(
|
||||||
@@ -97,6 +113,15 @@ def process_notification(n_object, datastore):
|
|||||||
if log_value and 'WARNING' in log_value or 'ERROR' in log_value:
|
if log_value and 'WARNING' in log_value or 'ERROR' in log_value:
|
||||||
raise Exception(log_value)
|
raise Exception(log_value)
|
||||||
|
|
||||||
|
sent_objs.append({'title': n_title,
|
||||||
|
'body': n_body,
|
||||||
|
'url' : url,
|
||||||
|
'body_format': n_format})
|
||||||
|
|
||||||
|
# Return what was sent for better logging - after the for loop
|
||||||
|
return sent_objs
|
||||||
|
|
||||||
|
|
||||||
# Notification title + body content parameters get created here.
|
# Notification title + body content parameters get created here.
|
||||||
def create_notification_parameters(n_object, datastore):
|
def create_notification_parameters(n_object, datastore):
|
||||||
from copy import deepcopy
|
from copy import deepcopy
|
||||||
|
|||||||
@@ -9,6 +9,8 @@
|
|||||||
# exit when any command fails
|
# exit when any command fails
|
||||||
set -e
|
set -e
|
||||||
|
|
||||||
|
SCRIPT_DIR=$( cd -- "$( dirname -- "${BASH_SOURCE[0]}" )" &> /dev/null && pwd )
|
||||||
|
|
||||||
find tests/test_*py -type f|while read test_name
|
find tests/test_*py -type f|while read test_name
|
||||||
do
|
do
|
||||||
echo "TEST RUNNING $test_name"
|
echo "TEST RUNNING $test_name"
|
||||||
@@ -22,3 +24,75 @@ echo "RUNNING WITH BASE_URL SET"
|
|||||||
export BASE_URL="https://really-unique-domain.io"
|
export BASE_URL="https://really-unique-domain.io"
|
||||||
pytest tests/test_notification.py
|
pytest tests/test_notification.py
|
||||||
|
|
||||||
|
# Now for the selenium and playwright/browserless fetchers
|
||||||
|
# Note - this is not UI functional tests - just checking that each one can fetch the content
|
||||||
|
|
||||||
|
echo "TESTING WEBDRIVER FETCH > SELENIUM/WEBDRIVER..."
|
||||||
|
docker run -d --name $$-test_selenium -p 4444:4444 --rm --shm-size="2g" selenium/standalone-chrome-debug:3.141.59
|
||||||
|
# takes a while to spin up
|
||||||
|
sleep 5
|
||||||
|
export WEBDRIVER_URL=http://localhost:4444/wd/hub
|
||||||
|
pytest tests/fetchers/test_content.py
|
||||||
|
pytest tests/test_errorhandling.py
|
||||||
|
unset WEBDRIVER_URL
|
||||||
|
docker kill $$-test_selenium
|
||||||
|
|
||||||
|
echo "TESTING WEBDRIVER FETCH > PLAYWRIGHT/BROWSERLESS..."
|
||||||
|
# Not all platforms support playwright (not ARM/rPI), so it's not packaged in requirements.txt
|
||||||
|
PLAYWRIGHT_VERSION=$(grep -i -E "RUN pip install.+" "$SCRIPT_DIR/../Dockerfile" | grep --only-matching -i -E "playwright[=><~+]+[0-9\.]+")
|
||||||
|
echo "using $PLAYWRIGHT_VERSION"
|
||||||
|
pip3 install "$PLAYWRIGHT_VERSION"
|
||||||
|
docker run -d --name $$-test_browserless -e "DEFAULT_LAUNCH_ARGS=[\"--window-size=1920,1080\"]" --rm -p 3000:3000 --shm-size="2g" browserless/chrome:1.53-chrome-stable
|
||||||
|
# takes a while to spin up
|
||||||
|
sleep 5
|
||||||
|
export PLAYWRIGHT_DRIVER_URL=ws://127.0.0.1:3000
|
||||||
|
pytest tests/fetchers/test_content.py
|
||||||
|
pytest tests/test_errorhandling.py
|
||||||
|
pytest tests/visualselector/test_fetch_data.py
|
||||||
|
|
||||||
|
unset PLAYWRIGHT_DRIVER_URL
|
||||||
|
docker kill $$-test_browserless
|
||||||
|
|
||||||
|
# Test proxy list handling, starting two squids on different ports
|
||||||
|
# Each squid adds a different header to the response, which is the main thing we test for.
|
||||||
|
docker run -d --name $$-squid-one --rm -v `pwd`/tests/proxy_list/squid.conf:/etc/squid/conf.d/debian.conf -p 3128:3128 ubuntu/squid:4.13-21.10_edge
|
||||||
|
docker run -d --name $$-squid-two --rm -v `pwd`/tests/proxy_list/squid.conf:/etc/squid/conf.d/debian.conf -p 3129:3128 ubuntu/squid:4.13-21.10_edge
|
||||||
|
|
||||||
|
|
||||||
|
# So, basic HTTP as env var test
|
||||||
|
export HTTP_PROXY=http://localhost:3128
|
||||||
|
export HTTPS_PROXY=http://localhost:3128
|
||||||
|
pytest tests/proxy_list/test_proxy.py
|
||||||
|
docker logs $$-squid-one 2>/dev/null|grep one.changedetection.io
|
||||||
|
if [ $? -ne 0 ]
|
||||||
|
then
|
||||||
|
echo "Did not see a request to one.changedetection.io in the squid logs (while checking env vars HTTP_PROXY/HTTPS_PROXY)"
|
||||||
|
fi
|
||||||
|
unset HTTP_PROXY
|
||||||
|
unset HTTPS_PROXY
|
||||||
|
|
||||||
|
|
||||||
|
# 2nd test actually choose the preferred proxy from proxies.json
|
||||||
|
cp tests/proxy_list/proxies.json-example ./test-datastore/proxies.json
|
||||||
|
# Makes a watch use a preferred proxy
|
||||||
|
pytest tests/proxy_list/test_multiple_proxy.py
|
||||||
|
|
||||||
|
# Should be a request in the default "first" squid
|
||||||
|
docker logs $$-squid-one 2>/dev/null|grep chosen.changedetection.io
|
||||||
|
if [ $? -ne 0 ]
|
||||||
|
then
|
||||||
|
echo "Did not see a request to chosen.changedetection.io in the squid logs (while checking preferred proxy)"
|
||||||
|
fi
|
||||||
|
|
||||||
|
# And one in the 'second' squid (user selects this as preferred)
|
||||||
|
docker logs $$-squid-two 2>/dev/null|grep chosen.changedetection.io
|
||||||
|
if [ $? -ne 0 ]
|
||||||
|
then
|
||||||
|
echo "Did not see a request to chosen.changedetection.io in the squid logs (while checking preferred proxy)"
|
||||||
|
fi
|
||||||
|
|
||||||
|
# @todo - test system override proxy selection and watch defaults, setup a 3rd squid?
|
||||||
|
docker kill $$-squid-one
|
||||||
|
docker kill $$-squid-two
|
||||||
|
|
||||||
|
|
||||||
|
|||||||
BIN
changedetectionio/static/images/Playwright-icon.png
Normal file
BIN
changedetectionio/static/images/Playwright-icon.png
Normal file
Binary file not shown.
|
After Width: | Height: | Size: 6.2 KiB |
42
changedetectionio/static/images/bell-off.svg
Normal file
42
changedetectionio/static/images/bell-off.svg
Normal file
@@ -0,0 +1,42 @@
|
|||||||
|
<?xml version="1.0" encoding="UTF-8" standalone="no"?>
|
||||||
|
<svg
|
||||||
|
width="15"
|
||||||
|
height="16.363636"
|
||||||
|
viewBox="0 0 15 16.363636"
|
||||||
|
version="1.1"
|
||||||
|
id="svg4"
|
||||||
|
sodipodi:docname="bell-off.svg"
|
||||||
|
inkscape:version="1.1.1 (1:1.1+202109281949+c3084ef5ed)"
|
||||||
|
xmlns:inkscape="http://www.inkscape.org/namespaces/inkscape"
|
||||||
|
xmlns:sodipodi="http://sodipodi.sourceforge.net/DTD/sodipodi-0.dtd"
|
||||||
|
xmlns="http://www.w3.org/2000/svg"
|
||||||
|
xmlns:svg="http://www.w3.org/2000/svg">
|
||||||
|
<sodipodi:namedview
|
||||||
|
id="namedview5"
|
||||||
|
pagecolor="#ffffff"
|
||||||
|
bordercolor="#666666"
|
||||||
|
borderopacity="1.0"
|
||||||
|
inkscape:pageshadow="2"
|
||||||
|
inkscape:pageopacity="0.0"
|
||||||
|
inkscape:pagecheckerboard="0"
|
||||||
|
showgrid="false"
|
||||||
|
fit-margin-top="0"
|
||||||
|
fit-margin-left="0"
|
||||||
|
fit-margin-right="0"
|
||||||
|
fit-margin-bottom="0"
|
||||||
|
inkscape:zoom="28.416667"
|
||||||
|
inkscape:cx="-0.59824046"
|
||||||
|
inkscape:cy="12"
|
||||||
|
inkscape:window-width="1554"
|
||||||
|
inkscape:window-height="896"
|
||||||
|
inkscape:window-x="2095"
|
||||||
|
inkscape:window-y="107"
|
||||||
|
inkscape:window-maximized="0"
|
||||||
|
inkscape:current-layer="svg4" />
|
||||||
|
<defs
|
||||||
|
id="defs8" />
|
||||||
|
<path
|
||||||
|
d="m 14.318182,11.762045 v 1.1925 H 5.4102273 L 11.849318,7.1140909 C 12.234545,9.1561364 12.54,11.181818 14.318182,11.762045 Z m -6.7984093,4.601591 c 1.0759091,0 2.0256823,-0.955909 2.0256823,-2.045454 H 5.4545455 c 0,1.089545 0.9879545,2.045454 2.0652272,2.045454 z M 15,2.8622727 0.9177273,15.636136 0,14.627045 l 1.8443182,-1.6725 h -1.1625 v -1.1925 C 4.0070455,10.677273 2.1784091,4.5388636 5.3611364,2.6897727 5.8009091,2.4347727 6.0709091,1.9609091 6.0702273,1.4488636 v -0.00205 C 6.0702273,0.64772727 6.7104545,0 7.5,0 8.2895455,0 8.9297727,0.64772727 8.9297727,1.4468182 v 0.00205 C 8.9290909,1.9602319 9.199773,2.4354591 9.638864,2.6897773 10.364318,3.111141 10.827273,3.7568228 11.1525,4.5129591 L 14.085682,1.8531818 Z M 6.8181818,1.3636364 C 6.8181818,1.74 7.1236364,2.0454545 7.5,2.0454545 7.8763636,2.0454545 8.1818182,1.74 8.1818182,1.3636364 8.1818182,0.98795455 7.8763636,0.68181818 7.5,0.68181818 c -0.3763636,0 -0.6818182,0.30613637 -0.6818182,0.68181822 z"
|
||||||
|
id="path2"
|
||||||
|
style="fill:#f8321b;stroke-width:0.681818;fill-opacity:1" />
|
||||||
|
</svg>
|
||||||
|
After Width: | Height: | Size: 2.1 KiB |
BIN
changedetectionio/static/images/beta-logo.png
Normal file
BIN
changedetectionio/static/images/beta-logo.png
Normal file
Binary file not shown.
|
After Width: | Height: | Size: 12 KiB |
51
changedetectionio/static/images/notice.svg
Normal file
51
changedetectionio/static/images/notice.svg
Normal file
@@ -0,0 +1,51 @@
|
|||||||
|
<?xml version="1.0" encoding="UTF-8" standalone="no"?>
|
||||||
|
<!-- Created with Inkscape (http://www.inkscape.org/) -->
|
||||||
|
|
||||||
|
<svg
|
||||||
|
width="20.108334mm"
|
||||||
|
height="21.43125mm"
|
||||||
|
viewBox="0 0 20.108334 21.43125"
|
||||||
|
version="1.1"
|
||||||
|
id="svg5"
|
||||||
|
xmlns:xlink="http://www.w3.org/1999/xlink"
|
||||||
|
xmlns="http://www.w3.org/2000/svg"
|
||||||
|
xmlns:svg="http://www.w3.org/2000/svg">
|
||||||
|
<defs
|
||||||
|
id="defs2" />
|
||||||
|
<g
|
||||||
|
id="layer1"
|
||||||
|
transform="translate(-141.05873,-76.816635)">
|
||||||
|
<image
|
||||||
|
width="20.108334"
|
||||||
|
height="21.43125"
|
||||||
|
preserveAspectRatio="none"
|
||||||
|
style="image-rendering:optimizeQuality"
|
||||||
|
xlink:href="data:image/png;base64,iVBORw0KGgoAAAANSUhEUgAAAEwAAABRCAYAAAB430BuAAAABHNCSVQICAgIfAhkiAAABLxJREFU
|
||||||
|
eJztnN2Z2jgUhl8Z7petIGwF0WMXsFBBoIKwFWS2gmQryKSCJRXsTAUDBTDRVBCmgkAB9tkLexh+
|
||||||
|
bIONLGwP7xU2RjafpaOjoyNBCxHNQAJEfG5sl+3ZLrAWeAyST5/sF91mFH3bRbZbsAq4ClaQq2B7
|
||||||
|
iKYnmg9Z318F20ICRnj8pMOd6E3HscNVsATxmQD/oeghPCnDLO26q2AkYin+TQ7XREyyrn3zgu2J
|
||||||
|
BSEjZTBZ179pwQ7EEv7KaoovvFnBUsV6ZHrsd+0WTHhKPV1SLGivYEsA1KEtEs2grFitRjQ65VxP
|
||||||
|
fH5JgEjAKsvXupKwFfYxaYJeSeHcWqVSCuwD7/HQQD8lRHLWDStBWG3slbAElkTc5/lTZdkIJhpN
|
||||||
|
h6/UUZDyzAgZK8PKVoEKErE8HlD0bBVcI2ZqwdBWYbFgAT+g1UZwrBbcvRyIpofHJ1Sh1rQCZt1k
|
||||||
|
lN5msQAm8CoYoFF8KVHOsFtQ5aayExBUhpnopJl6J/3/FREGWCrxmaH40/4z1oyQ320Yf5dDozXC
|
||||||
|
P4QMCRkCY4S5w/tbMTtd4L2Ngo6wJmSQ4hfdScAU+OjgGazgOXEl8oJyof3Z6Spx0iTzgnLKsMoK
|
||||||
|
w9SRuoR3rHniVVMXwRpDXQR7d+kHOJV6CFZB0khVOBGsTcE6VzWsNVGQizfJptU+N4LlD3AbVfsu
|
||||||
|
XsOahhvB8nrB08IrtcGNYNIct+EYl2+S6mr0D8kLUMrV6BfFRTzOGs4Ey8p1aNrUnssaliaMO/vV
|
||||||
|
sfNi3AmW5j54DgUTO/dyJ1hab9iwHhLcNskP23ZMND0kewFBXek6vZvHg/hMiUPSN00z+OBasFig
|
||||||
|
y8wSRfnZ0adSBz+sUVwFK4jbJhnPP06To1ETczpcCnavHhltHd82LU0AXDbJMGXBU8PSBAA8Jxk0
|
||||||
|
wnNaqlGSJuAyg+dsXIV38iZqXU3iWsmodhetSNlDQgJGriZxbWVSe1hS/gQ+S/C6j4QEfES21vxU
|
||||||
|
icXsoC4vC5mqJvbybyXgduucG/YWaYmmj+IdHvpoxFdt8ltRP5h3iZjRqfBh60C4t1rNY7rxAU95
|
||||||
|
aYnhEp+/u8pgxGfeRCfyJIR5SkLfFOHYXMMzu63PEDF9WQnSo8MUmhduyUWYEzGyvnRmU3683ugG
|
||||||
|
GAG/2bqJU4RnFDNCpsfWb5chswUnwb5Xg+hxiyo9w7MGJoSVpmYulam+A8scS+5nPYtf+s9mpZw7
|
||||||
|
J1nayDnCVuu4Ck+E6DqIBYDHHR1+is/n8kVUhfBExMBFMzm4taafkXcWL9BSfBG/nNN8sutYcE3S
|
||||||
|
d7XI3o6lSpIe/xcAIX/svzDxMVu22BAyLNKL2q9hwrdLiZWwXbP6B99GDLaGSpoOD6JPn4yxK1i8
|
||||||
|
B0StY1zKsCJiQNxzQ0HRbAm2BsZN2TBDGVaE5USzIVjsNix2VrzWHmUwB6J5fD32uyKCzQ7OxG5D
|
||||||
|
vzZuQ0E2osXjRlBMjvWe5WtYPE4b2BynXQJlMEToTUegmEiwM1mzQ1nBvqvH5ov1wlZHcA+AZHdc
|
||||||
|
xQW7vNuQS9kBtzKs1IIRMM7b0q/YvGTzto4qbFutdV5FnLtLk2x3JVWUfXKTbIu9Opc2J6Osj19S
|
||||||
|
HLfJKO64r6rg/wFBX3+2ZapW8wAAAABJRU5ErkJggg==
|
||||||
|
"
|
||||||
|
id="image832"
|
||||||
|
x="141.05873"
|
||||||
|
y="76.816635" />
|
||||||
|
</g>
|
||||||
|
</svg>
|
||||||
|
After Width: | Height: | Size: 2.4 KiB |
122
changedetectionio/static/images/play.svg
Normal file
122
changedetectionio/static/images/play.svg
Normal file
@@ -0,0 +1,122 @@
|
|||||||
|
<?xml version="1.0" encoding="UTF-8" standalone="no"?>
|
||||||
|
<svg
|
||||||
|
version="1.1"
|
||||||
|
id="Capa_1"
|
||||||
|
x="0px"
|
||||||
|
y="0px"
|
||||||
|
viewBox="0 0 15 14.998326"
|
||||||
|
xml:space="preserve"
|
||||||
|
width="15"
|
||||||
|
height="14.998326"
|
||||||
|
sodipodi:docname="play.svg"
|
||||||
|
inkscape:version="1.1.1 (1:1.1+202109281949+c3084ef5ed)"
|
||||||
|
xmlns:inkscape="http://www.inkscape.org/namespaces/inkscape"
|
||||||
|
xmlns:sodipodi="http://sodipodi.sourceforge.net/DTD/sodipodi-0.dtd"
|
||||||
|
xmlns="http://www.w3.org/2000/svg"
|
||||||
|
xmlns:svg="http://www.w3.org/2000/svg"
|
||||||
|
xmlns:rdf="http://www.w3.org/1999/02/22-rdf-syntax-ns#"
|
||||||
|
xmlns:cc="http://creativecommons.org/ns#"
|
||||||
|
xmlns:dc="http://purl.org/dc/elements/1.1/"><sodipodi:namedview
|
||||||
|
id="namedview21"
|
||||||
|
pagecolor="#ffffff"
|
||||||
|
bordercolor="#666666"
|
||||||
|
borderopacity="1.0"
|
||||||
|
inkscape:pageshadow="2"
|
||||||
|
inkscape:pageopacity="0.0"
|
||||||
|
inkscape:pagecheckerboard="0"
|
||||||
|
showgrid="false"
|
||||||
|
inkscape:zoom="45.47174"
|
||||||
|
inkscape:cx="7.4991632"
|
||||||
|
inkscape:cy="7.4991632"
|
||||||
|
inkscape:window-width="1554"
|
||||||
|
inkscape:window-height="896"
|
||||||
|
inkscape:window-x="3048"
|
||||||
|
inkscape:window-y="227"
|
||||||
|
inkscape:window-maximized="0"
|
||||||
|
inkscape:current-layer="Capa_1" /><metadata
|
||||||
|
id="metadata39"><rdf:RDF><cc:Work
|
||||||
|
rdf:about=""><dc:format>image/svg+xml</dc:format><dc:type
|
||||||
|
rdf:resource="http://purl.org/dc/dcmitype/StillImage" /></cc:Work></rdf:RDF></metadata><defs
|
||||||
|
id="defs37" />
|
||||||
|
<path
|
||||||
|
id="path2"
|
||||||
|
style="fill:#1b98f8;fill-opacity:1;stroke-width:0.0292893"
|
||||||
|
d="M 7.4980469,0 C 4.5496028,-0.04093755 1.7047721,1.8547661 0.58789062,4.5800781 -0.57819305,7.2574082 0.02636631,10.583252 2.0703125,12.671875 4.0368718,14.788335 7.2754393,15.560096 9.9882812,14.572266 12.800219,13.617028 14.874915,10.855516 14.986328,7.8847656 15.172991,4.9968456 13.497714,2.109448 10.910156,0.8203125 9.858961,0.28011352 8.6796569,-0.00179908 7.4980469,0 Z"
|
||||||
|
sodipodi:nodetypes="ccccccc" />
|
||||||
|
<g
|
||||||
|
id="g4"
|
||||||
|
transform="translate(-0.01903604,0.02221043)">
|
||||||
|
</g>
|
||||||
|
<g
|
||||||
|
id="g6"
|
||||||
|
transform="translate(-0.01903604,0.02221043)">
|
||||||
|
</g>
|
||||||
|
<g
|
||||||
|
id="g8"
|
||||||
|
transform="translate(-0.01903604,0.02221043)">
|
||||||
|
</g>
|
||||||
|
<g
|
||||||
|
id="g10"
|
||||||
|
transform="translate(-0.01903604,0.02221043)">
|
||||||
|
</g>
|
||||||
|
<g
|
||||||
|
id="g12"
|
||||||
|
transform="translate(-0.01903604,0.02221043)">
|
||||||
|
</g>
|
||||||
|
<g
|
||||||
|
id="g14"
|
||||||
|
transform="translate(-0.01903604,0.02221043)">
|
||||||
|
</g>
|
||||||
|
<g
|
||||||
|
id="g16"
|
||||||
|
transform="translate(-0.01903604,0.02221043)">
|
||||||
|
</g>
|
||||||
|
<g
|
||||||
|
id="g18"
|
||||||
|
transform="translate(-0.01903604,0.02221043)">
|
||||||
|
</g>
|
||||||
|
<g
|
||||||
|
id="g20"
|
||||||
|
transform="translate(-0.01903604,0.02221043)">
|
||||||
|
</g>
|
||||||
|
<g
|
||||||
|
id="g22"
|
||||||
|
transform="translate(-0.01903604,0.02221043)">
|
||||||
|
</g>
|
||||||
|
<g
|
||||||
|
id="g24"
|
||||||
|
transform="translate(-0.01903604,0.02221043)">
|
||||||
|
</g>
|
||||||
|
<g
|
||||||
|
id="g26"
|
||||||
|
transform="translate(-0.01903604,0.02221043)">
|
||||||
|
</g>
|
||||||
|
<g
|
||||||
|
id="g28"
|
||||||
|
transform="translate(-0.01903604,0.02221043)">
|
||||||
|
</g>
|
||||||
|
<g
|
||||||
|
id="g30"
|
||||||
|
transform="translate(-0.01903604,0.02221043)">
|
||||||
|
</g>
|
||||||
|
<g
|
||||||
|
id="g32"
|
||||||
|
transform="translate(-0.01903604,0.02221043)">
|
||||||
|
</g>
|
||||||
|
<path
|
||||||
|
sodipodi:type="star"
|
||||||
|
style="fill:#ffffff;fill-opacity:1;stroke-width:37.7953;paint-order:stroke fill markers"
|
||||||
|
id="path1203"
|
||||||
|
inkscape:flatsided="false"
|
||||||
|
sodipodi:sides="3"
|
||||||
|
sodipodi:cx="7.2964563"
|
||||||
|
sodipodi:cy="7.3240671"
|
||||||
|
sodipodi:r1="3.805218"
|
||||||
|
sodipodi:r2="1.9026089"
|
||||||
|
sodipodi:arg1="-0.0017436774"
|
||||||
|
sodipodi:arg2="1.0454539"
|
||||||
|
inkscape:rounded="0"
|
||||||
|
inkscape:randomized="0"
|
||||||
|
d="M 11.101669,7.317432 8.2506324,8.9701135 5.3995964,10.622795 5.3938504,7.3273846 5.3881041,4.0319742 8.2448863,5.6747033 Z"
|
||||||
|
inkscape:transform-center-x="-0.94843001"
|
||||||
|
inkscape:transform-center-y="0.0033175346" /></svg>
|
||||||
|
After Width: | Height: | Size: 3.5 KiB |
20
changedetectionio/static/images/spread-white.svg
Normal file
20
changedetectionio/static/images/spread-white.svg
Normal file
@@ -0,0 +1,20 @@
|
|||||||
|
<?xml version="1.0" encoding="UTF-8" standalone="no"?>
|
||||||
|
<svg
|
||||||
|
width="18"
|
||||||
|
height="19.92"
|
||||||
|
viewBox="0 0 18 19.92"
|
||||||
|
version="1.1"
|
||||||
|
id="svg6"
|
||||||
|
xmlns="http://www.w3.org/2000/svg"
|
||||||
|
xmlns:svg="http://www.w3.org/2000/svg">
|
||||||
|
<defs
|
||||||
|
id="defs10" />
|
||||||
|
<path
|
||||||
|
d="M -3,-2 H 21 V 22 H -3 Z"
|
||||||
|
fill="none"
|
||||||
|
id="path2" />
|
||||||
|
<path
|
||||||
|
d="m 15,14.08 c -0.76,0 -1.44,0.3 -1.96,0.77 L 5.91,10.7 C 5.96,10.47 6,10.24 6,10 6,9.76 5.96,9.53 5.91,9.3 L 12.96,5.19 C 13.5,5.69 14.21,6 15,6 16.66,6 18,4.66 18,3 18,1.34 16.66,0 15,0 c -1.66,0 -3,1.34 -3,3 0,0.24 0.04,0.47 0.09,0.7 L 5.04,7.81 C 4.5,7.31 3.79,7 3,7 1.34,7 0,8.34 0,10 c 0,1.66 1.34,3 3,3 0.79,0 1.5,-0.31 2.04,-0.81 l 7.12,4.16 c -0.05,0.21 -0.08,0.43 -0.08,0.65 0,1.61 1.31,2.92 2.92,2.92 1.61,0 2.92,-1.31 2.92,-2.92 0,-1.61 -1.31,-2.92 -2.92,-2.92 z"
|
||||||
|
id="path4"
|
||||||
|
style="fill:#ffffff;fill-opacity:1" />
|
||||||
|
</svg>
|
||||||
|
After Width: | Height: | Size: 892 B |
23
changedetectionio/static/js/diff-overview.js
Normal file
23
changedetectionio/static/js/diff-overview.js
Normal file
@@ -0,0 +1,23 @@
|
|||||||
|
$(document).ready(function () {
|
||||||
|
// Load it when the #screenshot tab is in use, so we dont give a slow experience when waiting for the text diff to load
|
||||||
|
window.addEventListener('hashchange', function (e) {
|
||||||
|
toggle(location.hash);
|
||||||
|
}, false);
|
||||||
|
|
||||||
|
toggle(location.hash);
|
||||||
|
|
||||||
|
function toggle(hash_name) {
|
||||||
|
if (hash_name === '#screenshot') {
|
||||||
|
$("img#screenshot-img").attr('src', screenshot_url);
|
||||||
|
$("#settings").hide();
|
||||||
|
} else if (hash_name === '#error-screenshot') {
|
||||||
|
$("img#error-screenshot-img").attr('src', error_screenshot_url);
|
||||||
|
$("#settings").hide();
|
||||||
|
}
|
||||||
|
|
||||||
|
|
||||||
|
else {
|
||||||
|
$("#settings").show();
|
||||||
|
}
|
||||||
|
}
|
||||||
|
});
|
||||||
@@ -1,4 +1,4 @@
|
|||||||
$(document).ready(function() {
|
$(document).ready(function () {
|
||||||
function toggle() {
|
function toggle() {
|
||||||
if ($('input[name="application-fetch_backend"]:checked').val() != 'html_requests') {
|
if ($('input[name="application-fetch_backend"]:checked').val() != 'html_requests') {
|
||||||
$('#requests-override-options').hide();
|
$('#requests-override-options').hide();
|
||||||
@@ -8,9 +8,29 @@ $(document).ready(function() {
|
|||||||
$('#webdriver-override-options').hide();
|
$('#webdriver-override-options').hide();
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
$('input[name="application-fetch_backend"]').click(function (e) {
|
$('input[name="application-fetch_backend"]').click(function (e) {
|
||||||
toggle();
|
toggle();
|
||||||
});
|
});
|
||||||
toggle();
|
toggle();
|
||||||
|
|
||||||
|
$("#api-key").hover(
|
||||||
|
function () {
|
||||||
|
$("#api-key-copy").html('copy').fadeIn();
|
||||||
|
},
|
||||||
|
function () {
|
||||||
|
$("#api-key-copy").hide();
|
||||||
|
}
|
||||||
|
).click(function (e) {
|
||||||
|
$("#api-key-copy").html('copied');
|
||||||
|
var range = document.createRange();
|
||||||
|
var n = $("#api-key")[0];
|
||||||
|
range.selectNode(n);
|
||||||
|
window.getSelection().removeAllRanges();
|
||||||
|
window.getSelection().addRange(range);
|
||||||
|
document.execCommand("copy");
|
||||||
|
window.getSelection().removeAllRanges();
|
||||||
|
|
||||||
|
});
|
||||||
});
|
});
|
||||||
|
|
||||||
|
|||||||
56
changedetectionio/static/js/limit.js
Normal file
56
changedetectionio/static/js/limit.js
Normal file
@@ -0,0 +1,56 @@
|
|||||||
|
/**
|
||||||
|
* debounce
|
||||||
|
* @param {integer} milliseconds This param indicates the number of milliseconds
|
||||||
|
* to wait after the last call before calling the original function.
|
||||||
|
* @param {object} What "this" refers to in the returned function.
|
||||||
|
* @return {function} This returns a function that when called will wait the
|
||||||
|
* indicated number of milliseconds after the last call before
|
||||||
|
* calling the original function.
|
||||||
|
*/
|
||||||
|
Function.prototype.debounce = function (milliseconds, context) {
|
||||||
|
var baseFunction = this,
|
||||||
|
timer = null,
|
||||||
|
wait = milliseconds;
|
||||||
|
|
||||||
|
return function () {
|
||||||
|
var self = context || this,
|
||||||
|
args = arguments;
|
||||||
|
|
||||||
|
function complete() {
|
||||||
|
baseFunction.apply(self, args);
|
||||||
|
timer = null;
|
||||||
|
}
|
||||||
|
|
||||||
|
if (timer) {
|
||||||
|
clearTimeout(timer);
|
||||||
|
}
|
||||||
|
|
||||||
|
timer = setTimeout(complete, wait);
|
||||||
|
};
|
||||||
|
};
|
||||||
|
|
||||||
|
/**
|
||||||
|
* throttle
|
||||||
|
* @param {integer} milliseconds This param indicates the number of milliseconds
|
||||||
|
* to wait between calls before calling the original function.
|
||||||
|
* @param {object} What "this" refers to in the returned function.
|
||||||
|
* @return {function} This returns a function that when called will wait the
|
||||||
|
* indicated number of milliseconds between calls before
|
||||||
|
* calling the original function.
|
||||||
|
*/
|
||||||
|
Function.prototype.throttle = function (milliseconds, context) {
|
||||||
|
var baseFunction = this,
|
||||||
|
lastEventTimestamp = null,
|
||||||
|
limit = milliseconds;
|
||||||
|
|
||||||
|
return function () {
|
||||||
|
var self = context || this,
|
||||||
|
args = arguments,
|
||||||
|
now = Date.now();
|
||||||
|
|
||||||
|
if (!lastEventTimestamp || now - lastEventTimestamp >= limit) {
|
||||||
|
lastEventTimestamp = now;
|
||||||
|
baseFunction.apply(self, args);
|
||||||
|
}
|
||||||
|
};
|
||||||
|
};
|
||||||
@@ -40,13 +40,19 @@ $(document).ready(function() {
|
|||||||
$.ajax({
|
$.ajax({
|
||||||
type: "POST",
|
type: "POST",
|
||||||
url: notification_base_url,
|
url: notification_base_url,
|
||||||
data : data
|
data : data,
|
||||||
|
statusCode: {
|
||||||
|
400: function() {
|
||||||
|
// More than likely the CSRF token was lost when the server restarted
|
||||||
|
alert("There was a problem processing the request, please reload the page.");
|
||||||
|
}
|
||||||
|
}
|
||||||
}).done(function(data){
|
}).done(function(data){
|
||||||
console.log(data);
|
console.log(data);
|
||||||
alert('Sent');
|
alert('Sent');
|
||||||
}).fail(function(data){
|
}).fail(function(data){
|
||||||
console.log(data);
|
console.log(data);
|
||||||
alert('Error: '+data.responseJSON.error);
|
alert('There was an error communicating with the server.');
|
||||||
})
|
})
|
||||||
});
|
});
|
||||||
});
|
});
|
||||||
|
|||||||
@@ -1,12 +1,6 @@
|
|||||||
// Rewrite this is a plugin.. is all this JS really 'worth it?'
|
// Rewrite this is a plugin.. is all this JS really 'worth it?'
|
||||||
|
|
||||||
|
window.addEventListener('hashchange', function () {
|
||||||
if(!window.location.hash) {
|
|
||||||
var tab=document.querySelectorAll("#default-tab a");
|
|
||||||
tab[0].click();
|
|
||||||
}
|
|
||||||
|
|
||||||
window.addEventListener('hashchange', function() {
|
|
||||||
var tabs = document.getElementsByClassName('active');
|
var tabs = document.getElementsByClassName('active');
|
||||||
while (tabs[0]) {
|
while (tabs[0]) {
|
||||||
tabs[0].classList.remove('active')
|
tabs[0].classList.remove('active')
|
||||||
@@ -14,11 +8,10 @@ window.addEventListener('hashchange', function() {
|
|||||||
set_active_tab();
|
set_active_tab();
|
||||||
}, false);
|
}, false);
|
||||||
|
|
||||||
var has_errors=document.querySelectorAll(".messages .error");
|
var has_errors = document.querySelectorAll(".messages .error");
|
||||||
if (!has_errors.length) {
|
if (!has_errors.length) {
|
||||||
if (document.location.hash == "" ) {
|
if (document.location.hash == "") {
|
||||||
document.location.hash = "#general";
|
document.querySelector(".tabs ul li:first-child a").click();
|
||||||
document.getElementById("default-tab").className = "active";
|
|
||||||
} else {
|
} else {
|
||||||
set_active_tab();
|
set_active_tab();
|
||||||
}
|
}
|
||||||
@@ -27,9 +20,9 @@ if (!has_errors.length) {
|
|||||||
}
|
}
|
||||||
|
|
||||||
function set_active_tab() {
|
function set_active_tab() {
|
||||||
var tab=document.querySelectorAll("a[href='"+location.hash+"']");
|
var tab = document.querySelectorAll("a[href='" + location.hash + "']");
|
||||||
if (tab.length) {
|
if (tab.length) {
|
||||||
tab[0].parentElement.className="active";
|
tab[0].parentElement.className = "active";
|
||||||
}
|
}
|
||||||
// hash could move the page down
|
// hash could move the page down
|
||||||
window.scrollTo(0, 0);
|
window.scrollTo(0, 0);
|
||||||
@@ -38,12 +31,12 @@ function set_active_tab() {
|
|||||||
function focus_error_tab() {
|
function focus_error_tab() {
|
||||||
// time to use jquery or vuejs really,
|
// time to use jquery or vuejs really,
|
||||||
// activate the tab with the error
|
// activate the tab with the error
|
||||||
var tabs = document.querySelectorAll('.tabs li a'),i;
|
var tabs = document.querySelectorAll('.tabs li a'), i;
|
||||||
for (i = 0; i < tabs.length; ++i) {
|
for (i = 0; i < tabs.length; ++i) {
|
||||||
var tab_name=tabs[i].hash.replace('#','');
|
var tab_name = tabs[i].hash.replace('#', '');
|
||||||
var pane_errors=document.querySelectorAll('#'+tab_name+' .error')
|
var pane_errors = document.querySelectorAll('#' + tab_name + ' .error')
|
||||||
if (pane_errors.length) {
|
if (pane_errors.length) {
|
||||||
document.location.hash = '#'+tab_name;
|
document.location.hash = '#' + tab_name;
|
||||||
return true;
|
return true;
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|||||||
230
changedetectionio/static/js/visual-selector.js
Normal file
230
changedetectionio/static/js/visual-selector.js
Normal file
@@ -0,0 +1,230 @@
|
|||||||
|
// Horrible proof of concept code :)
|
||||||
|
// yes - this is really a hack, if you are a front-ender and want to help, please get in touch!
|
||||||
|
|
||||||
|
$(document).ready(function() {
|
||||||
|
|
||||||
|
var current_selected_i;
|
||||||
|
var state_clicked=false;
|
||||||
|
|
||||||
|
var c;
|
||||||
|
|
||||||
|
// greyed out fill context
|
||||||
|
var xctx;
|
||||||
|
// redline highlight context
|
||||||
|
var ctx;
|
||||||
|
|
||||||
|
var current_default_xpath;
|
||||||
|
var x_scale=1;
|
||||||
|
var y_scale=1;
|
||||||
|
var selector_image;
|
||||||
|
var selector_image_rect;
|
||||||
|
var selector_data;
|
||||||
|
|
||||||
|
$('#visualselector-tab').click(function () {
|
||||||
|
$("img#selector-background").off('load');
|
||||||
|
state_clicked = false;
|
||||||
|
current_selected_i = false;
|
||||||
|
bootstrap_visualselector();
|
||||||
|
});
|
||||||
|
|
||||||
|
$(document).on('keydown', function(event) {
|
||||||
|
if ($("img#selector-background").is(":visible")) {
|
||||||
|
if (event.key == "Escape") {
|
||||||
|
state_clicked=false;
|
||||||
|
ctx.clearRect(0, 0, c.width, c.height);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
});
|
||||||
|
|
||||||
|
// For when the page loads
|
||||||
|
if(!window.location.hash || window.location.hash != '#visualselector') {
|
||||||
|
$("img#selector-background").attr('src','');
|
||||||
|
return;
|
||||||
|
}
|
||||||
|
|
||||||
|
// Handle clearing button/link
|
||||||
|
$('#clear-selector').on('click', function(event) {
|
||||||
|
if(!state_clicked) {
|
||||||
|
alert('Oops, Nothing selected!');
|
||||||
|
}
|
||||||
|
state_clicked=false;
|
||||||
|
ctx.clearRect(0, 0, c.width, c.height);
|
||||||
|
xctx.clearRect(0, 0, c.width, c.height);
|
||||||
|
$("#include_filters").val('');
|
||||||
|
});
|
||||||
|
|
||||||
|
|
||||||
|
bootstrap_visualselector();
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
|
function bootstrap_visualselector() {
|
||||||
|
if ( 1 ) {
|
||||||
|
// bootstrap it, this will trigger everything else
|
||||||
|
$("img#selector-background").bind('load', function () {
|
||||||
|
console.log("Loaded background...");
|
||||||
|
c = document.getElementById("selector-canvas");
|
||||||
|
// greyed out fill context
|
||||||
|
xctx = c.getContext("2d");
|
||||||
|
// redline highlight context
|
||||||
|
ctx = c.getContext("2d");
|
||||||
|
current_default_xpath =$("#include_filters").val();
|
||||||
|
fetch_data();
|
||||||
|
$('#selector-canvas').off("mousemove mousedown");
|
||||||
|
// screenshot_url defined in the edit.html template
|
||||||
|
}).attr("src", screenshot_url);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
function fetch_data() {
|
||||||
|
// Image is ready
|
||||||
|
$('.fetching-update-notice').html("Fetching element data..");
|
||||||
|
|
||||||
|
$.ajax({
|
||||||
|
url: watch_visual_selector_data_url,
|
||||||
|
context: document.body
|
||||||
|
}).done(function (data) {
|
||||||
|
$('.fetching-update-notice').html("Rendering..");
|
||||||
|
selector_data = data;
|
||||||
|
console.log("Reported browser width from backend: "+data['browser_width']);
|
||||||
|
state_clicked=false;
|
||||||
|
set_scale();
|
||||||
|
reflow_selector();
|
||||||
|
$('.fetching-update-notice').fadeOut();
|
||||||
|
});
|
||||||
|
};
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
|
function set_scale() {
|
||||||
|
|
||||||
|
// some things to check if the scaling doesnt work
|
||||||
|
// - that the widths/sizes really are about the actual screen size cat elements.json |grep -o width......|sort|uniq
|
||||||
|
selector_image = $("img#selector-background")[0];
|
||||||
|
selector_image_rect = selector_image.getBoundingClientRect();
|
||||||
|
|
||||||
|
// make the canvas the same size as the image
|
||||||
|
$('#selector-canvas').attr('height', selector_image_rect.height);
|
||||||
|
$('#selector-canvas').attr('width', selector_image_rect.width);
|
||||||
|
$('#selector-wrapper').attr('width', selector_image_rect.width);
|
||||||
|
x_scale = selector_image_rect.width / selector_data['browser_width'];
|
||||||
|
y_scale = selector_image_rect.height / selector_image.naturalHeight;
|
||||||
|
ctx.strokeStyle = 'rgba(255,0,0, 0.9)';
|
||||||
|
ctx.fillStyle = 'rgba(255,0,0, 0.1)';
|
||||||
|
ctx.lineWidth = 3;
|
||||||
|
console.log("scaling set x: "+x_scale+" by y:"+y_scale);
|
||||||
|
$("#selector-current-xpath").css('max-width', selector_image_rect.width);
|
||||||
|
}
|
||||||
|
|
||||||
|
function reflow_selector() {
|
||||||
|
$(window).resize(function() {
|
||||||
|
set_scale();
|
||||||
|
highlight_current_selected_i();
|
||||||
|
});
|
||||||
|
var selector_currnt_xpath_text=$("#selector-current-xpath span");
|
||||||
|
|
||||||
|
set_scale();
|
||||||
|
|
||||||
|
console.log(selector_data['size_pos'].length + " selectors found");
|
||||||
|
|
||||||
|
// highlight the default one if we can find it in the xPath list
|
||||||
|
// or the xpath matches the default one
|
||||||
|
found = false;
|
||||||
|
if(current_default_xpath.length) {
|
||||||
|
for (var i = selector_data['size_pos'].length; i!==0; i--) {
|
||||||
|
var sel = selector_data['size_pos'][i-1];
|
||||||
|
if(selector_data['size_pos'][i - 1].xpath == current_default_xpath) {
|
||||||
|
console.log("highlighting "+current_default_xpath);
|
||||||
|
current_selected_i = i-1;
|
||||||
|
highlight_current_selected_i();
|
||||||
|
found = true;
|
||||||
|
break;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
if(!found) {
|
||||||
|
alert("Unfortunately your existing CSS/xPath Filter was no longer found!");
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
|
||||||
|
$('#selector-canvas').bind('mousemove', function (e) {
|
||||||
|
if(state_clicked) {
|
||||||
|
return;
|
||||||
|
}
|
||||||
|
ctx.clearRect(0, 0, c.width, c.height);
|
||||||
|
current_selected_i=null;
|
||||||
|
|
||||||
|
// Add in offset
|
||||||
|
if ((typeof e.offsetX === "undefined" || typeof e.offsetY === "undefined") || (e.offsetX === 0 && e.offsetY === 0)) {
|
||||||
|
var targetOffset = $(e.target).offset();
|
||||||
|
e.offsetX = e.pageX - targetOffset.left;
|
||||||
|
e.offsetY = e.pageY - targetOffset.top;
|
||||||
|
}
|
||||||
|
|
||||||
|
// Reverse order - the most specific one should be deeper/"laster"
|
||||||
|
// Basically, find the most 'deepest'
|
||||||
|
var found=0;
|
||||||
|
ctx.fillStyle = 'rgba(205,0,0,0.35)';
|
||||||
|
for (var i = selector_data['size_pos'].length; i!==0; i--) {
|
||||||
|
// draw all of them? let them choose somehow?
|
||||||
|
var sel = selector_data['size_pos'][i-1];
|
||||||
|
// If we are in a bounding-box
|
||||||
|
if (e.offsetY > sel.top * y_scale && e.offsetY < sel.top * y_scale + sel.height * y_scale
|
||||||
|
&&
|
||||||
|
e.offsetX > sel.left * y_scale && e.offsetX < sel.left * y_scale + sel.width * y_scale
|
||||||
|
|
||||||
|
) {
|
||||||
|
|
||||||
|
// FOUND ONE
|
||||||
|
set_current_selected_text(sel.xpath);
|
||||||
|
ctx.strokeRect(sel.left * x_scale, sel.top * y_scale, sel.width * x_scale, sel.height * y_scale);
|
||||||
|
ctx.fillRect(sel.left * x_scale, sel.top * y_scale, sel.width * x_scale, sel.height * y_scale);
|
||||||
|
|
||||||
|
// no need to keep digging
|
||||||
|
// @todo or, O to go out/up, I to go in
|
||||||
|
// or double click to go up/out the selector?
|
||||||
|
current_selected_i=i-1;
|
||||||
|
found+=1;
|
||||||
|
break;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
}.debounce(5));
|
||||||
|
|
||||||
|
function set_current_selected_text(s) {
|
||||||
|
selector_currnt_xpath_text[0].innerHTML=s;
|
||||||
|
}
|
||||||
|
|
||||||
|
function highlight_current_selected_i() {
|
||||||
|
if(state_clicked) {
|
||||||
|
state_clicked=false;
|
||||||
|
xctx.clearRect(0,0,c.width, c.height);
|
||||||
|
return;
|
||||||
|
}
|
||||||
|
|
||||||
|
var sel = selector_data['size_pos'][current_selected_i];
|
||||||
|
if (sel[0] == '/') {
|
||||||
|
// @todo - not sure just checking / is right
|
||||||
|
$("#include_filters").val('xpath:'+sel.xpath);
|
||||||
|
} else {
|
||||||
|
$("#include_filters").val(sel.xpath);
|
||||||
|
}
|
||||||
|
xctx.fillStyle = 'rgba(205,205,205,0.95)';
|
||||||
|
xctx.strokeStyle = 'rgba(225,0,0,0.9)';
|
||||||
|
xctx.lineWidth = 3;
|
||||||
|
xctx.fillRect(0,0,c.width, c.height);
|
||||||
|
// Clear out what only should be seen (make a clear/clean spot)
|
||||||
|
xctx.clearRect(sel.left * x_scale, sel.top * y_scale, sel.width * x_scale, sel.height * y_scale);
|
||||||
|
xctx.strokeRect(sel.left * x_scale, sel.top * y_scale, sel.width * x_scale, sel.height * y_scale);
|
||||||
|
state_clicked=true;
|
||||||
|
set_current_selected_text(sel.xpath);
|
||||||
|
|
||||||
|
}
|
||||||
|
|
||||||
|
|
||||||
|
$('#selector-canvas').bind('mousedown', function (e) {
|
||||||
|
highlight_current_selected_i();
|
||||||
|
});
|
||||||
|
}
|
||||||
|
|
||||||
|
});
|
||||||
@@ -4,6 +4,7 @@ $(function () {
|
|||||||
$(this).closest('.unviewed').removeClass('unviewed');
|
$(this).closest('.unviewed').removeClass('unviewed');
|
||||||
});
|
});
|
||||||
|
|
||||||
|
|
||||||
$('.with-share-link > *').click(function () {
|
$('.with-share-link > *').click(function () {
|
||||||
$("#copied-clipboard").remove();
|
$("#copied-clipboard").remove();
|
||||||
|
|
||||||
@@ -20,5 +21,19 @@ $(function () {
|
|||||||
$(this).remove();
|
$(this).remove();
|
||||||
});
|
});
|
||||||
});
|
});
|
||||||
|
|
||||||
|
// checkboxes - check all
|
||||||
|
$("#check-all").click(function (e) {
|
||||||
|
$('input[type=checkbox]').not(this).prop('checked', this.checked);
|
||||||
|
});
|
||||||
|
// checkboxes - show/hide buttons
|
||||||
|
$("input[type=checkbox]").click(function (e) {
|
||||||
|
if ($('input[type=checkbox]:checked').length) {
|
||||||
|
$('#checkbox-operations').slideDown();
|
||||||
|
} else {
|
||||||
|
$('#checkbox-operations').slideUp();
|
||||||
|
}
|
||||||
|
});
|
||||||
|
|
||||||
});
|
});
|
||||||
|
|
||||||
|
|||||||
@@ -1,16 +1,40 @@
|
|||||||
$(document).ready(function() {
|
$(document).ready(function() {
|
||||||
function toggle() {
|
function toggle() {
|
||||||
if ($('input[name="fetch_backend"]:checked').val() != 'html_requests') {
|
if ($('input[name="fetch_backend"]:checked').val() == 'html_webdriver') {
|
||||||
$('#requests-override-options').hide();
|
if(playwright_enabled) {
|
||||||
$('#webdriver-override-options').show();
|
// playwright supports headers, so hide everything else
|
||||||
|
// See #664
|
||||||
|
$('#requests-override-options #request-method').hide();
|
||||||
|
$('#requests-override-options #request-body').hide();
|
||||||
|
|
||||||
|
// @todo connect this one up
|
||||||
|
$('#ignore-status-codes-option').hide();
|
||||||
} else {
|
} else {
|
||||||
|
// selenium/webdriver doesnt support anything afaik, hide it all
|
||||||
|
$('#requests-override-options').hide();
|
||||||
|
}
|
||||||
|
|
||||||
|
|
||||||
|
$('#webdriver-override-options').show();
|
||||||
|
|
||||||
|
} else {
|
||||||
|
|
||||||
$('#requests-override-options').show();
|
$('#requests-override-options').show();
|
||||||
|
$('#requests-override-options *:hidden').show();
|
||||||
$('#webdriver-override-options').hide();
|
$('#webdriver-override-options').hide();
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
$('input[name="fetch_backend"]').click(function (e) {
|
$('input[name="fetch_backend"]').click(function (e) {
|
||||||
toggle();
|
toggle();
|
||||||
});
|
});
|
||||||
toggle();
|
toggle();
|
||||||
|
|
||||||
|
$('#notification-setting-reset-to-default').click(function (e) {
|
||||||
|
$('#notification_title').val('');
|
||||||
|
$('#notification_body').val('');
|
||||||
|
$('#notification_format').val('System default');
|
||||||
|
$('#notification_urls').val('');
|
||||||
|
e.preventDefault();
|
||||||
|
});
|
||||||
});
|
});
|
||||||
|
|||||||
2
changedetectionio/static/styles/.gitignore
vendored
2
changedetectionio/static/styles/.gitignore
vendored
@@ -1 +1,3 @@
|
|||||||
node_modules
|
node_modules
|
||||||
|
package-lock.json
|
||||||
|
|
||||||
|
|||||||
3719
changedetectionio/static/styles/package-lock.json
generated
3719
changedetectionio/static/styles/package-lock.json
generated
File diff suppressed because it is too large
Load Diff
26
changedetectionio/static/styles/parts/_arrows.scss
Normal file
26
changedetectionio/static/styles/parts/_arrows.scss
Normal file
@@ -0,0 +1,26 @@
|
|||||||
|
.arrow {
|
||||||
|
border: solid #1b98f8;
|
||||||
|
border-width: 0 2px 2px 0;
|
||||||
|
display: inline-block;
|
||||||
|
padding: 3px;
|
||||||
|
|
||||||
|
&.right {
|
||||||
|
transform: rotate(-45deg);
|
||||||
|
-webkit-transform: rotate(-45deg);
|
||||||
|
}
|
||||||
|
|
||||||
|
&.left {
|
||||||
|
transform: rotate(135deg);
|
||||||
|
-webkit-transform: rotate(135deg);
|
||||||
|
}
|
||||||
|
|
||||||
|
&.up, &.asc {
|
||||||
|
transform: rotate(-135deg);
|
||||||
|
-webkit-transform: rotate(-135deg);
|
||||||
|
}
|
||||||
|
|
||||||
|
&.down, &.desc {
|
||||||
|
transform: rotate(45deg);
|
||||||
|
-webkit-transform: rotate(45deg);
|
||||||
|
}
|
||||||
|
}
|
||||||
@@ -1,11 +1,27 @@
|
|||||||
/*
|
/*
|
||||||
* -- BASE STYLES --
|
* -- BASE STYLES --
|
||||||
* Most of these are inherited from Base, but I want to change a few.
|
* Most of these are inherited from Base, but I want to change a few.
|
||||||
* nvm use v14.18.1
|
* nvm use v14.18.1 && npm install && npm run build
|
||||||
* npm install
|
|
||||||
* npm run build
|
|
||||||
* or npm run watch
|
* or npm run watch
|
||||||
*/
|
*/
|
||||||
|
.arrow {
|
||||||
|
border: solid #1b98f8;
|
||||||
|
border-width: 0 2px 2px 0;
|
||||||
|
display: inline-block;
|
||||||
|
padding: 3px; }
|
||||||
|
.arrow.right {
|
||||||
|
transform: rotate(-45deg);
|
||||||
|
-webkit-transform: rotate(-45deg); }
|
||||||
|
.arrow.left {
|
||||||
|
transform: rotate(135deg);
|
||||||
|
-webkit-transform: rotate(135deg); }
|
||||||
|
.arrow.up, .arrow.asc {
|
||||||
|
transform: rotate(-135deg);
|
||||||
|
-webkit-transform: rotate(-135deg); }
|
||||||
|
.arrow.down, .arrow.desc {
|
||||||
|
transform: rotate(45deg);
|
||||||
|
-webkit-transform: rotate(45deg); }
|
||||||
|
|
||||||
body {
|
body {
|
||||||
color: #333;
|
color: #333;
|
||||||
background: #262626; }
|
background: #262626; }
|
||||||
@@ -55,6 +71,12 @@ code {
|
|||||||
white-space: normal; }
|
white-space: normal; }
|
||||||
.watch-table th {
|
.watch-table th {
|
||||||
white-space: nowrap; }
|
white-space: nowrap; }
|
||||||
|
.watch-table th a {
|
||||||
|
font-weight: normal; }
|
||||||
|
.watch-table th a.active {
|
||||||
|
font-weight: bolder; }
|
||||||
|
.watch-table th a.inactive .arrow {
|
||||||
|
display: none; }
|
||||||
.watch-table .title-col a[target="_blank"]::after, .watch-table .current-diff-url::after {
|
.watch-table .title-col a[target="_blank"]::after, .watch-table .current-diff-url::after {
|
||||||
content: url(data:image/png;base64,iVBORw0KGgoAAAANSUhEUgAAAAoAAAAKCAYAAACNMs+9AAAAQElEQVR42qXKwQkAIAxDUUdxtO6/RBQkQZvSi8I/pL4BoGw/XPkh4XigPmsUgh0626AjRsgxHTkUThsG2T/sIlzdTsp52kSS1wAAAABJRU5ErkJggg==);
|
content: url(data:image/png;base64,iVBORw0KGgoAAAANSUhEUgAAAAoAAAAKCAYAAACNMs+9AAAAQElEQVR42qXKwQkAIAxDUUdxtO6/RBQkQZvSi8I/pL4BoGw/XPkh4XigPmsUgh0626AjRsgxHTkUThsG2T/sIlzdTsp52kSS1wAAAABJRU5ErkJggg==);
|
||||||
margin: 0 3px 0 5px; }
|
margin: 0 3px 0 5px; }
|
||||||
@@ -105,24 +127,6 @@ body:after, body:before {
|
|||||||
-webkit-clip-path: polygon(100% 0, 0 0, 0 77.5%, 1% 77.4%, 2% 77.1%, 3% 76.6%, 4% 75.9%, 5% 75.05%, 6% 74.05%, 7% 72.95%, 8% 71.75%, 9% 70.55%, 10% 69.3%, 11% 68.05%, 12% 66.9%, 13% 65.8%, 14% 64.8%, 15% 64%, 16% 63.35%, 17% 62.85%, 18% 62.6%, 19% 62.5%, 20% 62.65%, 21% 63%, 22% 63.5%, 23% 64.2%, 24% 65.1%, 25% 66.1%, 26% 67.2%, 27% 68.4%, 28% 69.65%, 29% 70.9%, 30% 72.15%, 31% 73.3%, 32% 74.35%, 33% 75.3%, 34% 76.1%, 35% 76.75%, 36% 77.2%, 37% 77.45%, 38% 77.5%, 39% 77.3%, 40% 76.95%, 41% 76.4%, 42% 75.65%, 43% 74.75%, 44% 73.75%, 45% 72.6%, 46% 71.4%, 47% 70.15%, 48% 68.9%, 49% 67.7%, 50% 66.55%, 51% 65.5%, 52% 64.55%, 53% 63.75%, 54% 63.15%, 55% 62.75%, 56% 62.55%, 57% 62.5%, 58% 62.7%, 59% 63.1%, 60% 63.7%, 61% 64.45%, 62% 65.4%, 63% 66.45%, 64% 67.6%, 65% 68.8%, 66% 70.05%, 67% 71.3%, 68% 72.5%, 69% 73.6%, 70% 74.65%, 71% 75.55%, 72% 76.35%, 73% 76.9%, 74% 77.3%, 75% 77.5%, 76% 77.45%, 77% 77.25%, 78% 76.8%, 79% 76.2%, 80% 75.4%, 81% 74.45%, 82% 73.4%, 83% 72.25%, 84% 71.05%, 85% 69.8%, 86% 68.55%, 87% 67.35%, 88% 66.2%, 89% 65.2%, 90% 64.3%, 91% 63.55%, 92% 63%, 93% 62.65%, 94% 62.5%, 95% 62.55%, 96% 62.8%, 97% 63.3%, 98% 63.9%, 99% 64.75%, 100% 65.7%);
|
-webkit-clip-path: polygon(100% 0, 0 0, 0 77.5%, 1% 77.4%, 2% 77.1%, 3% 76.6%, 4% 75.9%, 5% 75.05%, 6% 74.05%, 7% 72.95%, 8% 71.75%, 9% 70.55%, 10% 69.3%, 11% 68.05%, 12% 66.9%, 13% 65.8%, 14% 64.8%, 15% 64%, 16% 63.35%, 17% 62.85%, 18% 62.6%, 19% 62.5%, 20% 62.65%, 21% 63%, 22% 63.5%, 23% 64.2%, 24% 65.1%, 25% 66.1%, 26% 67.2%, 27% 68.4%, 28% 69.65%, 29% 70.9%, 30% 72.15%, 31% 73.3%, 32% 74.35%, 33% 75.3%, 34% 76.1%, 35% 76.75%, 36% 77.2%, 37% 77.45%, 38% 77.5%, 39% 77.3%, 40% 76.95%, 41% 76.4%, 42% 75.65%, 43% 74.75%, 44% 73.75%, 45% 72.6%, 46% 71.4%, 47% 70.15%, 48% 68.9%, 49% 67.7%, 50% 66.55%, 51% 65.5%, 52% 64.55%, 53% 63.75%, 54% 63.15%, 55% 62.75%, 56% 62.55%, 57% 62.5%, 58% 62.7%, 59% 63.1%, 60% 63.7%, 61% 64.45%, 62% 65.4%, 63% 66.45%, 64% 67.6%, 65% 68.8%, 66% 70.05%, 67% 71.3%, 68% 72.5%, 69% 73.6%, 70% 74.65%, 71% 75.55%, 72% 76.35%, 73% 76.9%, 74% 77.3%, 75% 77.5%, 76% 77.45%, 77% 77.25%, 78% 76.8%, 79% 76.2%, 80% 75.4%, 81% 74.45%, 82% 73.4%, 83% 72.25%, 84% 71.05%, 85% 69.8%, 86% 68.55%, 87% 67.35%, 88% 66.2%, 89% 65.2%, 90% 64.3%, 91% 63.55%, 92% 63%, 93% 62.65%, 94% 62.5%, 95% 62.55%, 96% 62.8%, 97% 63.3%, 98% 63.9%, 99% 64.75%, 100% 65.7%);
|
||||||
clip-path: polygon(100% 0, 0 0, 0 77.5%, 1% 77.4%, 2% 77.1%, 3% 76.6%, 4% 75.9%, 5% 75.05%, 6% 74.05%, 7% 72.95%, 8% 71.75%, 9% 70.55%, 10% 69.3%, 11% 68.05%, 12% 66.9%, 13% 65.8%, 14% 64.8%, 15% 64%, 16% 63.35%, 17% 62.85%, 18% 62.6%, 19% 62.5%, 20% 62.65%, 21% 63%, 22% 63.5%, 23% 64.2%, 24% 65.1%, 25% 66.1%, 26% 67.2%, 27% 68.4%, 28% 69.65%, 29% 70.9%, 30% 72.15%, 31% 73.3%, 32% 74.35%, 33% 75.3%, 34% 76.1%, 35% 76.75%, 36% 77.2%, 37% 77.45%, 38% 77.5%, 39% 77.3%, 40% 76.95%, 41% 76.4%, 42% 75.65%, 43% 74.75%, 44% 73.75%, 45% 72.6%, 46% 71.4%, 47% 70.15%, 48% 68.9%, 49% 67.7%, 50% 66.55%, 51% 65.5%, 52% 64.55%, 53% 63.75%, 54% 63.15%, 55% 62.75%, 56% 62.55%, 57% 62.5%, 58% 62.7%, 59% 63.1%, 60% 63.7%, 61% 64.45%, 62% 65.4%, 63% 66.45%, 64% 67.6%, 65% 68.8%, 66% 70.05%, 67% 71.3%, 68% 72.5%, 69% 73.6%, 70% 74.65%, 71% 75.55%, 72% 76.35%, 73% 76.9%, 74% 77.3%, 75% 77.5%, 76% 77.45%, 77% 77.25%, 78% 76.8%, 79% 76.2%, 80% 75.4%, 81% 74.45%, 82% 73.4%, 83% 72.25%, 84% 71.05%, 85% 69.8%, 86% 68.55%, 87% 67.35%, 88% 66.2%, 89% 65.2%, 90% 64.3%, 91% 63.55%, 92% 63%, 93% 62.65%, 94% 62.5%, 95% 62.55%, 96% 62.8%, 97% 63.3%, 98% 63.9%, 99% 64.75%, 100% 65.7%); }
|
clip-path: polygon(100% 0, 0 0, 0 77.5%, 1% 77.4%, 2% 77.1%, 3% 76.6%, 4% 75.9%, 5% 75.05%, 6% 74.05%, 7% 72.95%, 8% 71.75%, 9% 70.55%, 10% 69.3%, 11% 68.05%, 12% 66.9%, 13% 65.8%, 14% 64.8%, 15% 64%, 16% 63.35%, 17% 62.85%, 18% 62.6%, 19% 62.5%, 20% 62.65%, 21% 63%, 22% 63.5%, 23% 64.2%, 24% 65.1%, 25% 66.1%, 26% 67.2%, 27% 68.4%, 28% 69.65%, 29% 70.9%, 30% 72.15%, 31% 73.3%, 32% 74.35%, 33% 75.3%, 34% 76.1%, 35% 76.75%, 36% 77.2%, 37% 77.45%, 38% 77.5%, 39% 77.3%, 40% 76.95%, 41% 76.4%, 42% 75.65%, 43% 74.75%, 44% 73.75%, 45% 72.6%, 46% 71.4%, 47% 70.15%, 48% 68.9%, 49% 67.7%, 50% 66.55%, 51% 65.5%, 52% 64.55%, 53% 63.75%, 54% 63.15%, 55% 62.75%, 56% 62.55%, 57% 62.5%, 58% 62.7%, 59% 63.1%, 60% 63.7%, 61% 64.45%, 62% 65.4%, 63% 66.45%, 64% 67.6%, 65% 68.8%, 66% 70.05%, 67% 71.3%, 68% 72.5%, 69% 73.6%, 70% 74.65%, 71% 75.55%, 72% 76.35%, 73% 76.9%, 74% 77.3%, 75% 77.5%, 76% 77.45%, 77% 77.25%, 78% 76.8%, 79% 76.2%, 80% 75.4%, 81% 74.45%, 82% 73.4%, 83% 72.25%, 84% 71.05%, 85% 69.8%, 86% 68.55%, 87% 67.35%, 88% 66.2%, 89% 65.2%, 90% 64.3%, 91% 63.55%, 92% 63%, 93% 62.65%, 94% 62.5%, 95% 62.55%, 96% 62.8%, 97% 63.3%, 98% 63.9%, 99% 64.75%, 100% 65.7%); }
|
||||||
|
|
||||||
.arrow {
|
|
||||||
border: solid black;
|
|
||||||
border-width: 0 3px 3px 0;
|
|
||||||
display: inline-block;
|
|
||||||
padding: 3px; }
|
|
||||||
.arrow.right {
|
|
||||||
transform: rotate(-45deg);
|
|
||||||
-webkit-transform: rotate(-45deg); }
|
|
||||||
.arrow.left {
|
|
||||||
transform: rotate(135deg);
|
|
||||||
-webkit-transform: rotate(135deg); }
|
|
||||||
.arrow.up {
|
|
||||||
transform: rotate(-135deg);
|
|
||||||
-webkit-transform: rotate(-135deg); }
|
|
||||||
.arrow.down {
|
|
||||||
transform: rotate(45deg);
|
|
||||||
-webkit-transform: rotate(45deg); }
|
|
||||||
|
|
||||||
.button-small {
|
.button-small {
|
||||||
font-size: 85%; }
|
font-size: 85%; }
|
||||||
|
|
||||||
@@ -203,13 +207,18 @@ body:after, body:before {
|
|||||||
border-radius: 10px;
|
border-radius: 10px;
|
||||||
margin-bottom: 1em; }
|
margin-bottom: 1em; }
|
||||||
#new-watch-form input {
|
#new-watch-form input {
|
||||||
width: auto !important;
|
display: inline-block;
|
||||||
display: inline-block; }
|
margin-bottom: 5px; }
|
||||||
#new-watch-form .label {
|
#new-watch-form .label {
|
||||||
display: none; }
|
display: none; }
|
||||||
#new-watch-form legend {
|
#new-watch-form legend {
|
||||||
color: #fff;
|
color: #fff;
|
||||||
font-weight: bold; }
|
font-weight: bold; }
|
||||||
|
#new-watch-form #watch-add-wrapper-zone > div {
|
||||||
|
display: inline-block; }
|
||||||
|
@media only screen and (max-width: 760px) {
|
||||||
|
#new-watch-form #watch-add-wrapper-zone #url {
|
||||||
|
width: 100%; } }
|
||||||
|
|
||||||
#diff-col {
|
#diff-col {
|
||||||
padding-left: 40px; }
|
padding-left: 40px; }
|
||||||
@@ -268,10 +277,14 @@ footer {
|
|||||||
#new-version-text a {
|
#new-version-text a {
|
||||||
color: #e07171; }
|
color: #e07171; }
|
||||||
|
|
||||||
.paused-state.state-False img {
|
.watch-controls {
|
||||||
|
/* default */ }
|
||||||
|
.watch-controls .state-on img {
|
||||||
|
opacity: 0.8; }
|
||||||
|
.watch-controls img {
|
||||||
opacity: 0.2; }
|
opacity: 0.2; }
|
||||||
|
.watch-controls img:hover {
|
||||||
.paused-state.state-False:hover img {
|
transition: opacity 0.3s;
|
||||||
opacity: 0.8; }
|
opacity: 0.8; }
|
||||||
|
|
||||||
.monospaced-textarea textarea {
|
.monospaced-textarea textarea {
|
||||||
@@ -284,6 +297,11 @@ footer {
|
|||||||
.pure-form {
|
.pure-form {
|
||||||
/* The input fields with errors */
|
/* The input fields with errors */
|
||||||
/* The list of errors */ }
|
/* The list of errors */ }
|
||||||
|
.pure-form fieldset {
|
||||||
|
padding-top: 0px; }
|
||||||
|
.pure-form fieldset ul {
|
||||||
|
padding-bottom: 0px;
|
||||||
|
margin-bottom: 0px; }
|
||||||
.pure-form .pure-control-group, .pure-form .pure-group, .pure-form .pure-controls {
|
.pure-form .pure-control-group, .pure-form .pure-group, .pure-form .pure-controls {
|
||||||
padding-bottom: 1em; }
|
padding-bottom: 1em; }
|
||||||
.pure-form .pure-control-group div, .pure-form .pure-group div, .pure-form .pure-controls div {
|
.pure-form .pure-control-group div, .pure-form .pure-group div, .pure-form .pure-controls div {
|
||||||
@@ -333,7 +351,8 @@ footer {
|
|||||||
padding-top: 110px; }
|
padding-top: 110px; }
|
||||||
div.tabs.collapsable ul li {
|
div.tabs.collapsable ul li {
|
||||||
display: block;
|
display: block;
|
||||||
border-radius: 0px; }
|
border-radius: 0px;
|
||||||
|
margin-right: 0px; }
|
||||||
input[type='text'] {
|
input[type='text'] {
|
||||||
width: 100%; }
|
width: 100%; }
|
||||||
/*
|
/*
|
||||||
@@ -347,6 +366,8 @@ and also iPads specifically.
|
|||||||
/* Hide table headers (but not display: none;, for accessibility) */ }
|
/* Hide table headers (but not display: none;, for accessibility) */ }
|
||||||
.watch-table thead, .watch-table tbody, .watch-table th, .watch-table td, .watch-table tr {
|
.watch-table thead, .watch-table tbody, .watch-table th, .watch-table td, .watch-table tr {
|
||||||
display: block; }
|
display: block; }
|
||||||
|
.watch-table .last-checked > span {
|
||||||
|
vertical-align: middle; }
|
||||||
.watch-table .last-checked::before {
|
.watch-table .last-checked::before {
|
||||||
color: #555;
|
color: #555;
|
||||||
content: "Last Checked "; }
|
content: "Last Checked "; }
|
||||||
@@ -364,7 +385,8 @@ and also iPads specifically.
|
|||||||
.watch-table td {
|
.watch-table td {
|
||||||
/* Behave like a "row" */
|
/* Behave like a "row" */
|
||||||
border: none;
|
border: none;
|
||||||
border-bottom: 1px solid #eee; }
|
border-bottom: 1px solid #eee;
|
||||||
|
vertical-align: middle; }
|
||||||
.watch-table td:before {
|
.watch-table td:before {
|
||||||
/* Top/left values mimic padding */
|
/* Top/left values mimic padding */
|
||||||
top: 6px;
|
top: 6px;
|
||||||
@@ -424,6 +446,15 @@ and also iPads specifically.
|
|||||||
.tab-pane-inner:target {
|
.tab-pane-inner:target {
|
||||||
display: block; }
|
display: block; }
|
||||||
|
|
||||||
|
#beta-logo {
|
||||||
|
height: 50px;
|
||||||
|
right: -3px;
|
||||||
|
top: -3px;
|
||||||
|
position: absolute; }
|
||||||
|
|
||||||
|
#selector-header {
|
||||||
|
padding-bottom: 1em; }
|
||||||
|
|
||||||
.edit-form {
|
.edit-form {
|
||||||
min-width: 70%;
|
min-width: 70%;
|
||||||
/* so it cant overflow */
|
/* so it cant overflow */
|
||||||
@@ -447,4 +478,103 @@ ul {
|
|||||||
.time-check-widget tr {
|
.time-check-widget tr {
|
||||||
display: inline; }
|
display: inline; }
|
||||||
.time-check-widget tr input[type="number"] {
|
.time-check-widget tr input[type="number"] {
|
||||||
width: 4em; }
|
width: 5em; }
|
||||||
|
|
||||||
|
#selector-wrapper {
|
||||||
|
height: 600px;
|
||||||
|
overflow-y: scroll;
|
||||||
|
position: relative; }
|
||||||
|
#selector-wrapper > img {
|
||||||
|
position: absolute;
|
||||||
|
z-index: 4;
|
||||||
|
max-width: 100%; }
|
||||||
|
#selector-wrapper > canvas {
|
||||||
|
position: relative;
|
||||||
|
z-index: 5;
|
||||||
|
max-width: 100%; }
|
||||||
|
#selector-wrapper > canvas:hover {
|
||||||
|
cursor: pointer; }
|
||||||
|
|
||||||
|
#selector-current-xpath {
|
||||||
|
font-size: 80%; }
|
||||||
|
|
||||||
|
#webdriver-override-options input[type="number"] {
|
||||||
|
width: 5em; }
|
||||||
|
|
||||||
|
#api-key:hover {
|
||||||
|
cursor: pointer; }
|
||||||
|
|
||||||
|
#api-key-copy {
|
||||||
|
color: #0078e7; }
|
||||||
|
|
||||||
|
/* spinner */
|
||||||
|
.loader,
|
||||||
|
.loader:after {
|
||||||
|
border-radius: 50%;
|
||||||
|
width: 10px;
|
||||||
|
height: 10px; }
|
||||||
|
|
||||||
|
.loader {
|
||||||
|
margin: 0px auto;
|
||||||
|
font-size: 3px;
|
||||||
|
vertical-align: middle;
|
||||||
|
display: inline-block;
|
||||||
|
text-indent: -9999em;
|
||||||
|
border-top: 1.1em solid rgba(38, 104, 237, 0.2);
|
||||||
|
border-right: 1.1em solid rgba(38, 104, 237, 0.2);
|
||||||
|
border-bottom: 1.1em solid rgba(38, 104, 237, 0.2);
|
||||||
|
border-left: 1.1em solid #2668ed;
|
||||||
|
-webkit-transform: translateZ(0);
|
||||||
|
-ms-transform: translateZ(0);
|
||||||
|
transform: translateZ(0);
|
||||||
|
-webkit-animation: load8 1.1s infinite linear;
|
||||||
|
animation: load8 1.1s infinite linear; }
|
||||||
|
|
||||||
|
@-webkit-keyframes load8 {
|
||||||
|
0% {
|
||||||
|
-webkit-transform: rotate(0deg);
|
||||||
|
transform: rotate(0deg); }
|
||||||
|
100% {
|
||||||
|
-webkit-transform: rotate(360deg);
|
||||||
|
transform: rotate(360deg); } }
|
||||||
|
|
||||||
|
@keyframes load8 {
|
||||||
|
0% {
|
||||||
|
-webkit-transform: rotate(0deg);
|
||||||
|
transform: rotate(0deg); }
|
||||||
|
100% {
|
||||||
|
-webkit-transform: rotate(360deg);
|
||||||
|
transform: rotate(360deg); } }
|
||||||
|
|
||||||
|
.snapshot-age {
|
||||||
|
padding: 4px;
|
||||||
|
background-color: #dfdfdf;
|
||||||
|
border-radius: 3px;
|
||||||
|
font-weight: bold;
|
||||||
|
margin-bottom: 4px; }
|
||||||
|
.snapshot-age.error {
|
||||||
|
background-color: #ff0000;
|
||||||
|
color: #fff; }
|
||||||
|
|
||||||
|
#checkbox-operations {
|
||||||
|
background: rgba(0, 0, 0, 0.05);
|
||||||
|
padding: 1em;
|
||||||
|
border-radius: 10px;
|
||||||
|
margin-bottom: 1em;
|
||||||
|
display: none; }
|
||||||
|
|
||||||
|
.checkbox-uuid > * {
|
||||||
|
vertical-align: middle; }
|
||||||
|
|
||||||
|
.inline-warning {
|
||||||
|
border: 1px solid #ff3300;
|
||||||
|
padding: 0.5rem;
|
||||||
|
border-radius: 5px;
|
||||||
|
color: #ff3300; }
|
||||||
|
.inline-warning > span {
|
||||||
|
display: inline-block;
|
||||||
|
vertical-align: middle; }
|
||||||
|
.inline-warning img.inline-warning-icon {
|
||||||
|
display: inline;
|
||||||
|
height: 26px;
|
||||||
|
vertical-align: middle; }
|
||||||
|
|||||||
@@ -1,11 +1,11 @@
|
|||||||
/*
|
/*
|
||||||
* -- BASE STYLES --
|
* -- BASE STYLES --
|
||||||
* Most of these are inherited from Base, but I want to change a few.
|
* Most of these are inherited from Base, but I want to change a few.
|
||||||
* nvm use v14.18.1
|
* nvm use v14.18.1 && npm install && npm run build
|
||||||
* npm install
|
|
||||||
* npm run build
|
|
||||||
* or npm run watch
|
* or npm run watch
|
||||||
*/
|
*/
|
||||||
|
@import "parts/_arrows.scss";
|
||||||
|
|
||||||
body {
|
body {
|
||||||
color: #333;
|
color: #333;
|
||||||
background: #262626;
|
background: #262626;
|
||||||
@@ -70,6 +70,17 @@ code {
|
|||||||
|
|
||||||
th {
|
th {
|
||||||
white-space: nowrap;
|
white-space: nowrap;
|
||||||
|
a {
|
||||||
|
font-weight: normal;
|
||||||
|
&.active {
|
||||||
|
font-weight: bolder;
|
||||||
|
}
|
||||||
|
&.inactive {
|
||||||
|
.arrow {
|
||||||
|
display: none;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
.title-col a[target="_blank"]::after, .current-diff-url::after {
|
.title-col a[target="_blank"]::after, .current-diff-url::after {
|
||||||
@@ -139,36 +150,13 @@ body:after, body:before {
|
|||||||
clip-path: polygon(100% 0, 0 0, 0 77.5%, 1% 77.4%, 2% 77.1%, 3% 76.6%, 4% 75.9%, 5% 75.05%, 6% 74.05%, 7% 72.95%, 8% 71.75%, 9% 70.55%, 10% 69.3%, 11% 68.05%, 12% 66.9%, 13% 65.8%, 14% 64.8%, 15% 64%, 16% 63.35%, 17% 62.85%, 18% 62.6%, 19% 62.5%, 20% 62.65%, 21% 63%, 22% 63.5%, 23% 64.2%, 24% 65.1%, 25% 66.1%, 26% 67.2%, 27% 68.4%, 28% 69.65%, 29% 70.9%, 30% 72.15%, 31% 73.3%, 32% 74.35%, 33% 75.3%, 34% 76.1%, 35% 76.75%, 36% 77.2%, 37% 77.45%, 38% 77.5%, 39% 77.3%, 40% 76.95%, 41% 76.4%, 42% 75.65%, 43% 74.75%, 44% 73.75%, 45% 72.6%, 46% 71.4%, 47% 70.15%, 48% 68.9%, 49% 67.7%, 50% 66.55%, 51% 65.5%, 52% 64.55%, 53% 63.75%, 54% 63.15%, 55% 62.75%, 56% 62.55%, 57% 62.5%, 58% 62.7%, 59% 63.1%, 60% 63.7%, 61% 64.45%, 62% 65.4%, 63% 66.45%, 64% 67.6%, 65% 68.8%, 66% 70.05%, 67% 71.3%, 68% 72.5%, 69% 73.6%, 70% 74.65%, 71% 75.55%, 72% 76.35%, 73% 76.9%, 74% 77.3%, 75% 77.5%, 76% 77.45%, 77% 77.25%, 78% 76.8%, 79% 76.2%, 80% 75.4%, 81% 74.45%, 82% 73.4%, 83% 72.25%, 84% 71.05%, 85% 69.8%, 86% 68.55%, 87% 67.35%, 88% 66.2%, 89% 65.2%, 90% 64.3%, 91% 63.55%, 92% 63%, 93% 62.65%, 94% 62.5%, 95% 62.55%, 96% 62.8%, 97% 63.3%, 98% 63.9%, 99% 64.75%, 100% 65.7%)
|
clip-path: polygon(100% 0, 0 0, 0 77.5%, 1% 77.4%, 2% 77.1%, 3% 76.6%, 4% 75.9%, 5% 75.05%, 6% 74.05%, 7% 72.95%, 8% 71.75%, 9% 70.55%, 10% 69.3%, 11% 68.05%, 12% 66.9%, 13% 65.8%, 14% 64.8%, 15% 64%, 16% 63.35%, 17% 62.85%, 18% 62.6%, 19% 62.5%, 20% 62.65%, 21% 63%, 22% 63.5%, 23% 64.2%, 24% 65.1%, 25% 66.1%, 26% 67.2%, 27% 68.4%, 28% 69.65%, 29% 70.9%, 30% 72.15%, 31% 73.3%, 32% 74.35%, 33% 75.3%, 34% 76.1%, 35% 76.75%, 36% 77.2%, 37% 77.45%, 38% 77.5%, 39% 77.3%, 40% 76.95%, 41% 76.4%, 42% 75.65%, 43% 74.75%, 44% 73.75%, 45% 72.6%, 46% 71.4%, 47% 70.15%, 48% 68.9%, 49% 67.7%, 50% 66.55%, 51% 65.5%, 52% 64.55%, 53% 63.75%, 54% 63.15%, 55% 62.75%, 56% 62.55%, 57% 62.5%, 58% 62.7%, 59% 63.1%, 60% 63.7%, 61% 64.45%, 62% 65.4%, 63% 66.45%, 64% 67.6%, 65% 68.8%, 66% 70.05%, 67% 71.3%, 68% 72.5%, 69% 73.6%, 70% 74.65%, 71% 75.55%, 72% 76.35%, 73% 76.9%, 74% 77.3%, 75% 77.5%, 76% 77.45%, 77% 77.25%, 78% 76.8%, 79% 76.2%, 80% 75.4%, 81% 74.45%, 82% 73.4%, 83% 72.25%, 84% 71.05%, 85% 69.8%, 86% 68.55%, 87% 67.35%, 88% 66.2%, 89% 65.2%, 90% 64.3%, 91% 63.55%, 92% 63%, 93% 62.65%, 94% 62.5%, 95% 62.55%, 96% 62.8%, 97% 63.3%, 98% 63.9%, 99% 64.75%, 100% 65.7%)
|
||||||
}
|
}
|
||||||
|
|
||||||
.arrow {
|
|
||||||
border: solid black;
|
|
||||||
border-width: 0 3px 3px 0;
|
|
||||||
display: inline-block;
|
|
||||||
padding: 3px;
|
|
||||||
&.right {
|
|
||||||
transform: rotate(-45deg);
|
|
||||||
-webkit-transform: rotate(-45deg);
|
|
||||||
}
|
|
||||||
&.left {
|
|
||||||
transform: rotate(135deg);
|
|
||||||
-webkit-transform: rotate(135deg);
|
|
||||||
}
|
|
||||||
&.up {
|
|
||||||
transform: rotate(-135deg);
|
|
||||||
-webkit-transform: rotate(-135deg);
|
|
||||||
}
|
|
||||||
&.down {
|
|
||||||
transform: rotate(45deg);
|
|
||||||
-webkit-transform: rotate(45deg);
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
.button-small {
|
.button-small {
|
||||||
font-size: 85%;
|
font-size: 85%;
|
||||||
}
|
}
|
||||||
|
|
||||||
.fetch-error {
|
.fetch-error {
|
||||||
padding-top: 1em;
|
padding-top: 1em;
|
||||||
font-size: 60%;
|
font-size: 80%;
|
||||||
max-width: 400px;
|
max-width: 400px;
|
||||||
display: block;
|
display: block;
|
||||||
}
|
}
|
||||||
@@ -269,8 +257,8 @@ body:after, body:before {
|
|||||||
border-radius: 10px;
|
border-radius: 10px;
|
||||||
margin-bottom: 1em;
|
margin-bottom: 1em;
|
||||||
input {
|
input {
|
||||||
width: auto !important;
|
|
||||||
display: inline-block;
|
display: inline-block;
|
||||||
|
margin-bottom: 5px;
|
||||||
}
|
}
|
||||||
.label {
|
.label {
|
||||||
display: none;
|
display: none;
|
||||||
@@ -279,6 +267,17 @@ body:after, body:before {
|
|||||||
color: #fff;
|
color: #fff;
|
||||||
font-weight: bold;
|
font-weight: bold;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
#watch-add-wrapper-zone {
|
||||||
|
> div {
|
||||||
|
display: inline-block;
|
||||||
|
}
|
||||||
|
@media only screen and (max-width: 760px) {
|
||||||
|
#url {
|
||||||
|
width: 100%;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
||||||
@@ -353,14 +352,25 @@ footer {
|
|||||||
color: #e07171;
|
color: #e07171;
|
||||||
}
|
}
|
||||||
|
|
||||||
.paused-state {
|
.watch-controls {
|
||||||
&.state-False img {
|
.state-on {
|
||||||
|
img {
|
||||||
|
opacity: 0.8;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
/* default */
|
||||||
|
img {
|
||||||
opacity: 0.2;
|
opacity: 0.2;
|
||||||
}
|
}
|
||||||
|
|
||||||
&.state-False:hover img {
|
img {
|
||||||
|
&:hover {
|
||||||
|
transition: opacity 0.3s;
|
||||||
opacity: 0.8;
|
opacity: 0.8;
|
||||||
}
|
}
|
||||||
|
}
|
||||||
|
|
||||||
}
|
}
|
||||||
|
|
||||||
.monospaced-textarea {
|
.monospaced-textarea {
|
||||||
@@ -375,6 +385,13 @@ footer {
|
|||||||
|
|
||||||
|
|
||||||
.pure-form {
|
.pure-form {
|
||||||
|
fieldset {
|
||||||
|
padding-top: 0px;
|
||||||
|
ul {
|
||||||
|
padding-bottom: 0px;
|
||||||
|
margin-bottom: 0px;
|
||||||
|
}
|
||||||
|
}
|
||||||
.pure-control-group, .pure-group, .pure-controls {
|
.pure-control-group, .pure-group, .pure-controls {
|
||||||
padding-bottom: 1em;
|
padding-bottom: 1em;
|
||||||
div {
|
div {
|
||||||
@@ -462,6 +479,7 @@ footer {
|
|||||||
div.tabs.collapsable ul li {
|
div.tabs.collapsable ul li {
|
||||||
display: block;
|
display: block;
|
||||||
border-radius: 0px;
|
border-radius: 0px;
|
||||||
|
margin-right: 0px;
|
||||||
}
|
}
|
||||||
|
|
||||||
input[type='text'] {
|
input[type='text'] {
|
||||||
@@ -479,6 +497,12 @@ and also iPads specifically.
|
|||||||
display: block;
|
display: block;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
.last-checked {
|
||||||
|
> span {
|
||||||
|
vertical-align: middle;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
.last-checked::before {
|
.last-checked::before {
|
||||||
color: #555;
|
color: #555;
|
||||||
content: "Last Checked ";
|
content: "Last Checked ";
|
||||||
@@ -509,7 +533,7 @@ and also iPads specifically.
|
|||||||
/* Behave like a "row" */
|
/* Behave like a "row" */
|
||||||
border: none;
|
border: none;
|
||||||
border-bottom: 1px solid #eee;
|
border-bottom: 1px solid #eee;
|
||||||
|
vertical-align: middle;
|
||||||
&:before {
|
&:before {
|
||||||
/* Top/left values mimic padding */
|
/* Top/left values mimic padding */
|
||||||
top: 6px;
|
top: 6px;
|
||||||
@@ -606,6 +630,18 @@ $form-edge-padding: 20px;
|
|||||||
padding: 0px;
|
padding: 0px;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
#beta-logo {
|
||||||
|
height: 50px;
|
||||||
|
// looks better when it's hanging off a little
|
||||||
|
right: -3px;
|
||||||
|
top: -3px;
|
||||||
|
position: absolute;
|
||||||
|
}
|
||||||
|
|
||||||
|
#selector-header {
|
||||||
|
padding-bottom: 1em;
|
||||||
|
}
|
||||||
|
|
||||||
.edit-form {
|
.edit-form {
|
||||||
min-width: 70%;
|
min-width: 70%;
|
||||||
/* so it cant overflow */
|
/* so it cant overflow */
|
||||||
@@ -637,7 +673,134 @@ ul {
|
|||||||
tr {
|
tr {
|
||||||
display: inline;
|
display: inline;
|
||||||
input[type="number"] {
|
input[type="number"] {
|
||||||
width: 4em;
|
width: 5em;
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
#selector-wrapper {
|
||||||
|
height: 600px;
|
||||||
|
overflow-y: scroll;
|
||||||
|
position: relative;
|
||||||
|
//width: 100%;
|
||||||
|
> img {
|
||||||
|
position: absolute;
|
||||||
|
z-index: 4;
|
||||||
|
max-width: 100%;
|
||||||
|
}
|
||||||
|
>canvas {
|
||||||
|
position: relative;
|
||||||
|
z-index: 5;
|
||||||
|
max-width: 100%;
|
||||||
|
&:hover {
|
||||||
|
cursor: pointer;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
#selector-current-xpath {
|
||||||
|
font-size: 80%;
|
||||||
|
}
|
||||||
|
|
||||||
|
#webdriver-override-options {
|
||||||
|
input[type="number"] {
|
||||||
|
width: 5em;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
#api-key {
|
||||||
|
&:hover {
|
||||||
|
cursor: pointer;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
#api-key-copy {
|
||||||
|
color: #0078e7;
|
||||||
|
}
|
||||||
|
|
||||||
|
/* spinner */
|
||||||
|
.loader,
|
||||||
|
.loader:after {
|
||||||
|
border-radius: 50%;
|
||||||
|
width: 10px;
|
||||||
|
height: 10px;
|
||||||
|
}
|
||||||
|
.loader {
|
||||||
|
margin: 0px auto;
|
||||||
|
font-size: 3px;
|
||||||
|
vertical-align: middle;
|
||||||
|
display: inline-block;
|
||||||
|
text-indent: -9999em;
|
||||||
|
border-top: 1.1em solid rgba(38,104,237, 0.2);
|
||||||
|
border-right: 1.1em solid rgba(38,104,237, 0.2);
|
||||||
|
border-bottom: 1.1em solid rgba(38,104,237, 0.2);
|
||||||
|
border-left: 1.1em solid #2668ed;
|
||||||
|
-webkit-transform: translateZ(0);
|
||||||
|
-ms-transform: translateZ(0);
|
||||||
|
transform: translateZ(0);
|
||||||
|
-webkit-animation: load8 1.1s infinite linear;
|
||||||
|
animation: load8 1.1s infinite linear;
|
||||||
|
}
|
||||||
|
@-webkit-keyframes load8 {
|
||||||
|
0% {
|
||||||
|
-webkit-transform: rotate(0deg);
|
||||||
|
transform: rotate(0deg);
|
||||||
|
}
|
||||||
|
100% {
|
||||||
|
-webkit-transform: rotate(360deg);
|
||||||
|
transform: rotate(360deg);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
@keyframes load8 {
|
||||||
|
0% {
|
||||||
|
-webkit-transform: rotate(0deg);
|
||||||
|
transform: rotate(0deg);
|
||||||
|
}
|
||||||
|
100% {
|
||||||
|
-webkit-transform: rotate(360deg);
|
||||||
|
transform: rotate(360deg);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
.snapshot-age {
|
||||||
|
padding: 4px;
|
||||||
|
background-color: #dfdfdf;
|
||||||
|
border-radius: 3px;
|
||||||
|
font-weight: bold;
|
||||||
|
margin-bottom: 4px;
|
||||||
|
&.error {
|
||||||
|
background-color: #ff0000;
|
||||||
|
color: #fff;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
#checkbox-operations {
|
||||||
|
background: rgba(0, 0, 0, 0.05);
|
||||||
|
padding: 1em;
|
||||||
|
border-radius: 10px;
|
||||||
|
margin-bottom: 1em;
|
||||||
|
display: none;
|
||||||
|
}
|
||||||
|
.checkbox-uuid {
|
||||||
|
> * {
|
||||||
|
vertical-align: middle;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
.inline-warning {
|
||||||
|
> span {
|
||||||
|
display: inline-block;
|
||||||
|
vertical-align: middle;
|
||||||
|
}
|
||||||
|
|
||||||
|
img.inline-warning-icon {
|
||||||
|
display: inline;
|
||||||
|
height: 26px;
|
||||||
|
vertical-align: middle;
|
||||||
|
}
|
||||||
|
|
||||||
|
border: 1px solid #ff3300;
|
||||||
|
padding: 0.5rem;
|
||||||
|
border-radius: 5px;
|
||||||
|
color: #ff3300;
|
||||||
|
}
|
||||||
|
|||||||
@@ -8,10 +8,11 @@ import threading
|
|||||||
import time
|
import time
|
||||||
import uuid as uuid_builder
|
import uuid as uuid_builder
|
||||||
from copy import deepcopy
|
from copy import deepcopy
|
||||||
from os import mkdir, path, unlink
|
from os import path, unlink
|
||||||
from threading import Lock
|
from threading import Lock
|
||||||
import re
|
import re
|
||||||
import requests
|
import requests
|
||||||
|
import secrets
|
||||||
|
|
||||||
from . model import App, Watch
|
from . model import App, Watch
|
||||||
|
|
||||||
@@ -26,20 +27,21 @@ class ChangeDetectionStore:
|
|||||||
# For when we edit, we should write to disk
|
# For when we edit, we should write to disk
|
||||||
needs_write_urgent = False
|
needs_write_urgent = False
|
||||||
|
|
||||||
|
__version_check = True
|
||||||
|
|
||||||
def __init__(self, datastore_path="/datastore", include_default_watches=True, version_tag="0.0.0"):
|
def __init__(self, datastore_path="/datastore", include_default_watches=True, version_tag="0.0.0"):
|
||||||
# Should only be active for docker
|
# Should only be active for docker
|
||||||
# logging.basicConfig(filename='/dev/stdout', level=logging.INFO)
|
# logging.basicConfig(filename='/dev/stdout', level=logging.INFO)
|
||||||
self.needs_write = False
|
self.__data = App.model()
|
||||||
self.datastore_path = datastore_path
|
self.datastore_path = datastore_path
|
||||||
self.json_store_path = "{}/url-watches.json".format(self.datastore_path)
|
self.json_store_path = "{}/url-watches.json".format(self.datastore_path)
|
||||||
|
self.needs_write = False
|
||||||
self.proxy_list = None
|
self.proxy_list = None
|
||||||
|
self.start_time = time.time()
|
||||||
self.stop_thread = False
|
self.stop_thread = False
|
||||||
|
|
||||||
self.__data = App.model()
|
|
||||||
|
|
||||||
# Base definition for all watchers
|
# Base definition for all watchers
|
||||||
# deepcopy part of #569 - not sure why its needed exactly
|
# deepcopy part of #569 - not sure why its needed exactly
|
||||||
self.generic_definition = deepcopy(Watch.model())
|
self.generic_definition = deepcopy(Watch.model(datastore_path = datastore_path, default={}))
|
||||||
|
|
||||||
if path.isfile('changedetectionio/source.txt'):
|
if path.isfile('changedetectionio/source.txt'):
|
||||||
with open('changedetectionio/source.txt') as f:
|
with open('changedetectionio/source.txt') as f:
|
||||||
@@ -70,24 +72,23 @@ class ChangeDetectionStore:
|
|||||||
if 'application' in from_disk['settings']:
|
if 'application' in from_disk['settings']:
|
||||||
self.__data['settings']['application'].update(from_disk['settings']['application'])
|
self.__data['settings']['application'].update(from_disk['settings']['application'])
|
||||||
|
|
||||||
# Reinitialise each `watching` with our generic_definition in the case that we add a new var in the future.
|
# Convert each existing watch back to the Watch.model object
|
||||||
# @todo pretty sure theres a python we todo this with an abstracted(?) object!
|
|
||||||
for uuid, watch in self.__data['watching'].items():
|
for uuid, watch in self.__data['watching'].items():
|
||||||
_blank = deepcopy(self.generic_definition)
|
watch['uuid']=uuid
|
||||||
_blank.update(watch)
|
self.__data['watching'][uuid] = Watch.model(datastore_path=self.datastore_path, default=watch)
|
||||||
self.__data['watching'].update({uuid: _blank})
|
|
||||||
self.__data['watching'][uuid]['newest_history_key'] = self.get_newest_history_key(uuid)
|
|
||||||
print("Watching:", uuid, self.__data['watching'][uuid]['url'])
|
print("Watching:", uuid, self.__data['watching'][uuid]['url'])
|
||||||
|
|
||||||
# First time ran, doesnt exist.
|
# First time ran, doesnt exist.
|
||||||
except (FileNotFoundError, json.decoder.JSONDecodeError):
|
except (FileNotFoundError, json.decoder.JSONDecodeError):
|
||||||
if include_default_watches:
|
if include_default_watches:
|
||||||
print("Creating JSON store at", self.datastore_path)
|
print("Creating JSON store at", self.datastore_path)
|
||||||
|
self.add_watch(url='https://news.ycombinator.com/',
|
||||||
|
tag='Tech news',
|
||||||
|
extras={'fetch_backend': 'html_requests'})
|
||||||
|
|
||||||
self.add_watch(url='http://www.quotationspage.com/random.php', tag='test')
|
self.add_watch(url='https://changedetection.io/CHANGELOG.txt',
|
||||||
self.add_watch(url='https://news.ycombinator.com/', tag='Tech news')
|
tag='changedetection.io',
|
||||||
self.add_watch(url='https://www.gov.uk/coronavirus', tag='Covid')
|
extras={'fetch_backend': 'html_requests'})
|
||||||
self.add_watch(url='https://changedetection.io/CHANGELOG.txt')
|
|
||||||
|
|
||||||
self.__data['version_tag'] = version_tag
|
self.__data['version_tag'] = version_tag
|
||||||
|
|
||||||
@@ -107,15 +108,16 @@ class ChangeDetectionStore:
|
|||||||
|
|
||||||
# Generate the URL access token for RSS feeds
|
# Generate the URL access token for RSS feeds
|
||||||
if not 'rss_access_token' in self.__data['settings']['application']:
|
if not 'rss_access_token' in self.__data['settings']['application']:
|
||||||
import secrets
|
|
||||||
secret = secrets.token_hex(16)
|
secret = secrets.token_hex(16)
|
||||||
self.__data['settings']['application']['rss_access_token'] = secret
|
self.__data['settings']['application']['rss_access_token'] = secret
|
||||||
|
|
||||||
|
# Generate the API access token
|
||||||
|
if not 'api_access_token' in self.__data['settings']['application']:
|
||||||
|
secret = secrets.token_hex(16)
|
||||||
|
self.__data['settings']['application']['api_access_token'] = secret
|
||||||
|
|
||||||
# Proxy list support - available as a selection in settings when text file is imported
|
# Proxy list support - available as a selection in settings when text file is imported
|
||||||
# CSV list
|
proxy_list_file = "{}/proxies.json".format(self.datastore_path)
|
||||||
# "name, address", or just "name"
|
|
||||||
proxy_list_file = "{}/proxies.txt".format(self.datastore_path)
|
|
||||||
if path.isfile(proxy_list_file):
|
if path.isfile(proxy_list_file):
|
||||||
self.import_proxy_list(proxy_list_file)
|
self.import_proxy_list(proxy_list_file)
|
||||||
|
|
||||||
@@ -127,23 +129,8 @@ class ChangeDetectionStore:
|
|||||||
# Finally start the thread that will manage periodic data saves to JSON
|
# Finally start the thread that will manage periodic data saves to JSON
|
||||||
save_data_thread = threading.Thread(target=self.save_datastore).start()
|
save_data_thread = threading.Thread(target=self.save_datastore).start()
|
||||||
|
|
||||||
# Returns the newest key, but if theres only 1 record, then it's counted as not being new, so return 0.
|
|
||||||
def get_newest_history_key(self, uuid):
|
|
||||||
if len(self.__data['watching'][uuid]['history']) == 1:
|
|
||||||
return 0
|
|
||||||
|
|
||||||
dates = list(self.__data['watching'][uuid]['history'].keys())
|
|
||||||
# Convert to int, sort and back to str again
|
|
||||||
# @todo replace datastore getter that does this automatically
|
|
||||||
dates = [int(i) for i in dates]
|
|
||||||
dates.sort(reverse=True)
|
|
||||||
if len(dates):
|
|
||||||
# always keyed as str
|
|
||||||
return str(dates[0])
|
|
||||||
|
|
||||||
return 0
|
|
||||||
|
|
||||||
def set_last_viewed(self, uuid, timestamp):
|
def set_last_viewed(self, uuid, timestamp):
|
||||||
|
logging.debug("Setting watch UUID: {} last viewed to {}".format(uuid, int(timestamp)))
|
||||||
self.data['watching'][uuid].update({'last_viewed': int(timestamp)})
|
self.data['watching'][uuid].update({'last_viewed': int(timestamp)})
|
||||||
self.needs_write = True
|
self.needs_write = True
|
||||||
|
|
||||||
@@ -167,35 +154,32 @@ class ChangeDetectionStore:
|
|||||||
del (update_obj[dict_key])
|
del (update_obj[dict_key])
|
||||||
|
|
||||||
self.__data['watching'][uuid].update(update_obj)
|
self.__data['watching'][uuid].update(update_obj)
|
||||||
self.__data['watching'][uuid]['newest_history_key'] = self.get_newest_history_key(uuid)
|
|
||||||
|
|
||||||
self.needs_write = True
|
self.needs_write = True
|
||||||
|
|
||||||
@property
|
@property
|
||||||
def threshold_seconds(self):
|
def threshold_seconds(self):
|
||||||
seconds = 0
|
seconds = 0
|
||||||
mtable = {'seconds': 1, 'minutes': 60, 'hours': 3600, 'days': 86400, 'weeks': 86400 * 7}
|
for m, n in Watch.mtable.items():
|
||||||
minimum_seconds_recheck_time = int(os.getenv('MINIMUM_SECONDS_RECHECK_TIME', 60))
|
|
||||||
for m, n in mtable.items():
|
|
||||||
x = self.__data['settings']['requests']['time_between_check'].get(m)
|
x = self.__data['settings']['requests']['time_between_check'].get(m)
|
||||||
if x:
|
if x:
|
||||||
seconds += x * n
|
seconds += x * n
|
||||||
return max(seconds, minimum_seconds_recheck_time)
|
return seconds
|
||||||
|
|
||||||
|
@property
|
||||||
|
def has_unviewed(self):
|
||||||
|
for uuid, watch in self.__data['watching'].items():
|
||||||
|
if watch.viewed == False:
|
||||||
|
return True
|
||||||
|
return False
|
||||||
|
|
||||||
@property
|
@property
|
||||||
def data(self):
|
def data(self):
|
||||||
has_unviewed = False
|
has_unviewed = False
|
||||||
for uuid, v in self.__data['watching'].items():
|
for uuid, watch in self.__data['watching'].items():
|
||||||
self.__data['watching'][uuid]['newest_history_key'] = self.get_newest_history_key(uuid)
|
|
||||||
if int(v['newest_history_key']) <= int(v['last_viewed']):
|
|
||||||
self.__data['watching'][uuid]['viewed'] = True
|
|
||||||
|
|
||||||
else:
|
|
||||||
self.__data['watching'][uuid]['viewed'] = False
|
|
||||||
has_unviewed = True
|
|
||||||
|
|
||||||
# #106 - Be sure this is None on empty string, False, None, etc
|
# #106 - Be sure this is None on empty string, False, None, etc
|
||||||
# Default var for fetch_backend
|
# Default var for fetch_backend
|
||||||
|
# @todo this may not be needed anymore, or could be easily removed
|
||||||
if not self.__data['watching'][uuid]['fetch_backend']:
|
if not self.__data['watching'][uuid]['fetch_backend']:
|
||||||
self.__data['watching'][uuid]['fetch_backend'] = self.__data['settings']['application']['fetch_backend']
|
self.__data['watching'][uuid]['fetch_backend'] = self.__data['settings']['application']['fetch_backend']
|
||||||
|
|
||||||
@@ -204,14 +188,13 @@ class ChangeDetectionStore:
|
|||||||
if not self.__data['settings']['application']['base_url']:
|
if not self.__data['settings']['application']['base_url']:
|
||||||
self.__data['settings']['application']['base_url'] = env_base_url.strip('" ')
|
self.__data['settings']['application']['base_url'] = env_base_url.strip('" ')
|
||||||
|
|
||||||
self.__data['has_unviewed'] = has_unviewed
|
|
||||||
|
|
||||||
return self.__data
|
return self.__data
|
||||||
|
|
||||||
def get_all_tags(self):
|
def get_all_tags(self):
|
||||||
tags = []
|
tags = []
|
||||||
for uuid, watch in self.data['watching'].items():
|
for uuid, watch in self.data['watching'].items():
|
||||||
|
if watch['tag'] is None:
|
||||||
|
continue
|
||||||
# Support for comma separated list of tags.
|
# Support for comma separated list of tags.
|
||||||
for tag in watch['tag'].split(','):
|
for tag in watch['tag'].split(','):
|
||||||
tag = tag.strip()
|
tag = tag.strip()
|
||||||
@@ -235,11 +218,11 @@ class ChangeDetectionStore:
|
|||||||
|
|
||||||
# GitHub #30 also delete history records
|
# GitHub #30 also delete history records
|
||||||
for uuid in self.data['watching']:
|
for uuid in self.data['watching']:
|
||||||
for path in self.data['watching'][uuid]['history'].values():
|
for path in self.data['watching'][uuid].history.values():
|
||||||
self.unlink_history_file(path)
|
self.unlink_history_file(path)
|
||||||
|
|
||||||
else:
|
else:
|
||||||
for path in self.data['watching'][uuid]['history'].values():
|
for path in self.data['watching'][uuid].history.values():
|
||||||
self.unlink_history_file(path)
|
self.unlink_history_file(path)
|
||||||
|
|
||||||
del self.data['watching'][uuid]
|
del self.data['watching'][uuid]
|
||||||
@@ -263,23 +246,34 @@ class ChangeDetectionStore:
|
|||||||
|
|
||||||
return False
|
return False
|
||||||
|
|
||||||
def get_val(self, uuid, val):
|
|
||||||
# Probably their should be dict...
|
|
||||||
return self.data['watching'][uuid].get(val)
|
|
||||||
|
|
||||||
# Remove a watchs data but keep the entry (URL etc)
|
# Remove a watchs data but keep the entry (URL etc)
|
||||||
def scrub_watch(self, uuid):
|
def clear_watch_history(self, uuid):
|
||||||
import pathlib
|
import pathlib
|
||||||
|
|
||||||
self.__data['watching'][uuid].update({'history': {}, 'last_checked': 0, 'last_changed': 0, 'newest_history_key': 0, 'previous_md5': False})
|
self.__data['watching'][uuid].update(
|
||||||
self.needs_write_urgent = True
|
{'last_checked': 0,
|
||||||
|
'last_viewed': 0,
|
||||||
|
'previous_md5': False,
|
||||||
|
'last_notification_error': False,
|
||||||
|
'last_error': False})
|
||||||
|
|
||||||
for item in pathlib.Path(self.datastore_path).rglob(uuid+"/*.txt"):
|
# JSON Data, Screenshots, Textfiles (history index and snapshots), HTML in the future etc
|
||||||
|
for item in pathlib.Path(os.path.join(self.datastore_path, uuid)).rglob("*.*"):
|
||||||
unlink(item)
|
unlink(item)
|
||||||
|
|
||||||
|
# Force the attr to recalculate
|
||||||
|
bump = self.__data['watching'][uuid].history
|
||||||
|
|
||||||
|
self.needs_write_urgent = True
|
||||||
|
|
||||||
def add_watch(self, url, tag="", extras=None, write_to_disk_now=True):
|
def add_watch(self, url, tag="", extras=None, write_to_disk_now=True):
|
||||||
|
|
||||||
if extras is None:
|
if extras is None:
|
||||||
extras = {}
|
extras = {}
|
||||||
|
# should always be str
|
||||||
|
if tag is None or not tag:
|
||||||
|
tag = ''
|
||||||
|
|
||||||
# Incase these are copied across, assume it's a reference and deepcopy()
|
# Incase these are copied across, assume it's a reference and deepcopy()
|
||||||
apply_extras = deepcopy(extras)
|
apply_extras = deepcopy(extras)
|
||||||
|
|
||||||
@@ -292,16 +286,32 @@ class ChangeDetectionStore:
|
|||||||
headers={'App-Guid': self.__data['app_guid']})
|
headers={'App-Guid': self.__data['app_guid']})
|
||||||
res = r.json()
|
res = r.json()
|
||||||
|
|
||||||
# List of permisable stuff we accept from the wild internet
|
# List of permissible attributes we accept from the wild internet
|
||||||
for k in ['url', 'tag',
|
for k in [
|
||||||
'paused', 'title',
|
'body',
|
||||||
'previous_md5', 'headers',
|
'css_filter',
|
||||||
'body', 'method',
|
'extract_text',
|
||||||
'ignore_text', 'css_filter',
|
'extract_title_as_title',
|
||||||
'subtractive_selectors', 'trigger_text',
|
'headers',
|
||||||
'extract_title_as_title']:
|
'ignore_text',
|
||||||
|
'include_filters',
|
||||||
|
'method',
|
||||||
|
'paused',
|
||||||
|
'previous_md5',
|
||||||
|
'subtractive_selectors',
|
||||||
|
'tag',
|
||||||
|
'text_should_not_be_present',
|
||||||
|
'title',
|
||||||
|
'trigger_text',
|
||||||
|
'webdriver_js_execute_code',
|
||||||
|
'url',
|
||||||
|
]:
|
||||||
if res.get(k):
|
if res.get(k):
|
||||||
|
if k != 'css_filter':
|
||||||
apply_extras[k] = res[k]
|
apply_extras[k] = res[k]
|
||||||
|
else:
|
||||||
|
# We renamed the field and made it a list
|
||||||
|
apply_extras['include_filters'] = [res['css_filter']]
|
||||||
|
|
||||||
except Exception as e:
|
except Exception as e:
|
||||||
logging.error("Error fetching metadata for shared watch link", url, str(e))
|
logging.error("Error fetching metadata for shared watch link", url, str(e))
|
||||||
@@ -309,68 +319,76 @@ class ChangeDetectionStore:
|
|||||||
return False
|
return False
|
||||||
|
|
||||||
with self.lock:
|
with self.lock:
|
||||||
# @todo use a common generic version of this
|
|
||||||
new_uuid = str(uuid_builder.uuid4())
|
|
||||||
# #Re 569
|
# #Re 569
|
||||||
# Not sure why deepcopy was needed here, sometimes new watches would appear to already have 'history' set
|
new_watch = Watch.model(datastore_path=self.datastore_path, default={
|
||||||
# I assumed this would instantiate a new object but somehow an existing dict was getting used
|
|
||||||
new_watch = deepcopy(Watch.model({
|
|
||||||
'url': url,
|
'url': url,
|
||||||
'tag': tag
|
'tag': tag
|
||||||
}))
|
})
|
||||||
|
|
||||||
|
new_uuid = new_watch['uuid']
|
||||||
|
logging.debug("Added URL {} - {}".format(url, new_uuid))
|
||||||
|
|
||||||
for k in ['uuid', 'history', 'last_checked', 'last_changed', 'newest_history_key', 'previous_md5', 'viewed']:
|
for k in ['uuid', 'history', 'last_checked', 'last_changed', 'newest_history_key', 'previous_md5', 'viewed']:
|
||||||
if k in apply_extras:
|
if k in apply_extras:
|
||||||
del apply_extras[k]
|
del apply_extras[k]
|
||||||
|
|
||||||
new_watch.update(apply_extras)
|
new_watch.update(apply_extras)
|
||||||
self.__data['watching'][new_uuid]=new_watch
|
self.__data['watching'][new_uuid] = new_watch
|
||||||
|
|
||||||
# Get the directory ready
|
self.__data['watching'][new_uuid].ensure_data_dir_exists()
|
||||||
output_path = "{}/{}".format(self.datastore_path, new_uuid)
|
|
||||||
try:
|
|
||||||
mkdir(output_path)
|
|
||||||
except FileExistsError:
|
|
||||||
print(output_path, "already exists.")
|
|
||||||
|
|
||||||
if write_to_disk_now:
|
if write_to_disk_now:
|
||||||
self.sync_to_json()
|
self.sync_to_json()
|
||||||
|
|
||||||
return new_uuid
|
return new_uuid
|
||||||
|
|
||||||
# Save some text file to the appropriate path and bump the history
|
def visualselector_data_is_ready(self, watch_uuid):
|
||||||
# result_obj from fetch_site_status.run()
|
|
||||||
def save_history_text(self, watch_uuid, contents):
|
|
||||||
import uuid
|
|
||||||
|
|
||||||
output_path = "{}/{}".format(self.datastore_path, watch_uuid)
|
output_path = "{}/{}".format(self.datastore_path, watch_uuid)
|
||||||
# Incase the operator deleted it, check and create.
|
screenshot_filename = "{}/last-screenshot.png".format(output_path)
|
||||||
if not os.path.isdir(output_path):
|
elements_index_filename = "{}/elements.json".format(output_path)
|
||||||
mkdir(output_path)
|
if path.isfile(screenshot_filename) and path.isfile(elements_index_filename) :
|
||||||
|
return True
|
||||||
fname = "{}/{}.stripped.txt".format(output_path, uuid.uuid4())
|
|
||||||
with open(fname, 'wb') as f:
|
|
||||||
f.write(contents)
|
|
||||||
f.close()
|
|
||||||
|
|
||||||
return fname
|
|
||||||
|
|
||||||
def get_screenshot(self, watch_uuid):
|
|
||||||
output_path = "{}/{}".format(self.datastore_path, watch_uuid)
|
|
||||||
fname = "{}/last-screenshot.png".format(output_path)
|
|
||||||
if path.isfile(fname):
|
|
||||||
return fname
|
|
||||||
|
|
||||||
return False
|
return False
|
||||||
|
|
||||||
# Save as PNG, PNG is larger but better for doing visual diff in the future
|
# Save as PNG, PNG is larger but better for doing visual diff in the future
|
||||||
def save_screenshot(self, watch_uuid, screenshot: bytes):
|
def save_screenshot(self, watch_uuid, screenshot: bytes, as_error=False):
|
||||||
output_path = "{}/{}".format(self.datastore_path, watch_uuid)
|
if not self.data['watching'].get(watch_uuid):
|
||||||
fname = "{}/last-screenshot.png".format(output_path)
|
return
|
||||||
with open(fname, 'wb') as f:
|
|
||||||
|
if as_error:
|
||||||
|
target_path = os.path.join(self.datastore_path, watch_uuid, "last-error-screenshot.png")
|
||||||
|
else:
|
||||||
|
target_path = os.path.join(self.datastore_path, watch_uuid, "last-screenshot.png")
|
||||||
|
|
||||||
|
self.data['watching'][watch_uuid].ensure_data_dir_exists()
|
||||||
|
|
||||||
|
with open(target_path, 'wb') as f:
|
||||||
f.write(screenshot)
|
f.write(screenshot)
|
||||||
f.close()
|
f.close()
|
||||||
|
|
||||||
|
def save_error_text(self, watch_uuid, contents):
|
||||||
|
if not self.data['watching'].get(watch_uuid):
|
||||||
|
return
|
||||||
|
target_path = os.path.join(self.datastore_path, watch_uuid, "last-error.txt")
|
||||||
|
|
||||||
|
with open(target_path, 'w') as f:
|
||||||
|
f.write(contents)
|
||||||
|
|
||||||
|
def save_xpath_data(self, watch_uuid, data, as_error=False):
|
||||||
|
if not self.data['watching'].get(watch_uuid):
|
||||||
|
return
|
||||||
|
if as_error:
|
||||||
|
target_path = os.path.join(self.datastore_path, watch_uuid, "elements-error.json")
|
||||||
|
else:
|
||||||
|
target_path = os.path.join(self.datastore_path, watch_uuid, "elements.json")
|
||||||
|
|
||||||
|
with open(target_path, 'w') as f:
|
||||||
|
f.write(json.dumps(data))
|
||||||
|
f.close()
|
||||||
|
|
||||||
|
|
||||||
def sync_to_json(self):
|
def sync_to_json(self):
|
||||||
logging.info("Saving JSON..")
|
logging.info("Saving JSON..")
|
||||||
print("Saving JSON..")
|
print("Saving JSON..")
|
||||||
@@ -423,8 +441,8 @@ class ChangeDetectionStore:
|
|||||||
|
|
||||||
index=[]
|
index=[]
|
||||||
for uuid in self.data['watching']:
|
for uuid in self.data['watching']:
|
||||||
for id in self.data['watching'][uuid]['history']:
|
for id in self.data['watching'][uuid].history:
|
||||||
index.append(self.data['watching'][uuid]['history'][str(id)])
|
index.append(self.data['watching'][uuid].history[str(id)])
|
||||||
|
|
||||||
import pathlib
|
import pathlib
|
||||||
|
|
||||||
@@ -436,20 +454,42 @@ class ChangeDetectionStore:
|
|||||||
unlink(item)
|
unlink(item)
|
||||||
|
|
||||||
def import_proxy_list(self, filename):
|
def import_proxy_list(self, filename):
|
||||||
import csv
|
with open(filename) as f:
|
||||||
with open(filename, newline='') as f:
|
self.proxy_list = json.load(f)
|
||||||
reader = csv.reader(f, skipinitialspace=True)
|
print ("Registered proxy list", list(self.proxy_list.keys()))
|
||||||
# @todo This loop can could be improved
|
|
||||||
l = []
|
|
||||||
for row in reader:
|
|
||||||
if len(row):
|
|
||||||
if len(row)>=2:
|
|
||||||
l.append(tuple(row[:2]))
|
|
||||||
else:
|
|
||||||
l.append(tuple([row[0], row[0]]))
|
|
||||||
self.proxy_list = l if len(l) else None
|
|
||||||
|
|
||||||
|
|
||||||
|
def get_preferred_proxy_for_watch(self, uuid):
|
||||||
|
"""
|
||||||
|
Returns the preferred proxy by ID key
|
||||||
|
:param uuid: UUID
|
||||||
|
:return: proxy "key" id
|
||||||
|
"""
|
||||||
|
|
||||||
|
proxy_id = None
|
||||||
|
if self.proxy_list is None:
|
||||||
|
return None
|
||||||
|
|
||||||
|
# If its a valid one
|
||||||
|
watch = self.data['watching'].get(uuid)
|
||||||
|
|
||||||
|
if watch.get('proxy') and watch.get('proxy') in list(self.proxy_list.keys()):
|
||||||
|
return watch.get('proxy')
|
||||||
|
|
||||||
|
# not valid (including None), try the system one
|
||||||
|
else:
|
||||||
|
system_proxy_id = self.data['settings']['requests'].get('proxy')
|
||||||
|
# Is not None and exists
|
||||||
|
if self.proxy_list.get(system_proxy_id):
|
||||||
|
return system_proxy_id
|
||||||
|
|
||||||
|
# Fallback - Did not resolve anything, use the first available
|
||||||
|
if system_proxy_id is None:
|
||||||
|
first_default = list(self.proxy_list)[0]
|
||||||
|
return first_default
|
||||||
|
|
||||||
|
return None
|
||||||
|
|
||||||
# Run all updates
|
# Run all updates
|
||||||
# IMPORTANT - Each update could be run even when they have a new install and the schema is correct
|
# IMPORTANT - Each update could be run even when they have a new install and the schema is correct
|
||||||
# So therefor - each `update_n` should be very careful about checking if it needs to actually run
|
# So therefor - each `update_n` should be very careful about checking if it needs to actually run
|
||||||
@@ -495,3 +535,83 @@ class ChangeDetectionStore:
|
|||||||
# Only upgrade individual watch time if it was set
|
# Only upgrade individual watch time if it was set
|
||||||
if watch.get('minutes_between_check', False):
|
if watch.get('minutes_between_check', False):
|
||||||
self.data['watching'][uuid]['time_between_check']['minutes'] = watch['minutes_between_check']
|
self.data['watching'][uuid]['time_between_check']['minutes'] = watch['minutes_between_check']
|
||||||
|
|
||||||
|
# Move the history list to a flat text file index
|
||||||
|
# Better than SQLite because this list is only appended to, and works across NAS / NFS type setups
|
||||||
|
def update_2(self):
|
||||||
|
# @todo test running this on a newly updated one (when this already ran)
|
||||||
|
for uuid, watch in self.data['watching'].items():
|
||||||
|
history = []
|
||||||
|
|
||||||
|
if watch.get('history', False):
|
||||||
|
for d, p in watch['history'].items():
|
||||||
|
d = int(d) # Used to be keyed as str, we'll fix this now too
|
||||||
|
history.append("{},{}\n".format(d,p))
|
||||||
|
|
||||||
|
if len(history):
|
||||||
|
target_path = os.path.join(self.datastore_path, uuid)
|
||||||
|
if os.path.exists(target_path):
|
||||||
|
with open(os.path.join(target_path, "history.txt"), "w") as f:
|
||||||
|
f.writelines(history)
|
||||||
|
else:
|
||||||
|
logging.warning("Datastore history directory {} does not exist, skipping history import.".format(target_path))
|
||||||
|
|
||||||
|
# No longer needed, dynamically pulled from the disk when needed.
|
||||||
|
# But we should set it back to a empty dict so we don't break if this schema runs on an earlier version.
|
||||||
|
# In the distant future we can remove this entirely
|
||||||
|
self.data['watching'][uuid]['history'] = {}
|
||||||
|
|
||||||
|
# We incorrectly stored last_changed when there was not a change, and then confused the output list table
|
||||||
|
def update_3(self):
|
||||||
|
# see https://github.com/dgtlmoon/changedetection.io/pull/835
|
||||||
|
return
|
||||||
|
|
||||||
|
# `last_changed` not needed, we pull that information from the history.txt index
|
||||||
|
def update_4(self):
|
||||||
|
for uuid, watch in self.data['watching'].items():
|
||||||
|
try:
|
||||||
|
# Remove it from the struct
|
||||||
|
del(watch['last_changed'])
|
||||||
|
except:
|
||||||
|
continue
|
||||||
|
return
|
||||||
|
|
||||||
|
def update_5(self):
|
||||||
|
# If the watch notification body, title look the same as the global one, unset it, so the watch defaults back to using the main settings
|
||||||
|
# In other words - the watch notification_title and notification_body are not needed if they are the same as the default one
|
||||||
|
current_system_body = self.data['settings']['application']['notification_body'].translate(str.maketrans('', '', "\r\n "))
|
||||||
|
current_system_title = self.data['settings']['application']['notification_body'].translate(str.maketrans('', '', "\r\n "))
|
||||||
|
for uuid, watch in self.data['watching'].items():
|
||||||
|
try:
|
||||||
|
watch_body = watch.get('notification_body', '')
|
||||||
|
if watch_body and watch_body.translate(str.maketrans('', '', "\r\n ")) == current_system_body:
|
||||||
|
# Looks the same as the default one, so unset it
|
||||||
|
watch['notification_body'] = None
|
||||||
|
|
||||||
|
watch_title = watch.get('notification_title', '')
|
||||||
|
if watch_title and watch_title.translate(str.maketrans('', '', "\r\n ")) == current_system_title:
|
||||||
|
# Looks the same as the default one, so unset it
|
||||||
|
watch['notification_title'] = None
|
||||||
|
except Exception as e:
|
||||||
|
continue
|
||||||
|
return
|
||||||
|
|
||||||
|
|
||||||
|
# We incorrectly used common header overrides that should only apply to Requests
|
||||||
|
# These are now handled in content_fetcher::html_requests and shouldnt be passed to Playwright/Selenium
|
||||||
|
def update_7(self):
|
||||||
|
# These were hard-coded in early versions
|
||||||
|
for v in ['User-Agent', 'Accept', 'Accept-Encoding', 'Accept-Language']:
|
||||||
|
if self.data['settings']['headers'].get(v):
|
||||||
|
del self.data['settings']['headers'][v]
|
||||||
|
|
||||||
|
# Convert filters to a list of filters css_filter -> include_filters
|
||||||
|
def update_8(self):
|
||||||
|
for uuid, watch in self.data['watching'].items():
|
||||||
|
try:
|
||||||
|
existing_filter = watch.get('css_filter', '')
|
||||||
|
if existing_filter:
|
||||||
|
watch['include_filters'] = [existing_filter]
|
||||||
|
except:
|
||||||
|
continue
|
||||||
|
return
|
||||||
@@ -1,20 +1,21 @@
|
|||||||
|
|
||||||
{% from '_helpers.jinja' import render_field %}
|
{% from '_helpers.jinja' import render_field %}
|
||||||
|
|
||||||
{% macro render_common_settings_form(form, current_base_url, emailprefix) %}
|
{% macro render_common_settings_form(form, emailprefix, settings_application) %}
|
||||||
<div class="pure-control-group">
|
<div class="pure-control-group">
|
||||||
{{ render_field(form.notification_urls, rows=5, placeholder="Examples:
|
{{ render_field(form.notification_urls, rows=5, placeholder="Examples:
|
||||||
Gitter - gitter://token/room
|
Gitter - gitter://token/room
|
||||||
Office365 - o365://TenantID:AccountEmail/ClientID/ClientSecret/TargetEmail
|
Office365 - o365://TenantID:AccountEmail/ClientID/ClientSecret/TargetEmail
|
||||||
AWS SNS - sns://AccessKeyID/AccessSecretKey/RegionName/+PhoneNo
|
AWS SNS - sns://AccessKeyID/AccessSecretKey/RegionName/+PhoneNo
|
||||||
SMTPS - mailtos://user:pass@mail.domain.com?to=receivingAddress@example.com", class="notification-urls")
|
SMTPS - mailtos://user:pass@mail.domain.com?to=receivingAddress@example.com",
|
||||||
|
class="notification-urls" )
|
||||||
}}
|
}}
|
||||||
<div class="pure-form-message-inline">
|
<div class="pure-form-message-inline">
|
||||||
<ul>
|
<ul>
|
||||||
<li>Use <a target=_new href="https://github.com/caronc/apprise">AppRise URLs</a> for notification to just about any service! <i><a target=_new href="https://github.com/dgtlmoon/changedetection.io/wiki/Notification-configuration-notes">Please read the notification services wiki here for important configuration notes</a></i>.</li>
|
<li>Use <a target=_new href="https://github.com/caronc/apprise">AppRise URLs</a> for notification to just about any service! <i><a target=_new href="https://github.com/dgtlmoon/changedetection.io/wiki/Notification-configuration-notes">Please read the notification services wiki here for important configuration notes</a></i>.</li>
|
||||||
<li><code>discord://</code> only supports a maximum <strong>2,000 characters</strong> of notification text, including the title.</li>
|
<li><code>discord://</code> only supports a maximum <strong>2,000 characters</strong> of notification text, including the title.</li>
|
||||||
<li><code>tgram://</code> bots cant send messages to other bots, so you should specify chat ID of non-bot user.</li>
|
<li><code>tgram://</code> bots cant send messages to other bots, so you should specify chat ID of non-bot user.</li>
|
||||||
<li>Go here for <a href="{{url_for('notification_logs')}}">notification debug logs</a></li>
|
<li><code>tgram://</code> only supports very limited HTML and can fail when extra tags are sent, <a href="https://core.telegram.org/bots/api#html-style">read more here</a> (or use plaintext/markdown format)</li>
|
||||||
</ul>
|
</ul>
|
||||||
</div>
|
</div>
|
||||||
<br/>
|
<br/>
|
||||||
@@ -22,18 +23,20 @@
|
|||||||
{% if emailprefix %}
|
{% if emailprefix %}
|
||||||
<a id="add-email-helper" class="pure-button button-secondary button-xsmall" style="font-size: 70%">Add email</a>
|
<a id="add-email-helper" class="pure-button button-secondary button-xsmall" style="font-size: 70%">Add email</a>
|
||||||
{% endif %}
|
{% endif %}
|
||||||
|
<a href="{{url_for('notification_logs')}}" class="pure-button button-secondary button-xsmall" style="font-size: 70%">Notification debug logs</a>
|
||||||
</div>
|
</div>
|
||||||
<div id="notification-customisation" class="pure-control-group">
|
<div id="notification-customisation" class="pure-control-group">
|
||||||
<div class="pure-control-group">
|
<div class="pure-control-group">
|
||||||
{{ render_field(form.notification_title, class="m-d notification-title") }}
|
{{ render_field(form.notification_title, class="m-d notification-title", placeholder=settings_application['notification_title']) }}
|
||||||
<span class="pure-form-message-inline">Title for all notifications</span>
|
<span class="pure-form-message-inline">Title for all notifications</span>
|
||||||
</div>
|
</div>
|
||||||
<div class="pure-control-group">
|
<div class="pure-control-group">
|
||||||
{{ render_field(form.notification_body , rows=5, class="notification-body") }}
|
{{ render_field(form.notification_body , rows=5, class="notification-body", placeholder=settings_application['notification_body']) }}
|
||||||
<span class="pure-form-message-inline">Body for all notifications</span>
|
<span class="pure-form-message-inline">Body for all notifications</span>
|
||||||
</div>
|
</div>
|
||||||
<div class="pure-control-group">
|
<div class="pure-control-group">
|
||||||
{{ render_field(form.notification_format , rows=5, class="notification-format") }}
|
<!-- unsure -->
|
||||||
|
{{ render_field(form.notification_format , class="notification-format") }}
|
||||||
<span class="pure-form-message-inline">Format for all notifications</span>
|
<span class="pure-form-message-inline">Format for all notifications</span>
|
||||||
</div>
|
</div>
|
||||||
<div class="pure-controls">
|
<div class="pure-controls">
|
||||||
@@ -93,7 +96,7 @@
|
|||||||
</table>
|
</table>
|
||||||
<br/>
|
<br/>
|
||||||
URLs generated by changedetection.io (such as <code>{diff_url}</code>) require the <code>BASE_URL</code> environment variable set.<br/>
|
URLs generated by changedetection.io (such as <code>{diff_url}</code>) require the <code>BASE_URL</code> environment variable set.<br/>
|
||||||
Your <code>BASE_URL</code> var is currently "{{current_base_url}}"
|
Your <code>BASE_URL</code> var is currently "{{settings_application['current_base_url']}}"
|
||||||
</span>
|
</span>
|
||||||
</div>
|
</div>
|
||||||
</div>
|
</div>
|
||||||
|
|||||||
7
changedetectionio/templates/_pagination.jinja
Normal file
7
changedetectionio/templates/_pagination.jinja
Normal file
@@ -0,0 +1,7 @@
|
|||||||
|
{% macro pagination(sorted_watches, total_per_page, current_page) %}
|
||||||
|
{{ sorted_watches|length }}
|
||||||
|
|
||||||
|
{% for row in sorted_watches|batch(total_per_page, ' ') %}
|
||||||
|
{{ loop.index}}
|
||||||
|
{% endfor %}
|
||||||
|
{% endmacro %}
|
||||||
@@ -3,22 +3,22 @@
|
|||||||
{% block content %}
|
{% block content %}
|
||||||
<div class="edit-form">
|
<div class="edit-form">
|
||||||
<div class="box-wrap inner">
|
<div class="box-wrap inner">
|
||||||
<form class="pure-form pure-form-stacked" action="{{url_for('scrub_page')}}" method="POST">
|
<form class="pure-form pure-form-stacked" action="{{url_for('clear_all_history')}}" method="POST">
|
||||||
<input type="hidden" name="csrf_token" value="{{ csrf_token() }}"/>
|
<input type="hidden" name="csrf_token" value="{{ csrf_token() }}"/>
|
||||||
<fieldset>
|
<fieldset>
|
||||||
<div class="pure-control-group">
|
<div class="pure-control-group">
|
||||||
This will remove ALL version snapshots/data, but keep your list of URLs. <br/>
|
This will remove version history (snapshots) for ALL watches, but keep your list of URLs! <br/>
|
||||||
You may like to use the <strong>BACKUP</strong> link first.<br/>
|
You may like to use the <strong>BACKUP</strong> link first.<br/>
|
||||||
</div>
|
</div>
|
||||||
<br/>
|
<br/>
|
||||||
<div class="pure-control-group">
|
<div class="pure-control-group">
|
||||||
<label for="confirmtext">Confirmation text</label>
|
<label for="confirmtext">Confirmation text</label>
|
||||||
<input type="text" id="confirmtext" required="" name="confirmtext" value="" size="10"/>
|
<input type="text" id="confirmtext" required="" name="confirmtext" value="" size="10"/>
|
||||||
<span class="pure-form-message-inline">Type in the word <strong>scrub</strong> to confirm that you understand!</span>
|
<span class="pure-form-message-inline">Type in the word <strong>clear</strong> to confirm that you understand.</span>
|
||||||
</div>
|
</div>
|
||||||
<br/>
|
<br/>
|
||||||
<div class="pure-control-group">
|
<div class="pure-control-group">
|
||||||
<button type="submit" class="pure-button pure-button-primary">Scrub!</button>
|
<button type="submit" class="pure-button pure-button-primary">Clear History!</button>
|
||||||
</div>
|
</div>
|
||||||
<br/>
|
<br/>
|
||||||
<div class="pure-control-group">
|
<div class="pure-control-group">
|
||||||
@@ -1,6 +1,14 @@
|
|||||||
{% extends 'base.html' %}
|
{% extends 'base.html' %}
|
||||||
|
|
||||||
{% block content %}
|
{% block content %}
|
||||||
|
<script>
|
||||||
|
const screenshot_url="{{url_for('static_content', group='screenshot', filename=uuid)}}";
|
||||||
|
{% if last_error_screenshot %}
|
||||||
|
const error_screenshot_url="{{url_for('static_content', group='screenshot', filename=uuid, error_screenshot=1) }}";
|
||||||
|
{% endif %}
|
||||||
|
</script>
|
||||||
|
<script type="text/javascript" src="{{url_for('static_content', group='js', filename='diff-overview.js')}}" defer></script>
|
||||||
|
|
||||||
<div id="settings">
|
<div id="settings">
|
||||||
<h1>Differences</h1>
|
<h1>Differences</h1>
|
||||||
<form class="pure-form " action="" method="GET">
|
<form class="pure-form " action="" method="GET">
|
||||||
@@ -17,7 +25,7 @@
|
|||||||
{% if versions|length >= 1 %}
|
{% if versions|length >= 1 %}
|
||||||
<label for="diff-version">Compare newest (<span id="current-v-date"></span>) with</label>
|
<label for="diff-version">Compare newest (<span id="current-v-date"></span>) with</label>
|
||||||
<select id="diff-version" name="previous_version">
|
<select id="diff-version" name="previous_version">
|
||||||
{% for version in versions %}
|
{% for version in versions|reverse %}
|
||||||
<option value="{{version}}" {% if version== current_previous_version %} selected="" {% endif %}>
|
<option value="{{version}}" {% if version== current_previous_version %} selected="" {% endif %}>
|
||||||
{{version}}
|
{{version}}
|
||||||
</option>
|
</option>
|
||||||
@@ -38,17 +46,31 @@
|
|||||||
<script type="text/javascript" src="{{url_for('static_content', group='js', filename='tabs.js')}}" defer></script>
|
<script type="text/javascript" src="{{url_for('static_content', group='js', filename='tabs.js')}}" defer></script>
|
||||||
<div class="tabs">
|
<div class="tabs">
|
||||||
<ul>
|
<ul>
|
||||||
<li class="tab" id="default-tab"><a href="#text">Text</a></li>
|
{% if last_error_text %}<li class="tab" id="error-text-tab"><a href="#error-text">Error Text</a></li> {% endif %}
|
||||||
{% if screenshot %}
|
{% if last_error_screenshot %}<li class="tab" id="error-screenshot-tab"><a href="#error-screenshot">Error Screenshot</a></li> {% endif %}
|
||||||
<li class="tab"><a href="#screenshot">Current screenshot</a></li>
|
<li class="tab" id=""><a href="#text">Text</a></li>
|
||||||
{% endif %}
|
<li class="tab" id="screenshot-tab"><a href="#screenshot">Screenshot</a></li>
|
||||||
</ul>
|
</ul>
|
||||||
</div>
|
</div>
|
||||||
|
|
||||||
<div id="diff-ui">
|
<div id="diff-ui">
|
||||||
|
<div class="tab-pane-inner" id="error-text">
|
||||||
|
<div class="snapshot-age error">{{watch_a.error_text_ctime|format_seconds_ago}} seconds ago</div>
|
||||||
|
<pre>
|
||||||
|
{{ last_error_text }}
|
||||||
|
</pre>
|
||||||
|
</div>
|
||||||
|
|
||||||
|
<div class="tab-pane-inner" id="error-screenshot">
|
||||||
|
<div class="snapshot-age error">{{watch_a.snapshot_error_screenshot_ctime|format_seconds_ago}} seconds ago</div>
|
||||||
|
<img id="error-screenshot-img" style="max-width: 80%" alt="Current error-ing screenshot from most recent request"/>
|
||||||
|
</div>
|
||||||
|
|
||||||
<div class="tab-pane-inner" id="text">
|
<div class="tab-pane-inner" id="text">
|
||||||
<div class="tip">Pro-tip: Use <strong>show current snapshot</strong> tab to visualise what will be ignored.
|
<div class="tip">Pro-tip: Use <strong>show current snapshot</strong> tab to visualise what will be ignored.
|
||||||
</div>
|
</div>
|
||||||
|
<div class="snapshot-age">{{watch_a.snapshot_text_ctime|format_timestamp_timeago}}</div>
|
||||||
|
|
||||||
<table>
|
<table>
|
||||||
<tbody>
|
<tbody>
|
||||||
<tr>
|
<tr>
|
||||||
@@ -63,17 +85,21 @@
|
|||||||
</table>
|
</table>
|
||||||
Diff algorithm from the amazing <a href="https://github.com/kpdecker/jsdiff">github.com/kpdecker/jsdiff</a>
|
Diff algorithm from the amazing <a href="https://github.com/kpdecker/jsdiff">github.com/kpdecker/jsdiff</a>
|
||||||
</div>
|
</div>
|
||||||
|
|
||||||
{% if screenshot %}
|
|
||||||
<div class="tab-pane-inner" id="screenshot">
|
<div class="tab-pane-inner" id="screenshot">
|
||||||
<p>
|
<div class="tip">
|
||||||
<i>For now, only the most recent screenshot is saved and displayed.</i>
|
For now, Differences are performed on text, not graphically, only the latest screenshot is available.
|
||||||
</p>
|
</div>
|
||||||
|
{% if is_html_webdriver %}
|
||||||
<img src="{{url_for('static_content', group='screenshot', filename=uuid)}}">
|
{% if screenshot %}
|
||||||
|
<div class="snapshot-age">{{watch_a.snapshot_screenshot_ctime|format_timestamp_timeago}}</div>
|
||||||
|
<img style="max-width: 80%" id="screenshot-img" alt="Current screenshot from most recent request"/>
|
||||||
|
{% else %}
|
||||||
|
No screenshot available just yet! Try rechecking the page.
|
||||||
|
{% endif %}
|
||||||
|
{% else %}
|
||||||
|
<strong>Screenshot requires Playwright/WebDriver enabled</strong>
|
||||||
|
{% endif %}
|
||||||
</div>
|
</div>
|
||||||
{% endif %}
|
|
||||||
|
|
||||||
</div>
|
</div>
|
||||||
|
|
||||||
|
|
||||||
@@ -81,7 +107,6 @@
|
|||||||
|
|
||||||
<script defer="">
|
<script defer="">
|
||||||
|
|
||||||
|
|
||||||
var a = document.getElementById('a');
|
var a = document.getElementById('a');
|
||||||
var b = document.getElementById('b');
|
var b = document.getElementById('b');
|
||||||
var result = document.getElementById('result');
|
var result = document.getElementById('result');
|
||||||
|
|||||||
@@ -5,19 +5,27 @@
|
|||||||
<script type="text/javascript" src="{{url_for('static_content', group='js', filename='tabs.js')}}" defer></script>
|
<script type="text/javascript" src="{{url_for('static_content', group='js', filename='tabs.js')}}" defer></script>
|
||||||
<script>
|
<script>
|
||||||
const notification_base_url="{{url_for('ajax_callback_send_notification_test')}}";
|
const notification_base_url="{{url_for('ajax_callback_send_notification_test')}}";
|
||||||
|
const watch_visual_selector_data_url="{{url_for('static_content', group='visual_selector_data', filename=uuid)}}";
|
||||||
|
const screenshot_url="{{url_for('static_content', group='screenshot', filename=uuid)}}";
|
||||||
|
const playwright_enabled={% if playwright_enabled %} true {% else %} false {% endif %};
|
||||||
|
|
||||||
{% if emailprefix %}
|
{% if emailprefix %}
|
||||||
const email_notification_prefix=JSON.parse('{{ emailprefix|tojson }}');
|
const email_notification_prefix=JSON.parse('{{ emailprefix|tojson }}');
|
||||||
{% endif %}
|
{% endif %}
|
||||||
|
|
||||||
</script>
|
</script>
|
||||||
<script type="text/javascript" src="{{url_for('static_content', group='js', filename='watch-settings.js')}}" defer></script>
|
<script type="text/javascript" src="{{url_for('static_content', group='js', filename='watch-settings.js')}}" defer></script>
|
||||||
<script type="text/javascript" src="{{url_for('static_content', group='js', filename='notifications.js')}}" defer></script>
|
<script type="text/javascript" src="{{url_for('static_content', group='js', filename='notifications.js')}}" defer></script>
|
||||||
|
<script type="text/javascript" src="{{url_for('static_content', group='js', filename='visual-selector.js')}}" defer></script>
|
||||||
|
<script type="text/javascript" src="{{url_for('static_content', group='js', filename='limit.js')}}" defer></script>
|
||||||
|
|
||||||
<div class="edit-form monospaced-textarea">
|
<div class="edit-form monospaced-textarea">
|
||||||
|
|
||||||
<div class="tabs collapsable">
|
<div class="tabs collapsable">
|
||||||
<ul>
|
<ul>
|
||||||
<li class="tab" id="default-tab"><a href="#general">General</a></li>
|
<li class="tab" id=""><a href="#general">General</a></li>
|
||||||
<li class="tab"><a href="#request">Request</a></li>
|
<li class="tab"><a href="#request">Request</a></li>
|
||||||
|
<li class="tab"><a id="visualselector-tab" href="#visualselector">Visual Filter Selector</a></li>
|
||||||
<li class="tab"><a href="#filters-and-triggers">Filters & Triggers</a></li>
|
<li class="tab"><a href="#filters-and-triggers">Filters & Triggers</a></li>
|
||||||
<li class="tab"><a href="#notifications">Notifications</a></li>
|
<li class="tab"><a href="#notifications">Notifications</a></li>
|
||||||
</ul>
|
</ul>
|
||||||
@@ -25,14 +33,15 @@
|
|||||||
|
|
||||||
<div class="box-wrap inner">
|
<div class="box-wrap inner">
|
||||||
<form class="pure-form pure-form-stacked"
|
<form class="pure-form pure-form-stacked"
|
||||||
action="{{ url_for('edit_page', uuid=uuid, next = request.args.get('next') ) }}" method="POST">
|
action="{{ url_for('edit_page', uuid=uuid, next = request.args.get('next'), unpause_on_save = request.args.get('unpause_on_save')) }}" method="POST">
|
||||||
<input type="hidden" name="csrf_token" value="{{ csrf_token() }}"/>
|
<input type="hidden" name="csrf_token" value="{{ csrf_token() }}"/>
|
||||||
|
|
||||||
<div class="tab-pane-inner" id="general">
|
<div class="tab-pane-inner" id="general">
|
||||||
<fieldset>
|
<fieldset>
|
||||||
<div class="pure-control-group">
|
<div class="pure-control-group">
|
||||||
{{ render_field(form.url, placeholder="https://...", required=true, class="m-d") }}
|
{{ render_field(form.url, placeholder="https://...", required=true, class="m-d") }}
|
||||||
<span class="pure-form-message-inline">Some sites use JavaScript to create the content, for this you should <a href="https://github.com/dgtlmoon/changedetection.io/wiki/Fetching-pages-with-WebDriver">use the Chrome/WebDriver Fetcher</a></span>
|
<span class="pure-form-message-inline">Some sites use JavaScript to create the content, for this you should <a href="https://github.com/dgtlmoon/changedetection.io/wiki/Fetching-pages-with-WebDriver">use the Chrome/WebDriver Fetcher</a></span><br/>
|
||||||
|
<span class="pure-form-message-inline">You can use variables in the URL, perfect for inserting the current date and other logic, <a href="https://github.com/dgtlmoon/changedetection.io/wiki/Handling-variables-in-the-watched-URL">help and examples here</a></span><br/>
|
||||||
</div>
|
</div>
|
||||||
<div class="pure-control-group">
|
<div class="pure-control-group">
|
||||||
{{ render_field(form.title, class="m-d") }}
|
{{ render_field(form.title, class="m-d") }}
|
||||||
@@ -54,6 +63,12 @@
|
|||||||
<div class="pure-control-group">
|
<div class="pure-control-group">
|
||||||
{{ render_checkbox_field(form.extract_title_as_title) }}
|
{{ render_checkbox_field(form.extract_title_as_title) }}
|
||||||
</div>
|
</div>
|
||||||
|
<div class="pure-control-group">
|
||||||
|
{{ render_checkbox_field(form.filter_failure_notification_send) }}
|
||||||
|
<span class="pure-form-message-inline">
|
||||||
|
Sends a notification when the filter can no longer be seen on the page, good for knowing when the page changed and your filter will not work anymore.
|
||||||
|
</span>
|
||||||
|
</div>
|
||||||
</fieldset>
|
</fieldset>
|
||||||
</div>
|
</div>
|
||||||
|
|
||||||
@@ -63,6 +78,7 @@
|
|||||||
<span class="pure-form-message-inline">
|
<span class="pure-form-message-inline">
|
||||||
<p>Use the <strong>Basic</strong> method (default) where your watched site doesn't need Javascript to render.</p>
|
<p>Use the <strong>Basic</strong> method (default) where your watched site doesn't need Javascript to render.</p>
|
||||||
<p>The <strong>Chrome/Javascript</strong> method requires a network connection to a running WebDriver+Chrome server, set by the ENV var 'WEBDRIVER_URL'. </p>
|
<p>The <strong>Chrome/Javascript</strong> method requires a network connection to a running WebDriver+Chrome server, set by the ENV var 'WEBDRIVER_URL'. </p>
|
||||||
|
Tip: <a href="https://github.com/dgtlmoon/changedetection.io/wiki/Proxy-configuration#brightdata-proxy-support">Connect using BrightData Proxies, find out more here.</a>
|
||||||
</span>
|
</span>
|
||||||
</div>
|
</div>
|
||||||
{% if form.proxy %}
|
{% if form.proxy %}
|
||||||
@@ -73,34 +89,43 @@
|
|||||||
</span>
|
</span>
|
||||||
</div>
|
</div>
|
||||||
{% endif %}
|
{% endif %}
|
||||||
<fieldset class="pure-group" id="webdriver-override-options">
|
<div class="pure-control-group inline-radio">
|
||||||
|
{{ render_checkbox_field(form.ignore_status_codes) }}
|
||||||
|
</div>
|
||||||
|
<fieldset id="webdriver-override-options">
|
||||||
|
<div class="pure-control-group">
|
||||||
|
{{ render_field(form.webdriver_delay) }}
|
||||||
<div class="pure-form-message-inline">
|
<div class="pure-form-message-inline">
|
||||||
<strong>If you're having trouble waiting for the page to be fully rendered (text missing etc), try increasing the 'wait' time here.</strong>
|
<strong>If you're having trouble waiting for the page to be fully rendered (text missing etc), try increasing the 'wait' time here.</strong>
|
||||||
<br/>
|
<br/>
|
||||||
This will wait <i>n</i> seconds before extracting the text.
|
This will wait <i>n</i> seconds before extracting the text.
|
||||||
|
{% if using_global_webdriver_wait %}
|
||||||
|
<br/><strong>Using the current global default settings</strong>
|
||||||
|
{% endif %}
|
||||||
|
</div>
|
||||||
</div>
|
</div>
|
||||||
<div class="pure-control-group">
|
<div class="pure-control-group">
|
||||||
{{ render_field(form.webdriver_delay) }}
|
{{ render_field(form.webdriver_js_execute_code) }}
|
||||||
</div>
|
|
||||||
{% if using_global_webdriver_wait %}
|
|
||||||
<div class="pure-form-message-inline">
|
<div class="pure-form-message-inline">
|
||||||
<strong>Using the current global default settings</strong>
|
Run this code before performing change detection, handy for filling in fields and other actions <a href="https://github.com/dgtlmoon/changedetection.io/wiki/Run-JavaScript-before-change-detection">More help and examples here</a>
|
||||||
|
</div>
|
||||||
</div>
|
</div>
|
||||||
{% endif %}
|
|
||||||
</fieldset>
|
</fieldset>
|
||||||
<fieldset class="pure-group" id="requests-override-options">
|
<fieldset class="pure-group" id="requests-override-options">
|
||||||
|
{% if not playwright_enabled %}
|
||||||
<div class="pure-form-message-inline">
|
<div class="pure-form-message-inline">
|
||||||
<strong>Request override is currently only used by the <i>Basic fast Plaintext/HTTP Client</i> method.</strong>
|
<strong>Request override is currently only used by the <i>Basic fast Plaintext/HTTP Client</i> method.</strong>
|
||||||
</div>
|
</div>
|
||||||
<div class="pure-control-group">
|
{% endif %}
|
||||||
|
<div class="pure-control-group" id="request-method">
|
||||||
{{ render_field(form.method) }}
|
{{ render_field(form.method) }}
|
||||||
</div>
|
</div>
|
||||||
<div class="pure-control-group">
|
<div class="pure-control-group" id="request-headers">
|
||||||
{{ render_field(form.headers, rows=5, placeholder="Example
|
{{ render_field(form.headers, rows=5, placeholder="Example
|
||||||
Cookie: foobar
|
Cookie: foobar
|
||||||
User-Agent: wonderbra 1.0") }}
|
User-Agent: wonderbra 1.0") }}
|
||||||
</div>
|
</div>
|
||||||
<div class="pure-control-group">
|
<div class="pure-control-group" id="request-body">
|
||||||
{{ render_field(form.body, rows=5, placeholder="Example
|
{{ render_field(form.body, rows=5, placeholder="Example
|
||||||
{
|
{
|
||||||
\"name\":\"John\",
|
\"name\":\"John\",
|
||||||
@@ -108,24 +133,29 @@ User-Agent: wonderbra 1.0") }}
|
|||||||
\"car\":null
|
\"car\":null
|
||||||
}") }}
|
}") }}
|
||||||
</div>
|
</div>
|
||||||
<div>
|
|
||||||
{{ render_checkbox_field(form.ignore_status_codes) }}
|
|
||||||
</div>
|
|
||||||
</fieldset>
|
</fieldset>
|
||||||
<br/>
|
|
||||||
</div>
|
</div>
|
||||||
|
|
||||||
<div class="tab-pane-inner" id="notifications">
|
<div class="tab-pane-inner" id="notifications">
|
||||||
<strong>Note: <i>These settings override the global settings for this watch.</i></strong>
|
|
||||||
<fieldset>
|
<fieldset>
|
||||||
<div class="field-group">
|
<div class="pure-control-group inline-radio">
|
||||||
{{ render_common_settings_form(form, current_base_url, emailprefix) }}
|
{{ render_checkbox_field(form.notification_muted) }}
|
||||||
|
</div>
|
||||||
|
<div class="field-group" id="notification-field-group">
|
||||||
|
{% if has_default_notification_urls %}
|
||||||
|
<div class="inline-warning">
|
||||||
|
<img class="inline-warning-icon" src="{{url_for('static_content', group='images', filename='notice.svg')}}" alt="Look out!" title="Lookout!"/>
|
||||||
|
There are <a href="{{ url_for('settings_page')}}#notifications">system-wide notification URLs enabled</a>, this form will override notification settings for this watch only ‐ an empty Notification URL list here will still send notifications.
|
||||||
|
</div>
|
||||||
|
{% endif %}
|
||||||
|
<a href="#notifications" id="notification-setting-reset-to-default" class="pure-button button-xsmall" style="right: 20px; top: 20px; position: absolute; background-color: #5f42dd; border-radius: 4px; font-size: 70%; color: #fff">Use system defaults</a>
|
||||||
|
|
||||||
|
{{ render_common_settings_form(form, emailprefix, settings_application) }}
|
||||||
</div>
|
</div>
|
||||||
</fieldset>
|
</fieldset>
|
||||||
</div>
|
</div>
|
||||||
|
|
||||||
<div class="tab-pane-inner" id="filters-and-triggers">
|
<div class="tab-pane-inner" id="filters-and-triggers">
|
||||||
<fieldset>
|
|
||||||
<div class="pure-control-group">
|
<div class="pure-control-group">
|
||||||
<strong>Pro-tips:</strong><br/>
|
<strong>Pro-tips:</strong><br/>
|
||||||
<ul>
|
<ul>
|
||||||
@@ -137,23 +167,49 @@ User-Agent: wonderbra 1.0") }}
|
|||||||
</li>
|
</li>
|
||||||
</ul>
|
</ul>
|
||||||
</div>
|
</div>
|
||||||
|
<fieldset>
|
||||||
<div class="pure-control-group">
|
<div class="pure-control-group">
|
||||||
{{ render_field(form.css_filter, placeholder=".class-name or #some-id, or other CSS selector rule.",
|
{{ render_checkbox_field(form.check_unique_lines) }}
|
||||||
class="m-d") }}
|
<span class="pure-form-message-inline">Good for websites that just move the content around, and you want to know when NEW content is added, compares new lines against all history for this watch.</span>
|
||||||
<span class="pure-form-message-inline">
|
</div>
|
||||||
|
</fieldset>
|
||||||
|
<div class="pure-control-group">
|
||||||
|
{% set field = render_field(form.include_filters,
|
||||||
|
rows=5,
|
||||||
|
placeholder="#example
|
||||||
|
xpath://body/div/span[contains(@class, 'example-class')]",
|
||||||
|
class="m-d")
|
||||||
|
%}
|
||||||
|
{{ field }}
|
||||||
|
{% if '/text()' in field %}
|
||||||
|
<span class="pure-form-message-inline"><strong>Note!: //text() function does not work where the <element> contains <![CDATA[]]></strong></span><br/>
|
||||||
|
{% endif %}
|
||||||
|
<span class="pure-form-message-inline">One rule per line, <i>any</i> rules that matches will be used.<br/>
|
||||||
<ul>
|
<ul>
|
||||||
<li>CSS - Limit text to this CSS rule, only text matching this CSS rule is included.</li>
|
<li>CSS - Limit text to this CSS rule, only text matching this CSS rule is included.</li>
|
||||||
<li>JSON - Limit text to this JSON rule, using <a href="https://pypi.org/project/jsonpath-ng/">JSONPath</a>, prefix with <code>"json:"</code>, use <code>json:$</code> to force re-formatting if required, <a
|
<li>JSON - Limit text to this JSON rule, using either <a href="https://pypi.org/project/jsonpath-ng/" target="new">JSONPath</a> or <a href="https://stedolan.github.io/jq/" target="new">jq</a> (if installed).
|
||||||
href="https://jsonpath.com/" target="new">test your JSONPath here</a></li>
|
<ul>
|
||||||
<li>XPath - Limit text to this XPath rule, simply start with a forward-slash, example <code>//*[contains(@class, 'sametext')]</code> or <code>xpath://*[contains(@class, 'sametext')]</code>, <a
|
<li>JSONPath: Prefix with <code>json:</code>, use <code>json:$</code> to force re-formatting if required, <a href="https://jsonpath.com/" target="new">test your JSONPath here</a>.</li>
|
||||||
href="http://xpather.com/" target="new">test your XPath here</a></li>
|
{% if jq_support %}
|
||||||
|
<li>jq: Prefix with <code>jq:</code> and <a href="https://jqplay.org/" target="new">test your jq here</a>. Using <a href="https://stedolan.github.io/jq/" target="new">jq</a> allows for complex filtering and processing of JSON data with built-in functions, regex, filtering, and more. See examples and documentation <a href="https://stedolan.github.io/jq/manual/" target="new">here</a>.</li>
|
||||||
|
{% else %}
|
||||||
|
<li>jq support not installed</li>
|
||||||
|
{% endif %}
|
||||||
</ul>
|
</ul>
|
||||||
Please be sure that you thoroughly understand how to write CSS or JSONPath, XPath selector rules before filing an issue on GitHub! <a
|
</li>
|
||||||
|
<li>XPath - Limit text to this XPath rule, simply start with a forward-slash,
|
||||||
|
<ul>
|
||||||
|
<li>Example: <code>//*[contains(@class, 'sametext')]</code> or <code>xpath://*[contains(@class, 'sametext')]</code>, <a
|
||||||
|
href="http://xpather.com/" target="new">test your XPath here</a></li>
|
||||||
|
<li>Example: Get all titles from an RSS feed <code>//title/text()</code></li>
|
||||||
|
</ul>
|
||||||
|
</li>
|
||||||
|
</ul>
|
||||||
|
Please be sure that you thoroughly understand how to write CSS, JSONPath, XPath{% if jq_support %}, or jq selector{%endif%} rules before filing an issue on GitHub! <a
|
||||||
href="https://github.com/dgtlmoon/changedetection.io/wiki/CSS-Selector-help">here for more CSS selector help</a>.<br/>
|
href="https://github.com/dgtlmoon/changedetection.io/wiki/CSS-Selector-help">here for more CSS selector help</a>.<br/>
|
||||||
</span>
|
</span>
|
||||||
</div>
|
</div>
|
||||||
<fieldset class="pure-group">
|
<div class="pure-control-group">
|
||||||
{{ render_field(form.subtractive_selectors, rows=5, placeholder="header
|
{{ render_field(form.subtractive_selectors, rows=5, placeholder="header
|
||||||
footer
|
footer
|
||||||
nav
|
nav
|
||||||
@@ -164,8 +220,7 @@ nav
|
|||||||
<li> Add multiple elements or CSS selectors per line to ignore multiple parts of the HTML. </li>
|
<li> Add multiple elements or CSS selectors per line to ignore multiple parts of the HTML. </li>
|
||||||
</ul>
|
</ul>
|
||||||
</span>
|
</span>
|
||||||
</fieldset>
|
</div>
|
||||||
</fieldset>
|
|
||||||
<fieldset class="pure-group">
|
<fieldset class="pure-group">
|
||||||
{{ render_field(form.ignore_text, rows=5, placeholder="Some text to ignore in a line
|
{{ render_field(form.ignore_text, rows=5, placeholder="Some text to ignore in a line
|
||||||
/some.regex\d{2}/ for case-INsensitive regex
|
/some.regex\d{2}/ for case-INsensitive regex
|
||||||
@@ -173,7 +228,7 @@ nav
|
|||||||
<span class="pure-form-message-inline">
|
<span class="pure-form-message-inline">
|
||||||
<ul>
|
<ul>
|
||||||
<li>Each line processed separately, any line matching will be ignored (removed before creating the checksum)</li>
|
<li>Each line processed separately, any line matching will be ignored (removed before creating the checksum)</li>
|
||||||
<li>Regular Expression support, wrap the line in forward slash <code>/regex/</code></li>
|
<li>Regular Expression support, wrap the entire line in forward slash <code>/regex/</code></li>
|
||||||
<li>Changing this will affect the comparison checksum which may trigger an alert</li>
|
<li>Changing this will affect the comparison checksum which may trigger an alert</li>
|
||||||
<li>Use the preview/show current tab to see ignores</li>
|
<li>Use the preview/show current tab to see ignores</li>
|
||||||
</ul>
|
</ul>
|
||||||
@@ -195,16 +250,90 @@ nav
|
|||||||
</span>
|
</span>
|
||||||
</div>
|
</div>
|
||||||
</fieldset>
|
</fieldset>
|
||||||
|
<fieldset>
|
||||||
|
<div class="pure-control-group">
|
||||||
|
{{ render_field(form.text_should_not_be_present, rows=5, placeholder="For example: Out of stock
|
||||||
|
Sold out
|
||||||
|
Not in stock
|
||||||
|
Unavailable") }}
|
||||||
|
<span class="pure-form-message-inline">
|
||||||
|
<ul>
|
||||||
|
<li>Block change-detection while this text is on the page, all text and regex are tested <i>case-insensitive</i>, good for waiting for when a product is available again</li>
|
||||||
|
<li>Block text is processed from the result-text that comes out of any CSS/JSON Filters for this watch</li>
|
||||||
|
<li>All lines here must not exist (think of each line as "OR")</li>
|
||||||
|
<li>Note: Wrap in forward slash / to use regex example: <code>/foo\d/</code></li>
|
||||||
|
</ul>
|
||||||
|
</span>
|
||||||
|
</div>
|
||||||
|
</fieldset>
|
||||||
|
<fieldset>
|
||||||
|
<div class="pure-control-group">
|
||||||
|
{{ render_field(form.extract_text, rows=5, placeholder="\d+ online") }}
|
||||||
|
<span class="pure-form-message-inline">
|
||||||
|
<ul>
|
||||||
|
<li>Extracts text in the final output (line by line) after other filters using regular expressions;
|
||||||
|
<ul>
|
||||||
|
<li>Regular expression ‐ example <code>/reports.+?2022/i</code></li>
|
||||||
|
<li>Use <code>//(?aiLmsux))</code> type flags (more <a href="https://docs.python.org/3/library/re.html#index-15">information here</a>)<br/></li>
|
||||||
|
<li>Keyword example ‐ example <code>Out of stock</code></li>
|
||||||
|
<li>Use groups to extract just that text ‐ example <code>/reports.+?(\d+)/i</code> returns a list of years only</li>
|
||||||
|
</ul>
|
||||||
|
</li>
|
||||||
|
<li>One line per regular-expression/ string match</li>
|
||||||
|
</ul>
|
||||||
|
</span>
|
||||||
|
</div>
|
||||||
|
</fieldset>
|
||||||
|
</div>
|
||||||
|
|
||||||
|
<div class="tab-pane-inner visual-selector-ui" id="visualselector">
|
||||||
|
<img id="beta-logo" src="{{url_for('static_content', group='images', filename='beta-logo.png')}}">
|
||||||
|
<strong>Pro-tip:</strong> This tool is only for limiting which elements will be included on a change-detection, not for interacting with browser directly.
|
||||||
|
<fieldset>
|
||||||
|
<div class="pure-control-group">
|
||||||
|
{% if visualselector_enabled %}
|
||||||
|
{% if visualselector_data_is_ready %}
|
||||||
|
<div id="selector-header">
|
||||||
|
<a id="clear-selector" class="pure-button button-secondary button-xsmall" style="font-size: 70%">Clear selection</a>
|
||||||
|
<i class="fetching-update-notice" style="font-size: 80%;">One moment, fetching screenshot and element information..</i>
|
||||||
|
</div>
|
||||||
|
<div id="selector-wrapper">
|
||||||
|
<!-- request the screenshot and get the element offset info ready -->
|
||||||
|
<!-- use img src ready load to know everything is ready to map out -->
|
||||||
|
<!-- @todo: maybe something interesting like a field to select 'elements that contain text... and their parents n' -->
|
||||||
|
<img id="selector-background" />
|
||||||
|
<canvas id="selector-canvas"></canvas>
|
||||||
|
|
||||||
|
</div>
|
||||||
|
<div id="selector-current-xpath" style="overflow-x: hidden"><strong>Currently:</strong> <span class="text">Loading...</span></div>
|
||||||
|
|
||||||
|
<span class="pure-form-message-inline">
|
||||||
|
<p><span style="font-weight: bold">Beta!</span> The Visual Selector is new and there may be minor bugs, please report pages that dont work, help us to improve this software!</p>
|
||||||
|
</span>
|
||||||
|
|
||||||
|
{% else %}
|
||||||
|
<span class="pure-form-message-inline">Screenshot and element data is not available or not yet ready.</span>
|
||||||
|
{% endif %}
|
||||||
|
{% else %}
|
||||||
|
<span class="pure-form-message-inline">
|
||||||
|
<p>Sorry, this functionality only works with Playwright/Chrome enabled watches.</p>
|
||||||
|
<p>Enable the Playwright Chrome fetcher, or alternatively try our <a href="https://lemonade.changedetection.io/start">very affordable subscription based service</a>.</p>
|
||||||
|
<p>This is because Selenium/WebDriver can not extract full page screenshots reliably.</p>
|
||||||
|
|
||||||
|
</span>
|
||||||
|
{% endif %}
|
||||||
|
</div>
|
||||||
|
</fieldset>
|
||||||
</div>
|
</div>
|
||||||
|
|
||||||
<div id="actions">
|
<div id="actions">
|
||||||
<div class="pure-control-group">
|
<div class="pure-control-group">
|
||||||
|
{{ render_button(form.save_button) }}
|
||||||
{{ render_button(form.save_button) }} {{ render_button(form.save_and_preview_button) }}
|
<a href="{{url_for('form_delete', uuid=uuid)}}"
|
||||||
|
|
||||||
<a href="{{url_for('api_delete', uuid=uuid)}}"
|
|
||||||
class="pure-button button-small button-error ">Delete</a>
|
class="pure-button button-small button-error ">Delete</a>
|
||||||
<a href="{{url_for('api_clone', uuid=uuid)}}"
|
<a href="{{url_for('clear_watch_history', uuid=uuid)}}"
|
||||||
|
class="pure-button button-small button-error ">Clear History</a>
|
||||||
|
<a href="{{url_for('form_clone', uuid=uuid)}}"
|
||||||
class="pure-button button-small ">Create Copy</a>
|
class="pure-button button-small ">Create Copy</a>
|
||||||
</div>
|
</div>
|
||||||
</div>
|
</div>
|
||||||
|
|||||||
@@ -5,7 +5,7 @@
|
|||||||
|
|
||||||
<div class="tabs collapsable">
|
<div class="tabs collapsable">
|
||||||
<ul>
|
<ul>
|
||||||
<li class="tab" id="default-tab"><a href="#url-list">URL List</a></li>
|
<li class="tab" id=""><a href="#url-list">URL List</a></li>
|
||||||
<li class="tab"><a href="#distill-io">Distill.io</a></li>
|
<li class="tab"><a href="#distill-io">Distill.io</a></li>
|
||||||
</ul>
|
</ul>
|
||||||
</div>
|
</div>
|
||||||
|
|||||||
@@ -4,7 +4,7 @@
|
|||||||
<div class="edit-form">
|
<div class="edit-form">
|
||||||
<div class="inner">
|
<div class="inner">
|
||||||
|
|
||||||
<h4 style="margin-top: 0px;">The following issues were detected when sending notifications</h4>
|
<h4 style="margin-top: 0px;">Notification debug log</h4>
|
||||||
<div id="notification-error-log">
|
<div id="notification-error-log">
|
||||||
<ul style="font-size: 80%; margin:0px; padding: 0 0 0 7px">
|
<ul style="font-size: 80%; margin:0px; padding: 0 0 0 7px">
|
||||||
{% for log in logs|reverse %}
|
{% for log in logs|reverse %}
|
||||||
|
|||||||
@@ -1,23 +1,41 @@
|
|||||||
{% extends 'base.html' %}
|
{% extends 'base.html' %}
|
||||||
|
|
||||||
{% block content %}
|
{% block content %}
|
||||||
|
<script>
|
||||||
<div id="settings">
|
const screenshot_url="{{url_for('static_content', group='screenshot', filename=uuid)}}";
|
||||||
<h1>Current - {{watch.last_checked|format_timestamp_timeago}}</h1>
|
{% if last_error_screenshot %}
|
||||||
</div>
|
const error_screenshot_url="{{url_for('static_content', group='screenshot', filename=uuid, error_screenshot=1) }}";
|
||||||
|
{% endif %}
|
||||||
|
</script>
|
||||||
|
<script type="text/javascript" src="{{url_for('static_content', group='js', filename='diff-overview.js')}}" defer></script>
|
||||||
|
|
||||||
<script type="text/javascript" src="{{url_for('static_content', group='js', filename='tabs.js')}}" defer></script>
|
<script type="text/javascript" src="{{url_for('static_content', group='js', filename='tabs.js')}}" defer></script>
|
||||||
<div class="tabs">
|
<div class="tabs">
|
||||||
<ul>
|
<ul>
|
||||||
<li class="tab" id="default-tab"><a href="#text">Text</a></li>
|
{% if last_error_text %}<li class="tab" id="error-text-tab"><a href="#error-text">Error Text</a></li> {% endif %}
|
||||||
{% if screenshot %}
|
{% if last_error_screenshot %}<li class="tab" id="error-screenshot-tab"><a href="#error-screenshot">Error Screenshot</a></li> {% endif %}
|
||||||
<li class="tab"><a href="#screenshot">Current screenshot</a></li>
|
{% if history_n > 0 %}
|
||||||
{% endif %}
|
<li class="tab" id="text-tab"><a href="#text">Text</a></li>
|
||||||
|
<li class="tab" id="screenshot-tab"><a href="#screenshot">Screenshot</a></li>
|
||||||
|
{% endif %}
|
||||||
</ul>
|
</ul>
|
||||||
</div>
|
</div>
|
||||||
|
|
||||||
<div id="diff-ui">
|
<div id="diff-ui">
|
||||||
|
<div class="tab-pane-inner" id="error-text">
|
||||||
|
<div class="snapshot-age error">{{watch.error_text_ctime|format_seconds_ago}} seconds ago</div>
|
||||||
|
<pre>
|
||||||
|
{{ last_error_text }}
|
||||||
|
</pre>
|
||||||
|
</div>
|
||||||
|
|
||||||
|
<div class="tab-pane-inner" id="error-screenshot">
|
||||||
|
<div class="snapshot-age error">{{watch.snapshot_error_screenshot_ctime|format_seconds_ago}} seconds ago</div>
|
||||||
|
<img id="error-screenshot-img" style="max-width: 80%" alt="Current erroring screenshot from most recent request"/>
|
||||||
|
</div>
|
||||||
|
|
||||||
<div class="tab-pane-inner" id="text">
|
<div class="tab-pane-inner" id="text">
|
||||||
|
<div class="snapshot-age">{{watch.snapshot_text_ctime|format_timestamp_timeago}}</div>
|
||||||
<span class="ignored">Grey lines are ignored</span> <span class="triggered">Blue lines are triggers</span>
|
<span class="ignored">Grey lines are ignored</span> <span class="triggered">Blue lines are triggers</span>
|
||||||
<table>
|
<table>
|
||||||
<tbody>
|
<tbody>
|
||||||
@@ -32,14 +50,21 @@
|
|||||||
</table>
|
</table>
|
||||||
</div>
|
</div>
|
||||||
|
|
||||||
{% if screenshot %}
|
|
||||||
<div class="tab-pane-inner" id="screenshot">
|
<div class="tab-pane-inner" id="screenshot">
|
||||||
<p>
|
<div class="tip">
|
||||||
<i>For now, only the most recent screenshot is saved and displayed.</i>
|
For now, Differences are performed on text, not graphically, only the latest screenshot is available.
|
||||||
</p>
|
</div>
|
||||||
|
</br>
|
||||||
<img src="{{url_for('static_content', group='screenshot', filename=uuid)}}">
|
{% if is_html_webdriver %}
|
||||||
|
{% if screenshot %}
|
||||||
|
<div class="snapshot-age">{{watch.snapshot_screenshot_ctime|format_timestamp_timeago}}</div>
|
||||||
|
<img style="max-width: 80%" id="screenshot-img" alt="Current screenshot from most recent request"/>
|
||||||
|
{% else %}
|
||||||
|
No screenshot available just yet! Try rechecking the page.
|
||||||
|
{% endif %}
|
||||||
|
{% else %}
|
||||||
|
<strong>Screenshot requires Playwright/WebDriver enabled</strong>
|
||||||
|
{% endif %}
|
||||||
</div>
|
</div>
|
||||||
{% endif %}
|
|
||||||
</div>
|
</div>
|
||||||
{% endblock %}
|
{% endblock %}
|
||||||
@@ -16,10 +16,11 @@
|
|||||||
<div class="edit-form">
|
<div class="edit-form">
|
||||||
<div class="tabs collapsable">
|
<div class="tabs collapsable">
|
||||||
<ul>
|
<ul>
|
||||||
<li class="tab" id="default-tab"><a href="#general">General</a></li>
|
<li class="tab" id=""><a href="#general">General</a></li>
|
||||||
<li class="tab"><a href="#notifications">Notifications</a></li>
|
<li class="tab"><a href="#notifications">Notifications</a></li>
|
||||||
<li class="tab"><a href="#fetching">Fetching</a></li>
|
<li class="tab"><a href="#fetching">Fetching</a></li>
|
||||||
<li class="tab"><a href="#filters">Global Filters</a></li>
|
<li class="tab"><a href="#filters">Global Filters</a></li>
|
||||||
|
<li class="tab"><a href="#api">API</a></li>
|
||||||
</ul>
|
</ul>
|
||||||
</div>
|
</div>
|
||||||
<div class="box-wrap inner">
|
<div class="box-wrap inner">
|
||||||
@@ -31,6 +32,17 @@
|
|||||||
{{ render_field(form.requests.form.time_between_check, class="time-check-widget") }}
|
{{ render_field(form.requests.form.time_between_check, class="time-check-widget") }}
|
||||||
<span class="pure-form-message-inline">Default time for all watches, when the watch does not have a specific time setting.</span>
|
<span class="pure-form-message-inline">Default time for all watches, when the watch does not have a specific time setting.</span>
|
||||||
</div>
|
</div>
|
||||||
|
<div class="pure-control-group">
|
||||||
|
{{ render_field(form.requests.form.jitter_seconds, class="jitter_seconds") }}
|
||||||
|
<span class="pure-form-message-inline">Example - 3 seconds random jitter could trigger up to 3 seconds earlier or up to 3 seconds later</span>
|
||||||
|
</div>
|
||||||
|
<div class="pure-control-group">
|
||||||
|
{{ render_field(form.application.form.filter_failure_notification_threshold_attempts, class="filter_failure_notification_threshold_attempts") }}
|
||||||
|
<span class="pure-form-message-inline">After this many consecutive times that the CSS/xPath filter is missing, send a notification
|
||||||
|
<br/>
|
||||||
|
Set to <strong>0</strong> to disable
|
||||||
|
</span>
|
||||||
|
</div>
|
||||||
<div class="pure-control-group">
|
<div class="pure-control-group">
|
||||||
{% if not hide_remove_pass %}
|
{% if not hide_remove_pass %}
|
||||||
{% if current_user.is_authenticated %}
|
{% if current_user.is_authenticated %}
|
||||||
@@ -43,11 +55,12 @@
|
|||||||
<span class="pure-form-message-inline">Password is locked.</span>
|
<span class="pure-form-message-inline">Password is locked.</span>
|
||||||
{% endif %}
|
{% endif %}
|
||||||
</div>
|
</div>
|
||||||
|
|
||||||
<div class="pure-control-group">
|
<div class="pure-control-group">
|
||||||
{{ render_field(form.application.form.base_url, placeholder="http://yoursite.com:5000/",
|
{{ render_field(form.application.form.base_url, placeholder="http://yoursite.com:5000/",
|
||||||
class="m-d") }}
|
class="m-d") }}
|
||||||
<span class="pure-form-message-inline">
|
<span class="pure-form-message-inline">
|
||||||
Base URL used for the {base_url} token in notifications and RSS links.<br/>Default value is the ENV var 'BASE_URL' (Currently "{{current_base_url}}"),
|
Base URL used for the <code>{base_url}</code> token in notifications and RSS links.<br/>Default value is the ENV var 'BASE_URL' (Currently "{{settings_application['current_base_url']}}"),
|
||||||
<a href="https://github.com/dgtlmoon/changedetection.io/wiki/Configurable-BASE_URL-setting">read more here</a>.
|
<a href="https://github.com/dgtlmoon/changedetection.io/wiki/Configurable-BASE_URL-setting">read more here</a>.
|
||||||
</span>
|
</span>
|
||||||
</div>
|
</div>
|
||||||
@@ -56,10 +69,9 @@
|
|||||||
{{ render_checkbox_field(form.application.form.extract_title_as_title) }}
|
{{ render_checkbox_field(form.application.form.extract_title_as_title) }}
|
||||||
<span class="pure-form-message-inline">Note: This will automatically apply to all existing watches.</span>
|
<span class="pure-form-message-inline">Note: This will automatically apply to all existing watches.</span>
|
||||||
</div>
|
</div>
|
||||||
|
|
||||||
<div class="pure-control-group">
|
<div class="pure-control-group">
|
||||||
{{ render_checkbox_field(form.application.form.real_browser_save_screenshot) }}
|
{{ render_checkbox_field(form.application.form.empty_pages_are_a_change) }}
|
||||||
<span class="pure-form-message-inline">When using a Chrome browser, a screenshot from the last check will be available on the Diff page</span>
|
<span class="pure-form-message-inline">When a page contains HTML, but no renderable text appears (empty page), is this considered a change?</span>
|
||||||
</div>
|
</div>
|
||||||
{% if form.requests.proxy %}
|
{% if form.requests.proxy %}
|
||||||
<div class="pure-control-group inline-radio">
|
<div class="pure-control-group inline-radio">
|
||||||
@@ -75,7 +87,7 @@
|
|||||||
<div class="tab-pane-inner" id="notifications">
|
<div class="tab-pane-inner" id="notifications">
|
||||||
<fieldset>
|
<fieldset>
|
||||||
<div class="field-group">
|
<div class="field-group">
|
||||||
{{ render_common_settings_form(form.application.form, current_base_url, emailprefix) }}
|
{{ render_common_settings_form(form.application.form, emailprefix, settings_application) }}
|
||||||
</div>
|
</div>
|
||||||
</fieldset>
|
</fieldset>
|
||||||
</div>
|
</div>
|
||||||
@@ -87,6 +99,8 @@
|
|||||||
<p>Use the <strong>Basic</strong> method (default) where your watched sites don't need Javascript to render.</p>
|
<p>Use the <strong>Basic</strong> method (default) where your watched sites don't need Javascript to render.</p>
|
||||||
<p>The <strong>Chrome/Javascript</strong> method requires a network connection to a running WebDriver+Chrome server, set by the ENV var 'WEBDRIVER_URL'. </p>
|
<p>The <strong>Chrome/Javascript</strong> method requires a network connection to a running WebDriver+Chrome server, set by the ENV var 'WEBDRIVER_URL'. </p>
|
||||||
</span>
|
</span>
|
||||||
|
<br/>
|
||||||
|
Tip: <a href="https://github.com/dgtlmoon/changedetection.io/wiki/Proxy-configuration#brightdata-proxy-support">Connect using BrightData Proxies, find out more here.</a>
|
||||||
</div>
|
</div>
|
||||||
<fieldset class="pure-group" id="webdriver-override-options">
|
<fieldset class="pure-group" id="webdriver-override-options">
|
||||||
<div class="pure-form-message-inline">
|
<div class="pure-form-message-inline">
|
||||||
@@ -100,7 +114,6 @@
|
|||||||
</fieldset>
|
</fieldset>
|
||||||
</div>
|
</div>
|
||||||
|
|
||||||
|
|
||||||
<div class="tab-pane-inner" id="filters">
|
<div class="tab-pane-inner" id="filters">
|
||||||
|
|
||||||
<fieldset class="pure-group">
|
<fieldset class="pure-group">
|
||||||
@@ -137,7 +150,7 @@ nav
|
|||||||
<ul>
|
<ul>
|
||||||
<li>Note: This is applied globally in addition to the per-watch rules.</li>
|
<li>Note: This is applied globally in addition to the per-watch rules.</li>
|
||||||
<li>Each line processed separately, any line matching will be ignored (removed before creating the checksum)</li>
|
<li>Each line processed separately, any line matching will be ignored (removed before creating the checksum)</li>
|
||||||
<li>Regular Expression support, wrap the line in forward slash <code>/regex/</code></li>
|
<li>Regular Expression support, wrap the entire line in forward slash <code>/regex/</code></li>
|
||||||
<li>Changing this will affect the comparison checksum which may trigger an alert</li>
|
<li>Changing this will affect the comparison checksum which may trigger an alert</li>
|
||||||
<li>Use the preview/show current tab to see ignores</li>
|
<li>Use the preview/show current tab to see ignores</li>
|
||||||
</ul>
|
</ul>
|
||||||
@@ -145,12 +158,26 @@ nav
|
|||||||
</fieldset>
|
</fieldset>
|
||||||
</div>
|
</div>
|
||||||
|
|
||||||
|
<div class="tab-pane-inner" id="api">
|
||||||
|
|
||||||
|
<p>Drive your changedetection.io via API, More about <a href="https://github.com/dgtlmoon/changedetection.io/wiki/API-Reference">API access here</a></p>
|
||||||
|
|
||||||
|
<div class="pure-control-group">
|
||||||
|
{{ render_checkbox_field(form.application.form.api_access_token_enabled) }}
|
||||||
|
<div class="pure-form-message-inline">Restrict API access limit by using <code>x-api-key</code> header</div><br/>
|
||||||
|
<div class="pure-form-message-inline"><br/>API Key <span id="api-key">{{api_key}}</span>
|
||||||
|
<span style="display:none;" id="api-key-copy" >copy</span>
|
||||||
|
</div>
|
||||||
|
</div>
|
||||||
|
</div>
|
||||||
|
|
||||||
<div id="actions">
|
<div id="actions">
|
||||||
<div class="pure-control-group">
|
<div class="pure-control-group">
|
||||||
{{ render_button(form.save_button) }}
|
{{ render_button(form.save_button) }}
|
||||||
<a href="{{url_for('index')}}" class="pure-button button-small button-cancel">Back</a>
|
<a href="{{url_for('index')}}" class="pure-button button-small button-cancel">Back</a>
|
||||||
<a href="{{url_for('scrub_page')}}" class="pure-button button-small button-cancel">Delete History Snapshot Data</a>
|
<a href="{{url_for('clear_all_history')}}" class="pure-button button-small button-cancel">Clear Snapshot History</a>
|
||||||
</div>
|
</div>
|
||||||
|
|
||||||
</div>
|
</div>
|
||||||
</form>
|
</form>
|
||||||
</div>
|
</div>
|
||||||
|
|||||||
@@ -1,20 +1,40 @@
|
|||||||
{% extends 'base.html' %}
|
{% extends 'base.html' %}
|
||||||
{% block content %}
|
{% block content %}
|
||||||
{% from '_helpers.jinja' import render_simple_field %}
|
{% from '_helpers.jinja' import render_simple_field, render_field %}
|
||||||
|
{% from '_pagination.jinja' import pagination %}
|
||||||
<script type="text/javascript" src="{{url_for('static_content', group='js', filename='jquery-3.6.0.min.js')}}"></script>
|
<script type="text/javascript" src="{{url_for('static_content', group='js', filename='jquery-3.6.0.min.js')}}"></script>
|
||||||
<script type="text/javascript" src="{{url_for('static_content', group='js', filename='watch-overview.js')}}" defer></script>
|
<script type="text/javascript" src="{{url_for('static_content', group='js', filename='watch-overview.js')}}" defer></script>
|
||||||
|
|
||||||
<div class="box">
|
<div class="box">
|
||||||
|
|
||||||
<form class="pure-form" action="{{ url_for('api_watch_add') }}" method="POST" id="new-watch-form">
|
<form class="pure-form" action="{{ url_for('form_quick_watch_add') }}" method="POST" id="new-watch-form">
|
||||||
<input type="hidden" name="csrf_token" value="{{ csrf_token() }}"/>
|
<input type="hidden" name="csrf_token" value="{{ csrf_token() }}"/>
|
||||||
<fieldset>
|
<fieldset>
|
||||||
<legend>Add a new change detection watch</legend>
|
<legend>Add a new change detection watch</legend>
|
||||||
|
<div id="watch-add-wrapper-zone">
|
||||||
|
<div>
|
||||||
{{ render_simple_field(form.url, placeholder="https://...", required=true) }}
|
{{ render_simple_field(form.url, placeholder="https://...", required=true) }}
|
||||||
{{ render_simple_field(form.tag, value=active_tag if active_tag else '', placeholder="watch group") }}
|
{{ render_simple_field(form.tag, value=active_tag if active_tag else '', placeholder="watch group") }}
|
||||||
<button type="submit" class="pure-button pure-button-primary">Watch</button>
|
</div>
|
||||||
|
<div>
|
||||||
|
{{ render_simple_field(form.watch_submit_button, title="Watch this URL!" ) }}
|
||||||
|
{{ render_simple_field(form.edit_and_watch_submit_button, title="Edit first then Watch") }}
|
||||||
|
</div>
|
||||||
|
</div>
|
||||||
</fieldset>
|
</fieldset>
|
||||||
<span style="color:#eee; font-size: 80%;"><img style="height: 1em;display:inline-block;" src="{{url_for('static_content', group='images', filename='spread.svg')}}" /> Tip: You can also add 'shared' watches. <a href="#">More info</a></a></span>
|
<span style="color:#eee; font-size: 80%;"><img style="height: 1em;display:inline-block;" src="{{url_for('static_content', group='images', filename='spread-white.svg')}}" /> Tip: You can also add 'shared' watches. <a href="https://github.com/dgtlmoon/changedetection.io/wiki/Sharing-a-Watch">More info</a></a></span>
|
||||||
</form>
|
</form>
|
||||||
|
|
||||||
|
<form class="pure-form" action="{{ url_for('form_watch_list_checkbox_operations') }}" method="POST" id="watch-list-form">
|
||||||
|
<input type="hidden" name="csrf_token" value="{{ csrf_token() }}"/>
|
||||||
|
<div id="checkbox-operations">
|
||||||
|
<button class="pure-button button-secondary button-xsmall" style="font-size: 70%" name="op" value="pause">Pause</button>
|
||||||
|
<button class="pure-button button-secondary button-xsmall" style="font-size: 70%" name="op" value="unpause">UnPause</button>
|
||||||
|
<button class="pure-button button-secondary button-xsmall" style="font-size: 70%" name="op" value="mute">Mute</button>
|
||||||
|
<button class="pure-button button-secondary button-xsmall" style="font-size: 70%" name="op" value="unmute">UnMute</button>
|
||||||
|
<button class="pure-button button-secondary button-xsmall" style="font-size: 70%" name="op" value="notification-default">Use default notification</button>
|
||||||
|
<button class="pure-button button-secondary button-xsmall" style="background: #dd4242; font-size: 70%" name="op" value="delete">Delete</button>
|
||||||
|
</div>
|
||||||
<div>
|
<div>
|
||||||
<a href="{{url_for('index')}}" class="pure-button button-tag {{'active' if not active_tag }}">All</a>
|
<a href="{{url_for('index')}}" class="pure-button button-tag {{'active' if not active_tag }}">All</a>
|
||||||
{% for tag in tags %}
|
{% for tag in tags %}
|
||||||
@@ -24,35 +44,51 @@
|
|||||||
{% endfor %}
|
{% endfor %}
|
||||||
</div>
|
</div>
|
||||||
|
|
||||||
|
{% set sort_order = request.args.get('order', 'asc') == 'asc' %}
|
||||||
|
{% set sort_attribute = request.args.get('sort', 'last_changed') %}
|
||||||
|
{% set pagination_page = request.args.get('page', 0) %}
|
||||||
|
|
||||||
<div id="watch-table-wrapper">
|
<div id="watch-table-wrapper">
|
||||||
<table class="pure-table pure-table-striped watch-table">
|
<table class="pure-table pure-table-striped watch-table">
|
||||||
<thead>
|
<thead>
|
||||||
<tr>
|
<tr>
|
||||||
<th>#</th>
|
<th><input style="vertical-align: middle" type="checkbox" id="check-all"/> #</th>
|
||||||
<th></th>
|
<th></th>
|
||||||
<th></th>
|
{% set link_order = "desc" if sort_order else "asc" %}
|
||||||
<th>Last Checked</th>
|
{% set arrow_span = "" %}
|
||||||
<th>Last Changed</th>
|
<th><a class="{{ 'active '+link_order if sort_attribute == 'label' else 'inactive' }}" href="{{url_for('index', sort='label', order=link_order)}}">Website <span class='arrow {{link_order}}'></span></a></th>
|
||||||
|
<th><a class="{{ 'active '+link_order if sort_attribute == 'last_checked' else 'inactive' }}" href="{{url_for('index', sort='last_checked', order=link_order)}}">Last Checked <span class='arrow {{link_order}}'></span></a></th>
|
||||||
|
<th><a class="{{ 'active '+link_order if sort_attribute == 'last_changed' else 'inactive' }}" href="{{url_for('index', sort='last_changed', order=link_order)}}">Last Changed <span class='arrow {{link_order}}'></span></a></th>
|
||||||
<th></th>
|
<th></th>
|
||||||
</tr>
|
</tr>
|
||||||
</thead>
|
</thead>
|
||||||
<tbody>
|
<tbody>
|
||||||
|
|
||||||
|
{% set sorted_watches = watches|sort(attribute=sort_attribute, reverse=sort_order) %}
|
||||||
|
{% for watch in sorted_watches %}
|
||||||
|
|
||||||
{% for watch in watches %}
|
{# WIP for pagination, disabled for now
|
||||||
|
{% if not ( loop.index >= 3 and loop.index <=4) %}{% continue %}{% endif %} -->
|
||||||
|
#}
|
||||||
<tr id="{{ watch.uuid }}"
|
<tr id="{{ watch.uuid }}"
|
||||||
class="{{ loop.cycle('pure-table-odd', 'pure-table-even') }}
|
class="{{ loop.cycle('pure-table-odd', 'pure-table-even') }}
|
||||||
{% if watch.last_error is defined and watch.last_error != False %}error{% endif %}
|
{% if watch.last_error is defined and watch.last_error != False %}error{% endif %}
|
||||||
{% if watch.last_notification_error is defined and watch.last_notification_error != False %}error{% endif %}
|
{% if watch.last_notification_error is defined and watch.last_notification_error != False %}error{% endif %}
|
||||||
{% if watch.paused is defined and watch.paused != False %}paused{% endif %}
|
{% if watch.paused is defined and watch.paused != False %}paused{% endif %}
|
||||||
{% if watch.newest_history_key| int > watch.last_viewed| int %}unviewed{% endif %}
|
{% if watch.newest_history_key| int > watch.last_viewed and watch.history_n>=2 %}unviewed{% endif %}
|
||||||
{% if watch.uuid in queued_uuids %}queued{% endif %}">
|
{% if watch.uuid in queued_uuids %}queued{% endif %}">
|
||||||
<td class="inline">{{ loop.index }}</td>
|
<td class="inline checkbox-uuid" ><input name="uuids" type="checkbox" value="{{ watch.uuid}} "/> <span>{{ loop.index }}</span></td>
|
||||||
<td class="inline paused-state state-{{watch.paused}}"><a href="{{url_for('index', pause=watch.uuid, tag=active_tag)}}"><img src="{{url_for('static_content', group='images', filename='pause.svg')}}" alt="Pause" title="Pause"/></a></td>
|
<td class="inline watch-controls">
|
||||||
|
{% if not watch.paused %}
|
||||||
|
<a class="state-off" href="{{url_for('index', op='pause', uuid=watch.uuid, tag=active_tag)}}"><img src="{{url_for('static_content', group='images', filename='pause.svg')}}" alt="Pause checks" title="Pause checks"/></a>
|
||||||
|
{% else %}
|
||||||
|
<a class="state-on" href="{{url_for('index', op='pause', uuid=watch.uuid, tag=active_tag)}}"><img src="{{url_for('static_content', group='images', filename='play.svg')}}" alt="UnPause checks" title="UnPause checks"/></a>
|
||||||
|
{% endif %}
|
||||||
|
<a class="state-{{'on' if watch.notification_muted}}" href="{{url_for('index', op='mute', uuid=watch.uuid, tag=active_tag)}}"><img src="{{url_for('static_content', group='images', filename='bell-off.svg')}}" alt="Mute notifications" title="Mute notifications"/></a>
|
||||||
|
</td>
|
||||||
<td class="title-col inline">{{watch.title if watch.title is not none and watch.title|length > 0 else watch.url}}
|
<td class="title-col inline">{{watch.title if watch.title is not none and watch.title|length > 0 else watch.url}}
|
||||||
<a class="external" target="_blank" rel="noopener" href="{{ watch.url.replace('source:','') }}"></a>
|
<a class="external" target="_blank" rel="noopener" href="{{ watch.link.replace('source:','') }}"></a>
|
||||||
<a href="{{url_for('api_share_put_watch', uuid=watch.uuid)}}"><img style="height: 1em;display:inline-block;" src="{{url_for('static_content', group='images', filename='spread.svg')}}" /></a>
|
<a href="{{url_for('form_share_put_watch', uuid=watch.uuid)}}"><img style="height: 1em;display:inline-block;" src="{{url_for('static_content', group='images', filename='spread.svg')}}" /></a>
|
||||||
|
|
||||||
{%if watch.fetch_backend == "html_webdriver" %}<img style="height: 1em; display:inline-block;" src="{{url_for('static_content', group='images', filename='Google-Chrome-icon.png')}}" />{% endif %}
|
{%if watch.fetch_backend == "html_webdriver" %}<img style="height: 1em; display:inline-block;" src="{{url_for('static_content', group='images', filename='Google-Chrome-icon.png')}}" />{% endif %}
|
||||||
|
|
||||||
@@ -66,21 +102,21 @@
|
|||||||
<span class="watch-tag-list">{{ watch.tag}}</span>
|
<span class="watch-tag-list">{{ watch.tag}}</span>
|
||||||
{% endif %}
|
{% endif %}
|
||||||
</td>
|
</td>
|
||||||
<td class="last-checked">{{watch|format_last_checked_time}}</td>
|
<td class="last-checked">{{watch|format_last_checked_time|safe}}</td>
|
||||||
<td class="last-changed">{% if watch.history|length >= 2 and watch.last_changed %}
|
<td class="last-changed">{% if watch.history_n >=2 and watch.last_changed >0 %}
|
||||||
{{watch.last_changed|format_timestamp_timeago}}
|
{{watch.last_changed|format_timestamp_timeago}}
|
||||||
{% else %}
|
{% else %}
|
||||||
Not yet
|
Not yet
|
||||||
{% endif %}
|
{% endif %}
|
||||||
</td>
|
</td>
|
||||||
<td>
|
<td>
|
||||||
<a {% if watch.uuid in queued_uuids %}disabled="true"{% endif %} href="{{ url_for('api_watch_checknow', uuid=watch.uuid, tag=request.args.get('tag')) }}"
|
<a {% if watch.uuid in queued_uuids %}disabled="true"{% endif %} href="{{ url_for('form_watch_checknow', uuid=watch.uuid, tag=request.args.get('tag')) }}"
|
||||||
class="recheck pure-button button-small pure-button-primary">{% if watch.uuid in queued_uuids %}Queued{% else %}Recheck{% endif %}</a>
|
class="recheck pure-button button-small pure-button-primary">{% if watch.uuid in queued_uuids %}Queued{% else %}Recheck{% endif %}</a>
|
||||||
<a href="{{ url_for('edit_page', uuid=watch.uuid)}}" class="pure-button button-small pure-button-primary">Edit</a>
|
<a href="{{ url_for('edit_page', uuid=watch.uuid)}}" class="pure-button button-small pure-button-primary">Edit</a>
|
||||||
{% if watch.history|length >= 2 %}
|
{% if watch.history_n >= 2 %}
|
||||||
<a href="{{ url_for('diff_history_page', uuid=watch.uuid) }}" target="{{watch.uuid}}" class="pure-button button-small pure-button-primary diff-link">Diff</a>
|
<a href="{{ url_for('diff_history_page', uuid=watch.uuid) }}" target="{{watch.uuid}}" class="pure-button button-small pure-button-primary diff-link">Diff</a>
|
||||||
{% else %}
|
{% else %}
|
||||||
{% if watch.history|length == 1 %}
|
{% if watch.history_n == 1 or (watch.history_n ==0 and watch.error_text_ctime )%}
|
||||||
<a href="{{ url_for('preview_page', uuid=watch.uuid)}}" target="{{watch.uuid}}" class="pure-button button-small pure-button-primary">Preview</a>
|
<a href="{{ url_for('preview_page', uuid=watch.uuid)}}" target="{{watch.uuid}}" class="pure-button button-small pure-button-primary">Preview</a>
|
||||||
{% endif %}
|
{% endif %}
|
||||||
{% endif %}
|
{% endif %}
|
||||||
@@ -96,13 +132,18 @@
|
|||||||
</li>
|
</li>
|
||||||
{% endif %}
|
{% endif %}
|
||||||
<li>
|
<li>
|
||||||
<a href="{{ url_for('api_watch_checknow', tag=active_tag) }}" class="pure-button button-tag ">Recheck
|
<a href="{{ url_for('form_watch_checknow', tag=active_tag) }}" class="pure-button button-tag ">Recheck
|
||||||
all {% if active_tag%}in "{{active_tag}}"{%endif%}</a>
|
all {% if active_tag%}in "{{active_tag}}"{%endif%}</a>
|
||||||
</li>
|
</li>
|
||||||
<li>
|
<li>
|
||||||
<a href="{{ url_for('rss', tag=active_tag , token=app_rss_token)}}"><img alt="RSS Feed" id="feed-icon" src="{{url_for('static_content', group='images', filename='Generic_Feed-icon.svg')}}" height="15"></a>
|
<a href="{{ url_for('rss', tag=active_tag , token=app_rss_token)}}"><img alt="RSS Feed" id="feed-icon" src="{{url_for('static_content', group='images', filename='Generic_Feed-icon.svg')}}" height="15"></a>
|
||||||
</li>
|
</li>
|
||||||
</ul>
|
</ul>
|
||||||
|
{# WIP for pagination, disabled for now
|
||||||
|
{{ pagination(sorted_watches,3, pagination_page) }}
|
||||||
|
#}
|
||||||
|
|
||||||
</div>
|
</div>
|
||||||
|
</form>
|
||||||
</div>
|
</div>
|
||||||
{% endblock %}
|
{% endblock %}
|
||||||
|
|||||||
@@ -32,6 +32,8 @@ def app(request):
|
|||||||
"""Create application for the tests."""
|
"""Create application for the tests."""
|
||||||
datastore_path = "./test-datastore"
|
datastore_path = "./test-datastore"
|
||||||
|
|
||||||
|
# So they don't delay in fetching
|
||||||
|
os.environ["MINIMUM_SECONDS_RECHECK_TIME"] = "0"
|
||||||
try:
|
try:
|
||||||
os.mkdir(datastore_path)
|
os.mkdir(datastore_path)
|
||||||
except FileExistsError:
|
except FileExistsError:
|
||||||
@@ -39,7 +41,7 @@ def app(request):
|
|||||||
|
|
||||||
cleanup(datastore_path)
|
cleanup(datastore_path)
|
||||||
|
|
||||||
app_config = {'datastore_path': datastore_path}
|
app_config = {'datastore_path': datastore_path, 'disable_checkver' : True}
|
||||||
cleanup(app_config['datastore_path'])
|
cleanup(app_config['datastore_path'])
|
||||||
datastore = store.ChangeDetectionStore(datastore_path=app_config['datastore_path'], include_default_watches=False)
|
datastore = store.ChangeDetectionStore(datastore_path=app_config['datastore_path'], include_default_watches=False)
|
||||||
app = changedetection_app(app_config, datastore)
|
app = changedetection_app(app_config, datastore)
|
||||||
|
|||||||
2
changedetectionio/tests/fetchers/__init__.py
Normal file
2
changedetectionio/tests/fetchers/__init__.py
Normal file
@@ -0,0 +1,2 @@
|
|||||||
|
"""Tests for the app."""
|
||||||
|
|
||||||
3
changedetectionio/tests/fetchers/conftest.py
Normal file
3
changedetectionio/tests/fetchers/conftest.py
Normal file
@@ -0,0 +1,3 @@
|
|||||||
|
#!/usr/bin/python3
|
||||||
|
|
||||||
|
from .. import conftest
|
||||||
42
changedetectionio/tests/fetchers/test_content.py
Normal file
42
changedetectionio/tests/fetchers/test_content.py
Normal file
@@ -0,0 +1,42 @@
|
|||||||
|
#!/usr/bin/python3
|
||||||
|
|
||||||
|
import time
|
||||||
|
from flask import url_for
|
||||||
|
from ..util import live_server_setup, wait_for_all_checks
|
||||||
|
import logging
|
||||||
|
|
||||||
|
|
||||||
|
def test_fetch_webdriver_content(client, live_server):
|
||||||
|
live_server_setup(live_server)
|
||||||
|
|
||||||
|
#####################
|
||||||
|
res = client.post(
|
||||||
|
url_for("settings_page"),
|
||||||
|
data={"application-empty_pages_are_a_change": "",
|
||||||
|
"requests-time_between_check-minutes": 180,
|
||||||
|
'application-fetch_backend': "html_webdriver"},
|
||||||
|
follow_redirects=True
|
||||||
|
)
|
||||||
|
|
||||||
|
assert b"Settings updated." in res.data
|
||||||
|
|
||||||
|
# Add our URL to the import page
|
||||||
|
res = client.post(
|
||||||
|
url_for("import_page"),
|
||||||
|
data={"urls": "https://changedetection.io/ci-test.html"},
|
||||||
|
follow_redirects=True
|
||||||
|
)
|
||||||
|
|
||||||
|
assert b"1 Imported" in res.data
|
||||||
|
time.sleep(3)
|
||||||
|
|
||||||
|
wait_for_all_checks(client)
|
||||||
|
|
||||||
|
|
||||||
|
res = client.get(
|
||||||
|
url_for("preview_page", uuid="first"),
|
||||||
|
follow_redirects=True
|
||||||
|
)
|
||||||
|
logging.getLogger().info("Looking for correct fetched HTML (text) from server")
|
||||||
|
|
||||||
|
assert b'cool it works' in res.data
|
||||||
2
changedetectionio/tests/proxy_list/__init__.py
Normal file
2
changedetectionio/tests/proxy_list/__init__.py
Normal file
@@ -0,0 +1,2 @@
|
|||||||
|
"""Tests for the app."""
|
||||||
|
|
||||||
14
changedetectionio/tests/proxy_list/conftest.py
Normal file
14
changedetectionio/tests/proxy_list/conftest.py
Normal file
@@ -0,0 +1,14 @@
|
|||||||
|
#!/usr/bin/python3
|
||||||
|
|
||||||
|
from .. import conftest
|
||||||
|
|
||||||
|
#def pytest_addoption(parser):
|
||||||
|
# parser.addoption("--url_suffix", action="store", default="identifier for request")
|
||||||
|
|
||||||
|
|
||||||
|
#def pytest_generate_tests(metafunc):
|
||||||
|
# # This is called for every test. Only get/set command line arguments
|
||||||
|
# # if the argument is specified in the list of test "fixturenames".
|
||||||
|
# option_value = metafunc.config.option.url_suffix
|
||||||
|
# if 'url_suffix' in metafunc.fixturenames and option_value is not None:
|
||||||
|
# metafunc.parametrize("url_suffix", [option_value])
|
||||||
10
changedetectionio/tests/proxy_list/proxies.json-example
Normal file
10
changedetectionio/tests/proxy_list/proxies.json-example
Normal file
@@ -0,0 +1,10 @@
|
|||||||
|
{
|
||||||
|
"proxy-one": {
|
||||||
|
"label": "One",
|
||||||
|
"url": "http://127.0.0.1:3128"
|
||||||
|
},
|
||||||
|
"proxy-two": {
|
||||||
|
"label": "two",
|
||||||
|
"url": "http://127.0.0.1:3129"
|
||||||
|
}
|
||||||
|
}
|
||||||
41
changedetectionio/tests/proxy_list/squid.conf
Normal file
41
changedetectionio/tests/proxy_list/squid.conf
Normal file
@@ -0,0 +1,41 @@
|
|||||||
|
acl localnet src 0.0.0.1-0.255.255.255 # RFC 1122 "this" network (LAN)
|
||||||
|
acl localnet src 10.0.0.0/8 # RFC 1918 local private network (LAN)
|
||||||
|
acl localnet src 100.64.0.0/10 # RFC 6598 shared address space (CGN)
|
||||||
|
acl localnet src 169.254.0.0/16 # RFC 3927 link-local (directly plugged) machines
|
||||||
|
acl localnet src 172.16.0.0/12 # RFC 1918 local private network (LAN)
|
||||||
|
acl localnet src 192.168.0.0/16 # RFC 1918 local private network (LAN)
|
||||||
|
acl localnet src fc00::/7 # RFC 4193 local private network range
|
||||||
|
acl localnet src fe80::/10 # RFC 4291 link-local (directly plugged) machines
|
||||||
|
acl localnet src 159.65.224.174
|
||||||
|
acl SSL_ports port 443
|
||||||
|
acl Safe_ports port 80 # http
|
||||||
|
acl Safe_ports port 21 # ftp
|
||||||
|
acl Safe_ports port 443 # https
|
||||||
|
acl Safe_ports port 70 # gopher
|
||||||
|
acl Safe_ports port 210 # wais
|
||||||
|
acl Safe_ports port 1025-65535 # unregistered ports
|
||||||
|
acl Safe_ports port 280 # http-mgmt
|
||||||
|
acl Safe_ports port 488 # gss-http
|
||||||
|
acl Safe_ports port 591 # filemaker
|
||||||
|
acl Safe_ports port 777 # multiling http
|
||||||
|
acl CONNECT method CONNECT
|
||||||
|
|
||||||
|
http_access deny !Safe_ports
|
||||||
|
http_access deny CONNECT !SSL_ports
|
||||||
|
http_access allow localhost manager
|
||||||
|
http_access deny manager
|
||||||
|
http_access allow localhost
|
||||||
|
http_access allow localnet
|
||||||
|
http_access deny all
|
||||||
|
http_port 3128
|
||||||
|
coredump_dir /var/spool/squid
|
||||||
|
refresh_pattern ^ftp: 1440 20% 10080
|
||||||
|
refresh_pattern ^gopher: 1440 0% 1440
|
||||||
|
refresh_pattern -i (/cgi-bin/|\?) 0 0% 0
|
||||||
|
refresh_pattern \/(Packages|Sources)(|\.bz2|\.gz|\.xz)$ 0 0% 0 refresh-ims
|
||||||
|
refresh_pattern \/Release(|\.gpg)$ 0 0% 0 refresh-ims
|
||||||
|
refresh_pattern \/InRelease$ 0 0% 0 refresh-ims
|
||||||
|
refresh_pattern \/(Translation-.*)(|\.bz2|\.gz|\.xz)$ 0 0% 0 refresh-ims
|
||||||
|
refresh_pattern . 0 20% 4320
|
||||||
|
logfile_rotate 0
|
||||||
|
|
||||||
38
changedetectionio/tests/proxy_list/test_multiple_proxy.py
Normal file
38
changedetectionio/tests/proxy_list/test_multiple_proxy.py
Normal file
@@ -0,0 +1,38 @@
|
|||||||
|
#!/usr/bin/python3
|
||||||
|
|
||||||
|
import time
|
||||||
|
from flask import url_for
|
||||||
|
from ..util import live_server_setup
|
||||||
|
|
||||||
|
def test_preferred_proxy(client, live_server):
|
||||||
|
time.sleep(1)
|
||||||
|
live_server_setup(live_server)
|
||||||
|
time.sleep(1)
|
||||||
|
url = "http://chosen.changedetection.io"
|
||||||
|
|
||||||
|
res = client.post(
|
||||||
|
url_for("import_page"),
|
||||||
|
# Because a URL wont show in squid/proxy logs due it being SSLed
|
||||||
|
# Use plain HTTP or a specific domain-name here
|
||||||
|
data={"urls": url},
|
||||||
|
follow_redirects=True
|
||||||
|
)
|
||||||
|
|
||||||
|
assert b"1 Imported" in res.data
|
||||||
|
|
||||||
|
time.sleep(2)
|
||||||
|
res = client.post(
|
||||||
|
url_for("edit_page", uuid="first"),
|
||||||
|
data={
|
||||||
|
"include_filters": "",
|
||||||
|
"fetch_backend": "html_requests",
|
||||||
|
"headers": "",
|
||||||
|
"proxy": "proxy-two",
|
||||||
|
"tag": "",
|
||||||
|
"url": url,
|
||||||
|
},
|
||||||
|
follow_redirects=True
|
||||||
|
)
|
||||||
|
assert b"Updated watch." in res.data
|
||||||
|
time.sleep(2)
|
||||||
|
# Now the request should appear in the second-squid logs
|
||||||
19
changedetectionio/tests/proxy_list/test_proxy.py
Normal file
19
changedetectionio/tests/proxy_list/test_proxy.py
Normal file
@@ -0,0 +1,19 @@
|
|||||||
|
#!/usr/bin/python3
|
||||||
|
|
||||||
|
import time
|
||||||
|
from flask import url_for
|
||||||
|
from ..util import live_server_setup, wait_for_all_checks, extract_UUID_from_client
|
||||||
|
|
||||||
|
# just make a request, we will grep in the docker logs to see it actually got called
|
||||||
|
def test_check_basic_change_detection_functionality(client, live_server):
|
||||||
|
live_server_setup(live_server)
|
||||||
|
res = client.post(
|
||||||
|
url_for("import_page"),
|
||||||
|
# Because a URL wont show in squid/proxy logs due it being SSLed
|
||||||
|
# Use plain HTTP or a specific domain-name here
|
||||||
|
data={"urls": "http://one.changedetection.io"},
|
||||||
|
follow_redirects=True
|
||||||
|
)
|
||||||
|
|
||||||
|
assert b"1 Imported" in res.data
|
||||||
|
time.sleep(3)
|
||||||
@@ -19,7 +19,6 @@ def test_check_access_control(app, client):
|
|||||||
)
|
)
|
||||||
|
|
||||||
assert b"Password protection enabled." in res.data
|
assert b"Password protection enabled." in res.data
|
||||||
assert b"LOG OUT" not in res.data
|
|
||||||
|
|
||||||
# Check we hit the login
|
# Check we hit the login
|
||||||
res = c.get(url_for("index"), follow_redirects=True)
|
res = c.get(url_for("index"), follow_redirects=True)
|
||||||
@@ -38,7 +37,40 @@ def test_check_access_control(app, client):
|
|||||||
follow_redirects=True
|
follow_redirects=True
|
||||||
)
|
)
|
||||||
|
|
||||||
|
# Yes we are correctly logged in
|
||||||
assert b"LOG OUT" in res.data
|
assert b"LOG OUT" in res.data
|
||||||
|
|
||||||
|
# 598 - Password should be set and not accidently removed
|
||||||
|
res = c.post(
|
||||||
|
url_for("settings_page"),
|
||||||
|
data={
|
||||||
|
"requests-time_between_check-minutes": 180,
|
||||||
|
'application-fetch_backend': "html_requests"},
|
||||||
|
follow_redirects=True
|
||||||
|
)
|
||||||
|
|
||||||
|
res = c.get(url_for("logout"),
|
||||||
|
follow_redirects=True)
|
||||||
|
|
||||||
|
res = c.get(url_for("settings_page"),
|
||||||
|
follow_redirects=True)
|
||||||
|
|
||||||
|
|
||||||
|
assert b"Login" in res.data
|
||||||
|
|
||||||
|
res = c.get(url_for("login"))
|
||||||
|
assert b"Login" in res.data
|
||||||
|
|
||||||
|
|
||||||
|
res = c.post(
|
||||||
|
url_for("login"),
|
||||||
|
data={"password": "foobar"},
|
||||||
|
follow_redirects=True
|
||||||
|
)
|
||||||
|
|
||||||
|
# Yes we are correctly logged in
|
||||||
|
assert b"LOG OUT" in res.data
|
||||||
|
|
||||||
res = c.get(url_for("settings_page"))
|
res = c.get(url_for("settings_page"))
|
||||||
|
|
||||||
# Menu should be available now
|
# Menu should be available now
|
||||||
|
|||||||
@@ -2,73 +2,204 @@
|
|||||||
|
|
||||||
import time
|
import time
|
||||||
from flask import url_for
|
from flask import url_for
|
||||||
from . util import live_server_setup
|
from .util import live_server_setup, extract_api_key_from_UI
|
||||||
|
|
||||||
def test_setup(live_server):
|
import json
|
||||||
live_server_setup(live_server)
|
import uuid
|
||||||
|
|
||||||
|
|
||||||
def set_response_data(test_return_data):
|
def set_original_response():
|
||||||
|
test_return_data = """<html>
|
||||||
|
<body>
|
||||||
|
Some initial text</br>
|
||||||
|
<p>Which is across multiple lines</p>
|
||||||
|
</br>
|
||||||
|
So let's see what happens. </br>
|
||||||
|
<div id="sametext">Some text thats the same</div>
|
||||||
|
<div id="changetext">Some text that will change</div>
|
||||||
|
</body>
|
||||||
|
</html>
|
||||||
|
"""
|
||||||
|
|
||||||
|
with open("test-datastore/endpoint-content.txt", "w") as f:
|
||||||
|
f.write(test_return_data)
|
||||||
|
return None
|
||||||
|
|
||||||
|
|
||||||
|
def set_modified_response():
|
||||||
|
test_return_data = """<html>
|
||||||
|
<body>
|
||||||
|
Some initial text</br>
|
||||||
|
<p>which has this one new line</p>
|
||||||
|
</br>
|
||||||
|
So let's see what happens. </br>
|
||||||
|
<div id="sametext">Some text thats the same</div>
|
||||||
|
<div id="changetext">Some text that changes</div>
|
||||||
|
</body>
|
||||||
|
</html>
|
||||||
|
"""
|
||||||
|
|
||||||
with open("test-datastore/endpoint-content.txt", "w") as f:
|
with open("test-datastore/endpoint-content.txt", "w") as f:
|
||||||
f.write(test_return_data)
|
f.write(test_return_data)
|
||||||
|
|
||||||
|
return None
|
||||||
|
|
||||||
def test_snapshot_api_detects_change(client, live_server):
|
|
||||||
test_return_data = "Some initial text"
|
|
||||||
|
|
||||||
test_return_data_modified = "Some NEW nice initial text"
|
def is_valid_uuid(val):
|
||||||
|
try:
|
||||||
|
uuid.UUID(str(val))
|
||||||
|
return True
|
||||||
|
except ValueError:
|
||||||
|
return False
|
||||||
|
|
||||||
sleep_time_for_fetch_thread = 3
|
|
||||||
|
|
||||||
set_response_data(test_return_data)
|
def test_api_simple(client, live_server):
|
||||||
|
live_server_setup(live_server)
|
||||||
|
|
||||||
# Give the endpoint time to spin up
|
api_key = extract_api_key_from_UI(client)
|
||||||
time.sleep(1)
|
|
||||||
|
|
||||||
# Add our URL to the import page
|
# Create a watch
|
||||||
test_url = url_for('test_endpoint', content_type="text/plain",
|
set_original_response()
|
||||||
_external=True)
|
watch_uuid = None
|
||||||
|
|
||||||
|
# Validate bad URL
|
||||||
|
test_url = url_for('test_endpoint', _external=True,
|
||||||
|
headers={'x-api-key': api_key}, )
|
||||||
res = client.post(
|
res = client.post(
|
||||||
url_for("import_page"),
|
url_for("createwatch"),
|
||||||
data={"urls": test_url},
|
data=json.dumps({"url": "h://xxxxxxxxxom"}),
|
||||||
|
headers={'content-type': 'application/json', 'x-api-key': api_key},
|
||||||
follow_redirects=True
|
follow_redirects=True
|
||||||
)
|
)
|
||||||
assert b"1 Imported" in res.data
|
|
||||||
|
|
||||||
# Trigger a check
|
|
||||||
client.get(url_for("api_watch_checknow"), follow_redirects=True)
|
|
||||||
|
|
||||||
# Give the thread time to pick it up
|
|
||||||
time.sleep(sleep_time_for_fetch_thread)
|
|
||||||
|
|
||||||
res = client.get(
|
|
||||||
url_for("api_snapshot", uuid="first"),
|
|
||||||
follow_redirects=True
|
|
||||||
)
|
|
||||||
|
|
||||||
assert test_return_data.encode() == res.data
|
|
||||||
|
|
||||||
# Make a change
|
|
||||||
set_response_data(test_return_data_modified)
|
|
||||||
|
|
||||||
# Trigger a check
|
|
||||||
client.get(url_for("api_watch_checknow"), follow_redirects=True)
|
|
||||||
# Give the thread time to pick it up
|
|
||||||
time.sleep(sleep_time_for_fetch_thread)
|
|
||||||
|
|
||||||
res = client.get(
|
|
||||||
url_for("api_snapshot", uuid="first"),
|
|
||||||
follow_redirects=True
|
|
||||||
)
|
|
||||||
|
|
||||||
assert test_return_data_modified.encode() == res.data
|
|
||||||
|
|
||||||
def test_snapshot_api_invalid_uuid(client, live_server):
|
|
||||||
|
|
||||||
res = client.get(
|
|
||||||
url_for("api_snapshot", uuid="invalid"),
|
|
||||||
follow_redirects=True
|
|
||||||
)
|
|
||||||
|
|
||||||
assert res.status_code == 400
|
assert res.status_code == 400
|
||||||
|
|
||||||
|
# Create new
|
||||||
|
res = client.post(
|
||||||
|
url_for("createwatch"),
|
||||||
|
data=json.dumps({"url": test_url, 'tag': "One, Two", "title": "My test URL"}),
|
||||||
|
headers={'content-type': 'application/json', 'x-api-key': api_key},
|
||||||
|
follow_redirects=True
|
||||||
|
)
|
||||||
|
s = json.loads(res.data)
|
||||||
|
assert is_valid_uuid(s['uuid'])
|
||||||
|
watch_uuid = s['uuid']
|
||||||
|
assert res.status_code == 201
|
||||||
|
|
||||||
|
time.sleep(3)
|
||||||
|
|
||||||
|
# Verify its in the list and that recheck worked
|
||||||
|
res = client.get(
|
||||||
|
url_for("createwatch"),
|
||||||
|
headers={'x-api-key': api_key}
|
||||||
|
)
|
||||||
|
assert watch_uuid in json.loads(res.data).keys()
|
||||||
|
before_recheck_info = json.loads(res.data)[watch_uuid]
|
||||||
|
assert before_recheck_info['last_checked'] != 0
|
||||||
|
#705 `last_changed` should be zero on the first check
|
||||||
|
assert before_recheck_info['last_changed'] == 0
|
||||||
|
assert before_recheck_info['title'] == 'My test URL'
|
||||||
|
|
||||||
|
set_modified_response()
|
||||||
|
# Trigger recheck of all ?recheck_all=1
|
||||||
|
client.get(
|
||||||
|
url_for("createwatch", recheck_all='1'),
|
||||||
|
headers={'x-api-key': api_key},
|
||||||
|
)
|
||||||
|
time.sleep(3)
|
||||||
|
|
||||||
|
# Did the recheck fire?
|
||||||
|
res = client.get(
|
||||||
|
url_for("createwatch"),
|
||||||
|
headers={'x-api-key': api_key},
|
||||||
|
)
|
||||||
|
after_recheck_info = json.loads(res.data)[watch_uuid]
|
||||||
|
assert after_recheck_info['last_checked'] != before_recheck_info['last_checked']
|
||||||
|
assert after_recheck_info['last_changed'] != 0
|
||||||
|
|
||||||
|
# Check history index list
|
||||||
|
res = client.get(
|
||||||
|
url_for("watchhistory", uuid=watch_uuid),
|
||||||
|
headers={'x-api-key': api_key},
|
||||||
|
)
|
||||||
|
history = json.loads(res.data)
|
||||||
|
assert len(history) == 2, "Should have two history entries (the original and the changed)"
|
||||||
|
|
||||||
|
# Fetch a snapshot by timestamp, check the right one was found
|
||||||
|
res = client.get(
|
||||||
|
url_for("watchsinglehistory", uuid=watch_uuid, timestamp=list(history.keys())[-1]),
|
||||||
|
headers={'x-api-key': api_key},
|
||||||
|
)
|
||||||
|
assert b'which has this one new line' in res.data
|
||||||
|
|
||||||
|
# Fetch a snapshot by 'latest'', check the right one was found
|
||||||
|
res = client.get(
|
||||||
|
url_for("watchsinglehistory", uuid=watch_uuid, timestamp='latest'),
|
||||||
|
headers={'x-api-key': api_key},
|
||||||
|
)
|
||||||
|
assert b'which has this one new line' in res.data
|
||||||
|
|
||||||
|
# Fetch the whole watch
|
||||||
|
res = client.get(
|
||||||
|
url_for("watch", uuid=watch_uuid),
|
||||||
|
headers={'x-api-key': api_key}
|
||||||
|
)
|
||||||
|
watch = json.loads(res.data)
|
||||||
|
# @todo how to handle None/default global values?
|
||||||
|
assert watch['history_n'] == 2, "Found replacement history section, which is in its own API"
|
||||||
|
|
||||||
|
# basic systeminfo check
|
||||||
|
res = client.get(
|
||||||
|
url_for("systeminfo"),
|
||||||
|
headers={'x-api-key': api_key},
|
||||||
|
)
|
||||||
|
info = json.loads(res.data)
|
||||||
|
assert info.get('watch_count') == 1
|
||||||
|
assert info.get('uptime') > 0.5
|
||||||
|
|
||||||
|
|
||||||
|
# Finally delete the watch
|
||||||
|
res = client.delete(
|
||||||
|
url_for("watch", uuid=watch_uuid),
|
||||||
|
headers={'x-api-key': api_key},
|
||||||
|
)
|
||||||
|
assert res.status_code == 204
|
||||||
|
|
||||||
|
# Check via a relist
|
||||||
|
res = client.get(
|
||||||
|
url_for("createwatch"),
|
||||||
|
headers={'x-api-key': api_key}
|
||||||
|
)
|
||||||
|
watch_list = json.loads(res.data)
|
||||||
|
assert len(watch_list) == 0, "Watch list should be empty"
|
||||||
|
|
||||||
|
|
||||||
|
def test_access_denied(client, live_server):
|
||||||
|
# `config_api_token_enabled` Should be On by default
|
||||||
|
res = client.get(
|
||||||
|
url_for("createwatch")
|
||||||
|
)
|
||||||
|
assert res.status_code == 403
|
||||||
|
|
||||||
|
res = client.get(
|
||||||
|
url_for("createwatch"),
|
||||||
|
headers={'x-api-key': "something horrible"}
|
||||||
|
)
|
||||||
|
assert res.status_code == 403
|
||||||
|
|
||||||
|
# Disable config_api_token_enabled and it should work
|
||||||
|
res = client.post(
|
||||||
|
url_for("settings_page"),
|
||||||
|
data={
|
||||||
|
"requests-time_between_check-minutes": 180,
|
||||||
|
"application-fetch_backend": "html_requests",
|
||||||
|
"application-api_access_token_enabled": ""
|
||||||
|
},
|
||||||
|
follow_redirects=True
|
||||||
|
)
|
||||||
|
|
||||||
|
assert b"Settings updated." in res.data
|
||||||
|
|
||||||
|
res = client.get(
|
||||||
|
url_for("createwatch")
|
||||||
|
)
|
||||||
|
assert res.status_code == 200
|
||||||
|
|||||||
@@ -23,13 +23,13 @@ def test_basic_auth(client, live_server):
|
|||||||
# Check form validation
|
# Check form validation
|
||||||
res = client.post(
|
res = client.post(
|
||||||
url_for("edit_page", uuid="first"),
|
url_for("edit_page", uuid="first"),
|
||||||
data={"css_filter": "", "url": test_url, "tag": "", "headers": "", 'fetch_backend': "html_requests"},
|
data={"include_filters": "", "url": test_url, "tag": "", "headers": "", 'fetch_backend': "html_requests"},
|
||||||
follow_redirects=True
|
follow_redirects=True
|
||||||
)
|
)
|
||||||
assert b"Updated watch." in res.data
|
assert b"Updated watch." in res.data
|
||||||
|
|
||||||
# Trigger a check
|
# Trigger a check
|
||||||
client.get(url_for("api_watch_checknow"), follow_redirects=True)
|
client.get(url_for("form_watch_checknow"), follow_redirects=True)
|
||||||
time.sleep(1)
|
time.sleep(1)
|
||||||
res = client.get(
|
res = client.get(
|
||||||
url_for("preview_page", uuid="first"),
|
url_for("preview_page", uuid="first"),
|
||||||
|
|||||||
@@ -3,14 +3,15 @@
|
|||||||
import time
|
import time
|
||||||
from flask import url_for
|
from flask import url_for
|
||||||
from urllib.request import urlopen
|
from urllib.request import urlopen
|
||||||
from . util import set_original_response, set_modified_response, live_server_setup
|
from .util import set_original_response, set_modified_response, live_server_setup, wait_for_all_checks
|
||||||
|
|
||||||
sleep_time_for_fetch_thread = 3
|
sleep_time_for_fetch_thread = 3
|
||||||
|
|
||||||
|
|
||||||
# Basic test to check inscriptus is not adding return line chars, basically works etc
|
# Basic test to check inscriptus is not adding return line chars, basically works etc
|
||||||
def test_inscriptus():
|
def test_inscriptus():
|
||||||
from inscriptis import get_text
|
from inscriptis import get_text
|
||||||
html_content="<html><body>test!<br/>ok man</body></html>"
|
html_content = "<html><body>test!<br/>ok man</body></html>"
|
||||||
stripped_text_from_html = get_text(html_content)
|
stripped_text_from_html = get_text(html_content)
|
||||||
assert stripped_text_from_html == 'test!\nok man'
|
assert stripped_text_from_html == 'test!\nok man'
|
||||||
|
|
||||||
@@ -32,10 +33,10 @@ def test_check_basic_change_detection_functionality(client, live_server):
|
|||||||
|
|
||||||
# Do this a few times.. ensures we dont accidently set the status
|
# Do this a few times.. ensures we dont accidently set the status
|
||||||
for n in range(3):
|
for n in range(3):
|
||||||
client.get(url_for("api_watch_checknow"), follow_redirects=True)
|
client.get(url_for("form_watch_checknow"), follow_redirects=True)
|
||||||
|
|
||||||
# Give the thread time to pick it up
|
# Give the thread time to pick it up
|
||||||
time.sleep(sleep_time_for_fetch_thread)
|
wait_for_all_checks(client)
|
||||||
|
|
||||||
# It should report nothing found (no new 'unviewed' class)
|
# It should report nothing found (no new 'unviewed' class)
|
||||||
res = client.get(url_for("index"))
|
res = client.get(url_for("index"))
|
||||||
@@ -65,10 +66,10 @@ def test_check_basic_change_detection_functionality(client, live_server):
|
|||||||
assert b'which has this one new line' in res.read()
|
assert b'which has this one new line' in res.read()
|
||||||
|
|
||||||
# Force recheck
|
# Force recheck
|
||||||
res = client.get(url_for("api_watch_checknow"), follow_redirects=True)
|
res = client.get(url_for("form_watch_checknow"), follow_redirects=True)
|
||||||
assert b'1 watches are queued for rechecking.' in res.data
|
assert b'1 watches are queued for rechecking.' in res.data
|
||||||
|
|
||||||
time.sleep(sleep_time_for_fetch_thread)
|
wait_for_all_checks(client)
|
||||||
|
|
||||||
# Now something should be ready, indicated by having a 'unviewed' class
|
# Now something should be ready, indicated by having a 'unviewed' class
|
||||||
res = client.get(url_for("index"))
|
res = client.get(url_for("index"))
|
||||||
@@ -89,18 +90,27 @@ def test_check_basic_change_detection_functionality(client, live_server):
|
|||||||
res = client.get(url_for("diff_history_page", uuid="first"))
|
res = client.get(url_for("diff_history_page", uuid="first"))
|
||||||
assert b'Compare newest' in res.data
|
assert b'Compare newest' in res.data
|
||||||
|
|
||||||
time.sleep(2)
|
# Check the [preview] pulls the right one
|
||||||
|
res = client.get(
|
||||||
|
url_for("preview_page", uuid="first"),
|
||||||
|
follow_redirects=True
|
||||||
|
)
|
||||||
|
assert b'which has this one new line' in res.data
|
||||||
|
assert b'Which is across multiple lines' not in res.data
|
||||||
|
|
||||||
|
wait_for_all_checks(client)
|
||||||
|
|
||||||
# Do this a few times.. ensures we dont accidently set the status
|
# Do this a few times.. ensures we dont accidently set the status
|
||||||
for n in range(2):
|
for n in range(2):
|
||||||
client.get(url_for("api_watch_checknow"), follow_redirects=True)
|
client.get(url_for("form_watch_checknow"), follow_redirects=True)
|
||||||
|
|
||||||
# Give the thread time to pick it up
|
# Give the thread time to pick it up
|
||||||
time.sleep(sleep_time_for_fetch_thread)
|
wait_for_all_checks(client)
|
||||||
|
|
||||||
# It should report nothing found (no new 'unviewed' class)
|
# It should report nothing found (no new 'unviewed' class)
|
||||||
res = client.get(url_for("index"))
|
res = client.get(url_for("index"))
|
||||||
assert b'unviewed' not in res.data
|
assert b'unviewed' not in res.data
|
||||||
|
assert b'Mark all viewed' not in res.data
|
||||||
assert b'head title' not in res.data # Should not be present because this is off by default
|
assert b'head title' not in res.data # Should not be present because this is off by default
|
||||||
assert b'test-endpoint' in res.data
|
assert b'test-endpoint' in res.data
|
||||||
|
|
||||||
@@ -109,20 +119,28 @@ def test_check_basic_change_detection_functionality(client, live_server):
|
|||||||
# Enable auto pickup of <title> in settings
|
# Enable auto pickup of <title> in settings
|
||||||
res = client.post(
|
res = client.post(
|
||||||
url_for("settings_page"),
|
url_for("settings_page"),
|
||||||
data={"application-extract_title_as_title": "1", "requests-time_between_check-minutes": 180, 'application-fetch_backend': "html_requests"},
|
data={"application-extract_title_as_title": "1", "requests-time_between_check-minutes": 180,
|
||||||
|
'application-fetch_backend': "html_requests"},
|
||||||
follow_redirects=True
|
follow_redirects=True
|
||||||
)
|
)
|
||||||
|
|
||||||
client.get(url_for("api_watch_checknow"), follow_redirects=True)
|
client.get(url_for("form_watch_checknow"), follow_redirects=True)
|
||||||
time.sleep(sleep_time_for_fetch_thread)
|
wait_for_all_checks(client)
|
||||||
|
|
||||||
res = client.get(url_for("index"))
|
res = client.get(url_for("index"))
|
||||||
assert b'unviewed' in res.data
|
assert b'unviewed' in res.data
|
||||||
|
assert b'Mark all viewed' in res.data
|
||||||
|
|
||||||
# It should have picked up the <title>
|
# It should have picked up the <title>
|
||||||
assert b'head title' in res.data
|
assert b'head title' in res.data
|
||||||
|
|
||||||
|
# hit the mark all viewed link
|
||||||
|
res = client.get(url_for("mark_all_viewed"), follow_redirects=True)
|
||||||
|
|
||||||
|
assert b'Mark all viewed' not in res.data
|
||||||
|
assert b'unviewed' not in res.data
|
||||||
|
|
||||||
#
|
#
|
||||||
# Cleanup everything
|
# Cleanup everything
|
||||||
res = client.get(url_for("api_delete", uuid="all"), follow_redirects=True)
|
res = client.get(url_for("form_delete", uuid="all"), follow_redirects=True)
|
||||||
assert b'Deleted' in res.data
|
assert b'Deleted' in res.data
|
||||||
|
|
||||||
|
|||||||
@@ -1,18 +1,31 @@
|
|||||||
#!/usr/bin/python3
|
#!/usr/bin/python3
|
||||||
|
|
||||||
import time
|
from .util import set_original_response, set_modified_response, live_server_setup
|
||||||
from flask import url_for
|
from flask import url_for
|
||||||
from urllib.request import urlopen
|
from urllib.request import urlopen
|
||||||
from . util import set_original_response, set_modified_response, live_server_setup
|
from zipfile import ZipFile
|
||||||
|
import re
|
||||||
|
import time
|
||||||
|
|
||||||
|
|
||||||
def test_backup(client, live_server):
|
def test_backup(client, live_server):
|
||||||
|
|
||||||
live_server_setup(live_server)
|
live_server_setup(live_server)
|
||||||
|
|
||||||
|
set_original_response()
|
||||||
|
|
||||||
# Give the endpoint time to spin up
|
# Give the endpoint time to spin up
|
||||||
time.sleep(1)
|
time.sleep(1)
|
||||||
|
|
||||||
|
# Add our URL to the import page
|
||||||
|
res = client.post(
|
||||||
|
url_for("import_page"),
|
||||||
|
data={"urls": url_for('test_endpoint', _external=True)},
|
||||||
|
follow_redirects=True
|
||||||
|
)
|
||||||
|
|
||||||
|
assert b"1 Imported" in res.data
|
||||||
|
time.sleep(3)
|
||||||
|
|
||||||
res = client.get(
|
res = client.get(
|
||||||
url_for("get_backup"),
|
url_for("get_backup"),
|
||||||
follow_redirects=True
|
follow_redirects=True
|
||||||
@@ -20,6 +33,19 @@ def test_backup(client, live_server):
|
|||||||
|
|
||||||
# Should get the right zip content type
|
# Should get the right zip content type
|
||||||
assert res.content_type == "application/zip"
|
assert res.content_type == "application/zip"
|
||||||
|
|
||||||
# Should be PK/ZIP stream
|
# Should be PK/ZIP stream
|
||||||
assert res.data.count(b'PK') >= 2
|
assert res.data.count(b'PK') >= 2
|
||||||
|
|
||||||
|
# ZipFile from buffer seems non-obvious, just save it instead
|
||||||
|
with open("download.zip", 'wb') as f:
|
||||||
|
f.write(res.data)
|
||||||
|
|
||||||
|
zip = ZipFile('download.zip')
|
||||||
|
l = zip.namelist()
|
||||||
|
uuid4hex = re.compile('^[a-f0-9]{8}-?[a-f0-9]{4}-?4[a-f0-9]{3}-?[89ab][a-f0-9]{3}-?[a-f0-9]{12}.*txt', re.I)
|
||||||
|
newlist = list(filter(uuid4hex.match, l)) # Read Note below
|
||||||
|
|
||||||
|
# Should be two txt files in the archive (history and the snapshot)
|
||||||
|
assert len(newlist) == 2
|
||||||
|
|
||||||
|
|||||||
137
changedetectionio/tests/test_block_while_text_present.py
Normal file
137
changedetectionio/tests/test_block_while_text_present.py
Normal file
@@ -0,0 +1,137 @@
|
|||||||
|
#!/usr/bin/python3
|
||||||
|
|
||||||
|
import time
|
||||||
|
from flask import url_for
|
||||||
|
from . util import live_server_setup
|
||||||
|
from changedetectionio import html_tools
|
||||||
|
|
||||||
|
def set_original_ignore_response():
|
||||||
|
test_return_data = """<html>
|
||||||
|
<body>
|
||||||
|
Some initial text</br>
|
||||||
|
<p>Which is across multiple lines</p>
|
||||||
|
</br>
|
||||||
|
So let's see what happens. </br>
|
||||||
|
</body>
|
||||||
|
</html>
|
||||||
|
|
||||||
|
"""
|
||||||
|
|
||||||
|
with open("test-datastore/endpoint-content.txt", "w") as f:
|
||||||
|
f.write(test_return_data)
|
||||||
|
|
||||||
|
|
||||||
|
def set_modified_original_ignore_response():
|
||||||
|
test_return_data = """<html>
|
||||||
|
<body>
|
||||||
|
Some NEW nice initial text</br>
|
||||||
|
<p>Which is across multiple lines</p>
|
||||||
|
</br>
|
||||||
|
So let's see what happens. </br>
|
||||||
|
<p>new ignore stuff</p>
|
||||||
|
<p>out of stock</p>
|
||||||
|
<p>blah</p>
|
||||||
|
</body>
|
||||||
|
</html>
|
||||||
|
|
||||||
|
"""
|
||||||
|
|
||||||
|
with open("test-datastore/endpoint-content.txt", "w") as f:
|
||||||
|
f.write(test_return_data)
|
||||||
|
|
||||||
|
|
||||||
|
# Is the same but includes ZZZZZ, 'ZZZZZ' is the last line in ignore_text
|
||||||
|
def set_modified_response_minus_block_text():
|
||||||
|
test_return_data = """<html>
|
||||||
|
<body>
|
||||||
|
Some NEW nice initial text</br>
|
||||||
|
<p>Which is across multiple lines</p>
|
||||||
|
<p>now on sale $2/p>
|
||||||
|
</br>
|
||||||
|
So let's see what happens. </br>
|
||||||
|
<p>new ignore stuff</p>
|
||||||
|
<p>blah</p>
|
||||||
|
</body>
|
||||||
|
</html>
|
||||||
|
|
||||||
|
"""
|
||||||
|
|
||||||
|
with open("test-datastore/endpoint-content.txt", "w") as f:
|
||||||
|
f.write(test_return_data)
|
||||||
|
|
||||||
|
|
||||||
|
def test_check_block_changedetection_text_NOT_present(client, live_server):
|
||||||
|
sleep_time_for_fetch_thread = 3
|
||||||
|
live_server_setup(live_server)
|
||||||
|
# Use a mix of case in ZzZ to prove it works case-insensitive.
|
||||||
|
ignore_text = "out of stoCk\r\nfoobar"
|
||||||
|
|
||||||
|
set_original_ignore_response()
|
||||||
|
|
||||||
|
# Give the endpoint time to spin up
|
||||||
|
time.sleep(1)
|
||||||
|
|
||||||
|
# Add our URL to the import page
|
||||||
|
test_url = url_for('test_endpoint', _external=True)
|
||||||
|
res = client.post(
|
||||||
|
url_for("import_page"),
|
||||||
|
data={"urls": test_url},
|
||||||
|
follow_redirects=True
|
||||||
|
)
|
||||||
|
assert b"1 Imported" in res.data
|
||||||
|
|
||||||
|
# Give the thread time to pick it up
|
||||||
|
time.sleep(sleep_time_for_fetch_thread)
|
||||||
|
|
||||||
|
# Goto the edit page, add our ignore text
|
||||||
|
# Add our URL to the import page
|
||||||
|
res = client.post(
|
||||||
|
url_for("edit_page", uuid="first"),
|
||||||
|
data={"text_should_not_be_present": ignore_text, "url": test_url, 'fetch_backend': "html_requests"},
|
||||||
|
follow_redirects=True
|
||||||
|
)
|
||||||
|
assert b"Updated watch." in res.data
|
||||||
|
|
||||||
|
# Give the thread time to pick it up
|
||||||
|
time.sleep(sleep_time_for_fetch_thread)
|
||||||
|
# Check it saved
|
||||||
|
res = client.get(
|
||||||
|
url_for("edit_page", uuid="first"),
|
||||||
|
)
|
||||||
|
assert bytes(ignore_text.encode('utf-8')) in res.data
|
||||||
|
|
||||||
|
# Trigger a check
|
||||||
|
client.get(url_for("form_watch_checknow"), follow_redirects=True)
|
||||||
|
|
||||||
|
# Give the thread time to pick it up
|
||||||
|
time.sleep(sleep_time_for_fetch_thread)
|
||||||
|
|
||||||
|
# It should report nothing found (no new 'unviewed' class)
|
||||||
|
res = client.get(url_for("index"))
|
||||||
|
assert b'unviewed' not in res.data
|
||||||
|
assert b'/test-endpoint' in res.data
|
||||||
|
|
||||||
|
# The page changed, BUT the text is still there, just the rest of it changes, we should not see a change
|
||||||
|
set_modified_original_ignore_response()
|
||||||
|
|
||||||
|
# Trigger a check
|
||||||
|
client.get(url_for("form_watch_checknow"), follow_redirects=True)
|
||||||
|
# Give the thread time to pick it up
|
||||||
|
time.sleep(sleep_time_for_fetch_thread)
|
||||||
|
|
||||||
|
# It should report nothing found (no new 'unviewed' class)
|
||||||
|
res = client.get(url_for("index"))
|
||||||
|
assert b'unviewed' not in res.data
|
||||||
|
assert b'/test-endpoint' in res.data
|
||||||
|
|
||||||
|
|
||||||
|
# Now we set a change where the text is gone, it should now trigger
|
||||||
|
set_modified_response_minus_block_text()
|
||||||
|
client.get(url_for("form_watch_checknow"), follow_redirects=True)
|
||||||
|
time.sleep(sleep_time_for_fetch_thread)
|
||||||
|
|
||||||
|
res = client.get(url_for("index"))
|
||||||
|
assert b'unviewed' in res.data
|
||||||
|
|
||||||
|
res = client.get(url_for("form_delete", uuid="all"), follow_redirects=True)
|
||||||
|
assert b'Deleted' in res.data
|
||||||
@@ -23,7 +23,7 @@ def test_trigger_functionality(client, live_server):
|
|||||||
|
|
||||||
|
|
||||||
res = client.get(
|
res = client.get(
|
||||||
url_for("api_clone", uuid="first"),
|
url_for("form_clone", uuid="first"),
|
||||||
follow_redirects=True
|
follow_redirects=True
|
||||||
)
|
)
|
||||||
|
|
||||||
|
|||||||
@@ -46,22 +46,23 @@ def set_modified_response():
|
|||||||
|
|
||||||
|
|
||||||
# Test that the CSS extraction works how we expect, important here is the right placing of new lines \n's
|
# Test that the CSS extraction works how we expect, important here is the right placing of new lines \n's
|
||||||
def test_css_filter_output():
|
def test_include_filters_output():
|
||||||
from changedetectionio import fetch_site_status
|
|
||||||
from inscriptis import get_text
|
from inscriptis import get_text
|
||||||
|
|
||||||
# Check text with sub-parts renders correctly
|
# Check text with sub-parts renders correctly
|
||||||
content = """<html> <body><div id="thingthing" > Some really <b>bold</b> text </div> </body> </html>"""
|
content = """<html> <body><div id="thingthing" > Some really <b>bold</b> text </div> </body> </html>"""
|
||||||
html_blob = css_filter(css_filter="#thingthing", html_content=content)
|
html_blob = include_filters(include_filters="#thingthing", html_content=content)
|
||||||
text = get_text(html_blob)
|
text = get_text(html_blob)
|
||||||
assert text == " Some really bold text"
|
assert text == " Some really bold text"
|
||||||
|
|
||||||
content = """<html> <body>
|
content = """<html> <body>
|
||||||
<p>foo bar blah</p>
|
<p>foo bar blah</p>
|
||||||
<div class="parts">Block A</div> <div class="parts">Block B</div></body>
|
<DIV class="parts">Block A</DiV> <div class="parts">Block B</DIV></body>
|
||||||
</html>
|
</html>
|
||||||
"""
|
"""
|
||||||
html_blob = css_filter(css_filter=".parts", html_content=content)
|
|
||||||
|
# in xPath this would be //*[@class='parts']
|
||||||
|
html_blob = include_filters(include_filters=".parts", html_content=content)
|
||||||
text = get_text(html_blob)
|
text = get_text(html_blob)
|
||||||
|
|
||||||
# Divs are converted to 4 whitespaces by inscriptis
|
# Divs are converted to 4 whitespaces by inscriptis
|
||||||
@@ -69,10 +70,10 @@ def test_css_filter_output():
|
|||||||
|
|
||||||
|
|
||||||
# Tests the whole stack works with the CSS Filter
|
# Tests the whole stack works with the CSS Filter
|
||||||
def test_check_markup_css_filter_restriction(client, live_server):
|
def test_check_markup_include_filters_restriction(client, live_server):
|
||||||
sleep_time_for_fetch_thread = 3
|
sleep_time_for_fetch_thread = 3
|
||||||
|
|
||||||
css_filter = "#sametext"
|
include_filters = "#sametext"
|
||||||
|
|
||||||
set_original_response()
|
set_original_response()
|
||||||
|
|
||||||
@@ -89,7 +90,7 @@ def test_check_markup_css_filter_restriction(client, live_server):
|
|||||||
assert b"1 Imported" in res.data
|
assert b"1 Imported" in res.data
|
||||||
|
|
||||||
# Trigger a check
|
# Trigger a check
|
||||||
client.get(url_for("api_watch_checknow"), follow_redirects=True)
|
client.get(url_for("form_watch_checknow"), follow_redirects=True)
|
||||||
|
|
||||||
# Give the thread time to pick it up
|
# Give the thread time to pick it up
|
||||||
time.sleep(sleep_time_for_fetch_thread)
|
time.sleep(sleep_time_for_fetch_thread)
|
||||||
@@ -98,7 +99,7 @@ def test_check_markup_css_filter_restriction(client, live_server):
|
|||||||
# Add our URL to the import page
|
# Add our URL to the import page
|
||||||
res = client.post(
|
res = client.post(
|
||||||
url_for("edit_page", uuid="first"),
|
url_for("edit_page", uuid="first"),
|
||||||
data={"css_filter": css_filter, "url": test_url, "tag": "", "headers": "", 'fetch_backend': "html_requests"},
|
data={"include_filters": include_filters, "url": test_url, "tag": "", "headers": "", 'fetch_backend': "html_requests"},
|
||||||
follow_redirects=True
|
follow_redirects=True
|
||||||
)
|
)
|
||||||
assert b"Updated watch." in res.data
|
assert b"Updated watch." in res.data
|
||||||
@@ -107,10 +108,10 @@ def test_check_markup_css_filter_restriction(client, live_server):
|
|||||||
res = client.get(
|
res = client.get(
|
||||||
url_for("edit_page", uuid="first"),
|
url_for("edit_page", uuid="first"),
|
||||||
)
|
)
|
||||||
assert bytes(css_filter.encode('utf-8')) in res.data
|
assert bytes(include_filters.encode('utf-8')) in res.data
|
||||||
|
|
||||||
# Trigger a check
|
# Trigger a check
|
||||||
client.get(url_for("api_watch_checknow"), follow_redirects=True)
|
client.get(url_for("form_watch_checknow"), follow_redirects=True)
|
||||||
|
|
||||||
# Give the thread time to pick it up
|
# Give the thread time to pick it up
|
||||||
time.sleep(sleep_time_for_fetch_thread)
|
time.sleep(sleep_time_for_fetch_thread)
|
||||||
@@ -118,7 +119,7 @@ def test_check_markup_css_filter_restriction(client, live_server):
|
|||||||
set_modified_response()
|
set_modified_response()
|
||||||
|
|
||||||
# Trigger a check
|
# Trigger a check
|
||||||
client.get(url_for("api_watch_checknow"), follow_redirects=True)
|
client.get(url_for("form_watch_checknow"), follow_redirects=True)
|
||||||
# Give the thread time to pick it up
|
# Give the thread time to pick it up
|
||||||
time.sleep(sleep_time_for_fetch_thread)
|
time.sleep(sleep_time_for_fetch_thread)
|
||||||
|
|
||||||
@@ -126,3 +127,58 @@ def test_check_markup_css_filter_restriction(client, live_server):
|
|||||||
# Because it should be looking at only that 'sametext' id
|
# Because it should be looking at only that 'sametext' id
|
||||||
res = client.get(url_for("index"))
|
res = client.get(url_for("index"))
|
||||||
assert b'unviewed' in res.data
|
assert b'unviewed' in res.data
|
||||||
|
|
||||||
|
|
||||||
|
# Tests the whole stack works with the CSS Filter
|
||||||
|
def test_check_multiple_filters(client, live_server):
|
||||||
|
sleep_time_for_fetch_thread = 3
|
||||||
|
|
||||||
|
include_filters = "#blob-a\r\nxpath://*[contains(@id,'blob-b')]"
|
||||||
|
|
||||||
|
with open("test-datastore/endpoint-content.txt", "w") as f:
|
||||||
|
f.write("""<html><body>
|
||||||
|
<div id="blob-a">Blob A</div>
|
||||||
|
<div id="blob-b">Blob B</div>
|
||||||
|
<div id="blob-c">Blob C</div>
|
||||||
|
</body>
|
||||||
|
</html>
|
||||||
|
""")
|
||||||
|
|
||||||
|
# Give the endpoint time to spin up
|
||||||
|
time.sleep(1)
|
||||||
|
|
||||||
|
# Add our URL to the import page
|
||||||
|
test_url = url_for('test_endpoint', _external=True)
|
||||||
|
res = client.post(
|
||||||
|
url_for("import_page"),
|
||||||
|
data={"urls": test_url},
|
||||||
|
follow_redirects=True
|
||||||
|
)
|
||||||
|
assert b"1 Imported" in res.data
|
||||||
|
time.sleep(1)
|
||||||
|
|
||||||
|
# Goto the edit page, add our ignore text
|
||||||
|
# Add our URL to the import page
|
||||||
|
res = client.post(
|
||||||
|
url_for("edit_page", uuid="first"),
|
||||||
|
data={"include_filters": include_filters,
|
||||||
|
"url": test_url,
|
||||||
|
"tag": "",
|
||||||
|
"headers": "",
|
||||||
|
'fetch_backend': "html_requests"},
|
||||||
|
follow_redirects=True
|
||||||
|
)
|
||||||
|
assert b"Updated watch." in res.data
|
||||||
|
|
||||||
|
# Give the thread time to pick it up
|
||||||
|
time.sleep(sleep_time_for_fetch_thread)
|
||||||
|
|
||||||
|
res = client.get(
|
||||||
|
url_for("preview_page", uuid="first"),
|
||||||
|
follow_redirects=True
|
||||||
|
)
|
||||||
|
|
||||||
|
# Only the two blobs should be here
|
||||||
|
assert b"Blob A" in res.data # CSS was ok
|
||||||
|
assert b"Blob B" in res.data # xPath was ok
|
||||||
|
assert b"Blob C" not in res.data # Should not be included
|
||||||
|
|||||||
@@ -145,20 +145,19 @@ def test_element_removal_full(client, live_server):
|
|||||||
assert bytes(subtractive_selectors_data.encode("utf-8")) in res.data
|
assert bytes(subtractive_selectors_data.encode("utf-8")) in res.data
|
||||||
|
|
||||||
# Trigger a check
|
# Trigger a check
|
||||||
client.get(url_for("api_watch_checknow"), follow_redirects=True)
|
client.get(url_for("form_watch_checknow"), follow_redirects=True)
|
||||||
|
|
||||||
# Give the thread time to pick it up
|
# Give the thread time to pick it up
|
||||||
time.sleep(sleep_time_for_fetch_thread)
|
time.sleep(sleep_time_for_fetch_thread)
|
||||||
|
|
||||||
# No change yet - first check
|
# so that we set the state to 'unviewed' after all the edits
|
||||||
res = client.get(url_for("index"))
|
client.get(url_for("diff_history_page", uuid="first"))
|
||||||
assert b"unviewed" not in res.data
|
|
||||||
|
|
||||||
# Make a change to header/footer/nav
|
# Make a change to header/footer/nav
|
||||||
set_modified_response()
|
set_modified_response()
|
||||||
|
|
||||||
# Trigger a check
|
# Trigger a check
|
||||||
client.get(url_for("api_watch_checknow"), follow_redirects=True)
|
client.get(url_for("form_watch_checknow"), follow_redirects=True)
|
||||||
|
|
||||||
# Give the thread time to pick it up
|
# Give the thread time to pick it up
|
||||||
time.sleep(sleep_time_for_fetch_thread)
|
time.sleep(sleep_time_for_fetch_thread)
|
||||||
|
|||||||
@@ -39,7 +39,7 @@ def test_check_encoding_detection(client, live_server):
|
|||||||
)
|
)
|
||||||
|
|
||||||
# Trigger a check
|
# Trigger a check
|
||||||
client.get(url_for("api_watch_checknow"), follow_redirects=True)
|
client.get(url_for("form_watch_checknow"), follow_redirects=True)
|
||||||
|
|
||||||
# Give the thread time to pick it up
|
# Give the thread time to pick it up
|
||||||
time.sleep(2)
|
time.sleep(2)
|
||||||
@@ -71,7 +71,7 @@ def test_check_encoding_detection_missing_content_type_header(client, live_serve
|
|||||||
)
|
)
|
||||||
|
|
||||||
# Trigger a check
|
# Trigger a check
|
||||||
client.get(url_for("api_watch_checknow"), follow_redirects=True)
|
client.get(url_for("form_watch_checknow"), follow_redirects=True)
|
||||||
|
|
||||||
# Give the thread time to pick it up
|
# Give the thread time to pick it up
|
||||||
time.sleep(2)
|
time.sleep(2)
|
||||||
|
|||||||
@@ -11,16 +11,17 @@ def test_setup(live_server):
|
|||||||
live_server_setup(live_server)
|
live_server_setup(live_server)
|
||||||
|
|
||||||
|
|
||||||
def test_error_handler(client, live_server):
|
def _runner_test_http_errors(client, live_server, http_code, expected_text):
|
||||||
|
|
||||||
|
with open("test-datastore/endpoint-content.txt", "w") as f:
|
||||||
|
f.write("Now you going to get a {} error code\n".format(http_code))
|
||||||
|
|
||||||
# Give the endpoint time to spin up
|
|
||||||
time.sleep(1)
|
|
||||||
|
|
||||||
# Add our URL to the import page
|
# Add our URL to the import page
|
||||||
test_url = url_for('test_endpoint',
|
test_url = url_for('test_endpoint',
|
||||||
status_code=403,
|
status_code=http_code,
|
||||||
_external=True)
|
_external=True)
|
||||||
|
|
||||||
res = client.post(
|
res = client.post(
|
||||||
url_for("import_page"),
|
url_for("import_page"),
|
||||||
data={"urls": test_url},
|
data={"urls": test_url},
|
||||||
@@ -28,20 +29,39 @@ def test_error_handler(client, live_server):
|
|||||||
)
|
)
|
||||||
assert b"1 Imported" in res.data
|
assert b"1 Imported" in res.data
|
||||||
|
|
||||||
# Trigger a check
|
|
||||||
client.get(url_for("api_watch_checknow"), follow_redirects=True)
|
|
||||||
|
|
||||||
# Give the thread time to pick it up
|
# Give the thread time to pick it up
|
||||||
time.sleep(3)
|
time.sleep(2)
|
||||||
|
|
||||||
|
|
||||||
res = client.get(url_for("index"))
|
res = client.get(url_for("index"))
|
||||||
|
# no change
|
||||||
assert b'unviewed' not in res.data
|
assert b'unviewed' not in res.data
|
||||||
assert b'Status Code 403' in res.data
|
assert bytes(expected_text.encode('utf-8')) in res.data
|
||||||
assert bytes("just now".encode('utf-8')) in res.data
|
|
||||||
|
|
||||||
|
# Error viewing tabs should appear
|
||||||
|
res = client.get(
|
||||||
|
url_for("preview_page", uuid="first"),
|
||||||
|
follow_redirects=True
|
||||||
|
)
|
||||||
|
|
||||||
|
assert b'Error Text' in res.data
|
||||||
|
|
||||||
|
# 'Error Screenshot' only when in playwright mode
|
||||||
|
#assert b'Error Screenshot' in res.data
|
||||||
|
|
||||||
|
|
||||||
|
res = client.get(url_for("form_delete", uuid="all"), follow_redirects=True)
|
||||||
|
assert b'Deleted' in res.data
|
||||||
|
|
||||||
|
|
||||||
|
def test_http_error_handler(client, live_server):
|
||||||
|
_runner_test_http_errors(client, live_server, 403, 'Access denied')
|
||||||
|
_runner_test_http_errors(client, live_server, 404, 'Page not found')
|
||||||
|
_runner_test_http_errors(client, live_server, 500, '(Internal server Error) received')
|
||||||
|
_runner_test_http_errors(client, live_server, 400, 'Error - Request returned a HTTP error code 400')
|
||||||
|
|
||||||
# Just to be sure error text is properly handled
|
# Just to be sure error text is properly handled
|
||||||
def test_error_text_handler(client, live_server):
|
def test_DNS_errors(client, live_server):
|
||||||
# Give the endpoint time to spin up
|
# Give the endpoint time to spin up
|
||||||
time.sleep(1)
|
time.sleep(1)
|
||||||
|
|
||||||
@@ -53,13 +73,11 @@ def test_error_text_handler(client, live_server):
|
|||||||
)
|
)
|
||||||
assert b"1 Imported" in res.data
|
assert b"1 Imported" in res.data
|
||||||
|
|
||||||
# Trigger a check
|
|
||||||
client.get(url_for("api_watch_checknow"), follow_redirects=True)
|
|
||||||
|
|
||||||
# Give the thread time to pick it up
|
# Give the thread time to pick it up
|
||||||
time.sleep(3)
|
time.sleep(3)
|
||||||
|
|
||||||
res = client.get(url_for("index"))
|
res = client.get(url_for("index"))
|
||||||
assert b'Name or service not known' in res.data
|
assert b'Name or service not known' in res.data
|
||||||
|
# Should always record that we tried
|
||||||
assert bytes("just now".encode('utf-8')) in res.data
|
assert bytes("just now".encode('utf-8')) in res.data
|
||||||
|
|
||||||
|
|||||||
198
changedetectionio/tests/test_extract_regex.py
Normal file
198
changedetectionio/tests/test_extract_regex.py
Normal file
@@ -0,0 +1,198 @@
|
|||||||
|
#!/usr/bin/python3
|
||||||
|
|
||||||
|
import time
|
||||||
|
from flask import url_for
|
||||||
|
from .util import live_server_setup
|
||||||
|
|
||||||
|
from ..html_tools import *
|
||||||
|
|
||||||
|
|
||||||
|
def set_original_response():
|
||||||
|
test_return_data = """<html>
|
||||||
|
<body>
|
||||||
|
Some initial text</br>
|
||||||
|
<p>Which is across multiple lines</p>
|
||||||
|
</br>
|
||||||
|
So let's see what happens. </br>
|
||||||
|
<div id="sametext">Some text thats the same</div>
|
||||||
|
<div class="changetext">Some text that will change</div>
|
||||||
|
</body>
|
||||||
|
</html>
|
||||||
|
"""
|
||||||
|
|
||||||
|
with open("test-datastore/endpoint-content.txt", "w") as f:
|
||||||
|
f.write(test_return_data)
|
||||||
|
return None
|
||||||
|
|
||||||
|
|
||||||
|
def set_modified_response():
|
||||||
|
test_return_data = """<html>
|
||||||
|
<body>
|
||||||
|
Some initial text</br>
|
||||||
|
<p>which has this one new line</p>
|
||||||
|
</br>
|
||||||
|
So let's see what happens. </br>
|
||||||
|
<div id="sametext">Some text thats the same</div>
|
||||||
|
<div class="changetext">Some text that did change ( 1000 online <br/> 80 guests<br/> 2000 online )</div>
|
||||||
|
<div class="changetext">SomeCase insensitive 3456</div>
|
||||||
|
</body>
|
||||||
|
</html>
|
||||||
|
"""
|
||||||
|
|
||||||
|
with open("test-datastore/endpoint-content.txt", "w") as f:
|
||||||
|
f.write(test_return_data)
|
||||||
|
|
||||||
|
return None
|
||||||
|
|
||||||
|
|
||||||
|
def set_multiline_response():
|
||||||
|
test_return_data = """<html>
|
||||||
|
<body>
|
||||||
|
|
||||||
|
<p>Something <br/>
|
||||||
|
across 6 billion multiple<br/>
|
||||||
|
lines
|
||||||
|
</p>
|
||||||
|
|
||||||
|
<div>aaand something lines</div>
|
||||||
|
</body>
|
||||||
|
</html>
|
||||||
|
"""
|
||||||
|
|
||||||
|
with open("test-datastore/endpoint-content.txt", "w") as f:
|
||||||
|
f.write(test_return_data)
|
||||||
|
|
||||||
|
return None
|
||||||
|
|
||||||
|
|
||||||
|
def test_setup(client, live_server):
|
||||||
|
|
||||||
|
live_server_setup(live_server)
|
||||||
|
|
||||||
|
def test_check_filter_multiline(client, live_server):
|
||||||
|
|
||||||
|
set_multiline_response()
|
||||||
|
|
||||||
|
# Add our URL to the import page
|
||||||
|
test_url = url_for('test_endpoint', _external=True)
|
||||||
|
res = client.post(
|
||||||
|
url_for("import_page"),
|
||||||
|
data={"urls": test_url},
|
||||||
|
follow_redirects=True
|
||||||
|
)
|
||||||
|
assert b"1 Imported" in res.data
|
||||||
|
|
||||||
|
time.sleep(3)
|
||||||
|
|
||||||
|
# Goto the edit page, add our ignore text
|
||||||
|
# Add our URL to the import page
|
||||||
|
res = client.post(
|
||||||
|
url_for("edit_page", uuid="first"),
|
||||||
|
data={"include_filters": '',
|
||||||
|
'extract_text': '/something.+?6 billion.+?lines/si',
|
||||||
|
"url": test_url,
|
||||||
|
"tag": "",
|
||||||
|
"headers": "",
|
||||||
|
'fetch_backend': "html_requests"
|
||||||
|
},
|
||||||
|
follow_redirects=True
|
||||||
|
)
|
||||||
|
|
||||||
|
assert b"Updated watch." in res.data
|
||||||
|
time.sleep(3)
|
||||||
|
|
||||||
|
res = client.get(
|
||||||
|
url_for("preview_page", uuid="first"),
|
||||||
|
follow_redirects=True
|
||||||
|
)
|
||||||
|
|
||||||
|
|
||||||
|
assert b'<div class="">Something' in res.data
|
||||||
|
assert b'<div class="">across 6 billion multiple' in res.data
|
||||||
|
assert b'<div class="">lines' in res.data
|
||||||
|
|
||||||
|
# but the last one, which also says 'lines' shouldnt be here (non-greedy match checking)
|
||||||
|
assert b'aaand something lines' not in res.data
|
||||||
|
|
||||||
|
def test_check_filter_and_regex_extract(client, live_server):
|
||||||
|
sleep_time_for_fetch_thread = 3
|
||||||
|
include_filters = ".changetext"
|
||||||
|
|
||||||
|
set_original_response()
|
||||||
|
|
||||||
|
# Give the endpoint time to spin up
|
||||||
|
time.sleep(1)
|
||||||
|
|
||||||
|
# Add our URL to the import page
|
||||||
|
test_url = url_for('test_endpoint', _external=True)
|
||||||
|
res = client.post(
|
||||||
|
url_for("import_page"),
|
||||||
|
data={"urls": test_url},
|
||||||
|
follow_redirects=True
|
||||||
|
)
|
||||||
|
assert b"1 Imported" in res.data
|
||||||
|
|
||||||
|
time.sleep(1)
|
||||||
|
# Trigger a check
|
||||||
|
client.get(url_for("form_watch_checknow"), follow_redirects=True)
|
||||||
|
|
||||||
|
# Give the thread time to pick it up
|
||||||
|
time.sleep(sleep_time_for_fetch_thread)
|
||||||
|
|
||||||
|
# Goto the edit page, add our ignore text
|
||||||
|
# Add our URL to the import page
|
||||||
|
res = client.post(
|
||||||
|
url_for("edit_page", uuid="first"),
|
||||||
|
data={"include_filters": include_filters,
|
||||||
|
'extract_text': '\d+ online\r\n\d+ guests\r\n/somecase insensitive \d+/i\r\n/somecase insensitive (345\d)/i',
|
||||||
|
"url": test_url,
|
||||||
|
"tag": "",
|
||||||
|
"headers": "",
|
||||||
|
'fetch_backend': "html_requests"
|
||||||
|
},
|
||||||
|
follow_redirects=True
|
||||||
|
)
|
||||||
|
|
||||||
|
assert b"Updated watch." in res.data
|
||||||
|
|
||||||
|
# Give the thread time to pick it up
|
||||||
|
time.sleep(sleep_time_for_fetch_thread)
|
||||||
|
|
||||||
|
# Make a change
|
||||||
|
set_modified_response()
|
||||||
|
|
||||||
|
# Trigger a check
|
||||||
|
client.get(url_for("form_watch_checknow"), follow_redirects=True)
|
||||||
|
# Give the thread time to pick it up
|
||||||
|
time.sleep(sleep_time_for_fetch_thread)
|
||||||
|
|
||||||
|
# It should have 'unviewed' still
|
||||||
|
# Because it should be looking at only that 'sametext' id
|
||||||
|
res = client.get(url_for("index"))
|
||||||
|
assert b'unviewed' in res.data
|
||||||
|
|
||||||
|
# Check HTML conversion detected and workd
|
||||||
|
res = client.get(
|
||||||
|
url_for("preview_page", uuid="first"),
|
||||||
|
follow_redirects=True
|
||||||
|
)
|
||||||
|
|
||||||
|
# Class will be blank for now because the frontend didnt apply the diff
|
||||||
|
assert b'<div class="">1000 online' in res.data
|
||||||
|
|
||||||
|
# All regex matching should be here
|
||||||
|
assert b'<div class="">2000 online' in res.data
|
||||||
|
|
||||||
|
# Both regexs should be here
|
||||||
|
assert b'<div class="">80 guests' in res.data
|
||||||
|
|
||||||
|
# Regex with flag handling should be here
|
||||||
|
assert b'<div class="">SomeCase insensitive 3456' in res.data
|
||||||
|
|
||||||
|
# Singular group from /somecase insensitive (345\d)/i
|
||||||
|
assert b'<div class="">3456' in res.data
|
||||||
|
|
||||||
|
# Regex with multiline flag handling should be here
|
||||||
|
|
||||||
|
# Should not be here
|
||||||
|
assert b'Some text that did change' not in res.data
|
||||||
134
changedetectionio/tests/test_filter_exist_changes.py
Normal file
134
changedetectionio/tests/test_filter_exist_changes.py
Normal file
@@ -0,0 +1,134 @@
|
|||||||
|
#!/usr/bin/python3
|
||||||
|
|
||||||
|
# https://www.reddit.com/r/selfhosted/comments/wa89kp/comment/ii3a4g7/?context=3
|
||||||
|
import os
|
||||||
|
import time
|
||||||
|
from flask import url_for
|
||||||
|
from .util import set_original_response, live_server_setup
|
||||||
|
from changedetectionio.model import App
|
||||||
|
|
||||||
|
|
||||||
|
def set_response_without_filter():
|
||||||
|
test_return_data = """<html>
|
||||||
|
<body>
|
||||||
|
Some initial text</br>
|
||||||
|
<p>Which is across multiple lines</p>
|
||||||
|
</br>
|
||||||
|
So let's see what happens. </br>
|
||||||
|
<div id="nope-doesnt-exist">Some text thats the same</div>
|
||||||
|
</body>
|
||||||
|
</html>
|
||||||
|
"""
|
||||||
|
|
||||||
|
with open("test-datastore/endpoint-content.txt", "w") as f:
|
||||||
|
f.write(test_return_data)
|
||||||
|
return None
|
||||||
|
|
||||||
|
|
||||||
|
def set_response_with_filter():
|
||||||
|
test_return_data = """<html>
|
||||||
|
<body>
|
||||||
|
Some initial text</br>
|
||||||
|
<p>Which is across multiple lines</p>
|
||||||
|
</br>
|
||||||
|
So let's see what happens. </br>
|
||||||
|
<div class="ticket-available">Ticket now on sale!</div>
|
||||||
|
</body>
|
||||||
|
</html>
|
||||||
|
"""
|
||||||
|
|
||||||
|
with open("test-datastore/endpoint-content.txt", "w") as f:
|
||||||
|
f.write(test_return_data)
|
||||||
|
return None
|
||||||
|
|
||||||
|
def test_filter_doesnt_exist_then_exists_should_get_notification(client, live_server):
|
||||||
|
# Filter knowingly doesn't exist, like someone setting up a known filter to see if some cinema tickets are on sale again
|
||||||
|
# And the page has that filter available
|
||||||
|
# Then I should get a notification
|
||||||
|
|
||||||
|
live_server_setup(live_server)
|
||||||
|
|
||||||
|
# Give the endpoint time to spin up
|
||||||
|
time.sleep(1)
|
||||||
|
set_response_without_filter()
|
||||||
|
|
||||||
|
# Add our URL to the import page
|
||||||
|
test_url = url_for('test_endpoint', _external=True)
|
||||||
|
res = client.post(
|
||||||
|
url_for("form_quick_watch_add"),
|
||||||
|
data={"url": test_url, "tag": 'cinema'},
|
||||||
|
follow_redirects=True
|
||||||
|
)
|
||||||
|
assert b"Watch added" in res.data
|
||||||
|
|
||||||
|
# Give the thread time to pick up the first version
|
||||||
|
time.sleep(3)
|
||||||
|
|
||||||
|
# Goto the edit page, add our ignore text
|
||||||
|
# Add our URL to the import page
|
||||||
|
url = url_for('test_notification_endpoint', _external=True)
|
||||||
|
notification_url = url.replace('http', 'json')
|
||||||
|
|
||||||
|
print(">>>> Notification URL: " + notification_url)
|
||||||
|
|
||||||
|
# Just a regular notification setting, this will be used by the special 'filter not found' notification
|
||||||
|
notification_form_data = {"notification_urls": notification_url,
|
||||||
|
"notification_title": "New ChangeDetection.io Notification - {watch_url}",
|
||||||
|
"notification_body": "BASE URL: {base_url}\n"
|
||||||
|
"Watch URL: {watch_url}\n"
|
||||||
|
"Watch UUID: {watch_uuid}\n"
|
||||||
|
"Watch title: {watch_title}\n"
|
||||||
|
"Watch tag: {watch_tag}\n"
|
||||||
|
"Preview: {preview_url}\n"
|
||||||
|
"Diff URL: {diff_url}\n"
|
||||||
|
"Snapshot: {current_snapshot}\n"
|
||||||
|
"Diff: {diff}\n"
|
||||||
|
"Diff Full: {diff_full}\n"
|
||||||
|
":-)",
|
||||||
|
"notification_format": "Text"}
|
||||||
|
|
||||||
|
notification_form_data.update({
|
||||||
|
"url": test_url,
|
||||||
|
"tag": "my tag",
|
||||||
|
"title": "my title",
|
||||||
|
"headers": "",
|
||||||
|
"include_filters": '.ticket-available',
|
||||||
|
"fetch_backend": "html_requests"})
|
||||||
|
|
||||||
|
res = client.post(
|
||||||
|
url_for("edit_page", uuid="first"),
|
||||||
|
data=notification_form_data,
|
||||||
|
follow_redirects=True
|
||||||
|
)
|
||||||
|
assert b"Updated watch." in res.data
|
||||||
|
time.sleep(3)
|
||||||
|
|
||||||
|
# Shouldn't exist, shouldn't have fired
|
||||||
|
assert not os.path.isfile("test-datastore/notification.txt")
|
||||||
|
# Now the filter should exist
|
||||||
|
set_response_with_filter()
|
||||||
|
client.get(url_for("form_watch_checknow"), follow_redirects=True)
|
||||||
|
time.sleep(3)
|
||||||
|
|
||||||
|
assert os.path.isfile("test-datastore/notification.txt")
|
||||||
|
|
||||||
|
with open("test-datastore/notification.txt", 'r') as f:
|
||||||
|
notification = f.read()
|
||||||
|
|
||||||
|
assert 'Ticket now on sale' in notification
|
||||||
|
os.unlink("test-datastore/notification.txt")
|
||||||
|
|
||||||
|
|
||||||
|
# Test that if it gets removed, then re-added, we get a notification
|
||||||
|
# Remove the target and re-add it, we should get a new notification
|
||||||
|
set_response_without_filter()
|
||||||
|
client.get(url_for("form_watch_checknow"), follow_redirects=True)
|
||||||
|
time.sleep(3)
|
||||||
|
assert not os.path.isfile("test-datastore/notification.txt")
|
||||||
|
|
||||||
|
set_response_with_filter()
|
||||||
|
client.get(url_for("form_watch_checknow"), follow_redirects=True)
|
||||||
|
time.sleep(3)
|
||||||
|
assert os.path.isfile("test-datastore/notification.txt")
|
||||||
|
|
||||||
|
# Also test that the filter was updated after the first one was requested
|
||||||
144
changedetectionio/tests/test_filter_failure_notification.py
Normal file
144
changedetectionio/tests/test_filter_failure_notification.py
Normal file
@@ -0,0 +1,144 @@
|
|||||||
|
import os
|
||||||
|
import time
|
||||||
|
import re
|
||||||
|
from flask import url_for
|
||||||
|
from .util import set_original_response, live_server_setup
|
||||||
|
from changedetectionio.model import App
|
||||||
|
|
||||||
|
|
||||||
|
def set_response_with_filter():
|
||||||
|
test_return_data = """<html>
|
||||||
|
<body>
|
||||||
|
Some initial text</br>
|
||||||
|
<p>Which is across multiple lines</p>
|
||||||
|
</br>
|
||||||
|
So let's see what happens. </br>
|
||||||
|
<div id="nope-doesnt-exist">Some text thats the same</div>
|
||||||
|
</body>
|
||||||
|
</html>
|
||||||
|
"""
|
||||||
|
|
||||||
|
with open("test-datastore/endpoint-content.txt", "w") as f:
|
||||||
|
f.write(test_return_data)
|
||||||
|
return None
|
||||||
|
|
||||||
|
def run_filter_test(client, content_filter):
|
||||||
|
|
||||||
|
# Give the endpoint time to spin up
|
||||||
|
time.sleep(1)
|
||||||
|
# cleanup for the next
|
||||||
|
client.get(
|
||||||
|
url_for("form_delete", uuid="all"),
|
||||||
|
follow_redirects=True
|
||||||
|
)
|
||||||
|
if os.path.isfile("test-datastore/notification.txt"):
|
||||||
|
os.unlink("test-datastore/notification.txt")
|
||||||
|
|
||||||
|
# Add our URL to the import page
|
||||||
|
test_url = url_for('test_endpoint', _external=True)
|
||||||
|
res = client.post(
|
||||||
|
url_for("form_quick_watch_add"),
|
||||||
|
data={"url": test_url, "tag": ''},
|
||||||
|
follow_redirects=True
|
||||||
|
)
|
||||||
|
|
||||||
|
assert b"Watch added" in res.data
|
||||||
|
|
||||||
|
# Give the thread time to pick up the first version
|
||||||
|
time.sleep(3)
|
||||||
|
|
||||||
|
# Goto the edit page, add our ignore text
|
||||||
|
# Add our URL to the import page
|
||||||
|
url = url_for('test_notification_endpoint', _external=True)
|
||||||
|
notification_url = url.replace('http', 'json')
|
||||||
|
|
||||||
|
print(">>>> Notification URL: " + notification_url)
|
||||||
|
|
||||||
|
# Just a regular notification setting, this will be used by the special 'filter not found' notification
|
||||||
|
notification_form_data = {"notification_urls": notification_url,
|
||||||
|
"notification_title": "New ChangeDetection.io Notification - {watch_url}",
|
||||||
|
"notification_body": "BASE URL: {base_url}\n"
|
||||||
|
"Watch URL: {watch_url}\n"
|
||||||
|
"Watch UUID: {watch_uuid}\n"
|
||||||
|
"Watch title: {watch_title}\n"
|
||||||
|
"Watch tag: {watch_tag}\n"
|
||||||
|
"Preview: {preview_url}\n"
|
||||||
|
"Diff URL: {diff_url}\n"
|
||||||
|
"Snapshot: {current_snapshot}\n"
|
||||||
|
"Diff: {diff}\n"
|
||||||
|
"Diff Full: {diff_full}\n"
|
||||||
|
":-)",
|
||||||
|
"notification_format": "Text"}
|
||||||
|
|
||||||
|
notification_form_data.update({
|
||||||
|
"url": test_url,
|
||||||
|
"tag": "my tag",
|
||||||
|
"title": "my title",
|
||||||
|
"headers": "",
|
||||||
|
"filter_failure_notification_send": 'y',
|
||||||
|
"include_filters": content_filter,
|
||||||
|
"fetch_backend": "html_requests"})
|
||||||
|
|
||||||
|
res = client.post(
|
||||||
|
url_for("edit_page", uuid="first"),
|
||||||
|
data=notification_form_data,
|
||||||
|
follow_redirects=True
|
||||||
|
)
|
||||||
|
assert b"Updated watch." in res.data
|
||||||
|
time.sleep(3)
|
||||||
|
|
||||||
|
# Now the notification should not exist, because we didnt reach the threshold
|
||||||
|
assert not os.path.isfile("test-datastore/notification.txt")
|
||||||
|
|
||||||
|
for i in range(0, App._FILTER_FAILURE_THRESHOLD_ATTEMPTS_DEFAULT):
|
||||||
|
res = client.get(url_for("form_watch_checknow"), follow_redirects=True)
|
||||||
|
time.sleep(3)
|
||||||
|
|
||||||
|
# We should see something in the frontend
|
||||||
|
assert b'Warning, no filters were found' in res.data
|
||||||
|
|
||||||
|
# Now it should exist and contain our "filter not found" alert
|
||||||
|
assert os.path.isfile("test-datastore/notification.txt")
|
||||||
|
notification = False
|
||||||
|
with open("test-datastore/notification.txt", 'r') as f:
|
||||||
|
notification = f.read()
|
||||||
|
assert 'CSS/xPath filter was not present in the page' in notification
|
||||||
|
assert content_filter.replace('"', '\\"') in notification
|
||||||
|
|
||||||
|
# Remove it and prove that it doesnt trigger when not expected
|
||||||
|
os.unlink("test-datastore/notification.txt")
|
||||||
|
set_response_with_filter()
|
||||||
|
|
||||||
|
for i in range(0, App._FILTER_FAILURE_THRESHOLD_ATTEMPTS_DEFAULT):
|
||||||
|
client.get(url_for("form_watch_checknow"), follow_redirects=True)
|
||||||
|
time.sleep(3)
|
||||||
|
|
||||||
|
# It should have sent a notification, but..
|
||||||
|
assert os.path.isfile("test-datastore/notification.txt")
|
||||||
|
# but it should not contain the info about the failed filter
|
||||||
|
with open("test-datastore/notification.txt", 'r') as f:
|
||||||
|
notification = f.read()
|
||||||
|
assert not 'CSS/xPath filter was not present in the page' in notification
|
||||||
|
|
||||||
|
# cleanup for the next
|
||||||
|
client.get(
|
||||||
|
url_for("form_delete", uuid="all"),
|
||||||
|
follow_redirects=True
|
||||||
|
)
|
||||||
|
os.unlink("test-datastore/notification.txt")
|
||||||
|
|
||||||
|
|
||||||
|
def test_setup(live_server):
|
||||||
|
live_server_setup(live_server)
|
||||||
|
|
||||||
|
def test_check_include_filters_failure_notification(client, live_server):
|
||||||
|
set_original_response()
|
||||||
|
time.sleep(1)
|
||||||
|
run_filter_test(client, '#nope-doesnt-exist')
|
||||||
|
|
||||||
|
def test_check_xpath_filter_failure_notification(client, live_server):
|
||||||
|
set_original_response()
|
||||||
|
time.sleep(1)
|
||||||
|
run_filter_test(client, '//*[@id="nope-doesnt-exist"]')
|
||||||
|
|
||||||
|
# Test that notification is never sent
|
||||||
84
changedetectionio/tests/test_history_consistency.py
Normal file
84
changedetectionio/tests/test_history_consistency.py
Normal file
@@ -0,0 +1,84 @@
|
|||||||
|
#!/usr/bin/python3
|
||||||
|
|
||||||
|
import time
|
||||||
|
import os
|
||||||
|
import json
|
||||||
|
import logging
|
||||||
|
from flask import url_for
|
||||||
|
from .util import live_server_setup
|
||||||
|
from urllib.parse import urlparse, parse_qs
|
||||||
|
|
||||||
|
def test_consistent_history(client, live_server):
|
||||||
|
live_server_setup(live_server)
|
||||||
|
|
||||||
|
# Give the endpoint time to spin up
|
||||||
|
time.sleep(1)
|
||||||
|
r = range(1, 50)
|
||||||
|
|
||||||
|
for one in r:
|
||||||
|
test_url = url_for('test_endpoint', content_type="text/html", content=str(one), _external=True)
|
||||||
|
res = client.post(
|
||||||
|
url_for("import_page"),
|
||||||
|
data={"urls": test_url},
|
||||||
|
follow_redirects=True
|
||||||
|
)
|
||||||
|
|
||||||
|
assert b"1 Imported" in res.data
|
||||||
|
|
||||||
|
time.sleep(3)
|
||||||
|
while True:
|
||||||
|
res = client.get(url_for("index"))
|
||||||
|
logging.debug("Waiting for 'Checking now' to go away..")
|
||||||
|
if b'Checking now' not in res.data:
|
||||||
|
break
|
||||||
|
time.sleep(0.5)
|
||||||
|
|
||||||
|
time.sleep(3)
|
||||||
|
# Essentially just triggers the DB write/update
|
||||||
|
res = client.post(
|
||||||
|
url_for("settings_page"),
|
||||||
|
data={"application-empty_pages_are_a_change": "",
|
||||||
|
"requests-time_between_check-minutes": 180,
|
||||||
|
'application-fetch_backend': "html_requests"},
|
||||||
|
follow_redirects=True
|
||||||
|
)
|
||||||
|
assert b"Settings updated." in res.data
|
||||||
|
|
||||||
|
# Give it time to write it out
|
||||||
|
time.sleep(3)
|
||||||
|
json_db_file = os.path.join(live_server.app.config['DATASTORE'].datastore_path, 'url-watches.json')
|
||||||
|
|
||||||
|
json_obj = None
|
||||||
|
with open(json_db_file, 'r') as f:
|
||||||
|
json_obj = json.load(f)
|
||||||
|
|
||||||
|
# assert the right amount of watches was found in the JSON
|
||||||
|
assert len(json_obj['watching']) == len(r), "Correct number of watches was found in the JSON"
|
||||||
|
|
||||||
|
# each one should have a history.txt containing just one line
|
||||||
|
for w in json_obj['watching'].keys():
|
||||||
|
history_txt_index_file = os.path.join(live_server.app.config['DATASTORE'].datastore_path, w, 'history.txt')
|
||||||
|
assert os.path.isfile(history_txt_index_file), "History.txt should exist where I expect it - {}".format(history_txt_index_file)
|
||||||
|
|
||||||
|
# Same like in model.Watch
|
||||||
|
with open(history_txt_index_file, "r") as f:
|
||||||
|
tmp_history = dict(i.strip().split(',', 2) for i in f.readlines())
|
||||||
|
assert len(tmp_history) == 1, "History.txt should contain 1 line"
|
||||||
|
|
||||||
|
# Should be two files,. the history.txt , and the snapshot.txt
|
||||||
|
files_in_watch_dir = os.listdir(os.path.join(live_server.app.config['DATASTORE'].datastore_path,
|
||||||
|
w))
|
||||||
|
# Find the snapshot one
|
||||||
|
for fname in files_in_watch_dir:
|
||||||
|
if fname != 'history.txt':
|
||||||
|
# contents should match what we requested as content returned from the test url
|
||||||
|
with open(os.path.join(live_server.app.config['DATASTORE'].datastore_path, w, fname), 'r') as snapshot_f:
|
||||||
|
contents = snapshot_f.read()
|
||||||
|
watch_url = json_obj['watching'][w]['url']
|
||||||
|
u = urlparse(watch_url)
|
||||||
|
q = parse_qs(u[4])
|
||||||
|
assert q['content'][0] == contents.strip(), "Snapshot file {} should contain {}".format(fname, q['content'][0])
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
|
assert len(files_in_watch_dir) == 2, "Should be just two files in the dir, history.txt and the snapshot"
|
||||||
@@ -102,7 +102,7 @@ def test_check_ignore_text_functionality(client, live_server):
|
|||||||
assert b"1 Imported" in res.data
|
assert b"1 Imported" in res.data
|
||||||
|
|
||||||
# Trigger a check
|
# Trigger a check
|
||||||
client.get(url_for("api_watch_checknow"), follow_redirects=True)
|
client.get(url_for("form_watch_checknow"), follow_redirects=True)
|
||||||
|
|
||||||
# Give the thread time to pick it up
|
# Give the thread time to pick it up
|
||||||
time.sleep(sleep_time_for_fetch_thread)
|
time.sleep(sleep_time_for_fetch_thread)
|
||||||
@@ -123,7 +123,7 @@ def test_check_ignore_text_functionality(client, live_server):
|
|||||||
assert bytes(ignore_text.encode('utf-8')) in res.data
|
assert bytes(ignore_text.encode('utf-8')) in res.data
|
||||||
|
|
||||||
# Trigger a check
|
# Trigger a check
|
||||||
client.get(url_for("api_watch_checknow"), follow_redirects=True)
|
client.get(url_for("form_watch_checknow"), follow_redirects=True)
|
||||||
|
|
||||||
# Give the thread time to pick it up
|
# Give the thread time to pick it up
|
||||||
time.sleep(sleep_time_for_fetch_thread)
|
time.sleep(sleep_time_for_fetch_thread)
|
||||||
@@ -137,7 +137,7 @@ def test_check_ignore_text_functionality(client, live_server):
|
|||||||
set_modified_ignore_response()
|
set_modified_ignore_response()
|
||||||
|
|
||||||
# Trigger a check
|
# Trigger a check
|
||||||
client.get(url_for("api_watch_checknow"), follow_redirects=True)
|
client.get(url_for("form_watch_checknow"), follow_redirects=True)
|
||||||
# Give the thread time to pick it up
|
# Give the thread time to pick it up
|
||||||
time.sleep(sleep_time_for_fetch_thread)
|
time.sleep(sleep_time_for_fetch_thread)
|
||||||
|
|
||||||
@@ -152,7 +152,7 @@ def test_check_ignore_text_functionality(client, live_server):
|
|||||||
|
|
||||||
# Just to be sure.. set a regular modified change..
|
# Just to be sure.. set a regular modified change..
|
||||||
set_modified_original_ignore_response()
|
set_modified_original_ignore_response()
|
||||||
client.get(url_for("api_watch_checknow"), follow_redirects=True)
|
client.get(url_for("form_watch_checknow"), follow_redirects=True)
|
||||||
time.sleep(sleep_time_for_fetch_thread)
|
time.sleep(sleep_time_for_fetch_thread)
|
||||||
|
|
||||||
res = client.get(url_for("index"))
|
res = client.get(url_for("index"))
|
||||||
@@ -165,7 +165,7 @@ def test_check_ignore_text_functionality(client, live_server):
|
|||||||
# We should be able to see what we ignored
|
# We should be able to see what we ignored
|
||||||
assert b'<div class="ignored">new ignore stuff' in res.data
|
assert b'<div class="ignored">new ignore stuff' in res.data
|
||||||
|
|
||||||
res = client.get(url_for("api_delete", uuid="all"), follow_redirects=True)
|
res = client.get(url_for("form_delete", uuid="all"), follow_redirects=True)
|
||||||
assert b'Deleted' in res.data
|
assert b'Deleted' in res.data
|
||||||
|
|
||||||
def test_check_global_ignore_text_functionality(client, live_server):
|
def test_check_global_ignore_text_functionality(client, live_server):
|
||||||
@@ -200,7 +200,7 @@ def test_check_global_ignore_text_functionality(client, live_server):
|
|||||||
assert b"1 Imported" in res.data
|
assert b"1 Imported" in res.data
|
||||||
|
|
||||||
# Trigger a check
|
# Trigger a check
|
||||||
client.get(url_for("api_watch_checknow"), follow_redirects=True)
|
client.get(url_for("form_watch_checknow"), follow_redirects=True)
|
||||||
|
|
||||||
# Give the thread time to pick it up
|
# Give the thread time to pick it up
|
||||||
time.sleep(sleep_time_for_fetch_thread)
|
time.sleep(sleep_time_for_fetch_thread)
|
||||||
@@ -222,7 +222,7 @@ def test_check_global_ignore_text_functionality(client, live_server):
|
|||||||
assert bytes(ignore_text.encode('utf-8')) in res.data
|
assert bytes(ignore_text.encode('utf-8')) in res.data
|
||||||
|
|
||||||
# Trigger a check
|
# Trigger a check
|
||||||
client.get(url_for("api_watch_checknow"), follow_redirects=True)
|
client.get(url_for("form_watch_checknow"), follow_redirects=True)
|
||||||
|
|
||||||
# Give the thread time to pick it up
|
# Give the thread time to pick it up
|
||||||
time.sleep(sleep_time_for_fetch_thread)
|
time.sleep(sleep_time_for_fetch_thread)
|
||||||
@@ -240,7 +240,7 @@ def test_check_global_ignore_text_functionality(client, live_server):
|
|||||||
set_modified_ignore_response()
|
set_modified_ignore_response()
|
||||||
|
|
||||||
# Trigger a check
|
# Trigger a check
|
||||||
client.get(url_for("api_watch_checknow"), follow_redirects=True)
|
client.get(url_for("form_watch_checknow"), follow_redirects=True)
|
||||||
# Give the thread time to pick it up
|
# Give the thread time to pick it up
|
||||||
time.sleep(sleep_time_for_fetch_thread)
|
time.sleep(sleep_time_for_fetch_thread)
|
||||||
|
|
||||||
@@ -251,10 +251,10 @@ def test_check_global_ignore_text_functionality(client, live_server):
|
|||||||
|
|
||||||
# Just to be sure.. set a regular modified change that will trigger it
|
# Just to be sure.. set a regular modified change that will trigger it
|
||||||
set_modified_original_ignore_response()
|
set_modified_original_ignore_response()
|
||||||
client.get(url_for("api_watch_checknow"), follow_redirects=True)
|
client.get(url_for("form_watch_checknow"), follow_redirects=True)
|
||||||
time.sleep(sleep_time_for_fetch_thread)
|
time.sleep(sleep_time_for_fetch_thread)
|
||||||
res = client.get(url_for("index"))
|
res = client.get(url_for("index"))
|
||||||
assert b'unviewed' in res.data
|
assert b'unviewed' in res.data
|
||||||
|
|
||||||
res = client.get(url_for("api_delete", uuid="all"), follow_redirects=True)
|
res = client.get(url_for("form_delete", uuid="all"), follow_redirects=True)
|
||||||
assert b'Deleted' in res.data
|
assert b'Deleted' in res.data
|
||||||
|
|||||||
@@ -72,14 +72,14 @@ def test_render_anchor_tag_content_true(client, live_server):
|
|||||||
|
|
||||||
time.sleep(sleep_time_for_fetch_thread)
|
time.sleep(sleep_time_for_fetch_thread)
|
||||||
# Trigger a check
|
# Trigger a check
|
||||||
client.get(url_for("api_watch_checknow"), follow_redirects=True)
|
client.get(url_for("form_watch_checknow"), follow_redirects=True)
|
||||||
|
|
||||||
# set a new html text with a modified link
|
# set a new html text with a modified link
|
||||||
set_modified_ignore_response()
|
set_modified_ignore_response()
|
||||||
time.sleep(sleep_time_for_fetch_thread)
|
time.sleep(sleep_time_for_fetch_thread)
|
||||||
|
|
||||||
# Trigger a check
|
# Trigger a check
|
||||||
client.get(url_for("api_watch_checknow"), follow_redirects=True)
|
client.get(url_for("form_watch_checknow"), follow_redirects=True)
|
||||||
|
|
||||||
# Give the thread time to pick it up
|
# Give the thread time to pick it up
|
||||||
time.sleep(sleep_time_for_fetch_thread)
|
time.sleep(sleep_time_for_fetch_thread)
|
||||||
@@ -101,7 +101,7 @@ def test_render_anchor_tag_content_true(client, live_server):
|
|||||||
assert b"Settings updated." in res.data
|
assert b"Settings updated." in res.data
|
||||||
|
|
||||||
# Trigger a check
|
# Trigger a check
|
||||||
client.get(url_for("api_watch_checknow"), follow_redirects=True)
|
client.get(url_for("form_watch_checknow"), follow_redirects=True)
|
||||||
|
|
||||||
# Give the thread time to pick it up
|
# Give the thread time to pick it up
|
||||||
time.sleep(sleep_time_for_fetch_thread)
|
time.sleep(sleep_time_for_fetch_thread)
|
||||||
@@ -119,7 +119,7 @@ def test_render_anchor_tag_content_true(client, live_server):
|
|||||||
assert b"/test-endpoint" in res.data
|
assert b"/test-endpoint" in res.data
|
||||||
|
|
||||||
# Cleanup everything
|
# Cleanup everything
|
||||||
res = client.get(url_for("api_delete", uuid="all"),
|
res = client.get(url_for("form_delete", uuid="all"),
|
||||||
follow_redirects=True)
|
follow_redirects=True)
|
||||||
assert b'Deleted' in res.data
|
assert b'Deleted' in res.data
|
||||||
|
|
||||||
|
|||||||
@@ -70,12 +70,12 @@ def test_normal_page_check_works_with_ignore_status_code(client, live_server):
|
|||||||
|
|
||||||
time.sleep(sleep_time_for_fetch_thread)
|
time.sleep(sleep_time_for_fetch_thread)
|
||||||
# Trigger a check
|
# Trigger a check
|
||||||
client.get(url_for("api_watch_checknow"), follow_redirects=True)
|
client.get(url_for("form_watch_checknow"), follow_redirects=True)
|
||||||
|
|
||||||
set_some_changed_response()
|
set_some_changed_response()
|
||||||
time.sleep(sleep_time_for_fetch_thread)
|
time.sleep(sleep_time_for_fetch_thread)
|
||||||
# Trigger a check
|
# Trigger a check
|
||||||
client.get(url_for("api_watch_checknow"), follow_redirects=True)
|
client.get(url_for("form_watch_checknow"), follow_redirects=True)
|
||||||
|
|
||||||
# Give the thread time to pick it up
|
# Give the thread time to pick it up
|
||||||
time.sleep(sleep_time_for_fetch_thread)
|
time.sleep(sleep_time_for_fetch_thread)
|
||||||
@@ -105,7 +105,7 @@ def test_403_page_check_works_with_ignore_status_code(client, live_server):
|
|||||||
assert b"1 Imported" in res.data
|
assert b"1 Imported" in res.data
|
||||||
|
|
||||||
# Trigger a check
|
# Trigger a check
|
||||||
client.get(url_for("api_watch_checknow"), follow_redirects=True)
|
client.get(url_for("form_watch_checknow"), follow_redirects=True)
|
||||||
|
|
||||||
# Give the thread time to pick it up
|
# Give the thread time to pick it up
|
||||||
time.sleep(sleep_time_for_fetch_thread)
|
time.sleep(sleep_time_for_fetch_thread)
|
||||||
@@ -120,7 +120,7 @@ def test_403_page_check_works_with_ignore_status_code(client, live_server):
|
|||||||
assert b"Updated watch." in res.data
|
assert b"Updated watch." in res.data
|
||||||
|
|
||||||
# Trigger a check
|
# Trigger a check
|
||||||
client.get(url_for("api_watch_checknow"), follow_redirects=True)
|
client.get(url_for("form_watch_checknow"), follow_redirects=True)
|
||||||
|
|
||||||
# Give the thread time to pick it up
|
# Give the thread time to pick it up
|
||||||
time.sleep(sleep_time_for_fetch_thread)
|
time.sleep(sleep_time_for_fetch_thread)
|
||||||
@@ -128,7 +128,7 @@ def test_403_page_check_works_with_ignore_status_code(client, live_server):
|
|||||||
set_some_changed_response()
|
set_some_changed_response()
|
||||||
|
|
||||||
# Trigger a check
|
# Trigger a check
|
||||||
client.get(url_for("api_watch_checknow"), follow_redirects=True)
|
client.get(url_for("form_watch_checknow"), follow_redirects=True)
|
||||||
# Give the thread time to pick it up
|
# Give the thread time to pick it up
|
||||||
time.sleep(sleep_time_for_fetch_thread)
|
time.sleep(sleep_time_for_fetch_thread)
|
||||||
|
|
||||||
@@ -137,54 +137,3 @@ def test_403_page_check_works_with_ignore_status_code(client, live_server):
|
|||||||
res = client.get(url_for("index"))
|
res = client.get(url_for("index"))
|
||||||
assert b'unviewed' in res.data
|
assert b'unviewed' in res.data
|
||||||
|
|
||||||
|
|
||||||
# Tests the whole stack works with staus codes ignored
|
|
||||||
def test_403_page_check_fails_without_ignore_status_code(client, live_server):
|
|
||||||
sleep_time_for_fetch_thread = 3
|
|
||||||
|
|
||||||
set_original_response()
|
|
||||||
|
|
||||||
# Give the endpoint time to spin up
|
|
||||||
time.sleep(1)
|
|
||||||
|
|
||||||
# Add our URL to the import page
|
|
||||||
test_url = url_for('test_endpoint', status_code=403, _external=True)
|
|
||||||
res = client.post(
|
|
||||||
url_for("import_page"),
|
|
||||||
data={"urls": test_url},
|
|
||||||
follow_redirects=True
|
|
||||||
)
|
|
||||||
assert b"1 Imported" in res.data
|
|
||||||
|
|
||||||
# Trigger a check
|
|
||||||
client.get(url_for("api_watch_checknow"), follow_redirects=True)
|
|
||||||
|
|
||||||
# Give the thread time to pick it up
|
|
||||||
time.sleep(sleep_time_for_fetch_thread)
|
|
||||||
|
|
||||||
# Goto the edit page, check our ignore option
|
|
||||||
# Add our URL to the import page
|
|
||||||
res = client.post(
|
|
||||||
url_for("edit_page", uuid="first"),
|
|
||||||
data={"url": test_url, "tag": "", "headers": "", 'fetch_backend': "html_requests"},
|
|
||||||
follow_redirects=True
|
|
||||||
)
|
|
||||||
assert b"Updated watch." in res.data
|
|
||||||
|
|
||||||
# Trigger a check
|
|
||||||
client.get(url_for("api_watch_checknow"), follow_redirects=True)
|
|
||||||
|
|
||||||
# Give the thread time to pick it up
|
|
||||||
time.sleep(sleep_time_for_fetch_thread)
|
|
||||||
# Make a change
|
|
||||||
set_some_changed_response()
|
|
||||||
|
|
||||||
# Trigger a check
|
|
||||||
client.get(url_for("api_watch_checknow"), follow_redirects=True)
|
|
||||||
# Give the thread time to pick it up
|
|
||||||
time.sleep(sleep_time_for_fetch_thread)
|
|
||||||
|
|
||||||
# It should have 'unviewed' still
|
|
||||||
# Because it should be looking at only that 'sametext' id
|
|
||||||
res = client.get(url_for("index"))
|
|
||||||
assert b'Status Code 403' in res.data
|
|
||||||
|
|||||||
@@ -80,12 +80,12 @@ def test_check_ignore_whitespace(client, live_server):
|
|||||||
|
|
||||||
time.sleep(sleep_time_for_fetch_thread)
|
time.sleep(sleep_time_for_fetch_thread)
|
||||||
# Trigger a check
|
# Trigger a check
|
||||||
client.get(url_for("api_watch_checknow"), follow_redirects=True)
|
client.get(url_for("form_watch_checknow"), follow_redirects=True)
|
||||||
|
|
||||||
set_original_ignore_response_but_with_whitespace()
|
set_original_ignore_response_but_with_whitespace()
|
||||||
time.sleep(sleep_time_for_fetch_thread)
|
time.sleep(sleep_time_for_fetch_thread)
|
||||||
# Trigger a check
|
# Trigger a check
|
||||||
client.get(url_for("api_watch_checknow"), follow_redirects=True)
|
client.get(url_for("form_watch_checknow"), follow_redirects=True)
|
||||||
|
|
||||||
# Give the thread time to pick it up
|
# Give the thread time to pick it up
|
||||||
time.sleep(sleep_time_for_fetch_thread)
|
time.sleep(sleep_time_for_fetch_thread)
|
||||||
|
|||||||
@@ -25,7 +25,7 @@ https://example.com tag1, other tag"""
|
|||||||
assert b"3 Imported" in res.data
|
assert b"3 Imported" in res.data
|
||||||
assert b"tag1" in res.data
|
assert b"tag1" in res.data
|
||||||
assert b"other tag" in res.data
|
assert b"other tag" in res.data
|
||||||
res = client.get(url_for("api_delete", uuid="all"), follow_redirects=True)
|
res = client.get(url_for("form_delete", uuid="all"), follow_redirects=True)
|
||||||
|
|
||||||
# Clear flask alerts
|
# Clear flask alerts
|
||||||
res = client.get( url_for("index"))
|
res = client.get( url_for("index"))
|
||||||
@@ -50,7 +50,7 @@ def xtest_import_skip_url(client, live_server):
|
|||||||
assert b"1 Imported" in res.data
|
assert b"1 Imported" in res.data
|
||||||
assert b"ht000000broken" in res.data
|
assert b"ht000000broken" in res.data
|
||||||
assert b"1 Skipped" in res.data
|
assert b"1 Skipped" in res.data
|
||||||
res = client.get(url_for("api_delete", uuid="all"), follow_redirects=True)
|
res = client.get(url_for("form_delete", uuid="all"), follow_redirects=True)
|
||||||
# Clear flask alerts
|
# Clear flask alerts
|
||||||
res = client.get( url_for("index"))
|
res = client.get( url_for("index"))
|
||||||
|
|
||||||
@@ -79,7 +79,7 @@ def test_import_distillio(client, live_server):
|
|||||||
|
|
||||||
# Give the endpoint time to spin up
|
# Give the endpoint time to spin up
|
||||||
time.sleep(1)
|
time.sleep(1)
|
||||||
client.get(url_for("api_delete", uuid="all"), follow_redirects=True)
|
client.get(url_for("form_delete", uuid="all"), follow_redirects=True)
|
||||||
res = client.post(
|
res = client.post(
|
||||||
url_for("import_page"),
|
url_for("import_page"),
|
||||||
data={
|
data={
|
||||||
@@ -115,6 +115,6 @@ def test_import_distillio(client, live_server):
|
|||||||
assert b"nice stuff" in res.data
|
assert b"nice stuff" in res.data
|
||||||
assert b"nerd-news" in res.data
|
assert b"nerd-news" in res.data
|
||||||
|
|
||||||
res = client.get(url_for("api_delete", uuid="all"), follow_redirects=True)
|
res = client.get(url_for("form_delete", uuid="all"), follow_redirects=True)
|
||||||
# Clear flask alerts
|
# Clear flask alerts
|
||||||
res = client.get(url_for("index"))
|
res = client.get(url_for("index"))
|
||||||
|
|||||||
33
changedetectionio/tests/test_jinja2.py
Normal file
33
changedetectionio/tests/test_jinja2.py
Normal file
@@ -0,0 +1,33 @@
|
|||||||
|
#!/usr/bin/python3
|
||||||
|
|
||||||
|
import time
|
||||||
|
from flask import url_for
|
||||||
|
from .util import live_server_setup
|
||||||
|
|
||||||
|
|
||||||
|
# If there was only a change in the whitespacing, then we shouldnt have a change detected
|
||||||
|
def test_jinja2_in_url_query(client, live_server):
|
||||||
|
live_server_setup(live_server)
|
||||||
|
|
||||||
|
# Give the endpoint time to spin up
|
||||||
|
time.sleep(1)
|
||||||
|
|
||||||
|
# Add our URL to the import page
|
||||||
|
test_url = url_for('test_return_query', _external=True)
|
||||||
|
|
||||||
|
# because url_for() will URL-encode the var, but we dont here
|
||||||
|
full_url = "{}?{}".format(test_url,
|
||||||
|
"date={% now 'Europe/Berlin', '%Y' %}.{% now 'Europe/Berlin', '%m' %}.{% now 'Europe/Berlin', '%d' %}", )
|
||||||
|
res = client.post(
|
||||||
|
url_for("form_quick_watch_add"),
|
||||||
|
data={"url": full_url, "tag": "test"},
|
||||||
|
follow_redirects=True
|
||||||
|
)
|
||||||
|
assert b"Watch added" in res.data
|
||||||
|
time.sleep(3)
|
||||||
|
# It should report nothing found (no new 'unviewed' class)
|
||||||
|
res = client.get(
|
||||||
|
url_for("preview_page", uuid="first"),
|
||||||
|
follow_redirects=True
|
||||||
|
)
|
||||||
|
assert b'date=2' in res.data
|
||||||
@@ -2,10 +2,15 @@
|
|||||||
# coding=utf-8
|
# coding=utf-8
|
||||||
|
|
||||||
import time
|
import time
|
||||||
from flask import url_for
|
from flask import url_for, escape
|
||||||
from . util import live_server_setup
|
from . util import live_server_setup
|
||||||
import pytest
|
import pytest
|
||||||
|
jq_support = True
|
||||||
|
|
||||||
|
try:
|
||||||
|
import jq
|
||||||
|
except ModuleNotFoundError:
|
||||||
|
jq_support = False
|
||||||
|
|
||||||
def test_setup(live_server):
|
def test_setup(live_server):
|
||||||
live_server_setup(live_server)
|
live_server_setup(live_server)
|
||||||
@@ -36,16 +41,28 @@ and it can also be repeated
|
|||||||
from .. import html_tools
|
from .. import html_tools
|
||||||
|
|
||||||
# See that we can find the second <script> one, which is not broken, and matches our filter
|
# See that we can find the second <script> one, which is not broken, and matches our filter
|
||||||
text = html_tools.extract_json_as_string(content, "$.offers.price")
|
text = html_tools.extract_json_as_string(content, "json:$.offers.price")
|
||||||
assert text == "23.5"
|
assert text == "23.5"
|
||||||
|
|
||||||
text = html_tools.extract_json_as_string('{"id":5}', "$.id")
|
# also check for jq
|
||||||
|
if jq_support:
|
||||||
|
text = html_tools.extract_json_as_string(content, "jq:.offers.price")
|
||||||
|
assert text == "23.5"
|
||||||
|
|
||||||
|
text = html_tools.extract_json_as_string('{"id":5}', "jq:.id")
|
||||||
|
assert text == "5"
|
||||||
|
|
||||||
|
text = html_tools.extract_json_as_string('{"id":5}', "json:$.id")
|
||||||
assert text == "5"
|
assert text == "5"
|
||||||
|
|
||||||
# When nothing at all is found, it should throw JSONNOTFound
|
# When nothing at all is found, it should throw JSONNOTFound
|
||||||
# Which is caught and shown to the user in the watch-overview table
|
# Which is caught and shown to the user in the watch-overview table
|
||||||
with pytest.raises(html_tools.JSONNotFound) as e_info:
|
with pytest.raises(html_tools.JSONNotFound) as e_info:
|
||||||
html_tools.extract_json_as_string('COMPLETE GIBBERISH, NO JSON!', "$.id")
|
html_tools.extract_json_as_string('COMPLETE GIBBERISH, NO JSON!', "json:$.id")
|
||||||
|
|
||||||
|
if jq_support:
|
||||||
|
with pytest.raises(html_tools.JSONNotFound) as e_info:
|
||||||
|
html_tools.extract_json_as_string('COMPLETE GIBBERISH, NO JSON!', "jq:.id")
|
||||||
|
|
||||||
def set_original_ext_response():
|
def set_original_ext_response():
|
||||||
data = """
|
data = """
|
||||||
@@ -66,6 +83,7 @@ def set_original_ext_response():
|
|||||||
|
|
||||||
with open("test-datastore/endpoint-content.txt", "w") as f:
|
with open("test-datastore/endpoint-content.txt", "w") as f:
|
||||||
f.write(data)
|
f.write(data)
|
||||||
|
return None
|
||||||
|
|
||||||
def set_modified_ext_response():
|
def set_modified_ext_response():
|
||||||
data = """
|
data = """
|
||||||
@@ -86,6 +104,7 @@ def set_modified_ext_response():
|
|||||||
|
|
||||||
with open("test-datastore/endpoint-content.txt", "w") as f:
|
with open("test-datastore/endpoint-content.txt", "w") as f:
|
||||||
f.write(data)
|
f.write(data)
|
||||||
|
return None
|
||||||
|
|
||||||
def set_original_response():
|
def set_original_response():
|
||||||
test_return_data = """
|
test_return_data = """
|
||||||
@@ -113,7 +132,7 @@ def set_original_response():
|
|||||||
return None
|
return None
|
||||||
|
|
||||||
|
|
||||||
def set_response_with_html():
|
def set_json_response_with_html():
|
||||||
test_return_data = """
|
test_return_data = """
|
||||||
{
|
{
|
||||||
"test": [
|
"test": [
|
||||||
@@ -157,7 +176,7 @@ def set_modified_response():
|
|||||||
def test_check_json_without_filter(client, live_server):
|
def test_check_json_without_filter(client, live_server):
|
||||||
# Request a JSON document from a application/json source containing HTML
|
# Request a JSON document from a application/json source containing HTML
|
||||||
# and be sure it doesn't get chewed up by instriptis
|
# and be sure it doesn't get chewed up by instriptis
|
||||||
set_response_with_html()
|
set_json_response_with_html()
|
||||||
|
|
||||||
# Give the endpoint time to spin up
|
# Give the endpoint time to spin up
|
||||||
time.sleep(1)
|
time.sleep(1)
|
||||||
@@ -170,9 +189,6 @@ def test_check_json_without_filter(client, live_server):
|
|||||||
follow_redirects=True
|
follow_redirects=True
|
||||||
)
|
)
|
||||||
|
|
||||||
# Trigger a check
|
|
||||||
client.get(url_for("api_watch_checknow"), follow_redirects=True)
|
|
||||||
|
|
||||||
# Give the thread time to pick it up
|
# Give the thread time to pick it up
|
||||||
time.sleep(3)
|
time.sleep(3)
|
||||||
|
|
||||||
@@ -181,13 +197,14 @@ def test_check_json_without_filter(client, live_server):
|
|||||||
follow_redirects=True
|
follow_redirects=True
|
||||||
)
|
)
|
||||||
|
|
||||||
|
# Should still see '"html": "<b>"'
|
||||||
assert b'"<b>' in res.data
|
assert b'"<b>' in res.data
|
||||||
assert res.data.count(b'{\n') >= 2
|
assert res.data.count(b'{\n') >= 2
|
||||||
|
|
||||||
|
res = client.get(url_for("form_delete", uuid="all"), follow_redirects=True)
|
||||||
|
assert b'Deleted' in res.data
|
||||||
|
|
||||||
def test_check_json_filter(client, live_server):
|
def check_json_filter(json_filter, client, live_server):
|
||||||
json_filter = 'json:boss.name'
|
|
||||||
|
|
||||||
set_original_response()
|
set_original_response()
|
||||||
|
|
||||||
# Give the endpoint time to spin up
|
# Give the endpoint time to spin up
|
||||||
@@ -202,9 +219,6 @@ def test_check_json_filter(client, live_server):
|
|||||||
)
|
)
|
||||||
assert b"1 Imported" in res.data
|
assert b"1 Imported" in res.data
|
||||||
|
|
||||||
# Trigger a check
|
|
||||||
client.get(url_for("api_watch_checknow"), follow_redirects=True)
|
|
||||||
|
|
||||||
# Give the thread time to pick it up
|
# Give the thread time to pick it up
|
||||||
time.sleep(3)
|
time.sleep(3)
|
||||||
|
|
||||||
@@ -212,7 +226,7 @@ def test_check_json_filter(client, live_server):
|
|||||||
# Add our URL to the import page
|
# Add our URL to the import page
|
||||||
res = client.post(
|
res = client.post(
|
||||||
url_for("edit_page", uuid="first"),
|
url_for("edit_page", uuid="first"),
|
||||||
data={"css_filter": json_filter,
|
data={"include_filters": json_filter,
|
||||||
"url": test_url,
|
"url": test_url,
|
||||||
"tag": "",
|
"tag": "",
|
||||||
"headers": "",
|
"headers": "",
|
||||||
@@ -226,10 +240,7 @@ def test_check_json_filter(client, live_server):
|
|||||||
res = client.get(
|
res = client.get(
|
||||||
url_for("edit_page", uuid="first"),
|
url_for("edit_page", uuid="first"),
|
||||||
)
|
)
|
||||||
assert bytes(json_filter.encode('utf-8')) in res.data
|
assert bytes(escape(json_filter).encode('utf-8')) in res.data
|
||||||
|
|
||||||
# Trigger a check
|
|
||||||
client.get(url_for("api_watch_checknow"), follow_redirects=True)
|
|
||||||
|
|
||||||
# Give the thread time to pick it up
|
# Give the thread time to pick it up
|
||||||
time.sleep(3)
|
time.sleep(3)
|
||||||
@@ -237,7 +248,7 @@ def test_check_json_filter(client, live_server):
|
|||||||
set_modified_response()
|
set_modified_response()
|
||||||
|
|
||||||
# Trigger a check
|
# Trigger a check
|
||||||
client.get(url_for("api_watch_checknow"), follow_redirects=True)
|
client.get(url_for("form_watch_checknow"), follow_redirects=True)
|
||||||
# Give the thread time to pick it up
|
# Give the thread time to pick it up
|
||||||
time.sleep(4)
|
time.sleep(4)
|
||||||
|
|
||||||
@@ -252,10 +263,17 @@ def test_check_json_filter(client, live_server):
|
|||||||
# And #462 - check we see the proper utf-8 string there
|
# And #462 - check we see the proper utf-8 string there
|
||||||
assert "Örnsköldsvik".encode('utf-8') in res.data
|
assert "Örnsköldsvik".encode('utf-8') in res.data
|
||||||
|
|
||||||
|
res = client.get(url_for("form_delete", uuid="all"), follow_redirects=True)
|
||||||
|
assert b'Deleted' in res.data
|
||||||
|
|
||||||
def test_check_json_filter_bool_val(client, live_server):
|
def test_check_jsonpath_filter(client, live_server):
|
||||||
json_filter = "json:$['available']"
|
check_json_filter('json:boss.name', client, live_server)
|
||||||
|
|
||||||
|
def test_check_jq_filter(client, live_server):
|
||||||
|
if jq_support:
|
||||||
|
check_json_filter('jq:.boss.name', client, live_server)
|
||||||
|
|
||||||
|
def check_json_filter_bool_val(json_filter, client, live_server):
|
||||||
set_original_response()
|
set_original_response()
|
||||||
|
|
||||||
# Give the endpoint time to spin up
|
# Give the endpoint time to spin up
|
||||||
@@ -275,7 +293,7 @@ def test_check_json_filter_bool_val(client, live_server):
|
|||||||
# Add our URL to the import page
|
# Add our URL to the import page
|
||||||
res = client.post(
|
res = client.post(
|
||||||
url_for("edit_page", uuid="first"),
|
url_for("edit_page", uuid="first"),
|
||||||
data={"css_filter": json_filter,
|
data={"include_filters": json_filter,
|
||||||
"url": test_url,
|
"url": test_url,
|
||||||
"tag": "",
|
"tag": "",
|
||||||
"headers": "",
|
"headers": "",
|
||||||
@@ -285,18 +303,13 @@ def test_check_json_filter_bool_val(client, live_server):
|
|||||||
)
|
)
|
||||||
assert b"Updated watch." in res.data
|
assert b"Updated watch." in res.data
|
||||||
|
|
||||||
time.sleep(3)
|
|
||||||
|
|
||||||
# Trigger a check
|
|
||||||
client.get(url_for("api_watch_checknow"), follow_redirects=True)
|
|
||||||
|
|
||||||
# Give the thread time to pick it up
|
# Give the thread time to pick it up
|
||||||
time.sleep(3)
|
time.sleep(3)
|
||||||
# Make a change
|
# Make a change
|
||||||
set_modified_response()
|
set_modified_response()
|
||||||
|
|
||||||
# Trigger a check
|
# Trigger a check
|
||||||
client.get(url_for("api_watch_checknow"), follow_redirects=True)
|
client.get(url_for("form_watch_checknow"), follow_redirects=True)
|
||||||
# Give the thread time to pick it up
|
# Give the thread time to pick it up
|
||||||
time.sleep(3)
|
time.sleep(3)
|
||||||
|
|
||||||
@@ -304,14 +317,22 @@ def test_check_json_filter_bool_val(client, live_server):
|
|||||||
# But the change should be there, tho its hard to test the change was detected because it will show old and new versions
|
# But the change should be there, tho its hard to test the change was detected because it will show old and new versions
|
||||||
assert b'false' in res.data
|
assert b'false' in res.data
|
||||||
|
|
||||||
|
res = client.get(url_for("form_delete", uuid="all"), follow_redirects=True)
|
||||||
|
assert b'Deleted' in res.data
|
||||||
|
|
||||||
|
def test_check_jsonpath_filter_bool_val(client, live_server):
|
||||||
|
check_json_filter_bool_val("json:$['available']", client, live_server)
|
||||||
|
|
||||||
|
def test_check_jq_filter_bool_val(client, live_server):
|
||||||
|
if jq_support:
|
||||||
|
check_json_filter_bool_val("jq:.available", client, live_server)
|
||||||
|
|
||||||
# Re #265 - Extended JSON selector test
|
# Re #265 - Extended JSON selector test
|
||||||
# Stuff to consider here
|
# Stuff to consider here
|
||||||
# - Selector should be allowed to return empty when it doesnt match (people might wait for some condition)
|
# - Selector should be allowed to return empty when it doesnt match (people might wait for some condition)
|
||||||
# - The 'diff' tab could show the old and new content
|
# - The 'diff' tab could show the old and new content
|
||||||
# - Form should let us enter a selector that doesnt (yet) match anything
|
# - Form should let us enter a selector that doesnt (yet) match anything
|
||||||
def test_check_json_ext_filter(client, live_server):
|
def check_json_ext_filter(json_filter, client, live_server):
|
||||||
json_filter = 'json:$[?(@.status==Sold)]'
|
|
||||||
|
|
||||||
set_original_ext_response()
|
set_original_ext_response()
|
||||||
|
|
||||||
# Give the endpoint time to spin up
|
# Give the endpoint time to spin up
|
||||||
@@ -326,9 +347,6 @@ def test_check_json_ext_filter(client, live_server):
|
|||||||
)
|
)
|
||||||
assert b"1 Imported" in res.data
|
assert b"1 Imported" in res.data
|
||||||
|
|
||||||
# Trigger a check
|
|
||||||
client.get(url_for("api_watch_checknow"), follow_redirects=True)
|
|
||||||
|
|
||||||
# Give the thread time to pick it up
|
# Give the thread time to pick it up
|
||||||
time.sleep(3)
|
time.sleep(3)
|
||||||
|
|
||||||
@@ -336,7 +354,7 @@ def test_check_json_ext_filter(client, live_server):
|
|||||||
# Add our URL to the import page
|
# Add our URL to the import page
|
||||||
res = client.post(
|
res = client.post(
|
||||||
url_for("edit_page", uuid="first"),
|
url_for("edit_page", uuid="first"),
|
||||||
data={"css_filter": json_filter,
|
data={"include_filters": json_filter,
|
||||||
"url": test_url,
|
"url": test_url,
|
||||||
"tag": "",
|
"tag": "",
|
||||||
"headers": "",
|
"headers": "",
|
||||||
@@ -350,10 +368,7 @@ def test_check_json_ext_filter(client, live_server):
|
|||||||
res = client.get(
|
res = client.get(
|
||||||
url_for("edit_page", uuid="first"),
|
url_for("edit_page", uuid="first"),
|
||||||
)
|
)
|
||||||
assert bytes(json_filter.encode('utf-8')) in res.data
|
assert bytes(escape(json_filter).encode('utf-8')) in res.data
|
||||||
|
|
||||||
# Trigger a check
|
|
||||||
client.get(url_for("api_watch_checknow"), follow_redirects=True)
|
|
||||||
|
|
||||||
# Give the thread time to pick it up
|
# Give the thread time to pick it up
|
||||||
time.sleep(3)
|
time.sleep(3)
|
||||||
@@ -361,7 +376,7 @@ def test_check_json_ext_filter(client, live_server):
|
|||||||
set_modified_ext_response()
|
set_modified_ext_response()
|
||||||
|
|
||||||
# Trigger a check
|
# Trigger a check
|
||||||
client.get(url_for("api_watch_checknow"), follow_redirects=True)
|
client.get(url_for("form_watch_checknow"), follow_redirects=True)
|
||||||
# Give the thread time to pick it up
|
# Give the thread time to pick it up
|
||||||
time.sleep(4)
|
time.sleep(4)
|
||||||
|
|
||||||
@@ -376,3 +391,12 @@ def test_check_json_ext_filter(client, live_server):
|
|||||||
assert b'ForSale' not in res.data
|
assert b'ForSale' not in res.data
|
||||||
assert b'Sold' in res.data
|
assert b'Sold' in res.data
|
||||||
|
|
||||||
|
res = client.get(url_for("form_delete", uuid="all"), follow_redirects=True)
|
||||||
|
assert b'Deleted' in res.data
|
||||||
|
|
||||||
|
def test_check_jsonpath_ext_filter(client, live_server):
|
||||||
|
check_json_ext_filter('json:$[?(@.status==Sold)]', client, live_server)
|
||||||
|
|
||||||
|
def test_check_jq_ext_filter(client, live_server):
|
||||||
|
if jq_support:
|
||||||
|
check_json_ext_filter('jq:.[] | select(.status | contains("Sold"))', client, live_server)
|
||||||
102
changedetectionio/tests/test_nonrenderable_pages.py
Normal file
102
changedetectionio/tests/test_nonrenderable_pages.py
Normal file
@@ -0,0 +1,102 @@
|
|||||||
|
#!/usr/bin/python3
|
||||||
|
|
||||||
|
import time
|
||||||
|
from flask import url_for
|
||||||
|
from urllib.request import urlopen
|
||||||
|
from .util import set_original_response, set_modified_response, live_server_setup
|
||||||
|
|
||||||
|
sleep_time_for_fetch_thread = 3
|
||||||
|
|
||||||
|
|
||||||
|
def set_nonrenderable_response():
|
||||||
|
test_return_data = """<html>
|
||||||
|
<head><title>modified head title</title></head>
|
||||||
|
<!-- like when some angular app was broken and doesnt render or whatever -->
|
||||||
|
<body>
|
||||||
|
</body>
|
||||||
|
</html>
|
||||||
|
"""
|
||||||
|
|
||||||
|
with open("test-datastore/endpoint-content.txt", "w") as f:
|
||||||
|
f.write(test_return_data)
|
||||||
|
|
||||||
|
return None
|
||||||
|
|
||||||
|
def test_check_basic_change_detection_functionality(client, live_server):
|
||||||
|
set_original_response()
|
||||||
|
live_server_setup(live_server)
|
||||||
|
|
||||||
|
# Add our URL to the import page
|
||||||
|
res = client.post(
|
||||||
|
url_for("import_page"),
|
||||||
|
data={"urls": url_for('test_endpoint', _external=True)},
|
||||||
|
follow_redirects=True
|
||||||
|
)
|
||||||
|
|
||||||
|
assert b"1 Imported" in res.data
|
||||||
|
|
||||||
|
time.sleep(sleep_time_for_fetch_thread)
|
||||||
|
|
||||||
|
# Do this a few times.. ensures we dont accidently set the status
|
||||||
|
for n in range(3):
|
||||||
|
client.get(url_for("form_watch_checknow"), follow_redirects=True)
|
||||||
|
|
||||||
|
# Give the thread time to pick it up
|
||||||
|
time.sleep(sleep_time_for_fetch_thread)
|
||||||
|
|
||||||
|
# It should report nothing found (no new 'unviewed' class)
|
||||||
|
res = client.get(url_for("index"))
|
||||||
|
assert b'unviewed' not in res.data
|
||||||
|
|
||||||
|
|
||||||
|
#####################
|
||||||
|
client.post(
|
||||||
|
url_for("settings_page"),
|
||||||
|
data={"application-empty_pages_are_a_change": "",
|
||||||
|
"requests-time_between_check-minutes": 180,
|
||||||
|
'application-fetch_backend': "html_requests"},
|
||||||
|
follow_redirects=True
|
||||||
|
)
|
||||||
|
|
||||||
|
# this should not trigger a change, because no good text could be converted from the HTML
|
||||||
|
set_nonrenderable_response()
|
||||||
|
|
||||||
|
client.get(url_for("form_watch_checknow"), follow_redirects=True)
|
||||||
|
|
||||||
|
# Give the thread time to pick it up
|
||||||
|
time.sleep(sleep_time_for_fetch_thread)
|
||||||
|
|
||||||
|
# It should report nothing found (no new 'unviewed' class)
|
||||||
|
res = client.get(url_for("index"))
|
||||||
|
assert b'unviewed' not in res.data
|
||||||
|
|
||||||
|
|
||||||
|
# ok now do the opposite
|
||||||
|
|
||||||
|
client.post(
|
||||||
|
url_for("settings_page"),
|
||||||
|
data={"application-empty_pages_are_a_change": "y",
|
||||||
|
"requests-time_between_check-minutes": 180,
|
||||||
|
'application-fetch_backend': "html_requests"},
|
||||||
|
follow_redirects=True
|
||||||
|
)
|
||||||
|
set_modified_response()
|
||||||
|
|
||||||
|
|
||||||
|
client.get(url_for("form_watch_checknow"), follow_redirects=True)
|
||||||
|
|
||||||
|
# Give the thread time to pick it up
|
||||||
|
time.sleep(sleep_time_for_fetch_thread)
|
||||||
|
|
||||||
|
# It should report nothing found (no new 'unviewed' class)
|
||||||
|
res = client.get(url_for("index"))
|
||||||
|
assert b'unviewed' in res.data
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
|
#
|
||||||
|
# Cleanup everything
|
||||||
|
res = client.get(url_for("form_delete", uuid="all"), follow_redirects=True)
|
||||||
|
assert b'Deleted' in res.data
|
||||||
|
|
||||||
@@ -4,7 +4,13 @@ import re
|
|||||||
from flask import url_for
|
from flask import url_for
|
||||||
from . util import set_original_response, set_modified_response, set_more_modified_response, live_server_setup
|
from . util import set_original_response, set_modified_response, set_more_modified_response, live_server_setup
|
||||||
import logging
|
import logging
|
||||||
from changedetectionio.notification import default_notification_body, default_notification_title
|
|
||||||
|
from changedetectionio.notification import (
|
||||||
|
default_notification_body,
|
||||||
|
default_notification_format,
|
||||||
|
default_notification_title,
|
||||||
|
valid_notification_formats,
|
||||||
|
)
|
||||||
|
|
||||||
def test_setup(live_server):
|
def test_setup(live_server):
|
||||||
live_server_setup(live_server)
|
live_server_setup(live_server)
|
||||||
@@ -20,9 +26,26 @@ def test_check_notification(client, live_server):
|
|||||||
|
|
||||||
# Re 360 - new install should have defaults set
|
# Re 360 - new install should have defaults set
|
||||||
res = client.get(url_for("settings_page"))
|
res = client.get(url_for("settings_page"))
|
||||||
|
notification_url = url_for('test_notification_endpoint', _external=True).replace('http', 'json')
|
||||||
|
|
||||||
assert default_notification_body.encode() in res.data
|
assert default_notification_body.encode() in res.data
|
||||||
assert default_notification_title.encode() in res.data
|
assert default_notification_title.encode() in res.data
|
||||||
|
|
||||||
|
#####################
|
||||||
|
# Set this up for when we remove the notification from the watch, it should fallback with these details
|
||||||
|
res = client.post(
|
||||||
|
url_for("settings_page"),
|
||||||
|
data={"application-notification_urls": notification_url,
|
||||||
|
"application-notification_title": "fallback-title "+default_notification_title,
|
||||||
|
"application-notification_body": "fallback-body "+default_notification_body,
|
||||||
|
"application-notification_format": default_notification_format,
|
||||||
|
"requests-time_between_check-minutes": 180,
|
||||||
|
'application-fetch_backend': "html_requests"},
|
||||||
|
follow_redirects=True
|
||||||
|
)
|
||||||
|
|
||||||
|
assert b"Settings updated." in res.data
|
||||||
|
|
||||||
# When test mode is in BASE_URL env mode, we should see this already configured
|
# When test mode is in BASE_URL env mode, we should see this already configured
|
||||||
env_base_url = os.getenv('BASE_URL', '').strip()
|
env_base_url = os.getenv('BASE_URL', '').strip()
|
||||||
if len(env_base_url):
|
if len(env_base_url):
|
||||||
@@ -36,7 +59,7 @@ def test_check_notification(client, live_server):
|
|||||||
# Add our URL to the import page
|
# Add our URL to the import page
|
||||||
test_url = url_for('test_endpoint', _external=True)
|
test_url = url_for('test_endpoint', _external=True)
|
||||||
res = client.post(
|
res = client.post(
|
||||||
url_for("api_watch_add"),
|
url_for("form_quick_watch_add"),
|
||||||
data={"url": test_url, "tag": ''},
|
data={"url": test_url, "tag": ''},
|
||||||
follow_redirects=True
|
follow_redirects=True
|
||||||
)
|
)
|
||||||
@@ -47,8 +70,6 @@ def test_check_notification(client, live_server):
|
|||||||
|
|
||||||
# Goto the edit page, add our ignore text
|
# Goto the edit page, add our ignore text
|
||||||
# Add our URL to the import page
|
# Add our URL to the import page
|
||||||
url = url_for('test_notification_endpoint', _external=True)
|
|
||||||
notification_url = url.replace('http', 'json')
|
|
||||||
|
|
||||||
print (">>>> Notification URL: "+notification_url)
|
print (">>>> Notification URL: "+notification_url)
|
||||||
|
|
||||||
@@ -98,7 +119,7 @@ def test_check_notification(client, live_server):
|
|||||||
notification_submission = None
|
notification_submission = None
|
||||||
|
|
||||||
# Trigger a check
|
# Trigger a check
|
||||||
client.get(url_for("api_watch_checknow"), follow_redirects=True)
|
client.get(url_for("form_watch_checknow"), follow_redirects=True)
|
||||||
time.sleep(3)
|
time.sleep(3)
|
||||||
# Verify what was sent as a notification, this file should exist
|
# Verify what was sent as a notification, this file should exist
|
||||||
with open("test-datastore/notification.txt", "r") as f:
|
with open("test-datastore/notification.txt", "r") as f:
|
||||||
@@ -133,7 +154,7 @@ def test_check_notification(client, live_server):
|
|||||||
|
|
||||||
# This should insert the {current_snapshot}
|
# This should insert the {current_snapshot}
|
||||||
set_more_modified_response()
|
set_more_modified_response()
|
||||||
client.get(url_for("api_watch_checknow"), follow_redirects=True)
|
client.get(url_for("form_watch_checknow"), follow_redirects=True)
|
||||||
time.sleep(3)
|
time.sleep(3)
|
||||||
# Verify what was sent as a notification, this file should exist
|
# Verify what was sent as a notification, this file should exist
|
||||||
with open("test-datastore/notification.txt", "r") as f:
|
with open("test-datastore/notification.txt", "r") as f:
|
||||||
@@ -146,17 +167,45 @@ def test_check_notification(client, live_server):
|
|||||||
os.unlink("test-datastore/notification.txt")
|
os.unlink("test-datastore/notification.txt")
|
||||||
|
|
||||||
# Trigger a check
|
# Trigger a check
|
||||||
client.get(url_for("api_watch_checknow"), follow_redirects=True)
|
client.get(url_for("form_watch_checknow"), follow_redirects=True)
|
||||||
time.sleep(1)
|
time.sleep(1)
|
||||||
client.get(url_for("api_watch_checknow"), follow_redirects=True)
|
client.get(url_for("form_watch_checknow"), follow_redirects=True)
|
||||||
time.sleep(1)
|
time.sleep(1)
|
||||||
client.get(url_for("api_watch_checknow"), follow_redirects=True)
|
client.get(url_for("form_watch_checknow"), follow_redirects=True)
|
||||||
time.sleep(1)
|
time.sleep(1)
|
||||||
assert os.path.exists("test-datastore/notification.txt") == False
|
assert os.path.exists("test-datastore/notification.txt") == False
|
||||||
|
|
||||||
|
res = client.get(url_for("notification_logs"))
|
||||||
|
# be sure we see it in the output log
|
||||||
|
assert b'New ChangeDetection.io Notification - ' + test_url.encode('utf-8') in res.data
|
||||||
|
|
||||||
|
set_original_response()
|
||||||
|
res = client.post(
|
||||||
|
url_for("edit_page", uuid="first"),
|
||||||
|
data={
|
||||||
|
"url": test_url,
|
||||||
|
"tag": "my tag",
|
||||||
|
"title": "my title",
|
||||||
|
"notification_urls": '',
|
||||||
|
"notification_title": '',
|
||||||
|
"notification_body": '',
|
||||||
|
"notification_format": default_notification_format,
|
||||||
|
"fetch_backend": "html_requests"},
|
||||||
|
follow_redirects=True
|
||||||
|
)
|
||||||
|
assert b"Updated watch." in res.data
|
||||||
|
|
||||||
|
time.sleep(2)
|
||||||
|
|
||||||
|
# Verify what was sent as a notification, this file should exist
|
||||||
|
with open("test-datastore/notification.txt", "r") as f:
|
||||||
|
notification_submission = f.read()
|
||||||
|
assert "fallback-title" in notification_submission
|
||||||
|
assert "fallback-body" in notification_submission
|
||||||
|
|
||||||
# cleanup for the next
|
# cleanup for the next
|
||||||
client.get(
|
client.get(
|
||||||
url_for("api_delete", uuid="all"),
|
url_for("form_delete", uuid="all"),
|
||||||
follow_redirects=True
|
follow_redirects=True
|
||||||
)
|
)
|
||||||
|
|
||||||
@@ -168,7 +217,7 @@ def test_notification_validation(client, live_server):
|
|||||||
# Add our URL to the import page
|
# Add our URL to the import page
|
||||||
test_url = url_for('test_endpoint', _external=True)
|
test_url = url_for('test_endpoint', _external=True)
|
||||||
res = client.post(
|
res = client.post(
|
||||||
url_for("api_watch_add"),
|
url_for("form_quick_watch_add"),
|
||||||
data={"url": test_url, "tag": 'nice one'},
|
data={"url": test_url, "tag": 'nice one'},
|
||||||
follow_redirects=True
|
follow_redirects=True
|
||||||
)
|
)
|
||||||
@@ -176,20 +225,20 @@ def test_notification_validation(client, live_server):
|
|||||||
assert b"Watch added" in res.data
|
assert b"Watch added" in res.data
|
||||||
|
|
||||||
# Re #360 some validation
|
# Re #360 some validation
|
||||||
res = client.post(
|
# res = client.post(
|
||||||
url_for("edit_page", uuid="first"),
|
# url_for("edit_page", uuid="first"),
|
||||||
data={"notification_urls": 'json://localhost/foobar',
|
# data={"notification_urls": 'json://localhost/foobar',
|
||||||
"notification_title": "",
|
# "notification_title": "",
|
||||||
"notification_body": "",
|
# "notification_body": "",
|
||||||
"notification_format": "Text",
|
# "notification_format": "Text",
|
||||||
"url": test_url,
|
# "url": test_url,
|
||||||
"tag": "my tag",
|
# "tag": "my tag",
|
||||||
"title": "my title",
|
# "title": "my title",
|
||||||
"headers": "",
|
# "headers": "",
|
||||||
"fetch_backend": "html_requests"},
|
# "fetch_backend": "html_requests"},
|
||||||
follow_redirects=True
|
# follow_redirects=True
|
||||||
)
|
# )
|
||||||
assert b"Notification Body and Title is required when a Notification URL is used" in res.data
|
# assert b"Notification Body and Title is required when a Notification URL is used" in res.data
|
||||||
|
|
||||||
# Now adding a wrong token should give us an error
|
# Now adding a wrong token should give us an error
|
||||||
res = client.post(
|
res = client.post(
|
||||||
@@ -208,6 +257,8 @@ def test_notification_validation(client, live_server):
|
|||||||
|
|
||||||
# cleanup for the next
|
# cleanup for the next
|
||||||
client.get(
|
client.get(
|
||||||
url_for("api_delete", uuid="all"),
|
url_for("form_delete", uuid="all"),
|
||||||
follow_redirects=True
|
follow_redirects=True
|
||||||
)
|
)
|
||||||
|
|
||||||
|
|
||||||
|
|||||||
@@ -16,7 +16,7 @@ def test_check_notification_error_handling(client, live_server):
|
|||||||
# use a different URL so that it doesnt interfere with the actual check until we are ready
|
# use a different URL so that it doesnt interfere with the actual check until we are ready
|
||||||
test_url = url_for('test_endpoint', _external=True)
|
test_url = url_for('test_endpoint', _external=True)
|
||||||
res = client.post(
|
res = client.post(
|
||||||
url_for("api_watch_add"),
|
url_for("form_quick_watch_add"),
|
||||||
data={"url": "https://changedetection.io/CHANGELOG.txt", "tag": ''},
|
data={"url": "https://changedetection.io/CHANGELOG.txt", "tag": ''},
|
||||||
follow_redirects=True
|
follow_redirects=True
|
||||||
)
|
)
|
||||||
|
|||||||
43
changedetectionio/tests/test_obfuscations.py
Normal file
43
changedetectionio/tests/test_obfuscations.py
Normal file
@@ -0,0 +1,43 @@
|
|||||||
|
#!/usr/bin/python3
|
||||||
|
|
||||||
|
import time
|
||||||
|
from flask import url_for
|
||||||
|
from .util import live_server_setup
|
||||||
|
|
||||||
|
|
||||||
|
def set_original_ignore_response():
|
||||||
|
test_return_data = """<html>
|
||||||
|
<body>
|
||||||
|
<span>The price is</span><span>$<!-- -->90<!-- -->.<!-- -->74</span>
|
||||||
|
</body>
|
||||||
|
</html>
|
||||||
|
|
||||||
|
"""
|
||||||
|
|
||||||
|
with open("test-datastore/endpoint-content.txt", "w") as f:
|
||||||
|
f.write(test_return_data)
|
||||||
|
|
||||||
|
|
||||||
|
def test_obfuscations(client, live_server):
|
||||||
|
set_original_ignore_response()
|
||||||
|
live_server_setup(live_server)
|
||||||
|
time.sleep(1)
|
||||||
|
# Add our URL to the import page
|
||||||
|
test_url = url_for('test_endpoint', _external=True)
|
||||||
|
res = client.post(
|
||||||
|
url_for("import_page"),
|
||||||
|
data={"urls": test_url},
|
||||||
|
follow_redirects=True
|
||||||
|
)
|
||||||
|
assert b"1 Imported" in res.data
|
||||||
|
|
||||||
|
# Give the thread time to pick it up
|
||||||
|
time.sleep(3)
|
||||||
|
|
||||||
|
# Check HTML conversion detected and workd
|
||||||
|
res = client.get(
|
||||||
|
url_for("preview_page", uuid="first"),
|
||||||
|
follow_redirects=True
|
||||||
|
)
|
||||||
|
|
||||||
|
assert b'$90.74' in res.data
|
||||||
@@ -14,7 +14,7 @@ def test_share_watch(client, live_server):
|
|||||||
live_server_setup(live_server)
|
live_server_setup(live_server)
|
||||||
|
|
||||||
test_url = url_for('test_endpoint', _external=True)
|
test_url = url_for('test_endpoint', _external=True)
|
||||||
css_filter = ".nice-filter"
|
include_filters = ".nice-filter"
|
||||||
|
|
||||||
# Add our URL to the import page
|
# Add our URL to the import page
|
||||||
res = client.post(
|
res = client.post(
|
||||||
@@ -29,7 +29,7 @@ def test_share_watch(client, live_server):
|
|||||||
# Add our URL to the import page
|
# Add our URL to the import page
|
||||||
res = client.post(
|
res = client.post(
|
||||||
url_for("edit_page", uuid="first"),
|
url_for("edit_page", uuid="first"),
|
||||||
data={"css_filter": css_filter, "url": test_url, "tag": "", "headers": "", 'fetch_backend': "html_requests"},
|
data={"include_filters": include_filters, "url": test_url, "tag": "", "headers": "", 'fetch_backend': "html_requests"},
|
||||||
follow_redirects=True
|
follow_redirects=True
|
||||||
)
|
)
|
||||||
assert b"Updated watch." in res.data
|
assert b"Updated watch." in res.data
|
||||||
@@ -37,11 +37,11 @@ def test_share_watch(client, live_server):
|
|||||||
res = client.get(
|
res = client.get(
|
||||||
url_for("edit_page", uuid="first"),
|
url_for("edit_page", uuid="first"),
|
||||||
)
|
)
|
||||||
assert bytes(css_filter.encode('utf-8')) in res.data
|
assert bytes(include_filters.encode('utf-8')) in res.data
|
||||||
|
|
||||||
# click share the link
|
# click share the link
|
||||||
res = client.get(
|
res = client.get(
|
||||||
url_for("api_share_put_watch", uuid="first"),
|
url_for("form_share_put_watch", uuid="first"),
|
||||||
follow_redirects=True
|
follow_redirects=True
|
||||||
)
|
)
|
||||||
|
|
||||||
@@ -54,7 +54,7 @@ def test_share_watch(client, live_server):
|
|||||||
|
|
||||||
# Now delete what we have, we will try to re-import it
|
# Now delete what we have, we will try to re-import it
|
||||||
# Cleanup everything
|
# Cleanup everything
|
||||||
res = client.get(url_for("api_delete", uuid="all"), follow_redirects=True)
|
res = client.get(url_for("form_delete", uuid="all"), follow_redirects=True)
|
||||||
assert b'Deleted' in res.data
|
assert b'Deleted' in res.data
|
||||||
|
|
||||||
# Add our URL to the import page
|
# Add our URL to the import page
|
||||||
@@ -73,4 +73,8 @@ def test_share_watch(client, live_server):
|
|||||||
res = client.get(
|
res = client.get(
|
||||||
url_for("edit_page", uuid="first"),
|
url_for("edit_page", uuid="first"),
|
||||||
)
|
)
|
||||||
assert bytes(css_filter.encode('utf-8')) in res.data
|
assert bytes(include_filters.encode('utf-8')) in res.data
|
||||||
|
|
||||||
|
# Check it saved the URL
|
||||||
|
res = client.get(url_for("index"))
|
||||||
|
assert bytes(test_url.encode('utf-8')) in res.data
|
||||||
|
|||||||
@@ -39,7 +39,7 @@ def test_check_basic_change_detection_functionality_source(client, live_server):
|
|||||||
set_modified_response()
|
set_modified_response()
|
||||||
|
|
||||||
# Force recheck
|
# Force recheck
|
||||||
res = client.get(url_for("api_watch_checknow"), follow_redirects=True)
|
res = client.get(url_for("form_watch_checknow"), follow_redirects=True)
|
||||||
assert b'1 watches are queued for rechecking.' in res.data
|
assert b'1 watches are queued for rechecking.' in res.data
|
||||||
|
|
||||||
time.sleep(5)
|
time.sleep(5)
|
||||||
@@ -57,10 +57,9 @@ def test_check_basic_change_detection_functionality_source(client, live_server):
|
|||||||
|
|
||||||
|
|
||||||
|
|
||||||
|
# `subtractive_selectors` should still work in `source:` type requests
|
||||||
def test_check_ignore_elements(client, live_server):
|
def test_check_ignore_elements(client, live_server):
|
||||||
set_original_response()
|
set_original_response()
|
||||||
|
|
||||||
time.sleep(2)
|
time.sleep(2)
|
||||||
test_url = 'source:'+url_for('test_endpoint', _external=True)
|
test_url = 'source:'+url_for('test_endpoint', _external=True)
|
||||||
# Add our URL to the import page
|
# Add our URL to the import page
|
||||||
@@ -77,9 +76,9 @@ def test_check_ignore_elements(client, live_server):
|
|||||||
#####################
|
#####################
|
||||||
# We want <span> and <p> ONLY, but ignore span with .foobar-detection
|
# We want <span> and <p> ONLY, but ignore span with .foobar-detection
|
||||||
|
|
||||||
res = client.post(
|
client.post(
|
||||||
url_for("edit_page", uuid="first"),
|
url_for("edit_page", uuid="first"),
|
||||||
data={"css_filter": 'span,p', "url": test_url, "tag": "", "subtractive_selectors": ".foobar-detection", 'fetch_backend': "html_requests"},
|
data={"include_filters": 'span,p', "url": test_url, "tag": "", "subtractive_selectors": ".foobar-detection", 'fetch_backend': "html_requests"},
|
||||||
follow_redirects=True
|
follow_redirects=True
|
||||||
)
|
)
|
||||||
|
|
||||||
@@ -89,7 +88,6 @@ def test_check_ignore_elements(client, live_server):
|
|||||||
url_for("preview_page", uuid="first"),
|
url_for("preview_page", uuid="first"),
|
||||||
follow_redirects=True
|
follow_redirects=True
|
||||||
)
|
)
|
||||||
|
|
||||||
assert b'foobar-detection' not in res.data
|
assert b'foobar-detection' not in res.data
|
||||||
assert b'<br' not in res.data
|
assert b'<br' not in res.data
|
||||||
assert b'<p' in res.data
|
assert b'<p' in res.data
|
||||||
@@ -43,7 +43,7 @@ def set_modified_with_trigger_text_response():
|
|||||||
Some NEW nice initial text</br>
|
Some NEW nice initial text</br>
|
||||||
<p>Which is across multiple lines</p>
|
<p>Which is across multiple lines</p>
|
||||||
</br>
|
</br>
|
||||||
foobar123
|
Add to cart
|
||||||
<br/>
|
<br/>
|
||||||
So let's see what happens. </br>
|
So let's see what happens. </br>
|
||||||
</body>
|
</body>
|
||||||
@@ -60,7 +60,7 @@ def test_trigger_functionality(client, live_server):
|
|||||||
live_server_setup(live_server)
|
live_server_setup(live_server)
|
||||||
|
|
||||||
sleep_time_for_fetch_thread = 3
|
sleep_time_for_fetch_thread = 3
|
||||||
trigger_text = "foobar123"
|
trigger_text = "Add to cart"
|
||||||
set_original_ignore_response()
|
set_original_ignore_response()
|
||||||
|
|
||||||
# Give the endpoint time to spin up
|
# Give the endpoint time to spin up
|
||||||
@@ -76,10 +76,7 @@ def test_trigger_functionality(client, live_server):
|
|||||||
assert b"1 Imported" in res.data
|
assert b"1 Imported" in res.data
|
||||||
|
|
||||||
# Trigger a check
|
# Trigger a check
|
||||||
client.get(url_for("api_watch_checknow"), follow_redirects=True)
|
client.get(url_for("form_watch_checknow"), follow_redirects=True)
|
||||||
|
|
||||||
# Give the thread time to pick it up
|
|
||||||
time.sleep(sleep_time_for_fetch_thread)
|
|
||||||
|
|
||||||
# Goto the edit page, add our ignore text
|
# Goto the edit page, add our ignore text
|
||||||
# Add our URL to the import page
|
# Add our URL to the import page
|
||||||
@@ -98,8 +95,14 @@ def test_trigger_functionality(client, live_server):
|
|||||||
)
|
)
|
||||||
assert bytes(trigger_text.encode('utf-8')) in res.data
|
assert bytes(trigger_text.encode('utf-8')) in res.data
|
||||||
|
|
||||||
|
# Give the thread time to pick it up
|
||||||
|
time.sleep(sleep_time_for_fetch_thread)
|
||||||
|
|
||||||
|
# so that we set the state to 'unviewed' after all the edits
|
||||||
|
client.get(url_for("diff_history_page", uuid="first"))
|
||||||
|
|
||||||
# Trigger a check
|
# Trigger a check
|
||||||
client.get(url_for("api_watch_checknow"), follow_redirects=True)
|
client.get(url_for("form_watch_checknow"), follow_redirects=True)
|
||||||
|
|
||||||
# Give the thread time to pick it up
|
# Give the thread time to pick it up
|
||||||
time.sleep(sleep_time_for_fetch_thread)
|
time.sleep(sleep_time_for_fetch_thread)
|
||||||
@@ -113,7 +116,7 @@ def test_trigger_functionality(client, live_server):
|
|||||||
set_modified_original_ignore_response()
|
set_modified_original_ignore_response()
|
||||||
|
|
||||||
# Trigger a check
|
# Trigger a check
|
||||||
client.get(url_for("api_watch_checknow"), follow_redirects=True)
|
client.get(url_for("form_watch_checknow"), follow_redirects=True)
|
||||||
# Give the thread time to pick it up
|
# Give the thread time to pick it up
|
||||||
time.sleep(sleep_time_for_fetch_thread)
|
time.sleep(sleep_time_for_fetch_thread)
|
||||||
|
|
||||||
@@ -121,16 +124,22 @@ def test_trigger_functionality(client, live_server):
|
|||||||
res = client.get(url_for("index"))
|
res = client.get(url_for("index"))
|
||||||
assert b'unviewed' not in res.data
|
assert b'unviewed' not in res.data
|
||||||
|
|
||||||
# Just to be sure.. set a regular modified change..
|
# Now set the content which contains the trigger text
|
||||||
time.sleep(sleep_time_for_fetch_thread)
|
time.sleep(sleep_time_for_fetch_thread)
|
||||||
set_modified_with_trigger_text_response()
|
set_modified_with_trigger_text_response()
|
||||||
|
|
||||||
client.get(url_for("api_watch_checknow"), follow_redirects=True)
|
client.get(url_for("form_watch_checknow"), follow_redirects=True)
|
||||||
time.sleep(sleep_time_for_fetch_thread)
|
time.sleep(sleep_time_for_fetch_thread)
|
||||||
res = client.get(url_for("index"))
|
res = client.get(url_for("index"))
|
||||||
assert b'unviewed' in res.data
|
assert b'unviewed' in res.data
|
||||||
|
|
||||||
|
# https://github.com/dgtlmoon/changedetection.io/issues/616
|
||||||
|
# Apparently the actual snapshot that contains the trigger never shows
|
||||||
|
res = client.get(url_for("diff_history_page", uuid="first"))
|
||||||
|
assert b'Add to cart' in res.data
|
||||||
|
|
||||||
# Check the preview/highlighter, we should be able to see what we triggered on, but it should be highlighted
|
# Check the preview/highlighter, we should be able to see what we triggered on, but it should be highlighted
|
||||||
res = client.get(url_for("preview_page", uuid="first"))
|
res = client.get(url_for("preview_page", uuid="first"))
|
||||||
# We should be able to see what we ignored
|
|
||||||
assert b'<div class="triggered">foobar' in res.data
|
# We should be able to see what we triggered on
|
||||||
|
assert b'<div class="triggered">Add to cart' in res.data
|
||||||
@@ -42,9 +42,6 @@ def test_trigger_regex_functionality(client, live_server):
|
|||||||
)
|
)
|
||||||
assert b"1 Imported" in res.data
|
assert b"1 Imported" in res.data
|
||||||
|
|
||||||
# Trigger a check
|
|
||||||
client.get(url_for("api_watch_checknow"), follow_redirects=True)
|
|
||||||
|
|
||||||
# Give the thread time to pick it up
|
# Give the thread time to pick it up
|
||||||
time.sleep(sleep_time_for_fetch_thread)
|
time.sleep(sleep_time_for_fetch_thread)
|
||||||
|
|
||||||
@@ -60,12 +57,14 @@ def test_trigger_regex_functionality(client, live_server):
|
|||||||
"fetch_backend": "html_requests"},
|
"fetch_backend": "html_requests"},
|
||||||
follow_redirects=True
|
follow_redirects=True
|
||||||
)
|
)
|
||||||
|
time.sleep(sleep_time_for_fetch_thread)
|
||||||
|
# so that we set the state to 'unviewed' after all the edits
|
||||||
|
client.get(url_for("diff_history_page", uuid="first"))
|
||||||
|
|
||||||
with open("test-datastore/endpoint-content.txt", "w") as f:
|
with open("test-datastore/endpoint-content.txt", "w") as f:
|
||||||
f.write("some new noise")
|
f.write("some new noise")
|
||||||
|
|
||||||
client.get(url_for("api_watch_checknow"), follow_redirects=True)
|
client.get(url_for("form_watch_checknow"), follow_redirects=True)
|
||||||
time.sleep(sleep_time_for_fetch_thread)
|
time.sleep(sleep_time_for_fetch_thread)
|
||||||
|
|
||||||
# It should report nothing found (nothing should match the regex)
|
# It should report nothing found (nothing should match the regex)
|
||||||
@@ -75,7 +74,11 @@ def test_trigger_regex_functionality(client, live_server):
|
|||||||
with open("test-datastore/endpoint-content.txt", "w") as f:
|
with open("test-datastore/endpoint-content.txt", "w") as f:
|
||||||
f.write("regex test123<br/>\nsomething 123")
|
f.write("regex test123<br/>\nsomething 123")
|
||||||
|
|
||||||
client.get(url_for("api_watch_checknow"), follow_redirects=True)
|
client.get(url_for("form_watch_checknow"), follow_redirects=True)
|
||||||
time.sleep(sleep_time_for_fetch_thread)
|
time.sleep(sleep_time_for_fetch_thread)
|
||||||
res = client.get(url_for("index"))
|
res = client.get(url_for("index"))
|
||||||
assert b'unviewed' in res.data
|
assert b'unviewed' in res.data
|
||||||
|
|
||||||
|
# Cleanup everything
|
||||||
|
res = client.get(url_for("form_delete", uuid="all"), follow_redirects=True)
|
||||||
|
assert b'Deleted' in res.data
|
||||||
@@ -22,10 +22,9 @@ def set_original_ignore_response():
|
|||||||
|
|
||||||
|
|
||||||
|
|
||||||
def test_trigger_regex_functionality(client, live_server):
|
def test_trigger_regex_functionality_with_filter(client, live_server):
|
||||||
|
|
||||||
live_server_setup(live_server)
|
live_server_setup(live_server)
|
||||||
|
|
||||||
sleep_time_for_fetch_thread = 3
|
sleep_time_for_fetch_thread = 3
|
||||||
|
|
||||||
set_original_ignore_response()
|
set_original_ignore_response()
|
||||||
@@ -42,43 +41,44 @@ def test_trigger_regex_functionality(client, live_server):
|
|||||||
)
|
)
|
||||||
assert b"1 Imported" in res.data
|
assert b"1 Imported" in res.data
|
||||||
|
|
||||||
# Trigger a check
|
# it needs time to save the original version
|
||||||
client.get(url_for("api_watch_checknow"), follow_redirects=True)
|
|
||||||
|
|
||||||
# Give the thread time to pick it up
|
|
||||||
time.sleep(sleep_time_for_fetch_thread)
|
time.sleep(sleep_time_for_fetch_thread)
|
||||||
|
|
||||||
# It should report nothing found (just a new one shouldnt have anything)
|
|
||||||
res = client.get(url_for("index"))
|
|
||||||
assert b'unviewed' not in res.data
|
|
||||||
|
|
||||||
### test regex with filter
|
### test regex with filter
|
||||||
res = client.post(
|
res = client.post(
|
||||||
url_for("edit_page", uuid="first"),
|
url_for("edit_page", uuid="first"),
|
||||||
data={"trigger_text": "/cool.stuff\d/",
|
data={"trigger_text": "/cool.stuff/",
|
||||||
"url": test_url,
|
"url": test_url,
|
||||||
"css_filter": '#in-here',
|
"include_filters": '#in-here',
|
||||||
"fetch_backend": "html_requests"},
|
"fetch_backend": "html_requests"},
|
||||||
follow_redirects=True
|
follow_redirects=True
|
||||||
)
|
)
|
||||||
|
|
||||||
|
# Give the thread time to pick it up
|
||||||
|
time.sleep(sleep_time_for_fetch_thread)
|
||||||
|
|
||||||
|
client.get(url_for("diff_history_page", uuid="first"))
|
||||||
|
|
||||||
# Check that we have the expected text.. but it's not in the css filter we want
|
# Check that we have the expected text.. but it's not in the css filter we want
|
||||||
with open("test-datastore/endpoint-content.txt", "w") as f:
|
with open("test-datastore/endpoint-content.txt", "w") as f:
|
||||||
f.write("<html>some new noise with cool stuff2 ok</html>")
|
f.write("<html>some new noise with cool stuff2 ok</html>")
|
||||||
|
|
||||||
client.get(url_for("api_watch_checknow"), follow_redirects=True)
|
client.get(url_for("form_watch_checknow"), follow_redirects=True)
|
||||||
time.sleep(sleep_time_for_fetch_thread)
|
time.sleep(sleep_time_for_fetch_thread)
|
||||||
|
|
||||||
# It should report nothing found (nothing should match the regex and filter)
|
# It should report nothing found (nothing should match the regex and filter)
|
||||||
res = client.get(url_for("index"))
|
res = client.get(url_for("index"))
|
||||||
assert b'unviewed' not in res.data
|
assert b'unviewed' not in res.data
|
||||||
|
|
||||||
|
# now this should trigger something
|
||||||
with open("test-datastore/endpoint-content.txt", "w") as f:
|
with open("test-datastore/endpoint-content.txt", "w") as f:
|
||||||
f.write("<html>some new noise with <span id=in-here>cool stuff6</span> ok</html>")
|
f.write("<html>some new noise with <span id=in-here>cool stuff6</span> ok</html>")
|
||||||
|
|
||||||
client.get(url_for("api_watch_checknow"), follow_redirects=True)
|
client.get(url_for("form_watch_checknow"), follow_redirects=True)
|
||||||
time.sleep(sleep_time_for_fetch_thread)
|
time.sleep(sleep_time_for_fetch_thread)
|
||||||
res = client.get(url_for("index"))
|
res = client.get(url_for("index"))
|
||||||
assert b'unviewed' in res.data
|
assert b'unviewed' in res.data
|
||||||
|
|
||||||
|
# Cleanup everything
|
||||||
|
res = client.get(url_for("form_delete", uuid="all"), follow_redirects=True)
|
||||||
|
assert b'Deleted' in res.data
|
||||||
|
|||||||
Some files were not shown because too many files have changed in this diff Show More
Reference in New Issue
Block a user