mirror of
https://github.com/dgtlmoon/changedetection.io.git
synced 2025-11-26 11:23:23 +00:00
Compare commits
2 Commits
1899-extra
...
xpath-extr
| Author | SHA1 | Date | |
|---|---|---|---|
|
|
67396b2b75 | ||
|
|
67c77922fc |
4
.github/workflows/test-only.yml
vendored
4
.github/workflows/test-only.yml
vendored
@@ -29,8 +29,8 @@ jobs:
|
|||||||
docker network create changedet-network
|
docker network create changedet-network
|
||||||
|
|
||||||
# Selenium+browserless
|
# Selenium+browserless
|
||||||
docker run --network changedet-network -d --hostname selenium -p 4444:4444 --rm --shm-size="2g" selenium/standalone-chrome:4.14.1
|
docker run --network changedet-network -d --hostname selenium -p 4444:4444 --rm --shm-size="2g" selenium/standalone-chrome-debug:3.141.59
|
||||||
docker run --network changedet-network -d --hostname browserless -e "FUNCTION_BUILT_INS=[\"fs\",\"crypto\"]" -e "DEFAULT_LAUNCH_ARGS=[\"--window-size=1920,1080\"]" --rm -p 3000:3000 --shm-size="2g" browserless/chrome:1.60-chrome-stable
|
docker run --network changedet-network -d --hostname browserless -e "FUNCTION_BUILT_INS=[\"fs\",\"crypto\"]" -e "DEFAULT_LAUNCH_ARGS=[\"--window-size=1920,1080\"]" --rm -p 3000:3000 --shm-size="2g" browserless/chrome:1.53-chrome-stable
|
||||||
|
|
||||||
- name: Build changedetection.io container for testing
|
- name: Build changedetection.io container for testing
|
||||||
run: |
|
run: |
|
||||||
|
|||||||
@@ -20,11 +20,6 @@ WORKDIR /install
|
|||||||
|
|
||||||
COPY requirements.txt /requirements.txt
|
COPY requirements.txt /requirements.txt
|
||||||
|
|
||||||
# Instructing pip to fetch wheels from piwheels.org" on ARMv6 and ARMv7 machines
|
|
||||||
RUN if [ "$(dpkg --print-architecture)" = "armhf" ] || [ "$(dpkg --print-architecture)" = "armel" ]; then \
|
|
||||||
printf "[global]\nextra-index-url=https://www.piwheels.org/simple\n" > /etc/pip.conf; \
|
|
||||||
fi;
|
|
||||||
|
|
||||||
RUN pip install --target=/dependencies -r /requirements.txt
|
RUN pip install --target=/dependencies -r /requirements.txt
|
||||||
|
|
||||||
# Playwright is an alternative to Selenium
|
# Playwright is an alternative to Selenium
|
||||||
|
|||||||
@@ -416,18 +416,11 @@ def changedetection_app(config=None, datastore_o=None):
|
|||||||
|
|
||||||
# Sort by last_changed and add the uuid which is usually the key..
|
# Sort by last_changed and add the uuid which is usually the key..
|
||||||
sorted_watches = []
|
sorted_watches = []
|
||||||
with_errors = request.args.get('with_errors') == "1"
|
|
||||||
errored_count = 0
|
|
||||||
search_q = request.args.get('q').strip().lower() if request.args.get('q') else False
|
search_q = request.args.get('q').strip().lower() if request.args.get('q') else False
|
||||||
for uuid, watch in datastore.data['watching'].items():
|
for uuid, watch in datastore.data['watching'].items():
|
||||||
if with_errors and not watch.get('last_error'):
|
|
||||||
continue
|
|
||||||
|
|
||||||
if limit_tag and not limit_tag in watch['tags']:
|
if limit_tag and not limit_tag in watch['tags']:
|
||||||
continue
|
continue
|
||||||
if watch.get('last_error'):
|
|
||||||
errored_count += 1
|
|
||||||
|
|
||||||
if search_q:
|
if search_q:
|
||||||
if (watch.get('title') and search_q in watch.get('title').lower()) or search_q in watch.get('url', '').lower():
|
if (watch.get('title') and search_q in watch.get('title').lower()) or search_q in watch.get('url', '').lower():
|
||||||
sorted_watches.append(watch)
|
sorted_watches.append(watch)
|
||||||
@@ -449,7 +442,6 @@ def changedetection_app(config=None, datastore_o=None):
|
|||||||
active_tag=limit_tag,
|
active_tag=limit_tag,
|
||||||
app_rss_token=datastore.data['settings']['application']['rss_access_token'],
|
app_rss_token=datastore.data['settings']['application']['rss_access_token'],
|
||||||
datastore=datastore,
|
datastore=datastore,
|
||||||
errored_count=errored_count,
|
|
||||||
form=form,
|
form=form,
|
||||||
guid=datastore.data['app_guid'],
|
guid=datastore.data['app_guid'],
|
||||||
has_proxies=datastore.proxy_list,
|
has_proxies=datastore.proxy_list,
|
||||||
@@ -630,6 +622,7 @@ def changedetection_app(config=None, datastore_o=None):
|
|||||||
|
|
||||||
if request.args.get('unpause_on_save'):
|
if request.args.get('unpause_on_save'):
|
||||||
extra_update_obj['paused'] = False
|
extra_update_obj['paused'] = False
|
||||||
|
|
||||||
# Re #110, if they submit the same as the default value, set it to None, so we continue to follow the default
|
# Re #110, if they submit the same as the default value, set it to None, so we continue to follow the default
|
||||||
# Assume we use the default value, unless something relevant is different, then use the form value
|
# Assume we use the default value, unless something relevant is different, then use the form value
|
||||||
# values could be None, 0 etc.
|
# values could be None, 0 etc.
|
||||||
@@ -715,6 +708,7 @@ def changedetection_app(config=None, datastore_o=None):
|
|||||||
|
|
||||||
# Only works reliably with Playwright
|
# Only works reliably with Playwright
|
||||||
visualselector_enabled = os.getenv('PLAYWRIGHT_DRIVER_URL', False) and is_html_webdriver
|
visualselector_enabled = os.getenv('PLAYWRIGHT_DRIVER_URL', False) and is_html_webdriver
|
||||||
|
|
||||||
output = render_template("edit.html",
|
output = render_template("edit.html",
|
||||||
available_processors=processors.available_processors(),
|
available_processors=processors.available_processors(),
|
||||||
browser_steps_config=browser_step_ui_config,
|
browser_steps_config=browser_step_ui_config,
|
||||||
@@ -863,10 +857,7 @@ def changedetection_app(config=None, datastore_o=None):
|
|||||||
def mark_all_viewed():
|
def mark_all_viewed():
|
||||||
|
|
||||||
# Save the current newest history as the most recently viewed
|
# Save the current newest history as the most recently viewed
|
||||||
with_errors = request.args.get('with_errors') == "1"
|
|
||||||
for watch_uuid, watch in datastore.data['watching'].items():
|
for watch_uuid, watch in datastore.data['watching'].items():
|
||||||
if with_errors and not watch.get('last_error'):
|
|
||||||
continue
|
|
||||||
datastore.set_last_viewed(watch_uuid, int(time.time()))
|
datastore.set_last_viewed(watch_uuid, int(time.time()))
|
||||||
|
|
||||||
return redirect(url_for('index'))
|
return redirect(url_for('index'))
|
||||||
@@ -1275,8 +1266,6 @@ def changedetection_app(config=None, datastore_o=None):
|
|||||||
# Forced recheck will skip the 'skip if content is the same' rule (, 'reprocess_existing_data': True})))
|
# Forced recheck will skip the 'skip if content is the same' rule (, 'reprocess_existing_data': True})))
|
||||||
tag = request.args.get('tag')
|
tag = request.args.get('tag')
|
||||||
uuid = request.args.get('uuid')
|
uuid = request.args.get('uuid')
|
||||||
with_errors = request.args.get('with_errors') == "1"
|
|
||||||
|
|
||||||
i = 0
|
i = 0
|
||||||
|
|
||||||
running_uuids = []
|
running_uuids = []
|
||||||
@@ -1292,8 +1281,6 @@ def changedetection_app(config=None, datastore_o=None):
|
|||||||
# Items that have this current tag
|
# Items that have this current tag
|
||||||
for watch_uuid, watch in datastore.data['watching'].items():
|
for watch_uuid, watch in datastore.data['watching'].items():
|
||||||
if tag in watch.get('tags', {}):
|
if tag in watch.get('tags', {}):
|
||||||
if with_errors and not watch.get('last_error'):
|
|
||||||
continue
|
|
||||||
if watch_uuid not in running_uuids and not datastore.data['watching'][watch_uuid]['paused']:
|
if watch_uuid not in running_uuids and not datastore.data['watching'][watch_uuid]['paused']:
|
||||||
update_q.put(
|
update_q.put(
|
||||||
queuedWatchMetaData.PrioritizedItem(priority=1, item={'uuid': watch_uuid, 'skip_when_checksum_same': False})
|
queuedWatchMetaData.PrioritizedItem(priority=1, item={'uuid': watch_uuid, 'skip_when_checksum_same': False})
|
||||||
@@ -1304,11 +1291,8 @@ def changedetection_app(config=None, datastore_o=None):
|
|||||||
# No tag, no uuid, add everything.
|
# No tag, no uuid, add everything.
|
||||||
for watch_uuid, watch in datastore.data['watching'].items():
|
for watch_uuid, watch in datastore.data['watching'].items():
|
||||||
if watch_uuid not in running_uuids and not datastore.data['watching'][watch_uuid]['paused']:
|
if watch_uuid not in running_uuids and not datastore.data['watching'][watch_uuid]['paused']:
|
||||||
if with_errors and not watch.get('last_error'):
|
|
||||||
continue
|
|
||||||
update_q.put(queuedWatchMetaData.PrioritizedItem(priority=1, item={'uuid': watch_uuid, 'skip_when_checksum_same': False}))
|
update_q.put(queuedWatchMetaData.PrioritizedItem(priority=1, item={'uuid': watch_uuid, 'skip_when_checksum_same': False}))
|
||||||
i += 1
|
i += 1
|
||||||
|
|
||||||
flash("{} watches queued for rechecking.".format(i))
|
flash("{} watches queued for rechecking.".format(i))
|
||||||
return redirect(url_for('index', tag=tag))
|
return redirect(url_for('index', tag=tag))
|
||||||
|
|
||||||
|
|||||||
@@ -23,10 +23,8 @@
|
|||||||
|
|
||||||
from distutils.util import strtobool
|
from distutils.util import strtobool
|
||||||
from flask import Blueprint, request, make_response
|
from flask import Blueprint, request, make_response
|
||||||
import logging
|
|
||||||
import os
|
import os
|
||||||
import re
|
import logging
|
||||||
|
|
||||||
from changedetectionio.store import ChangeDetectionStore
|
from changedetectionio.store import ChangeDetectionStore
|
||||||
from changedetectionio import login_optionally_required
|
from changedetectionio import login_optionally_required
|
||||||
|
|
||||||
@@ -46,7 +44,7 @@ def construct_blueprint(datastore: ChangeDetectionStore):
|
|||||||
|
|
||||||
|
|
||||||
# We keep the playwright session open for many minutes
|
# We keep the playwright session open for many minutes
|
||||||
keepalive_seconds = int(os.getenv('BROWSERSTEPS_MINUTES_KEEPALIVE', 10)) * 60
|
seconds_keepalive = int(os.getenv('BROWSERSTEPS_MINUTES_KEEPALIVE', 10)) * 60
|
||||||
|
|
||||||
browsersteps_start_session = {'start_time': time.time()}
|
browsersteps_start_session = {'start_time': time.time()}
|
||||||
|
|
||||||
@@ -58,18 +56,16 @@ def construct_blueprint(datastore: ChangeDetectionStore):
|
|||||||
# Start the Playwright context, which is actually a nodejs sub-process and communicates over STDIN/STDOUT pipes
|
# Start the Playwright context, which is actually a nodejs sub-process and communicates over STDIN/STDOUT pipes
|
||||||
io_interface_context = io_interface_context.start()
|
io_interface_context = io_interface_context.start()
|
||||||
|
|
||||||
keepalive_ms = ((keepalive_seconds + 3) * 1000)
|
|
||||||
base_url = os.getenv('PLAYWRIGHT_DRIVER_URL', '')
|
|
||||||
a = "?" if not '?' in base_url else '&'
|
|
||||||
base_url += a + f"timeout={keepalive_ms}"
|
|
||||||
|
|
||||||
|
# keep it alive for 10 seconds more than we advertise, sometimes it helps to keep it shutting down cleanly
|
||||||
|
keepalive = "&timeout={}".format(((seconds_keepalive + 3) * 1000))
|
||||||
try:
|
try:
|
||||||
browsersteps_start_session['browser'] = io_interface_context.chromium.connect_over_cdp(base_url)
|
browsersteps_start_session['browser'] = io_interface_context.chromium.connect_over_cdp(
|
||||||
|
os.getenv('PLAYWRIGHT_DRIVER_URL', '') + keepalive)
|
||||||
except Exception as e:
|
except Exception as e:
|
||||||
if 'ECONNREFUSED' in str(e):
|
if 'ECONNREFUSED' in str(e):
|
||||||
return make_response('Unable to start the Playwright Browser session, is it running?', 401)
|
return make_response('Unable to start the Playwright Browser session, is it running?', 401)
|
||||||
else:
|
else:
|
||||||
# Other errors, bad URL syntax, bad reply etc
|
|
||||||
return make_response(str(e), 401)
|
return make_response(str(e), 401)
|
||||||
|
|
||||||
proxy_id = datastore.get_preferred_proxy_for_watch(uuid=watch_uuid)
|
proxy_id = datastore.get_preferred_proxy_for_watch(uuid=watch_uuid)
|
||||||
@@ -122,31 +118,6 @@ def construct_blueprint(datastore: ChangeDetectionStore):
|
|||||||
print("Starting connection with playwright - done")
|
print("Starting connection with playwright - done")
|
||||||
return {'browsersteps_session_id': browsersteps_session_id}
|
return {'browsersteps_session_id': browsersteps_session_id}
|
||||||
|
|
||||||
@login_optionally_required
|
|
||||||
@browser_steps_blueprint.route("/browsersteps_image", methods=['GET'])
|
|
||||||
def browser_steps_fetch_screenshot_image():
|
|
||||||
from flask import (
|
|
||||||
make_response,
|
|
||||||
request,
|
|
||||||
send_from_directory,
|
|
||||||
)
|
|
||||||
uuid = request.args.get('uuid')
|
|
||||||
step_n = int(request.args.get('step_n'))
|
|
||||||
|
|
||||||
watch = datastore.data['watching'].get(uuid)
|
|
||||||
filename = f"step_before-{step_n}.jpeg" if request.args.get('type', '') == 'before' else f"step_{step_n}.jpeg"
|
|
||||||
|
|
||||||
if step_n and watch and os.path.isfile(os.path.join(watch.watch_data_dir, filename)):
|
|
||||||
response = make_response(send_from_directory(directory=watch.watch_data_dir, path=filename))
|
|
||||||
response.headers['Content-type'] = 'image/jpeg'
|
|
||||||
response.headers['Cache-Control'] = 'no-cache, no-store, must-revalidate'
|
|
||||||
response.headers['Pragma'] = 'no-cache'
|
|
||||||
response.headers['Expires'] = 0
|
|
||||||
return response
|
|
||||||
|
|
||||||
else:
|
|
||||||
return make_response('Unable to fetch image, is the URL correct? does the watch exist? does the step_type-n.jpeg exist?', 401)
|
|
||||||
|
|
||||||
# A request for an action was received
|
# A request for an action was received
|
||||||
@login_optionally_required
|
@login_optionally_required
|
||||||
@browser_steps_blueprint.route("/browsersteps_update", methods=['POST'])
|
@browser_steps_blueprint.route("/browsersteps_update", methods=['POST'])
|
||||||
|
|||||||
@@ -77,13 +77,13 @@ class steppable_browser_interface():
|
|||||||
def action_goto_url(self, selector=None, value=None):
|
def action_goto_url(self, selector=None, value=None):
|
||||||
# self.page.set_viewport_size({"width": 1280, "height": 5000})
|
# self.page.set_viewport_size({"width": 1280, "height": 5000})
|
||||||
now = time.time()
|
now = time.time()
|
||||||
response = self.page.goto(value, timeout=0, wait_until='load')
|
response = self.page.goto(value, timeout=0, wait_until='commit')
|
||||||
# Should be the same as the puppeteer_fetch.js methods, means, load with no timeout set (skip timeout)
|
|
||||||
#and also wait for seconds ?
|
# Wait_until = commit
|
||||||
#await page.waitForTimeout(1000);
|
# - `'commit'` - consider operation to be finished when network response is received and the document started loading.
|
||||||
#await page.waitForTimeout(extra_wait_ms);
|
# Better to not use any smarts from Playwright and just wait an arbitrary number of seconds
|
||||||
|
# This seemed to solve nearly all 'TimeoutErrors'
|
||||||
print("Time to goto URL ", time.time() - now)
|
print("Time to goto URL ", time.time() - now)
|
||||||
return response
|
|
||||||
|
|
||||||
def action_click_element_containing_text(self, selector=None, value=''):
|
def action_click_element_containing_text(self, selector=None, value=''):
|
||||||
if not len(value.strip()):
|
if not len(value.strip()):
|
||||||
@@ -99,8 +99,7 @@ class steppable_browser_interface():
|
|||||||
self.page.fill(selector, value, timeout=10 * 1000)
|
self.page.fill(selector, value, timeout=10 * 1000)
|
||||||
|
|
||||||
def action_execute_js(self, selector, value):
|
def action_execute_js(self, selector, value):
|
||||||
response = self.page.evaluate(value)
|
self.page.evaluate(value)
|
||||||
return response
|
|
||||||
|
|
||||||
def action_click_element(self, selector, value):
|
def action_click_element(self, selector, value):
|
||||||
print("Clicking element")
|
print("Clicking element")
|
||||||
@@ -139,13 +138,13 @@ class steppable_browser_interface():
|
|||||||
def action_wait_for_text(self, selector, value):
|
def action_wait_for_text(self, selector, value):
|
||||||
import json
|
import json
|
||||||
v = json.dumps(value)
|
v = json.dumps(value)
|
||||||
self.page.wait_for_function(f'document.querySelector("body").innerText.includes({v});', timeout=30000)
|
self.page.wait_for_function(f'document.querySelector("body").innerText.includes({v});', timeout=90000)
|
||||||
|
|
||||||
def action_wait_for_text_in_element(self, selector, value):
|
def action_wait_for_text_in_element(self, selector, value):
|
||||||
import json
|
import json
|
||||||
s = json.dumps(selector)
|
s = json.dumps(selector)
|
||||||
v = json.dumps(value)
|
v = json.dumps(value)
|
||||||
self.page.wait_for_function(f'document.querySelector({s}).innerText.includes({v});', timeout=30000)
|
self.page.wait_for_function(f'document.querySelector({s}).innerText.includes({v});', timeout=90000)
|
||||||
|
|
||||||
# @todo - in the future make some popout interface to capture what needs to be set
|
# @todo - in the future make some popout interface to capture what needs to be set
|
||||||
# https://playwright.dev/python/docs/api/class-keyboard
|
# https://playwright.dev/python/docs/api/class-keyboard
|
||||||
|
|||||||
@@ -464,19 +464,38 @@ class base_html_playwright(Fetcher):
|
|||||||
if len(request_headers):
|
if len(request_headers):
|
||||||
context.set_extra_http_headers(request_headers)
|
context.set_extra_http_headers(request_headers)
|
||||||
|
|
||||||
# Listen for all console events and handle errors
|
self.page.set_default_navigation_timeout(90000)
|
||||||
self.page.on("console", lambda msg: print(f"Playwright console: Watch URL: {url} {msg.type}: {msg.text} {msg.args}"))
|
self.page.set_default_timeout(90000)
|
||||||
|
|
||||||
# Re-use as much code from browser steps as possible so its the same
|
# Listen for all console events and handle errors
|
||||||
from changedetectionio.blueprint.browser_steps.browser_steps import steppable_browser_interface
|
self.page.on("console", lambda msg: print(f"Playwright console: Watch URL: {url} {msg.type}: {msg.text} {msg.args}"))
|
||||||
browsersteps_interface = steppable_browser_interface()
|
|
||||||
browsersteps_interface.page = self.page
|
|
||||||
|
|
||||||
|
# Goto page
|
||||||
try:
|
try:
|
||||||
response = browsersteps_interface.action_goto_url(value=url)
|
# Wait_until = commit
|
||||||
|
# - `'commit'` - consider operation to be finished when network response is received and the document started loading.
|
||||||
|
# Better to not use any smarts from Playwright and just wait an arbitrary number of seconds
|
||||||
|
# This seemed to solve nearly all 'TimeoutErrors'
|
||||||
|
response = self.page.goto(url, wait_until='commit')
|
||||||
|
except playwright._impl._api_types.Error as e:
|
||||||
|
# Retry once - https://github.com/browserless/chrome/issues/2485
|
||||||
|
# Sometimes errors related to invalid cert's and other can be random
|
||||||
|
print("Content Fetcher > retrying request got error - ", str(e))
|
||||||
|
time.sleep(1)
|
||||||
|
response = self.page.goto(url, wait_until='commit')
|
||||||
|
except Exception as e:
|
||||||
|
print("Content Fetcher > Other exception when page.goto", str(e))
|
||||||
|
context.close()
|
||||||
|
browser.close()
|
||||||
|
raise PageUnloadable(url=url, status_code=None, message=str(e))
|
||||||
|
|
||||||
|
# Execute any browser steps
|
||||||
|
try:
|
||||||
|
extra_wait = int(os.getenv("WEBDRIVER_DELAY_BEFORE_CONTENT_READY", 5)) + self.render_extract_delay
|
||||||
|
self.page.wait_for_timeout(extra_wait * 1000)
|
||||||
|
|
||||||
if self.webdriver_js_execute_code is not None and len(self.webdriver_js_execute_code):
|
if self.webdriver_js_execute_code is not None and len(self.webdriver_js_execute_code):
|
||||||
browsersteps_interface.action_execute_js(value=self.webdriver_js_execute_code, selector=None)
|
self.page.evaluate(self.webdriver_js_execute_code)
|
||||||
|
|
||||||
except playwright._impl._api_types.TimeoutError as e:
|
except playwright._impl._api_types.TimeoutError as e:
|
||||||
context.close()
|
context.close()
|
||||||
@@ -495,14 +514,11 @@ class base_html_playwright(Fetcher):
|
|||||||
print("Content Fetcher > Response object was none")
|
print("Content Fetcher > Response object was none")
|
||||||
raise EmptyReply(url=url, status_code=None)
|
raise EmptyReply(url=url, status_code=None)
|
||||||
|
|
||||||
extra_wait = int(os.getenv("WEBDRIVER_DELAY_BEFORE_CONTENT_READY", 5)) + self.render_extract_delay
|
|
||||||
self.page.wait_for_timeout(extra_wait * 1000)
|
|
||||||
|
|
||||||
# Run Browser Steps here
|
# Run Browser Steps here
|
||||||
self.iterate_browser_steps()
|
self.iterate_browser_steps()
|
||||||
|
|
||||||
extra_wait = int(os.getenv("WEBDRIVER_DELAY_BEFORE_CONTENT_READY", 5)) + self.render_extract_delay
|
extra_wait = int(os.getenv("WEBDRIVER_DELAY_BEFORE_CONTENT_READY", 5)) + self.render_extract_delay
|
||||||
self.page.wait_for_timeout(extra_wait * 1000)
|
time.sleep(extra_wait)
|
||||||
|
|
||||||
self.content = self.page.content()
|
self.content = self.page.content()
|
||||||
self.status_code = response.status
|
self.status_code = response.status
|
||||||
@@ -598,17 +614,14 @@ class base_html_webdriver(Fetcher):
|
|||||||
is_binary=False):
|
is_binary=False):
|
||||||
|
|
||||||
from selenium import webdriver
|
from selenium import webdriver
|
||||||
from selenium.webdriver.chrome.options import Options as ChromeOptions
|
from selenium.webdriver.common.desired_capabilities import DesiredCapabilities
|
||||||
from selenium.common.exceptions import WebDriverException
|
from selenium.common.exceptions import WebDriverException
|
||||||
# request_body, request_method unused for now, until some magic in the future happens.
|
# request_body, request_method unused for now, until some magic in the future happens.
|
||||||
|
|
||||||
options = ChromeOptions()
|
|
||||||
if self.proxy:
|
|
||||||
options.proxy = self.proxy
|
|
||||||
|
|
||||||
self.driver = webdriver.Remote(
|
self.driver = webdriver.Remote(
|
||||||
command_executor=self.command_executor,
|
command_executor=self.command_executor,
|
||||||
options=options)
|
desired_capabilities=DesiredCapabilities.CHROME,
|
||||||
|
proxy=self.proxy)
|
||||||
|
|
||||||
try:
|
try:
|
||||||
self.driver.get(url)
|
self.driver.get(url)
|
||||||
@@ -640,11 +653,11 @@ class base_html_webdriver(Fetcher):
|
|||||||
# Does the connection to the webdriver work? run a test connection.
|
# Does the connection to the webdriver work? run a test connection.
|
||||||
def is_ready(self):
|
def is_ready(self):
|
||||||
from selenium import webdriver
|
from selenium import webdriver
|
||||||
from selenium.webdriver.chrome.options import Options as ChromeOptions
|
from selenium.webdriver.common.desired_capabilities import DesiredCapabilities
|
||||||
|
|
||||||
self.driver = webdriver.Remote(
|
self.driver = webdriver.Remote(
|
||||||
command_executor=self.command_executor,
|
command_executor=self.command_executor,
|
||||||
options=ChromeOptions())
|
desired_capabilities=DesiredCapabilities.CHROME)
|
||||||
|
|
||||||
# driver.quit() seems to cause better exceptions
|
# driver.quit() seems to cause better exceptions
|
||||||
self.quit()
|
self.quit()
|
||||||
|
|||||||
@@ -22,8 +22,7 @@ from wtforms.validators import ValidationError
|
|||||||
# each select <option data-enabled="enabled-0-0"
|
# each select <option data-enabled="enabled-0-0"
|
||||||
from changedetectionio.blueprint.browser_steps.browser_steps import browser_step_ui_config
|
from changedetectionio.blueprint.browser_steps.browser_steps import browser_step_ui_config
|
||||||
|
|
||||||
from changedetectionio import content_fetcher, html_tools
|
from changedetectionio import content_fetcher
|
||||||
|
|
||||||
from changedetectionio.notification import (
|
from changedetectionio.notification import (
|
||||||
valid_notification_formats,
|
valid_notification_formats,
|
||||||
)
|
)
|
||||||
@@ -285,10 +284,11 @@ class ValidateListRegex(object):
|
|||||||
def __call__(self, form, field):
|
def __call__(self, form, field):
|
||||||
|
|
||||||
for line in field.data:
|
for line in field.data:
|
||||||
if re.search(html_tools.PERL_STYLE_REGEX, line, re.IGNORECASE):
|
if line[0] == '/' and line[-1] == '/':
|
||||||
|
# Because internally we dont wrap in /
|
||||||
|
line = line.strip('/')
|
||||||
try:
|
try:
|
||||||
regex = html_tools.perl_style_slash_enclosed_regex_to_options(line)
|
re.compile(line)
|
||||||
re.compile(regex)
|
|
||||||
except re.error:
|
except re.error:
|
||||||
message = field.gettext('RegEx \'%s\' is not a valid regular expression.')
|
message = field.gettext('RegEx \'%s\' is not a valid regular expression.')
|
||||||
raise ValidationError(message % (line))
|
raise ValidationError(message % (line))
|
||||||
|
|||||||
@@ -1,12 +1,9 @@
|
|||||||
|
|
||||||
from bs4 import BeautifulSoup
|
from bs4 import BeautifulSoup
|
||||||
from inscriptis import get_text
|
from inscriptis import get_text
|
||||||
|
from inscriptis.model.config import ParserConfig
|
||||||
from jsonpath_ng.ext import parse
|
from jsonpath_ng.ext import parse
|
||||||
from typing import List
|
from typing import List
|
||||||
from inscriptis.css_profiles import CSS_PROFILES, HtmlElement
|
|
||||||
from inscriptis.html_properties import Display
|
|
||||||
from inscriptis.model.config import ParserConfig
|
|
||||||
from xml.sax.saxutils import escape as xml_escape
|
|
||||||
import json
|
import json
|
||||||
import re
|
import re
|
||||||
|
|
||||||
@@ -71,15 +68,10 @@ def element_removal(selectors: List[str], html_content):
|
|||||||
|
|
||||||
|
|
||||||
# Return str Utf-8 of matched rules
|
# Return str Utf-8 of matched rules
|
||||||
def xpath_filter(xpath_filter, html_content, append_pretty_line_formatting=False, is_rss=False):
|
def xpath_filter(xpath_filter, html_content, append_pretty_line_formatting=False):
|
||||||
from lxml import etree, html
|
from lxml import etree, html
|
||||||
|
|
||||||
parser = None
|
tree = html.fromstring(bytes(html_content, encoding='utf-8'))
|
||||||
if is_rss:
|
|
||||||
# So that we can keep CDATA for cdata_in_document_to_text() to process
|
|
||||||
parser = etree.XMLParser(strip_cdata=False)
|
|
||||||
|
|
||||||
tree = html.fromstring(bytes(html_content, encoding='utf-8'), parser=parser)
|
|
||||||
html_block = ""
|
html_block = ""
|
||||||
|
|
||||||
r = tree.xpath(xpath_filter.strip(), namespaces={'re': 'http://exslt.org/regular-expressions'})
|
r = tree.xpath(xpath_filter.strip(), namespaces={'re': 'http://exslt.org/regular-expressions'})
|
||||||
@@ -102,6 +94,7 @@ def xpath_filter(xpath_filter, html_content, append_pretty_line_formatting=False
|
|||||||
|
|
||||||
return html_block
|
return html_block
|
||||||
|
|
||||||
|
|
||||||
# Extract/find element
|
# Extract/find element
|
||||||
def extract_element(find='title', html_content=''):
|
def extract_element(find='title', html_content=''):
|
||||||
|
|
||||||
@@ -267,15 +260,8 @@ def strip_ignore_text(content, wordlist, mode="content"):
|
|||||||
|
|
||||||
return "\n".encode('utf8').join(output)
|
return "\n".encode('utf8').join(output)
|
||||||
|
|
||||||
def cdata_in_document_to_text(html_content: str, render_anchor_tag_content=False) -> str:
|
|
||||||
pattern = '<!\[CDATA\[(\s*(?:.(?<!\]\]>)\s*)*)\]\]>'
|
|
||||||
def repl(m):
|
|
||||||
text = m.group(1)
|
|
||||||
return xml_escape(html_to_text(html_content=text)).strip()
|
|
||||||
|
|
||||||
return re.sub(pattern, repl, html_content)
|
def html_to_text(html_content: str, render_anchor_tag_content=False) -> str:
|
||||||
|
|
||||||
def html_to_text(html_content: str, render_anchor_tag_content=False, is_rss=False) -> str:
|
|
||||||
"""Converts html string to a string with just the text. If ignoring
|
"""Converts html string to a string with just the text. If ignoring
|
||||||
rendering anchor tag content is enable, anchor tag content are also
|
rendering anchor tag content is enable, anchor tag content are also
|
||||||
included in the text
|
included in the text
|
||||||
@@ -291,21 +277,16 @@ def html_to_text(html_content: str, render_anchor_tag_content=False, is_rss=Fals
|
|||||||
# if anchor tag content flag is set to True define a config for
|
# if anchor tag content flag is set to True define a config for
|
||||||
# extracting this content
|
# extracting this content
|
||||||
if render_anchor_tag_content:
|
if render_anchor_tag_content:
|
||||||
|
|
||||||
parser_config = ParserConfig(
|
parser_config = ParserConfig(
|
||||||
annotation_rules={"a": ["hyperlink"]},
|
annotation_rules={"a": ["hyperlink"]}, display_links=True
|
||||||
display_links=True
|
|
||||||
)
|
)
|
||||||
# otherwise set config to None/default
|
|
||||||
|
# otherwise set config to None
|
||||||
else:
|
else:
|
||||||
parser_config = None
|
parser_config = None
|
||||||
|
|
||||||
# RSS Mode - Inscriptis will treat `title` as something else.
|
# get text and annotations via inscriptis
|
||||||
# Make it as a regular block display element (//item/title)
|
|
||||||
# This is a bit of a hack - the real way it to use XSLT to convert it to HTML #1874
|
|
||||||
if is_rss:
|
|
||||||
html_content = re.sub(r'<title([\s>])', r'<h1\1', html_content)
|
|
||||||
html_content = re.sub(r'</title>', r'</h1>', html_content)
|
|
||||||
|
|
||||||
text_content = get_text(html_content, config=parser_config)
|
text_content = get_text(html_content, config=parser_config)
|
||||||
|
|
||||||
return text_content
|
return text_content
|
||||||
|
|||||||
@@ -4,7 +4,6 @@ import os
|
|||||||
import re
|
import re
|
||||||
import time
|
import time
|
||||||
import uuid
|
import uuid
|
||||||
from pathlib import Path
|
|
||||||
|
|
||||||
# Allowable protocols, protects against javascript: etc
|
# Allowable protocols, protects against javascript: etc
|
||||||
# file:// is further checked by ALLOW_FILE_URI
|
# file:// is further checked by ALLOW_FILE_URI
|
||||||
@@ -19,7 +18,6 @@ from changedetectionio.notification import (
|
|||||||
|
|
||||||
base_config = {
|
base_config = {
|
||||||
'body': None,
|
'body': None,
|
||||||
'browser_steps_last_error_step': None,
|
|
||||||
'check_unique_lines': False, # On change-detected, compare against all history if its something new
|
'check_unique_lines': False, # On change-detected, compare against all history if its something new
|
||||||
'check_count': 0,
|
'check_count': 0,
|
||||||
'date_created': None,
|
'date_created': None,
|
||||||
@@ -27,7 +25,6 @@ base_config = {
|
|||||||
'extract_text': [], # Extract text by regex after filters
|
'extract_text': [], # Extract text by regex after filters
|
||||||
'extract_title_as_title': False,
|
'extract_title_as_title': False,
|
||||||
'fetch_backend': 'system', # plaintext, playwright etc
|
'fetch_backend': 'system', # plaintext, playwright etc
|
||||||
'fetch_time': 0.0,
|
|
||||||
'processor': 'text_json_diff', # could be restock_diff or others from .processors
|
'processor': 'text_json_diff', # could be restock_diff or others from .processors
|
||||||
'filter_failure_notification_send': strtobool(os.getenv('FILTER_FAILURE_NOTIFICATION_SEND_DEFAULT', 'True')),
|
'filter_failure_notification_send': strtobool(os.getenv('FILTER_FAILURE_NOTIFICATION_SEND_DEFAULT', 'True')),
|
||||||
'filter_text_added': True,
|
'filter_text_added': True,
|
||||||
@@ -492,13 +489,3 @@ class model(dict):
|
|||||||
filepath = os.path.join(self.watch_data_dir, 'last-fetched.br')
|
filepath = os.path.join(self.watch_data_dir, 'last-fetched.br')
|
||||||
with open(filepath, 'wb') as f:
|
with open(filepath, 'wb') as f:
|
||||||
f.write(brotli.compress(contents, mode=brotli.MODE_TEXT))
|
f.write(brotli.compress(contents, mode=brotli.MODE_TEXT))
|
||||||
|
|
||||||
@property
|
|
||||||
def get_browsersteps_available_screenshots(self):
|
|
||||||
"For knowing which screenshots are available to show the user in BrowserSteps UI"
|
|
||||||
available = []
|
|
||||||
for f in Path(self.watch_data_dir).glob('step_before-*.jpeg'):
|
|
||||||
step_n=re.search(r'step_before-(\d+)', f.name)
|
|
||||||
if step_n:
|
|
||||||
available.append(step_n.group(1))
|
|
||||||
return available
|
|
||||||
|
|||||||
@@ -11,7 +11,7 @@ from changedetectionio import content_fetcher, html_tools
|
|||||||
from changedetectionio.blueprint.price_data_follower import PRICE_DATA_TRACK_ACCEPT, PRICE_DATA_TRACK_REJECT
|
from changedetectionio.blueprint.price_data_follower import PRICE_DATA_TRACK_ACCEPT, PRICE_DATA_TRACK_REJECT
|
||||||
from copy import deepcopy
|
from copy import deepcopy
|
||||||
from . import difference_detection_processor
|
from . import difference_detection_processor
|
||||||
from ..html_tools import PERL_STYLE_REGEX, cdata_in_document_to_text
|
from ..html_tools import PERL_STYLE_REGEX
|
||||||
|
|
||||||
urllib3.disable_warnings(urllib3.exceptions.InsecureRequestWarning)
|
urllib3.disable_warnings(urllib3.exceptions.InsecureRequestWarning)
|
||||||
|
|
||||||
@@ -153,22 +153,13 @@ class perform_site_check(difference_detection_processor):
|
|||||||
|
|
||||||
is_json = 'application/json' in fetcher.get_all_headers().get('content-type', '').lower()
|
is_json = 'application/json' in fetcher.get_all_headers().get('content-type', '').lower()
|
||||||
is_html = not is_json
|
is_html = not is_json
|
||||||
is_rss = False
|
|
||||||
|
|
||||||
ctype_header = fetcher.get_all_headers().get('content-type', '').lower()
|
|
||||||
# Go into RSS preprocess for converting CDATA/comment to usable text
|
|
||||||
if any(substring in ctype_header for substring in ['application/xml', 'application/rss', 'text/xml']):
|
|
||||||
if '<rss' in fetcher.content[:100].lower():
|
|
||||||
fetcher.content = cdata_in_document_to_text(html_content=fetcher.content)
|
|
||||||
is_rss = True
|
|
||||||
|
|
||||||
# source: support, basically treat it as plaintext
|
# source: support, basically treat it as plaintext
|
||||||
if is_source:
|
if is_source:
|
||||||
is_html = False
|
is_html = False
|
||||||
is_json = False
|
is_json = False
|
||||||
|
|
||||||
inline_pdf = fetcher.get_all_headers().get('content-disposition', '') and '%PDF-1' in fetcher.content[:10]
|
if watch.is_pdf or 'application/pdf' in fetcher.get_all_headers().get('content-type', '').lower():
|
||||||
if watch.is_pdf or 'application/pdf' in fetcher.get_all_headers().get('content-type', '').lower() or inline_pdf:
|
|
||||||
from shutil import which
|
from shutil import which
|
||||||
tool = os.getenv("PDF_TO_HTML_TOOL", "pdftohtml")
|
tool = os.getenv("PDF_TO_HTML_TOOL", "pdftohtml")
|
||||||
if not which(tool):
|
if not which(tool):
|
||||||
@@ -251,8 +242,7 @@ class perform_site_check(difference_detection_processor):
|
|||||||
if filter_rule[0] == '/' or filter_rule.startswith('xpath:'):
|
if filter_rule[0] == '/' or filter_rule.startswith('xpath:'):
|
||||||
html_content += html_tools.xpath_filter(xpath_filter=filter_rule.replace('xpath:', ''),
|
html_content += html_tools.xpath_filter(xpath_filter=filter_rule.replace('xpath:', ''),
|
||||||
html_content=fetcher.content,
|
html_content=fetcher.content,
|
||||||
append_pretty_line_formatting=not is_source,
|
append_pretty_line_formatting=not is_source)
|
||||||
is_rss=is_rss)
|
|
||||||
else:
|
else:
|
||||||
# CSS Filter, extract the HTML that matches and feed that into the existing inscriptis::get_text
|
# CSS Filter, extract the HTML that matches and feed that into the existing inscriptis::get_text
|
||||||
html_content += html_tools.include_filters(include_filters=filter_rule,
|
html_content += html_tools.include_filters(include_filters=filter_rule,
|
||||||
@@ -272,9 +262,8 @@ class perform_site_check(difference_detection_processor):
|
|||||||
do_anchor = self.datastore.data["settings"]["application"].get("render_anchor_tag_content", False)
|
do_anchor = self.datastore.data["settings"]["application"].get("render_anchor_tag_content", False)
|
||||||
stripped_text_from_html = \
|
stripped_text_from_html = \
|
||||||
html_tools.html_to_text(
|
html_tools.html_to_text(
|
||||||
html_content=html_content,
|
html_content,
|
||||||
render_anchor_tag_content=do_anchor,
|
render_anchor_tag_content=do_anchor
|
||||||
is_rss=is_rss # #1874 activate the <title workaround hack
|
|
||||||
)
|
)
|
||||||
|
|
||||||
# Re #340 - return the content before the 'ignore text' was applied
|
# Re #340 - return the content before the 'ignore text' was applied
|
||||||
@@ -338,8 +327,7 @@ class perform_site_check(difference_detection_processor):
|
|||||||
# incase they specified something in '/.../x'
|
# incase they specified something in '/.../x'
|
||||||
if re.search(PERL_STYLE_REGEX, s_re, re.IGNORECASE):
|
if re.search(PERL_STYLE_REGEX, s_re, re.IGNORECASE):
|
||||||
regex = html_tools.perl_style_slash_enclosed_regex_to_options(s_re)
|
regex = html_tools.perl_style_slash_enclosed_regex_to_options(s_re)
|
||||||
# Both the regex (r) and (stripped_text_from_html) should be bytes by now
|
result = re.findall(regex.encode('utf-8'), stripped_text_from_html)
|
||||||
result = re.findall(rf"{regex}".encode('utf-8'), stripped_text_from_html)
|
|
||||||
|
|
||||||
for l in result:
|
for l in result:
|
||||||
if type(l) is tuple:
|
if type(l) is tuple:
|
||||||
|
|||||||
@@ -321,14 +321,8 @@ $(document).ready(function () {
|
|||||||
var s = '<div class="control">' + '<a data-step-index=' + i + ' class="pure-button button-secondary button-green button-xsmall apply" >Apply</a> ';
|
var s = '<div class="control">' + '<a data-step-index=' + i + ' class="pure-button button-secondary button-green button-xsmall apply" >Apply</a> ';
|
||||||
if (i > 0) {
|
if (i > 0) {
|
||||||
// The first step never gets these (Goto-site)
|
// The first step never gets these (Goto-site)
|
||||||
s += `<a data-step-index="${i}" class="pure-button button-secondary button-xsmall clear" >Clear</a> ` +
|
s += '<a data-step-index=' + i + ' class="pure-button button-secondary button-xsmall clear" >Clear</a> ' +
|
||||||
`<a data-step-index="${i}" class="pure-button button-secondary button-red button-xsmall remove" >Remove</a>`;
|
'<a data-step-index=' + i + ' class="pure-button button-secondary button-red button-xsmall remove" >Remove</a>';
|
||||||
|
|
||||||
// if a screenshot is available
|
|
||||||
if (browser_steps_available_screenshots.includes(i.toString())) {
|
|
||||||
var d = (browser_steps_last_error_step === i+1) ? 'before' : 'after';
|
|
||||||
s += ` <a data-step-index="${i}" class="pure-button button-secondary button-xsmall show-screenshot" title="Show screenshot from last run" data-type="${d}">Pic</a> `;
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
s += '</div>';
|
s += '</div>';
|
||||||
$(this).append(s)
|
$(this).append(s)
|
||||||
@@ -443,24 +437,6 @@ $(document).ready(function () {
|
|||||||
|
|
||||||
});
|
});
|
||||||
|
|
||||||
$('ul#browser_steps li .control .show-screenshot').click(function (element) {
|
|
||||||
var step_n = $(event.currentTarget).data('step-index');
|
|
||||||
w = window.open(this.href, "_blank", "width=640,height=480");
|
|
||||||
const t = $(event.currentTarget).data('type');
|
|
||||||
|
|
||||||
const url = browser_steps_fetch_screenshot_image_url + `&step_n=${step_n}&type=${t}`;
|
|
||||||
w.document.body.innerHTML = `<!DOCTYPE html>
|
|
||||||
<html lang="en">
|
|
||||||
<body>
|
|
||||||
<img src="${url}" style="width: 100%" alt="Browser Step at step ${step_n} from last run." title="Browser Step at step ${step_n} from last run."/>
|
|
||||||
</body>
|
|
||||||
</html>`;
|
|
||||||
w.document.title = `Browser Step at step ${step_n} from last run.`;
|
|
||||||
});
|
|
||||||
|
|
||||||
if (browser_steps_last_error_step) {
|
|
||||||
$("ul#browser_steps>li:nth-child("+browser_steps_last_error_step+")").addClass("browser-step-with-error");
|
|
||||||
}
|
|
||||||
|
|
||||||
$("ul#browser_steps select").change(function () {
|
$("ul#browser_steps select").change(function () {
|
||||||
set_greyed_state();
|
set_greyed_state();
|
||||||
|
|||||||
@@ -2,7 +2,8 @@ $(document).ready(function () {
|
|||||||
var a = document.getElementById("a");
|
var a = document.getElementById("a");
|
||||||
var b = document.getElementById("b");
|
var b = document.getElementById("b");
|
||||||
var result = document.getElementById("result");
|
var result = document.getElementById("result");
|
||||||
var inputs;
|
var inputs = document.getElementsByClassName("change");
|
||||||
|
inputs.current = 0;
|
||||||
|
|
||||||
$('#jump-next-diff').click(function () {
|
$('#jump-next-diff').click(function () {
|
||||||
|
|
||||||
@@ -58,6 +59,9 @@ $(document).ready(function () {
|
|||||||
result.textContent = "";
|
result.textContent = "";
|
||||||
result.appendChild(fragment);
|
result.appendChild(fragment);
|
||||||
|
|
||||||
|
// Jump at start
|
||||||
|
inputs.current = 0;
|
||||||
|
|
||||||
// For nice mouse-over hover/title information
|
// For nice mouse-over hover/title information
|
||||||
const removed_current_option = $('#diff-version option:selected')
|
const removed_current_option = $('#diff-version option:selected')
|
||||||
if (removed_current_option) {
|
if (removed_current_option) {
|
||||||
@@ -71,12 +75,8 @@ $(document).ready(function () {
|
|||||||
$(this).prop('title', 'Inserted '+inserted_current_option[0].label);
|
$(this).prop('title', 'Inserted '+inserted_current_option[0].label);
|
||||||
});
|
});
|
||||||
}
|
}
|
||||||
// Set the list of possible differences to jump to
|
|
||||||
inputs = document.querySelectorAll('#diff-ui .change')
|
next_diff();
|
||||||
// Set the "current" diff pointer
|
|
||||||
inputs.current = 0;
|
|
||||||
// Goto diff
|
|
||||||
$('#jump-next-diff').click();
|
|
||||||
}
|
}
|
||||||
|
|
||||||
$('.needs-localtime').each(function () {
|
$('.needs-localtime').each(function () {
|
||||||
|
|||||||
@@ -4,14 +4,6 @@ $(function () {
|
|||||||
$(this).closest('.unviewed').removeClass('unviewed');
|
$(this).closest('.unviewed').removeClass('unviewed');
|
||||||
});
|
});
|
||||||
|
|
||||||
$('td[data-timestamp]').each(function () {
|
|
||||||
$(this).prop('title', new Intl.DateTimeFormat(undefined,
|
|
||||||
{
|
|
||||||
dateStyle: 'full',
|
|
||||||
timeStyle: 'long'
|
|
||||||
}).format($(this).data('timestamp') * 1000));
|
|
||||||
})
|
|
||||||
|
|
||||||
$("#checkbox-assign-tag").click(function (e) {
|
$("#checkbox-assign-tag").click(function (e) {
|
||||||
$('#op_extradata').val(prompt("Enter a tag name"));
|
$('#op_extradata').val(prompt("Enter a tag name"));
|
||||||
});
|
});
|
||||||
|
|||||||
@@ -6,10 +6,6 @@
|
|||||||
}
|
}
|
||||||
|
|
||||||
li {
|
li {
|
||||||
&.browser-step-with-error {
|
|
||||||
background-color: #ffd6d6;
|
|
||||||
border-radius: 4px;
|
|
||||||
}
|
|
||||||
&:not(:first-child) {
|
&:not(:first-child) {
|
||||||
&:hover {
|
&:hover {
|
||||||
opacity: 1.0;
|
opacity: 1.0;
|
||||||
|
|||||||
@@ -1,28 +0,0 @@
|
|||||||
|
|
||||||
#selector-wrapper {
|
|
||||||
height: 100%;
|
|
||||||
max-height: 70vh;
|
|
||||||
overflow-y: scroll;
|
|
||||||
position: relative;
|
|
||||||
|
|
||||||
//width: 100%;
|
|
||||||
>img {
|
|
||||||
position: absolute;
|
|
||||||
z-index: 4;
|
|
||||||
max-width: 100%;
|
|
||||||
}
|
|
||||||
|
|
||||||
>canvas {
|
|
||||||
position: relative;
|
|
||||||
z-index: 5;
|
|
||||||
max-width: 100%;
|
|
||||||
|
|
||||||
&:hover {
|
|
||||||
cursor: pointer;
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
#selector-current-xpath {
|
|
||||||
font-size: 80%;
|
|
||||||
}
|
|
||||||
@@ -943,7 +943,32 @@ ul {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
@import "parts/_visualselector";
|
#selector-wrapper {
|
||||||
|
height: 100%;
|
||||||
|
overflow-y: scroll;
|
||||||
|
position: relative;
|
||||||
|
|
||||||
|
//width: 100%;
|
||||||
|
>img {
|
||||||
|
position: absolute;
|
||||||
|
z-index: 4;
|
||||||
|
max-width: 100%;
|
||||||
|
}
|
||||||
|
|
||||||
|
>canvas {
|
||||||
|
position: relative;
|
||||||
|
z-index: 5;
|
||||||
|
max-width: 100%;
|
||||||
|
|
||||||
|
&:hover {
|
||||||
|
cursor: pointer;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
#selector-current-xpath {
|
||||||
|
font-size: 80%;
|
||||||
|
}
|
||||||
|
|
||||||
#webdriver-override-options {
|
#webdriver-override-options {
|
||||||
input[type="number"] {
|
input[type="number"] {
|
||||||
|
|||||||
@@ -26,9 +26,6 @@
|
|||||||
#browser_steps li {
|
#browser_steps li {
|
||||||
list-style: decimal;
|
list-style: decimal;
|
||||||
padding: 5px; }
|
padding: 5px; }
|
||||||
#browser_steps li.browser-step-with-error {
|
|
||||||
background-color: #ffd6d6;
|
|
||||||
border-radius: 4px; }
|
|
||||||
#browser_steps li:not(:first-child):hover {
|
#browser_steps li:not(:first-child):hover {
|
||||||
opacity: 1.0; }
|
opacity: 1.0; }
|
||||||
#browser_steps li .control {
|
#browser_steps li .control {
|
||||||
@@ -983,7 +980,6 @@ ul {
|
|||||||
|
|
||||||
#selector-wrapper {
|
#selector-wrapper {
|
||||||
height: 100%;
|
height: 100%;
|
||||||
max-height: 70vh;
|
|
||||||
overflow-y: scroll;
|
overflow-y: scroll;
|
||||||
position: relative; }
|
position: relative; }
|
||||||
#selector-wrapper > img {
|
#selector-wrapper > img {
|
||||||
|
|||||||
@@ -244,16 +244,12 @@ class ChangeDetectionStore:
|
|||||||
import pathlib
|
import pathlib
|
||||||
|
|
||||||
self.__data['watching'][uuid].update({
|
self.__data['watching'][uuid].update({
|
||||||
'browser_steps_last_error_step' : None,
|
|
||||||
'check_count': 0,
|
|
||||||
'fetch_time' : 0.0,
|
|
||||||
'has_ldjson_price_data': None,
|
|
||||||
'last_checked': 0,
|
'last_checked': 0,
|
||||||
|
'has_ldjson_price_data': None,
|
||||||
'last_error': False,
|
'last_error': False,
|
||||||
'last_notification_error': False,
|
'last_notification_error': False,
|
||||||
'last_viewed': 0,
|
'last_viewed': 0,
|
||||||
'previous_md5': False,
|
'previous_md5': False,
|
||||||
'previous_md5_before_filters': False,
|
|
||||||
'track_ldjson_price_data': None,
|
'track_ldjson_price_data': None,
|
||||||
})
|
})
|
||||||
|
|
||||||
|
|||||||
@@ -85,7 +85,6 @@
|
|||||||
<a href="{{url_for('logout')}}" class="pure-menu-link">LOG OUT</a>
|
<a href="{{url_for('logout')}}" class="pure-menu-link">LOG OUT</a>
|
||||||
</li>
|
</li>
|
||||||
{% endif %}
|
{% endif %}
|
||||||
{% if current_user.is_authenticated or not has_password %}
|
|
||||||
<li class="pure-menu-item pure-form" id="search-menu-item">
|
<li class="pure-menu-item pure-form" id="search-menu-item">
|
||||||
<!-- We use GET here so it offers people a chance to set bookmarks etc -->
|
<!-- We use GET here so it offers people a chance to set bookmarks etc -->
|
||||||
<form name="searchForm" action="" method="GET">
|
<form name="searchForm" action="" method="GET">
|
||||||
@@ -96,7 +95,6 @@
|
|||||||
</button>
|
</button>
|
||||||
</form>
|
</form>
|
||||||
</li>
|
</li>
|
||||||
{% endif %}
|
|
||||||
<li class="pure-menu-item">
|
<li class="pure-menu-item">
|
||||||
<button class="toggle-button" id ="toggle-light-mode" type="button" title="Toggle Light/Dark Mode">
|
<button class="toggle-button" id ="toggle-light-mode" type="button" title="Toggle Light/Dark Mode">
|
||||||
<span class="visually-hidden">Toggle light/dark mode</span>
|
<span class="visually-hidden">Toggle light/dark mode</span>
|
||||||
|
|||||||
@@ -60,7 +60,7 @@
|
|||||||
</div>
|
</div>
|
||||||
|
|
||||||
<div id="diff-jump">
|
<div id="diff-jump">
|
||||||
<a id="jump-next-diff" title="Jump to next difference">Jump</a>
|
<a id="jump-next-diff">Jump</a>
|
||||||
</div>
|
</div>
|
||||||
|
|
||||||
<script src="{{url_for('static_content', group='js', filename='tabs.js')}}" defer></script>
|
<script src="{{url_for('static_content', group='js', filename='tabs.js')}}" defer></script>
|
||||||
|
|||||||
@@ -4,10 +4,8 @@
|
|||||||
{% from '_common_fields.jinja' import render_common_settings_form %}
|
{% from '_common_fields.jinja' import render_common_settings_form %}
|
||||||
<script src="{{url_for('static_content', group='js', filename='tabs.js')}}" defer></script>
|
<script src="{{url_for('static_content', group='js', filename='tabs.js')}}" defer></script>
|
||||||
<script>
|
<script>
|
||||||
const browser_steps_available_screenshots=JSON.parse('{{ watch.get_browsersteps_available_screenshots|tojson }}');
|
|
||||||
const browser_steps_config=JSON.parse('{{ browser_steps_config|tojson }}');
|
const browser_steps_config=JSON.parse('{{ browser_steps_config|tojson }}');
|
||||||
const browser_steps_fetch_screenshot_image_url="{{url_for('browser_steps.browser_steps_fetch_screenshot_image', uuid=uuid)}}";
|
|
||||||
const browser_steps_last_error_step={{ watch.browser_steps_last_error_step|tojson }};
|
|
||||||
const browser_steps_start_url="{{url_for('browser_steps.browsersteps_start_session', uuid=uuid)}}";
|
const browser_steps_start_url="{{url_for('browser_steps.browsersteps_start_session', uuid=uuid)}}";
|
||||||
const browser_steps_sync_url="{{url_for('browser_steps.browsersteps_ui_update', uuid=uuid)}}";
|
const browser_steps_sync_url="{{url_for('browser_steps.browsersteps_ui_update', uuid=uuid)}}";
|
||||||
{% if emailprefix %}
|
{% if emailprefix %}
|
||||||
@@ -51,7 +49,6 @@
|
|||||||
<li class="tab"><a href="#restock">Restock Detection</a></li>
|
<li class="tab"><a href="#restock">Restock Detection</a></li>
|
||||||
{% endif %}
|
{% endif %}
|
||||||
<li class="tab"><a href="#notifications">Notifications</a></li>
|
<li class="tab"><a href="#notifications">Notifications</a></li>
|
||||||
<li class="tab"><a href="#stats">Stats</a></li>
|
|
||||||
</ul>
|
</ul>
|
||||||
</div>
|
</div>
|
||||||
|
|
||||||
@@ -127,7 +124,7 @@
|
|||||||
<div class="pure-control-group inline-radio">
|
<div class="pure-control-group inline-radio">
|
||||||
{{ render_checkbox_field(form.ignore_status_codes) }}
|
{{ render_checkbox_field(form.ignore_status_codes) }}
|
||||||
</div>
|
</div>
|
||||||
<fieldset id="webdriver-override-options" class="hideable show-html_webdriver">
|
<fieldset id="webdriver-override-options">
|
||||||
<div class="pure-control-group">
|
<div class="pure-control-group">
|
||||||
{{ render_field(form.webdriver_delay) }}
|
{{ render_field(form.webdriver_delay) }}
|
||||||
<div class="pure-form-message-inline">
|
<div class="pure-form-message-inline">
|
||||||
@@ -139,8 +136,6 @@
|
|||||||
{% endif %}
|
{% endif %}
|
||||||
</div>
|
</div>
|
||||||
</div>
|
</div>
|
||||||
</fieldset>
|
|
||||||
<fieldset class="hideable show-html_webdriver">
|
|
||||||
<div class="pure-control-group">
|
<div class="pure-control-group">
|
||||||
{{ render_field(form.webdriver_js_execute_code) }}
|
{{ render_field(form.webdriver_js_execute_code) }}
|
||||||
<div class="pure-form-message-inline">
|
<div class="pure-form-message-inline">
|
||||||
@@ -148,8 +143,13 @@
|
|||||||
</div>
|
</div>
|
||||||
</div>
|
</div>
|
||||||
</fieldset>
|
</fieldset>
|
||||||
<fieldset class="pure-group hideable show-html_requests" id="requests-override-options">
|
<fieldset class="pure-group" id="requests-override-options">
|
||||||
<div class="pure-control-group " id="request-method">
|
{% if not playwright_enabled %}
|
||||||
|
<div class="pure-form-message-inline">
|
||||||
|
<strong>Request override is currently only used by the <i>Basic fast Plaintext/HTTP Client</i> method.</strong>
|
||||||
|
</div>
|
||||||
|
{% endif %}
|
||||||
|
<div class="pure-control-group" id="request-method">
|
||||||
{{ render_field(form.method) }}
|
{{ render_field(form.method) }}
|
||||||
</div>
|
</div>
|
||||||
<div class="pure-control-group" id="request-headers">
|
<div class="pure-control-group" id="request-headers">
|
||||||
@@ -381,7 +381,6 @@ Unavailable") }}
|
|||||||
<li>Extracts text in the final output (line by line) after other filters using regular expressions or string match;
|
<li>Extracts text in the final output (line by line) after other filters using regular expressions or string match;
|
||||||
<ul>
|
<ul>
|
||||||
<li>Regular expression ‐ example <code>/reports.+?2022/i</code></li>
|
<li>Regular expression ‐ example <code>/reports.+?2022/i</code></li>
|
||||||
<li>Regular expression to match across lines ‐ example <code>/FromHere.+?ToHere/is</code></li>
|
|
||||||
<li>Don't forget to consider the white-space at the start of a line <code>/.+?reports.+?2022/i</code></li>
|
<li>Don't forget to consider the white-space at the start of a line <code>/.+?reports.+?2022/i</code></li>
|
||||||
<li>Use <code>//(?aiLmsux))</code> type flags (more <a href="https://docs.python.org/3/library/re.html#index-15">information here</a>)<br></li>
|
<li>Use <code>//(?aiLmsux))</code> type flags (more <a href="https://docs.python.org/3/library/re.html#index-15">information here</a>)<br></li>
|
||||||
<li>Keyword example ‐ example <code>Out of stock</code></li>
|
<li>Keyword example ‐ example <code>Out of stock</code></li>
|
||||||
@@ -442,35 +441,7 @@ Unavailable") }}
|
|||||||
</fieldset>
|
</fieldset>
|
||||||
</div>
|
</div>
|
||||||
{% endif %}
|
{% endif %}
|
||||||
<div class="tab-pane-inner" id="stats">
|
|
||||||
<div class="pure-control-group">
|
|
||||||
<style>
|
|
||||||
#stats-table tr > td:first-child {
|
|
||||||
font-weight: bold;
|
|
||||||
}
|
|
||||||
</style>
|
|
||||||
<table class="pure-table" id="stats-table">
|
|
||||||
<tbody>
|
|
||||||
<tr>
|
|
||||||
<td>Check count</td>
|
|
||||||
<td>{{ watch.check_count }}</td>
|
|
||||||
</tr>
|
|
||||||
<tr>
|
|
||||||
<td>Consecutive filter failures</td>
|
|
||||||
<td>{{ watch.consecutive_filter_failures }}</td>
|
|
||||||
</tr>
|
|
||||||
<tr>
|
|
||||||
<td>History length</td>
|
|
||||||
<td>{{ watch.history|length }}</td>
|
|
||||||
</tr>
|
|
||||||
<tr>
|
|
||||||
<td>Last fetch time</td>
|
|
||||||
<td>{{ watch.fetch_time }}s</td>
|
|
||||||
</tr>
|
|
||||||
</tbody>
|
|
||||||
</table>
|
|
||||||
</div>
|
|
||||||
</div>
|
|
||||||
<div id="actions">
|
<div id="actions">
|
||||||
<div class="pure-control-group">
|
<div class="pure-control-group">
|
||||||
{{ render_button(form.save_button) }}
|
{{ render_button(form.save_button) }}
|
||||||
|
|||||||
@@ -154,8 +154,8 @@
|
|||||||
{% endfor %}
|
{% endfor %}
|
||||||
|
|
||||||
</td>
|
</td>
|
||||||
<td class="last-checked" data-timestamp="{{ watch.last_checked }}">{{watch|format_last_checked_time|safe}}</td>
|
<td class="last-checked">{{watch|format_last_checked_time|safe}}</td>
|
||||||
<td class="last-changed" data-timestamp="{{ watch.last_changed }}">{% if watch.history_n >=2 and watch.last_changed >0 %}
|
<td class="last-changed">{% if watch.history_n >=2 and watch.last_changed >0 %}
|
||||||
{{watch.last_changed|format_timestamp_timeago}}
|
{{watch.last_changed|format_timestamp_timeago}}
|
||||||
{% else %}
|
{% else %}
|
||||||
Not yet
|
Not yet
|
||||||
@@ -178,18 +178,13 @@
|
|||||||
</tbody>
|
</tbody>
|
||||||
</table>
|
</table>
|
||||||
<ul id="post-list-buttons">
|
<ul id="post-list-buttons">
|
||||||
{% if errored_count %}
|
|
||||||
<li>
|
|
||||||
<a href="{{url_for('index', with_errors=1, tag=request.args.get('tag')) }}" class="pure-button button-tag button-error ">With errors ({{ errored_count }})</a>
|
|
||||||
</li>
|
|
||||||
{% endif %}
|
|
||||||
{% if has_unviewed %}
|
{% if has_unviewed %}
|
||||||
<li>
|
<li>
|
||||||
<a href="{{url_for('mark_all_viewed',with_errors=request.args.get('with_errors',0)) }}" class="pure-button button-tag ">Mark all viewed</a>
|
<a href="{{url_for('mark_all_viewed', tag=request.args.get('tag')) }}" class="pure-button button-tag ">Mark all viewed</a>
|
||||||
</li>
|
</li>
|
||||||
{% endif %}
|
{% endif %}
|
||||||
<li>
|
<li>
|
||||||
<a href="{{ url_for('form_watch_checknow', tag=active_tag, with_errors=request.args.get('with_errors',0)) }}" class="pure-button button-tag ">Recheck
|
<a href="{{ url_for('form_watch_checknow', tag=active_tag) }}" class="pure-button button-tag ">Recheck
|
||||||
all {% if active_tag%} in "{{tags[active_tag].title}}"{%endif%}</a>
|
all {% if active_tag%} in "{{tags[active_tag].title}}"{%endif%}</a>
|
||||||
</li>
|
</li>
|
||||||
<li>
|
<li>
|
||||||
|
|||||||
@@ -202,35 +202,3 @@ def test_check_filter_and_regex_extract(client, live_server):
|
|||||||
|
|
||||||
# Should not be here
|
# Should not be here
|
||||||
assert b'Some text that did change' not in res.data
|
assert b'Some text that did change' not in res.data
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
def test_regex_error_handling(client, live_server):
|
|
||||||
|
|
||||||
#live_server_setup(live_server)
|
|
||||||
|
|
||||||
# Add our URL to the import page
|
|
||||||
test_url = url_for('test_endpoint', _external=True)
|
|
||||||
res = client.post(
|
|
||||||
url_for("import_page"),
|
|
||||||
data={"urls": test_url},
|
|
||||||
follow_redirects=True
|
|
||||||
)
|
|
||||||
assert b"1 Imported" in res.data
|
|
||||||
|
|
||||||
### test regex error handling
|
|
||||||
res = client.post(
|
|
||||||
url_for("edit_page", uuid="first"),
|
|
||||||
data={"extract_text": '/something bad\d{3/XYZ',
|
|
||||||
"url": test_url,
|
|
||||||
"fetch_backend": "html_requests"},
|
|
||||||
follow_redirects=True
|
|
||||||
)
|
|
||||||
|
|
||||||
with open('/tmp/fuck.html', 'wb') as f:
|
|
||||||
f.write(res.data)
|
|
||||||
|
|
||||||
assert b'is not a valid regular expression.' in res.data
|
|
||||||
|
|
||||||
res = client.get(url_for("form_delete", uuid="all"), follow_redirects=True)
|
|
||||||
assert b'Deleted' in res.data
|
|
||||||
|
|||||||
@@ -2,61 +2,12 @@
|
|||||||
|
|
||||||
import time
|
import time
|
||||||
from flask import url_for
|
from flask import url_for
|
||||||
from .util import set_original_response, set_modified_response, live_server_setup, wait_for_all_checks, extract_rss_token_from_UI, \
|
from .util import set_original_response, set_modified_response, live_server_setup, wait_for_all_checks, extract_rss_token_from_UI
|
||||||
extract_UUID_from_client
|
|
||||||
|
|
||||||
|
|
||||||
def set_original_cdata_xml():
|
|
||||||
test_return_data = """<rss xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:media="http://search.yahoo.com/mrss/" xmlns:atom="http://www.w3.org/2005/Atom" version="2.0">
|
|
||||||
<channel>
|
|
||||||
<title>Gizi</title>
|
|
||||||
<link>https://test.com</link>
|
|
||||||
<atom:link href="https://testsite.com" rel="self" type="application/rss+xml"/>
|
|
||||||
<description>
|
|
||||||
<![CDATA[ The Future Could Be Here ]]>
|
|
||||||
</description>
|
|
||||||
<language>en</language>
|
|
||||||
<item>
|
|
||||||
<title>
|
|
||||||
<![CDATA[ <img src="https://testsite.com/hacked.jpg"> Hackers can access your computer ]]>
|
|
||||||
</title>
|
|
||||||
<link>https://testsite.com/news/12341234234</link>
|
|
||||||
<description>
|
|
||||||
<![CDATA[ <img class="type:primaryImage" src="https://testsite.com/701c981da04869e.jpg"/><p>The days of Terminator and The Matrix could be closer. But be positive.</p><p><a href="https://testsite.com">Read more link...</a></p> ]]>
|
|
||||||
</description>
|
|
||||||
<category>cybernetics</category>
|
|
||||||
<category>rand corporation</category>
|
|
||||||
<pubDate>Tue, 17 Oct 2023 15:10:00 GMT</pubDate>
|
|
||||||
<guid isPermaLink="false">1850933241</guid>
|
|
||||||
<dc:creator>
|
|
||||||
<![CDATA[ Mr Hacker News ]]>
|
|
||||||
</dc:creator>
|
|
||||||
<media:thumbnail url="https://testsite.com/thumbnail-c224e10d81488e818701c981da04869e.jpg"/>
|
|
||||||
</item>
|
|
||||||
|
|
||||||
<item>
|
|
||||||
<title> Some other title </title>
|
|
||||||
<link>https://testsite.com/news/12341234236</link>
|
|
||||||
<description>
|
|
||||||
Some other description
|
|
||||||
</description>
|
|
||||||
</item>
|
|
||||||
</channel>
|
|
||||||
</rss>
|
|
||||||
"""
|
|
||||||
|
|
||||||
with open("test-datastore/endpoint-content.txt", "w") as f:
|
|
||||||
f.write(test_return_data)
|
|
||||||
|
|
||||||
|
|
||||||
def test_setup(client, live_server):
|
|
||||||
live_server_setup(live_server)
|
|
||||||
|
|
||||||
def test_rss_and_token(client, live_server):
|
def test_rss_and_token(client, live_server):
|
||||||
# live_server_setup(live_server)
|
|
||||||
|
|
||||||
set_original_response()
|
set_original_response()
|
||||||
rss_token = extract_rss_token_from_UI(client)
|
live_server_setup(live_server)
|
||||||
|
|
||||||
# Add our URL to the import page
|
# Add our URL to the import page
|
||||||
res = client.post(
|
res = client.post(
|
||||||
@@ -66,11 +17,11 @@ def test_rss_and_token(client, live_server):
|
|||||||
)
|
)
|
||||||
|
|
||||||
assert b"1 Imported" in res.data
|
assert b"1 Imported" in res.data
|
||||||
|
rss_token = extract_rss_token_from_UI(client)
|
||||||
|
|
||||||
wait_for_all_checks(client)
|
time.sleep(2)
|
||||||
set_modified_response()
|
|
||||||
client.get(url_for("form_watch_checknow"), follow_redirects=True)
|
client.get(url_for("form_watch_checknow"), follow_redirects=True)
|
||||||
wait_for_all_checks(client)
|
time.sleep(2)
|
||||||
|
|
||||||
# Add our URL to the import page
|
# Add our URL to the import page
|
||||||
res = client.get(
|
res = client.get(
|
||||||
@@ -86,80 +37,3 @@ def test_rss_and_token(client, live_server):
|
|||||||
)
|
)
|
||||||
assert b"Access denied, bad token" not in res.data
|
assert b"Access denied, bad token" not in res.data
|
||||||
assert b"Random content" in res.data
|
assert b"Random content" in res.data
|
||||||
|
|
||||||
res = client.get(url_for("form_delete", uuid="all"), follow_redirects=True)
|
|
||||||
|
|
||||||
def test_basic_cdata_rss_markup(client, live_server):
|
|
||||||
#live_server_setup(live_server)
|
|
||||||
|
|
||||||
set_original_cdata_xml()
|
|
||||||
|
|
||||||
test_url = url_for('test_endpoint', content_type="application/xml", _external=True)
|
|
||||||
|
|
||||||
# Add our URL to the import page
|
|
||||||
res = client.post(
|
|
||||||
url_for("import_page"),
|
|
||||||
data={"urls": test_url},
|
|
||||||
follow_redirects=True
|
|
||||||
)
|
|
||||||
|
|
||||||
assert b"1 Imported" in res.data
|
|
||||||
|
|
||||||
wait_for_all_checks(client)
|
|
||||||
|
|
||||||
res = client.get(
|
|
||||||
url_for("preview_page", uuid="first"),
|
|
||||||
follow_redirects=True
|
|
||||||
)
|
|
||||||
assert b'CDATA' not in res.data
|
|
||||||
assert b'<![' not in res.data
|
|
||||||
assert b'Hackers can access your computer' in res.data
|
|
||||||
assert b'The days of Terminator' in res.data
|
|
||||||
res = client.get(url_for("form_delete", uuid="all"), follow_redirects=True)
|
|
||||||
|
|
||||||
def test_rss_xpath_filtering(client, live_server):
|
|
||||||
#live_server_setup(live_server)
|
|
||||||
|
|
||||||
set_original_cdata_xml()
|
|
||||||
|
|
||||||
test_url = url_for('test_endpoint', content_type="application/xml", _external=True)
|
|
||||||
|
|
||||||
res = client.post(
|
|
||||||
url_for("form_quick_watch_add"),
|
|
||||||
data={"url": test_url, "tags": '', 'edit_and_watch_submit_button': 'Edit > Watch'},
|
|
||||||
follow_redirects=True
|
|
||||||
)
|
|
||||||
assert b"Watch added in Paused state, saving will unpause" in res.data
|
|
||||||
|
|
||||||
uuid = extract_UUID_from_client(client)
|
|
||||||
res = client.post(
|
|
||||||
url_for("edit_page", uuid=uuid, unpause_on_save=1),
|
|
||||||
data={
|
|
||||||
"include_filters": "//item/title",
|
|
||||||
"fetch_backend": "html_requests",
|
|
||||||
"headers": "",
|
|
||||||
"proxy": "no-proxy",
|
|
||||||
"tags": "",
|
|
||||||
"url": test_url,
|
|
||||||
},
|
|
||||||
follow_redirects=True
|
|
||||||
)
|
|
||||||
assert b"unpaused" in res.data
|
|
||||||
|
|
||||||
wait_for_all_checks(client)
|
|
||||||
|
|
||||||
res = client.get(
|
|
||||||
url_for("preview_page", uuid="first"),
|
|
||||||
follow_redirects=True
|
|
||||||
)
|
|
||||||
assert b'CDATA' not in res.data
|
|
||||||
assert b'<![' not in res.data
|
|
||||||
# #1874 All but the first <title was getting selected
|
|
||||||
# Convert any HTML with just a top level <title> to <h1> to be sure title renders
|
|
||||||
|
|
||||||
assert b'Hackers can access your computer' in res.data # Should ONLY be selected by the xpath
|
|
||||||
assert b'Some other title' in res.data # Should ONLY be selected by the xpath
|
|
||||||
assert b'The days of Terminator' not in res.data # Should NOT be selected by the xpath
|
|
||||||
assert b'Some other description' not in res.data # Should NOT be selected by the xpath
|
|
||||||
|
|
||||||
res = client.get(url_for("form_delete", uuid="all"), follow_redirects=True)
|
|
||||||
|
|||||||
@@ -238,9 +238,7 @@ class update_worker(threading.Thread):
|
|||||||
# Used as a default and also by some tests
|
# Used as a default and also by some tests
|
||||||
update_handler = text_json_diff.perform_site_check(datastore=self.datastore)
|
update_handler = text_json_diff.perform_site_check(datastore=self.datastore)
|
||||||
|
|
||||||
self.datastore.data['watching'][uuid]['browser_steps_last_error_step'] = None
|
|
||||||
changed_detected, update_obj, contents = update_handler.run(uuid, skip_when_checksum_same=queued_item_data.item.get('skip_when_checksum_same'))
|
changed_detected, update_obj, contents = update_handler.run(uuid, skip_when_checksum_same=queued_item_data.item.get('skip_when_checksum_same'))
|
||||||
|
|
||||||
# Re #342
|
# Re #342
|
||||||
# In Python 3, all strings are sequences of Unicode characters. There is a bytes type that holds raw bytes.
|
# In Python 3, all strings are sequences of Unicode characters. There is a bytes type that holds raw bytes.
|
||||||
# We then convert/.decode('utf-8') for the notification etc
|
# We then convert/.decode('utf-8') for the notification etc
|
||||||
@@ -326,13 +324,8 @@ class update_worker(threading.Thread):
|
|||||||
if not self.datastore.data['watching'].get(uuid):
|
if not self.datastore.data['watching'].get(uuid):
|
||||||
continue
|
continue
|
||||||
|
|
||||||
error_step = e.step_n + 1
|
err_text = "Warning, browser step at position {} could not run, target not found, check the watch, add a delay if necessary.".format(e.step_n+1)
|
||||||
err_text = f"Warning, browser step at position {error_step} could not run, target not found, check the watch, add a delay if necessary, view Browser Steps to see screenshot at that step"
|
self.datastore.update_watch(uuid=uuid, update_obj={'last_error': err_text})
|
||||||
self.datastore.update_watch(uuid=uuid,
|
|
||||||
update_obj={'last_error': err_text,
|
|
||||||
'browser_steps_last_error_step': error_step
|
|
||||||
}
|
|
||||||
)
|
|
||||||
|
|
||||||
|
|
||||||
if self.datastore.data['watching'][uuid].get('filter_failure_notification_send', False):
|
if self.datastore.data['watching'][uuid].get('filter_failure_notification_send', False):
|
||||||
|
|||||||
@@ -33,7 +33,7 @@ dnspython<2.3.0
|
|||||||
# jq not available on Windows so must be installed manually
|
# jq not available on Windows so must be installed manually
|
||||||
|
|
||||||
# Notification library
|
# Notification library
|
||||||
apprise~=1.6.0
|
apprise~=1.5.0
|
||||||
|
|
||||||
# apprise mqtt https://github.com/dgtlmoon/changedetection.io/issues/315
|
# apprise mqtt https://github.com/dgtlmoon/changedetection.io/issues/315
|
||||||
paho-mqtt
|
paho-mqtt
|
||||||
@@ -49,7 +49,8 @@ beautifulsoup4
|
|||||||
# XPath filtering, lxml is required by bs4 anyway, but put it here to be safe.
|
# XPath filtering, lxml is required by bs4 anyway, but put it here to be safe.
|
||||||
lxml
|
lxml
|
||||||
|
|
||||||
selenium~=4.14.0
|
# 3.141 was missing socksVersion, 3.150 was not in pypi, so we try 4.1.0
|
||||||
|
selenium~=4.1.0
|
||||||
|
|
||||||
# https://stackoverflow.com/questions/71652965/importerror-cannot-import-name-safe-str-cmp-from-werkzeug-security/71653849#71653849
|
# https://stackoverflow.com/questions/71652965/importerror-cannot-import-name-safe-str-cmp-from-werkzeug-security/71653849#71653849
|
||||||
# ImportError: cannot import name 'safe_str_cmp' from 'werkzeug.security'
|
# ImportError: cannot import name 'safe_str_cmp' from 'werkzeug.security'
|
||||||
@@ -62,8 +63,7 @@ jinja2-time
|
|||||||
|
|
||||||
# https://peps.python.org/pep-0508/#environment-markers
|
# https://peps.python.org/pep-0508/#environment-markers
|
||||||
# https://github.com/dgtlmoon/changedetection.io/pull/1009
|
# https://github.com/dgtlmoon/changedetection.io/pull/1009
|
||||||
jq~=1.3; python_version >= "3.8" and sys_platform == "darwin"
|
jq~=1.3 ;python_version >= "3.8" and sys_platform == "linux"
|
||||||
jq~=1.3; python_version >= "3.8" and sys_platform == "linux"
|
|
||||||
|
|
||||||
# Any current modern version, required so far for screenshot PNG->JPEG conversion but will be used more in the future
|
# Any current modern version, required so far for screenshot PNG->JPEG conversion but will be used more in the future
|
||||||
pillow
|
pillow
|
||||||
|
|||||||
Reference in New Issue
Block a user