mirror of
https://github.com/jrnl-org/jrnl.git
synced 2025-05-10 08:38:32 +02:00
In pytest-bdd v8.0.0 it is no longer possible to match based on multiple lines, which breaks essentially all steps that support docstrings. Solve this by adding a wrapper-function for each of these instances, that matches the docstring step, and calls the original function. So, what used to be: @then(parse("the output should match\n{regex}")) @then(parse('the output should match "{regex}"')) def output_should_match(regex, cli_run): ... Is now: @then(parse("the output should match")) def output_should_match_docstring(cli_run, docstring): output_should_match(docstring, cli_run) @then(parse('the output should match "{regex}"')) def output_should_match(regex, cli_run): ... There is possibly a way around this that is much better than what I've done here, but this is a start at least.
464 lines
15 KiB
Python
464 lines
15 KiB
Python
# Copyright © 2012-2023 jrnl contributors
|
|
# License: https://www.gnu.org/licenses/gpl-3.0.html
|
|
|
|
import json
|
|
import os
|
|
import re
|
|
from xml.etree import ElementTree as ET
|
|
|
|
from pytest_bdd import then
|
|
from pytest_bdd.parsers import parse
|
|
from pytest_bdd.parsers import re as pytest_bdd_parsers_re
|
|
from ruamel.yaml import YAML
|
|
|
|
from jrnl.config import scope_config
|
|
from tests.lib.helpers import assert_equal_tags_ignoring_order
|
|
from tests.lib.helpers import does_directory_contain_files
|
|
from tests.lib.helpers import does_directory_contain_n_files
|
|
from tests.lib.helpers import get_nested_val
|
|
from tests.lib.type_builders import should_choice
|
|
|
|
SHOULD_DICT = {"Should": should_choice}
|
|
|
|
|
|
@then("we should get no error")
|
|
def should_get_no_error(cli_run):
|
|
assert cli_run["status"] == 0, cli_run["status"]
|
|
|
|
|
|
@then("we should get an error")
|
|
def should_get_an_error(cli_run):
|
|
assert cli_run["status"] != 0, cli_run["status"]
|
|
|
|
|
|
@then(parse("the output should match"))
|
|
def output_should_match_docstring(cli_run, docstring):
|
|
output_should_match(docstring, cli_run)
|
|
|
|
|
|
@then(parse('the output should match "{regex}"'))
|
|
def output_should_match(regex, cli_run):
|
|
out = cli_run["stdout"]
|
|
matches = re.findall(regex, out)
|
|
assert matches, f"\nRegex didn't match:\n{regex}\n{str(out)}\n{str(matches)}"
|
|
|
|
|
|
@then(parse("the output {it_should:Should} contain", SHOULD_DICT))
|
|
@then(
|
|
parse(
|
|
"the {which_output_stream} output {it_should:Should} contain",
|
|
SHOULD_DICT,
|
|
)
|
|
)
|
|
def output_should_contain_docstring(which_output_stream, cli_run, it_should, docstring):
|
|
output_should_contain(docstring, which_output_stream, cli_run, it_should)
|
|
|
|
|
|
@then(parse('the output {it_should:Should} contain "{expected}"', SHOULD_DICT))
|
|
@then(
|
|
parse(
|
|
'the {which_output_stream} output {it_should:Should} contain "{expected}"',
|
|
SHOULD_DICT,
|
|
)
|
|
)
|
|
def output_should_contain(expected, which_output_stream, cli_run, it_should):
|
|
output_str = (
|
|
f"\nEXPECTED:\n{expected}\n\n"
|
|
f"ACTUAL STDOUT:\n{cli_run['stdout']}\n\n"
|
|
f"ACTUAL STDERR:\n{cli_run['stderr']}"
|
|
)
|
|
|
|
assert expected
|
|
if which_output_stream is None:
|
|
assert ((expected in cli_run["stdout"]) == it_should) or (
|
|
(expected in cli_run["stderr"]) == it_should
|
|
), output_str
|
|
|
|
elif which_output_stream == "standard":
|
|
assert (expected in cli_run["stdout"]) == it_should, output_str
|
|
|
|
elif which_output_stream == "error":
|
|
assert (expected in cli_run["stderr"]) == it_should, output_str
|
|
|
|
else:
|
|
assert (expected in cli_run[which_output_stream]) == it_should, output_str
|
|
|
|
|
|
@then(parse("the output should not contain"))
|
|
def output_should_not_contain_docstring(cli_run, docstring):
|
|
output_should_not_contain(docstring, cli_run)
|
|
|
|
|
|
@then(parse('the output should not contain "{expected_output}"'))
|
|
def output_should_not_contain(expected_output, cli_run):
|
|
assert expected_output not in cli_run["stdout"]
|
|
|
|
|
|
@then(parse("the output should be"))
|
|
def output_should_be_docstring(cli_run, docstring):
|
|
output_should_be(docstring, cli_run)
|
|
|
|
|
|
@then(parse('the output should be "{expected_output}"'))
|
|
def output_should_be(expected_output, cli_run):
|
|
actual = cli_run["stdout"].strip()
|
|
expected = expected_output.strip()
|
|
assert actual == expected
|
|
|
|
|
|
@then("the output should be empty")
|
|
def output_should_be_empty(cli_run):
|
|
actual = cli_run["stdout"].strip()
|
|
assert actual == ""
|
|
|
|
|
|
@then(parse('the output should contain the date "{date}"'))
|
|
def output_should_contain_date(date, cli_run):
|
|
assert date and date in cli_run["stdout"]
|
|
|
|
|
|
@then("the output should contain pyproject.toml version")
|
|
def output_should_contain_version(cli_run, toml_version):
|
|
out = cli_run["stdout"]
|
|
assert toml_version in out, toml_version
|
|
|
|
|
|
@then("the version in the config file should be up-to-date")
|
|
def config_file_version(config_on_disk, toml_version):
|
|
config_version = config_on_disk["version"]
|
|
assert config_version == toml_version
|
|
|
|
|
|
@then(parse("the output should be {width:d} columns wide"))
|
|
def output_should_be_columns_wide(cli_run, width):
|
|
out = cli_run["stdout"]
|
|
out_lines = out.splitlines()
|
|
for line in out_lines:
|
|
assert len(line) <= width
|
|
|
|
|
|
@then(
|
|
parse(
|
|
'the default journal "{journal_file}" '
|
|
'should be in the "{journal_dir}" directory'
|
|
)
|
|
)
|
|
def default_journal_location(journal_file, journal_dir, config_on_disk, temp_dir):
|
|
default_journal_path = config_on_disk["journals"]["default"]["journal"]
|
|
expected_journal_path = (
|
|
os.path.join(temp_dir.name, journal_file)
|
|
if journal_dir == "."
|
|
else os.path.join(temp_dir.name, journal_dir, journal_file)
|
|
)
|
|
# Use os.path.samefile here because both paths might not be fully expanded.
|
|
assert os.path.samefile(default_journal_path, expected_journal_path)
|
|
|
|
|
|
@then(
|
|
parse(
|
|
'the config for journal "{journal_name}" {it_should:Should} contain',
|
|
SHOULD_DICT,
|
|
)
|
|
)
|
|
@then(parse("the config {it_should:Should} contain", SHOULD_DICT))
|
|
def config_var_on_disk_docstring(config_on_disk, journal_name, it_should, docstring):
|
|
config_var_on_disk(config_on_disk, journal_name, it_should, docstring)
|
|
|
|
|
|
@then(
|
|
parse(
|
|
'the config for journal "{journal_name}" '
|
|
'{it_should:Should} contain "{some_yaml}"',
|
|
SHOULD_DICT,
|
|
)
|
|
)
|
|
@then(parse('the config {it_should:Should} contain "{some_yaml}"', SHOULD_DICT))
|
|
def config_var_on_disk(config_on_disk, journal_name, it_should, some_yaml):
|
|
actual = config_on_disk
|
|
if journal_name:
|
|
actual = actual["journals"][journal_name]
|
|
|
|
expected = YAML(typ="safe").load(some_yaml)
|
|
|
|
actual_slice = actual
|
|
if isinstance(actual, dict):
|
|
# `expected` objects formatted in yaml only compare one level deep
|
|
actual_slice = {key: actual.get(key) for key in expected.keys()}
|
|
|
|
assert (expected == actual_slice) == it_should
|
|
|
|
|
|
@then(
|
|
parse(
|
|
'the config in memory for journal "{journal_name}" '
|
|
"{it_should:Should} contain",
|
|
SHOULD_DICT,
|
|
)
|
|
)
|
|
@then(parse("the config in memory {it_should:Should} contain", SHOULD_DICT))
|
|
def config_var_in_memory_docstring(
|
|
config_in_memory, journal_name, it_should, docstring
|
|
):
|
|
config_var_in_memory(config_in_memory, journal_name, it_should, docstring)
|
|
|
|
|
|
@then(
|
|
parse(
|
|
'the config in memory for journal "{journal_name}" '
|
|
'{it_should:Should} contain "{some_yaml}"',
|
|
SHOULD_DICT,
|
|
)
|
|
)
|
|
@then(
|
|
parse('the config in memory {it_should:Should} contain "{some_yaml}"', SHOULD_DICT)
|
|
)
|
|
def config_var_in_memory(config_in_memory, journal_name, it_should, some_yaml):
|
|
actual = config_in_memory["overrides"]
|
|
if journal_name:
|
|
actual = actual["journals"][journal_name]
|
|
|
|
expected = YAML(typ="safe").load(some_yaml)
|
|
|
|
actual_slice = actual
|
|
if isinstance(actual, dict):
|
|
# `expected` objects formatted in yaml only compare one level deep
|
|
actual_slice = {key: get_nested_val(actual, key) for key in expected.keys()}
|
|
|
|
assert (expected == actual_slice) == it_should
|
|
|
|
|
|
@then("we should be prompted for a password")
|
|
def password_was_called(cli_run):
|
|
assert cli_run["mocks"]["user_input"].return_value.input.called
|
|
|
|
|
|
@then("we should not be prompted for a password")
|
|
def password_was_not_called(cli_run):
|
|
assert not cli_run["mocks"]["user_input"].return_value.input.called
|
|
|
|
|
|
@then(parse("the cache directory should contain the files"))
|
|
def assert_dir_contains_files(cache_dir, docstring):
|
|
assert does_directory_contain_files(docstring, cache_dir["path"])
|
|
|
|
|
|
@then(
|
|
pytest_bdd_parsers_re(r"the cache directory should contain (?P<number>\d+) files")
|
|
)
|
|
def assert_dir_contains_n_files(cache_dir, number):
|
|
assert does_directory_contain_n_files(cache_dir["path"], number)
|
|
|
|
|
|
@then(parse("the journal directory should contain"))
|
|
def journal_directory_should_contain(config_on_disk, docstring):
|
|
scoped_config = scope_config(config_on_disk, "default")
|
|
|
|
assert does_directory_contain_files(docstring, scoped_config["journal"])
|
|
|
|
|
|
@then(parse('journal "{journal_name}" should not exist'))
|
|
def journal_directory_should_not_exist(config_on_disk, journal_name):
|
|
scoped_config = scope_config(config_on_disk, journal_name)
|
|
|
|
assert not does_directory_contain_files(
|
|
scoped_config["journal"], "."
|
|
), f'Journal "{journal_name}" does exist'
|
|
|
|
|
|
@then(parse("the journal {it_should:Should} exist", SHOULD_DICT))
|
|
def journal_should_not_exist(config_on_disk, it_should):
|
|
scoped_config = scope_config(config_on_disk, "default")
|
|
expected_path = scoped_config["journal"]
|
|
|
|
contains_files = does_directory_contain_files(expected_path, ".")
|
|
|
|
assert contains_files == it_should
|
|
|
|
|
|
@then(
|
|
parse(
|
|
'the journal "{journal_name}" directory {it_should:Should} exist', SHOULD_DICT
|
|
)
|
|
)
|
|
def directory_should_not_exist(config_on_disk, it_should, journal_name):
|
|
scoped_config = scope_config(config_on_disk, journal_name)
|
|
expected_path = scoped_config["journal"]
|
|
dir_exists = os.path.isdir(expected_path)
|
|
|
|
assert dir_exists == it_should
|
|
|
|
|
|
@then(parse('the content of file "{file_path}" in the cache should be'))
|
|
def content_of_file_should_be(file_path, cache_dir, docstring):
|
|
assert cache_dir["exists"]
|
|
expected_content = docstring.strip().splitlines()
|
|
|
|
with open(os.path.join(cache_dir["path"], file_path), "r") as f:
|
|
actual_content = f.read().strip().splitlines()
|
|
|
|
for actual_line, expected_line in zip(actual_content, expected_content):
|
|
if actual_line.startswith("tags: ") and expected_line.startswith("tags: "):
|
|
assert_equal_tags_ignoring_order(
|
|
actual_line, expected_line, actual_content, expected_content
|
|
)
|
|
else:
|
|
assert actual_line.strip() == expected_line.strip(), [
|
|
[actual_line.strip(), expected_line.strip()],
|
|
[actual_content, expected_content],
|
|
]
|
|
|
|
|
|
@then(parse("the cache should contain the files"))
|
|
def cache_dir_contains_files(cache_dir, docstring):
|
|
assert cache_dir["exists"]
|
|
|
|
actual_files = os.listdir(cache_dir["path"])
|
|
expected_files = docstring.split("\n")
|
|
|
|
# sort to deal with inconsistent default file ordering on different OS's
|
|
actual_files.sort()
|
|
expected_files.sort()
|
|
|
|
assert actual_files == expected_files, [actual_files, expected_files]
|
|
|
|
|
|
@then(parse("the output should be valid {language_name}"))
|
|
def assert_output_is_valid_language(cli_run, language_name):
|
|
language_name = language_name.upper()
|
|
if language_name == "XML":
|
|
xml_tree = ET.fromstring(cli_run["stdout"])
|
|
assert xml_tree, "Invalid XML"
|
|
elif language_name == "JSON":
|
|
assert json.loads(cli_run["stdout"]), "Invalid JSON"
|
|
else:
|
|
assert False, f"Language name {language_name} not recognized"
|
|
|
|
|
|
@then(parse('"{node_name}" in the parsed output should have {number:d} elements'))
|
|
def assert_parsed_output_item_count(node_name, number, parsed_output):
|
|
lang = parsed_output["lang"]
|
|
obj = parsed_output["obj"]
|
|
|
|
if lang == "XML":
|
|
xml_node_names = (node.tag for node in obj)
|
|
assert node_name in xml_node_names, str(list(xml_node_names))
|
|
|
|
actual_entry_count = len(obj.find(node_name))
|
|
assert actual_entry_count == number, actual_entry_count
|
|
|
|
elif lang == "JSON":
|
|
my_obj = obj
|
|
|
|
for node in node_name.split("."):
|
|
try:
|
|
my_obj = my_obj[int(node)]
|
|
except ValueError:
|
|
assert node in my_obj
|
|
my_obj = my_obj[node]
|
|
|
|
assert len(my_obj) == number, len(my_obj)
|
|
|
|
else:
|
|
assert False, f"Language name {lang} not recognized"
|
|
|
|
|
|
@then(parse('"{field_name}" in the parsed output should {comparison}'))
|
|
def assert_output_field_content(field_name, comparison, parsed_output, docstring):
|
|
lang = parsed_output["lang"]
|
|
obj = parsed_output["obj"]
|
|
expected_keys = docstring.split("\n")
|
|
if len(expected_keys) == 1:
|
|
expected_keys = expected_keys[0]
|
|
|
|
if lang == "XML":
|
|
xml_node_names = (node.tag for node in obj)
|
|
assert field_name in xml_node_names, str(list(xml_node_names))
|
|
|
|
if field_name == "tags":
|
|
actual_tags = set(t.attrib["name"] for t in obj.find("tags"))
|
|
assert set(actual_tags) == set(expected_keys), [
|
|
actual_tags,
|
|
set(expected_keys),
|
|
]
|
|
else:
|
|
assert False, "This test only works for tags in XML"
|
|
|
|
elif lang == "JSON":
|
|
my_obj = obj
|
|
|
|
for node in field_name.split("."):
|
|
try:
|
|
my_obj = my_obj[int(node)]
|
|
except ValueError:
|
|
assert node in my_obj, [my_obj.keys(), node]
|
|
my_obj = my_obj[node]
|
|
|
|
if comparison == "be":
|
|
if isinstance(my_obj, str):
|
|
assert expected_keys == my_obj, [my_obj, expected_keys]
|
|
else:
|
|
assert set(expected_keys) == set(my_obj), [
|
|
set(my_obj),
|
|
set(expected_keys),
|
|
]
|
|
elif comparison == "contain":
|
|
if isinstance(my_obj, str):
|
|
assert expected_keys in my_obj, [my_obj, expected_keys]
|
|
else:
|
|
assert all(elem in my_obj for elem in expected_keys), [
|
|
my_obj,
|
|
expected_keys,
|
|
]
|
|
else:
|
|
assert False, f"Language name {lang} not recognized"
|
|
|
|
|
|
@then(parse('there should be {number:d} "{item}" elements'))
|
|
def count_elements(number, item, cli_run):
|
|
actual_output = cli_run["stdout"]
|
|
xml_tree = ET.fromstring(actual_output)
|
|
assert len(xml_tree.findall(".//" + item)) == number
|
|
|
|
|
|
@then(parse("the editor {it_should:Should} have been called", SHOULD_DICT))
|
|
@then(
|
|
parse(
|
|
"the editor {it_should:Should} have been called with {num_args} arguments",
|
|
SHOULD_DICT,
|
|
)
|
|
)
|
|
def count_editor_args(num_args, cli_run, editor_state, it_should):
|
|
assert cli_run["mocks"]["editor"].called == it_should
|
|
|
|
if isinstance(num_args, int):
|
|
assert len(editor_state["command"]) == int(num_args)
|
|
|
|
|
|
@then(parse("the stdin prompt {it_should:Should} have been called", SHOULD_DICT))
|
|
def stdin_prompt_called(cli_run, it_should):
|
|
assert cli_run["mocks"]["stdin_input"].called == it_should
|
|
|
|
|
|
@then(parse('the editor filename should end with "{suffix}"'))
|
|
def editor_filename_suffix(suffix, editor_state):
|
|
editor_filename = editor_state["tmpfile"]["name"]
|
|
|
|
assert editor_state["tmpfile"]["name"].endswith(suffix), (editor_filename, suffix)
|
|
|
|
|
|
@then(parse("the editor file content should {comparison}"))
|
|
def contains_editor_file_docstring(comparison, editor_state, docstring):
|
|
contains_editor_file(comparison, docstring, editor_state)
|
|
|
|
|
|
@then(parse('the editor file content should {comparison} "{str_value}"'))
|
|
@then(parse("the editor file content should {comparison} empty"))
|
|
def contains_editor_file(comparison, str_value, editor_state):
|
|
content = editor_state["tmpfile"]["content"]
|
|
# content = f'\n"""\n{content}\n"""\n'
|
|
if comparison == "be":
|
|
assert content == str_value
|
|
elif comparison == "contain":
|
|
assert str_value in content
|
|
else:
|
|
assert False, f"Comparison '{comparison}' not supported"
|