1
0
Fork 0
mirror of https://github.com/dragonflydb/dragonfly.git synced 2024-12-14 11:58:02 +00:00

fix: debug crash inside parsing of ZRANGE (#3611)

Also, fix error msg for EXEC command and finally tune more fakeredis tests.

Signed-off-by: Roman Gershman <roman@dragonflydb.io>
This commit is contained in:
Roman Gershman 2024-08-31 08:02:21 +03:00 committed by GitHub
parent 06f6dcafcd
commit 5c48320496
No known key found for this signature in database
GPG key ID: B5690EEEBB952194
5 changed files with 25 additions and 866 deletions

View file

@ -75,7 +75,10 @@ uint64_t CommandId::Invoke(CmdArgList args, ConnectionContext* cntx) const {
optional<facade::ErrorReply> CommandId::Validate(CmdArgList tail_args) const {
if ((arity() > 0 && tail_args.size() + 1 != size_t(arity())) ||
(arity() < 0 && tail_args.size() + 1 < size_t(-arity()))) {
return facade::ErrorReply{facade::WrongNumArgsError(name()), kSyntaxErrType};
string prefix;
if (name() == "EXEC")
prefix = "-EXECABORT Transaction discarded because of: ";
return facade::ErrorReply{prefix + facade::WrongNumArgsError(name()), kSyntaxErrType};
}
if ((opt_mask() & CO::INTERLEAVED_KEYS)) {

View file

@ -2133,12 +2133,19 @@ void ZSetFamily::ZRange(CmdArgList args, ConnectionContext* cntx) {
void ZSetFamily::ZRangeGeneric(CmdArgList args, ConnectionContext* cntx, RangeParams range_params) {
facade::CmdArgParser parser{args.subspan(3)};
while (parser.HasNext()) {
while (true) {
if (auto err = parser.Error(); err)
return cntx->SendError(err->MakeReply());
if (!parser.HasNext())
break;
if (parser.Check("BYSCORE")) {
if (exchange(range_params.interval_type, RangeParams::SCORE) == RangeParams::LEX)
return cntx->SendError("BYSCORE and BYLEX options are not compatible");
continue;
}
if (parser.Check("BYLEX")) {
if (exchange(range_params.interval_type, RangeParams::LEX) == RangeParams::SCORE)
return cntx->SendError("BYSCORE and BYLEX options are not compatible");
@ -2152,7 +2159,8 @@ void ZSetFamily::ZRangeGeneric(CmdArgList args, ConnectionContext* cntx, RangePa
range_params.with_scores = true;
continue;
}
if (parser.Check("LIMIT").ExpectTail(2)) {
if (parser.Check("LIMIT")) {
int64_t limit;
tie(range_params.offset, limit) = parser.Next<uint32_t, int64_t>();
range_params.limit = limit < 0 ? UINT32_MAX : static_cast<uint32_t>(limit);
@ -2162,9 +2170,6 @@ void ZSetFamily::ZRangeGeneric(CmdArgList args, ConnectionContext* cntx, RangePa
return cntx->SendError(absl::StrCat("unsupported option ", parser.Peek()));
}
if (auto err = parser.Error(); err)
return cntx->SendError(err->MakeReply());
ZRangeInternal(args.subspan(0, 3), range_params, cntx);
}

View file

@ -1135,4 +1135,14 @@ TEST_F(ZSetFamilyTest, GeoRadiusByMember) {
"WITHHASH and WITHCOORDS options"));
}
TEST_F(ZSetFamilyTest, RangeLimit) {
auto resp = Run({"ZRANGEBYSCORE", "", "0.0", "0.0", "limit", "0"});
EXPECT_THAT(resp, ErrArg("syntax error"));
resp = Run({"ZRANGEBYSCORE", "", "0.0", "0.0", "limit", "0", "0"});
EXPECT_THAT(resp, ArrLen(0));
resp = Run({"ZRANGEBYSCORE", "", "0.0", "0.0", "foo"});
EXPECT_THAT(resp, ErrArg("unsupported option"));
}
} // namespace dfly

View file

@ -61,6 +61,7 @@ def test_bf_mexists(r: redis.Redis):
r.bf().add("key1", "v")
@pytest.mark.unsupported_server_types("dragonfly")
def test_bf_reserve(r: redis.Redis):
assert r.bf().reserve("bloom", 0.01, 1000)
assert r.bf().reserve("bloom_ns", 0.01, 1000, noScale=True)

View file

@ -1,860 +0,0 @@
import math
import time
from time import sleep
import pytest
import redis
from fakeredis import _msgs as msgs
timeseries_tests = pytest.importorskip("probables")
@pytest.mark.unsupported_server_types("dragonfly")
def test_add_ts_close(r: redis.Redis):
ts1 = r.ts().add(5, "*", 1)
time.sleep(0.001)
ts2 = r.ts().add(5, "*", 1)
assert abs(ts2 - ts1) < 5
@pytest.mark.unsupported_server_types("dragonfly")
def test_create_key_exist(r: redis.Redis):
assert r.ts().create(1)
with pytest.raises(redis.ResponseError) as e:
r.ts().create(1)
assert str(e.value) == msgs.TIMESERIES_KEY_EXISTS
@pytest.mark.unsupported_server_types("dragonfly")
def test_create_bad_duplicate_policy(r: redis.Redis):
with pytest.raises(redis.ResponseError) as e:
assert r.ts().create(1, duplicate_policy="bad")
assert str(e.value) == msgs.TIMESERIES_INVALID_DUPLICATE_POLICY
@pytest.mark.unsupported_server_types("dragonfly")
def test_create(r: redis.Redis):
assert r.ts().create(1)
assert r.ts().create(2, retention_msecs=5)
assert r.ts().create(3, labels={"Redis": "Labs"})
assert r.ts().create(4, retention_msecs=20, labels={"Time": "Series"})
info = r.ts().info(4)
assert 20 == info.get("retention_msecs")
assert "Series" == info["labels"]["Time"]
# Test for a chunk size of 128 Bytes
assert r.ts().create("time-serie-1", chunk_size=128)
info = r.ts().info("time-serie-1")
assert 128 == info.get("chunk_size")
@pytest.mark.unsupported_server_types("dragonfly")
def test_create_duplicate_policy(r: redis.Redis):
# Test for duplicate policy
for duplicate_policy in ["block", "last", "first", "min", "max"]:
ts_name = f"time-serie-ooo-{duplicate_policy}"
assert r.ts().create(ts_name, duplicate_policy=duplicate_policy)
info = r.ts().info(ts_name)
assert duplicate_policy == info.get("duplicate_policy")
@pytest.mark.unsupported_server_types("dragonfly")
def test_alter(r: redis.Redis):
assert r.ts().create(1)
info = r.ts().info(1)
assert 0 == info.get("retention_msecs")
assert r.ts().alter(1, retention_msecs=10)
assert {} == r.ts().info(1)["labels"]
info = r.ts().info(1)
assert 10 == info.get("retention_msecs")
assert r.ts().alter(1, labels={"Time": "Series"})
assert "Series" == r.ts().info(1)["labels"]["Time"]
info = r.ts().info(1)
assert 10 == info.get("retention_msecs")
# Test for a chunk size of 50 Bytes on TS.ALTER
with pytest.raises(redis.ResponseError) as e:
r.ts().alter(1, chunk_size=50)
assert (
str(e.value)
== "TSDB: CHUNK_SIZE value must be a multiple of 8 in the range [48 .. 1048576]"
)
@pytest.mark.unsupported_server_types("dragonfly")
def test_alter_diplicate_policy(r: redis.Redis):
assert r.ts().create(1)
info = r.ts().info(1)
assert info.get("duplicate_policy") is None
assert r.ts().alter(1, duplicate_policy="min")
info = r.ts().info(1)
assert "min" == info.get("duplicate_policy")
@pytest.mark.unsupported_server_types("dragonfly")
def test_add(r: redis.Redis):
assert 1 == r.ts().add(1, 1, 1)
assert 2 == r.ts().add(2, 2, 3, retention_msecs=10)
assert 3 == r.ts().add(3, 3, 2, labels={"Redis": "Labs"})
assert 4 == r.ts().add(
4, 4, 2, retention_msecs=10, labels={"Redis": "Labs", "Time": "Series"}
)
info = r.ts().info(4)
assert 10 == info.get("retention_msecs")
assert "Labs" == info["labels"]["Redis"]
# Test for a chunk size of 128 Bytes on TS.ADD
assert r.ts().add("time-serie-1", 1, 10.0, chunk_size=128)
info = r.ts().info("time-serie-1")
assert 128 == info.get("chunk_size")
@pytest.mark.unsupported_server_types("dragonfly")
def test_add_before_retention(r: redis.Redis):
r.ts().create("time-serie-1", retention_msecs=1000)
assert r.ts().add("time-serie-1", 10000, 10.0)
with pytest.raises(redis.ResponseError) as e:
r.ts().add("time-serie-1", 2, 20.0)
assert str(e.value) == msgs.TIMESERIES_TIMESTAMP_OLDER_THAN_RETENTION
@pytest.mark.unsupported_server_types("dragonfly")
def test_add_before_last(r: redis.Redis):
r.ts().create("time-serie-1", retention_msecs=1000)
assert r.ts().add("time-serie-1", 100, 10.0) == 100
assert r.ts().add("time-serie-1", 2, 20.0) == 2
assert r.ts().incrby("time-serie-1", 10.0, timestamp=100) == 100
with pytest.raises(redis.ResponseError) as e:
r.ts().incrby("time-serie-1", 20.0, timestamp=2)
assert (
str(e.value) == msgs.TIMESERIES_TIMESTAMP_LOWER_THAN_MAX_V7
or str(e.value) == msgs.TIMESERIES_TIMESTAMP_LOWER_THAN_MAX_V6
)
@pytest.mark.unsupported_server_types("dragonfly")
def test_add_duplicate_policy(r: redis.Redis):
# Test for duplicate policy BLOCK
assert 1 == r.ts().add("time-serie-add-ooo-block", 1, 5.0)
with pytest.raises(Exception) as e:
r.ts().add("time-serie-add-ooo-block", 1, 5.0, on_duplicate="block")
assert (
str(e.value)
== "TSDB: Error at upsert, update is not supported when DUPLICATE_POLICY is set to BLOCK mode"
)
# Test for duplicate policy LAST
assert 1 == r.ts().add("time-serie-add-ooo-last", 1, 5.0)
assert 1 == r.ts().add("time-serie-add-ooo-last", 1, 10.0, on_duplicate="last")
assert 10.0 == r.ts().get("time-serie-add-ooo-last")[1]
# Test for duplicate policy FIRST
assert 1 == r.ts().add("time-serie-add-ooo-first", 1, 5.0)
assert 1 == r.ts().add("time-serie-add-ooo-first", 1, 10.0, on_duplicate="first")
assert 5.0 == r.ts().get("time-serie-add-ooo-first")[1]
# Test for duplicate policy MAX
assert 1 == r.ts().add("time-serie-add-ooo-max", 1, 5.0)
assert 1 == r.ts().add("time-serie-add-ooo-max", 1, 10.0, on_duplicate="max")
assert 10.0 == r.ts().get("time-serie-add-ooo-max")[1]
# Test for duplicate policy MIN
assert 1 == r.ts().add("time-serie-add-ooo-min", 1, 5.0)
assert 1 == r.ts().add("time-serie-add-ooo-min", 1, 10.0, on_duplicate="min")
assert 5.0 == r.ts().get("time-serie-add-ooo-min")[1]
@pytest.mark.unsupported_server_types("dragonfly")
def test_madd(r: redis.Redis):
r.ts().create("a")
assert [1, 2, 3] == r.ts().madd([("a", 1, 5), ("a", 2, 10), ("a", 3, 15)])
@pytest.mark.unsupported_server_types("dragonfly")
def test_incrby_decrby(r: redis.Redis):
for _ in range(100):
assert r.ts().incrby(1, 1)
sleep(0.001)
assert 100 == r.ts().get(1)[1]
for _ in range(100):
assert r.ts().decrby(1, 1)
sleep(0.001)
assert 0 == r.ts().get(1)[1]
assert r.ts().incrby(2, 1.5, timestamp=5)
assert r.ts().get(2) == (5, 1.5)
assert r.ts().incrby(2, 2.25, timestamp=7)
assert r.ts().get(2) == (7, 3.75)
assert r.ts().decrby(2, 1.5, timestamp=15)
assert r.ts().get(2) == (15, 2.25)
# Test for a chunk size of 128 Bytes on TS.INCRBY
assert r.ts().incrby("time-serie-1", 10, chunk_size=128)
info = r.ts().info("time-serie-1")
assert 128 == info.get("chunk_size")
# Test for a chunk size of 128 Bytes on TS.DECRBY
assert r.ts().decrby("time-serie-2", 10, chunk_size=128)
info = r.ts().info("time-serie-2")
assert 128 == info.get("chunk_size")
@pytest.mark.unsupported_server_types("dragonfly")
def test_create_and_delete_rule(r: redis.Redis):
# test rule creation
time = 100
r.ts().create(1)
r.ts().create(2)
r.ts().createrule(1, 2, "avg", 100)
for i in range(50):
r.ts().add(1, time + i * 2, 1)
r.ts().add(1, time + i * 2 + 1, 2)
r.ts().add(1, time * 2, 1.5)
assert round(r.ts().get(2)[1], 1) == 1.5
info1 = r.ts().info(1)
assert info1.rules[0][1] == 100
info2 = r.ts().info(2)
assert info2["source_key"] == b"1"
# test rule deletion
r.ts().deleterule(1, 2)
info = r.ts().info(1)
assert not info["rules"]
info2 = r.ts().info(2)
assert info2["source_key"] is None
@pytest.mark.unsupported_server_types("dragonfly")
def test_del_range(r: redis.Redis):
with pytest.raises(redis.ResponseError) as e:
r.ts().delete("test", 0, 100)
assert str(e.value) == msgs.TIMESERIES_KEY_DOES_NOT_EXIST
for i in range(100):
r.ts().add(1, i, i % 7)
assert 22 == r.ts().delete(1, 0, 21)
assert [] == r.ts().range(1, 0, 21)
assert r.ts().range(1, 22, 22) == [(22, 1.0)]
assert r.ts().delete(1, 60, 3) == 0
@pytest.mark.unsupported_server_types("dragonfly")
def test_range(r: redis.Redis):
for i in range(100):
r.ts().add(1, i, i % 7)
assert 100 == len(r.ts().range(1, 0, 200))
for i in range(100):
r.ts().add(1, i + 200, i % 7)
assert 200 == len(r.ts().range(1, 0, 500))
range_with_count_result = r.ts().range(1, 0, 500, count=10)
assert 10 == len(range_with_count_result)
assert (0, 0) == range_with_count_result[0]
# last sample isn't returned
# assert 20 == len(r.ts().range(1, 0, 500, aggregation_type="avg", bucket_size_msec=10)) TODO
@pytest.mark.unsupported_server_types("dragonfly")
def test_range_advanced(r: redis.Redis):
for i in range(100):
r.ts().add(1, i, i % 7)
r.ts().add(1, i + 200, i % 7)
assert 2 == len(
r.ts().range(
1,
0,
500,
filter_by_ts=[i for i in range(10, 20)],
filter_by_min_value=1,
filter_by_max_value=2,
)
)
res = r.ts().range(1, 0, 10, aggregation_type="count", bucket_size_msec=10)
assert res == [(0, 10.0), (10, 1.0)]
res = r.ts().range(1, 0, 10, aggregation_type="twa", bucket_size_msec=10)
assert res == [(0, pytest.approx(2.55, 0.1)), (10, 3.0)]
@pytest.mark.unsupported_server_types("dragonfly")
def test_range_latest(r: redis.Redis):
timeseries = r.ts()
timeseries.create("t1")
timeseries.create("t2")
timeseries.createrule("t1", "t2", aggregation_type="sum", bucket_size_msec=10)
timeseries.add("t1", 1, 1)
timeseries.add("t1", 2, 3)
timeseries.add("t1", 11, 7)
timeseries.add("t1", 13, 1)
assert timeseries.range("t1", 0, 20) == [(1, 1.0), (2, 3.0), (11, 7.0), (13, 1.0)]
assert timeseries.range("t2", 0, 10) == [(0, 4.0)]
res = timeseries.range("t2", 0, 10, latest=True)
assert res == [(0, 4.0)]
assert timeseries.range("t2", 0, 9) == [(0, 4.0)]
@pytest.mark.unsupported_server_types("dragonfly")
def test_range_bucket_timestamp(r: redis.Redis):
timeseries = r.ts()
timeseries.create("t1")
timeseries.add("t1", 15, 1)
timeseries.add("t1", 17, 4)
timeseries.add("t1", 51, 3)
timeseries.add("t1", 73, 5)
timeseries.add("t1", 75, 3)
# assert timeseries.range("t1", 0, 100, align=0, aggregation_type="max", bucket_size_msec=10) == [
# (10, 4.0),
# (50, 3.0),
# (70, 5.0),
# ]
assert timeseries.range(
"t1",
0,
100,
align=0,
aggregation_type="max",
bucket_size_msec=10,
bucket_timestamp="+",
) == [(20, 4.0), (60, 3.0), (80, 5.0)]
@pytest.mark.unsupported_server_types("dragonfly")
def test_range_empty(r: redis.Redis):
timeseries = r.ts()
timeseries.create("t1")
timeseries.add("t1", 15, 1)
timeseries.add("t1", 17, 4)
timeseries.add("t1", 51, 3)
timeseries.add("t1", 73, 5)
timeseries.add("t1", 75, 3)
assert timeseries.range(
"t1", 0, 100, aggregation_type="max", bucket_size_msec=10
) == [
(10, 4.0),
(50, 3.0),
(70, 5.0),
]
res = timeseries.range(
"t1", 0, 100, aggregation_type="max", bucket_size_msec=10, empty=True
)
for i in range(len(res)):
if math.isnan(res[i][1]):
res[i] = (res[i][0], None)
resp2_expected = [
(10, 4.0),
(20, None),
(30, None),
(40, None),
(50, 3.0),
(60, None),
(70, 5.0),
]
assert res == resp2_expected
@pytest.mark.unsupported_server_types("dragonfly")
def test_rev_range(r: redis.Redis):
for i in range(100):
r.ts().add(1, i, i % 7)
assert 100 == len(r.ts().range(1, 0, 200))
for i in range(100):
r.ts().add(1, i + 200, i % 7)
assert 200 == len(r.ts().range(1, 0, 500))
assert 20 == len(
r.ts().revrange(1, 0, 500, aggregation_type="avg", bucket_size_msec=10)
)
assert 10 == len(r.ts().revrange(1, 0, 500, count=10))
assert 2 == len(
r.ts().revrange(
1,
0,
500,
filter_by_ts=[i for i in range(10, 20)],
filter_by_min_value=1,
filter_by_max_value=2,
)
)
assert r.ts().revrange(1, 0, 10, aggregation_type="count", bucket_size_msec=10) == [
(10, 1.0),
(0, 10.0),
]
assert r.ts().revrange(1, 0, 10, aggregation_type="twa", bucket_size_msec=10) == [
(10, pytest.approx(3.0, 0.1)),
(0, pytest.approx(2.55, 0.1)),
]
@pytest.mark.onlynoncluster
@pytest.mark.unsupported_server_types("dragonfly")
def test_revrange_latest(r: redis.Redis):
timeseries = r.ts()
timeseries.create("t1")
timeseries.create("t2")
timeseries.createrule("t1", "t2", aggregation_type="sum", bucket_size_msec=10)
timeseries.add("t1", 1, 1)
timeseries.add("t1", 2, 3)
timeseries.add("t1", 11, 7)
timeseries.add("t1", 13, 1)
assert timeseries.revrange("t2", 0, 10) == [(0, 4.0)]
assert timeseries.revrange("t2", 0, 10, latest=True) == [(0, 4.0)]
assert timeseries.revrange("t2", 0, 9, latest=True) == [(0, 4.0)]
@pytest.mark.unsupported_server_types("dragonfly")
def test_revrange_bucket_timestamp(r: redis.Redis):
timeseries = r.ts()
timeseries.create("t1")
timeseries.add("t1", 15, 1)
timeseries.add("t1", 17, 4)
timeseries.add("t1", 51, 3)
timeseries.add("t1", 73, 5)
timeseries.add("t1", 75, 3)
assert timeseries.revrange(
"t1", 0, 100, align=0, aggregation_type="max", bucket_size_msec=10
) == [
(70, 5.0),
(50, 3.0),
(10, 4.0),
]
assert timeseries.range(
"t1",
0,
100,
align=0,
aggregation_type="max",
bucket_size_msec=10,
bucket_timestamp="+",
) == [(20, 4.0), (60, 3.0), (80, 5.0)]
@pytest.mark.unsupported_server_types("dragonfly")
def test_revrange_empty(r: redis.Redis):
timeseries = r.ts()
timeseries.create("t1")
timeseries.add("t1", 15, 1)
timeseries.add("t1", 17, 4)
timeseries.add("t1", 51, 3)
timeseries.add("t1", 73, 5)
timeseries.add("t1", 75, 3)
assert timeseries.revrange(
"t1", 0, 100, align=0, aggregation_type="max", bucket_size_msec=10
) == [
(70, 5.0),
(50, 3.0),
(10, 4.0),
]
res = timeseries.revrange(
"t1", 0, 100, align=0, aggregation_type="max", bucket_size_msec=10, empty=True
)
for i in range(len(res)):
if math.isnan(res[i][1]):
res[i] = (res[i][0], None)
assert res == [
(70, 5.0),
(60, None),
(50, 3.0),
(40, None),
(30, None),
(20, None),
(10, 4.0),
]
@pytest.mark.onlynoncluster
@pytest.mark.unsupported_server_types("dragonfly")
def test_mrange(r: redis.Redis):
r.ts().create(1, labels={"Test": "This", "team": "ny"})
r.ts().create(2, labels={"Test": "This", "Taste": "That", "team": "sf"})
for i in range(100):
r.ts().add(1, i, i % 7)
r.ts().add(2, i, i % 11)
res = r.ts().mrange(0, 200, filters=["Test=This"])
assert 2 == len(res)
assert 100 == len(res[0]["1"][1])
res = r.ts().mrange(0, 200, filters=["Test=This"], count=10)
assert 10 == len(res[0]["1"][1])
for i in range(100):
r.ts().add(1, i + 200, i % 7)
res = r.ts().mrange(
0, 500, filters=["Test=This"], aggregation_type="avg", bucket_size_msec=10
)
assert 2 == len(res)
assert 20 == len(res[0]["1"][1])
# test withlabels
assert {} == res[0]["1"][0]
res = r.ts().mrange(0, 200, filters=["Test=This"], with_labels=True)
assert {"Test": "This", "team": "ny"} == res[0]["1"][0]
@pytest.mark.unsupported_server_types("dragonfly")
def test_multi_range_advanced(r: redis.Redis):
r.ts().create(1, labels={"Test": "This", "team": "ny"})
r.ts().create(2, labels={"Test": "This", "Taste": "That", "team": "sf"})
for i in range(100):
r.ts().add(1, i, i % 7)
r.ts().add(2, i, i % 11)
# test with selected labels
res = r.ts().mrange(0, 200, filters=["Test=This"], select_labels=["team"])
assert {"team": "ny"} == res[0]["1"][0]
assert {"team": "sf"} == res[1]["2"][0]
# test with filterby
res = r.ts().mrange(
0,
200,
filters=["Test=This"],
filter_by_ts=[i for i in range(10, 20)],
filter_by_min_value=1,
filter_by_max_value=2,
)
assert [(15, 1.0), (16, 2.0)] == res[0]["1"][1]
# test groupby
res = r.ts().mrange(0, 3, filters=["Test=This"], groupby="Test", reduce="sum")
assert [(0, 0.0), (1, 2.0), (2, 4.0), (3, 6.0)] == res[0]["Test=This"][1]
res = r.ts().mrange(0, 3, filters=["Test=This"], groupby="Test", reduce="max")
assert [(0, 0.0), (1, 1.0), (2, 2.0), (3, 3.0)] == res[0]["Test=This"][1]
res = r.ts().mrange(0, 3, filters=["Test=This"], groupby="team", reduce="min")
assert 2 == len(res)
assert [(0, 0.0), (1, 1.0), (2, 2.0), (3, 3.0)] == res[0]["team=ny"][1]
assert [(0, 0.0), (1, 1.0), (2, 2.0), (3, 3.0)] == res[1]["team=sf"][1]
# test align
res = r.ts().mrange(
0,
10,
filters=["team=ny"],
aggregation_type="count",
bucket_size_msec=10,
align="-",
)
assert [(0, 10.0), (10, 1.0)] == res[0]["1"][1]
# res = r.ts().mrange(
# 0,
# 10,
# filters=["team=ny"],
# aggregation_type="count",
# bucket_size_msec=10,
# align=5,
# )
# assert [(0, 5.0), (5, 6.0)] == res[0]["1"][1]
@pytest.mark.onlynoncluster
@pytest.mark.unsupported_server_types("dragonfly")
def test_mrange_latest(r: redis.Redis):
timeseries = r.ts()
timeseries.create("t1")
timeseries.create("t2", labels={"is_compaction": "true"})
timeseries.create("t3")
timeseries.create("t4", labels={"is_compaction": "true"})
timeseries.createrule("t1", "t2", aggregation_type="sum", bucket_size_msec=10)
timeseries.createrule("t3", "t4", aggregation_type="sum", bucket_size_msec=10)
timeseries.add("t1", 1, 1)
timeseries.add("t1", 2, 3)
timeseries.add("t1", 11, 7)
timeseries.add("t1", 13, 1)
timeseries.add("t3", 1, 1)
timeseries.add("t3", 2, 3)
timeseries.add("t3", 11, 7)
timeseries.add("t3", 13, 1)
assert r.ts().mrange(0, 10, filters=["is_compaction=true"], latest=True) == [
{"t2": [{}, [(0, 4.0)]]},
{"t4": [{}, [(0, 4.0)]]},
]
@pytest.mark.onlynoncluster
@pytest.mark.unsupported_server_types("dragonfly")
def test_multi_reverse_range(r: redis.Redis):
r.ts().create(1, labels={"Test": "This", "team": "ny"})
r.ts().create(2, labels={"Test": "This", "Taste": "That", "team": "sf"})
for i in range(100):
r.ts().add(1, i, i % 7)
r.ts().add(2, i, i % 11)
res = r.ts().mrange(0, 200, filters=["Test=This"])
assert 2 == len(res)
assert 100 == len(res[0]["1"][1])
res = r.ts().mrange(0, 200, filters=["Test=This"], count=10)
assert 10 == len(res[0]["1"][1])
for i in range(100):
r.ts().add(1, i + 200, i % 7)
res = r.ts().mrevrange(
0, 500, filters=["Test=This"], aggregation_type="avg", bucket_size_msec=10
)
assert 2 == len(res)
assert 20 == len(res[0]["1"][1])
assert {} == res[0]["1"][0]
# test withlabels
res = r.ts().mrevrange(0, 200, filters=["Test=This"], with_labels=True)
assert {"Test": "This", "team": "ny"} == res[0]["1"][0]
# test with selected labels
res = r.ts().mrevrange(0, 200, filters=["Test=This"], select_labels=["team"])
assert {"team": "ny"} == res[0]["1"][0]
assert {"team": "sf"} == res[1]["2"][0]
# test filterby
res = r.ts().mrevrange(
0,
200,
filters=["Test=This"],
filter_by_ts=[i for i in range(10, 20)],
filter_by_min_value=1,
filter_by_max_value=2,
)
assert [(16, 2.0), (15, 1.0)] == res[0]["1"][1]
# test groupby
res = r.ts().mrevrange(0, 3, filters=["Test=This"], groupby="Test", reduce="sum")
assert [(3, 6.0), (2, 4.0), (1, 2.0), (0, 0.0)] == res[0]["Test=This"][1]
res = r.ts().mrevrange(0, 3, filters=["Test=This"], groupby="Test", reduce="max")
assert [(3, 3.0), (2, 2.0), (1, 1.0), (0, 0.0)] == res[0]["Test=This"][1]
res = r.ts().mrevrange(0, 3, filters=["Test=This"], groupby="team", reduce="min")
assert 2 == len(res)
assert [(3, 3.0), (2, 2.0), (1, 1.0), (0, 0.0)] == res[0]["team=ny"][1]
assert [(3, 3.0), (2, 2.0), (1, 1.0), (0, 0.0)] == res[1]["team=sf"][1]
# test align
res = r.ts().mrevrange(
0,
10,
filters=["team=ny"],
aggregation_type="count",
bucket_size_msec=10,
align="-",
)
assert [(10, 1.0), (0, 10.0)] == res[0]["1"][1]
# res = r.ts().mrevrange(
# 0,
# 10,
# filters=["team=ny"],
# aggregation_type="count",
# bucket_size_msec=10,
# align=1,
# )
# assert [(1, 10.0), (0, 1.0)] == res[0]["1"][1]
@pytest.mark.onlynoncluster
@pytest.mark.unsupported_server_types("dragonfly")
def test_mrevrange_latest(r: redis.Redis):
timeseries = r.ts()
timeseries.create("t1")
timeseries.create("t2", labels={"is_compaction": "true"})
timeseries.create("t3")
timeseries.create("t4", labels={"is_compaction": "true"})
timeseries.createrule("t1", "t2", aggregation_type="sum", bucket_size_msec=10)
timeseries.createrule("t3", "t4", aggregation_type="sum", bucket_size_msec=10)
timeseries.add("t1", 1, 1)
timeseries.add("t1", 2, 3)
timeseries.add("t1", 11, 7)
timeseries.add("t1", 13, 1)
timeseries.add("t3", 1, 1)
timeseries.add("t3", 2, 3)
timeseries.add("t3", 11, 7)
timeseries.add("t3", 13, 1)
assert r.ts().mrevrange(0, 10, filters=["is_compaction=true"], latest=True) == [
{"t2": [{}, [(0, 4.0)]]},
{"t4": [{}, [(0, 4.0)]]},
]
@pytest.mark.unsupported_server_types("dragonfly")
def test_get(r: redis.Redis):
name = "test"
r.ts().create(name)
assert r.ts().get(name) is None
r.ts().add(name, 2, 3)
assert 2 == r.ts().get(name)[0]
r.ts().add(name, 3, 4)
assert 4 == r.ts().get(name)[1]
@pytest.mark.onlynoncluster
@pytest.mark.unsupported_server_types("dragonfly")
def test_get_latest(r: redis.Redis):
timeseries = r.ts()
timeseries.create("t1")
timeseries.create("t2")
timeseries.createrule("t1", "t2", aggregation_type="sum", bucket_size_msec=10)
timeseries.add("t1", 1, 1)
timeseries.add("t1", 2, 3)
timeseries.add("t1", 11, 7)
timeseries.add("t1", 13, 1)
assert timeseries.get("t2") == (0, 4.0)
assert timeseries.get("t2", latest=True) == (0, 4.0)
@pytest.mark.unsupported_server_types("dragonfly")
def test_mget_errors(r: redis.Redis):
r.ts().create(1, labels={"Test": "This"})
r.ts().create(2, labels={"Test": "This", "Taste": "That"})
with pytest.raises(redis.ResponseError) as e:
r.ts().mget([])
assert str(e.value).lower() == "wrong number of arguments for 'ts.mget' command"
with pytest.raises(redis.ResponseError) as e:
r.ts().mget(["Test=(Th=is"], with_labels="true")
assert str(e.value) == "TSDB: failed parsing labels"
@pytest.mark.unsupported_server_types("dragonfly")
def test_mget(r: redis.Redis):
r.ts().create(1, labels={"Test": "This"})
r.ts().create(2, labels={"Test": "This", "Taste": "That"})
act_res = r.ts().mget(["Test=This"])
exp_res = [{"1": [{}, None, None]}, {"2": [{}, None, None]}]
assert act_res == exp_res
r.ts().add(1, "*", 15)
r.ts().add(2, "*", 25)
res = r.ts().mget(["Test=This"])
assert 15 == res[0]["1"][2]
assert 25 == res[1]["2"][2]
res = r.ts().mget(["Taste=That"])
assert 25 == res[0]["2"][2]
# test with_labels
assert {} == res[0]["2"][0]
res = r.ts().mget(["Taste=That"], with_labels=True)
assert {"Taste": "That", "Test": "This"} == res[0]["2"][0]
@pytest.mark.onlynoncluster
@pytest.mark.unsupported_server_types("dragonfly")
def test_mget_latest(r: redis.Redis):
timeseries = r.ts()
timeseries.create("t1")
timeseries.create("t2", labels={"is_compaction": "true"})
timeseries.createrule("t1", "t2", aggregation_type="sum", bucket_size_msec=10)
timeseries.add("t1", 1, 1)
timeseries.add("t1", 2, 3)
timeseries.add("t1", 11, 7)
timeseries.add("t1", 13, 1)
res = timeseries.mget(filters=["is_compaction=true"])
assert res == [{"t2": [{}, 0, 4.0]}]
res = timeseries.mget(filters=["is_compaction=true"], latest=True)
assert res == [{"t2": [{}, 0, 4.0]}]
@pytest.mark.unsupported_server_types("dragonfly")
def test_info(r: redis.Redis):
r.ts().create(1, retention_msecs=5, labels={"currentLabel": "currentData"})
info = r.ts().info(1)
assert 5 == info.get("retention_msecs")
assert info["labels"]["currentLabel"] == "currentData"
@pytest.mark.unsupported_server_types("dragonfly")
def testInfoDuplicatePolicy(r: redis.Redis):
r.ts().create(1, retention_msecs=5, labels={"currentLabel": "currentData"})
info = r.ts().info(1)
assert info.get("duplicate_policy") is None
r.ts().create("time-serie-2", duplicate_policy="min")
info = r.ts().info("time-serie-2")
assert info.get("duplicate_policy") == "min"
@pytest.mark.onlynoncluster
@pytest.mark.unsupported_server_types("dragonfly")
def test_query_index(r: redis.Redis):
r.ts().create(1, labels={"Test": "This"})
r.ts().create(2, labels={"Test": "This", "Taste": "That"})
assert 2 == len(r.ts().queryindex(["Test=This"]))
assert 1 == len(r.ts().queryindex(["Taste=That"]))
assert r.ts().queryindex(["Taste=That"]) == [2]
@pytest.mark.unsupported_server_types("dragonfly")
def test_pipeline(r: redis.Redis):
pipeline = r.ts().pipeline()
pipeline.create("with_pipeline")
for i in range(100):
pipeline.add("with_pipeline", i, 1.1 * i)
pipeline.execute()
info = r.ts().info("with_pipeline")
assert 99 == info.get("last_timestamp")
assert 100 == info.get("total_samples")
assert r.ts().get("with_pipeline")[1] == 99 * 1.1
@pytest.mark.unsupported_server_types("dragonfly")
def test_uncompressed(r: redis.Redis):
r.ts().create("compressed")
r.ts().create("uncompressed", uncompressed=True)
compressed_info = r.ts().info("compressed")
uncompressed_info = r.ts().info("uncompressed")
assert compressed_info["memory_usage"] != uncompressed_info["memory_usage"]
@pytest.mark.unsupported_server_types("dragonfly")
def test_create_rule_green(r: redis.Redis):
r.ts().create(1)
r.ts().create(2)
r.ts().createrule(1, 2, "avg", 100)
for i in range(50):
r.ts().add(1, 100 + i * 2, 1)
r.ts().add(1, 100 + i * 2 + 1, 2)
r.ts().add(1, 200, 1.5)
last_sample = r.ts().get(2)
assert last_sample[0] == 100
assert round(last_sample[1], 5) == pytest.approx(1.5, 0.1)
info = r.ts().info(2)
assert info["source_key"] == b"1"
@pytest.mark.unsupported_server_types("dragonfly")
def test_create_rule_bad_aggregator(r: redis.Redis):
r.ts().create(1)
r.ts().create(2)
with pytest.raises(redis.ResponseError) as e:
r.ts().createrule(1, 2, "bad", 100, align_timestamp=50)
assert str(e.value) == msgs.TIMESERIES_BAD_AGGREGATION_TYPE
@pytest.mark.unsupported_server_types("dragonfly")
def test_create_rule_key_not_exist(r: redis.Redis):
with pytest.raises(redis.ResponseError) as e:
r.ts().createrule(1, 2, "avg", 100)
assert str(e.value) == msgs.TIMESERIES_KEY_DOES_NOT_EXIST
@pytest.mark.unsupported_server_types("dragonfly")
def test_create_rule_with_rule_to_dest_key_exists(r: redis.Redis):
r.ts().create(1)
r.ts().create(2)
r.ts().createrule(1, 2, "avg", 100)
with pytest.raises(redis.ResponseError) as e:
r.ts().createrule(1, 2, "avg", 100)
assert str(e.value) == msgs.TIMESERIES_RULE_EXISTS