File 0011-remove-rand_compat-and-time_compat-deps.patch of Package plumtree
From c0dd299c254f4f4aae88fd0d44747454e06f60c6 Mon Sep 17 00:00:00 2001
From: Led <ledest@gmail.com>
Date: Tue, 12 Dec 2017 18:40:57 +0200
Subject: [PATCH] remove rand_compat and time_compat deps
---
rebar.config | 2 --
rebar.lock | 4 ----
src/plumtree.app.src | 2 +-
src/plumtree_broadcast.erl | 4 +---
test/plumtree_SUITE.erl | 14 +++++++-------
test/plumtree_test_utils.erl | 2 +-
6 files changed, 10 insertions(+), 18 deletions(-)
diff --git a/rebar.config b/rebar.config
index a354e8b..2eb961d 100644
--- a/rebar.config
+++ b/rebar.config
@@ -1,8 +1,6 @@
{min_otp_version, "19.0"}.
{deps, [{lasp_support, "~>0.0"},
- time_compat,
- rand_compat,
{partisan, "~>1.2"},
{lager, "~>3.5"}]}.
diff --git a/rebar.lock b/rebar.lock
index 96b456e..683549b 100644
--- a/rebar.lock
+++ b/rebar.lock
@@ -4,8 +4,6 @@
{<<"lager">>,{pkg,<<"lager">>,<<"3.5.2">>},0},
{<<"lasp_support">>,{pkg,<<"lasp_support">>,<<"0.0.3">>},0},
{<<"partisan">>,{pkg,<<"partisan">>,<<"1.2.0">>},0},
- {<<"rand_compat">>,{pkg,<<"rand_compat">>,<<"0.0.3">>},0},
- {<<"time_compat">>,{pkg,<<"time_compat">>,<<"0.0.1">>},0},
{<<"types">>,{pkg,<<"types">>,<<"0.1.7">>},1}]}.
[
{pkg_hash,[
@@ -14,7 +12,5 @@
{<<"lager">>, <<"614A8C8F67BF99B69EB264EA22121AD25511C055AAEC09B086773D5108C6767F">>},
{<<"lasp_support">>, <<"C1B7E1A472037AE82C71D2D16A10B7D644A621B66AE5AFE834CECF170F2E9169">>},
{<<"partisan">>, <<"460DE0CB05694C5AC2BD56E1F8F46B639D68F74D5C8AFA5D617A7F3CCFE28E75">>},
- {<<"rand_compat">>, <<"011646BC1F0B0C432FE101B816F25B9BBB74A085713CEE1DAFD2D62E9415EAD3">>},
- {<<"time_compat">>, <<"23FE0AD1FDF3B5B88821B2D04B4B5E865BF587AE66056D671FE0F53514ED8139">>},
{<<"types">>, <<"7542CEE48E664521BCF0311EEC3B433E70D80DCDF34DC18D8F0D39921AFE6831">>}]}
].
diff --git a/src/plumtree.app.src b/src/plumtree.app.src
index 152b2b1..6a57471 100644
--- a/src/plumtree.app.src
+++ b/src/plumtree.app.src
@@ -2,7 +2,7 @@
[{description,"Epidemic Broadcast Trees"},
{vsn,"0.5.0"},
{registered,[]},
- {applications,[kernel,stdlib,crypto,lager,rand_compat,partisan]},
+ {applications,[kernel,stdlib,crypto,lager,partisan]},
{mod,{plumtree_app,[]}},
{modules,[]},
{env,[{plumtree_data_dir,"data"}]},
diff --git a/src/plumtree_broadcast.erl b/src/plumtree_broadcast.erl
index e3b526c..86ce710 100644
--- a/src/plumtree_broadcast.erl
+++ b/src/plumtree_broadcast.erl
@@ -561,9 +561,7 @@ random_other_node(OrdSet) ->
Size = ordsets:size(OrdSet),
case Size of
0 -> undefined;
- _ ->
- lists:nth(rand_compat:uniform(Size),
- ordsets:to_list(OrdSet))
+ _ -> lists:nth(rand:uniform(Size), ordsets:to_list(OrdSet))
end.
ack_outstanding(MessageId, Mod, Round, Root, From, State=#state{outstanding=All}) ->
diff --git a/test/plumtree_SUITE.erl b/test/plumtree_SUITE.erl
index 420d024..79d31b6 100644
--- a/test/plumtree_SUITE.erl
+++ b/test/plumtree_SUITE.erl
@@ -154,7 +154,7 @@ membership_test(Config) ->
%% check membership after cluster
check_membership(Nodes),
- BroadcastRounds1 = rand_compat:uniform(100),
+ BroadcastRounds1 = rand:uniform(100),
ct:pal("now doing ~p rounds of broadcast",
[BroadcastRounds1]),
%% do several rounds of broadcast from random nodes, then wait a bit for propagation
@@ -162,7 +162,7 @@ membership_test(Config) ->
{_, Node} = plumtree_test_utils:select_random(Nodes),
ok = rpc:call(Node,
plumtree_broadcast, broadcast,
- [{k, rand_compat:uniform()}, plumtree_test_broadcast_handler])
+ [{k, rand:uniform()}, plumtree_test_broadcast_handler])
end, lists:seq(1, BroadcastRounds1)),
%% allow 100ms per broadcast to settle
timer:sleep(100 * BroadcastRounds1),
@@ -173,12 +173,12 @@ membership_test(Config) ->
%% now inject partitions in the broadcast tree until the graph is no longer connected
%% do some rounds of broadcast in order to repair the tree
- BroadcastRounds2 = rand_compat:uniform(100),
+ BroadcastRounds2 = rand:uniform(100),
lists:foreach(fun(_) ->
{_, Node} = plumtree_test_utils:select_random(Nodes),
ok = rpc:call(Node,
plumtree_broadcast, broadcast,
- [{k, rand_compat:uniform()}, plumtree_test_broadcast_handler])
+ [{k, rand:uniform()}, plumtree_test_broadcast_handler])
end, lists:seq(1, BroadcastRounds2)),
%% allow 100ms per broadcast to settle
timer:sleep(100 * BroadcastRounds1),
@@ -221,12 +221,12 @@ broadcast_test(Config) ->
maybe_resolve_partition(Partition, Reference, Manager, Nodes),
%% do several rounds of broadcast from random nodes, then wait a bit for propagation
- BroadcastRounds1 = rand_compat:uniform(100),
+ BroadcastRounds1 = rand:uniform(100),
lists:foreach(fun(_) ->
{_, Node} = plumtree_test_utils:select_random(Nodes),
ok = rpc:call(Node,
plumtree_broadcast, broadcast,
- [{k, rand_compat:uniform()}, plumtree_test_broadcast_handler])
+ [{k, rand:uniform()}, plumtree_test_broadcast_handler])
end, lists:seq(1, BroadcastRounds1)),
%% allow 500ms per broadcast to settle
timer:sleep(200 * BroadcastRounds1),
@@ -235,7 +235,7 @@ broadcast_test(Config) ->
check_membership(Nodes),
%% do a final round of broadcast, also from a random node, which is the one we'll be checking
- Rand = rand_compat:uniform(),
+ Rand = rand:uniform(),
{_, RandomNode} = plumtree_test_utils:select_random(Nodes),
ok = rpc:call(RandomNode,
plumtree_broadcast, broadcast,
diff --git a/test/plumtree_test_utils.erl b/test/plumtree_test_utils.erl
index 6c69bb8..ab1fe32 100644
--- a/test/plumtree_test_utils.erl
+++ b/test/plumtree_test_utils.erl
@@ -159,7 +159,7 @@ select_random(List0, Omit) ->
List = List0 -- lists:flatten([Omit]),
%% Catch exceptions where there may not be enough members.
try
- Index = rand_compat:uniform(length(List)),
+ Index = rand:uniform(length(List)),
lists:nth(Index, List)
catch
_:_ ->
--
2.15.1