mirror of
https://github.com/fog/fog-aws.git
synced 2022-11-09 13:50:52 -05:00
Implement merge_shards
This commit is contained in:
parent
3ca6b75c86
commit
b5a1a3576c
4 changed files with 122 additions and 8 deletions
|
@ -23,7 +23,7 @@ module Fog
|
||||||
request :put_record
|
request :put_record
|
||||||
request :get_records
|
request :get_records
|
||||||
request :split_shard
|
request :split_shard
|
||||||
# request :merge_shards
|
request :merge_shards
|
||||||
# request :add_tags_to_stream
|
# request :add_tags_to_stream
|
||||||
# request :list_tags_for_stream
|
# request :list_tags_for_stream
|
||||||
# request :remove_tags_from_stream
|
# request :remove_tags_from_stream
|
||||||
|
|
86
lib/fog/aws/requests/kinesis/merge_shards.rb
Normal file
86
lib/fog/aws/requests/kinesis/merge_shards.rb
Normal file
|
@ -0,0 +1,86 @@
|
||||||
|
module Fog
|
||||||
|
module AWS
|
||||||
|
class Kinesis
|
||||||
|
class Real
|
||||||
|
# Merges two adjacent shards in a stream and combines them into a single shard to reduce the stream's capacity to ingest and transport data.
|
||||||
|
#
|
||||||
|
# ==== Options
|
||||||
|
# * AdjacentShardToMerge<~String>: The shard ID of the adjacent shard for the merge.
|
||||||
|
# * ShardToMerge<~String>: The shard ID of the shard to combine with the adjacent shard for the merge.
|
||||||
|
# * StreamName<~String>: The name of the stream for the merge.
|
||||||
|
# ==== Returns
|
||||||
|
# * response<~Excon::Response>:
|
||||||
|
#
|
||||||
|
# ==== See Also
|
||||||
|
# https://docs.aws.amazon.com/kinesis/latest/APIReference/API_MergeShards.html
|
||||||
|
#
|
||||||
|
def merge_shards(options={})
|
||||||
|
body = {
|
||||||
|
"AdjacentShardToMerge" => options.delete("AdjacentShardToMerge"),
|
||||||
|
"ShardToMerge" => options.delete("ShardToMerge"),
|
||||||
|
"StreamName" => options.delete("StreamName")
|
||||||
|
}.reject{ |_,v| v.nil? }
|
||||||
|
|
||||||
|
response = request({
|
||||||
|
'X-Amz-Target' => "Kinesis_#{@version}.MergeShards",
|
||||||
|
:body => body,
|
||||||
|
}.merge(options))
|
||||||
|
response
|
||||||
|
end
|
||||||
|
end
|
||||||
|
|
||||||
|
class Mock
|
||||||
|
def merge_shards(options={})
|
||||||
|
stream_name = options.delete("StreamName")
|
||||||
|
shard_to_merge_id = options.delete("ShardToMerge")
|
||||||
|
adjacent_shard_to_merge_id = options.delete("AdjacentShardToMerge")
|
||||||
|
|
||||||
|
unless stream = data[:kinesis_streams].detect{ |s| s["StreamName"] == stream_name }
|
||||||
|
raise 'unknown stream'
|
||||||
|
end
|
||||||
|
|
||||||
|
unless shard_to_merge = stream["Shards"].detect{ |shard| shard["ShardId"] == shard_to_merge_id }
|
||||||
|
raise 'unknown shard'
|
||||||
|
end
|
||||||
|
|
||||||
|
unless adjacent_shard_to_merge = stream["Shards"].detect{ |shard| shard["ShardId"] == adjacent_shard_to_merge_id }
|
||||||
|
raise 'unknown shard'
|
||||||
|
end
|
||||||
|
|
||||||
|
# Close shards (set an EndingSequenceNumber on them)
|
||||||
|
shard_to_merge["SequenceNumberRange"]["EndingSequenceNumber"] = next_sequence_number
|
||||||
|
adjacent_shard_to_merge["SequenceNumberRange"]["EndingSequenceNumber"] = next_sequence_number
|
||||||
|
|
||||||
|
new_starting_hash_key = [
|
||||||
|
shard_to_merge["HashKeyRange"]["StartingHashKey"].to_i,
|
||||||
|
adjacent_shard_to_merge["HashKeyRange"]["StartingHashKey"].to_i
|
||||||
|
].min.to_s
|
||||||
|
|
||||||
|
new_ending_hash_key = [
|
||||||
|
shard_to_merge["HashKeyRange"]["EndingHashKey"].to_i,
|
||||||
|
adjacent_shard_to_merge["HashKeyRange"]["EndingHashKey"].to_i
|
||||||
|
].max.to_s
|
||||||
|
|
||||||
|
# create a new shard with ParentShardId and AdjacentParentShardID
|
||||||
|
stream["Shards"] << {
|
||||||
|
"HashKeyRange"=> {
|
||||||
|
"EndingHashKey" => new_ending_hash_key,
|
||||||
|
"StartingHashKey" => new_starting_hash_key
|
||||||
|
},
|
||||||
|
"SequenceNumberRange" => {
|
||||||
|
"StartingSequenceNumber" => next_sequence_number
|
||||||
|
},
|
||||||
|
"ShardId" => next_shard_id,
|
||||||
|
"ParentShardId" => shard_to_merge_id,
|
||||||
|
"AdjacentParentShardId" => adjacent_shard_to_merge_id
|
||||||
|
}
|
||||||
|
|
||||||
|
response = Excon::Response.new
|
||||||
|
response.status = 200
|
||||||
|
response.body = ""
|
||||||
|
response
|
||||||
|
end
|
||||||
|
end
|
||||||
|
end
|
||||||
|
end
|
||||||
|
end
|
|
@ -35,6 +35,10 @@ module Fog
|
||||||
shard_id = options.delete("ShardToSplit")
|
shard_id = options.delete("ShardToSplit")
|
||||||
stream = data[:kinesis_streams].detect{ |s| s["StreamName"] == stream_name }
|
stream = data[:kinesis_streams].detect{ |s| s["StreamName"] == stream_name }
|
||||||
|
|
||||||
|
unless stream = data[:kinesis_streams].detect{ |s| s["StreamName"] == stream_name }
|
||||||
|
raise 'unknown stream'
|
||||||
|
end
|
||||||
|
|
||||||
unless shard = stream["Shards"].detect{ |shard| shard["ShardId"] == shard_id }
|
unless shard = stream["Shards"].detect{ |shard| shard["ShardId"] == shard_id }
|
||||||
raise 'unknown shard'
|
raise 'unknown shard'
|
||||||
end
|
end
|
||||||
|
@ -75,7 +79,6 @@ module Fog
|
||||||
response.status = 200
|
response.status = 200
|
||||||
response.body = ""
|
response.body = ""
|
||||||
response
|
response
|
||||||
|
|
||||||
end
|
end
|
||||||
end
|
end
|
||||||
end
|
end
|
||||||
|
|
|
@ -92,11 +92,11 @@ Shindo.tests('AWS::Kinesis | stream requests', ['aws', 'kinesis']) do
|
||||||
end
|
end
|
||||||
|
|
||||||
tests("#list_streams").formats(@list_streams_format, false) do
|
tests("#list_streams").formats(@list_streams_format, false) do
|
||||||
Fog::AWS[:kinesis].list_streams.body
|
Fog::AWS[:kinesis].list_streams.body.tap do
|
||||||
end
|
returns(true) {
|
||||||
|
Fog::AWS[:kinesis].list_streams.body["StreamNames"].include?(@stream_id)
|
||||||
tests("#list_streams").returns(true) do
|
}
|
||||||
Fog::AWS[:kinesis].list_streams.body["StreamNames"].include?(@stream_id)
|
end
|
||||||
end
|
end
|
||||||
|
|
||||||
tests("#describe_stream").formats(@describe_stream_format) do
|
tests("#describe_stream").formats(@describe_stream_format) do
|
||||||
|
@ -147,7 +147,7 @@ Shindo.tests('AWS::Kinesis | stream requests', ['aws', 'kinesis']) do
|
||||||
data
|
data
|
||||||
end
|
end
|
||||||
|
|
||||||
tests("#split_shard").formats("") do
|
tests("#split_shard").returns("") do
|
||||||
shard = Fog::AWS[:kinesis].describe_stream("StreamName" => @stream_id).body["StreamDescription"]["Shards"].first
|
shard = Fog::AWS[:kinesis].describe_stream("StreamName" => @stream_id).body["StreamDescription"]["Shards"].first
|
||||||
shard_id = shard["ShardId"]
|
shard_id = shard["ShardId"]
|
||||||
ending_hash_key = shard["HashKeyRange"]["EndingHashKey"]
|
ending_hash_key = shard["HashKeyRange"]["EndingHashKey"]
|
||||||
|
@ -180,6 +180,31 @@ Shindo.tests('AWS::Kinesis | stream requests', ['aws', 'kinesis']) do
|
||||||
result
|
result
|
||||||
end
|
end
|
||||||
|
|
||||||
|
tests("#merge_shards").returns("") do
|
||||||
|
shards = Fog::AWS[:kinesis].describe_stream("StreamName" => @stream_id).body["StreamDescription"]["Shards"]
|
||||||
|
child_shard_ids = shards.reject{ |shard| shard["SequenceNumberRange"].has_key?("EndingSequenceNumber") }.map{ |shard| shard["ShardId"] }.sort
|
||||||
|
result = Fog::AWS[:kinesis].merge_shards("StreamName" => @stream_id, "ShardToMerge" => child_shard_ids[0], "AdjacentShardToMerge" => child_shard_ids[1]).body
|
||||||
|
|
||||||
|
wait_for_status.call("ACTIVE")
|
||||||
|
shards = Fog::AWS[:kinesis].describe_stream("StreamName" => @stream_id).body["StreamDescription"]["Shards"]
|
||||||
|
parent_shards = shards.select{ |shard| child_shard_ids.include?(shard["ShardId"]) }
|
||||||
|
child_shard = shards.detect{ |shard|
|
||||||
|
shard["ParentShardId"] == child_shard_ids[0] &&
|
||||||
|
shard["AdjacentParentShardId"] == child_shard_ids[1]
|
||||||
|
}
|
||||||
|
|
||||||
|
returns(2) { parent_shards.size }
|
||||||
|
returns(false) { child_shard.nil? }
|
||||||
|
returns({
|
||||||
|
"EndingHashKey" => "340282366920938463463374607431768211455",
|
||||||
|
"StartingHashKey" => "0"
|
||||||
|
}) {
|
||||||
|
child_shard["HashKeyRange"]
|
||||||
|
}
|
||||||
|
|
||||||
|
result
|
||||||
|
end
|
||||||
|
|
||||||
tests("#delete_stream").returns("") do
|
tests("#delete_stream").returns("") do
|
||||||
Fog::AWS[:kinesis].delete_stream("StreamName" => @stream_id).body
|
Fog::AWS[:kinesis].delete_stream("StreamName" => @stream_id).body
|
||||||
end
|
end
|
||||||
|
|
Loading…
Reference in a new issue