101 lines
3.2 KiB
Ruby
101 lines
3.2 KiB
Ruby
# frozen_string_literal: true
|
|
|
|
require 'spec_helper'
|
|
|
|
RSpec.describe 'CarrierWave::Storage::Fog::File' do
|
|
let(:uploader_class) { Class.new(CarrierWave::Uploader::Base) }
|
|
let(:uploader) { uploader_class.new }
|
|
let(:storage) { CarrierWave::Storage::Fog.new(uploader) }
|
|
let(:bucket_name) { 'some-bucket' }
|
|
let(:connection) { ::Fog::Storage.new(connection_options) }
|
|
let(:bucket) { connection.directories.new(key: bucket_name )}
|
|
let(:test_filename) { 'test' }
|
|
let(:test_data) { File.read(Rails.root.join('spec/support/gitlab_stubs/gitlab_ci.yml')) }
|
|
|
|
subject { CarrierWave::Storage::Fog::File.new(uploader, storage, test_filename) }
|
|
|
|
before do
|
|
require 'fog/azurerm'
|
|
require 'fog/aws'
|
|
|
|
stub_object_storage(connection_params: connection_options, remote_directory: bucket_name)
|
|
|
|
allow(uploader).to receive(:fog_directory).and_return(bucket_name)
|
|
allow(uploader).to receive(:fog_credentials).and_return(connection_options)
|
|
|
|
bucket.files.create(key: test_filename, body: test_data) # rubocop:disable Rails/SaveBang
|
|
end
|
|
|
|
context 'AWS' do
|
|
let(:connection_options) do
|
|
{
|
|
provider: 'AWS',
|
|
aws_access_key_id: 'AWS_ACCESS_KEY',
|
|
aws_secret_access_key: 'AWS_SECRET_KEY'
|
|
}
|
|
end
|
|
|
|
describe '#copy_to' do
|
|
let(:dest_filename) { 'copied.txt'}
|
|
|
|
it 'copies the file' do
|
|
fog_file = subject.send(:file)
|
|
|
|
expect(fog_file).to receive(:concurrency=).with(10).and_call_original
|
|
# multipart_chunk_size must be explicitly set in order to leverage
|
|
# multithreaded, multipart transfers for files below 5GB.
|
|
expect(fog_file).to receive(:multipart_chunk_size=).with(10.megabytes).and_call_original
|
|
expect(fog_file).to receive(:copy).with(bucket_name, dest_filename, anything).and_call_original
|
|
|
|
result = subject.copy_to(dest_filename)
|
|
|
|
expect(result.exists?).to be true
|
|
expect(result.read).to eq(test_data)
|
|
|
|
# Sanity check that the file actually is there
|
|
copied = bucket.files.get(dest_filename)
|
|
expect(copied).to be_present
|
|
expect(copied.body).to eq(test_data)
|
|
end
|
|
end
|
|
end
|
|
|
|
context 'Azure' do
|
|
let(:connection_options) do
|
|
{
|
|
provider: 'AzureRM',
|
|
azure_storage_account_name: 'AZURE_ACCOUNT_NAME',
|
|
azure_storage_access_key: 'AZURE_ACCESS_KEY'
|
|
}
|
|
end
|
|
|
|
describe '#copy_to' do
|
|
let(:dest_filename) { 'copied.txt'}
|
|
|
|
it 'copies the file' do
|
|
result = subject.copy_to(dest_filename)
|
|
|
|
# Fog Azure provider doesn't mock the actual copied data
|
|
expect(result.exists?).to be true
|
|
end
|
|
end
|
|
|
|
describe '#authenticated_url' do
|
|
it 'has an authenticated URL' do
|
|
expect(subject.authenticated_url).to eq("https://sa.blob.core.windows.net/test_container/test_blob?token")
|
|
end
|
|
|
|
context 'with custom expire_at' do
|
|
it 'properly sets expires param' do
|
|
expire_at = 24.hours.from_now
|
|
|
|
expect_next_instance_of(Fog::Storage::AzureRM::File) do |file|
|
|
expect(file).to receive(:url).with(expire_at).and_call_original
|
|
end
|
|
|
|
expect(subject.authenticated_url(expire_at: expire_at)).to eq("https://sa.blob.core.windows.net/test_container/test_blob?token")
|
|
end
|
|
end
|
|
end
|
|
end
|
|
end
|