2018-03-05 06:16:17 -05:00
|
|
|
require 'spec_helper'
|
|
|
|
|
|
|
|
describe Gitlab::Verify::Uploads do
|
|
|
|
include GitlabVerifyHelpers
|
|
|
|
|
|
|
|
it_behaves_like 'Gitlab::Verify::BatchVerifier subclass' do
|
|
|
|
let(:projects) { create_list(:project, 3, :with_avatar) }
|
|
|
|
let!(:objects) { projects.flat_map(&:uploads) }
|
|
|
|
end
|
|
|
|
|
|
|
|
describe '#run_batches' do
|
|
|
|
let(:project) { create(:project, :with_avatar) }
|
|
|
|
let(:failures) { collect_failures }
|
|
|
|
let(:failure) { failures[upload] }
|
|
|
|
|
|
|
|
let!(:upload) { project.uploads.first }
|
|
|
|
|
|
|
|
it 'passes uploads with the correct file' do
|
|
|
|
expect(failures).to eq({})
|
|
|
|
end
|
|
|
|
|
|
|
|
it 'fails uploads with a missing file' do
|
|
|
|
FileUtils.rm_f(upload.absolute_path)
|
|
|
|
|
|
|
|
expect(failures.keys).to contain_exactly(upload)
|
2018-06-13 13:11:43 -04:00
|
|
|
expect(failure).to include('No such file or directory')
|
|
|
|
expect(failure).to include(upload.absolute_path)
|
2018-03-05 06:16:17 -05:00
|
|
|
end
|
|
|
|
|
|
|
|
it 'fails uploads with a mismatched checksum' do
|
|
|
|
upload.update!(checksum: 'something incorrect')
|
|
|
|
|
|
|
|
expect(failures.keys).to contain_exactly(upload)
|
2018-06-13 13:11:43 -04:00
|
|
|
expect(failure).to include('Checksum mismatch')
|
2018-03-05 06:16:17 -05:00
|
|
|
end
|
|
|
|
|
|
|
|
it 'fails uploads with a missing precalculated checksum' do
|
|
|
|
upload.update!(checksum: '')
|
|
|
|
|
|
|
|
expect(failures.keys).to contain_exactly(upload)
|
2018-06-13 13:11:43 -04:00
|
|
|
expect(failure).to include('Checksum missing')
|
2018-03-05 06:16:17 -05:00
|
|
|
end
|
|
|
|
|
|
|
|
context 'with remote files' do
|
2018-06-13 13:11:43 -04:00
|
|
|
let(:file) { double(:file) }
|
|
|
|
|
2018-03-05 06:16:17 -05:00
|
|
|
before do
|
|
|
|
stub_uploads_object_storage(AvatarUploader)
|
2018-06-13 13:11:43 -04:00
|
|
|
upload.update!(store: ObjectStorage::Store::REMOTE)
|
2018-03-05 06:16:17 -05:00
|
|
|
end
|
|
|
|
|
2018-06-19 15:32:01 -04:00
|
|
|
describe 'returned hash object' do
|
|
|
|
before do
|
|
|
|
expect(CarrierWave::Storage::Fog::File).to receive(:new).and_return(file)
|
|
|
|
end
|
2018-06-13 13:11:43 -04:00
|
|
|
|
2018-06-19 15:32:01 -04:00
|
|
|
it 'passes uploads in object storage that exist' do
|
|
|
|
expect(file).to receive(:exists?).and_return(true)
|
|
|
|
|
|
|
|
expect(failures).to eq({})
|
|
|
|
end
|
|
|
|
|
|
|
|
it 'fails uploads in object storage that do not exist' do
|
|
|
|
expect(file).to receive(:exists?).and_return(false)
|
|
|
|
|
|
|
|
expect(failures.keys).to contain_exactly(upload)
|
|
|
|
expect(failure).to include('Remote object does not exist')
|
|
|
|
end
|
2018-06-13 13:11:43 -04:00
|
|
|
end
|
2018-03-05 06:16:17 -05:00
|
|
|
|
2018-06-19 15:32:01 -04:00
|
|
|
describe 'performance' do
|
|
|
|
before do
|
|
|
|
allow(file).to receive(:exists?)
|
|
|
|
allow(CarrierWave::Storage::Fog::File).to receive(:new).and_return(file)
|
|
|
|
end
|
|
|
|
|
|
|
|
it "avoids N+1 queries" do
|
|
|
|
control_count = ActiveRecord::QueryRecorder.new { perform_task }
|
|
|
|
|
|
|
|
# Create additional uploads in object storage
|
|
|
|
projects = create_list(:project, 3, :with_avatar)
|
|
|
|
uploads = projects.flat_map(&:uploads)
|
|
|
|
uploads.each do |upload|
|
|
|
|
upload.update!(store: ObjectStorage::Store::REMOTE)
|
|
|
|
end
|
|
|
|
|
|
|
|
expect { perform_task }.not_to exceed_query_limit(control_count)
|
|
|
|
end
|
2018-03-05 06:16:17 -05:00
|
|
|
|
2018-06-19 15:32:01 -04:00
|
|
|
def perform_task
|
|
|
|
described_class.new(batch_size: 100).run_batches { }
|
|
|
|
end
|
2018-03-05 06:16:17 -05:00
|
|
|
end
|
|
|
|
end
|
|
|
|
end
|
|
|
|
end
|