403 lines
12 KiB
Ruby
403 lines
12 KiB
Ruby
# frozen_string_literal: true
|
|
|
|
require 'spec_helper'
|
|
|
|
describe Ci::JobArtifact do
|
|
let(:artifact) { create(:ci_job_artifact, :archive) }
|
|
|
|
describe "Associations" do
|
|
it { is_expected.to belong_to(:project) }
|
|
it { is_expected.to belong_to(:job) }
|
|
end
|
|
|
|
it { is_expected.to respond_to(:file) }
|
|
it { is_expected.to respond_to(:created_at) }
|
|
it { is_expected.to respond_to(:updated_at) }
|
|
|
|
it { is_expected.to delegate_method(:open).to(:file) }
|
|
it { is_expected.to delegate_method(:exists?).to(:file) }
|
|
|
|
it_behaves_like 'having unique enum values'
|
|
|
|
it_behaves_like 'UpdateProjectStatistics' do
|
|
subject { build(:ci_job_artifact, :archive, size: 107464) }
|
|
end
|
|
|
|
describe '.with_reports' do
|
|
let!(:artifact) { create(:ci_job_artifact, :archive) }
|
|
|
|
subject { described_class.with_reports }
|
|
|
|
it { is_expected.to be_empty }
|
|
|
|
context 'when there are reports' do
|
|
let!(:metrics_report) { create(:ci_job_artifact, :junit) }
|
|
let!(:codequality_report) { create(:ci_job_artifact, :codequality) }
|
|
|
|
it { is_expected.to eq([metrics_report, codequality_report]) }
|
|
end
|
|
end
|
|
|
|
describe '.test_reports' do
|
|
subject { described_class.test_reports }
|
|
|
|
context 'when there is a test report' do
|
|
let!(:artifact) { create(:ci_job_artifact, :junit) }
|
|
|
|
it { is_expected.to eq([artifact]) }
|
|
end
|
|
|
|
context 'when there are no test reports' do
|
|
let!(:artifact) { create(:ci_job_artifact, :archive) }
|
|
|
|
it { is_expected.to be_empty }
|
|
end
|
|
end
|
|
|
|
describe '.accessibility_reports' do
|
|
subject { described_class.accessibility_reports }
|
|
|
|
context 'when there is an accessibility report' do
|
|
let(:artifact) { create(:ci_job_artifact, :accessibility) }
|
|
|
|
it { is_expected.to eq([artifact]) }
|
|
end
|
|
|
|
context 'when there are no accessibility report' do
|
|
let(:artifact) { create(:ci_job_artifact, :archive) }
|
|
|
|
it { is_expected.to be_empty }
|
|
end
|
|
end
|
|
|
|
describe '.coverage_reports' do
|
|
subject { described_class.coverage_reports }
|
|
|
|
context 'when there is a coverage report' do
|
|
let!(:artifact) { create(:ci_job_artifact, :cobertura) }
|
|
|
|
it { is_expected.to eq([artifact]) }
|
|
end
|
|
|
|
context 'when there are no coverage reports' do
|
|
let!(:artifact) { create(:ci_job_artifact, :archive) }
|
|
|
|
it { is_expected.to be_empty }
|
|
end
|
|
end
|
|
|
|
describe '.terraform_reports' do
|
|
context 'when there is a terraform report' do
|
|
it 'return the job artifact' do
|
|
artifact = create(:ci_job_artifact, :terraform)
|
|
|
|
expect(described_class.terraform_reports).to eq([artifact])
|
|
end
|
|
end
|
|
|
|
context 'when there are no terraform reports' do
|
|
it 'return the an empty array' do
|
|
expect(described_class.terraform_reports).to eq([])
|
|
end
|
|
end
|
|
end
|
|
|
|
describe '.erasable' do
|
|
subject { described_class.erasable }
|
|
|
|
context 'when there is an erasable artifact' do
|
|
let!(:artifact) { create(:ci_job_artifact, :junit) }
|
|
|
|
it { is_expected.to eq([artifact]) }
|
|
end
|
|
|
|
context 'when there are no erasable artifacts' do
|
|
let!(:artifact) { create(:ci_job_artifact, :trace) }
|
|
|
|
it { is_expected.to be_empty }
|
|
end
|
|
end
|
|
|
|
describe '.archived_trace_exists_for?' do
|
|
subject { described_class.archived_trace_exists_for?(job_id) }
|
|
|
|
let!(:artifact) { create(:ci_job_artifact, :trace, job: job) }
|
|
let(:job) { create(:ci_build) }
|
|
|
|
context 'when the specified job_id exists' do
|
|
let(:job_id) { job.id }
|
|
|
|
it { is_expected.to be_truthy }
|
|
|
|
context 'when the job does have archived trace' do
|
|
let!(:artifact) { }
|
|
|
|
it { is_expected.to be_falsy }
|
|
end
|
|
end
|
|
|
|
context 'when the specified job_id does not exist' do
|
|
let(:job_id) { 10000 }
|
|
|
|
it { is_expected.to be_falsy }
|
|
end
|
|
end
|
|
|
|
describe '.for_sha' do
|
|
let(:first_pipeline) { create(:ci_pipeline) }
|
|
let(:second_pipeline) { create(:ci_pipeline, project: first_pipeline.project, sha: Digest::SHA1.hexdigest(SecureRandom.hex)) }
|
|
let!(:first_artifact) { create(:ci_job_artifact, job: create(:ci_build, pipeline: first_pipeline)) }
|
|
let!(:second_artifact) { create(:ci_job_artifact, job: create(:ci_build, pipeline: second_pipeline)) }
|
|
|
|
it 'returns job artifacts for a given pipeline sha' do
|
|
expect(described_class.for_sha(first_pipeline.sha, first_pipeline.project.id)).to eq([first_artifact])
|
|
expect(described_class.for_sha(second_pipeline.sha, first_pipeline.project.id)).to eq([second_artifact])
|
|
end
|
|
end
|
|
|
|
describe '.for_ref' do
|
|
let(:first_pipeline) { create(:ci_pipeline, ref: 'first_ref') }
|
|
let(:second_pipeline) { create(:ci_pipeline, ref: 'second_ref', project: first_pipeline.project) }
|
|
let!(:first_artifact) { create(:ci_job_artifact, job: create(:ci_build, pipeline: first_pipeline)) }
|
|
let!(:second_artifact) { create(:ci_job_artifact, job: create(:ci_build, pipeline: second_pipeline)) }
|
|
|
|
it 'returns job artifacts for a given pipeline ref' do
|
|
expect(described_class.for_ref(first_pipeline.ref, first_pipeline.project.id)).to eq([first_artifact])
|
|
expect(described_class.for_ref(second_pipeline.ref, first_pipeline.project.id)).to eq([second_artifact])
|
|
end
|
|
end
|
|
|
|
describe '.for_job_name' do
|
|
it 'returns job artifacts for a given job name' do
|
|
first_job = create(:ci_build, name: 'first')
|
|
second_job = create(:ci_build, name: 'second')
|
|
first_artifact = create(:ci_job_artifact, job: first_job)
|
|
second_artifact = create(:ci_job_artifact, job: second_job)
|
|
|
|
expect(described_class.for_job_name(first_job.name)).to eq([first_artifact])
|
|
expect(described_class.for_job_name(second_job.name)).to eq([second_artifact])
|
|
end
|
|
end
|
|
|
|
describe 'callbacks' do
|
|
describe '#schedule_background_upload' do
|
|
subject { create(:ci_job_artifact, :archive) }
|
|
|
|
context 'when object storage is disabled' do
|
|
before do
|
|
stub_artifacts_object_storage(enabled: false)
|
|
end
|
|
|
|
it 'does not schedule the migration' do
|
|
expect(ObjectStorage::BackgroundMoveWorker).not_to receive(:perform_async)
|
|
|
|
subject
|
|
end
|
|
end
|
|
|
|
context 'when object storage is enabled' do
|
|
context 'when background upload is enabled' do
|
|
before do
|
|
stub_artifacts_object_storage(background_upload: true)
|
|
end
|
|
|
|
it 'schedules the model for migration' do
|
|
expect(ObjectStorage::BackgroundMoveWorker).to receive(:perform_async).with('JobArtifactUploader', described_class.name, :file, kind_of(Numeric))
|
|
|
|
subject
|
|
end
|
|
end
|
|
|
|
context 'when background upload is disabled' do
|
|
before do
|
|
stub_artifacts_object_storage(background_upload: false)
|
|
end
|
|
|
|
it 'schedules the model for migration' do
|
|
expect(ObjectStorage::BackgroundMoveWorker).not_to receive(:perform_async)
|
|
|
|
subject
|
|
end
|
|
end
|
|
end
|
|
end
|
|
end
|
|
|
|
context 'creating the artifact' do
|
|
let(:project) { create(:project) }
|
|
let(:artifact) { create(:ci_job_artifact, :archive, project: project) }
|
|
|
|
it 'sets the size from the file size' do
|
|
expect(artifact.size).to eq(107464)
|
|
end
|
|
end
|
|
|
|
context 'updating the artifact file' do
|
|
it 'updates the artifact size' do
|
|
artifact.update!(file: fixture_file_upload('spec/fixtures/dk.png'))
|
|
expect(artifact.size).to eq(1062)
|
|
end
|
|
end
|
|
|
|
describe 'validates file format' do
|
|
subject { artifact }
|
|
|
|
described_class::TYPE_AND_FORMAT_PAIRS.except(:trace).each do |file_type, file_format|
|
|
context "when #{file_type} type with #{file_format} format" do
|
|
let(:artifact) { build(:ci_job_artifact, file_type: file_type, file_format: file_format) }
|
|
|
|
it { is_expected.to be_valid }
|
|
end
|
|
|
|
context "when #{file_type} type without format specification" do
|
|
let(:artifact) { build(:ci_job_artifact, file_type: file_type, file_format: nil) }
|
|
|
|
it { is_expected.not_to be_valid }
|
|
end
|
|
|
|
context "when #{file_type} type with other formats" do
|
|
described_class.file_formats.except(file_format).values.each do |other_format|
|
|
let(:artifact) { build(:ci_job_artifact, file_type: file_type, file_format: other_format) }
|
|
|
|
it { is_expected.not_to be_valid }
|
|
end
|
|
end
|
|
end
|
|
end
|
|
|
|
describe 'validates DEFAULT_FILE_NAMES' do
|
|
subject { described_class::DEFAULT_FILE_NAMES }
|
|
|
|
described_class.file_types.each do |file_type, _|
|
|
it "expects #{file_type} to be included" do
|
|
is_expected.to include(file_type.to_sym)
|
|
end
|
|
end
|
|
end
|
|
|
|
describe 'validates TYPE_AND_FORMAT_PAIRS' do
|
|
subject { described_class::TYPE_AND_FORMAT_PAIRS }
|
|
|
|
described_class.file_types.each do |file_type, _|
|
|
it "expects #{file_type} to be included" do
|
|
expect(described_class.file_formats).to include(subject[file_type.to_sym])
|
|
end
|
|
end
|
|
end
|
|
|
|
describe '#file' do
|
|
subject { artifact.file }
|
|
|
|
context 'the uploader api' do
|
|
it { is_expected.to respond_to(:store_dir) }
|
|
it { is_expected.to respond_to(:cache_dir) }
|
|
it { is_expected.to respond_to(:work_dir) }
|
|
end
|
|
end
|
|
|
|
describe '#each_blob' do
|
|
context 'when file format is gzip' do
|
|
context 'when gzip file contains one file' do
|
|
let(:artifact) { build(:ci_job_artifact, :junit) }
|
|
|
|
it 'iterates blob once' do
|
|
expect { |b| artifact.each_blob(&b) }.to yield_control.once
|
|
end
|
|
end
|
|
|
|
context 'when gzip file contains three files' do
|
|
let(:artifact) { build(:ci_job_artifact, :junit_with_three_testsuites) }
|
|
|
|
it 'iterates blob three times' do
|
|
expect { |b| artifact.each_blob(&b) }.to yield_control.exactly(3).times
|
|
end
|
|
end
|
|
end
|
|
|
|
context 'when file format is raw' do
|
|
let(:artifact) { build(:ci_job_artifact, :codequality, file_format: :raw) }
|
|
|
|
it 'iterates blob once' do
|
|
expect { |b| artifact.each_blob(&b) }.to yield_control.once
|
|
end
|
|
end
|
|
|
|
context 'when there are no adapters for the file format' do
|
|
let(:artifact) { build(:ci_job_artifact, :junit, file_format: :zip) }
|
|
|
|
it 'raises an error' do
|
|
expect { |b| artifact.each_blob(&b) }.to raise_error(described_class::NotSupportedAdapterError)
|
|
end
|
|
end
|
|
end
|
|
|
|
describe '#expire_in' do
|
|
subject { artifact.expire_in }
|
|
|
|
it { is_expected.to be_nil }
|
|
|
|
context 'when expire_at is specified' do
|
|
let(:expire_at) { Time.now + 7.days }
|
|
|
|
before do
|
|
artifact.expire_at = expire_at
|
|
end
|
|
|
|
it { is_expected.to be_within(5).of(expire_at - Time.now) }
|
|
end
|
|
end
|
|
|
|
describe '#expire_in=' do
|
|
subject { artifact.expire_in }
|
|
|
|
it 'when assigning valid duration' do
|
|
artifact.expire_in = '7 days'
|
|
|
|
is_expected.to be_within(10).of(7.days.to_i)
|
|
end
|
|
|
|
it 'when assigning invalid duration' do
|
|
expect { artifact.expire_in = '7 elephants' }.to raise_error(ChronicDuration::DurationParseError)
|
|
|
|
is_expected.to be_nil
|
|
end
|
|
|
|
it 'when resetting value' do
|
|
artifact.expire_in = nil
|
|
|
|
is_expected.to be_nil
|
|
end
|
|
|
|
it 'when setting to 0' do
|
|
artifact.expire_in = '0'
|
|
|
|
is_expected.to be_nil
|
|
end
|
|
end
|
|
|
|
describe 'file is being stored' do
|
|
subject { create(:ci_job_artifact, :archive) }
|
|
|
|
context 'when existing object has local store' do
|
|
it 'is stored locally' do
|
|
expect(subject.file_store).to be(ObjectStorage::Store::LOCAL)
|
|
expect(subject.file).to be_file_storage
|
|
expect(subject.file.object_store).to eq(ObjectStorage::Store::LOCAL)
|
|
end
|
|
end
|
|
|
|
context 'when direct upload is enabled' do
|
|
before do
|
|
stub_artifacts_object_storage(direct_upload: true)
|
|
end
|
|
|
|
context 'when file is stored' do
|
|
it 'is stored remotely' do
|
|
expect(subject.file_store).to eq(ObjectStorage::Store::REMOTE)
|
|
expect(subject.file).not_to be_file_storage
|
|
expect(subject.file.object_store).to eq(ObjectStorage::Store::REMOTE)
|
|
end
|
|
end
|
|
end
|
|
end
|
|
end
|