Welcome to mirror list, hosted at ThFree Co, Russian Federation.

gitlab.com/gitlab-org/gitlab-foss.git - Unnamed repository; edit this file 'description' to name the repository.
summaryrefslogtreecommitdiff
diff options
context:
space:
mode:
Diffstat (limited to 'spec/workers/bulk_import_worker_spec.rb')
-rw-r--r--spec/workers/bulk_import_worker_spec.rb306
1 files changed, 14 insertions, 292 deletions
diff --git a/spec/workers/bulk_import_worker_spec.rb b/spec/workers/bulk_import_worker_spec.rb
index c96e5ace124..8b73549e071 100644
--- a/spec/workers/bulk_import_worker_spec.rb
+++ b/spec/workers/bulk_import_worker_spec.rb
@@ -3,306 +3,28 @@
require 'spec_helper'
RSpec.describe BulkImportWorker, feature_category: :importers do
- describe '#perform' do
- context 'when no bulk import is found' do
- it 'does nothing' do
- expect(described_class).not_to receive(:perform_in)
-
- subject.perform(non_existing_record_id)
- end
- end
-
- context 'when bulk import is finished' do
- it 'does nothing' do
- bulk_import = create(:bulk_import, :finished)
-
- expect(described_class).not_to receive(:perform_in)
-
- subject.perform(bulk_import.id)
- end
- end
-
- context 'when bulk import is failed' do
- it 'does nothing' do
- bulk_import = create(:bulk_import, :failed)
-
- expect(described_class).not_to receive(:perform_in)
-
- subject.perform(bulk_import.id)
- end
- end
-
- context 'when all entities are processed' do
- it 'marks bulk import as finished' do
- bulk_import = create(:bulk_import, :started)
- create(:bulk_import_entity, :finished, bulk_import: bulk_import)
- create(:bulk_import_entity, :failed, bulk_import: bulk_import)
-
- subject.perform(bulk_import.id)
-
- expect(bulk_import.reload.finished?).to eq(true)
- end
- end
-
- context 'when all entities are failed' do
- it 'marks bulk import as failed' do
- bulk_import = create(:bulk_import, :started)
- create(:bulk_import_entity, :failed, bulk_import: bulk_import)
- create(:bulk_import_entity, :failed, bulk_import: bulk_import)
-
- subject.perform(bulk_import.id)
-
- expect(bulk_import.reload.failed?).to eq(true)
- end
- end
-
- context 'when maximum allowed number of import entities in progress' do
- it 'reenqueues itself' do
- bulk_import = create(:bulk_import, :started)
- create(:bulk_import_entity, :created, bulk_import: bulk_import)
- (described_class::DEFAULT_BATCH_SIZE + 1).times { |_| create(:bulk_import_entity, :started, bulk_import: bulk_import) }
-
- expect(described_class).to receive(:perform_in).with(described_class::PERFORM_DELAY, bulk_import.id)
- expect(BulkImports::ExportRequestWorker).not_to receive(:perform_async)
-
- subject.perform(bulk_import.id)
- end
- end
-
- context 'when bulk import is created' do
- it 'marks bulk import as started' do
- bulk_import = create(:bulk_import, :created)
- create(:bulk_import_entity, :created, bulk_import: bulk_import)
-
- subject.perform(bulk_import.id)
-
- expect(bulk_import.reload.started?).to eq(true)
- end
-
- it 'creates all the required pipeline trackers' do
- bulk_import = create(:bulk_import, :created)
- entity_1 = create(:bulk_import_entity, :created, bulk_import: bulk_import)
- entity_2 = create(:bulk_import_entity, :created, bulk_import: bulk_import)
-
- expect { subject.perform(bulk_import.id) }
- .to change { BulkImports::Tracker.count }
- .by(BulkImports::Groups::Stage.new(entity_1).pipelines.size * 2)
-
- expect(entity_1.trackers).not_to be_empty
- expect(entity_2.trackers).not_to be_empty
- end
-
- context 'when there are created entities to process' do
- let_it_be(:bulk_import) { create(:bulk_import, :created) }
-
- before do
- stub_const("#{described_class}::DEFAULT_BATCH_SIZE", 1)
- end
-
- it 'marks a batch of entities as started, enqueues EntityWorker, ExportRequestWorker and reenqueues' do
- create(:bulk_import_entity, :created, bulk_import: bulk_import)
- create(:bulk_import_entity, :created, bulk_import: bulk_import)
-
- expect(described_class).to receive(:perform_in).with(described_class::PERFORM_DELAY, bulk_import.id)
- expect(BulkImports::ExportRequestWorker).to receive(:perform_async).once
-
- subject.perform(bulk_import.id)
-
- expect(bulk_import.entities.map(&:status_name)).to contain_exactly(:created, :started)
- end
-
- context 'when there are project entities to process' do
- it 'enqueues ExportRequestWorker' do
- create(:bulk_import_entity, :created, :project_entity, bulk_import: bulk_import)
+ let_it_be(:bulk_import) { create(:bulk_import) }
+ let_it_be(:job_args) { [bulk_import.id] }
- expect(BulkImports::ExportRequestWorker).to receive(:perform_async).once
-
- subject.perform(bulk_import.id)
- end
- end
- end
-
- context 'when exception occurs' do
- it 'tracks the exception & marks import as failed' do
- bulk_import = create(:bulk_import, :created)
- create(:bulk_import_entity, :created, bulk_import: bulk_import)
-
- allow(BulkImports::ExportRequestWorker).to receive(:perform_async).and_raise(StandardError)
-
- expect(Gitlab::ErrorTracking).to receive(:track_exception).with(kind_of(StandardError), bulk_import_id: bulk_import.id)
-
- subject.perform(bulk_import.id)
-
- expect(bulk_import.reload.failed?).to eq(true)
- end
- end
- end
-
- context 'when importing a group' do
- it 'creates trackers for group entity' do
- bulk_import = create(:bulk_import)
- entity = create(:bulk_import_entity, :group_entity, bulk_import: bulk_import)
-
- subject.perform(bulk_import.id)
-
- expect(entity.trackers.to_a).to include(
- have_attributes(
- stage: 0, status_name: :created, relation: BulkImports::Groups::Pipelines::GroupPipeline.to_s
- ),
- have_attributes(
- stage: 1, status_name: :created, relation: BulkImports::Groups::Pipelines::GroupAttributesPipeline.to_s
- )
- )
- end
- end
-
- context 'when importing a project' do
- it 'creates trackers for project entity' do
- bulk_import = create(:bulk_import)
- entity = create(:bulk_import_entity, :project_entity, bulk_import: bulk_import)
-
- subject.perform(bulk_import.id)
-
- expect(entity.trackers.to_a).to include(
- have_attributes(
- stage: 0, status_name: :created, relation: BulkImports::Projects::Pipelines::ProjectPipeline.to_s
- ),
- have_attributes(
- stage: 1, status_name: :created, relation: BulkImports::Projects::Pipelines::RepositoryPipeline.to_s
- )
- )
- end
- end
-
- context 'when tracker configuration has a minimum version defined' do
- before do
- allow_next_instance_of(BulkImports::Groups::Stage) do |stage|
- allow(stage).to receive(:config).and_return(
- {
- pipeline1: { pipeline: 'PipelineClass1', stage: 0 },
- pipeline2: { pipeline: 'PipelineClass2', stage: 1, minimum_source_version: '14.10.0' },
- pipeline3: { pipeline: 'PipelineClass3', stage: 1, minimum_source_version: '15.0.0' },
- pipeline5: { pipeline: 'PipelineClass4', stage: 1, minimum_source_version: '15.1.0' },
- pipeline6: { pipeline: 'PipelineClass5', stage: 1, minimum_source_version: '16.0.0' }
- }
- )
- end
- end
-
- context 'when the source instance version is older than the tracker mininum version' do
- let_it_be(:bulk_import) { create(:bulk_import, source_version: '15.0.0') }
- let_it_be(:entity) { create(:bulk_import_entity, :group_entity, bulk_import: bulk_import) }
-
- it 'creates trackers as skipped if version requirement does not meet' do
- subject.perform(bulk_import.id)
-
- expect(entity.trackers.collect { |tracker| [tracker.status_name, tracker.relation] }).to contain_exactly(
- [:created, 'PipelineClass1'],
- [:created, 'PipelineClass2'],
- [:created, 'PipelineClass3'],
- [:skipped, 'PipelineClass4'],
- [:skipped, 'PipelineClass5']
- )
- end
-
- it 'logs an info message for the skipped pipelines' do
- expect_next_instance_of(Gitlab::Import::Logger) do |logger|
- expect(logger).to receive(:info).with({
- message: 'Pipeline skipped as source instance version not compatible with pipeline',
- bulk_import_entity_id: entity.id,
- bulk_import_id: entity.bulk_import_id,
- bulk_import_entity_type: entity.source_type,
- source_full_path: entity.source_full_path,
- importer: 'gitlab_migration',
- pipeline_name: 'PipelineClass4',
- minimum_source_version: '15.1.0',
- maximum_source_version: nil,
- source_version: '15.0.0'
- })
-
- expect(logger).to receive(:info).with({
- message: 'Pipeline skipped as source instance version not compatible with pipeline',
- bulk_import_entity_id: entity.id,
- bulk_import_id: entity.bulk_import_id,
- bulk_import_entity_type: entity.source_type,
- source_full_path: entity.source_full_path,
- importer: 'gitlab_migration',
- pipeline_name: 'PipelineClass5',
- minimum_source_version: '16.0.0',
- maximum_source_version: nil,
- source_version: '15.0.0'
- })
- end
-
- subject.perform(bulk_import.id)
- end
+ describe '#perform' do
+ it 'executes the BulkImports::ProcessService' do
+ expect_next_instance_of(BulkImports::ProcessService) do |process_service|
+ expect(process_service).to receive(:execute)
end
- context 'when the source instance version is undefined' do
- it 'creates trackers as created' do
- bulk_import = create(:bulk_import, source_version: nil)
- entity = create(:bulk_import_entity, :group_entity, bulk_import: bulk_import)
-
- subject.perform(bulk_import.id)
-
- expect(entity.trackers.collect { |tracker| [tracker.status_name, tracker.relation] }).to contain_exactly(
- [:created, 'PipelineClass1'],
- [:created, 'PipelineClass2'],
- [:created, 'PipelineClass3'],
- [:created, 'PipelineClass4'],
- [:created, 'PipelineClass5']
- )
- end
- end
+ described_class.new.perform(bulk_import.id)
end
- context 'when tracker configuration has a maximum version defined' do
- before do
- allow_next_instance_of(BulkImports::Groups::Stage) do |stage|
- allow(stage).to receive(:config).and_return(
- {
- pipeline1: { pipeline: 'PipelineClass1', stage: 0 },
- pipeline2: { pipeline: 'PipelineClass2', stage: 1, maximum_source_version: '14.10.0' },
- pipeline3: { pipeline: 'PipelineClass3', stage: 1, maximum_source_version: '15.0.0' },
- pipeline5: { pipeline: 'PipelineClass4', stage: 1, maximum_source_version: '15.1.0' },
- pipeline6: { pipeline: 'PipelineClass5', stage: 1, maximum_source_version: '16.0.0' }
- }
- )
- end
- end
-
- context 'when the source instance version is older than the tracker maximum version' do
- it 'creates trackers as skipped if version requirement does not meet' do
- bulk_import = create(:bulk_import, source_version: '15.0.0')
- entity = create(:bulk_import_entity, :group_entity, bulk_import: bulk_import)
-
- subject.perform(bulk_import.id)
+ context 'when no BulkImport is found' do
+ let(:job_args) { nil }
- expect(entity.trackers.collect { |tracker| [tracker.status_name, tracker.relation] }).to contain_exactly(
- [:created, 'PipelineClass1'],
- [:skipped, 'PipelineClass2'],
- [:created, 'PipelineClass3'],
- [:created, 'PipelineClass4'],
- [:created, 'PipelineClass5']
- )
- end
+ it 'returns without error' do
+ expect { described_class.new.perform(bulk_import.id) }.not_to raise_error
end
- context 'when the source instance version is a patch version' do
- it 'creates trackers with the same status as the non-patch source version' do
- bulk_import_1 = create(:bulk_import, source_version: '15.0.1')
- entity_1 = create(:bulk_import_entity, :group_entity, bulk_import: bulk_import_1)
-
- bulk_import_2 = create(:bulk_import, source_version: '15.0.0')
- entity_2 = create(:bulk_import_entity, :group_entity, bulk_import: bulk_import_2)
-
- described_class.perform_inline(bulk_import_1.id)
- described_class.perform_inline(bulk_import_2.id)
-
- trackers_1 = entity_1.trackers.collect { |tracker| [tracker.status_name, tracker.relation] }
- trackers_2 = entity_2.trackers.collect { |tracker| [tracker.status_name, tracker.relation] }
-
- expect(trackers_1).to eq(trackers_2)
+ it 'does not executes the BulkImports::ProcessService' do
+ expect_any_instance_of(BulkImports::ProcessService) do |process_service|
+ expect(process_service).not_to receive(:execute)
end
end
end