Fix race conditions for AuthorizedProjectsWorker
There were two cases that could be problematic:
1. Because sometimes AuthorizedProjectsWorker would be scheduled in a
transaction it was possible for a job to run/complete before a
COMMIT; resulting in it either producing an error, or producing no
new data.
2. When scheduling jobs the code would not wait until completion. This
could lead to a user creating a project and then immediately trying
to push to it. Usually this will work fine, but given enough load it
might take a few seconds before a user has access.
The first one is problematic, the second one is mostly just annoying
(but annoying enough to warrant a solution).
This commit changes two things to deal with this:
1. Sidekiq scheduling now takes places after a COMMIT, this is ensured
by scheduling using Rails' after_commit hook instead of doing so in
an arbitrary method.
2. When scheduling jobs the calling thread now waits for all jobs to
complete.
Solution 2 requires tracking of job completions. Sidekiq provides a way
to find a job by its ID, but this involves scanning over the entire
queue; something that is very in-efficient for large queues. As such a
more efficient solution is necessary. There are two main Gems that can
do this in a more efficient manner:
* sidekiq-status
* sidekiq_status
No, this is not a joke. Both Gems do a similar thing (but slightly
different), and the only difference in their name is a dash vs an
underscore. Both Gems however provide far more than just checking if a
job has been completed, and both have their problems. sidekiq-status
does not appear to be actively maintained, with the last release being
in 2015. It also has some issues during testing as API calls are not
stubbed in any way. sidekiq_status on the other hand does not appear to
be very popular, and introduces a similar amount of code.
Because of this I opted to write a simple home grown solution. After
all, all we need is storing a job ID somewhere so we can efficiently
look it up; we don't need extra web UIs (as provided by sidekiq-status)
or complex APIs to update progress, etc.
This is where Gitlab::SidekiqStatus comes in handy. This namespace
contains some code used for tracking, removing, and looking up job IDs;
all without having to scan over an entire queue. Data is removed
explicitly, but also expires automatically just in case.
Using this API we can now schedule jobs in a fork-join like manner: we
schedule the jobs in Sidekiq, process them in parallel, then wait for
completion. By using Sidekiq we can leverage all the benefits such as
being able to scale across multiple cores and hosts, retrying failed
jobs, etc.
The one downside is that we need to make sure we can deal with
unexpected increases in job processing timings. To deal with this the
class Gitlab::JobWaiter (used for waiting for jobs to complete) will
only wait a number of seconds (30 by default). Once this timeout is
reached it will simply return.
For GitLab.com almost all AuthorizedProjectWorker jobs complete in
seconds, only very rarely do we spike to job timings of around a minute.
These in turn seem to be the result of external factors (e.g. deploys),
in which case a user is most likely not able to use the system anyway.
In short, this new solution should ensure that jobs are processed
properly and that in almost all cases a user has access to their
resources whenever they need to have access.
2017-01-22 12:22:02 -05:00
|
|
|
require './spec/support/sidekiq'
|
2018-03-29 05:47:54 -04:00
|
|
|
require './spec/support/helpers/test_env'
|
2016-09-09 07:34:28 -04:00
|
|
|
|
|
|
|
class Gitlab::Seeder::CycleAnalytics
|
2016-09-14 10:10:31 -04:00
|
|
|
def initialize(project, perf: false)
|
2016-09-09 07:34:28 -04:00
|
|
|
@project = project
|
2017-11-16 16:26:12 -05:00
|
|
|
@user = User.admins.first
|
2016-09-14 10:10:31 -04:00
|
|
|
@issue_count = perf ? 1000 : 5
|
2016-09-09 07:34:28 -04:00
|
|
|
end
|
|
|
|
|
2016-09-15 10:12:12 -04:00
|
|
|
def seed_metrics!
|
|
|
|
@issue_count.times do |index|
|
|
|
|
# Issue
|
|
|
|
Timecop.travel 5.days.from_now
|
|
|
|
title = "#{FFaker::Product.brand}-#{FFaker::Product.brand}-#{rand(1000)}"
|
|
|
|
issue = Issue.create(project: @project, title: title, author: @user)
|
|
|
|
issue_metrics = issue.metrics
|
|
|
|
|
|
|
|
# Milestones / Labels
|
|
|
|
Timecop.travel 5.days.from_now
|
2019-09-03 06:04:11 -04:00
|
|
|
|
2016-09-15 10:12:12 -04:00
|
|
|
if index.even?
|
|
|
|
issue_metrics.first_associated_with_milestone_at = rand(6..12).hours.from_now
|
|
|
|
else
|
|
|
|
issue_metrics.first_added_to_board_at = rand(6..12).hours.from_now
|
|
|
|
end
|
|
|
|
|
|
|
|
# Commit
|
|
|
|
Timecop.travel 5.days.from_now
|
|
|
|
issue_metrics.first_mentioned_in_commit_at = rand(6..12).hours.from_now
|
|
|
|
|
|
|
|
# MR
|
|
|
|
Timecop.travel 5.days.from_now
|
|
|
|
branch_name = "#{FFaker::Product.brand}-#{FFaker::Product.brand}-#{rand(1000)}"
|
|
|
|
@project.repository.add_branch(@user, branch_name, 'master')
|
|
|
|
merge_request = MergeRequest.create(target_project: @project, source_project: @project, source_branch: branch_name, target_branch: 'master', title: branch_name, author: @user)
|
|
|
|
merge_request_metrics = merge_request.metrics
|
|
|
|
|
|
|
|
# MR closing issues
|
|
|
|
Timecop.travel 5.days.from_now
|
|
|
|
MergeRequestsClosingIssues.create!(issue: issue, merge_request: merge_request)
|
|
|
|
|
|
|
|
# Merge
|
|
|
|
Timecop.travel 5.days.from_now
|
|
|
|
merge_request_metrics.merged_at = rand(6..12).hours.from_now
|
|
|
|
|
|
|
|
# Start build
|
|
|
|
Timecop.travel 5.days.from_now
|
|
|
|
merge_request_metrics.latest_build_started_at = rand(6..12).hours.from_now
|
|
|
|
|
|
|
|
# Finish build
|
|
|
|
Timecop.travel 5.days.from_now
|
|
|
|
merge_request_metrics.latest_build_finished_at = rand(6..12).hours.from_now
|
|
|
|
|
|
|
|
# Deploy to production
|
|
|
|
Timecop.travel 5.days.from_now
|
|
|
|
merge_request_metrics.first_deployed_to_production_at = rand(6..12).hours.from_now
|
|
|
|
|
|
|
|
issue_metrics.save!
|
|
|
|
merge_request_metrics.save!
|
|
|
|
|
|
|
|
print '.'
|
|
|
|
end
|
|
|
|
end
|
|
|
|
|
2016-09-09 07:34:28 -04:00
|
|
|
def seed!
|
2017-11-16 16:26:12 -05:00
|
|
|
Sidekiq::Worker.skipping_transaction_check do
|
|
|
|
Sidekiq::Testing.inline! do
|
|
|
|
issues = create_issues
|
|
|
|
puts '.'
|
|
|
|
|
|
|
|
# Stage 1
|
|
|
|
Timecop.travel 5.days.from_now
|
|
|
|
add_milestones_and_list_labels(issues)
|
|
|
|
print '.'
|
|
|
|
|
|
|
|
# Stage 2
|
|
|
|
Timecop.travel 5.days.from_now
|
|
|
|
branches = mention_in_commits(issues)
|
|
|
|
print '.'
|
|
|
|
|
|
|
|
# Stage 3
|
|
|
|
Timecop.travel 5.days.from_now
|
|
|
|
merge_requests = create_merge_requests_closing_issues(issues, branches)
|
|
|
|
print '.'
|
|
|
|
|
|
|
|
# Stage 4
|
|
|
|
Timecop.travel 5.days.from_now
|
|
|
|
run_builds(merge_requests)
|
|
|
|
print '.'
|
|
|
|
|
|
|
|
# Stage 5
|
|
|
|
Timecop.travel 5.days.from_now
|
|
|
|
merge_merge_requests(merge_requests)
|
|
|
|
print '.'
|
|
|
|
|
|
|
|
# Stage 6 / 7
|
|
|
|
Timecop.travel 5.days.from_now
|
|
|
|
deploy_to_production(merge_requests)
|
|
|
|
print '.'
|
|
|
|
end
|
2016-09-09 07:34:28 -04:00
|
|
|
end
|
|
|
|
|
|
|
|
print '.'
|
|
|
|
end
|
|
|
|
|
|
|
|
private
|
|
|
|
|
2016-09-15 10:12:12 -04:00
|
|
|
def create_issues
|
2016-09-14 10:10:31 -04:00
|
|
|
Array.new(@issue_count) do
|
2016-09-09 07:34:28 -04:00
|
|
|
issue_params = {
|
|
|
|
title: "Cycle Analytics: #{FFaker::Lorem.sentence(6)}",
|
|
|
|
description: FFaker::Lorem.sentence,
|
|
|
|
state: 'opened',
|
2017-11-16 16:26:12 -05:00
|
|
|
assignees: [@project.team.users.sample]
|
2016-09-09 07:34:28 -04:00
|
|
|
}
|
|
|
|
|
|
|
|
Issues::CreateService.new(@project, @project.team.users.sample, issue_params).execute
|
|
|
|
end
|
|
|
|
end
|
|
|
|
|
|
|
|
def add_milestones_and_list_labels(issues)
|
|
|
|
issues.shuffle.map.with_index do |issue, index|
|
|
|
|
Timecop.travel 12.hours.from_now
|
|
|
|
|
|
|
|
if index.even?
|
|
|
|
issue.update(milestone: @project.milestones.sample)
|
|
|
|
else
|
|
|
|
label_name = "#{FFaker::Product.brand}-#{FFaker::Product.brand}-#{rand(1000)}"
|
2017-12-13 19:13:44 -05:00
|
|
|
list_label = FactoryBot.create(:label, title: label_name, project: issue.project)
|
|
|
|
FactoryBot.create(:list, board: FactoryBot.create(:board, project: issue.project), label: list_label)
|
2016-09-09 07:34:28 -04:00
|
|
|
issue.update(labels: [list_label])
|
|
|
|
end
|
|
|
|
|
|
|
|
issue
|
|
|
|
end
|
|
|
|
end
|
|
|
|
|
|
|
|
def mention_in_commits(issues)
|
|
|
|
issues.map do |issue|
|
|
|
|
Timecop.travel 12.hours.from_now
|
|
|
|
|
|
|
|
branch_name = filename = "#{FFaker::Product.brand}-#{FFaker::Product.brand}-#{rand(1000)}"
|
|
|
|
|
|
|
|
issue.project.repository.add_branch(@user, branch_name, 'master')
|
|
|
|
|
2017-11-16 16:26:12 -05:00
|
|
|
commit_sha = issue.project.repository.create_file(@user, filename, "content", message: "Commit for #{issue.to_reference}", branch_name: branch_name)
|
Improve performance of the cycle analytics page.
1. These changes bring down page load time for 100 issues from more than
a minute to about 1.5 seconds.
2. This entire commit is composed of these types of performance
enhancements:
- Cache relevant data in `IssueMetrics` wherever possible.
- Cache relevant data in `MergeRequestMetrics` wherever possible.
- Preload metrics
3. Given these improvements, we now only need to make 4 SQL calls:
- Load all issues
- Load all merge requests
- Load all metrics for the issues
- Load all metrics for the merge requests
4. A list of all the data points that are now being pre-calculated:
a. The first time an issue is mentioned in a commit
- In `GitPushService`, find all issues mentioned by the given commit
using `ReferenceExtractor`. Set the `first_mentioned_in_commit_at`
flag for each of them.
- There seems to be a (pre-existing) bug here - files (and
therefore commits) created using the Web CI don't have
cross-references created, and issues are not closed even when
the commit title is "Fixes #xx".
b. The first time a merge request is deployed to production
When a `Deployment` is created, find all merge requests that
were merged in before the deployment, and set the
`first_deployed_to_production_at` flag for each of them.
c. The start / end time for a merge request pipeline
Hook into the `Pipeline` state machine. When the `status` moves to
`running`, find the merge requests whose tip commit matches the
pipeline, and record the `latest_build_started_at` time for each
of them. When the `status` moves to `success`, record the
`latest_build_finished_at` time.
d. The merge requests that close an issue
- This was a big cause of the performance problems we were having
with Cycle Analytics. We need to use `ReferenceExtractor` to make
this calculation, which is slow when we have to run it on a large
number of merge requests.
- When a merge request is created, updated, or refreshed, find the
issues it closes, and create an instance of
`MergeRequestsClosingIssues`, which acts as a join model between
merge requests and issues.
- If a `MergeRequestsClosingIssues` instance links a merge request
and an issue, that issue closes that merge request.
5. The `Queries` module was changed into a class, so we can cache the
results of `issues` and `merge_requests_closing_issues` across
various cycle analytics stages.
6. The code added in this commit is untested. Tests will be added in the
next commit.
2016-09-15 04:59:36 -04:00
|
|
|
issue.project.repository.commit(commit_sha)
|
|
|
|
|
2019-09-03 06:04:11 -04:00
|
|
|
::Git::BranchPushService.new(
|
2019-07-01 05:19:04 -04:00
|
|
|
issue.project,
|
|
|
|
@user,
|
2019-10-03 11:07:07 -04:00
|
|
|
change: {
|
|
|
|
oldrev: issue.project.repository.commit("master").sha,
|
|
|
|
newrev: commit_sha,
|
|
|
|
ref: 'refs/heads/master'
|
|
|
|
}
|
2019-07-01 05:19:04 -04:00
|
|
|
).execute
|
2016-09-09 07:34:28 -04:00
|
|
|
|
|
|
|
branch_name
|
|
|
|
end
|
|
|
|
end
|
|
|
|
|
|
|
|
def create_merge_requests_closing_issues(issues, branches)
|
|
|
|
issues.zip(branches).map do |issue, branch|
|
|
|
|
Timecop.travel 12.hours.from_now
|
|
|
|
|
|
|
|
opts = {
|
|
|
|
title: 'Cycle Analytics merge_request',
|
|
|
|
description: "Fixes #{issue.to_reference}",
|
|
|
|
source_branch: branch,
|
|
|
|
target_branch: 'master'
|
|
|
|
}
|
|
|
|
|
|
|
|
MergeRequests::CreateService.new(issue.project, @user, opts).execute
|
|
|
|
end
|
|
|
|
end
|
|
|
|
|
|
|
|
def run_builds(merge_requests)
|
|
|
|
merge_requests.each do |merge_request|
|
|
|
|
Timecop.travel 12.hours.from_now
|
|
|
|
|
|
|
|
service = Ci::CreatePipelineService.new(merge_request.project,
|
|
|
|
@user,
|
|
|
|
ref: "refs/heads/#{merge_request.source_branch}")
|
2017-05-24 09:13:51 -04:00
|
|
|
pipeline = service.execute(:push, ignore_skip_ci: true, save_on_errors: false)
|
2016-09-09 07:34:28 -04:00
|
|
|
|
2019-09-03 06:04:11 -04:00
|
|
|
pipeline.builds.each(&:enqueue) # make sure all pipelines in pending state
|
|
|
|
pipeline.builds.each(&:run!)
|
2018-11-04 19:37:40 -05:00
|
|
|
pipeline.update_status
|
2016-09-09 07:34:28 -04:00
|
|
|
end
|
|
|
|
end
|
|
|
|
|
|
|
|
def merge_merge_requests(merge_requests)
|
|
|
|
merge_requests.each do |merge_request|
|
|
|
|
Timecop.travel 12.hours.from_now
|
|
|
|
|
|
|
|
MergeRequests::MergeService.new(merge_request.project, @user).execute(merge_request)
|
|
|
|
end
|
|
|
|
end
|
|
|
|
|
|
|
|
def deploy_to_production(merge_requests)
|
|
|
|
merge_requests.each do |merge_request|
|
2017-11-16 16:26:12 -05:00
|
|
|
next unless merge_request.head_pipeline
|
|
|
|
|
2016-09-09 07:34:28 -04:00
|
|
|
Timecop.travel 12.hours.from_now
|
|
|
|
|
2017-06-01 14:46:34 -04:00
|
|
|
job = merge_request.head_pipeline.builds.where.not(environment: nil).last
|
|
|
|
|
2018-11-04 19:37:40 -05:00
|
|
|
job.success!
|
|
|
|
pipeline.update_status
|
2016-09-09 07:34:28 -04:00
|
|
|
end
|
|
|
|
end
|
|
|
|
end
|
|
|
|
|
|
|
|
Gitlab::Seeder.quiet do
|
2017-04-10 16:18:58 -04:00
|
|
|
flag = 'SEED_CYCLE_ANALYTICS'
|
|
|
|
|
|
|
|
if ENV[flag]
|
2019-11-15 07:06:12 -05:00
|
|
|
Project.not_mass_generated.find_each do |project|
|
2017-11-16 16:26:12 -05:00
|
|
|
# This seed naively assumes that every project has a repository, and every
|
|
|
|
# repository has a `master` branch, which may be the case for a pristine
|
|
|
|
# GDK seed, but is almost never true for a GDK that's actually had
|
|
|
|
# development performed on it.
|
|
|
|
next unless project.repository_exists?
|
|
|
|
next unless project.repository.commit('master')
|
|
|
|
|
Improve performance of the cycle analytics page.
1. These changes bring down page load time for 100 issues from more than
a minute to about 1.5 seconds.
2. This entire commit is composed of these types of performance
enhancements:
- Cache relevant data in `IssueMetrics` wherever possible.
- Cache relevant data in `MergeRequestMetrics` wherever possible.
- Preload metrics
3. Given these improvements, we now only need to make 4 SQL calls:
- Load all issues
- Load all merge requests
- Load all metrics for the issues
- Load all metrics for the merge requests
4. A list of all the data points that are now being pre-calculated:
a. The first time an issue is mentioned in a commit
- In `GitPushService`, find all issues mentioned by the given commit
using `ReferenceExtractor`. Set the `first_mentioned_in_commit_at`
flag for each of them.
- There seems to be a (pre-existing) bug here - files (and
therefore commits) created using the Web CI don't have
cross-references created, and issues are not closed even when
the commit title is "Fixes #xx".
b. The first time a merge request is deployed to production
When a `Deployment` is created, find all merge requests that
were merged in before the deployment, and set the
`first_deployed_to_production_at` flag for each of them.
c. The start / end time for a merge request pipeline
Hook into the `Pipeline` state machine. When the `status` moves to
`running`, find the merge requests whose tip commit matches the
pipeline, and record the `latest_build_started_at` time for each
of them. When the `status` moves to `success`, record the
`latest_build_finished_at` time.
d. The merge requests that close an issue
- This was a big cause of the performance problems we were having
with Cycle Analytics. We need to use `ReferenceExtractor` to make
this calculation, which is slow when we have to run it on a large
number of merge requests.
- When a merge request is created, updated, or refreshed, find the
issues it closes, and create an instance of
`MergeRequestsClosingIssues`, which acts as a join model between
merge requests and issues.
- If a `MergeRequestsClosingIssues` instance links a merge request
and an issue, that issue closes that merge request.
5. The `Queries` module was changed into a class, so we can cache the
results of `issues` and `merge_requests_closing_issues` across
various cycle analytics stages.
6. The code added in this commit is untested. Tests will be added in the
next commit.
2016-09-15 04:59:36 -04:00
|
|
|
seeder = Gitlab::Seeder::CycleAnalytics.new(project)
|
2016-09-14 10:10:31 -04:00
|
|
|
seeder.seed!
|
Improve performance of the cycle analytics page.
1. These changes bring down page load time for 100 issues from more than
a minute to about 1.5 seconds.
2. This entire commit is composed of these types of performance
enhancements:
- Cache relevant data in `IssueMetrics` wherever possible.
- Cache relevant data in `MergeRequestMetrics` wherever possible.
- Preload metrics
3. Given these improvements, we now only need to make 4 SQL calls:
- Load all issues
- Load all merge requests
- Load all metrics for the issues
- Load all metrics for the merge requests
4. A list of all the data points that are now being pre-calculated:
a. The first time an issue is mentioned in a commit
- In `GitPushService`, find all issues mentioned by the given commit
using `ReferenceExtractor`. Set the `first_mentioned_in_commit_at`
flag for each of them.
- There seems to be a (pre-existing) bug here - files (and
therefore commits) created using the Web CI don't have
cross-references created, and issues are not closed even when
the commit title is "Fixes #xx".
b. The first time a merge request is deployed to production
When a `Deployment` is created, find all merge requests that
were merged in before the deployment, and set the
`first_deployed_to_production_at` flag for each of them.
c. The start / end time for a merge request pipeline
Hook into the `Pipeline` state machine. When the `status` moves to
`running`, find the merge requests whose tip commit matches the
pipeline, and record the `latest_build_started_at` time for each
of them. When the `status` moves to `success`, record the
`latest_build_finished_at` time.
d. The merge requests that close an issue
- This was a big cause of the performance problems we were having
with Cycle Analytics. We need to use `ReferenceExtractor` to make
this calculation, which is slow when we have to run it on a large
number of merge requests.
- When a merge request is created, updated, or refreshed, find the
issues it closes, and create an instance of
`MergeRequestsClosingIssues`, which acts as a join model between
merge requests and issues.
- If a `MergeRequestsClosingIssues` instance links a merge request
and an issue, that issue closes that merge request.
5. The `Queries` module was changed into a class, so we can cache the
results of `issues` and `merge_requests_closing_issues` across
various cycle analytics stages.
6. The code added in this commit is untested. Tests will be added in the
next commit.
2016-09-15 04:59:36 -04:00
|
|
|
end
|
2016-09-14 10:10:31 -04:00
|
|
|
elsif ENV['CYCLE_ANALYTICS_PERF_TEST']
|
Improve performance of the cycle analytics page.
1. These changes bring down page load time for 100 issues from more than
a minute to about 1.5 seconds.
2. This entire commit is composed of these types of performance
enhancements:
- Cache relevant data in `IssueMetrics` wherever possible.
- Cache relevant data in `MergeRequestMetrics` wherever possible.
- Preload metrics
3. Given these improvements, we now only need to make 4 SQL calls:
- Load all issues
- Load all merge requests
- Load all metrics for the issues
- Load all metrics for the merge requests
4. A list of all the data points that are now being pre-calculated:
a. The first time an issue is mentioned in a commit
- In `GitPushService`, find all issues mentioned by the given commit
using `ReferenceExtractor`. Set the `first_mentioned_in_commit_at`
flag for each of them.
- There seems to be a (pre-existing) bug here - files (and
therefore commits) created using the Web CI don't have
cross-references created, and issues are not closed even when
the commit title is "Fixes #xx".
b. The first time a merge request is deployed to production
When a `Deployment` is created, find all merge requests that
were merged in before the deployment, and set the
`first_deployed_to_production_at` flag for each of them.
c. The start / end time for a merge request pipeline
Hook into the `Pipeline` state machine. When the `status` moves to
`running`, find the merge requests whose tip commit matches the
pipeline, and record the `latest_build_started_at` time for each
of them. When the `status` moves to `success`, record the
`latest_build_finished_at` time.
d. The merge requests that close an issue
- This was a big cause of the performance problems we were having
with Cycle Analytics. We need to use `ReferenceExtractor` to make
this calculation, which is slow when we have to run it on a large
number of merge requests.
- When a merge request is created, updated, or refreshed, find the
issues it closes, and create an instance of
`MergeRequestsClosingIssues`, which acts as a join model between
merge requests and issues.
- If a `MergeRequestsClosingIssues` instance links a merge request
and an issue, that issue closes that merge request.
5. The `Queries` module was changed into a class, so we can cache the
results of `issues` and `merge_requests_closing_issues` across
various cycle analytics stages.
6. The code added in this commit is untested. Tests will be added in the
next commit.
2016-09-15 04:59:36 -04:00
|
|
|
seeder = Gitlab::Seeder::CycleAnalytics.new(Project.order(:id).first, perf: true)
|
2016-09-09 07:34:28 -04:00
|
|
|
seeder.seed!
|
2016-09-15 10:12:12 -04:00
|
|
|
elsif ENV['CYCLE_ANALYTICS_POPULATE_METRICS_DIRECTLY']
|
|
|
|
seeder = Gitlab::Seeder::CycleAnalytics.new(Project.order(:id).first, perf: true)
|
|
|
|
seeder.seed_metrics!
|
2016-09-14 10:10:31 -04:00
|
|
|
else
|
2017-04-10 16:18:58 -04:00
|
|
|
puts "Skipped. Use the `#{flag}` environment variable to enable."
|
2016-09-09 07:34:28 -04:00
|
|
|
end
|
|
|
|
end
|