module Ci class Pipeline < ActiveRecord::Base extend Ci::Model include HasStatus include Importable include AfterCommitQueue self.table_name = 'ci_commits' belongs_to :project, foreign_key: :gl_project_id belongs_to :user has_many :statuses, class_name: 'CommitStatus', foreign_key: :commit_id has_many :builds, foreign_key: :commit_id has_many :trigger_requests, dependent: :destroy, foreign_key: :commit_id validates_presence_of :sha, unless: :importing? validates_presence_of :ref, unless: :importing? validates_presence_of :status, unless: :importing? validate :valid_commit_sha, unless: :importing? after_create :keep_around_commits, unless: :importing? state_machine :status, initial: :created do event :enqueue do transition created: :pending transition [:success, :failed, :canceled, :skipped] => :running end event :run do transition any - [:running] => :running end event :skip do transition any - [:skipped] => :skipped end event :drop do transition any - [:failed] => :failed end event :succeed do transition any - [:success] => :success end event :cancel do transition any - [:canceled] => :canceled end # IMPORTANT # Do not add any operations to this state_machine # Create a separate worker for each new operation before_transition [:created, :pending] => :running do |pipeline| pipeline.started_at = Time.now end before_transition any => [:success, :failed, :canceled] do |pipeline| pipeline.finished_at = Time.now pipeline.update_duration end after_transition [:created, :pending] => :running do |pipeline| pipeline.run_after_commit { PipelineMetricsWorker.perform_async(id) } end after_transition any => [:success] do |pipeline| pipeline.run_after_commit { PipelineMetricsWorker.perform_async(id) } end after_transition [:created, :pending, :running] => :success do |pipeline| pipeline.run_after_commit { PipelineSuccessWorker.perform_async(id) } end after_transition do |pipeline, transition| next if transition.loopback? pipeline.run_after_commit do PipelineHooksWorker.perform_async(id) end end after_transition any => [:success, :failed] do |pipeline| pipeline.run_after_commit do PipelineNotificationWorker.perform_async(pipeline.id) end end end # ref can't be HEAD or SHA, can only be branch/tag name def self.latest_successful_for(ref) where(ref: ref).order(id: :desc).success.first end def self.truncate_sha(sha) sha[0...8] end def self.total_duration where.not(duration: nil).sum(:duration) end def stages_count statuses.select(:stage).distinct.count end def stages_name statuses.order(:stage_idx).distinct. pluck(:stage, :stage_idx).map(&:first) end def stages status_sql = statuses.latest.where('stage=sg.stage').status_sql stages_query = statuses.group('stage').select(:stage) .order('max(stage_idx)') stages_with_statuses = CommitStatus.from(stages_query, :sg). pluck('sg.stage', status_sql) stages_with_statuses.map do |stage| Ci::Stage.new(self, name: stage.first, status: stage.last) end end def artifacts builds.latest.with_artifacts_not_expired end def project_id project.id end # For now the only user who participates is the user who triggered def participants(_current_user = nil) Array(user) end def valid_commit_sha if self.sha == Gitlab::Git::BLANK_SHA self.errors.add(:sha, " cant be 00000000 (branch removal)") end end def git_author_name commit.try(:author_name) end def git_author_email commit.try(:author_email) end def git_commit_message commit.try(:message) end def git_commit_title commit.try(:title) end def short_sha Ci::Pipeline.truncate_sha(sha) end def commit @commit ||= project.commit(sha) rescue nil end def branch? !tag? end def manual_actions builds.latest.manual_actions end def retryable? builds.latest.failed_or_canceled.any?(&:retryable?) end def cancelable? statuses.cancelable.any? end def cancel_running Gitlab::OptimisticLocking.retry_lock( statuses.cancelable) do |cancelable| cancelable.each(&:cancel) end end def retry_failed(user) Gitlab::OptimisticLocking.retry_lock( builds.latest.failed_or_canceled) do |failed_or_canceled| failed_or_canceled.select(&:retryable?).each do |build| Ci::Build.retry(build, user) end end end def mark_as_processable_after_stage(stage_idx) builds.skipped.where('stage_idx > ?', stage_idx).find_each(&:process) end def latest? return false unless ref commit = project.commit(ref) return false unless commit commit.sha == sha end def triggered? trigger_requests.any? end def retried @retried ||= (statuses.order(id: :desc) - statuses.latest) end def coverage coverage_array = statuses.latest.map(&:coverage).compact if coverage_array.size >= 1 '%.2f' % (coverage_array.reduce(:+) / coverage_array.size) end end def config_builds_attributes return [] unless config_processor config_processor. builds_for_ref(ref, tag?, trigger_requests.first). sort_by { |build| build[:stage_idx] } end def has_warnings? builds.latest.failed_but_allowed.any? end def config_processor return nil unless ci_yaml_file return @config_processor if defined?(@config_processor) @config_processor ||= begin Ci::GitlabCiYamlProcessor.new(ci_yaml_file, project.path_with_namespace) rescue Ci::GitlabCiYamlProcessor::ValidationError, Psych::SyntaxError => e self.yaml_errors = e.message nil rescue self.yaml_errors = 'Undefined error' nil end end def ci_yaml_file return @ci_yaml_file if defined?(@ci_yaml_file) @ci_yaml_file ||= begin blob = project.repository.blob_at(sha, '.gitlab-ci.yml') blob.load_all_data!(project.repository) blob.data rescue nil end end def environments builds.where.not(environment: nil).success.pluck(:environment).uniq end # Manually set the notes for a Ci::Pipeline # There is no ActiveRecord relation between Ci::Pipeline and notes # as they are related to a commit sha. This method helps importing # them using the +Gitlab::ImportExport::RelationFactory+ class. def notes=(notes) notes.each do |note| note[:id] = nil note[:commit_id] = sha note[:noteable_id] = self['id'] note.save! end end def notes Note.for_commit_id(sha) end def process! Ci::ProcessPipelineService.new(project, user).execute(self) end def update_status Gitlab::OptimisticLocking.retry_lock(self) do case latest_builds_status when 'pending' then enqueue when 'running' then run when 'success' then succeed when 'failed' then drop when 'canceled' then cancel when 'skipped' then skip end end end def predefined_variables [ { key: 'CI_PIPELINE_ID', value: id.to_s, public: true } ] end def queued_duration return unless started_at seconds = (started_at - created_at).to_i seconds unless seconds.zero? end def update_duration return unless started_at self.duration = Gitlab::Ci::PipelineDuration.from_pipeline(self) end def execute_hooks data = pipeline_data project.execute_hooks(data, :pipeline_hooks) project.execute_services(data, :pipeline_hooks) end # Merge requests for which the current pipeline is running against # the merge request's latest commit. def merge_requests @merge_requests ||= project.merge_requests .where(source_branch: self.ref) .select { |merge_request| merge_request.head_pipeline.try(:id) == self.id } end def detailed_status Gitlab::Ci::Status::Pipeline::Factory.new(self).fabricate! end private def pipeline_data Gitlab::DataBuilder::Pipeline.build(self) end def latest_builds_status return 'failed' unless yaml_errors.blank? statuses.latest.status || 'skipped' end def keep_around_commits return unless project project.repository.keep_around(self.sha) project.repository.keep_around(self.before_sha) end end end