Commit 3e645f39 authored by Andrejs Cunskis's avatar Andrejs Cunskis

Improve logging and diff json

Fixup comment
parent 872a0e51
...@@ -3,11 +3,16 @@ ...@@ -3,11 +3,16 @@
# rubocop:disable Rails/Pluck # rubocop:disable Rails/Pluck
module QA module QA
# Only executes in custom job/pipeline # Only executes in custom job/pipeline
# https://gitlab.com/gitlab-org/manage/import/import-github-performance
#
RSpec.describe 'Manage', :github, :requires_admin, only: { job: 'large-github-import' } do RSpec.describe 'Manage', :github, :requires_admin, only: { job: 'large-github-import' } do
describe 'Project import' do describe 'Project import' do
let(:logger) { Runtime::Logger.logger } let(:logger) { Runtime::Logger.logger }
let(:differ) { RSpec::Support::Differ.new(color: true) } let(:differ) { RSpec::Support::Differ.new(color: true) }
let(:created_by_pattern) { /\*Created by: \S+\*\n\n/ }
let(:suggestion_pattern) { /suggestion:-\d+\+\d+/ }
let(:api_client) { Runtime::API::Client.as_admin } let(:api_client) { Runtime::API::Client.as_admin }
let(:user) do let(:user) do
...@@ -19,46 +24,57 @@ module QA ...@@ -19,46 +24,57 @@ module QA
let(:github_repo) { ENV['QA_LARGE_GH_IMPORT_REPO'] || 'rspec/rspec-core' } let(:github_repo) { ENV['QA_LARGE_GH_IMPORT_REPO'] || 'rspec/rspec-core' }
let(:import_max_duration) { ENV['QA_LARGE_GH_IMPORT_DURATION'] ? ENV['QA_LARGE_GH_IMPORT_DURATION'].to_i : 7200 } let(:import_max_duration) { ENV['QA_LARGE_GH_IMPORT_DURATION'] ? ENV['QA_LARGE_GH_IMPORT_DURATION'].to_i : 7200 }
let(:github_client) do let(:github_client) do
Octokit.middleware = Faraday::RackBuilder.new do |builder|
builder.response(:logger, logger, headers: false, bodies: false)
end
Octokit::Client.new( Octokit::Client.new(
access_token: ENV['QA_LARGE_GH_IMPORT_GH_TOKEN'] || Runtime::Env.github_access_token, access_token: ENV['QA_LARGE_GH_IMPORT_GH_TOKEN'] || Runtime::Env.github_access_token,
auto_paginate: true auto_paginate: true
) )
end end
let(:gh_branches) { github_client.branches(github_repo).map(&:name) }
let(:gh_commits) { github_client.commits(github_repo).map(&:sha) }
let(:gh_repo) { github_client.repository(github_repo) } let(:gh_repo) { github_client.repository(github_repo) }
let(:gh_branches) do
logger.debug("= Fetching branches =")
github_client.branches(github_repo).map(&:name)
end
let(:gh_commits) do
logger.debug("= Fetching commits =")
github_client.commits(github_repo).map(&:sha)
end
let(:gh_labels) do let(:gh_labels) do
logger.debug("= Fetching labels =")
github_client.labels(github_repo).map { |label| { name: label.name, color: "##{label.color}" } } github_client.labels(github_repo).map { |label| { name: label.name, color: "##{label.color}" } }
end end
let(:gh_milestones) do let(:gh_milestones) do
logger.debug("= Fetching milestones =")
github_client github_client
.list_milestones(github_repo, state: 'all') .list_milestones(github_repo, state: 'all')
.map { |ms| { title: ms.title, description: ms.description } } .map { |ms| { title: ms.title, description: ms.description } }
end end
let(:gh_all_issues) do let(:gh_all_issues) do
logger.debug("= Fetching issues and prs =")
github_client.list_issues(github_repo, state: 'all') github_client.list_issues(github_repo, state: 'all')
end end
let(:gh_prs) do let(:gh_prs) do
gh_all_issues.select(&:pull_request).each_with_object({}) do |pr, hash| gh_all_issues.select(&:pull_request).each_with_object({}) do |pr, hash|
hash[pr.title] = { hash[pr.number] = {
url: pr.html_url,
title: pr.title,
body: pr.body || '', body: pr.body || '',
comments: [*gh_pr_comments[pr.html_url], *gh_issue_comments[pr.html_url]].compact.sort comments: [*gh_pr_comments[pr.html_url], *gh_issue_comments[pr.html_url]].compact
} }
end end
end end
let(:gh_issues) do let(:gh_issues) do
gh_all_issues.reject(&:pull_request).each_with_object({}) do |issue, hash| gh_all_issues.reject(&:pull_request).each_with_object({}) do |issue, hash|
hash[issue.title] = { hash[issue.number] = {
url: issue.html_url,
title: issue.title,
body: issue.body || '', body: issue.body || '',
comments: gh_issue_comments[issue.html_url] comments: gh_issue_comments[issue.html_url]
} }
...@@ -66,12 +82,14 @@ module QA ...@@ -66,12 +82,14 @@ module QA
end end
let(:gh_issue_comments) do let(:gh_issue_comments) do
logger.debug("= Fetching issue comments =")
github_client.issues_comments(github_repo).each_with_object(Hash.new { |h, k| h[k] = [] }) do |c, hash| github_client.issues_comments(github_repo).each_with_object(Hash.new { |h, k| h[k] = [] }) do |c, hash|
hash[c.html_url.gsub(/\#\S+/, "")] << c.body # use base html url as key hash[c.html_url.gsub(/\#\S+/, "")] << c.body # use base html url as key
end end
end end
let(:gh_pr_comments) do let(:gh_pr_comments) do
logger.debug("= Fetching pr comments =")
github_client.pull_requests_comments(github_repo).each_with_object(Hash.new { |h, k| h[k] = [] }) do |c, hash| github_client.pull_requests_comments(github_repo).each_with_object(Hash.new { |h, k| h[k] = [] }) do |c, hash|
hash[c.html_url.gsub(/\#\S+/, "")] << c.body # use base html url as key hash[c.html_url.gsub(/\#\S+/, "")] << c.body # use base html url as key
end end
...@@ -97,6 +115,7 @@ module QA ...@@ -97,6 +115,7 @@ module QA
"data", "data",
{ {
import_time: @import_time, import_time: @import_time,
reported_stats: @stats,
github: { github: {
project_name: github_repo, project_name: github_repo,
branches: gh_branches.length, branches: gh_branches.length,
...@@ -104,9 +123,9 @@ module QA ...@@ -104,9 +123,9 @@ module QA
labels: gh_labels.length, labels: gh_labels.length,
milestones: gh_milestones.length, milestones: gh_milestones.length,
prs: gh_prs.length, prs: gh_prs.length,
pr_comments: gh_prs.sum { |_k, v| v.length }, pr_comments: gh_prs.sum { |_k, v| v[:comments].length },
issues: gh_issues.length, issues: gh_issues.length,
issue_comments: gh_issues.sum { |_k, v| v.length } issue_comments: gh_issues.sum { |_k, v| v[:comments].length }
}, },
gitlab: { gitlab: {
project_name: imported_project.path_with_namespace, project_name: imported_project.path_with_namespace,
...@@ -115,15 +134,15 @@ module QA ...@@ -115,15 +134,15 @@ module QA
labels: gl_labels.length, labels: gl_labels.length,
milestones: gl_milestones.length, milestones: gl_milestones.length,
mrs: mrs.length, mrs: mrs.length,
mr_comments: mrs.sum { |_k, v| v.length }, mr_comments: mrs.sum { |_k, v| v[:comments].length },
issues: gl_issues.length, issues: gl_issues.length,
issue_comments: gl_issues.sum { |_k, v| v.length } issue_comments: gl_issues.sum { |_k, v| v[:comments].length }
}, },
not_imported: { not_imported: {
mrs: @mr_diff, mrs: @mr_diff,
issues: @issue_diff issues: @issue_diff
} }
}.to_json }
) )
end end
...@@ -133,19 +152,25 @@ module QA ...@@ -133,19 +152,25 @@ module QA
) do ) do
start = Time.now start = Time.now
# import the project and log path # import the project and log gitlab path
Runtime::Logger.info("Importing project '#{imported_project.reload!.full_path}'") Runtime::Logger.info("== Importing project '#{github_repo}' in to '#{imported_project.reload!.full_path}' ==")
# fetch all objects right after import has started # fetch all objects right after import has started
fetch_github_objects fetch_github_objects
import_status = lambda do import_status = lambda do
imported_project.project_import_status[:import_status].tap do |status| imported_project.project_import_status.yield_self do |status|
@stats = status.dig(:stats, :imported)
# fail fast if import explicitly failed # fail fast if import explicitly failed
raise "Import of '#{imported_project.name}' failed!" if status == 'failed' raise "Import of '#{imported_project.name}' failed!" if status[:import_status] == 'failed'
status[:import_status]
end end
end end
logger.info("== Waiting for import to be finished ==")
expect(import_status).to eventually_eq('finished').within(max_duration: import_max_duration, sleep_interval: 30) expect(import_status).to eventually_eq('finished').within(max_duration: import_max_duration, sleep_interval: 30)
@import_time = Time.now - start @import_time = Time.now - start
aggregate_failures do aggregate_failures do
...@@ -161,22 +186,22 @@ module QA ...@@ -161,22 +186,22 @@ module QA
# #
# @return [void] # @return [void]
def fetch_github_objects def fetch_github_objects
logger.debug("== Fetching objects for github repo: '#{github_repo}' ==") logger.info("== Fetching github repo objects ==")
gh_repo gh_repo
gh_branches gh_branches
gh_commits gh_commits
gh_prs
gh_issues
gh_labels gh_labels
gh_milestones gh_milestones
gh_prs
gh_issues
end end
# Verify repository imported correctly # Verify repository imported correctly
# #
# @return [void] # @return [void]
def verify_repository_import def verify_repository_import
logger.debug("== Verifying repository import ==") logger.info("== Verifying repository import ==")
expect(imported_project.description).to eq(gh_repo.description) expect(imported_project.description).to eq(gh_repo.description)
# check via include, importer creates more branches # check via include, importer creates more branches
# https://gitlab.com/gitlab-org/gitlab/-/issues/332711 # https://gitlab.com/gitlab-org/gitlab/-/issues/332711
...@@ -184,42 +209,42 @@ module QA ...@@ -184,42 +209,42 @@ module QA
expect(gl_commits).to match_array(gh_commits) expect(gl_commits).to match_array(gh_commits)
end end
# Verify imported merge requests and mr issues # Verify imported labels
# #
# @return [void] # @return [void]
def verify_merge_requests_import def verify_labels_import
logger.debug("== Verifying merge request import ==") logger.info("== Verifying label import ==")
@mr_diff = verify_mrs_or_issues('mr') # check via include, additional labels can be inherited from parent group
expect(gl_labels).to include(*gh_labels)
end end
# Verify imported issues and issue comments # Verify milestones import
# #
# @return [void] # @return [void]
def verify_issues_import def verify_milestones_import
logger.debug("== Verifying issue import ==") logger.info("== Verifying milestones import ==")
@issue_diff = verify_mrs_or_issues('issue') expect(gl_milestones).to match_array(gh_milestones)
end end
# Verify imported labels # Verify imported merge requests and mr issues
# #
# @return [void] # @return [void]
def verify_labels_import def verify_merge_requests_import
logger.debug("== Verifying label import ==") logger.info("== Verifying merge request import ==")
# check via include, additional labels can be inherited from parent group @mr_diff = verify_mrs_or_issues('mr')
expect(gl_labels).to include(*gh_labels)
end end
# Verify milestones import # Verify imported issues and issue comments
# #
# @return [void] # @return [void]
def verify_milestones_import def verify_issues_import
logger.debug("== Verifying milestones import ==") logger.info("== Verifying issue import ==")
expect(gl_milestones).to match_array(gh_milestones) @issue_diff = verify_mrs_or_issues('issue')
end end
private private
# Verify imported mrs or issues and return diff # Verify imported mrs or issues and return missing items
# #
# @param [String] type verification object, 'mrs' or 'issues' # @param [String] type verification object, 'mrs' or 'issues'
# @return [Hash] # @return [Hash]
...@@ -231,11 +256,10 @@ module QA ...@@ -231,11 +256,10 @@ module QA
count_msg = "Expected to contain same amount of #{type}s. Gitlab: #{expected.length}, Github: #{actual.length}" count_msg = "Expected to contain same amount of #{type}s. Gitlab: #{expected.length}, Github: #{actual.length}"
expect(expected.length).to eq(actual.length), count_msg expect(expected.length).to eq(actual.length), count_msg
logger.debug("= Comparing #{type}s =")
missing_comments = verify_comments(type, actual, expected) missing_comments = verify_comments(type, actual, expected)
{ {
"#{type}s": actual.keys - expected.keys, "#{type}s": (actual.keys - expected.keys).map { |it| actual[it].slice(:title, :url) },
"#{type}_comments": missing_comments "#{type}_comments": missing_comments
} }
end end
...@@ -247,9 +271,10 @@ module QA ...@@ -247,9 +271,10 @@ module QA
# @param [Hash] expected # @param [Hash] expected
# @return [Hash] # @return [Hash]
def verify_comments(type, actual, expected) def verify_comments(type, actual, expected)
actual.each_with_object({}) do |(title, actual_item), missing_comments| actual.each_with_object([]) do |(key, actual_item), missing_comments|
expected_item = expected[key]
title = actual_item[:title]
msg = "expected #{type} with title '#{title}' to have" msg = "expected #{type} with title '#{title}' to have"
expected_item = expected[title]
# Print title in the error message to see which object is missing # Print title in the error message to see which object is missing
# #
...@@ -261,9 +286,9 @@ module QA ...@@ -261,9 +286,9 @@ module QA
expected_body = expected_item[:body] expected_body = expected_item[:body]
actual_body = actual_item[:body] actual_body = actual_item[:body]
body_msg = <<~MSG body_msg = <<~MSG
#{msg} same description. diff:\n#{differ.diff(expected_item[:body], actual_item[:body])} #{msg} same description. diff:\n#{differ.diff(expected_body, actual_body)}
MSG MSG
expect(expected_body).to include(actual_body), body_msg expect(expected_body).to eq(actual_body), body_msg
# Print amount difference first # Print amount difference first
# #
...@@ -278,7 +303,14 @@ module QA ...@@ -278,7 +303,14 @@ module QA
# Save missing comments # Save missing comments
# #
comment_diff = actual_comments - expected_comments comment_diff = actual_comments - expected_comments
missing_comments[title] = comment_diff unless comment_diff.empty? next if comment_diff.empty?
missing_comments << {
title: title,
github_url: actual_item[:url],
gitlab_url: expected_item[:url],
missing_comments: comment_diff
}
end end
end end
...@@ -329,20 +361,25 @@ module QA ...@@ -329,20 +361,25 @@ module QA
@mrs ||= begin @mrs ||= begin
logger.debug("= Fetching merge requests =") logger.debug("= Fetching merge requests =")
imported_mrs = imported_project.merge_requests(auto_paginate: true, attempts: 2) imported_mrs = imported_project.merge_requests(auto_paginate: true, attempts: 2)
logger.debug("= Transforming merge request objects for comparison =")
imported_mrs.each_with_object({}) do |mr, hash| logger.debug("= Fetching merge request comments =")
imported_mrs.each_with_object({}) do |mr, mrs_with_comments|
resource = Resource::MergeRequest.init do |resource| resource = Resource::MergeRequest.init do |resource|
resource.project = imported_project resource.project = imported_project
resource.iid = mr[:iid] resource.iid = mr[:iid]
resource.api_client = api_client resource.api_client = api_client
end end
hash[mr[:title]] = { logger.debug("Fetching comments for mr '#{mr[:title]}'")
body: mr[:description], mrs_with_comments[mr[:iid]] = {
comments: resource.comments(auto_paginate: true, attempts: 2) url: mr[:web_url],
title: mr[:title],
body: sanitize_description(mr[:description]) || '',
comments: resource
.comments(auto_paginate: true, attempts: 2)
# remove system notes # remove system notes
.reject { |c| c[:system] || c[:body].match?(/^(\*\*Review:\*\*)|(\*Merged by:).*/) } .reject { |c| c[:system] || c[:body].match?(/^(\*\*Review:\*\*)|(\*Merged by:).*/) }
.map { |c| sanitize(c[:body]) } .map { |c| sanitize_comment(c[:body]) }
} }
end end
end end
...@@ -355,37 +392,51 @@ module QA ...@@ -355,37 +392,51 @@ module QA
@gl_issues ||= begin @gl_issues ||= begin
logger.debug("= Fetching issues =") logger.debug("= Fetching issues =")
imported_issues = imported_project.issues(auto_paginate: true, attempts: 2) imported_issues = imported_project.issues(auto_paginate: true, attempts: 2)
logger.debug("= Transforming issue objects for comparison =")
imported_issues.each_with_object({}) do |issue, hash| logger.debug("= Fetching issue comments =")
imported_issues.each_with_object({}) do |issue, issues_with_comments|
resource = Resource::Issue.init do |issue_resource| resource = Resource::Issue.init do |issue_resource|
issue_resource.project = imported_project issue_resource.project = imported_project
issue_resource.iid = issue[:iid] issue_resource.iid = issue[:iid]
issue_resource.api_client = api_client issue_resource.api_client = api_client
end end
hash[issue[:title]] = { logger.debug("Fetching comments for issue '#{issue[:title]}'")
body: issue[:description], issues_with_comments[issue[:iid]] = {
comments: resource.comments(auto_paginate: true, attempts: 2).map { |c| sanitize(c[:body]) } url: issue[:web_url],
title: issue[:title],
body: sanitize_description(issue[:description]) || '',
comments: resource
.comments(auto_paginate: true, attempts: 2)
.map { |c| sanitize_comment(c[:body]) }
} }
end end
end end
end end
# Remove added prefixes and legacy diff format # Remove added prefixes and legacy diff format from comments
#
# @param [String] body
# @return [String]
def sanitize_comment(body)
body.gsub(created_by_pattern, "").gsub(suggestion_pattern, "suggestion\r")
end
# Remove created by prefix from descripion
# #
# @param [String] body # @param [String] body
# @return [String] # @return [String]
def sanitize(body) def sanitize_description(body)
body.gsub(/\*Created by: \S+\*\n\n/, "").gsub(/suggestion:-\d+\+\d+/, "suggestion\r") body&.gsub(created_by_pattern, "")
end end
# Save json as file # Save json as file
# #
# @param [String] name # @param [String] name
# @param [String] json # @param [Hash] json
# @return [void] # @return [void]
def save_json(name, json) def save_json(name, json)
File.open("tmp/#{name}.json", "w") { |file| file.write(json) } File.open("tmp/#{name}.json", "w") { |file| file.write(JSON.pretty_generate(json)) }
end end
end end
end end
......
Markdown is supported
0%
or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment