mirror of
https://gh.wpcy.net/https://github.com/discourse/discourse.git
synced 2026-05-03 21:32:46 +08:00
173 lines
5.7 KiB
Ruby
173 lines
5.7 KiB
Ruby
# frozen_string_literal: true
|
|
|
|
require_relative "../mixins/github_body"
|
|
require_relative "../mixins/github_auth_header"
|
|
|
|
module Onebox
|
|
module Engine
|
|
class GithubPullRequestOnebox
|
|
include Engine
|
|
include LayoutSupport
|
|
include JSON
|
|
include Onebox::Mixins::GithubBody
|
|
include Onebox::Mixins::GithubAuthHeader
|
|
|
|
matches_domain("github.com", "www.github.com")
|
|
always_https
|
|
|
|
def self.matches_path(path)
|
|
path.match?(%r{.*/pull})
|
|
end
|
|
|
|
def url
|
|
"https://api.github.com/repos/#{match[:org]}/#{match[:repository]}/pulls/#{match[:number]}"
|
|
end
|
|
|
|
private
|
|
|
|
def match
|
|
@match ||=
|
|
@url.match(%r{github\.com/(?<org>[^/]+)/(?<repository>[^/]+)/pull/(?<number>[^/]+)})
|
|
end
|
|
|
|
def data
|
|
result = raw(github_auth_header(match[:org])).clone
|
|
result["link"] = link
|
|
|
|
status_data = fetch_pr_status(result)
|
|
result["pr_status"] = status_data&.dig(:status)
|
|
result["pr_status_title"] = pr_status_title(result["pr_status"])
|
|
|
|
status_timestamp = status_data&.dig(:timestamp) || result["created_at"]
|
|
status_date = Time.parse(status_timestamp)
|
|
result["status_date"] = status_date.strftime("%I:%M%p - %d %b %y %Z")
|
|
result["status_date_date"] = status_date.strftime("%F")
|
|
result["status_date_time"] = status_date.strftime("%T")
|
|
|
|
ulink = URI(link)
|
|
_, org, repo = ulink.path.split("/")
|
|
result["domain"] = "#{ulink.host}/#{org}/#{repo}"
|
|
|
|
result["body"], result["excerpt"] = compute_body(result["body"])
|
|
|
|
if result["commit"] = load_commit(link)
|
|
result["body"], result["excerpt"] =
|
|
compute_body(result["commit"]["commit"]["message"].lines[1..].join)
|
|
elsif result["comment"] = load_comment(link)
|
|
result["body"], result["excerpt"] = compute_body(result["comment"]["body"])
|
|
elsif result["discussion"] = load_review(link)
|
|
result["body"], result["excerpt"] = compute_body(result["discussion"]["body"])
|
|
else
|
|
result["pr"] = true
|
|
end
|
|
|
|
result["number"] = match[:number]
|
|
result["i18n"] = i18n
|
|
result["i18n"]["status_date_label"] = status_date_label(result["pr_status"])
|
|
result["i18n"]["pr_summary"] = I18n.t(
|
|
"onebox.github.pr_summary",
|
|
{
|
|
commits: result["commits"],
|
|
changed_files: result["changed_files"],
|
|
additions: result["additions"],
|
|
deletions: result["deletions"],
|
|
},
|
|
)
|
|
result["is_private"] = result.dig("base", "repo", "private")
|
|
|
|
result["base"]["label"].sub!(/\A#{org}:/, "")
|
|
result["head"]["label"].sub!(/\A#{org}:/, "")
|
|
|
|
result
|
|
end
|
|
|
|
def i18n
|
|
{
|
|
opened: I18n.t("onebox.github.opened"),
|
|
commit_by: I18n.t("onebox.github.commit_by"),
|
|
comment_by: I18n.t("onebox.github.comment_by"),
|
|
review_by: I18n.t("onebox.github.review_by"),
|
|
}
|
|
end
|
|
|
|
def status_date_label(status)
|
|
key = status.presence || "open"
|
|
I18n.t("onebox.github.status_date.#{key}")
|
|
end
|
|
|
|
def pr_status_title(status)
|
|
key = status.presence || "default"
|
|
I18n.t("onebox.github.pr_title.#{key}")
|
|
end
|
|
|
|
def load_commit(link)
|
|
if commit_match = link.match(%r{commits/(\h+)})
|
|
load_json(
|
|
"https://api.github.com/repos/#{match[:org]}/#{match[:repository]}/commits/#{commit_match[1]}",
|
|
)
|
|
end
|
|
end
|
|
|
|
def load_comment(link)
|
|
if comment_match = link.match(/#issuecomment-(\d+)/)
|
|
load_json(
|
|
"https://api.github.com/repos/#{match[:org]}/#{match[:repository]}/issues/comments/#{comment_match[1]}",
|
|
)
|
|
end
|
|
end
|
|
|
|
def load_review(link)
|
|
if review_match = link.match(/#discussion_r(\d+)/)
|
|
load_json(
|
|
"https://api.github.com/repos/#{match[:org]}/#{match[:repository]}/pulls/comments/#{review_match[1]}",
|
|
)
|
|
end
|
|
end
|
|
|
|
def load_json(url)
|
|
::MultiJson.load(
|
|
URI.parse(url).open({ read_timeout: timeout }.merge(github_auth_header(match[:org]))),
|
|
)
|
|
rescue OpenURI::HTTPError => e
|
|
Rails.logger.warn("GitHub API error: #{e.io.status[0]} fetching #{url}")
|
|
raise
|
|
end
|
|
|
|
def fetch_pr_status(pr_data)
|
|
return unless SiteSetting.github_pr_status_enabled
|
|
|
|
return { status: "merged", timestamp: pr_data["merged_at"] } if pr_data["merged"]
|
|
return { status: "closed", timestamp: pr_data["closed_at"] } if pr_data["state"] == "closed"
|
|
return { status: "draft", timestamp: pr_data["created_at"] } if pr_data["draft"]
|
|
|
|
reviews_data = load_json(url + "/reviews")
|
|
latest_reviews = latest_review_states_with_timestamps(reviews_data)
|
|
|
|
%w[CHANGES_REQUESTED APPROVED].each do |state|
|
|
reviews = latest_reviews.select { |r| r[:state] == state }
|
|
if reviews.present?
|
|
return { status: state.downcase, timestamp: reviews.map { |r| r[:timestamp] }.max }
|
|
end
|
|
end
|
|
|
|
{ status: "open", timestamp: pr_data["created_at"] }
|
|
rescue StandardError => e
|
|
Rails.logger.warn("GitHub PR status fetch error: #{e.message}")
|
|
nil
|
|
end
|
|
|
|
def latest_review_states_with_timestamps(reviews)
|
|
return [] if reviews.blank?
|
|
|
|
reviews
|
|
.reject do |r|
|
|
r.dig("user", "id").nil? || !%w[CHANGES_REQUESTED APPROVED].include?(r["state"])
|
|
end
|
|
.group_by { |r| r.dig("user", "id") }
|
|
.transform_values { |rs| rs.max_by { |r| r["submitted_at"] } }
|
|
.values
|
|
.map { |r| { state: r["state"], timestamp: r["submitted_at"] } }
|
|
end
|
|
end
|
|
end
|
|
end
|