26b124a67c
* Add a bundled cacert.pem from curl's website * Using the --ssl-ca-file PATH option overrides ENV var SSL_CA_PATH * Having none of those allow you to use the default, the bundled cacert.pem * Octokit's SSL config uses this new setting * Rake task to update_ssl_ca_file * Rubocop target 2.2
386 lines
12 KiB
Ruby
386 lines
12 KiB
Ruby
# frozen_string_literal: true
|
|
|
|
require "tmpdir"
|
|
require "retriable"
|
|
module GitHubChangelogGenerator
|
|
# A Fetcher responsible for all requests to GitHub and all basic manipulation with related data
|
|
# (such as filtering, validating, e.t.c)
|
|
#
|
|
# Example:
|
|
# fetcher = GitHubChangelogGenerator::OctoFetcher.new(options)
|
|
class OctoFetcher
|
|
PER_PAGE_NUMBER = 100
|
|
MAX_THREAD_NUMBER = 25
|
|
MAX_FORBIDDEN_RETRIES = 100
|
|
CHANGELOG_GITHUB_TOKEN = "CHANGELOG_GITHUB_TOKEN"
|
|
GH_RATE_LIMIT_EXCEEDED_MSG = "Warning: Can't finish operation: GitHub API rate limit exceeded, change log may be " \
|
|
"missing some issues. You can limit the number of issues fetched using the `--max-issues NUM` argument."
|
|
NO_TOKEN_PROVIDED = "Warning: No token provided (-t option) and variable $CHANGELOG_GITHUB_TOKEN was not found. " \
|
|
"This script can make only 50 requests to GitHub API per hour without token!"
|
|
|
|
# @param options [Hash] Options passed in
|
|
# @option options [String] :user GitHub username
|
|
# @option options [String] :project GitHub project
|
|
# @option options [String] :since Only issues updated at or after this time are returned. This is a timestamp in ISO 8601 format: YYYY-MM-DDTHH:MM:SSZ. eg. Time.parse("2016-01-01 10:00:00").iso8601
|
|
# @option options [Boolean] :http_cache Use ActiveSupport::Cache::FileStore to cache http requests
|
|
# @option options [Boolean] :cache_file If using http_cache, this is the cache file path
|
|
# @option options [Boolean] :cache_log If using http_cache, this is the cache log file path
|
|
def initialize(options = {})
|
|
@options = options || {}
|
|
@user = @options[:user]
|
|
@project = @options[:project]
|
|
@since = @options[:since]
|
|
@http_cache = @options[:http_cache]
|
|
if @http_cache
|
|
@cache_file = @options.fetch(:cache_file) { File.join(Dir.tmpdir, "github-changelog-http-cache") }
|
|
@cache_log = @options.fetch(:cache_log) { File.join(Dir.tmpdir, "github-changelog-logger.log") }
|
|
init_cache
|
|
end
|
|
@github_token = fetch_github_token
|
|
|
|
@request_options = { per_page: PER_PAGE_NUMBER }
|
|
@github_options = {}
|
|
@github_options[:access_token] = @github_token unless @github_token.nil?
|
|
@github_options[:api_endpoint] = @options[:github_endpoint] unless @options[:github_endpoint].nil?
|
|
|
|
configure_octokit_ssl
|
|
|
|
@client = Octokit::Client.new(@github_options)
|
|
end
|
|
|
|
def configure_octokit_ssl
|
|
ca_file = @options[:ssl_ca_file] || ENV["SSL_CA_FILE"] || File.expand_path("../ssl_certs/cacert.pem", __FILE__)
|
|
Octokit.connection_options = { ssl: { ca_file: ca_file } }
|
|
end
|
|
|
|
def init_cache
|
|
middleware_opts = {
|
|
serializer: Marshal,
|
|
store: ActiveSupport::Cache::FileStore.new(@cache_file),
|
|
logger: Logger.new(@cache_log),
|
|
shared_cache: false
|
|
}
|
|
stack = Faraday::RackBuilder.new do |builder|
|
|
builder.use Faraday::HttpCache, middleware_opts
|
|
builder.use Octokit::Response::RaiseError
|
|
builder.adapter Faraday.default_adapter
|
|
# builder.response :logger
|
|
end
|
|
Octokit.middleware = stack
|
|
end
|
|
|
|
# Fetch all tags from repo
|
|
#
|
|
# @return [Array <Hash>] array of tags
|
|
def get_all_tags
|
|
print "Fetching tags...\r" if @options[:verbose]
|
|
|
|
check_github_response { github_fetch_tags }
|
|
end
|
|
|
|
# Returns the number of pages for a API call
|
|
#
|
|
# @return [Integer] number of pages for this API call in total
|
|
def calculate_pages(client, method, request_options)
|
|
# Makes the first API call so that we can call last_response
|
|
check_github_response do
|
|
client.send(method, user_project, @request_options.merge(request_options))
|
|
end
|
|
|
|
last_response = client.last_response
|
|
|
|
if (last_pg = last_response.rels[:last])
|
|
querystring_as_hash(last_pg.href)["page"].to_i
|
|
else
|
|
1
|
|
end
|
|
end
|
|
|
|
# Fill input array with tags
|
|
#
|
|
# @return [Array <Hash>] array of tags in repo
|
|
def github_fetch_tags
|
|
tags = []
|
|
page_i = 0
|
|
count_pages = calculate_pages(@client, "tags", {})
|
|
|
|
iterate_pages(@client, "tags", {}) do |new_tags|
|
|
page_i += PER_PAGE_NUMBER
|
|
print_in_same_line("Fetching tags... #{page_i}/#{count_pages * PER_PAGE_NUMBER}")
|
|
tags.concat(new_tags)
|
|
end
|
|
print_empty_line
|
|
|
|
if tags.count == 0
|
|
Helper.log.warn "Warning: Can't find any tags in repo. \
|
|
Make sure, that you push tags to remote repo via 'git push --tags'"
|
|
else
|
|
Helper.log.info "Found #{tags.count} tags"
|
|
end
|
|
# tags are a Sawyer::Resource. Convert to hash
|
|
tags = tags.map { |h| stringify_keys_deep(h.to_hash) }
|
|
tags
|
|
end
|
|
|
|
# This method fetch all closed issues and separate them to pull requests and pure issues
|
|
# (pull request is kind of issue in term of GitHub)
|
|
#
|
|
# @return [Tuple] with (issues [Array <Hash>], pull-requests [Array <Hash>])
|
|
def fetch_closed_issues_and_pr
|
|
print "Fetching closed issues...\r" if @options[:verbose]
|
|
issues = []
|
|
options = {
|
|
state: "closed",
|
|
filter: "all",
|
|
labels: nil
|
|
}
|
|
options[:since] = @since unless @since.nil?
|
|
|
|
page_i = 0
|
|
count_pages = calculate_pages(@client, "issues", options)
|
|
|
|
iterate_pages(@client, "issues", options) do |new_issues|
|
|
page_i += PER_PAGE_NUMBER
|
|
print_in_same_line("Fetching issues... #{page_i}/#{count_pages * PER_PAGE_NUMBER}")
|
|
issues.concat(new_issues)
|
|
break if @options[:max_issues] && issues.length >= @options[:max_issues]
|
|
end
|
|
print_empty_line
|
|
Helper.log.info "Received issues: #{issues.count}"
|
|
|
|
issues = issues.map { |h| stringify_keys_deep(h.to_hash) }
|
|
|
|
# separate arrays of issues and pull requests:
|
|
issues.partition do |x|
|
|
x["pull_request"].nil?
|
|
end
|
|
end
|
|
|
|
# Fetch all pull requests. We need them to detect :merged_at parameter
|
|
#
|
|
# @return [Array <Hash>] all pull requests
|
|
def fetch_closed_pull_requests
|
|
pull_requests = []
|
|
options = { state: "closed" }
|
|
|
|
unless @options[:release_branch].nil?
|
|
options[:base] = @options[:release_branch]
|
|
end
|
|
|
|
page_i = 0
|
|
count_pages = calculate_pages(@client, "pull_requests", options)
|
|
|
|
iterate_pages(@client, "pull_requests", options) do |new_pr|
|
|
page_i += PER_PAGE_NUMBER
|
|
log_string = "Fetching merged dates... #{page_i}/#{count_pages * PER_PAGE_NUMBER}"
|
|
print_in_same_line(log_string)
|
|
pull_requests.concat(new_pr)
|
|
end
|
|
print_empty_line
|
|
|
|
Helper.log.info "Pull Request count: #{pull_requests.count}"
|
|
pull_requests = pull_requests.map { |h| stringify_keys_deep(h.to_hash) }
|
|
pull_requests
|
|
end
|
|
|
|
# Fetch event for all issues and add them to 'events'
|
|
#
|
|
# @param [Array] issues
|
|
# @return [Void]
|
|
def fetch_events_async(issues)
|
|
i = 0
|
|
threads = []
|
|
|
|
issues.each_slice(MAX_THREAD_NUMBER) do |issues_slice|
|
|
issues_slice.each do |issue|
|
|
threads << Thread.new do
|
|
issue["events"] = []
|
|
iterate_pages(@client, "issue_events", issue["number"], {}) do |new_event|
|
|
issue["events"].concat(new_event)
|
|
end
|
|
issue["events"] = issue["events"].map { |h| stringify_keys_deep(h.to_hash) }
|
|
print_in_same_line("Fetching events for issues and PR: #{i + 1}/#{issues.count}")
|
|
i += 1
|
|
end
|
|
end
|
|
threads.each(&:join)
|
|
threads = []
|
|
end
|
|
|
|
# to clear line from prev print
|
|
print_empty_line
|
|
|
|
Helper.log.info "Fetching events for issues and PR: #{i}"
|
|
end
|
|
|
|
# Fetch tag time from repo
|
|
#
|
|
# @param [Hash] tag GitHub data item about a Tag
|
|
#
|
|
# @return [Time] time of specified tag
|
|
def fetch_date_of_tag(tag)
|
|
commit_data = check_github_response { @client.commit(user_project, tag["commit"]["sha"]) }
|
|
commit_data = stringify_keys_deep(commit_data.to_hash)
|
|
|
|
commit_data["commit"]["committer"]["date"]
|
|
end
|
|
|
|
# Fetch commit for specified event
|
|
#
|
|
# @return [Hash]
|
|
def fetch_commit(event)
|
|
check_github_response do
|
|
commit = @client.commit(user_project, event["commit_id"])
|
|
commit = stringify_keys_deep(commit.to_hash)
|
|
commit
|
|
end
|
|
end
|
|
|
|
private
|
|
|
|
def stringify_keys_deep(indata)
|
|
case indata
|
|
when Array
|
|
indata.map do |value|
|
|
stringify_keys_deep(value)
|
|
end
|
|
when Hash
|
|
indata.each_with_object({}) do |(k, v), output|
|
|
output[k.to_s] = stringify_keys_deep(v)
|
|
end
|
|
else
|
|
indata
|
|
end
|
|
end
|
|
|
|
MovedPermanentlyError = Class.new(RuntimeError)
|
|
|
|
# Iterates through all pages until there are no more :next pages to follow
|
|
# yields the result per page
|
|
#
|
|
# @param [Octokit::Client] client
|
|
# @param [String] method (eg. 'tags')
|
|
#
|
|
# @yield [Sawyer::Resource] An OctoKit-provided response (which can be empty)
|
|
#
|
|
# @return [Integer] total number of pages
|
|
def iterate_pages(client, method, *args)
|
|
request_opts = extract_request_args(args)
|
|
args.push(@request_options.merge(request_opts))
|
|
|
|
number_of_pages = 1
|
|
|
|
check_github_response { client.send(method, user_project, *args) }
|
|
last_response = client.last_response
|
|
if last_response.status == 301
|
|
raise MovedPermanentlyError, last_response.data[:url]
|
|
end
|
|
|
|
yield(last_response.data)
|
|
|
|
until (next_one = last_response.rels[:next]).nil?
|
|
number_of_pages += 1
|
|
|
|
last_response = check_github_response { next_one.get }
|
|
yield(last_response.data)
|
|
end
|
|
|
|
number_of_pages
|
|
end
|
|
|
|
def extract_request_args(args)
|
|
if args.size == 1 && args.first.is_a?(Hash)
|
|
args.delete_at(0)
|
|
elsif args.size > 1 && args.last.is_a?(Hash)
|
|
args.delete_at(args.length - 1)
|
|
else
|
|
{}
|
|
end
|
|
end
|
|
|
|
# This is wrapper with rescue block
|
|
#
|
|
# @return [Object] returns exactly the same, what you put in the block, but wrap it with begin-rescue block
|
|
def check_github_response
|
|
Retriable.retriable(retry_options) do
|
|
yield
|
|
end
|
|
rescue MovedPermanentlyError => e
|
|
Helper.log.error("#{e.class}: #{e.message}")
|
|
sys_abort("The repository has moved, please update your configuration")
|
|
rescue Octokit::Forbidden => e
|
|
Helper.log.error("#{e.class}: #{e.message}")
|
|
sys_abort("Exceeded retry limit")
|
|
rescue Octokit::Unauthorized => e
|
|
Helper.log.error("#{e.class}: #{e.message}")
|
|
sys_abort("Error: wrong GitHub token")
|
|
end
|
|
|
|
# Exponential backoff
|
|
def retry_options
|
|
{
|
|
on: [Octokit::Forbidden],
|
|
tries: MAX_FORBIDDEN_RETRIES,
|
|
base_interval: sleep_base_interval,
|
|
multiplier: 1.0,
|
|
rand_factor: 0.0,
|
|
on_retry: retry_callback
|
|
}
|
|
end
|
|
|
|
def sleep_base_interval
|
|
1.0
|
|
end
|
|
|
|
def retry_callback
|
|
proc do |exception, try, elapsed_time, next_interval|
|
|
Helper.log.warn("RETRY - #{exception.class}: '#{exception.message}'")
|
|
Helper.log.warn("#{try} tries in #{elapsed_time} seconds and #{next_interval} seconds until the next try")
|
|
Helper.log.warn GH_RATE_LIMIT_EXCEEDED_MSG
|
|
Helper.log.warn @client.rate_limit
|
|
end
|
|
end
|
|
|
|
def sys_abort(msg)
|
|
abort(msg)
|
|
end
|
|
|
|
# Print specified line on the same string
|
|
#
|
|
# @param [String] log_string
|
|
def print_in_same_line(log_string)
|
|
print log_string + "\r"
|
|
end
|
|
|
|
# Print long line with spaces on same line to clear prev message
|
|
def print_empty_line
|
|
print_in_same_line(" ")
|
|
end
|
|
|
|
# Returns GitHub token. First try to use variable, provided by --token option,
|
|
# otherwise try to fetch it from CHANGELOG_GITHUB_TOKEN env variable.
|
|
#
|
|
# @return [String]
|
|
def fetch_github_token
|
|
env_var = @options[:token] ? @options[:token] : (ENV.fetch CHANGELOG_GITHUB_TOKEN, nil)
|
|
|
|
Helper.log.warn NO_TOKEN_PROVIDED unless env_var
|
|
|
|
env_var
|
|
end
|
|
|
|
# @return [String] helper to return Github "user/project"
|
|
def user_project
|
|
"#{@options[:user]}/#{@options[:project]}"
|
|
end
|
|
|
|
# Returns Hash of all querystring variables in given URI.
|
|
#
|
|
# @param [String] uri eg. https://api.github.com/repositories/43914960/tags?page=37&foo=1
|
|
# @return [Hash] of all GET variables. eg. { 'page' => 37, 'foo' => 1 }
|
|
def querystring_as_hash(uri)
|
|
Hash[URI.decode_www_form(URI(uri).query || "")]
|
|
end
|
|
end
|
|
end
|