857 lines
28 KiB
Python
857 lines
28 KiB
Python
# -*- coding: utf-8 -*-
|
|
# Copyright (c) 2013 The Chromium OS Authors. All rights reserved.
|
|
# Use of this source code is governed by a BSD-style license that can be
|
|
# found in the LICENSE file.
|
|
|
|
"""Utilities for requesting information for a gerrit server via https.
|
|
|
|
https://gerrit-review.googlesource.com/Documentation/rest-api.html
|
|
"""
|
|
|
|
from __future__ import print_function
|
|
|
|
import datetime
|
|
import json
|
|
import os
|
|
import re
|
|
import socket
|
|
import sys
|
|
import warnings
|
|
|
|
import httplib2
|
|
try:
|
|
from oauth2client import gce
|
|
except ImportError: # Newer oauth2client versions put it in .contrib
|
|
# pylint: disable=import-error,no-name-in-module
|
|
from oauth2client.contrib import gce
|
|
import six
|
|
from six.moves import html_parser as HTMLParser
|
|
from six.moves import http_client as httplib
|
|
from six.moves import http_cookiejar as cookielib
|
|
from six.moves import urllib
|
|
|
|
from autotest_lib.utils.frozen_chromite.lib import auth
|
|
from autotest_lib.utils.frozen_chromite.lib import constants
|
|
from autotest_lib.utils.frozen_chromite.lib import cros_logging as logging
|
|
from autotest_lib.utils.frozen_chromite.lib import git
|
|
from autotest_lib.utils.frozen_chromite.lib import retry_util
|
|
from autotest_lib.utils.frozen_chromite.lib import timeout_util
|
|
from autotest_lib.utils.frozen_chromite.lib import cros_build_lib
|
|
from autotest_lib.utils.frozen_chromite.utils import memoize
|
|
|
|
|
|
_GAE_VERSION = 'GAE_VERSION'
|
|
|
|
|
|
class ErrorParser(HTMLParser.HTMLParser):
|
|
"""Class to parse GOB error message reported as HTML.
|
|
|
|
Only data inside <div id='af-error-container'> section is retrieved from the
|
|
GOB error message. Retrieved data is processed as follows:
|
|
|
|
- newlines are removed
|
|
- each <br> tag is replaced with '\n'
|
|
- each <p> tag is replaced with '\n\n'
|
|
"""
|
|
|
|
def __init__(self):
|
|
HTMLParser.HTMLParser.__init__(self)
|
|
self.in_div = False
|
|
self.err_data = ''
|
|
|
|
def handle_starttag(self, tag, attrs):
|
|
tag_id = [x[1] for x in attrs if x[0] == 'id']
|
|
if tag == 'div' and tag_id and tag_id[0] == 'af-error-container':
|
|
self.in_div = True
|
|
return
|
|
|
|
if self.in_div:
|
|
if tag == 'p':
|
|
self.err_data += '\n\n'
|
|
return
|
|
|
|
if tag == 'br':
|
|
self.err_data += '\n'
|
|
return
|
|
|
|
def handle_endtag(self, tag):
|
|
if tag == 'div':
|
|
self.in_div = False
|
|
|
|
def handle_data(self, data):
|
|
if self.in_div:
|
|
self.err_data += data.replace('\n', '')
|
|
|
|
def ParsedDiv(self):
|
|
return self.err_data.strip()
|
|
|
|
|
|
@memoize.Memoize
|
|
def _GetAppCredentials():
|
|
"""Returns the singleton Appengine credentials for gerrit code review."""
|
|
return gce.AppAssertionCredentials(
|
|
scope='https://www.googleapis.com/auth/gerritcodereview')
|
|
|
|
|
|
TRY_LIMIT = 11
|
|
SLEEP = 0.5
|
|
REQUEST_TIMEOUT_SECONDS = 120 # 2 minutes.
|
|
|
|
# Controls the transport protocol used to communicate with Gerrit servers using
|
|
# git. This is parameterized primarily to enable cros_test_lib.GerritTestCase.
|
|
GIT_PROTOCOL = 'https'
|
|
|
|
# The GOB conflict errors which could be ignorable.
|
|
GOB_CONFLICT_ERRORS = (
|
|
br'change is closed',
|
|
br'Cannot reduce vote on labels for closed change',
|
|
)
|
|
|
|
GOB_CONFLICT_ERRORS_RE = re.compile(br'|'.join(GOB_CONFLICT_ERRORS),
|
|
re.IGNORECASE)
|
|
|
|
GOB_ERROR_REASON_CLOSED_CHANGE = 'CLOSED CHANGE'
|
|
|
|
|
|
class GOBError(Exception):
|
|
"""Exception class for errors commuicating with the gerrit-on-borg service."""
|
|
def __init__(self, http_status=None, reason=None):
|
|
self.http_status = http_status
|
|
self.reason = reason
|
|
|
|
message = ''
|
|
if http_status is not None:
|
|
message += '(http_status): %d' % (http_status,)
|
|
if reason is not None:
|
|
message += '(reason): %s' % (reason,)
|
|
if not message:
|
|
message = 'Unknown error'
|
|
|
|
super(GOBError, self).__init__(message)
|
|
|
|
|
|
class InternalGOBError(GOBError):
|
|
"""Exception class for GOB errors with status >= 500"""
|
|
|
|
|
|
def _QueryString(param_dict, first_param=None):
|
|
"""Encodes query parameters in the key:val[+key:val...] format specified here:
|
|
|
|
https://gerrit-review.googlesource.com/Documentation/rest-api-changes.html#list-changes
|
|
"""
|
|
q = [urllib.parse.quote(first_param)] if first_param else []
|
|
q.extend(['%s:%s' % (key, val) for key, val in param_dict.items()])
|
|
return '+'.join(q)
|
|
|
|
|
|
def GetCookies(host, path, cookie_paths=None):
|
|
"""Returns cookies that should be set on a request.
|
|
|
|
Used by CreateHttpConn for any requests that do not already specify a Cookie
|
|
header. All requests made by this library are HTTPS.
|
|
|
|
Args:
|
|
host: The hostname of the Gerrit service.
|
|
path: The path on the Gerrit service, already including /a/ if applicable.
|
|
cookie_paths: Files to look in for cookies. Defaults to looking in the
|
|
standard places where GoB places cookies.
|
|
|
|
Returns:
|
|
A dict of cookie name to value, with no URL encoding applied.
|
|
"""
|
|
cookies = {}
|
|
if cookie_paths is None:
|
|
cookie_paths = (constants.GOB_COOKIE_PATH, constants.GITCOOKIES_PATH)
|
|
for cookie_path in cookie_paths:
|
|
if os.path.isfile(cookie_path):
|
|
with open(cookie_path) as f:
|
|
for line in f:
|
|
fields = line.strip().split('\t')
|
|
if line.strip().startswith('#') or len(fields) != 7:
|
|
continue
|
|
domain, xpath, key, value = fields[0], fields[2], fields[5], fields[6]
|
|
if cookielib.domain_match(host, domain) and path.startswith(xpath):
|
|
cookies[key] = value
|
|
return cookies
|
|
|
|
|
|
def CreateHttpConn(host, path, reqtype='GET', headers=None, body=None):
|
|
"""Opens an https connection to a gerrit service, and sends a request."""
|
|
path = '/a/' + path.lstrip('/')
|
|
headers = headers or {}
|
|
if _InAppengine():
|
|
# TODO(phobbs) how can we choose to only run this on GCE / AppEngine?
|
|
credentials = _GetAppCredentials()
|
|
try:
|
|
headers.setdefault(
|
|
'Authorization',
|
|
'Bearer %s' % credentials.get_access_token().access_token)
|
|
except gce.HttpAccessTokenRefreshError as e:
|
|
logging.debug('Failed to retreive gce access token: %s', e)
|
|
# Not in an Appengine or GCE environment.
|
|
except httplib2.ServerNotFoundError as e:
|
|
pass
|
|
|
|
cookies = GetCookies(host, path)
|
|
if 'Cookie' not in headers and cookies:
|
|
headers['Cookie'] = '; '.join('%s=%s' % (n, v) for n, v in cookies.items())
|
|
elif 'Authorization' not in headers:
|
|
try:
|
|
git_creds = auth.GitCreds()
|
|
except auth.AccessTokenError:
|
|
git_creds = None
|
|
if git_creds:
|
|
headers.setdefault('Authorization', 'Bearer %s' % git_creds)
|
|
else:
|
|
logging.debug(
|
|
'No gitcookies file, Appengine credentials, or LUCI git creds found.')
|
|
|
|
if 'User-Agent' not in headers:
|
|
# We may not be in a git repository.
|
|
try:
|
|
version = git.GetGitRepoRevision(
|
|
os.path.dirname(os.path.realpath(__file__)))
|
|
except cros_build_lib.RunCommandError:
|
|
version = 'unknown'
|
|
headers['User-Agent'] = ' '.join((
|
|
'autotest.chromite.lib.gob_util',
|
|
os.path.basename(sys.argv[0]),
|
|
version,
|
|
))
|
|
|
|
if body:
|
|
body = json.JSONEncoder().encode(body)
|
|
headers.setdefault('Content-Type', 'application/json')
|
|
if logging.getLogger().isEnabledFor(logging.DEBUG):
|
|
logging.debug('%s https://%s%s', reqtype, host, path)
|
|
for key, val in headers.items():
|
|
if key.lower() in ('authorization', 'cookie'):
|
|
val = 'HIDDEN'
|
|
logging.debug('%s: %s', key, val)
|
|
if body:
|
|
logging.debug(body)
|
|
conn = httplib.HTTPSConnection(host)
|
|
conn.req_host = host
|
|
conn.req_params = {
|
|
'url': path,
|
|
'method': reqtype,
|
|
'headers': headers,
|
|
'body': body,
|
|
}
|
|
conn.request(**conn.req_params)
|
|
return conn
|
|
|
|
|
|
def _InAppengine():
|
|
"""Returns whether we're in the Appengine environment."""
|
|
return _GAE_VERSION in os.environ
|
|
|
|
|
|
def FetchUrl(host, path, reqtype='GET', headers=None, body=None,
|
|
ignore_204=False, ignore_404=True):
|
|
"""Fetches the http response from the specified URL.
|
|
|
|
Args:
|
|
host: The hostname of the Gerrit service.
|
|
path: The path on the Gerrit service. This will be prefixed with '/a'
|
|
automatically.
|
|
reqtype: The request type. Can be GET or POST.
|
|
headers: A mapping of extra HTTP headers to pass in with the request.
|
|
body: A string of data to send after the headers are finished.
|
|
ignore_204: for some requests gerrit-on-borg will return 204 to confirm
|
|
proper processing of the request. When processing responses to
|
|
these requests we should expect this status.
|
|
ignore_404: For many requests, gerrit-on-borg will return 404 if the request
|
|
doesn't match the database contents. In most such cases, we
|
|
want the API to return None rather than raise an Exception.
|
|
|
|
Returns:
|
|
The connection's reply, as bytes.
|
|
"""
|
|
@timeout_util.TimeoutDecorator(REQUEST_TIMEOUT_SECONDS)
|
|
def _FetchUrlHelper():
|
|
err_prefix = 'A transient error occured while querying %s:\n' % (host,)
|
|
try:
|
|
conn = CreateHttpConn(host, path, reqtype=reqtype, headers=headers,
|
|
body=body)
|
|
response = conn.getresponse()
|
|
except socket.error as ex:
|
|
logging.warning('%s%s', err_prefix, str(ex))
|
|
raise
|
|
|
|
# Normal/good responses.
|
|
response_body = response.read()
|
|
if response.status == 204 and ignore_204:
|
|
# This exception is used to confirm expected response status.
|
|
raise GOBError(http_status=response.status, reason=response.reason)
|
|
if response.status == 404 and ignore_404:
|
|
return b''
|
|
elif response.status == 200:
|
|
return response_body
|
|
|
|
# Bad responses.
|
|
logging.debug('response msg:\n%s', response.msg)
|
|
http_version = 'HTTP/%s' % ('1.1' if response.version == 11 else '1.0')
|
|
msg = ('%s %s %s\n%s %d %s\nResponse body: %r' %
|
|
(reqtype, conn.req_params['url'], http_version,
|
|
http_version, response.status, response.reason,
|
|
response_body))
|
|
|
|
# Ones we can retry.
|
|
if response.status >= 500:
|
|
# A status >=500 is assumed to be a possible transient error; retry.
|
|
logging.warning('%s%s', err_prefix, msg)
|
|
raise InternalGOBError(
|
|
http_status=response.status,
|
|
reason=response.reason)
|
|
|
|
# Ones we cannot retry.
|
|
home = os.environ.get('HOME', '~')
|
|
url = 'https://%s/new-password' % host
|
|
if response.status in (302, 303, 307):
|
|
err_prefix = ('Redirect found; missing/bad %s/.gitcookies credentials or '
|
|
'permissions (0600)?\n See %s' % (home, url))
|
|
elif response.status in (400,):
|
|
err_prefix = 'Permission error; talk to the admins of the GoB instance'
|
|
elif response.status in (401,):
|
|
err_prefix = ('Authorization error; missing/bad %s/.gitcookies '
|
|
'credentials or permissions (0600)?\n See %s' % (home, url))
|
|
elif response.status in (422,):
|
|
err_prefix = ('Bad request body?')
|
|
|
|
logging.warning(err_prefix)
|
|
|
|
# If GOB output contained expected error message, reduce log visibility of
|
|
# raw GOB output reported below.
|
|
ep = ErrorParser()
|
|
ep.feed(response_body.decode('utf-8'))
|
|
ep.close()
|
|
parsed_div = ep.ParsedDiv()
|
|
if parsed_div:
|
|
logging.warning('GOB Error:\n%s', parsed_div)
|
|
logging_function = logging.debug
|
|
else:
|
|
logging_function = logging.warning
|
|
|
|
logging_function(msg)
|
|
if response.status >= 400:
|
|
# The 'X-ErrorId' header is set only on >= 400 response code.
|
|
logging_function('X-ErrorId: %s', response.getheader('X-ErrorId'))
|
|
|
|
try:
|
|
logging.warning('conn.sock.getpeername(): %s', conn.sock.getpeername())
|
|
except AttributeError:
|
|
logging.warning('peer name unavailable')
|
|
|
|
if response.status == httplib.CONFLICT:
|
|
# 409 conflict
|
|
if GOB_CONFLICT_ERRORS_RE.search(response_body):
|
|
raise GOBError(
|
|
http_status=response.status,
|
|
reason=GOB_ERROR_REASON_CLOSED_CHANGE)
|
|
else:
|
|
raise GOBError(http_status=response.status, reason=response.reason)
|
|
else:
|
|
raise GOBError(http_status=response.status, reason=response.reason)
|
|
|
|
return retry_util.RetryException(
|
|
(socket.error, InternalGOBError, timeout_util.TimeoutError),
|
|
TRY_LIMIT,
|
|
_FetchUrlHelper, sleep=SLEEP, backoff_factor=2)
|
|
|
|
|
|
def FetchUrlJson(*args, **kwargs):
|
|
"""Fetch the specified URL and parse it as JSON.
|
|
|
|
See FetchUrl for arguments.
|
|
"""
|
|
fh = FetchUrl(*args, **kwargs)
|
|
|
|
# In case ignore_404 is True, we want to return None instead of
|
|
# raising an exception.
|
|
if not fh:
|
|
return None
|
|
|
|
# The first line of the response should always be: )]}'
|
|
if not fh.startswith(b")]}'"):
|
|
raise GOBError(http_status=200, reason='Unexpected json output: %r' % fh)
|
|
|
|
_, _, json_data = fh.partition(b'\n')
|
|
return json.loads(json_data)
|
|
|
|
|
|
def QueryChanges(host, param_dict, first_param=None, limit=None, o_params=None,
|
|
start=None):
|
|
"""Queries a gerrit-on-borg server for changes matching query terms.
|
|
|
|
Args:
|
|
host: The Gerrit server hostname.
|
|
param_dict: A dictionary of search parameters, as documented here:
|
|
https://gerrit-review.googlesource.com/Documentation/user-search.html
|
|
first_param: A change identifier
|
|
limit: Maximum number of results to return.
|
|
o_params: A list of additional output specifiers, as documented here:
|
|
https://gerrit-review.googlesource.com/Documentation/rest-api-changes.html#list-changes
|
|
start: Offset in the result set to start at.
|
|
|
|
Returns:
|
|
A list of json-decoded query results.
|
|
"""
|
|
# Note that no attempt is made to escape special characters; YMMV.
|
|
if not param_dict and not first_param:
|
|
raise RuntimeError('QueryChanges requires search parameters')
|
|
path = 'changes/?q=%s' % _QueryString(param_dict, first_param)
|
|
if start:
|
|
path = '%s&S=%d' % (path, start)
|
|
if limit:
|
|
path = '%s&n=%d' % (path, limit)
|
|
if o_params:
|
|
path = '%s&%s' % (path, '&'.join(['o=%s' % p for p in o_params]))
|
|
# Don't ignore 404; a query should always return a list, even if it's empty.
|
|
return FetchUrlJson(host, path, ignore_404=False)
|
|
|
|
|
|
def MultiQueryChanges(host, param_dict, change_list, limit=None, o_params=None,
|
|
start=None):
|
|
"""Initiate a query composed of multiple sets of query parameters."""
|
|
if not change_list:
|
|
raise RuntimeError(
|
|
"MultiQueryChanges requires a list of change numbers/id's")
|
|
q = ['q=%s' % '+OR+'.join(urllib.parse.quote(str(x)) for x in change_list)]
|
|
if param_dict:
|
|
q.append(_QueryString(param_dict))
|
|
if limit:
|
|
q.append('n=%d' % limit)
|
|
if start:
|
|
q.append('S=%s' % start)
|
|
if o_params:
|
|
q.extend(['o=%s' % p for p in o_params])
|
|
path = 'changes/?%s' % '&'.join(q)
|
|
try:
|
|
result = FetchUrlJson(host, path, ignore_404=False)
|
|
except GOBError as e:
|
|
msg = '%s:\n%s' % (e, path)
|
|
raise GOBError(http_status=e.http_status, reason=msg)
|
|
return result
|
|
|
|
|
|
def GetGerritFetchUrl(host):
|
|
"""Given a gerrit host name returns URL of a gerrit instance to fetch from."""
|
|
return 'https://%s/' % host
|
|
|
|
|
|
def GetChangePageUrl(host, change_number):
|
|
"""Given a gerrit host name and change number, return change page url."""
|
|
return 'https://%s/#/c/%d/' % (host, change_number)
|
|
|
|
|
|
def _GetChangePath(change):
|
|
"""Given a change id, return a path prefix for the change."""
|
|
return 'changes/%s' % str(change).replace('/', '%2F')
|
|
|
|
|
|
def GetChangeUrl(host, change):
|
|
"""Given a gerrit host name and change id, return an url for the change."""
|
|
return 'https://%s/a/%s' % (host, _GetChangePath(change))
|
|
|
|
|
|
def GetChange(host, change):
|
|
"""Query a gerrit server for information about a single change."""
|
|
return FetchUrlJson(host, _GetChangePath(change))
|
|
|
|
|
|
def GetChangeReview(host, change, revision=None):
|
|
"""Get the current review information for a change."""
|
|
if revision is None:
|
|
revision = 'current'
|
|
path = '%s/revisions/%s/review' % (_GetChangePath(change), revision)
|
|
return FetchUrlJson(host, path)
|
|
|
|
|
|
def GetChangeCommit(host, change, revision=None):
|
|
"""Get the current review information for a change."""
|
|
if revision is None:
|
|
revision = 'current'
|
|
path = '%s/revisions/%s/commit' % (_GetChangePath(change), revision)
|
|
return FetchUrlJson(host, path)
|
|
|
|
|
|
def GetChangeCurrentRevision(host, change):
|
|
"""Get information about the latest revision for a given change."""
|
|
jmsg = GetChangeReview(host, change)
|
|
if jmsg:
|
|
return jmsg.get('current_revision')
|
|
|
|
|
|
def GetChangeDetail(host, change, o_params=None):
|
|
"""Query a gerrit server for extended information about a single change."""
|
|
path = '%s/detail' % _GetChangePath(change)
|
|
if o_params:
|
|
path = '%s?%s' % (path, '&'.join(['o=%s' % p for p in o_params]))
|
|
return FetchUrlJson(host, path)
|
|
|
|
|
|
def GetChangeReviewers(host, change):
|
|
"""Get information about all reviewers attached to a change.
|
|
|
|
Args:
|
|
host: The Gerrit host to interact with.
|
|
change: The Gerrit change ID.
|
|
"""
|
|
warnings.warn('GetChangeReviewers is deprecated; use GetReviewers instead.')
|
|
GetReviewers(host, change)
|
|
|
|
|
|
def ReviewedChange(host, change):
|
|
"""Mark a gerrit change as reviewed."""
|
|
path = '%s/reviewed' % _GetChangePath(change)
|
|
return FetchUrlJson(host, path, reqtype='PUT', ignore_404=False)
|
|
|
|
|
|
def UnreviewedChange(host, change):
|
|
"""Mark a gerrit change as unreviewed."""
|
|
path = '%s/unreviewed' % _GetChangePath(change)
|
|
return FetchUrlJson(host, path, reqtype='PUT', ignore_404=False)
|
|
|
|
|
|
def IgnoreChange(host, change):
|
|
"""Ignore a gerrit change."""
|
|
path = '%s/ignore' % _GetChangePath(change)
|
|
return FetchUrlJson(host, path, reqtype='PUT', ignore_404=False)
|
|
|
|
|
|
def UnignoreChange(host, change):
|
|
"""Unignore a gerrit change."""
|
|
path = '%s/unignore' % _GetChangePath(change)
|
|
return FetchUrlJson(host, path, reqtype='PUT', ignore_404=False)
|
|
|
|
|
|
def AbandonChange(host, change, msg=''):
|
|
"""Abandon a gerrit change."""
|
|
path = '%s/abandon' % _GetChangePath(change)
|
|
body = {'message': msg}
|
|
return FetchUrlJson(host, path, reqtype='POST', body=body, ignore_404=False)
|
|
|
|
|
|
def RestoreChange(host, change, msg=''):
|
|
"""Restore a previously abandoned change."""
|
|
path = '%s/restore' % _GetChangePath(change)
|
|
body = {'message': msg}
|
|
return FetchUrlJson(host, path, reqtype='POST', body=body, ignore_404=False)
|
|
|
|
|
|
def DeleteDraft(host, change):
|
|
"""Delete a gerrit draft change."""
|
|
path = _GetChangePath(change)
|
|
try:
|
|
FetchUrl(host, path, reqtype='DELETE', ignore_204=True, ignore_404=False)
|
|
except GOBError as e:
|
|
# On success, gerrit returns status 204; anything else is an error.
|
|
if e.http_status != 204:
|
|
raise
|
|
else:
|
|
raise GOBError(
|
|
http_status=200,
|
|
reason='Unexpectedly received a 200 http status while deleting draft '
|
|
' %r' % change)
|
|
|
|
|
|
def SubmitChange(host, change, revision=None, wait_for_merge=True):
|
|
"""Submits a gerrit change via Gerrit."""
|
|
if revision is None:
|
|
revision = 'current'
|
|
path = '%s/revisions/%s/submit' % (_GetChangePath(change), revision)
|
|
body = {'wait_for_merge': wait_for_merge}
|
|
return FetchUrlJson(host, path, reqtype='POST', body=body, ignore_404=False)
|
|
|
|
|
|
def CheckChange(host, change, sha1=None):
|
|
"""Performs consistency checks on the change, and fixes inconsistencies.
|
|
|
|
This is useful for forcing Gerrit to check whether a change has already been
|
|
merged into the git repo. Namely, if |sha1| is provided and the change is in
|
|
'NEW' status, Gerrit will check if a change with that |sha1| is in the repo
|
|
and mark the change as 'MERGED' if it exists.
|
|
|
|
Args:
|
|
host: The Gerrit host to interact with.
|
|
change: The Gerrit change ID.
|
|
sha1: An optional hint of the commit's SHA1 in Git.
|
|
"""
|
|
path = '%s/check' % (_GetChangePath(change),)
|
|
if sha1:
|
|
body, headers = {'expect_merged_as': sha1}, {}
|
|
else:
|
|
body, headers = {}, {'Content-Length': '0'}
|
|
|
|
return FetchUrlJson(host, path, reqtype='POST',
|
|
body=body, ignore_404=False,
|
|
headers=headers)
|
|
|
|
|
|
def GetAssignee(host, change):
|
|
"""Get assignee for a change."""
|
|
path = '%s/assignee' % _GetChangePath(change)
|
|
return FetchUrlJson(host, path)
|
|
|
|
|
|
def AddAssignee(host, change, assignee):
|
|
"""Add reviewers to a change.
|
|
|
|
Args:
|
|
host: The Gerrit host to interact with.
|
|
change: The Gerrit change ID.
|
|
assignee: Gerrit account email as a string
|
|
"""
|
|
path = '%s/assignee' % _GetChangePath(change)
|
|
body = {'assignee': assignee}
|
|
return FetchUrlJson(host, path, reqtype='PUT', body=body, ignore_404=False)
|
|
|
|
|
|
def MarkPrivate(host, change):
|
|
"""Marks the given CL as private.
|
|
|
|
Args:
|
|
host: The gob host to interact with.
|
|
change: CL number on the given host.
|
|
"""
|
|
path = '%s/private' % _GetChangePath(change)
|
|
try:
|
|
FetchUrlJson(host, path, reqtype='POST', ignore_404=False)
|
|
except GOBError as e:
|
|
# 201: created -- change was successfully marked private.
|
|
if e.http_status != 201:
|
|
raise
|
|
else:
|
|
raise GOBError(
|
|
http_status=200,
|
|
reason='Change was already marked private',
|
|
)
|
|
|
|
|
|
def MarkNotPrivate(host, change):
|
|
"""Sets the private bit on given CL to False.
|
|
|
|
Args:
|
|
host: The gob host to interact with.
|
|
change: CL number on the given host.
|
|
"""
|
|
path = '%s/private.delete' % _GetChangePath(change)
|
|
try:
|
|
FetchUrlJson(host, path, reqtype='POST', ignore_404=False, ignore_204=True)
|
|
except GOBError as e:
|
|
if e.http_status == 204:
|
|
# 204: no content -- change was successfully marked not private.
|
|
pass
|
|
elif e.http_status == 409:
|
|
raise GOBError(
|
|
http_status=e.http_status,
|
|
reason='Change was already marked not private',
|
|
)
|
|
else:
|
|
raise
|
|
else:
|
|
raise GOBError(
|
|
http_status=200,
|
|
reason='Got unexpected 200 when marking change not private.',
|
|
)
|
|
|
|
|
|
def GetReviewers(host, change):
|
|
"""Get information about all reviewers attached to a change.
|
|
|
|
Args:
|
|
host: The Gerrit host to interact with.
|
|
change: The Gerrit change ID.
|
|
"""
|
|
path = '%s/reviewers' % _GetChangePath(change)
|
|
return FetchUrlJson(host, path)
|
|
|
|
|
|
def AddReviewers(host, change, add=None, notify=None):
|
|
"""Add reviewers to a change."""
|
|
if not add:
|
|
return
|
|
if isinstance(add, six.string_types):
|
|
add = (add,)
|
|
body = {}
|
|
if notify:
|
|
body['notify'] = notify
|
|
path = '%s/reviewers' % _GetChangePath(change)
|
|
for r in add:
|
|
body['reviewer'] = r
|
|
jmsg = FetchUrlJson(host, path, reqtype='POST', body=body, ignore_404=False)
|
|
return jmsg
|
|
|
|
|
|
def RemoveReviewers(host, change, remove=None, notify=None):
|
|
"""Remove reveiewers from a change."""
|
|
if not remove:
|
|
return
|
|
if isinstance(remove, six.string_types):
|
|
remove = (remove,)
|
|
body = {}
|
|
if notify:
|
|
body['notify'] = notify
|
|
for r in remove:
|
|
path = '%s/reviewers/%s/delete' % (_GetChangePath(change), r)
|
|
try:
|
|
FetchUrl(host, path, reqtype='POST', body=body, ignore_204=True)
|
|
except GOBError as e:
|
|
# On success, gerrit returns status 204; anything else is an error.
|
|
if e.http_status != 204:
|
|
raise
|
|
|
|
|
|
def SetReview(host, change, revision=None, msg=None, labels=None, notify=None):
|
|
"""Set labels and/or add a message to a code review."""
|
|
if revision is None:
|
|
revision = 'current'
|
|
if not msg and not labels:
|
|
return
|
|
path = '%s/revisions/%s/review' % (_GetChangePath(change), revision)
|
|
body = {}
|
|
if msg:
|
|
body['message'] = msg
|
|
if labels:
|
|
body['labels'] = labels
|
|
if notify:
|
|
body['notify'] = notify
|
|
response = FetchUrlJson(host, path, reqtype='POST', body=body)
|
|
if response is None:
|
|
raise GOBError(
|
|
http_status=404,
|
|
reason='CL %s not found in %s' % (change, host))
|
|
if labels:
|
|
for key, val in labels.items():
|
|
if ('labels' not in response or key not in response['labels'] or
|
|
int(response['labels'][key] != int(val))):
|
|
raise GOBError(
|
|
http_status=200,
|
|
reason='Unable to set "%s" label on change %s.' % (key, change))
|
|
|
|
|
|
def SetTopic(host, change, topic):
|
|
"""Set |topic| for a change. If |topic| is empty, it will be deleted"""
|
|
path = '%s/topic' % _GetChangePath(change)
|
|
body = {'topic': topic}
|
|
return FetchUrlJson(host, path, reqtype='PUT', body=body, ignore_404=False)
|
|
|
|
|
|
def SetHashtags(host, change, add, remove):
|
|
"""Adds and / or removes hashtags from a change.
|
|
|
|
Args:
|
|
host: Hostname (without protocol prefix) of the gerrit server.
|
|
change: A gerrit change number.
|
|
add: a list of hashtags to be added.
|
|
remove: a list of hashtags to be removed.
|
|
"""
|
|
path = '%s/hashtags' % _GetChangePath(change)
|
|
return FetchUrlJson(host, path, reqtype='POST',
|
|
body={'add': add, 'remove': remove},
|
|
ignore_404=False)
|
|
|
|
|
|
def ResetReviewLabels(host, change, label, value='0', revision=None,
|
|
message=None, notify=None):
|
|
"""Reset the value of a given label for all reviewers on a change."""
|
|
if revision is None:
|
|
revision = 'current'
|
|
# This is tricky when working on the "current" revision, because there's
|
|
# always the risk that the "current" revision will change in between API
|
|
# calls. So, the code dereferences the "current" revision down to a literal
|
|
# sha1 at the beginning and uses it for all subsequent calls. As a sanity
|
|
# check, the "current" revision is dereferenced again at the end, and if it
|
|
# differs from the previous "current" revision, an exception is raised.
|
|
current = (revision == 'current')
|
|
jmsg = GetChangeDetail(
|
|
host, change, o_params=['CURRENT_REVISION', 'CURRENT_COMMIT'])
|
|
if current:
|
|
revision = jmsg['current_revision']
|
|
value = str(value)
|
|
path = '%s/revisions/%s/review' % (_GetChangePath(change), revision)
|
|
message = message or (
|
|
'%s label set to %s programmatically by chromite.' % (label, value))
|
|
for review in jmsg.get('labels', {}).get(label, {}).get('all', []):
|
|
if str(review.get('value', value)) != value:
|
|
body = {
|
|
'message': message,
|
|
'labels': {label: value},
|
|
'on_behalf_of': review['_account_id'],
|
|
}
|
|
if notify:
|
|
body['notify'] = notify
|
|
response = FetchUrlJson(host, path, reqtype='POST', body=body)
|
|
if str(response['labels'][label]) != value:
|
|
username = review.get('email', jmsg.get('name', ''))
|
|
raise GOBError(
|
|
http_status=200,
|
|
reason='Unable to set %s label for user "%s" on change %s.' % (
|
|
label, username, change))
|
|
if current:
|
|
new_revision = GetChangeCurrentRevision(host, change)
|
|
if not new_revision:
|
|
raise GOBError(
|
|
http_status=200,
|
|
reason='Could not get review information for change "%s"' % change)
|
|
elif new_revision != revision:
|
|
raise GOBError(
|
|
http_status=200,
|
|
reason='While resetting labels on change "%s", a new patchset was '
|
|
'uploaded.' % change)
|
|
|
|
|
|
def GetTipOfTrunkRevision(git_url):
|
|
"""Returns the current git revision on the master branch."""
|
|
parsed_url = urllib.parse.urlparse(git_url)
|
|
path = parsed_url[2].rstrip('/') + '/+log/master?n=1&format=JSON'
|
|
j = FetchUrlJson(parsed_url[1], path, ignore_404=False)
|
|
if not j:
|
|
raise GOBError(
|
|
reason='Could not find revision information from %s' % git_url)
|
|
try:
|
|
return j['log'][0]['commit']
|
|
except (IndexError, KeyError, TypeError):
|
|
msg = ('The json returned by https://%s%s has an unfamiliar structure:\n'
|
|
'%s\n' % (parsed_url[1], path, j))
|
|
raise GOBError(reason=msg)
|
|
|
|
|
|
def GetCommitDate(git_url, commit):
|
|
"""Returns the date of a particular git commit.
|
|
|
|
The returned object is naive in the sense that it doesn't carry any timezone
|
|
information - you should assume UTC.
|
|
|
|
Args:
|
|
git_url: URL for the repository to get the commit date from.
|
|
commit: A git commit identifier (e.g. a sha1).
|
|
|
|
Returns:
|
|
A datetime object.
|
|
"""
|
|
parsed_url = urllib.parse.urlparse(git_url)
|
|
path = '%s/+log/%s?n=1&format=JSON' % (parsed_url.path.rstrip('/'), commit)
|
|
j = FetchUrlJson(parsed_url.netloc, path, ignore_404=False)
|
|
if not j:
|
|
raise GOBError(
|
|
reason='Could not find revision information from %s' % git_url)
|
|
try:
|
|
commit_timestr = j['log'][0]['committer']['time']
|
|
except (IndexError, KeyError, TypeError):
|
|
msg = ('The json returned by https://%s%s has an unfamiliar structure:\n'
|
|
'%s\n' % (parsed_url.netloc, path, j))
|
|
raise GOBError(reason=msg)
|
|
try:
|
|
# We're parsing a string of the form 'Tue Dec 02 17:48:06 2014'.
|
|
return datetime.datetime.strptime(commit_timestr,
|
|
constants.GOB_COMMIT_TIME_FORMAT)
|
|
except ValueError:
|
|
raise GOBError(reason='Failed parsing commit time "%s"' % commit_timestr)
|
|
|
|
|
|
def GetAccount(host):
|
|
"""Get information about the user account."""
|
|
return FetchUrlJson(host, 'accounts/self')
|