mirror of
https://github.com/servo/servo.git
synced 2025-08-03 12:40:06 +01:00
tests/wpt/update: use urllib module from six package
This commit is contained in:
parent
179caed30f
commit
d8190a7933
3 changed files with 7 additions and 7 deletions
|
@ -8,7 +8,7 @@ import gzip
|
||||||
import json
|
import json
|
||||||
import os
|
import os
|
||||||
import requests
|
import requests
|
||||||
import urlparse
|
import six.moves.urllib as urllib
|
||||||
|
|
||||||
treeherder_base = "https://treeherder.mozilla.org/"
|
treeherder_base = "https://treeherder.mozilla.org/"
|
||||||
|
|
||||||
|
@ -53,7 +53,7 @@ def download(url, prefix, dest, force_suffix=True):
|
||||||
|
|
||||||
def get_blobber_url(branch, job):
|
def get_blobber_url(branch, job):
|
||||||
job_id = job["id"]
|
job_id = job["id"]
|
||||||
resp = requests.get(urlparse.urljoin(treeherder_base,
|
resp = requests.get(urllib.parse.urljoin(treeherder_base,
|
||||||
"/api/project/%s/artifact/?job_id=%i&name=Job%%20Info" % (branch,
|
"/api/project/%s/artifact/?job_id=%i&name=Job%%20Info" % (branch,
|
||||||
job_id)))
|
job_id)))
|
||||||
job_data = resp.json()
|
job_data = resp.json()
|
||||||
|
@ -71,13 +71,13 @@ def get_blobber_url(branch, job):
|
||||||
|
|
||||||
|
|
||||||
def get_structured_logs(branch, commit, dest=None):
|
def get_structured_logs(branch, commit, dest=None):
|
||||||
resp = requests.get(urlparse.urljoin(treeherder_base, "/api/project/%s/resultset/?revision=%s" % (branch, commit)))
|
resp = requests.get(urllib.parse.urljoin(treeherder_base, "/api/project/%s/resultset/?revision=%s" % (branch, commit)))
|
||||||
|
|
||||||
revision_data = resp.json()
|
revision_data = resp.json()
|
||||||
|
|
||||||
result_set = revision_data["results"][0]["id"]
|
result_set = revision_data["results"][0]["id"]
|
||||||
|
|
||||||
resp = requests.get(urlparse.urljoin(treeherder_base, "/api/project/%s/jobs/?result_set_id=%s&count=2000&exclusion_profile=false" % (branch, result_set)))
|
resp = requests.get(urllib.parse.urljoin(treeherder_base, "/api/project/%s/jobs/?result_set_id=%s&count=2000&exclusion_profile=false" % (branch, result_set)))
|
||||||
|
|
||||||
job_data = resp.json()
|
job_data = resp.json()
|
||||||
|
|
||||||
|
|
|
@ -5,7 +5,7 @@
|
||||||
from __future__ import print_function
|
from __future__ import print_function
|
||||||
|
|
||||||
import json
|
import json
|
||||||
from urlparse import urljoin
|
from six.moves.urllib.parse import urljoin
|
||||||
requests = None
|
requests = None
|
||||||
|
|
||||||
class GitHubError(Exception):
|
class GitHubError(Exception):
|
||||||
|
|
|
@ -4,7 +4,7 @@ import os
|
||||||
import re
|
import re
|
||||||
import subprocess
|
import subprocess
|
||||||
import sys
|
import sys
|
||||||
import urlparse
|
import six.moves.urllib as urllib
|
||||||
|
|
||||||
from wptrunner.update.sync import UpdateCheckout
|
from wptrunner.update.sync import UpdateCheckout
|
||||||
from wptrunner.update.tree import get_unique_name
|
from wptrunner.update.tree import get_unique_name
|
||||||
|
@ -264,7 +264,7 @@ class MergeUpstream(Step):
|
||||||
if "merge_index" not in state:
|
if "merge_index" not in state:
|
||||||
state.merge_index = 0
|
state.merge_index = 0
|
||||||
|
|
||||||
org, name = urlparse.urlsplit(state.sync["remote_url"]).path[1:].split("/")
|
org, name = urllib.parse.urlsplit(state.sync["remote_url"]).path[1:].split("/")
|
||||||
if name.endswith(".git"):
|
if name.endswith(".git"):
|
||||||
name = name[:-4]
|
name = name[:-4]
|
||||||
state.gh_repo = gh.repo(org, name)
|
state.gh_repo = gh.repo(org, name)
|
||||||
|
|
Loading…
Add table
Add a link
Reference in a new issue