-
Notifications
You must be signed in to change notification settings - Fork 0
/
Copy pathconfig.py.in
47 lines (38 loc) · 1.82 KB
/
config.py.in
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
# Copy this file to config.py and edit to suit.
# Debug mode. In this mode, no uploads or updates will be performed. Instead,
# the data that would be imported to Github will be output, and updates to
# Redmine skipped.
DEBUG = False
# The maximum number of issues to try to migrate. Useful for testing with
# a small initial batch.
MAX_ISSUES = 9999
# The issues to migrate, based on their status
ISSUE_STATUS = 'all' # One of 'all', 'open', 'closed'
# Track Status. If this is enabled, Redmine IDs will be written to
# migrated_ids.txt, one per line. The next time the migration is run,
# any IDs in the file will be skipped. Note that they will be discarded
# *after* filtering by MAX_ISSUES and ISSUE_STATE.
TRACK_STATUS = True
# Clear labels from Github before migrating?
CLEAR_LABELS = False
# Clear (and migrate) milestones from Github before migrating?
CLEAR_MILESTONES = False
# Redmine details
REDMINE_URL = 'https://redmine.mydomain.com' # No trailing /
REDMINE_VERSION = 'my_redmine_version' # e.g. 4.0.7
REDMINE_TOKEN = 'my_redmine_api_token'
REDMINE_PROJECT = 'my_redmine_project'
REDMINE_CUSTOM_FIELDS = ['my_custom_field1', 'my_custom_field2'] # Values will be added as labels
REDMINE_LINKBACK = False # Add a link back to the new Github ticket on Redmine
# Github details
GITHUB_TOKEN = 'my_github_api_token' # Must have repo access scope
GITHUB_OWNER = 'my_github_owner_or_org' # Repo owner; username or org name
GITHUB_REPO = 'my_github_repo_name'
# AWS Details
# There's no API on Github for uploading attachments, so we upload them to an
# S3 bucket instead. They will be stored under /redmine/<issue_num>/. Files
# must be readable!
AWS_CLI_PROFILE = 'my_aws_cli_profile'
S3_BUCKET_NAME = 'my_s3_bucket_name'
S3_BUCKET_URL = 'my_s3_bucket_public_url'
S3_SKIP_UPLOAD = False # Useful if attachments have been previously uploaded