commit
stringlengths 40
40
| old_file
stringlengths 4
118
| new_file
stringlengths 4
118
| old_contents
stringlengths 10
2.94k
| new_contents
stringlengths 21
3.18k
| subject
stringlengths 16
444
| message
stringlengths 17
2.63k
| lang
stringclasses 1
value | license
stringclasses 13
values | repos
stringlengths 5
43k
| ndiff
stringlengths 52
3.32k
| instruction
stringlengths 16
444
| content
stringlengths 133
4.32k
| fuzzy_diff
stringlengths 16
3.18k
|
---|---|---|---|---|---|---|---|---|---|---|---|---|---|
7104882ffcd35c24d8df5b9aa909e9bc9619cba7 | eli5/__init__.py | eli5/__init__.py | from __future__ import absolute_import
__version__ = '0.4.1'
from .formatters import format_as_html, format_html_styles, format_as_text
from .explain import explain_weights, explain_prediction
from .sklearn import explain_weights_sklearn, explain_prediction_sklearn
try:
from .ipython import show_weights, show_prediction
except ImportError:
pass # IPython is not installed
try:
from .lightning import (
explain_prediction_lightning,
explain_weights_lightning
)
except ImportError as e:
# lightning is not available
pass
try:
from .sklearn_crfsuite import (
explain_weights_sklearn_crfsuite
)
except ImportError as e:
# sklearn-crfsuite is not available
pass
try:
from .xgboost import explain_weights_xgboost
except ImportError:
# xgboost is not available
pass
| from __future__ import absolute_import
__version__ = '0.4.1'
from .formatters import format_as_html, format_html_styles, format_as_text
from .explain import explain_weights, explain_prediction
from .sklearn import explain_weights_sklearn, explain_prediction_sklearn
try:
from .ipython import show_weights, show_prediction
except ImportError:
pass # IPython is not installed
try:
from .lightning import (
explain_prediction_lightning,
explain_weights_lightning
)
except ImportError as e:
# lightning is not available
pass
try:
from .sklearn_crfsuite import (
explain_weights_sklearn_crfsuite
)
except ImportError as e:
# sklearn-crfsuite is not available
pass
try:
from .xgboost import explain_weights_xgboost
except ImportError:
# xgboost is not available
pass
except Exception as e:
if e.__class__.__name__ == 'XGBoostLibraryNotFound':
# improperly installed xgboost
pass
else:
raise
| Handle improperly installed xgboost. Fixes GH-162. | Handle improperly installed xgboost. Fixes GH-162.
| Python | mit | TeamHG-Memex/eli5,TeamHG-Memex/eli5,TeamHG-Memex/eli5 | from __future__ import absolute_import
__version__ = '0.4.1'
from .formatters import format_as_html, format_html_styles, format_as_text
from .explain import explain_weights, explain_prediction
from .sklearn import explain_weights_sklearn, explain_prediction_sklearn
try:
from .ipython import show_weights, show_prediction
except ImportError:
pass # IPython is not installed
try:
from .lightning import (
explain_prediction_lightning,
explain_weights_lightning
)
except ImportError as e:
# lightning is not available
pass
try:
from .sklearn_crfsuite import (
explain_weights_sklearn_crfsuite
)
except ImportError as e:
# sklearn-crfsuite is not available
pass
try:
from .xgboost import explain_weights_xgboost
except ImportError:
# xgboost is not available
pass
+ except Exception as e:
+ if e.__class__.__name__ == 'XGBoostLibraryNotFound':
+ # improperly installed xgboost
+ pass
+ else:
+ raise
| Handle improperly installed xgboost. Fixes GH-162. | ## Code Before:
from __future__ import absolute_import
__version__ = '0.4.1'
from .formatters import format_as_html, format_html_styles, format_as_text
from .explain import explain_weights, explain_prediction
from .sklearn import explain_weights_sklearn, explain_prediction_sklearn
try:
from .ipython import show_weights, show_prediction
except ImportError:
pass # IPython is not installed
try:
from .lightning import (
explain_prediction_lightning,
explain_weights_lightning
)
except ImportError as e:
# lightning is not available
pass
try:
from .sklearn_crfsuite import (
explain_weights_sklearn_crfsuite
)
except ImportError as e:
# sklearn-crfsuite is not available
pass
try:
from .xgboost import explain_weights_xgboost
except ImportError:
# xgboost is not available
pass
## Instruction:
Handle improperly installed xgboost. Fixes GH-162.
## Code After:
from __future__ import absolute_import
__version__ = '0.4.1'
from .formatters import format_as_html, format_html_styles, format_as_text
from .explain import explain_weights, explain_prediction
from .sklearn import explain_weights_sklearn, explain_prediction_sklearn
try:
from .ipython import show_weights, show_prediction
except ImportError:
pass # IPython is not installed
try:
from .lightning import (
explain_prediction_lightning,
explain_weights_lightning
)
except ImportError as e:
# lightning is not available
pass
try:
from .sklearn_crfsuite import (
explain_weights_sklearn_crfsuite
)
except ImportError as e:
# sklearn-crfsuite is not available
pass
try:
from .xgboost import explain_weights_xgboost
except ImportError:
# xgboost is not available
pass
except Exception as e:
if e.__class__.__name__ == 'XGBoostLibraryNotFound':
# improperly installed xgboost
pass
else:
raise
| # ... existing code ...
pass
except Exception as e:
if e.__class__.__name__ == 'XGBoostLibraryNotFound':
# improperly installed xgboost
pass
else:
raise
# ... rest of the code ... |
1eacbac722ca949518e1a8e9d6a0a957e193ba9e | tests/functional/staging_and_prod/test_admin.py | tests/functional/staging_and_prod/test_admin.py | from retry.api import retry_call
from config import config
from tests.pages import UploadCsvPage
from tests.postman import (
send_notification_via_csv,
get_notification_by_id_via_api,
)
from tests.test_utils import assert_notification_body, recordtime, NotificationStatuses
@recordtime
def test_admin(driver, client, login_user):
upload_csv_page = UploadCsvPage(driver)
csv_sms_notification_id = send_notification_via_csv(upload_csv_page, 'sms')
csv_sms_notification = retry_call(
get_notification_by_id_via_api,
fargs=[client, csv_sms_notification_id, NotificationStatuses.SENT],
tries=config['notification_retry_times'],
delay=config['notification_retry_interval']
)
assert_notification_body(csv_sms_notification_id, csv_sms_notification)
csv_email_notification_id = send_notification_via_csv(upload_csv_page, 'email')
csv_email_notification = retry_call(
get_notification_by_id_via_api,
fargs=[client, csv_email_notification_id, NotificationStatuses.SENT],
tries=config['notification_retry_times'],
delay=config['notification_retry_interval']
)
assert_notification_body(csv_email_notification_id, csv_email_notification)
upload_csv_page.sign_out()
| import pytest
from retry.api import retry_call
from config import config
from tests.pages import UploadCsvPage
from tests.postman import (
send_notification_via_csv,
get_notification_by_id_via_api,
)
from tests.test_utils import assert_notification_body, recordtime, NotificationStatuses
@pytest.mark.skip(reason="intermittent pager duty alerts due to queue backlog")
def test_admin(driver, client, login_user):
upload_csv_page = UploadCsvPage(driver)
csv_sms_notification_id = send_notification_via_csv(upload_csv_page, 'sms')
csv_sms_notification = retry_call(
get_notification_by_id_via_api,
fargs=[client, csv_sms_notification_id, NotificationStatuses.SENT],
tries=config['notification_retry_times'],
delay=config['notification_retry_interval']
)
assert_notification_body(csv_sms_notification_id, csv_sms_notification)
csv_email_notification_id = send_notification_via_csv(upload_csv_page, 'email')
csv_email_notification = retry_call(
get_notification_by_id_via_api,
fargs=[client, csv_email_notification_id, NotificationStatuses.SENT],
tries=config['notification_retry_times'],
delay=config['notification_retry_interval']
)
assert_notification_body(csv_email_notification_id, csv_email_notification)
upload_csv_page.sign_out()
| Disable CSV upload tests temporarily | Disable CSV upload tests temporarily
When the database tasks queue builds up we get false pager duty alerts due to the time it takes for the test csv to get through to the front of the queue. | Python | mit | alphagov/notifications-functional-tests,alphagov/notifications-functional-tests | + import pytest
+
from retry.api import retry_call
from config import config
from tests.pages import UploadCsvPage
from tests.postman import (
send_notification_via_csv,
get_notification_by_id_via_api,
)
from tests.test_utils import assert_notification_body, recordtime, NotificationStatuses
- @recordtime
+ @pytest.mark.skip(reason="intermittent pager duty alerts due to queue backlog")
def test_admin(driver, client, login_user):
upload_csv_page = UploadCsvPage(driver)
csv_sms_notification_id = send_notification_via_csv(upload_csv_page, 'sms')
csv_sms_notification = retry_call(
get_notification_by_id_via_api,
fargs=[client, csv_sms_notification_id, NotificationStatuses.SENT],
tries=config['notification_retry_times'],
delay=config['notification_retry_interval']
)
assert_notification_body(csv_sms_notification_id, csv_sms_notification)
csv_email_notification_id = send_notification_via_csv(upload_csv_page, 'email')
csv_email_notification = retry_call(
get_notification_by_id_via_api,
fargs=[client, csv_email_notification_id, NotificationStatuses.SENT],
tries=config['notification_retry_times'],
delay=config['notification_retry_interval']
)
assert_notification_body(csv_email_notification_id, csv_email_notification)
upload_csv_page.sign_out()
| Disable CSV upload tests temporarily | ## Code Before:
from retry.api import retry_call
from config import config
from tests.pages import UploadCsvPage
from tests.postman import (
send_notification_via_csv,
get_notification_by_id_via_api,
)
from tests.test_utils import assert_notification_body, recordtime, NotificationStatuses
@recordtime
def test_admin(driver, client, login_user):
upload_csv_page = UploadCsvPage(driver)
csv_sms_notification_id = send_notification_via_csv(upload_csv_page, 'sms')
csv_sms_notification = retry_call(
get_notification_by_id_via_api,
fargs=[client, csv_sms_notification_id, NotificationStatuses.SENT],
tries=config['notification_retry_times'],
delay=config['notification_retry_interval']
)
assert_notification_body(csv_sms_notification_id, csv_sms_notification)
csv_email_notification_id = send_notification_via_csv(upload_csv_page, 'email')
csv_email_notification = retry_call(
get_notification_by_id_via_api,
fargs=[client, csv_email_notification_id, NotificationStatuses.SENT],
tries=config['notification_retry_times'],
delay=config['notification_retry_interval']
)
assert_notification_body(csv_email_notification_id, csv_email_notification)
upload_csv_page.sign_out()
## Instruction:
Disable CSV upload tests temporarily
## Code After:
import pytest
from retry.api import retry_call
from config import config
from tests.pages import UploadCsvPage
from tests.postman import (
send_notification_via_csv,
get_notification_by_id_via_api,
)
from tests.test_utils import assert_notification_body, recordtime, NotificationStatuses
@pytest.mark.skip(reason="intermittent pager duty alerts due to queue backlog")
def test_admin(driver, client, login_user):
upload_csv_page = UploadCsvPage(driver)
csv_sms_notification_id = send_notification_via_csv(upload_csv_page, 'sms')
csv_sms_notification = retry_call(
get_notification_by_id_via_api,
fargs=[client, csv_sms_notification_id, NotificationStatuses.SENT],
tries=config['notification_retry_times'],
delay=config['notification_retry_interval']
)
assert_notification_body(csv_sms_notification_id, csv_sms_notification)
csv_email_notification_id = send_notification_via_csv(upload_csv_page, 'email')
csv_email_notification = retry_call(
get_notification_by_id_via_api,
fargs=[client, csv_email_notification_id, NotificationStatuses.SENT],
tries=config['notification_retry_times'],
delay=config['notification_retry_interval']
)
assert_notification_body(csv_email_notification_id, csv_email_notification)
upload_csv_page.sign_out()
| ...
import pytest
from retry.api import retry_call
...
@pytest.mark.skip(reason="intermittent pager duty alerts due to queue backlog")
def test_admin(driver, client, login_user):
... |
775170d69862aaff63231b669639a872596ed2cd | test_interpreter.py | test_interpreter.py | import unittest
import brainfuck
test_cases = [("++++++++[>++++[>++>+++>+++>+<<<<-]>+>+>->>+[<]<-]>>.>---.+++++++..+++.>>.<-.<.+++.------.--------.>>+.>++.", "Hello World!\n")]
class InterpreterTestCase(unittest.TestCase):
def setUp(self):
self.interpreter = brainfuck.BrainfuckInterpreter()
def runTest(self):
for case in test_cases:
self.assertEqual(case[1], self.interpreter.eval(case[0]))
| import unittest
import brainfuck
hello_case = ("++++++++[>++++[>++>+++>+++>+<<<<-]>+>+>->>+[<]<-]>>.>---.+++++++..+++.>>.<-.<.+++.------.--------.>>+.>++.", "Hello World!\n")
class InterpreterTestCase(unittest.TestCase):
def setUp(self):
self.interpreter = brainfuck.BrainfuckInterpreter()
def test_hello_world(self):
self.assertEqual(hello_case[1], self.interpreter.eval(hello_case[0]))
def test_missing_parenthesis(self):
self.assertRaises(SyntaxError, self.interpreter.eval, '[++]+]')
| Add unittest for missing parenthesis | Add unittest for missing parenthesis
| Python | bsd-3-clause | handrake/brainfuck | import unittest
import brainfuck
- test_cases = [("++++++++[>++++[>++>+++>+++>+<<<<-]>+>+>->>+[<]<-]>>.>---.+++++++..+++.>>.<-.<.+++.------.--------.>>+.>++.", "Hello World!\n")]
+ hello_case = ("++++++++[>++++[>++>+++>+++>+<<<<-]>+>+>->>+[<]<-]>>.>---.+++++++..+++.>>.<-.<.+++.------.--------.>>+.>++.", "Hello World!\n")
class InterpreterTestCase(unittest.TestCase):
def setUp(self):
self.interpreter = brainfuck.BrainfuckInterpreter()
+ def test_hello_world(self):
- def runTest(self):
- for case in test_cases:
- self.assertEqual(case[1], self.interpreter.eval(case[0]))
+ self.assertEqual(hello_case[1], self.interpreter.eval(hello_case[0]))
+ def test_missing_parenthesis(self):
+ self.assertRaises(SyntaxError, self.interpreter.eval, '[++]+]')
| Add unittest for missing parenthesis | ## Code Before:
import unittest
import brainfuck
test_cases = [("++++++++[>++++[>++>+++>+++>+<<<<-]>+>+>->>+[<]<-]>>.>---.+++++++..+++.>>.<-.<.+++.------.--------.>>+.>++.", "Hello World!\n")]
class InterpreterTestCase(unittest.TestCase):
def setUp(self):
self.interpreter = brainfuck.BrainfuckInterpreter()
def runTest(self):
for case in test_cases:
self.assertEqual(case[1], self.interpreter.eval(case[0]))
## Instruction:
Add unittest for missing parenthesis
## Code After:
import unittest
import brainfuck
hello_case = ("++++++++[>++++[>++>+++>+++>+<<<<-]>+>+>->>+[<]<-]>>.>---.+++++++..+++.>>.<-.<.+++.------.--------.>>+.>++.", "Hello World!\n")
class InterpreterTestCase(unittest.TestCase):
def setUp(self):
self.interpreter = brainfuck.BrainfuckInterpreter()
def test_hello_world(self):
self.assertEqual(hello_case[1], self.interpreter.eval(hello_case[0]))
def test_missing_parenthesis(self):
self.assertRaises(SyntaxError, self.interpreter.eval, '[++]+]')
| ...
hello_case = ("++++++++[>++++[>++>+++>+++>+<<<<-]>+>+>->>+[<]<-]>>.>---.+++++++..+++.>>.<-.<.+++.------.--------.>>+.>++.", "Hello World!\n")
...
self.interpreter = brainfuck.BrainfuckInterpreter()
def test_hello_world(self):
self.assertEqual(hello_case[1], self.interpreter.eval(hello_case[0]))
def test_missing_parenthesis(self):
self.assertRaises(SyntaxError, self.interpreter.eval, '[++]+]')
... |
cd9cb42c16d443a26c7054e27c3ebc254142fbfb | python/ligscore/__init__.py | python/ligscore/__init__.py | import saliweb.backend
class Job(saliweb.backend.Job):
runnercls = saliweb.backend.SGERunner
def run(self):
libs = {'PoseScore': 'protein_ligand_pose_score.lib',
'RankScore': 'protein_ligand_rank_score.lib'}
pdb, mol2, lib = open('input.txt').readline().strip().split(' ')
lib = libs[lib]
script = """
module load imp
lib=`python -c "import IMP.atom; print IMP.atom.get_data_path('%s')"`
ligand_score %s %s "$lib" > score.list 2> score.log
""" % (lib, mol2, pdb)
r = self.runnercls(script)
r.set_sge_options('-l arch=linux-x64')
return r
def get_web_service(config_file):
db = saliweb.backend.Database(Job)
config = saliweb.backend.Config(config_file)
return saliweb.backend.WebService(config, db)
| import saliweb.backend
class Job(saliweb.backend.Job):
runnercls = saliweb.backend.WyntonSGERunner
def run(self):
libs = {'PoseScore': 'protein_ligand_pose_score.lib',
'RankScore': 'protein_ligand_rank_score.lib'}
pdb, mol2, lib = open('input.txt').readline().strip().split(' ')
lib = libs[lib]
script = """
module load Sali
module load imp
lib=`python -c "import IMP.atom; print IMP.atom.get_data_path('%s')"`
ligand_score %s %s "$lib" > score.list 2> score.log
""" % (lib, mol2, pdb)
r = self.runnercls(script)
r.set_sge_options('-l arch=lx-amd64')
return r
def get_web_service(config_file):
db = saliweb.backend.Database(Job)
config = saliweb.backend.Config(config_file)
return saliweb.backend.WebService(config, db)
| Switch to new Wynton cluster | Switch to new Wynton cluster
| Python | lgpl-2.1 | salilab/ligscore,salilab/ligscore | import saliweb.backend
class Job(saliweb.backend.Job):
- runnercls = saliweb.backend.SGERunner
+ runnercls = saliweb.backend.WyntonSGERunner
def run(self):
libs = {'PoseScore': 'protein_ligand_pose_score.lib',
'RankScore': 'protein_ligand_rank_score.lib'}
pdb, mol2, lib = open('input.txt').readline().strip().split(' ')
lib = libs[lib]
script = """
+ module load Sali
module load imp
lib=`python -c "import IMP.atom; print IMP.atom.get_data_path('%s')"`
ligand_score %s %s "$lib" > score.list 2> score.log
""" % (lib, mol2, pdb)
r = self.runnercls(script)
- r.set_sge_options('-l arch=linux-x64')
+ r.set_sge_options('-l arch=lx-amd64')
return r
def get_web_service(config_file):
db = saliweb.backend.Database(Job)
config = saliweb.backend.Config(config_file)
return saliweb.backend.WebService(config, db)
| Switch to new Wynton cluster | ## Code Before:
import saliweb.backend
class Job(saliweb.backend.Job):
runnercls = saliweb.backend.SGERunner
def run(self):
libs = {'PoseScore': 'protein_ligand_pose_score.lib',
'RankScore': 'protein_ligand_rank_score.lib'}
pdb, mol2, lib = open('input.txt').readline().strip().split(' ')
lib = libs[lib]
script = """
module load imp
lib=`python -c "import IMP.atom; print IMP.atom.get_data_path('%s')"`
ligand_score %s %s "$lib" > score.list 2> score.log
""" % (lib, mol2, pdb)
r = self.runnercls(script)
r.set_sge_options('-l arch=linux-x64')
return r
def get_web_service(config_file):
db = saliweb.backend.Database(Job)
config = saliweb.backend.Config(config_file)
return saliweb.backend.WebService(config, db)
## Instruction:
Switch to new Wynton cluster
## Code After:
import saliweb.backend
class Job(saliweb.backend.Job):
runnercls = saliweb.backend.WyntonSGERunner
def run(self):
libs = {'PoseScore': 'protein_ligand_pose_score.lib',
'RankScore': 'protein_ligand_rank_score.lib'}
pdb, mol2, lib = open('input.txt').readline().strip().split(' ')
lib = libs[lib]
script = """
module load Sali
module load imp
lib=`python -c "import IMP.atom; print IMP.atom.get_data_path('%s')"`
ligand_score %s %s "$lib" > score.list 2> score.log
""" % (lib, mol2, pdb)
r = self.runnercls(script)
r.set_sge_options('-l arch=lx-amd64')
return r
def get_web_service(config_file):
db = saliweb.backend.Database(Job)
config = saliweb.backend.Config(config_file)
return saliweb.backend.WebService(config, db)
| ...
class Job(saliweb.backend.Job):
runnercls = saliweb.backend.WyntonSGERunner
...
script = """
module load Sali
module load imp
...
r = self.runnercls(script)
r.set_sge_options('-l arch=lx-amd64')
return r
... |
c1ed5eb96b04ca0af2ad8f26023d8cbaa4a75eda | rx/concurrency/threadpoolscheduler.py | rx/concurrency/threadpoolscheduler.py | import logging
from concurrent.futures import ThreadPoolExecutor
from rx.core import Scheduler, Disposable
from rx.disposables import SingleAssignmentDisposable, CompositeDisposable
from .timeoutscheduler import TimeoutScheduler
log = logging.getLogger("Rx")
class ThreadPoolScheduler(TimeoutScheduler):
"""A scheduler that schedules work via the thread pool and threading
timers."""
def __init__(self, max_workers=None):
self.executor = ThreadPoolExecutor(max_workers=max_workers)
def schedule(self, action, state=None):
"""Schedules an action to be executed."""
disposable = SingleAssignmentDisposable()
def run():
disposable.disposable = self.invoke_action(action, state)
future = self.executor.submit(run)
def dispose():
future.cancel()
return CompositeDisposable(disposable, Disposable.create(dispose))
Scheduler.thread_pool = thread_pool_scheduler = ThreadPoolScheduler()
| from concurrent.futures import ThreadPoolExecutor
from rx.core import Scheduler
from .newthreadscheduler import NewThreadScheduler
class ThreadPoolScheduler(NewThreadScheduler):
"""A scheduler that schedules work via the thread pool."""
class ThreadPoolThread:
"""Wraps a concurrent future as a thread."""
def __init__(self, executor, run):
self.run = run
self.future = None
self.executor = executor
def start(self):
self.future = self.executor.submit(self.run)
def cancel(self):
self.future.cancel()
def __init__(self, max_workers=None):
super(ThreadPoolScheduler, self).__init__(self.thread_factory)
self.executor = ThreadPoolExecutor(max_workers=max_workers)
def thread_factory(self, target, *args):
return self.ThreadPoolThread(self.executor, target)
Scheduler.thread_pool = thread_pool_scheduler = ThreadPoolScheduler()
| Make thread pool scheduler behave as a pooled new thread scheduler | Make thread pool scheduler behave as a pooled new thread scheduler
| Python | mit | ReactiveX/RxPY,ReactiveX/RxPY | - import logging
from concurrent.futures import ThreadPoolExecutor
- from rx.core import Scheduler, Disposable
+ from rx.core import Scheduler
- from rx.disposables import SingleAssignmentDisposable, CompositeDisposable
- from .timeoutscheduler import TimeoutScheduler
+ from .newthreadscheduler import NewThreadScheduler
-
- log = logging.getLogger("Rx")
- class ThreadPoolScheduler(TimeoutScheduler):
+ class ThreadPoolScheduler(NewThreadScheduler):
- """A scheduler that schedules work via the thread pool and threading
+ """A scheduler that schedules work via the thread pool."""
- timers."""
+
+ class ThreadPoolThread:
+ """Wraps a concurrent future as a thread."""
+
+ def __init__(self, executor, run):
+ self.run = run
+ self.future = None
+ self.executor = executor
+
+ def start(self):
+ self.future = self.executor.submit(self.run)
+
+ def cancel(self):
+ self.future.cancel()
def __init__(self, max_workers=None):
+ super(ThreadPoolScheduler, self).__init__(self.thread_factory)
self.executor = ThreadPoolExecutor(max_workers=max_workers)
+ def thread_factory(self, target, *args):
+ return self.ThreadPoolThread(self.executor, target)
- def schedule(self, action, state=None):
- """Schedules an action to be executed."""
-
- disposable = SingleAssignmentDisposable()
-
- def run():
- disposable.disposable = self.invoke_action(action, state)
- future = self.executor.submit(run)
-
- def dispose():
- future.cancel()
- return CompositeDisposable(disposable, Disposable.create(dispose))
Scheduler.thread_pool = thread_pool_scheduler = ThreadPoolScheduler()
| Make thread pool scheduler behave as a pooled new thread scheduler | ## Code Before:
import logging
from concurrent.futures import ThreadPoolExecutor
from rx.core import Scheduler, Disposable
from rx.disposables import SingleAssignmentDisposable, CompositeDisposable
from .timeoutscheduler import TimeoutScheduler
log = logging.getLogger("Rx")
class ThreadPoolScheduler(TimeoutScheduler):
"""A scheduler that schedules work via the thread pool and threading
timers."""
def __init__(self, max_workers=None):
self.executor = ThreadPoolExecutor(max_workers=max_workers)
def schedule(self, action, state=None):
"""Schedules an action to be executed."""
disposable = SingleAssignmentDisposable()
def run():
disposable.disposable = self.invoke_action(action, state)
future = self.executor.submit(run)
def dispose():
future.cancel()
return CompositeDisposable(disposable, Disposable.create(dispose))
Scheduler.thread_pool = thread_pool_scheduler = ThreadPoolScheduler()
## Instruction:
Make thread pool scheduler behave as a pooled new thread scheduler
## Code After:
from concurrent.futures import ThreadPoolExecutor
from rx.core import Scheduler
from .newthreadscheduler import NewThreadScheduler
class ThreadPoolScheduler(NewThreadScheduler):
"""A scheduler that schedules work via the thread pool."""
class ThreadPoolThread:
"""Wraps a concurrent future as a thread."""
def __init__(self, executor, run):
self.run = run
self.future = None
self.executor = executor
def start(self):
self.future = self.executor.submit(self.run)
def cancel(self):
self.future.cancel()
def __init__(self, max_workers=None):
super(ThreadPoolScheduler, self).__init__(self.thread_factory)
self.executor = ThreadPoolExecutor(max_workers=max_workers)
def thread_factory(self, target, *args):
return self.ThreadPoolThread(self.executor, target)
Scheduler.thread_pool = thread_pool_scheduler = ThreadPoolScheduler()
| # ... existing code ...
from concurrent.futures import ThreadPoolExecutor
# ... modified code ...
from rx.core import Scheduler
from .newthreadscheduler import NewThreadScheduler
...
class ThreadPoolScheduler(NewThreadScheduler):
"""A scheduler that schedules work via the thread pool."""
class ThreadPoolThread:
"""Wraps a concurrent future as a thread."""
def __init__(self, executor, run):
self.run = run
self.future = None
self.executor = executor
def start(self):
self.future = self.executor.submit(self.run)
def cancel(self):
self.future.cancel()
...
def __init__(self, max_workers=None):
super(ThreadPoolScheduler, self).__init__(self.thread_factory)
self.executor = ThreadPoolExecutor(max_workers=max_workers)
...
def thread_factory(self, target, *args):
return self.ThreadPoolThread(self.executor, target)
# ... rest of the code ... |
faba2bc98f08cddea51d2e0093aa5c2981c8bf15 | gdrived.py | gdrived.py |
import sys
import time
import daemon
class GDriveDaemon(daemon.Daemon):
def run(self):
while True:
time.sleep(1)
|
import sys
import time
import daemon
UPDATE_INTERVAL = 30 # Sync update interval in seconds.
class GDriveDaemon(daemon.Daemon, object):
def __init__(self):
"Class constructor."
# Use pidfile in Gdrive config directory.
pidfile = None
# Use loglevel from GDrive config.
loglevel = None
# Use logfile in GDrive config directory.
stdout = None
super(GDriveDaemon, self).__init__(pidfile, loglevel, stdout)
def run(self):
"Run the daemon."
while True:
time.sleep(UPDATE_INTERVAL)
| Add update interval constant. Add detail to constructor. | Add update interval constant. Add detail to constructor.
| Python | apache-2.0 | babycaseny/gdrive-linux,jimlawton/gdrive-linux-googlecode,jimlawton/gdrive-linux,jmfield2/gdrive-linux |
import sys
import time
import daemon
+ UPDATE_INTERVAL = 30 # Sync update interval in seconds.
+
- class GDriveDaemon(daemon.Daemon):
+ class GDriveDaemon(daemon.Daemon, object):
+
+ def __init__(self):
+ "Class constructor."
+
+ # Use pidfile in Gdrive config directory.
+ pidfile = None
+
+ # Use loglevel from GDrive config.
+ loglevel = None
+
+ # Use logfile in GDrive config directory.
+ stdout = None
+
+ super(GDriveDaemon, self).__init__(pidfile, loglevel, stdout)
+
def run(self):
+ "Run the daemon."
+
while True:
+
- time.sleep(1)
+ time.sleep(UPDATE_INTERVAL)
| Add update interval constant. Add detail to constructor. | ## Code Before:
import sys
import time
import daemon
class GDriveDaemon(daemon.Daemon):
def run(self):
while True:
time.sleep(1)
## Instruction:
Add update interval constant. Add detail to constructor.
## Code After:
import sys
import time
import daemon
UPDATE_INTERVAL = 30 # Sync update interval in seconds.
class GDriveDaemon(daemon.Daemon, object):
def __init__(self):
"Class constructor."
# Use pidfile in Gdrive config directory.
pidfile = None
# Use loglevel from GDrive config.
loglevel = None
# Use logfile in GDrive config directory.
stdout = None
super(GDriveDaemon, self).__init__(pidfile, loglevel, stdout)
def run(self):
"Run the daemon."
while True:
time.sleep(UPDATE_INTERVAL)
| ...
UPDATE_INTERVAL = 30 # Sync update interval in seconds.
class GDriveDaemon(daemon.Daemon, object):
def __init__(self):
"Class constructor."
# Use pidfile in Gdrive config directory.
pidfile = None
# Use loglevel from GDrive config.
loglevel = None
# Use logfile in GDrive config directory.
stdout = None
super(GDriveDaemon, self).__init__(pidfile, loglevel, stdout)
def run(self):
"Run the daemon."
while True:
time.sleep(UPDATE_INTERVAL)
... |
e86901ac2b074d42d2e388353bbe60fcdd8f0240 | wagtail/contrib/postgres_search/apps.py | wagtail/contrib/postgres_search/apps.py | from django.apps import AppConfig
from django.core.checks import Error, Tags, register
from .utils import get_postgresql_connections, set_weights
class PostgresSearchConfig(AppConfig):
name = 'wagtail.contrib.postgres_search'
def ready(self):
@register(Tags.compatibility, Tags.database)
def check_if_postgresql(app_configs, **kwargs):
if get_postgresql_connections():
return []
return [Error('You must use a PostgreSQL database '
'to use PostgreSQL search.',
id='wagtail.contrib.postgres_search.E001')]
set_weights()
from .models import IndexEntry
IndexEntry.add_generic_relations()
| from django.apps import AppConfig
from django.core.checks import Error, Tags, register
from .utils import get_postgresql_connections, set_weights
class PostgresSearchConfig(AppConfig):
name = 'wagtail.contrib.postgres_search'
default_auto_field = 'django.db.models.AutoField'
def ready(self):
@register(Tags.compatibility, Tags.database)
def check_if_postgresql(app_configs, **kwargs):
if get_postgresql_connections():
return []
return [Error('You must use a PostgreSQL database '
'to use PostgreSQL search.',
id='wagtail.contrib.postgres_search.E001')]
set_weights()
from .models import IndexEntry
IndexEntry.add_generic_relations()
| Set default_auto_field in wagtail.contrib.postgres_search AppConfig | Set default_auto_field in wagtail.contrib.postgres_search AppConfig
Add default_auto_field = 'django.db.models.AutoField'
Co-authored-by: Nick Moreton <[email protected]> | Python | bsd-3-clause | jnns/wagtail,zerolab/wagtail,gasman/wagtail,gasman/wagtail,gasman/wagtail,rsalmaso/wagtail,rsalmaso/wagtail,thenewguy/wagtail,thenewguy/wagtail,jnns/wagtail,jnns/wagtail,rsalmaso/wagtail,wagtail/wagtail,mixxorz/wagtail,torchbox/wagtail,jnns/wagtail,gasman/wagtail,thenewguy/wagtail,thenewguy/wagtail,wagtail/wagtail,mixxorz/wagtail,zerolab/wagtail,torchbox/wagtail,wagtail/wagtail,mixxorz/wagtail,mixxorz/wagtail,zerolab/wagtail,torchbox/wagtail,mixxorz/wagtail,rsalmaso/wagtail,wagtail/wagtail,wagtail/wagtail,gasman/wagtail,zerolab/wagtail,thenewguy/wagtail,zerolab/wagtail,torchbox/wagtail,rsalmaso/wagtail | from django.apps import AppConfig
from django.core.checks import Error, Tags, register
from .utils import get_postgresql_connections, set_weights
class PostgresSearchConfig(AppConfig):
name = 'wagtail.contrib.postgres_search'
+ default_auto_field = 'django.db.models.AutoField'
def ready(self):
@register(Tags.compatibility, Tags.database)
def check_if_postgresql(app_configs, **kwargs):
if get_postgresql_connections():
return []
return [Error('You must use a PostgreSQL database '
'to use PostgreSQL search.',
id='wagtail.contrib.postgres_search.E001')]
set_weights()
from .models import IndexEntry
IndexEntry.add_generic_relations()
| Set default_auto_field in wagtail.contrib.postgres_search AppConfig | ## Code Before:
from django.apps import AppConfig
from django.core.checks import Error, Tags, register
from .utils import get_postgresql_connections, set_weights
class PostgresSearchConfig(AppConfig):
name = 'wagtail.contrib.postgres_search'
def ready(self):
@register(Tags.compatibility, Tags.database)
def check_if_postgresql(app_configs, **kwargs):
if get_postgresql_connections():
return []
return [Error('You must use a PostgreSQL database '
'to use PostgreSQL search.',
id='wagtail.contrib.postgres_search.E001')]
set_weights()
from .models import IndexEntry
IndexEntry.add_generic_relations()
## Instruction:
Set default_auto_field in wagtail.contrib.postgres_search AppConfig
## Code After:
from django.apps import AppConfig
from django.core.checks import Error, Tags, register
from .utils import get_postgresql_connections, set_weights
class PostgresSearchConfig(AppConfig):
name = 'wagtail.contrib.postgres_search'
default_auto_field = 'django.db.models.AutoField'
def ready(self):
@register(Tags.compatibility, Tags.database)
def check_if_postgresql(app_configs, **kwargs):
if get_postgresql_connections():
return []
return [Error('You must use a PostgreSQL database '
'to use PostgreSQL search.',
id='wagtail.contrib.postgres_search.E001')]
set_weights()
from .models import IndexEntry
IndexEntry.add_generic_relations()
| ...
name = 'wagtail.contrib.postgres_search'
default_auto_field = 'django.db.models.AutoField'
... |
5decd7e68c6454e455bc1debe232ea37f7260c58 | mixins.py | mixins.py | class DepthSerializerMixin(object):
"""Custom method 'get_serializer_class', set attribute 'depth' based on query parameter in the url"""
def get_serializer_class(self):
serializer_class = self.serializer_class
query_params = self.request.QUERY_PARAMS
depth = query_params.get('__depth', None)
serializer_class.Meta.depth = int(depth) if(depth != None and depth.isdigit()) else 0
return serializer_class
| class DepthSerializerMixin(object):
"""Custom method 'get_serializer_class', set attribute 'depth' based on query parameter in the url"""
def get_serializer_class(self, *args, **kwargs):
serializer_class = super(DepthSerializerMixin, self).get_serializer_class(*args, **kwargs)
query_params = self.request.QUERY_PARAMS
depth = query_params.get('__depth', None)
serializer_class.Meta.depth = int(depth) if(depth != None and depth.isdigit()) else 0
return serializer_class
| Call method 'get_serializer_class' of the Class parent | Call method 'get_serializer_class' of the Class parent
| Python | mit | krescruz/depth-serializer-mixin | class DepthSerializerMixin(object):
"""Custom method 'get_serializer_class', set attribute 'depth' based on query parameter in the url"""
- def get_serializer_class(self):
+ def get_serializer_class(self, *args, **kwargs):
- serializer_class = self.serializer_class
+ serializer_class = super(DepthSerializerMixin, self).get_serializer_class(*args, **kwargs)
query_params = self.request.QUERY_PARAMS
depth = query_params.get('__depth', None)
serializer_class.Meta.depth = int(depth) if(depth != None and depth.isdigit()) else 0
return serializer_class
| Call method 'get_serializer_class' of the Class parent | ## Code Before:
class DepthSerializerMixin(object):
"""Custom method 'get_serializer_class', set attribute 'depth' based on query parameter in the url"""
def get_serializer_class(self):
serializer_class = self.serializer_class
query_params = self.request.QUERY_PARAMS
depth = query_params.get('__depth', None)
serializer_class.Meta.depth = int(depth) if(depth != None and depth.isdigit()) else 0
return serializer_class
## Instruction:
Call method 'get_serializer_class' of the Class parent
## Code After:
class DepthSerializerMixin(object):
"""Custom method 'get_serializer_class', set attribute 'depth' based on query parameter in the url"""
def get_serializer_class(self, *args, **kwargs):
serializer_class = super(DepthSerializerMixin, self).get_serializer_class(*args, **kwargs)
query_params = self.request.QUERY_PARAMS
depth = query_params.get('__depth', None)
serializer_class.Meta.depth = int(depth) if(depth != None and depth.isdigit()) else 0
return serializer_class
| ...
def get_serializer_class(self, *args, **kwargs):
serializer_class = super(DepthSerializerMixin, self).get_serializer_class(*args, **kwargs)
query_params = self.request.QUERY_PARAMS
... |
593bab981f36f7af52ae55914c18e368e8c1a94f | examples/app-on-ws-init.py | examples/app-on-ws-init.py |
from argparse import ArgumentParser
import i3ipc
i3 = i3ipc.Connection()
parser = ArgumentParser(description='Open an application on a given workspace when it is initialized')
parser.add_argument('--workspace', metavar='NAME', help='The name of the workspace')
parser.add_argument('--command', metavar='CMD', help='The command to run on the newly initted workspace')
args = parser.parse_args()
def on_workspace(i3, e):
if e.current.props.name == args.workspace and not len(e.current.leaves()):
i3.command('exec {}'.format(args.command))
i3.on('workspace::focus', on_workspace)
i3.main()
|
from argparse import ArgumentParser
import i3ipc
i3 = i3ipc.Connection()
parser = ArgumentParser(description="""Open the given application each time the
given workspace is created. For instance, running 'app-on-ws-init.py 6
i3-sensible-terminal' should open your terminal as soon as you create the
workspace 6.
""")
parser.add_argument('workspace', metavar='WS_NAME', help='The name of the workspace')
parser.add_argument('command', metavar='CMD', help='The command to run on the newly initted workspace')
args = parser.parse_args()
def on_workspace(i3, e):
if e.current.props.name == args.workspace and not len(e.current.leaves()):
i3.command('exec {}'.format(args.command))
i3.on('workspace::focus', on_workspace)
i3.main()
| Make the 2 mandatory parameters mandatory. Make the help message a bit clearer and provides an example. | Make the 2 mandatory parameters mandatory.
Make the help message a bit clearer and provides an example.
| Python | bsd-3-clause | xenomachina/i3ipc-python,nicoe/i3ipc-python,acrisci/i3ipc-python,chrsclmn/i3ipc-python |
from argparse import ArgumentParser
import i3ipc
i3 = i3ipc.Connection()
- parser = ArgumentParser(description='Open an application on a given workspace when it is initialized')
+ parser = ArgumentParser(description="""Open the given application each time the
+ given workspace is created. For instance, running 'app-on-ws-init.py 6
+ i3-sensible-terminal' should open your terminal as soon as you create the
+ workspace 6.
+ """)
- parser.add_argument('--workspace', metavar='NAME', help='The name of the workspace')
+ parser.add_argument('workspace', metavar='WS_NAME', help='The name of the workspace')
- parser.add_argument('--command', metavar='CMD', help='The command to run on the newly initted workspace')
+ parser.add_argument('command', metavar='CMD', help='The command to run on the newly initted workspace')
args = parser.parse_args()
def on_workspace(i3, e):
if e.current.props.name == args.workspace and not len(e.current.leaves()):
i3.command('exec {}'.format(args.command))
i3.on('workspace::focus', on_workspace)
i3.main()
| Make the 2 mandatory parameters mandatory. Make the help message a bit clearer and provides an example. | ## Code Before:
from argparse import ArgumentParser
import i3ipc
i3 = i3ipc.Connection()
parser = ArgumentParser(description='Open an application on a given workspace when it is initialized')
parser.add_argument('--workspace', metavar='NAME', help='The name of the workspace')
parser.add_argument('--command', metavar='CMD', help='The command to run on the newly initted workspace')
args = parser.parse_args()
def on_workspace(i3, e):
if e.current.props.name == args.workspace and not len(e.current.leaves()):
i3.command('exec {}'.format(args.command))
i3.on('workspace::focus', on_workspace)
i3.main()
## Instruction:
Make the 2 mandatory parameters mandatory. Make the help message a bit clearer and provides an example.
## Code After:
from argparse import ArgumentParser
import i3ipc
i3 = i3ipc.Connection()
parser = ArgumentParser(description="""Open the given application each time the
given workspace is created. For instance, running 'app-on-ws-init.py 6
i3-sensible-terminal' should open your terminal as soon as you create the
workspace 6.
""")
parser.add_argument('workspace', metavar='WS_NAME', help='The name of the workspace')
parser.add_argument('command', metavar='CMD', help='The command to run on the newly initted workspace')
args = parser.parse_args()
def on_workspace(i3, e):
if e.current.props.name == args.workspace and not len(e.current.leaves()):
i3.command('exec {}'.format(args.command))
i3.on('workspace::focus', on_workspace)
i3.main()
| ...
parser = ArgumentParser(description="""Open the given application each time the
given workspace is created. For instance, running 'app-on-ws-init.py 6
i3-sensible-terminal' should open your terminal as soon as you create the
workspace 6.
""")
parser.add_argument('workspace', metavar='WS_NAME', help='The name of the workspace')
parser.add_argument('command', metavar='CMD', help='The command to run on the newly initted workspace')
... |
0b8b32a044e92f4e996af734d44a2d93d1492684 | project_code/bulk_fitting.py | project_code/bulk_fitting.py |
'''
Bulk spectral line fitting with SDSS galaxy spectra
'''
import os
from astropy.io import fits
from pandas import concat
# Bring in the package funcs
from specfit import do_specfit
from download_spectra import download_spectra
def bulk_fit(obs_file, output_file, keep_spectra=False):
'''
Downloads files based off of the entries in the given file, performs
spectral line fitting and saves the results to a FITS table.
'''
# Open the file
data_file = fits.open(obs_file)
spectra_data = data_file[1].data
del data_file
num_spectra = spectra_data['Z'].shape[0]
for i in range(num_spectra):
spec_info = spectra_data[i]
# Download the spectrum
spec_name = \
download_spectra(spec_info['PLATEID'], spec_info['FIBREID'],
spec_info['MJD'], spec_info['SURVEY'])
spec_df = do_specfit(spec_name, verbose=False)
if i == 0:
df = spec_df
else:
df = concat([df, spec_df])
if not keep_spectra:
os.system('rm ' + spec_name)
df.write(output_file)
return
|
'''
Bulk spectral line fitting with SDSS galaxy spectra
'''
import os
from astropy.io import fits
from pandas import DataFrame
# Bring in the package funcs
from specfit import do_specfit
from download_spectra import download_spectra
def bulk_fit(obs_file, output_file, keep_spectra=False):
'''
Downloads files based off of the entries in the given file, performs
spectral line fitting and saves the results to a FITS table.
'''
# Open the file
data_file = fits.open(obs_file)
spectra_data = data_file[1].data
del data_file
num_spectra = spectra_data['Z'].shape[0]
for i in range(num_spectra):
spec_info = spectra_data[i]
# Download the spectrum
spec_name = \
download_spectra(spec_info['PLATE'], spec_info['FIBERID'],
spec_info['MJD'], spec_info['SURVEY'])
spec_df = do_specfit(spec_name, verbose=False)
if i == 0:
df = DataFrame(spec_df, columns=[spec_name[:-5]])
else:
df[spec_name[:-5]] = spec_df
if not keep_spectra:
os.system('rm ' + spec_name)
df.to_csv(output_file)
return
| Correct names, concat dataframes properly | Correct names, concat dataframes properly
| Python | mit | e-koch/Phys-595 |
'''
Bulk spectral line fitting with SDSS galaxy spectra
'''
import os
from astropy.io import fits
- from pandas import concat
+ from pandas import DataFrame
# Bring in the package funcs
from specfit import do_specfit
from download_spectra import download_spectra
def bulk_fit(obs_file, output_file, keep_spectra=False):
'''
Downloads files based off of the entries in the given file, performs
spectral line fitting and saves the results to a FITS table.
'''
# Open the file
data_file = fits.open(obs_file)
spectra_data = data_file[1].data
del data_file
num_spectra = spectra_data['Z'].shape[0]
for i in range(num_spectra):
spec_info = spectra_data[i]
# Download the spectrum
spec_name = \
- download_spectra(spec_info['PLATEID'], spec_info['FIBREID'],
+ download_spectra(spec_info['PLATE'], spec_info['FIBERID'],
spec_info['MJD'], spec_info['SURVEY'])
spec_df = do_specfit(spec_name, verbose=False)
if i == 0:
- df = spec_df
+ df = DataFrame(spec_df, columns=[spec_name[:-5]])
else:
- df = concat([df, spec_df])
+ df[spec_name[:-5]] = spec_df
if not keep_spectra:
os.system('rm ' + spec_name)
- df.write(output_file)
+ df.to_csv(output_file)
return
| Correct names, concat dataframes properly | ## Code Before:
'''
Bulk spectral line fitting with SDSS galaxy spectra
'''
import os
from astropy.io import fits
from pandas import concat
# Bring in the package funcs
from specfit import do_specfit
from download_spectra import download_spectra
def bulk_fit(obs_file, output_file, keep_spectra=False):
'''
Downloads files based off of the entries in the given file, performs
spectral line fitting and saves the results to a FITS table.
'''
# Open the file
data_file = fits.open(obs_file)
spectra_data = data_file[1].data
del data_file
num_spectra = spectra_data['Z'].shape[0]
for i in range(num_spectra):
spec_info = spectra_data[i]
# Download the spectrum
spec_name = \
download_spectra(spec_info['PLATEID'], spec_info['FIBREID'],
spec_info['MJD'], spec_info['SURVEY'])
spec_df = do_specfit(spec_name, verbose=False)
if i == 0:
df = spec_df
else:
df = concat([df, spec_df])
if not keep_spectra:
os.system('rm ' + spec_name)
df.write(output_file)
return
## Instruction:
Correct names, concat dataframes properly
## Code After:
'''
Bulk spectral line fitting with SDSS galaxy spectra
'''
import os
from astropy.io import fits
from pandas import DataFrame
# Bring in the package funcs
from specfit import do_specfit
from download_spectra import download_spectra
def bulk_fit(obs_file, output_file, keep_spectra=False):
'''
Downloads files based off of the entries in the given file, performs
spectral line fitting and saves the results to a FITS table.
'''
# Open the file
data_file = fits.open(obs_file)
spectra_data = data_file[1].data
del data_file
num_spectra = spectra_data['Z'].shape[0]
for i in range(num_spectra):
spec_info = spectra_data[i]
# Download the spectrum
spec_name = \
download_spectra(spec_info['PLATE'], spec_info['FIBERID'],
spec_info['MJD'], spec_info['SURVEY'])
spec_df = do_specfit(spec_name, verbose=False)
if i == 0:
df = DataFrame(spec_df, columns=[spec_name[:-5]])
else:
df[spec_name[:-5]] = spec_df
if not keep_spectra:
os.system('rm ' + spec_name)
df.to_csv(output_file)
return
| # ... existing code ...
from astropy.io import fits
from pandas import DataFrame
# ... modified code ...
spec_name = \
download_spectra(spec_info['PLATE'], spec_info['FIBERID'],
spec_info['MJD'], spec_info['SURVEY'])
...
if i == 0:
df = DataFrame(spec_df, columns=[spec_name[:-5]])
else:
df[spec_name[:-5]] = spec_df
...
df.to_csv(output_file)
# ... rest of the code ... |
f622e11536c4ebf8f82985329d06efc58c2fe60e | blog/tests/test_views.py | blog/tests/test_views.py | from django.test import TestCase
class BlogViewsTestCase(TestCase):
def setUp(self):
| from django import test
from django.core.urlresolvers import reverse
from blog.models import Post, Category
class BlogViewsTestCase(test.TestCase):
def setUp(self):
# Add parent category and post category
parent = Category(name='Writing', parent=None)
parent.save()
category = Category(name='Thoughts', parent=parent)
category.save()
# Create a draft
_post = Post(title='Random thoughts of the author',
body='Thoughts turned to words.', category=category)
_post.save()
self.draft = _post
# Publish a post
post = Post(title='New thoughts from without',
body='A post fit to be published!', category=category)
post.save()
post.publish()
self.post = post
self.client = test.Client()
def test_index(self):
response = self.client.get('/')
self.assertEqual(response.status_code, 200)
posts = response.context['posts']
self.assertNotIn(self.draft, posts)
self.assertIn(self.post, posts)
def test_post_view(self):
post_url = reverse('blog:post', kwargs=dict(pid=self.post.id,
slug=self.post.slug))
response = self.client.get(post_url)
self.assertEqual(response.status_code, 200)
post = response.context['post']
posts = response.context['posts']
self.assertEqual(post, self.post)
self.assertEqual(posts.count(), 0)
def test_draft_view(self):
draft_url = reverse('blog:post', kwargs=dict(pid=self.draft.id,
slug=self.draft.slug))
response = self.client.get(draft_url)
self.assertEqual(response.status_code, 404)
| Add tests for blog index view and post view | Add tests for blog index view and post view
| Python | mit | ajoyoommen/weblog,ajoyoommen/weblog | - from django.test import TestCase
+ from django import test
+ from django.core.urlresolvers import reverse
+
+ from blog.models import Post, Category
- class BlogViewsTestCase(TestCase):
+ class BlogViewsTestCase(test.TestCase):
def setUp(self):
+ # Add parent category and post category
+ parent = Category(name='Writing', parent=None)
+ parent.save()
+ category = Category(name='Thoughts', parent=parent)
+ category.save()
+ # Create a draft
+ _post = Post(title='Random thoughts of the author',
+ body='Thoughts turned to words.', category=category)
+ _post.save()
+ self.draft = _post
+ # Publish a post
+ post = Post(title='New thoughts from without',
+ body='A post fit to be published!', category=category)
+ post.save()
+ post.publish()
+ self.post = post
+ self.client = test.Client()
+
+ def test_index(self):
+ response = self.client.get('/')
+ self.assertEqual(response.status_code, 200)
+ posts = response.context['posts']
+ self.assertNotIn(self.draft, posts)
+ self.assertIn(self.post, posts)
+
+ def test_post_view(self):
+ post_url = reverse('blog:post', kwargs=dict(pid=self.post.id,
+ slug=self.post.slug))
+ response = self.client.get(post_url)
+ self.assertEqual(response.status_code, 200)
+ post = response.context['post']
+ posts = response.context['posts']
+ self.assertEqual(post, self.post)
+ self.assertEqual(posts.count(), 0)
+
+ def test_draft_view(self):
+ draft_url = reverse('blog:post', kwargs=dict(pid=self.draft.id,
+ slug=self.draft.slug))
+ response = self.client.get(draft_url)
+ self.assertEqual(response.status_code, 404)
+ | Add tests for blog index view and post view | ## Code Before:
from django.test import TestCase
class BlogViewsTestCase(TestCase):
def setUp(self):
## Instruction:
Add tests for blog index view and post view
## Code After:
from django import test
from django.core.urlresolvers import reverse
from blog.models import Post, Category
class BlogViewsTestCase(test.TestCase):
def setUp(self):
# Add parent category and post category
parent = Category(name='Writing', parent=None)
parent.save()
category = Category(name='Thoughts', parent=parent)
category.save()
# Create a draft
_post = Post(title='Random thoughts of the author',
body='Thoughts turned to words.', category=category)
_post.save()
self.draft = _post
# Publish a post
post = Post(title='New thoughts from without',
body='A post fit to be published!', category=category)
post.save()
post.publish()
self.post = post
self.client = test.Client()
def test_index(self):
response = self.client.get('/')
self.assertEqual(response.status_code, 200)
posts = response.context['posts']
self.assertNotIn(self.draft, posts)
self.assertIn(self.post, posts)
def test_post_view(self):
post_url = reverse('blog:post', kwargs=dict(pid=self.post.id,
slug=self.post.slug))
response = self.client.get(post_url)
self.assertEqual(response.status_code, 200)
post = response.context['post']
posts = response.context['posts']
self.assertEqual(post, self.post)
self.assertEqual(posts.count(), 0)
def test_draft_view(self):
draft_url = reverse('blog:post', kwargs=dict(pid=self.draft.id,
slug=self.draft.slug))
response = self.client.get(draft_url)
self.assertEqual(response.status_code, 404)
| // ... existing code ...
from django import test
from django.core.urlresolvers import reverse
from blog.models import Post, Category
// ... modified code ...
class BlogViewsTestCase(test.TestCase):
def setUp(self):
# Add parent category and post category
parent = Category(name='Writing', parent=None)
parent.save()
category = Category(name='Thoughts', parent=parent)
category.save()
# Create a draft
_post = Post(title='Random thoughts of the author',
body='Thoughts turned to words.', category=category)
_post.save()
self.draft = _post
# Publish a post
post = Post(title='New thoughts from without',
body='A post fit to be published!', category=category)
post.save()
post.publish()
self.post = post
self.client = test.Client()
def test_index(self):
response = self.client.get('/')
self.assertEqual(response.status_code, 200)
posts = response.context['posts']
self.assertNotIn(self.draft, posts)
self.assertIn(self.post, posts)
def test_post_view(self):
post_url = reverse('blog:post', kwargs=dict(pid=self.post.id,
slug=self.post.slug))
response = self.client.get(post_url)
self.assertEqual(response.status_code, 200)
post = response.context['post']
posts = response.context['posts']
self.assertEqual(post, self.post)
self.assertEqual(posts.count(), 0)
def test_draft_view(self):
draft_url = reverse('blog:post', kwargs=dict(pid=self.draft.id,
slug=self.draft.slug))
response = self.client.get(draft_url)
self.assertEqual(response.status_code, 404)
// ... rest of the code ... |
edf151feea948ebf4a9f00a0248ab1f363cacfac | scaffolder/commands/install.py | scaffolder/commands/install.py |
from optparse import make_option
from optparse import OptionParser
from scaffolder import get_minion_path
from scaffolder.core.template import TemplateManager
from scaffolder.core.commands import BaseCommand
class InstallCommand(BaseCommand):
option_list = BaseCommand.option_list + (
make_option(
"-t",
"--target",
dest="target_dir",
default=get_minion_path('weaver'),
help='Project Templates directory.',
metavar="TEMPLATES_DIR"
),
)
def __init__(self, name, help='', aliases=(), stdout=None, stderr=None):
help = 'install: Installs a Project Template.'
parser = OptionParser(
version=self.get_version(),
option_list=self.get_option_list(),
usage='\n %prog {0} ACTION [OPTIONS]'.format(name)
)
aliases = ('tmp',)
BaseCommand.__init__(self, name, parser=parser, help=help, aliases=aliases)
def run(self, *args, **options):
src = args[0]
tgt = options.get('target_dir')
manager = TemplateManager()
manager.install(src=src, dest=tgt)
|
from optparse import make_option
from optparse import OptionParser
from scaffolder import get_minion_path
from scaffolder.core.template import TemplateManager
from scaffolder.core.commands import BaseCommand
class InstallCommand(BaseCommand):
option_list = BaseCommand.option_list + (
make_option(
"-t",
"--target",
dest="target_dir",
default=get_minion_path('weaver'),
help='Project Templates directory.',
metavar="TEMPLATES_DIR"
),
)
help = 'Installs a Project Template.'
def run(self, *args, **options):
src = args[0]
tgt = options.get('target_dir')
manager = TemplateManager()
manager.install(src=src, dest=tgt)
| Remove __init__ method, not needed. | InstallCommand: Remove __init__ method, not needed.
| Python | mit | goliatone/minions |
from optparse import make_option
from optparse import OptionParser
from scaffolder import get_minion_path
from scaffolder.core.template import TemplateManager
from scaffolder.core.commands import BaseCommand
class InstallCommand(BaseCommand):
option_list = BaseCommand.option_list + (
make_option(
"-t",
"--target",
dest="target_dir",
default=get_minion_path('weaver'),
help='Project Templates directory.',
metavar="TEMPLATES_DIR"
),
)
- def __init__(self, name, help='', aliases=(), stdout=None, stderr=None):
- help = 'install: Installs a Project Template.'
+ help = 'Installs a Project Template.'
+
- parser = OptionParser(
- version=self.get_version(),
- option_list=self.get_option_list(),
- usage='\n %prog {0} ACTION [OPTIONS]'.format(name)
- )
- aliases = ('tmp',)
- BaseCommand.__init__(self, name, parser=parser, help=help, aliases=aliases)
def run(self, *args, **options):
src = args[0]
tgt = options.get('target_dir')
manager = TemplateManager()
manager.install(src=src, dest=tgt)
| Remove __init__ method, not needed. | ## Code Before:
from optparse import make_option
from optparse import OptionParser
from scaffolder import get_minion_path
from scaffolder.core.template import TemplateManager
from scaffolder.core.commands import BaseCommand
class InstallCommand(BaseCommand):
option_list = BaseCommand.option_list + (
make_option(
"-t",
"--target",
dest="target_dir",
default=get_minion_path('weaver'),
help='Project Templates directory.',
metavar="TEMPLATES_DIR"
),
)
def __init__(self, name, help='', aliases=(), stdout=None, stderr=None):
help = 'install: Installs a Project Template.'
parser = OptionParser(
version=self.get_version(),
option_list=self.get_option_list(),
usage='\n %prog {0} ACTION [OPTIONS]'.format(name)
)
aliases = ('tmp',)
BaseCommand.__init__(self, name, parser=parser, help=help, aliases=aliases)
def run(self, *args, **options):
src = args[0]
tgt = options.get('target_dir')
manager = TemplateManager()
manager.install(src=src, dest=tgt)
## Instruction:
Remove __init__ method, not needed.
## Code After:
from optparse import make_option
from optparse import OptionParser
from scaffolder import get_minion_path
from scaffolder.core.template import TemplateManager
from scaffolder.core.commands import BaseCommand
class InstallCommand(BaseCommand):
option_list = BaseCommand.option_list + (
make_option(
"-t",
"--target",
dest="target_dir",
default=get_minion_path('weaver'),
help='Project Templates directory.',
metavar="TEMPLATES_DIR"
),
)
help = 'Installs a Project Template.'
def run(self, *args, **options):
src = args[0]
tgt = options.get('target_dir')
manager = TemplateManager()
manager.install(src=src, dest=tgt)
| // ... existing code ...
help = 'Installs a Project Template.'
// ... rest of the code ... |
796f9ff27f579557237c48196eb50d40269c6840 | glitch/__main__.py | glitch/__main__.py | from . import config
from . import apikeys
import argparse
import logging
parser = argparse.ArgumentParser(description="Invoke the Infinite Glitch server(s)")
parser.add_argument("server", help="Server to invoke", choices=["main", "renderer"], nargs="?", default="main")
parser.add_argument("-l", "--log", help="Logging level", type=lambda x: x.upper(),
choices=logging._nameToLevel, # NAUGHTY
default="INFO")
parser.add_argument("--dev", help="Dev mode (no logins)", action='store_true')
arguments = parser.parse_args()
log = logging.getLogger(__name__)
logging.basicConfig(level=getattr(logging, arguments.log), format='%(asctime)s:%(levelname)s:%(name)s:%(message)s')
if arguments.server == "renderer":
from . import renderer
renderer.run() # doesn't return
else:
from . import server
server.run(disable_logins=arguments.dev) # doesn't return
| from . import config
from . import apikeys
import argparse
# Hack: Allow "python -m glitch database" to be the same as "glitch.database"
import sys
if len(sys.argv) > 1 and sys.argv[1] == "database":
from . import database
import clize
sys.exit(clize.run(*database.commands, args=sys.argv[1:]))
import logging
parser = argparse.ArgumentParser(description="Invoke the Infinite Glitch server(s)")
parser.add_argument("server", help="Server to invoke", choices=["main", "renderer"], nargs="?", default="main")
parser.add_argument("-l", "--log", help="Logging level", type=lambda x: x.upper(),
choices=logging._nameToLevel, # NAUGHTY
default="INFO")
parser.add_argument("--dev", help="Dev mode (no logins)", action='store_true')
arguments = parser.parse_args()
log = logging.getLogger(__name__)
logging.basicConfig(level=getattr(logging, arguments.log), format='%(asctime)s:%(levelname)s:%(name)s:%(message)s')
if arguments.server == "renderer":
from . import renderer
renderer.run() # doesn't return
else:
from . import server
server.run(disable_logins=arguments.dev) # doesn't return
| Allow 'python -m glitch database' as well as with a dot | Allow 'python -m glitch database' as well as with a dot
| Python | artistic-2.0 | MikeiLL/appension,MikeiLL/appension,Rosuav/appension,Rosuav/appension,MikeiLL/appension,MikeiLL/appension,Rosuav/appension,Rosuav/appension | from . import config
from . import apikeys
import argparse
+
+ # Hack: Allow "python -m glitch database" to be the same as "glitch.database"
+ import sys
+ if len(sys.argv) > 1 and sys.argv[1] == "database":
+ from . import database
+ import clize
+ sys.exit(clize.run(*database.commands, args=sys.argv[1:]))
import logging
parser = argparse.ArgumentParser(description="Invoke the Infinite Glitch server(s)")
parser.add_argument("server", help="Server to invoke", choices=["main", "renderer"], nargs="?", default="main")
parser.add_argument("-l", "--log", help="Logging level", type=lambda x: x.upper(),
choices=logging._nameToLevel, # NAUGHTY
default="INFO")
parser.add_argument("--dev", help="Dev mode (no logins)", action='store_true')
arguments = parser.parse_args()
log = logging.getLogger(__name__)
logging.basicConfig(level=getattr(logging, arguments.log), format='%(asctime)s:%(levelname)s:%(name)s:%(message)s')
if arguments.server == "renderer":
from . import renderer
renderer.run() # doesn't return
else:
from . import server
server.run(disable_logins=arguments.dev) # doesn't return
| Allow 'python -m glitch database' as well as with a dot | ## Code Before:
from . import config
from . import apikeys
import argparse
import logging
parser = argparse.ArgumentParser(description="Invoke the Infinite Glitch server(s)")
parser.add_argument("server", help="Server to invoke", choices=["main", "renderer"], nargs="?", default="main")
parser.add_argument("-l", "--log", help="Logging level", type=lambda x: x.upper(),
choices=logging._nameToLevel, # NAUGHTY
default="INFO")
parser.add_argument("--dev", help="Dev mode (no logins)", action='store_true')
arguments = parser.parse_args()
log = logging.getLogger(__name__)
logging.basicConfig(level=getattr(logging, arguments.log), format='%(asctime)s:%(levelname)s:%(name)s:%(message)s')
if arguments.server == "renderer":
from . import renderer
renderer.run() # doesn't return
else:
from . import server
server.run(disable_logins=arguments.dev) # doesn't return
## Instruction:
Allow 'python -m glitch database' as well as with a dot
## Code After:
from . import config
from . import apikeys
import argparse
# Hack: Allow "python -m glitch database" to be the same as "glitch.database"
import sys
if len(sys.argv) > 1 and sys.argv[1] == "database":
from . import database
import clize
sys.exit(clize.run(*database.commands, args=sys.argv[1:]))
import logging
parser = argparse.ArgumentParser(description="Invoke the Infinite Glitch server(s)")
parser.add_argument("server", help="Server to invoke", choices=["main", "renderer"], nargs="?", default="main")
parser.add_argument("-l", "--log", help="Logging level", type=lambda x: x.upper(),
choices=logging._nameToLevel, # NAUGHTY
default="INFO")
parser.add_argument("--dev", help="Dev mode (no logins)", action='store_true')
arguments = parser.parse_args()
log = logging.getLogger(__name__)
logging.basicConfig(level=getattr(logging, arguments.log), format='%(asctime)s:%(levelname)s:%(name)s:%(message)s')
if arguments.server == "renderer":
from . import renderer
renderer.run() # doesn't return
else:
from . import server
server.run(disable_logins=arguments.dev) # doesn't return
| ...
import argparse
# Hack: Allow "python -m glitch database" to be the same as "glitch.database"
import sys
if len(sys.argv) > 1 and sys.argv[1] == "database":
from . import database
import clize
sys.exit(clize.run(*database.commands, args=sys.argv[1:]))
... |
4b93e5aa8c0ce90189fb852e75ee213d3be0d01a | flicks/base/urls.py | flicks/base/urls.py | from django.conf.urls.defaults import patterns, url
from flicks.base import views
urlpatterns = patterns('',
url(r'^/?$', views.home, name='flicks.base.home'),
url(r'^strings/?$', views.strings, name='flicks.base.strings'),
)
| from django.conf.urls.defaults import patterns, url
from flicks.base import views
urlpatterns = patterns('',
url(r'^/?$', views.home, name='flicks.base.home'),
url(r'^faq/?$', views.faq, name='flicks.base.faq'),
url(r'^strings/?$', views.strings, name='flicks.base.strings'),
)
| Add back in FAQ url that was removed accidentally. | Add back in FAQ url that was removed accidentally.
| Python | bsd-3-clause | mozilla/firefox-flicks,mozilla/firefox-flicks,mozilla/firefox-flicks,mozilla/firefox-flicks | from django.conf.urls.defaults import patterns, url
from flicks.base import views
urlpatterns = patterns('',
url(r'^/?$', views.home, name='flicks.base.home'),
+ url(r'^faq/?$', views.faq, name='flicks.base.faq'),
url(r'^strings/?$', views.strings, name='flicks.base.strings'),
)
| Add back in FAQ url that was removed accidentally. | ## Code Before:
from django.conf.urls.defaults import patterns, url
from flicks.base import views
urlpatterns = patterns('',
url(r'^/?$', views.home, name='flicks.base.home'),
url(r'^strings/?$', views.strings, name='flicks.base.strings'),
)
## Instruction:
Add back in FAQ url that was removed accidentally.
## Code After:
from django.conf.urls.defaults import patterns, url
from flicks.base import views
urlpatterns = patterns('',
url(r'^/?$', views.home, name='flicks.base.home'),
url(r'^faq/?$', views.faq, name='flicks.base.faq'),
url(r'^strings/?$', views.strings, name='flicks.base.strings'),
)
| # ... existing code ...
url(r'^/?$', views.home, name='flicks.base.home'),
url(r'^faq/?$', views.faq, name='flicks.base.faq'),
url(r'^strings/?$', views.strings, name='flicks.base.strings'),
# ... rest of the code ... |
9df3f3a2d0660b8e8166aa944bf45f261a51d987 | ies_base/serializers.py | ies_base/serializers.py | from rest_framework import serializers
class Tag(object):
def __init__(self, name, related_tags, equivalent_names):
self.name = name
self.related_tags = related_tags
self.equivalent_names = equivalent_names
class TagSerializer(serializers.Serializer):
name = serializers.CharField()
related_tags = serializers.ListField(child=serializers.CharField(), allow_null=True)
equivalent_names = serializers.ListField(child=serializers.CharField(), allow_null=True)
def create(self, validated_data):
return Tag(**validated_data)
class Followable(object):
def __init__(self, name, type, game, object_id, thumbnail_url="", **kwargs):
self.name = name
self.type = type
self.game = game
self.object_id = object_id
self.thumbnail_url = thumbnail_url
for field, value in kwargs.items():
setattr(self, field, value)
class FollowableSerializer(serializers.Serializer):
name = serializers.CharField()
type = serializers.IntegerField()
game = serializers.IntegerField()
object_id = serializers.IntegerField()
default_color = serializers.CharField(allow_null=True, allow_blank=True)
thumbnail_url = serializers.CharField(allow_blank=True, allow_null=True)
def create(self, validated_data):
return Followable(**validated_data)
| from rest_framework import serializers
class Tag(object):
def __init__(self, name, related_tags, equivalent_names):
self.name = name
self.related_tags = related_tags
self.equivalent_names = equivalent_names
class TagSerializer(serializers.Serializer):
name = serializers.CharField()
related_tags = serializers.ListField(child=serializers.CharField(), allow_null=True)
equivalent_names = serializers.ListField(child=serializers.CharField(), allow_null=True)
def create(self, validated_data):
return Tag(**validated_data)
class Followable(object):
def __init__(self, name, type, game, object_id, thumbnail_url="", **kwargs):
self.name = name
self.type = type
self.game = game
self.object_id = object_id
self.thumbnail_url = thumbnail_url
for field, value in kwargs.items():
setattr(self, field, value)
class FollowableSerializer(serializers.Serializer):
name = serializers.CharField()
type = serializers.IntegerField()
game = serializers.IntegerField()
object_id = serializers.IntegerField()
default_color = serializers.CharField(allow_null=True, allow_blank=True, required=False)
thumbnail_url = serializers.CharField(allow_blank=True, allow_null=True)
def create(self, validated_data):
return Followable(**validated_data)
| Make default color not required | Make default color not required
| Python | mit | InstanteSports/ies-django-base | from rest_framework import serializers
class Tag(object):
def __init__(self, name, related_tags, equivalent_names):
self.name = name
self.related_tags = related_tags
self.equivalent_names = equivalent_names
class TagSerializer(serializers.Serializer):
name = serializers.CharField()
related_tags = serializers.ListField(child=serializers.CharField(), allow_null=True)
equivalent_names = serializers.ListField(child=serializers.CharField(), allow_null=True)
def create(self, validated_data):
return Tag(**validated_data)
class Followable(object):
def __init__(self, name, type, game, object_id, thumbnail_url="", **kwargs):
self.name = name
self.type = type
self.game = game
self.object_id = object_id
self.thumbnail_url = thumbnail_url
for field, value in kwargs.items():
setattr(self, field, value)
class FollowableSerializer(serializers.Serializer):
name = serializers.CharField()
type = serializers.IntegerField()
game = serializers.IntegerField()
object_id = serializers.IntegerField()
- default_color = serializers.CharField(allow_null=True, allow_blank=True)
+ default_color = serializers.CharField(allow_null=True, allow_blank=True, required=False)
thumbnail_url = serializers.CharField(allow_blank=True, allow_null=True)
def create(self, validated_data):
return Followable(**validated_data)
| Make default color not required | ## Code Before:
from rest_framework import serializers
class Tag(object):
def __init__(self, name, related_tags, equivalent_names):
self.name = name
self.related_tags = related_tags
self.equivalent_names = equivalent_names
class TagSerializer(serializers.Serializer):
name = serializers.CharField()
related_tags = serializers.ListField(child=serializers.CharField(), allow_null=True)
equivalent_names = serializers.ListField(child=serializers.CharField(), allow_null=True)
def create(self, validated_data):
return Tag(**validated_data)
class Followable(object):
def __init__(self, name, type, game, object_id, thumbnail_url="", **kwargs):
self.name = name
self.type = type
self.game = game
self.object_id = object_id
self.thumbnail_url = thumbnail_url
for field, value in kwargs.items():
setattr(self, field, value)
class FollowableSerializer(serializers.Serializer):
name = serializers.CharField()
type = serializers.IntegerField()
game = serializers.IntegerField()
object_id = serializers.IntegerField()
default_color = serializers.CharField(allow_null=True, allow_blank=True)
thumbnail_url = serializers.CharField(allow_blank=True, allow_null=True)
def create(self, validated_data):
return Followable(**validated_data)
## Instruction:
Make default color not required
## Code After:
from rest_framework import serializers
class Tag(object):
def __init__(self, name, related_tags, equivalent_names):
self.name = name
self.related_tags = related_tags
self.equivalent_names = equivalent_names
class TagSerializer(serializers.Serializer):
name = serializers.CharField()
related_tags = serializers.ListField(child=serializers.CharField(), allow_null=True)
equivalent_names = serializers.ListField(child=serializers.CharField(), allow_null=True)
def create(self, validated_data):
return Tag(**validated_data)
class Followable(object):
def __init__(self, name, type, game, object_id, thumbnail_url="", **kwargs):
self.name = name
self.type = type
self.game = game
self.object_id = object_id
self.thumbnail_url = thumbnail_url
for field, value in kwargs.items():
setattr(self, field, value)
class FollowableSerializer(serializers.Serializer):
name = serializers.CharField()
type = serializers.IntegerField()
game = serializers.IntegerField()
object_id = serializers.IntegerField()
default_color = serializers.CharField(allow_null=True, allow_blank=True, required=False)
thumbnail_url = serializers.CharField(allow_blank=True, allow_null=True)
def create(self, validated_data):
return Followable(**validated_data)
| // ... existing code ...
object_id = serializers.IntegerField()
default_color = serializers.CharField(allow_null=True, allow_blank=True, required=False)
thumbnail_url = serializers.CharField(allow_blank=True, allow_null=True)
// ... rest of the code ... |
7519bebe1d9d87930275858a537dcc0a0a64f007 | tools/strip_filenames.py | tools/strip_filenames.py | import os
directory = os.listdir()
illegal_characters = "%?_'*+$!\""
tolowercase=True
for a in range(len(directory)):
newname=""
for c in directory[a]:
if c in illegal_characters:
continue
if c.isalnum() or c == '.':
newname=newname+c.lower()
print("convert {} to {}".format(directory[a],newname))
os.rename(directory[a], newname)
|
import sys
import os
from docopt import docopt
# docopt(doc, argv=None, help=True, version=None, options_first=False))
def main():
opt = docopt(__doc__, sys.argv[1:])
directory = opt.get("filename", os.listdir())
legal_characters = ""
list_N010 = list(range(size))
list_alpha = [ chr(x+97) for x in range(26) ]
list_ALPHA = [ chr(x+65) for x in range(26) ]
legal_characters += "".join(list_N010)
legal_characters += "".join(list_alpha)
if not opt.get("--lowercase", False):
legal_characters += "".join(list_N010)
for a in range(len(directory)):
newname=""
for c in directory[a]:
if c not in legal_characters:
continue
newname += c
print("convert {} to {}".format(directory[a],newname))
os.rename(directory[a], newname)
if __name__ == "__main__":
main()
| Use legal characters for stripping filenames | Use legal characters for stripping filenames
| Python | mit | dgengtek/scripts,dgengtek/scripts | +
+ import sys
import os
+ from docopt import docopt
- directory = os.listdir()
- illegal_characters = "%?_'*+$!\""
- tolowercase=True
- for a in range(len(directory)):
- newname=""
- for c in directory[a]:
- if c in illegal_characters:
- continue
- if c.isalnum() or c == '.':
- newname=newname+c.lower()
- print("convert {} to {}".format(directory[a],newname))
- os.rename(directory[a], newname)
+ # docopt(doc, argv=None, help=True, version=None, options_first=False))
+
+ def main():
+ opt = docopt(__doc__, sys.argv[1:])
+ directory = opt.get("filename", os.listdir())
+ legal_characters = ""
+ list_N010 = list(range(size))
+ list_alpha = [ chr(x+97) for x in range(26) ]
+ list_ALPHA = [ chr(x+65) for x in range(26) ]
+
+ legal_characters += "".join(list_N010)
+ legal_characters += "".join(list_alpha)
+ if not opt.get("--lowercase", False):
+ legal_characters += "".join(list_N010)
+ for a in range(len(directory)):
+ newname=""
+ for c in directory[a]:
+ if c not in legal_characters:
+ continue
+ newname += c
+ print("convert {} to {}".format(directory[a],newname))
+ os.rename(directory[a], newname)
+
+ if __name__ == "__main__":
+ main()
+ | Use legal characters for stripping filenames | ## Code Before:
import os
directory = os.listdir()
illegal_characters = "%?_'*+$!\""
tolowercase=True
for a in range(len(directory)):
newname=""
for c in directory[a]:
if c in illegal_characters:
continue
if c.isalnum() or c == '.':
newname=newname+c.lower()
print("convert {} to {}".format(directory[a],newname))
os.rename(directory[a], newname)
## Instruction:
Use legal characters for stripping filenames
## Code After:
import sys
import os
from docopt import docopt
# docopt(doc, argv=None, help=True, version=None, options_first=False))
def main():
opt = docopt(__doc__, sys.argv[1:])
directory = opt.get("filename", os.listdir())
legal_characters = ""
list_N010 = list(range(size))
list_alpha = [ chr(x+97) for x in range(26) ]
list_ALPHA = [ chr(x+65) for x in range(26) ]
legal_characters += "".join(list_N010)
legal_characters += "".join(list_alpha)
if not opt.get("--lowercase", False):
legal_characters += "".join(list_N010)
for a in range(len(directory)):
newname=""
for c in directory[a]:
if c not in legal_characters:
continue
newname += c
print("convert {} to {}".format(directory[a],newname))
os.rename(directory[a], newname)
if __name__ == "__main__":
main()
| // ... existing code ...
import sys
import os
from docopt import docopt
# docopt(doc, argv=None, help=True, version=None, options_first=False))
def main():
opt = docopt(__doc__, sys.argv[1:])
directory = opt.get("filename", os.listdir())
legal_characters = ""
list_N010 = list(range(size))
list_alpha = [ chr(x+97) for x in range(26) ]
list_ALPHA = [ chr(x+65) for x in range(26) ]
legal_characters += "".join(list_N010)
legal_characters += "".join(list_alpha)
if not opt.get("--lowercase", False):
legal_characters += "".join(list_N010)
for a in range(len(directory)):
newname=""
for c in directory[a]:
if c not in legal_characters:
continue
newname += c
print("convert {} to {}".format(directory[a],newname))
os.rename(directory[a], newname)
if __name__ == "__main__":
main()
// ... rest of the code ... |
4f8aed6ed3491e62911619eaa9aa4b86b30065e4 | leonardo/module/leonardo_auth/widget/userlogin/models.py | leonardo/module/leonardo_auth/widget/userlogin/models.py |
from django.db import models
from django.utils.translation import ugettext_lazy as _
from leonardo.module.web.models import Widget
LOGIN_TYPE_CHOICES = (
(1, _("Admin")),
(2, _("Public")),
)
class UserLoginWidget(Widget):
type = models.PositiveIntegerField(verbose_name=_(
"type"), choices=LOGIN_TYPE_CHOICES, default=2)
def get_context_data(self, request):
context = super(UserLoginWidget, self).get_context_data(request)
if 'next' in request.GET:
context['next'] = request.GET['next']
return context
class Meta:
abstract = True
verbose_name = _("user login")
verbose_name_plural = _("user logins")
|
from django.db import models
from django.utils.translation import ugettext_lazy as _
from leonardo.module.web.models import Widget
LOGIN_TYPE_CHOICES = (
(1, _("Admin")),
(2, _("Public")),
)
class UserLoginWidget(Widget):
type = models.PositiveIntegerField(verbose_name=_(
"type"), choices=LOGIN_TYPE_CHOICES, default=2)
def get_context_data(self, request):
context = super(UserLoginWidget, self).get_context_data(request)
if 'next' in request.GET:
context['next'] = request.GET['next']
else:
context['next'] = request.path
return context
class Meta:
abstract = True
verbose_name = _("user login")
verbose_name_plural = _("user logins")
| Fix missing next in context. | Fix missing next in context.
| Python | bsd-3-clause | django-leonardo/django-leonardo,django-leonardo/django-leonardo,django-leonardo/django-leonardo,django-leonardo/django-leonardo |
from django.db import models
from django.utils.translation import ugettext_lazy as _
from leonardo.module.web.models import Widget
LOGIN_TYPE_CHOICES = (
(1, _("Admin")),
(2, _("Public")),
)
class UserLoginWidget(Widget):
type = models.PositiveIntegerField(verbose_name=_(
"type"), choices=LOGIN_TYPE_CHOICES, default=2)
def get_context_data(self, request):
context = super(UserLoginWidget, self).get_context_data(request)
if 'next' in request.GET:
context['next'] = request.GET['next']
+ else:
+ context['next'] = request.path
return context
class Meta:
abstract = True
verbose_name = _("user login")
verbose_name_plural = _("user logins")
| Fix missing next in context. | ## Code Before:
from django.db import models
from django.utils.translation import ugettext_lazy as _
from leonardo.module.web.models import Widget
LOGIN_TYPE_CHOICES = (
(1, _("Admin")),
(2, _("Public")),
)
class UserLoginWidget(Widget):
type = models.PositiveIntegerField(verbose_name=_(
"type"), choices=LOGIN_TYPE_CHOICES, default=2)
def get_context_data(self, request):
context = super(UserLoginWidget, self).get_context_data(request)
if 'next' in request.GET:
context['next'] = request.GET['next']
return context
class Meta:
abstract = True
verbose_name = _("user login")
verbose_name_plural = _("user logins")
## Instruction:
Fix missing next in context.
## Code After:
from django.db import models
from django.utils.translation import ugettext_lazy as _
from leonardo.module.web.models import Widget
LOGIN_TYPE_CHOICES = (
(1, _("Admin")),
(2, _("Public")),
)
class UserLoginWidget(Widget):
type = models.PositiveIntegerField(verbose_name=_(
"type"), choices=LOGIN_TYPE_CHOICES, default=2)
def get_context_data(self, request):
context = super(UserLoginWidget, self).get_context_data(request)
if 'next' in request.GET:
context['next'] = request.GET['next']
else:
context['next'] = request.path
return context
class Meta:
abstract = True
verbose_name = _("user login")
verbose_name_plural = _("user logins")
| ...
context['next'] = request.GET['next']
else:
context['next'] = request.path
... |
b6ab579fa65f816704142716fbd68645ac5f2ff8 | zenaida/contrib/feedback/models.py | zenaida/contrib/feedback/models.py | from django.conf import settings
from django.db import models
class FeedbackItem(models.Model):
timestamp = models.DateTimeField(auto_now_add=True)
user = models.ForeignKey(settings.AUTH_USER_MODEL)
resolved = models.BooleanField(default=False)
content = models.TextField()
screenshot = models.FileField(blank=True, null=True, upload_to="feedback/screenshots")
# Request Data
view = models.CharField(max_length=255)
request_path = models.CharField(max_length=255)
# The longest methods should be 7 chars, but we'll allow custom methods up
# to 20 chars just in case.
request_method = models.CharField(max_length=20, blank=True, null=True)
# How long is the longest encoding name?
request_encoding = models.CharField(max_length=20, blank=True, null=True)
request_meta = models.TextField(blank=True, null=True)
request_get = models.TextField(blank=True, null=True)
request_post = models.TextField(blank=True, null=True)
request_files = models.TextField(blank=True, null=True)
def __unicode__(self):
return "{username} at {path}".format(
username=self.user.get_full_name(),
path = self.request_path
)
| from django.conf import settings
from django.db import models
class FeedbackItem(models.Model):
timestamp = models.DateTimeField(auto_now_add=True)
user = models.ForeignKey(settings.AUTH_USER_MODEL)
resolved = models.BooleanField(default=False)
content = models.TextField()
screenshot = models.FileField(blank=True, null=True, upload_to="feedback/screenshots")
# Request Data
view = models.CharField(max_length=255)
request_path = models.CharField(max_length=255)
# The longest methods should be 7 chars, but we'll allow custom methods up
# to 20 chars just in case.
request_method = models.CharField(max_length=20, blank=True, null=True)
# How long is the longest encoding name?
request_encoding = models.CharField(max_length=20, blank=True, null=True)
request_meta = models.TextField(blank=True, null=True)
request_get = models.TextField(blank=True, null=True)
request_post = models.TextField(blank=True, null=True)
request_files = models.TextField(blank=True, null=True)
def __unicode__(self):
return "{username} at {path}".format(
username=self.user.get_full_name(),
path = self.request_path
)
class Meta:
ordering = ["-timestamp"]
| Order feedback items by their timestamp. | Order feedback items by their timestamp.
| Python | bsd-3-clause | littleweaver/django-zenaida,littleweaver/django-zenaida,littleweaver/django-zenaida,littleweaver/django-zenaida | from django.conf import settings
from django.db import models
class FeedbackItem(models.Model):
timestamp = models.DateTimeField(auto_now_add=True)
user = models.ForeignKey(settings.AUTH_USER_MODEL)
resolved = models.BooleanField(default=False)
content = models.TextField()
screenshot = models.FileField(blank=True, null=True, upload_to="feedback/screenshots")
# Request Data
view = models.CharField(max_length=255)
request_path = models.CharField(max_length=255)
# The longest methods should be 7 chars, but we'll allow custom methods up
# to 20 chars just in case.
request_method = models.CharField(max_length=20, blank=True, null=True)
# How long is the longest encoding name?
request_encoding = models.CharField(max_length=20, blank=True, null=True)
request_meta = models.TextField(blank=True, null=True)
request_get = models.TextField(blank=True, null=True)
request_post = models.TextField(blank=True, null=True)
request_files = models.TextField(blank=True, null=True)
def __unicode__(self):
return "{username} at {path}".format(
username=self.user.get_full_name(),
path = self.request_path
)
+ class Meta:
+ ordering = ["-timestamp"]
+ | Order feedback items by their timestamp. | ## Code Before:
from django.conf import settings
from django.db import models
class FeedbackItem(models.Model):
timestamp = models.DateTimeField(auto_now_add=True)
user = models.ForeignKey(settings.AUTH_USER_MODEL)
resolved = models.BooleanField(default=False)
content = models.TextField()
screenshot = models.FileField(blank=True, null=True, upload_to="feedback/screenshots")
# Request Data
view = models.CharField(max_length=255)
request_path = models.CharField(max_length=255)
# The longest methods should be 7 chars, but we'll allow custom methods up
# to 20 chars just in case.
request_method = models.CharField(max_length=20, blank=True, null=True)
# How long is the longest encoding name?
request_encoding = models.CharField(max_length=20, blank=True, null=True)
request_meta = models.TextField(blank=True, null=True)
request_get = models.TextField(blank=True, null=True)
request_post = models.TextField(blank=True, null=True)
request_files = models.TextField(blank=True, null=True)
def __unicode__(self):
return "{username} at {path}".format(
username=self.user.get_full_name(),
path = self.request_path
)
## Instruction:
Order feedback items by their timestamp.
## Code After:
from django.conf import settings
from django.db import models
class FeedbackItem(models.Model):
timestamp = models.DateTimeField(auto_now_add=True)
user = models.ForeignKey(settings.AUTH_USER_MODEL)
resolved = models.BooleanField(default=False)
content = models.TextField()
screenshot = models.FileField(blank=True, null=True, upload_to="feedback/screenshots")
# Request Data
view = models.CharField(max_length=255)
request_path = models.CharField(max_length=255)
# The longest methods should be 7 chars, but we'll allow custom methods up
# to 20 chars just in case.
request_method = models.CharField(max_length=20, blank=True, null=True)
# How long is the longest encoding name?
request_encoding = models.CharField(max_length=20, blank=True, null=True)
request_meta = models.TextField(blank=True, null=True)
request_get = models.TextField(blank=True, null=True)
request_post = models.TextField(blank=True, null=True)
request_files = models.TextField(blank=True, null=True)
def __unicode__(self):
return "{username} at {path}".format(
username=self.user.get_full_name(),
path = self.request_path
)
class Meta:
ordering = ["-timestamp"]
| ...
)
class Meta:
ordering = ["-timestamp"]
... |
29cc59bc478c4c6bc936141d19a3386468ff8f07 | tests/test_general_attributes.py | tests/test_general_attributes.py | from jawa.attribute import get_attribute_classes
def test_mandatory_attributes():
for parser_class in get_attribute_classes().values():
assert hasattr(parser_class, 'ADDED_IN'), (
'Attribute parser missing mandatory ADDED_IN property'
)
assert hasattr(parser_class, 'MINIMUM_CLASS_VERSION'), (
'Attribute parser missing mandatory MINIMUM_CLASS_VERSION '
'property'
)
| from jawa.attribute import get_attribute_classes
def test_mandatory_attributes():
required_properities = ['ADDED_IN', 'MINIMUM_CLASS_VERSION']
for name, class_ in get_attribute_classes().items():
for p in required_properities:
assert hasattr(class_, p), (
'{name} parser missing mandatory {p} property'.format(
name=name,
p=p
)
)
def test_attribute_naming():
for name, class_ in get_attribute_classes().items():
if hasattr(class_, 'ATTRIBUTE_NAME'):
continue
assert class_.__name__.endswith('Attribute'), (
'{name} parser does not follow naming convention and does'
' not explicity set it.'.format(name=name)
)
| Add a simple test for Attribuet class naming conventions. | Add a simple test for Attribuet class naming conventions.
| Python | mit | TkTech/Jawa,TkTech/Jawa | from jawa.attribute import get_attribute_classes
def test_mandatory_attributes():
+ required_properities = ['ADDED_IN', 'MINIMUM_CLASS_VERSION']
- for parser_class in get_attribute_classes().values():
+ for name, class_ in get_attribute_classes().items():
- assert hasattr(parser_class, 'ADDED_IN'), (
- 'Attribute parser missing mandatory ADDED_IN property'
+ for p in required_properities:
+ assert hasattr(class_, p), (
+ '{name} parser missing mandatory {p} property'.format(
+ name=name,
+ p=p
+ )
- )
+ )
- assert hasattr(parser_class, 'MINIMUM_CLASS_VERSION'), (
- 'Attribute parser missing mandatory MINIMUM_CLASS_VERSION '
- 'property'
+
+
+ def test_attribute_naming():
+ for name, class_ in get_attribute_classes().items():
+ if hasattr(class_, 'ATTRIBUTE_NAME'):
+ continue
+
+ assert class_.__name__.endswith('Attribute'), (
+ '{name} parser does not follow naming convention and does'
+ ' not explicity set it.'.format(name=name)
)
| Add a simple test for Attribuet class naming conventions. | ## Code Before:
from jawa.attribute import get_attribute_classes
def test_mandatory_attributes():
for parser_class in get_attribute_classes().values():
assert hasattr(parser_class, 'ADDED_IN'), (
'Attribute parser missing mandatory ADDED_IN property'
)
assert hasattr(parser_class, 'MINIMUM_CLASS_VERSION'), (
'Attribute parser missing mandatory MINIMUM_CLASS_VERSION '
'property'
)
## Instruction:
Add a simple test for Attribuet class naming conventions.
## Code After:
from jawa.attribute import get_attribute_classes
def test_mandatory_attributes():
required_properities = ['ADDED_IN', 'MINIMUM_CLASS_VERSION']
for name, class_ in get_attribute_classes().items():
for p in required_properities:
assert hasattr(class_, p), (
'{name} parser missing mandatory {p} property'.format(
name=name,
p=p
)
)
def test_attribute_naming():
for name, class_ in get_attribute_classes().items():
if hasattr(class_, 'ATTRIBUTE_NAME'):
continue
assert class_.__name__.endswith('Attribute'), (
'{name} parser does not follow naming convention and does'
' not explicity set it.'.format(name=name)
)
| # ... existing code ...
def test_mandatory_attributes():
required_properities = ['ADDED_IN', 'MINIMUM_CLASS_VERSION']
for name, class_ in get_attribute_classes().items():
for p in required_properities:
assert hasattr(class_, p), (
'{name} parser missing mandatory {p} property'.format(
name=name,
p=p
)
)
def test_attribute_naming():
for name, class_ in get_attribute_classes().items():
if hasattr(class_, 'ATTRIBUTE_NAME'):
continue
assert class_.__name__.endswith('Attribute'), (
'{name} parser does not follow naming convention and does'
' not explicity set it.'.format(name=name)
)
# ... rest of the code ... |
0a3eb4b966dff69cbe582c60bf4444facb4b683d | tcconfig/_tc_command_helper.py | tcconfig/_tc_command_helper.py |
from __future__ import absolute_import, unicode_literals
import subprocrunner as spr
from ._common import find_bin_path
from ._const import TcSubCommand
from ._error import NetworkInterfaceNotFoundError
def get_tc_base_command(tc_subcommand):
if tc_subcommand not in TcSubCommand:
raise ValueError("the argument must be a TcSubCommand value")
return "{:s} {:s}".format(find_bin_path("tc"), tc_subcommand.value)
def run_tc_show(subcommand, device, tc_command_output):
from ._network import verify_network_interface
verify_network_interface(device, tc_command_output)
runner = spr.SubprocessRunner(
"{:s} show dev {:s}".format(get_tc_base_command(subcommand), device)
)
if runner.run() != 0 and runner.stderr.find("Cannot find device") != -1:
# reach here if the device does not exist at the system and netiface
# not installed.
raise NetworkInterfaceNotFoundError(target=device)
return runner.stdout
|
from __future__ import absolute_import, unicode_literals
import subprocrunner as spr
from ._common import find_bin_path
from ._const import TcSubCommand
from ._error import NetworkInterfaceNotFoundError
def get_tc_base_command(tc_subcommand):
if not isinstance(tc_subcommand, TcSubCommand):
raise ValueError("the argument must be a TcSubCommand value")
return "{:s} {:s}".format(find_bin_path("tc"), tc_subcommand.value)
def run_tc_show(subcommand, device, tc_command_output):
from ._network import verify_network_interface
verify_network_interface(device, tc_command_output)
runner = spr.SubprocessRunner(
"{:s} show dev {:s}".format(get_tc_base_command(subcommand), device)
)
if runner.run() != 0 and runner.stderr.find("Cannot find device") != -1:
# reach here if the device does not exist at the system and netiface
# not installed.
raise NetworkInterfaceNotFoundError(target=device)
return runner.stdout
| Change to avoid a DeprecationWarning | Change to avoid a DeprecationWarning
| Python | mit | thombashi/tcconfig,thombashi/tcconfig |
from __future__ import absolute_import, unicode_literals
import subprocrunner as spr
from ._common import find_bin_path
from ._const import TcSubCommand
from ._error import NetworkInterfaceNotFoundError
def get_tc_base_command(tc_subcommand):
- if tc_subcommand not in TcSubCommand:
+ if not isinstance(tc_subcommand, TcSubCommand):
raise ValueError("the argument must be a TcSubCommand value")
return "{:s} {:s}".format(find_bin_path("tc"), tc_subcommand.value)
def run_tc_show(subcommand, device, tc_command_output):
from ._network import verify_network_interface
verify_network_interface(device, tc_command_output)
runner = spr.SubprocessRunner(
"{:s} show dev {:s}".format(get_tc_base_command(subcommand), device)
)
if runner.run() != 0 and runner.stderr.find("Cannot find device") != -1:
# reach here if the device does not exist at the system and netiface
# not installed.
raise NetworkInterfaceNotFoundError(target=device)
return runner.stdout
| Change to avoid a DeprecationWarning | ## Code Before:
from __future__ import absolute_import, unicode_literals
import subprocrunner as spr
from ._common import find_bin_path
from ._const import TcSubCommand
from ._error import NetworkInterfaceNotFoundError
def get_tc_base_command(tc_subcommand):
if tc_subcommand not in TcSubCommand:
raise ValueError("the argument must be a TcSubCommand value")
return "{:s} {:s}".format(find_bin_path("tc"), tc_subcommand.value)
def run_tc_show(subcommand, device, tc_command_output):
from ._network import verify_network_interface
verify_network_interface(device, tc_command_output)
runner = spr.SubprocessRunner(
"{:s} show dev {:s}".format(get_tc_base_command(subcommand), device)
)
if runner.run() != 0 and runner.stderr.find("Cannot find device") != -1:
# reach here if the device does not exist at the system and netiface
# not installed.
raise NetworkInterfaceNotFoundError(target=device)
return runner.stdout
## Instruction:
Change to avoid a DeprecationWarning
## Code After:
from __future__ import absolute_import, unicode_literals
import subprocrunner as spr
from ._common import find_bin_path
from ._const import TcSubCommand
from ._error import NetworkInterfaceNotFoundError
def get_tc_base_command(tc_subcommand):
if not isinstance(tc_subcommand, TcSubCommand):
raise ValueError("the argument must be a TcSubCommand value")
return "{:s} {:s}".format(find_bin_path("tc"), tc_subcommand.value)
def run_tc_show(subcommand, device, tc_command_output):
from ._network import verify_network_interface
verify_network_interface(device, tc_command_output)
runner = spr.SubprocessRunner(
"{:s} show dev {:s}".format(get_tc_base_command(subcommand), device)
)
if runner.run() != 0 and runner.stderr.find("Cannot find device") != -1:
# reach here if the device does not exist at the system and netiface
# not installed.
raise NetworkInterfaceNotFoundError(target=device)
return runner.stdout
| # ... existing code ...
def get_tc_base_command(tc_subcommand):
if not isinstance(tc_subcommand, TcSubCommand):
raise ValueError("the argument must be a TcSubCommand value")
# ... rest of the code ... |
817b597f3a45a8b16de84d480458a66499604f5a | owned_models/models.py | owned_models/models.py | from django.conf import settings
from django.db import models
class UserOwnedModelManager(models.Manager):
def filter_for_user(self, user, *args, **kwargs):
return super(UserOwnedModelManager, self).get_queryset().filter(user = user, *args, **kwargs)
def get_for_user(self, user, *args, **kwargs):
if 'user' in kwargs:
kwargs.pop('user')
return super(UserOwnedModelManager, self).get_queryset().get(user = user, *args, **kwargs)
class UserOwnedModel(models.Model):
user = models.ForeignKey(settings.AUTH_USER_MODEL, editable = False)
objects = UserOwnedModelManager()
class Meta:
abstract = True | from django.conf import settings
from django.db import models
class UserOwnedModelManager(models.Manager):
def filter_for_user(self, user, *args, **kwargs):
return super(UserOwnedModelManager, self).get_queryset().filter(user = user, *args, **kwargs)
def get_for_user(self, user, *args, **kwargs):
if 'user' in kwargs:
kwargs.pop('user')
return super(UserOwnedModelManager, self).get_queryset().get(user = user, *args, **kwargs)
def get_or_create_for_user(self, user, **kwargs):
return super(UserOwnedModelManager, self).get_or_create(user = user, **kwargs)
class UserOwnedModel(models.Model):
user = models.ForeignKey(settings.AUTH_USER_MODEL, editable = False)
objects = UserOwnedModelManager()
class Meta:
abstract = True | Add `get_or_create_for_user` method to default Manager. | Add `get_or_create_for_user` method to default Manager. | Python | mit | discolabs/django-owned-models | from django.conf import settings
from django.db import models
class UserOwnedModelManager(models.Manager):
def filter_for_user(self, user, *args, **kwargs):
return super(UserOwnedModelManager, self).get_queryset().filter(user = user, *args, **kwargs)
def get_for_user(self, user, *args, **kwargs):
if 'user' in kwargs:
kwargs.pop('user')
return super(UserOwnedModelManager, self).get_queryset().get(user = user, *args, **kwargs)
+ def get_or_create_for_user(self, user, **kwargs):
+ return super(UserOwnedModelManager, self).get_or_create(user = user, **kwargs)
class UserOwnedModel(models.Model):
user = models.ForeignKey(settings.AUTH_USER_MODEL, editable = False)
objects = UserOwnedModelManager()
class Meta:
abstract = True | Add `get_or_create_for_user` method to default Manager. | ## Code Before:
from django.conf import settings
from django.db import models
class UserOwnedModelManager(models.Manager):
def filter_for_user(self, user, *args, **kwargs):
return super(UserOwnedModelManager, self).get_queryset().filter(user = user, *args, **kwargs)
def get_for_user(self, user, *args, **kwargs):
if 'user' in kwargs:
kwargs.pop('user')
return super(UserOwnedModelManager, self).get_queryset().get(user = user, *args, **kwargs)
class UserOwnedModel(models.Model):
user = models.ForeignKey(settings.AUTH_USER_MODEL, editable = False)
objects = UserOwnedModelManager()
class Meta:
abstract = True
## Instruction:
Add `get_or_create_for_user` method to default Manager.
## Code After:
from django.conf import settings
from django.db import models
class UserOwnedModelManager(models.Manager):
def filter_for_user(self, user, *args, **kwargs):
return super(UserOwnedModelManager, self).get_queryset().filter(user = user, *args, **kwargs)
def get_for_user(self, user, *args, **kwargs):
if 'user' in kwargs:
kwargs.pop('user')
return super(UserOwnedModelManager, self).get_queryset().get(user = user, *args, **kwargs)
def get_or_create_for_user(self, user, **kwargs):
return super(UserOwnedModelManager, self).get_or_create(user = user, **kwargs)
class UserOwnedModel(models.Model):
user = models.ForeignKey(settings.AUTH_USER_MODEL, editable = False)
objects = UserOwnedModelManager()
class Meta:
abstract = True | // ... existing code ...
def get_or_create_for_user(self, user, **kwargs):
return super(UserOwnedModelManager, self).get_or_create(user = user, **kwargs)
// ... rest of the code ... |
15c8215415d36da4fac9c7333e62239f7b81c12d | test/support/mock_definitions.py | test/support/mock_definitions.py | class MockDefinitions(object):
def __init__(self, session_key=None):
self.session_key = session_key if session_key is not None else '123456789'
@property
def metadata(self):
host = os.getenv('SPLUNK_API_HOST', 'localhost')
port = os.getenv('SPLUNK_API_PORT', 8089),
return {'server_uri': 'https://{host}:{port}/', 'session_key': self.session_key,
'name': 'amp4e_events_test_input'}
| import os
# Generates validation/input definitions as if they were created by splunk for tests
class MockDefinitions(object):
def __init__(self, session_key=None):
self.session_key = session_key if session_key is not None else '123456789'
@property
def metadata(self):
host = os.getenv('SPLUNK_API_HOST', '127.0.0.1')
return {'server_uri': 'https://{host}:8089/'.format(host=host), 'session_key': self.session_key,
'name': 'amp4e_events_test_input'}
| Change mock to be env dependant | Change mock to be env dependant
| Python | bsd-2-clause | Cisco-AMP/amp4e_splunk_events_input,Cisco-AMP/amp4e_splunk_events_input,Cisco-AMP/amp4e_splunk_events_input,Cisco-AMP/amp4e_splunk_events_input | + import os
+ # Generates validation/input definitions as if they were created by splunk for tests
class MockDefinitions(object):
def __init__(self, session_key=None):
self.session_key = session_key if session_key is not None else '123456789'
@property
def metadata(self):
- host = os.getenv('SPLUNK_API_HOST', 'localhost')
+ host = os.getenv('SPLUNK_API_HOST', '127.0.0.1')
- port = os.getenv('SPLUNK_API_PORT', 8089),
- return {'server_uri': 'https://{host}:{port}/', 'session_key': self.session_key,
+ return {'server_uri': 'https://{host}:8089/'.format(host=host), 'session_key': self.session_key,
'name': 'amp4e_events_test_input'}
| Change mock to be env dependant | ## Code Before:
class MockDefinitions(object):
def __init__(self, session_key=None):
self.session_key = session_key if session_key is not None else '123456789'
@property
def metadata(self):
host = os.getenv('SPLUNK_API_HOST', 'localhost')
port = os.getenv('SPLUNK_API_PORT', 8089),
return {'server_uri': 'https://{host}:{port}/', 'session_key': self.session_key,
'name': 'amp4e_events_test_input'}
## Instruction:
Change mock to be env dependant
## Code After:
import os
# Generates validation/input definitions as if they were created by splunk for tests
class MockDefinitions(object):
def __init__(self, session_key=None):
self.session_key = session_key if session_key is not None else '123456789'
@property
def metadata(self):
host = os.getenv('SPLUNK_API_HOST', '127.0.0.1')
return {'server_uri': 'https://{host}:8089/'.format(host=host), 'session_key': self.session_key,
'name': 'amp4e_events_test_input'}
| # ... existing code ...
import os
# Generates validation/input definitions as if they were created by splunk for tests
class MockDefinitions(object):
# ... modified code ...
def metadata(self):
host = os.getenv('SPLUNK_API_HOST', '127.0.0.1')
return {'server_uri': 'https://{host}:8089/'.format(host=host), 'session_key': self.session_key,
'name': 'amp4e_events_test_input'}
# ... rest of the code ... |
1e182ec0fd7cf550c809f2e6792629caeb8d5553 | sauce/lib/helpers.py | sauce/lib/helpers.py |
from datetime import datetime
from tg import url as tgurl
#from webhelpers import date, feedgenerator, html, number, misc, text
import webhelpers as w
from webhelpers.html.tags import link_to
from webhelpers.text import truncate
from webhelpers.date import distance_of_time_in_words
import re
#log = logging.getLogger(__name__)
cut = lambda text, max=200: truncate(text, max, whole_word=True)
strftimedelta = lambda delta, granularity='minute': distance_of_time_in_words(datetime.now(), datetime.now()+delta, granularity)
def link(label, url='', **attrs):
return link_to(label, tgurl(url), **attrs)
def striphtml(text):
return re.sub('<[^<]+?>', ' ', text).strip()
|
from datetime import datetime
from tg import url as tgurl
#from webhelpers import date, feedgenerator, html, number, misc, text
import webhelpers as w
from webhelpers.html.tags import link_to
from webhelpers.text import truncate
from webhelpers.date import distance_of_time_in_words
from genshi.core import striptags
import re
#log = logging.getLogger(__name__)
cut = lambda text, max=200: truncate(text, max, whole_word=True)
strftimedelta = lambda delta, granularity='minute': distance_of_time_in_words(datetime.now(), datetime.now()+delta, granularity)
striphtml = striptags
def link(label, url='', **attrs):
return link_to(label, tgurl(url), **attrs)
| Use striptags from genshi for striphtml, since we have to have genshi anyway | Use striptags from genshi for striphtml, since we have to have genshi anyway
| Python | agpl-3.0 | moschlar/SAUCE,moschlar/SAUCE,moschlar/SAUCE,moschlar/SAUCE |
from datetime import datetime
from tg import url as tgurl
#from webhelpers import date, feedgenerator, html, number, misc, text
import webhelpers as w
from webhelpers.html.tags import link_to
from webhelpers.text import truncate
from webhelpers.date import distance_of_time_in_words
+ from genshi.core import striptags
+
import re
#log = logging.getLogger(__name__)
cut = lambda text, max=200: truncate(text, max, whole_word=True)
strftimedelta = lambda delta, granularity='minute': distance_of_time_in_words(datetime.now(), datetime.now()+delta, granularity)
+ striphtml = striptags
def link(label, url='', **attrs):
return link_to(label, tgurl(url), **attrs)
- def striphtml(text):
- return re.sub('<[^<]+?>', ' ', text).strip()
| Use striptags from genshi for striphtml, since we have to have genshi anyway | ## Code Before:
from datetime import datetime
from tg import url as tgurl
#from webhelpers import date, feedgenerator, html, number, misc, text
import webhelpers as w
from webhelpers.html.tags import link_to
from webhelpers.text import truncate
from webhelpers.date import distance_of_time_in_words
import re
#log = logging.getLogger(__name__)
cut = lambda text, max=200: truncate(text, max, whole_word=True)
strftimedelta = lambda delta, granularity='minute': distance_of_time_in_words(datetime.now(), datetime.now()+delta, granularity)
def link(label, url='', **attrs):
return link_to(label, tgurl(url), **attrs)
def striphtml(text):
return re.sub('<[^<]+?>', ' ', text).strip()
## Instruction:
Use striptags from genshi for striphtml, since we have to have genshi anyway
## Code After:
from datetime import datetime
from tg import url as tgurl
#from webhelpers import date, feedgenerator, html, number, misc, text
import webhelpers as w
from webhelpers.html.tags import link_to
from webhelpers.text import truncate
from webhelpers.date import distance_of_time_in_words
from genshi.core import striptags
import re
#log = logging.getLogger(__name__)
cut = lambda text, max=200: truncate(text, max, whole_word=True)
strftimedelta = lambda delta, granularity='minute': distance_of_time_in_words(datetime.now(), datetime.now()+delta, granularity)
striphtml = striptags
def link(label, url='', **attrs):
return link_to(label, tgurl(url), **attrs)
| // ... existing code ...
from genshi.core import striptags
import re
// ... modified code ...
strftimedelta = lambda delta, granularity='minute': distance_of_time_in_words(datetime.now(), datetime.now()+delta, granularity)
striphtml = striptags
...
// ... rest of the code ... |
cd00388bdc4c1963ac8ff81f9b7132ba32272fc8 | adwords_client/__init__.py | adwords_client/__init__.py | __version__ = '17.07' # Date based versioning
# See: https://packaging.python.org/tutorials/distributing-packages/#date-based-versioning
|
__version__ = '17.07' # Date based versioning
# See: https://packaging.python.org/tutorials/distributing-packages/#date-based-versioning
import sys
print(__doc__, file=sys.stderr, flush=True)
| Print license on each import | Print license on each import
| Python | apache-2.0 | getninjas/adwords-client | +
__version__ = '17.07' # Date based versioning
# See: https://packaging.python.org/tutorials/distributing-packages/#date-based-versioning
+ import sys
+ print(__doc__, file=sys.stderr, flush=True)
+ | Print license on each import | ## Code Before:
__version__ = '17.07' # Date based versioning
# See: https://packaging.python.org/tutorials/distributing-packages/#date-based-versioning
## Instruction:
Print license on each import
## Code After:
__version__ = '17.07' # Date based versioning
# See: https://packaging.python.org/tutorials/distributing-packages/#date-based-versioning
import sys
print(__doc__, file=sys.stderr, flush=True)
| // ... existing code ...
__version__ = '17.07' # Date based versioning
// ... modified code ...
# See: https://packaging.python.org/tutorials/distributing-packages/#date-based-versioning
import sys
print(__doc__, file=sys.stderr, flush=True)
// ... rest of the code ... |
14d6955118893c532c1d9f8f6037d1da1b18dbbb | analysis/plot-skeleton.py | analysis/plot-skeleton.py |
import climate
import database
import plots
@climate.annotate(
root='plot data rooted at this path',
pattern=('plot data from files matching this pattern', 'option'),
)
def main(root, pattern='*/*block02/*trial00*.csv.gz'):
for trial in database.Experiment(root).trials_matching(pattern):
with plots.space() as ax:
plots.skeleton(ax, trial, 100)
if __name__ == '__main__':
climate.call(main)
|
import climate
import pandas as pd
import database
import plots
@climate.annotate(
root='plot data rooted at this path',
pattern=('plot data from files matching this pattern', 'option'),
)
def main(root, pattern='*/*block03/*trial00*.csv.gz'):
for trial in database.Experiment(root).trials_matching(pattern):
with plots.space() as ax:
for i in range(3):
plots.skeleton(ax, trial, 1000 + 300 * i, lw=2, color='#fd3220', alpha=0.3)
#trial.rotate_heading(pd.Series([-6.28 / 10] * len(trial.df)))
trial.make_body_relative()
for i in range(3):
plots.skeleton(ax, trial, 1000 + 300 * i, offset=(0.5 * i, 0.5 * i), lw=2, color='#111111', alpha=0.3)
if __name__ == '__main__':
climate.call(main)
| Add multiple skeletons for the moment. | Add multiple skeletons for the moment.
| Python | mit | lmjohns3/cube-experiment,lmjohns3/cube-experiment,lmjohns3/cube-experiment |
import climate
+ import pandas as pd
import database
import plots
@climate.annotate(
root='plot data rooted at this path',
pattern=('plot data from files matching this pattern', 'option'),
)
- def main(root, pattern='*/*block02/*trial00*.csv.gz'):
+ def main(root, pattern='*/*block03/*trial00*.csv.gz'):
for trial in database.Experiment(root).trials_matching(pattern):
with plots.space() as ax:
- plots.skeleton(ax, trial, 100)
+ for i in range(3):
+ plots.skeleton(ax, trial, 1000 + 300 * i, lw=2, color='#fd3220', alpha=0.3)
+ #trial.rotate_heading(pd.Series([-6.28 / 10] * len(trial.df)))
+ trial.make_body_relative()
+ for i in range(3):
+ plots.skeleton(ax, trial, 1000 + 300 * i, offset=(0.5 * i, 0.5 * i), lw=2, color='#111111', alpha=0.3)
if __name__ == '__main__':
climate.call(main)
| Add multiple skeletons for the moment. | ## Code Before:
import climate
import database
import plots
@climate.annotate(
root='plot data rooted at this path',
pattern=('plot data from files matching this pattern', 'option'),
)
def main(root, pattern='*/*block02/*trial00*.csv.gz'):
for trial in database.Experiment(root).trials_matching(pattern):
with plots.space() as ax:
plots.skeleton(ax, trial, 100)
if __name__ == '__main__':
climate.call(main)
## Instruction:
Add multiple skeletons for the moment.
## Code After:
import climate
import pandas as pd
import database
import plots
@climate.annotate(
root='plot data rooted at this path',
pattern=('plot data from files matching this pattern', 'option'),
)
def main(root, pattern='*/*block03/*trial00*.csv.gz'):
for trial in database.Experiment(root).trials_matching(pattern):
with plots.space() as ax:
for i in range(3):
plots.skeleton(ax, trial, 1000 + 300 * i, lw=2, color='#fd3220', alpha=0.3)
#trial.rotate_heading(pd.Series([-6.28 / 10] * len(trial.df)))
trial.make_body_relative()
for i in range(3):
plots.skeleton(ax, trial, 1000 + 300 * i, offset=(0.5 * i, 0.5 * i), lw=2, color='#111111', alpha=0.3)
if __name__ == '__main__':
climate.call(main)
| // ... existing code ...
import climate
import pandas as pd
// ... modified code ...
)
def main(root, pattern='*/*block03/*trial00*.csv.gz'):
for trial in database.Experiment(root).trials_matching(pattern):
...
with plots.space() as ax:
for i in range(3):
plots.skeleton(ax, trial, 1000 + 300 * i, lw=2, color='#fd3220', alpha=0.3)
#trial.rotate_heading(pd.Series([-6.28 / 10] * len(trial.df)))
trial.make_body_relative()
for i in range(3):
plots.skeleton(ax, trial, 1000 + 300 * i, offset=(0.5 * i, 0.5 * i), lw=2, color='#111111', alpha=0.3)
// ... rest of the code ... |
f2fc7f1015fc24fdbb69069ac74a21437e94657b | xmantissa/plugins/sineoff.py | xmantissa/plugins/sineoff.py | from axiom import iaxiom, userbase
from xmantissa import website, offering, provisioning
from sine import sipserver, sinetheme
sineproxy = provisioning.BenefactorFactory(
name = u'sineproxy',
description = u'Sine SIP Proxy',
benefactorClass = sipserver.SineBenefactor)
plugin = offering.Offering(
name = u"Sine",
description = u"""
The Sine SIP proxy and registrar.
""",
siteRequirements = (
(userbase.IRealm, userbase.LoginSystem),
(None, website.WebSite),
(None, sipserver.SIPServer)),
appPowerups = (sipserver.SinePublicPage,
),
benefactorFactories = (sineproxy,),
loginInterfaces=(),
themes = (sinetheme.XHTMLDirectoryTheme('base'),)
)
| from axiom import iaxiom, userbase
from xmantissa import website, offering, provisioning
from sine import sipserver, sinetheme
sineproxy = provisioning.BenefactorFactory(
name = u'sineproxy',
description = u'Sine SIP Proxy',
benefactorClass = sipserver.SineBenefactor)
plugin = offering.Offering(
name = u"Sine",
description = u"""
The Sine SIP proxy and registrar.
""",
siteRequirements = (
(userbase.IRealm, userbase.LoginSystem),
(None, website.WebSite),
(None, sipserver.SIPServer)),
appPowerups = (sipserver.SinePublicPage,
),
benefactorFactories = (sineproxy,),
themes = (sinetheme.XHTMLDirectoryTheme('base'),)
)
| Revert 5505 - introduced numerous regressions into the test suite | Revert 5505 - introduced numerous regressions into the test suite | Python | mit | habnabit/divmod-sine,twisted/sine | from axiom import iaxiom, userbase
from xmantissa import website, offering, provisioning
from sine import sipserver, sinetheme
sineproxy = provisioning.BenefactorFactory(
name = u'sineproxy',
description = u'Sine SIP Proxy',
benefactorClass = sipserver.SineBenefactor)
plugin = offering.Offering(
name = u"Sine",
description = u"""
The Sine SIP proxy and registrar.
""",
siteRequirements = (
(userbase.IRealm, userbase.LoginSystem),
(None, website.WebSite),
(None, sipserver.SIPServer)),
appPowerups = (sipserver.SinePublicPage,
),
benefactorFactories = (sineproxy,),
- loginInterfaces=(),
+
themes = (sinetheme.XHTMLDirectoryTheme('base'),)
)
| Revert 5505 - introduced numerous regressions into the test suite | ## Code Before:
from axiom import iaxiom, userbase
from xmantissa import website, offering, provisioning
from sine import sipserver, sinetheme
sineproxy = provisioning.BenefactorFactory(
name = u'sineproxy',
description = u'Sine SIP Proxy',
benefactorClass = sipserver.SineBenefactor)
plugin = offering.Offering(
name = u"Sine",
description = u"""
The Sine SIP proxy and registrar.
""",
siteRequirements = (
(userbase.IRealm, userbase.LoginSystem),
(None, website.WebSite),
(None, sipserver.SIPServer)),
appPowerups = (sipserver.SinePublicPage,
),
benefactorFactories = (sineproxy,),
loginInterfaces=(),
themes = (sinetheme.XHTMLDirectoryTheme('base'),)
)
## Instruction:
Revert 5505 - introduced numerous regressions into the test suite
## Code After:
from axiom import iaxiom, userbase
from xmantissa import website, offering, provisioning
from sine import sipserver, sinetheme
sineproxy = provisioning.BenefactorFactory(
name = u'sineproxy',
description = u'Sine SIP Proxy',
benefactorClass = sipserver.SineBenefactor)
plugin = offering.Offering(
name = u"Sine",
description = u"""
The Sine SIP proxy and registrar.
""",
siteRequirements = (
(userbase.IRealm, userbase.LoginSystem),
(None, website.WebSite),
(None, sipserver.SIPServer)),
appPowerups = (sipserver.SinePublicPage,
),
benefactorFactories = (sineproxy,),
themes = (sinetheme.XHTMLDirectoryTheme('base'),)
)
| ...
benefactorFactories = (sineproxy,),
themes = (sinetheme.XHTMLDirectoryTheme('base'),)
... |
ca8fa466638c0ef405a82dfc3cfecfdb400faaa7 | sublime_jedi/helper.py | sublime_jedi/helper.py | import sublime
import sublime_plugin
from .utils import ask_daemon
class HelpMessageCommand(sublime_plugin.TextCommand):
def run(self, edit, docstring):
self.view.close()
self.view.insert(edit, self.view.size(), docstring)
class SublimeJediDocstring(sublime_plugin.TextCommand):
"""
Show doctring in output panel
"""
def run(self, edit):
ask_daemon(self.view, self.show_docstring, 'docstring')
def show_docstring(self, view, docstring):
window = sublime.active_window()
if docstring:
output = window.get_output_panel('help_panel')
output.set_read_only(False)
output.run_command('help_message', {'docstring': docstring})
output.set_read_only(True)
window.run_command("show_panel", {"panel": "output.help_panel"})
else:
window.run_command("hide_panel", {"panel": "output.help_panel"})
sublime.status_message('Jedi: No results!')
class SublimeJediSignature(sublime_plugin.TextCommand):
"""
Show signature in statusbar
"""
def run(self, edit):
ask_daemon(self.view, self.show_signature, 'signature')
def show_signature(self, view, signature):
if signature:
sublime.status_message('Jedi: {0}'.format(signature))
| import sublime
import sublime_plugin
from .utils import ask_daemon, PythonCommandMixin
class HelpMessageCommand(sublime_plugin.TextCommand):
def run(self, edit, docstring):
self.view.close()
self.view.insert(edit, self.view.size(), docstring)
class SublimeJediDocstring(PythonCommandMixin, sublime_plugin.TextCommand):
"""
Show doctring in output panel
"""
def run(self, edit):
ask_daemon(self.view, self.show_docstring, 'docstring')
def show_docstring(self, view, docstring):
window = sublime.active_window()
if docstring:
output = window.get_output_panel('help_panel')
output.set_read_only(False)
output.run_command('help_message', {'docstring': docstring})
output.set_read_only(True)
window.run_command("show_panel", {"panel": "output.help_panel"})
else:
window.run_command("hide_panel", {"panel": "output.help_panel"})
sublime.status_message('Jedi: No results!')
class SublimeJediSignature(PythonCommandMixin, sublime_plugin.TextCommand):
"""
Show signature in statusbar
"""
def run(self, edit):
ask_daemon(self.view, self.show_signature, 'signature')
def show_signature(self, view, signature):
if signature:
sublime.status_message('Jedi: {0}'.format(signature))
| Hide documentation commands in non-python scope | Hide documentation commands in non-python scope
| Python | mit | srusskih/SublimeJEDI,edelvalle/SublimeJEDI | import sublime
import sublime_plugin
- from .utils import ask_daemon
+ from .utils import ask_daemon, PythonCommandMixin
class HelpMessageCommand(sublime_plugin.TextCommand):
def run(self, edit, docstring):
self.view.close()
self.view.insert(edit, self.view.size(), docstring)
- class SublimeJediDocstring(sublime_plugin.TextCommand):
+ class SublimeJediDocstring(PythonCommandMixin, sublime_plugin.TextCommand):
"""
Show doctring in output panel
"""
def run(self, edit):
ask_daemon(self.view, self.show_docstring, 'docstring')
def show_docstring(self, view, docstring):
window = sublime.active_window()
if docstring:
output = window.get_output_panel('help_panel')
output.set_read_only(False)
output.run_command('help_message', {'docstring': docstring})
output.set_read_only(True)
window.run_command("show_panel", {"panel": "output.help_panel"})
else:
window.run_command("hide_panel", {"panel": "output.help_panel"})
sublime.status_message('Jedi: No results!')
- class SublimeJediSignature(sublime_plugin.TextCommand):
+ class SublimeJediSignature(PythonCommandMixin, sublime_plugin.TextCommand):
"""
Show signature in statusbar
"""
def run(self, edit):
ask_daemon(self.view, self.show_signature, 'signature')
def show_signature(self, view, signature):
if signature:
sublime.status_message('Jedi: {0}'.format(signature))
| Hide documentation commands in non-python scope | ## Code Before:
import sublime
import sublime_plugin
from .utils import ask_daemon
class HelpMessageCommand(sublime_plugin.TextCommand):
def run(self, edit, docstring):
self.view.close()
self.view.insert(edit, self.view.size(), docstring)
class SublimeJediDocstring(sublime_plugin.TextCommand):
"""
Show doctring in output panel
"""
def run(self, edit):
ask_daemon(self.view, self.show_docstring, 'docstring')
def show_docstring(self, view, docstring):
window = sublime.active_window()
if docstring:
output = window.get_output_panel('help_panel')
output.set_read_only(False)
output.run_command('help_message', {'docstring': docstring})
output.set_read_only(True)
window.run_command("show_panel", {"panel": "output.help_panel"})
else:
window.run_command("hide_panel", {"panel": "output.help_panel"})
sublime.status_message('Jedi: No results!')
class SublimeJediSignature(sublime_plugin.TextCommand):
"""
Show signature in statusbar
"""
def run(self, edit):
ask_daemon(self.view, self.show_signature, 'signature')
def show_signature(self, view, signature):
if signature:
sublime.status_message('Jedi: {0}'.format(signature))
## Instruction:
Hide documentation commands in non-python scope
## Code After:
import sublime
import sublime_plugin
from .utils import ask_daemon, PythonCommandMixin
class HelpMessageCommand(sublime_plugin.TextCommand):
def run(self, edit, docstring):
self.view.close()
self.view.insert(edit, self.view.size(), docstring)
class SublimeJediDocstring(PythonCommandMixin, sublime_plugin.TextCommand):
"""
Show doctring in output panel
"""
def run(self, edit):
ask_daemon(self.view, self.show_docstring, 'docstring')
def show_docstring(self, view, docstring):
window = sublime.active_window()
if docstring:
output = window.get_output_panel('help_panel')
output.set_read_only(False)
output.run_command('help_message', {'docstring': docstring})
output.set_read_only(True)
window.run_command("show_panel", {"panel": "output.help_panel"})
else:
window.run_command("hide_panel", {"panel": "output.help_panel"})
sublime.status_message('Jedi: No results!')
class SublimeJediSignature(PythonCommandMixin, sublime_plugin.TextCommand):
"""
Show signature in statusbar
"""
def run(self, edit):
ask_daemon(self.view, self.show_signature, 'signature')
def show_signature(self, view, signature):
if signature:
sublime.status_message('Jedi: {0}'.format(signature))
| ...
from .utils import ask_daemon, PythonCommandMixin
...
class SublimeJediDocstring(PythonCommandMixin, sublime_plugin.TextCommand):
"""
...
class SublimeJediSignature(PythonCommandMixin, sublime_plugin.TextCommand):
"""
... |
276f22927890051f66976468585d8351c0ccf5b9 | sum-of-multiples/sum_of_multiples.py | sum-of-multiples/sum_of_multiples.py | def sum_of_multiples(limit, factors):
return sum(filter(lambda n: n < limit,
{f*i for i in range(1, limit) for f in factors}))
| def sum_of_multiples(limit, factors):
return sum({n for f in factors for n in range(f, limit, f)})
| Use more optimal method of getting multiples | Use more optimal method of getting multiples
| Python | agpl-3.0 | CubicComet/exercism-python-solutions | def sum_of_multiples(limit, factors):
+ return sum({n for f in factors for n in range(f, limit, f)})
- return sum(filter(lambda n: n < limit,
- {f*i for i in range(1, limit) for f in factors}))
| Use more optimal method of getting multiples | ## Code Before:
def sum_of_multiples(limit, factors):
return sum(filter(lambda n: n < limit,
{f*i for i in range(1, limit) for f in factors}))
## Instruction:
Use more optimal method of getting multiples
## Code After:
def sum_of_multiples(limit, factors):
return sum({n for f in factors for n in range(f, limit, f)})
| ...
def sum_of_multiples(limit, factors):
return sum({n for f in factors for n in range(f, limit, f)})
... |
19354bd82a89383d795cdada8d6af78e8f12eed8 | src/server/test_client.py | src/server/test_client.py | import socket
import sys
from RemoteFunctionCaller import *
from SocketNetworker import SocketNetworker
HOST = 'localhost' # The remote host
PORT = 8553 # The same port as used by the server
s = None
for res in socket.getaddrinfo(HOST, PORT, socket.AF_UNSPEC, socket.SOCK_STREAM):
af, socktype, proto, canonname, sa = res
try:
s = socket.socket(af, socktype, proto)
except OSError as msg:
s = None
continue
try:
s.connect(sa)
except OSError as msg:
s.close()
s = None
continue
break
if s is None:
print('could not open socket')
sys.exit(1)
nw = SocketNetworker(s)
caller = RemoteFunctionCaller(nw)
try:
caller.setData("test", "success")
print(caller.getData("test", default="failish"))
except TimeoutError:
print("Timed out.")
nw.close()
| import socket
import sys
from RemoteFunctionCaller import *
from SocketNetworker import SocketNetworker
HOST = 'localhost' # The remote host
PORT = 8553 # The same port as used by the server
s = None
for res in socket.getaddrinfo(HOST, PORT, socket.AF_UNSPEC, socket.SOCK_STREAM):
af, socktype, proto, canonname, sa = res
try:
s = socket.socket(af, socktype, proto)
except OSError as msg:
s = None
continue
try:
s.connect(sa)
except OSError as msg:
s.close()
s = None
continue
break
if s is None:
print('could not open socket')
sys.exit(1)
nw = SocketNetworker(s)
caller = RemoteFunctionCaller(nw)
try:
print(caller.SharedClientDataStore__set("test", "success"))
print(caller.SharedClientDtaStore__get("test", default="failish"))
except TimeoutError:
print("Timed out.")
nw.close()
| Update call method in test client | Update call method in test client
| Python | mit | cnlohr/bridgesim,cnlohr/bridgesim,cnlohr/bridgesim,cnlohr/bridgesim | import socket
import sys
from RemoteFunctionCaller import *
from SocketNetworker import SocketNetworker
HOST = 'localhost' # The remote host
PORT = 8553 # The same port as used by the server
s = None
for res in socket.getaddrinfo(HOST, PORT, socket.AF_UNSPEC, socket.SOCK_STREAM):
af, socktype, proto, canonname, sa = res
try:
s = socket.socket(af, socktype, proto)
except OSError as msg:
s = None
continue
try:
s.connect(sa)
except OSError as msg:
s.close()
s = None
continue
break
if s is None:
print('could not open socket')
sys.exit(1)
nw = SocketNetworker(s)
caller = RemoteFunctionCaller(nw)
try:
- caller.setData("test", "success")
+ print(caller.SharedClientDataStore__set("test", "success"))
- print(caller.getData("test", default="failish"))
+ print(caller.SharedClientDtaStore__get("test", default="failish"))
except TimeoutError:
print("Timed out.")
nw.close()
| Update call method in test client | ## Code Before:
import socket
import sys
from RemoteFunctionCaller import *
from SocketNetworker import SocketNetworker
HOST = 'localhost' # The remote host
PORT = 8553 # The same port as used by the server
s = None
for res in socket.getaddrinfo(HOST, PORT, socket.AF_UNSPEC, socket.SOCK_STREAM):
af, socktype, proto, canonname, sa = res
try:
s = socket.socket(af, socktype, proto)
except OSError as msg:
s = None
continue
try:
s.connect(sa)
except OSError as msg:
s.close()
s = None
continue
break
if s is None:
print('could not open socket')
sys.exit(1)
nw = SocketNetworker(s)
caller = RemoteFunctionCaller(nw)
try:
caller.setData("test", "success")
print(caller.getData("test", default="failish"))
except TimeoutError:
print("Timed out.")
nw.close()
## Instruction:
Update call method in test client
## Code After:
import socket
import sys
from RemoteFunctionCaller import *
from SocketNetworker import SocketNetworker
HOST = 'localhost' # The remote host
PORT = 8553 # The same port as used by the server
s = None
for res in socket.getaddrinfo(HOST, PORT, socket.AF_UNSPEC, socket.SOCK_STREAM):
af, socktype, proto, canonname, sa = res
try:
s = socket.socket(af, socktype, proto)
except OSError as msg:
s = None
continue
try:
s.connect(sa)
except OSError as msg:
s.close()
s = None
continue
break
if s is None:
print('could not open socket')
sys.exit(1)
nw = SocketNetworker(s)
caller = RemoteFunctionCaller(nw)
try:
print(caller.SharedClientDataStore__set("test", "success"))
print(caller.SharedClientDtaStore__get("test", default="failish"))
except TimeoutError:
print("Timed out.")
nw.close()
| # ... existing code ...
try:
print(caller.SharedClientDataStore__set("test", "success"))
print(caller.SharedClientDtaStore__get("test", default="failish"))
except TimeoutError:
# ... rest of the code ... |
945d64464857581052e18d79e62a6fde8bdecb9b | fabfile.py | fabfile.py | import sys
from fabric.api import local, task
@task
def start_db():
if sys.platform.startswith('darwin'):
# Mac OSX
local('postgres -D /usr/local/var/postgres -s')
| import sys
from pathlib import Path
from fabric.api import local, task, lcd, env
from fabric.contrib.console import confirm
from fabric.utils import abort
src_p = Path(env.real_fabfile).parent / 'src'
@task
def start_db():
if sys.platform.startswith('darwin'):
# Mac OSX
local('postgres -D /usr/local/var/postgres -s')
@task
def backup():
cmd_dumpdata = 'python manage.py dumpdata '
with lcd(src_p):
local(
cmd_dumpdata + 'users.EmailUser data_sources.DataSource | '
'tee ../db_dump/user_sources.json'
)
local(
cmd_dumpdata + 'experiments | '
'tee ../db_dump/experiments.json'
)
local(
cmd_dumpdata + 'analyses.GenomeReference | '
'tee ../db_dump/genome_reference.json'
)
@task
def reborn():
with lcd(src_p.as_posix()):
db_dump_dir = Path(env.cwd, '../db_dump')
if not (
db_dump_dir.joinpath('user_sources.json').exists() and
db_dump_dir.joinpath('genome_reference.json').exists() and
db_dump_dir.joinpath('experiments.json').exists()
):
abort('Backup the import database content first!')
confirm('Destory and re-create the current database?', False)
local('dropdb biocloud')
local('createdb biocloud')
local('python manage.py migrate')
local('python manage.py loaddata ../db_dump/user_sources.json')
local('python manage.py loaddata ../db_dump/genome_reference.json')
local('python manage.py loaddata ../db_dump/experiments.json')
| Add fab command to backup and destroy database | Add fab command to backup and destroy database
| Python | mit | ccwang002/biocloud-server-kai,ccwang002/biocloud-server-kai,ccwang002/biocloud-server-kai | import sys
+ from pathlib import Path
- from fabric.api import local, task
+ from fabric.api import local, task, lcd, env
+ from fabric.contrib.console import confirm
+ from fabric.utils import abort
+
+ src_p = Path(env.real_fabfile).parent / 'src'
+
@task
def start_db():
if sys.platform.startswith('darwin'):
# Mac OSX
local('postgres -D /usr/local/var/postgres -s')
+
+ @task
+ def backup():
+ cmd_dumpdata = 'python manage.py dumpdata '
+ with lcd(src_p):
+ local(
+ cmd_dumpdata + 'users.EmailUser data_sources.DataSource | '
+ 'tee ../db_dump/user_sources.json'
+ )
+ local(
+ cmd_dumpdata + 'experiments | '
+ 'tee ../db_dump/experiments.json'
+ )
+ local(
+ cmd_dumpdata + 'analyses.GenomeReference | '
+ 'tee ../db_dump/genome_reference.json'
+ )
+
+
+ @task
+ def reborn():
+ with lcd(src_p.as_posix()):
+ db_dump_dir = Path(env.cwd, '../db_dump')
+ if not (
+ db_dump_dir.joinpath('user_sources.json').exists() and
+ db_dump_dir.joinpath('genome_reference.json').exists() and
+ db_dump_dir.joinpath('experiments.json').exists()
+ ):
+ abort('Backup the import database content first!')
+ confirm('Destory and re-create the current database?', False)
+
+ local('dropdb biocloud')
+ local('createdb biocloud')
+ local('python manage.py migrate')
+ local('python manage.py loaddata ../db_dump/user_sources.json')
+ local('python manage.py loaddata ../db_dump/genome_reference.json')
+ local('python manage.py loaddata ../db_dump/experiments.json')
+ | Add fab command to backup and destroy database | ## Code Before:
import sys
from fabric.api import local, task
@task
def start_db():
if sys.platform.startswith('darwin'):
# Mac OSX
local('postgres -D /usr/local/var/postgres -s')
## Instruction:
Add fab command to backup and destroy database
## Code After:
import sys
from pathlib import Path
from fabric.api import local, task, lcd, env
from fabric.contrib.console import confirm
from fabric.utils import abort
src_p = Path(env.real_fabfile).parent / 'src'
@task
def start_db():
if sys.platform.startswith('darwin'):
# Mac OSX
local('postgres -D /usr/local/var/postgres -s')
@task
def backup():
cmd_dumpdata = 'python manage.py dumpdata '
with lcd(src_p):
local(
cmd_dumpdata + 'users.EmailUser data_sources.DataSource | '
'tee ../db_dump/user_sources.json'
)
local(
cmd_dumpdata + 'experiments | '
'tee ../db_dump/experiments.json'
)
local(
cmd_dumpdata + 'analyses.GenomeReference | '
'tee ../db_dump/genome_reference.json'
)
@task
def reborn():
with lcd(src_p.as_posix()):
db_dump_dir = Path(env.cwd, '../db_dump')
if not (
db_dump_dir.joinpath('user_sources.json').exists() and
db_dump_dir.joinpath('genome_reference.json').exists() and
db_dump_dir.joinpath('experiments.json').exists()
):
abort('Backup the import database content first!')
confirm('Destory and re-create the current database?', False)
local('dropdb biocloud')
local('createdb biocloud')
local('python manage.py migrate')
local('python manage.py loaddata ../db_dump/user_sources.json')
local('python manage.py loaddata ../db_dump/genome_reference.json')
local('python manage.py loaddata ../db_dump/experiments.json')
| // ... existing code ...
import sys
from pathlib import Path
from fabric.api import local, task, lcd, env
from fabric.contrib.console import confirm
from fabric.utils import abort
src_p = Path(env.real_fabfile).parent / 'src'
// ... modified code ...
local('postgres -D /usr/local/var/postgres -s')
@task
def backup():
cmd_dumpdata = 'python manage.py dumpdata '
with lcd(src_p):
local(
cmd_dumpdata + 'users.EmailUser data_sources.DataSource | '
'tee ../db_dump/user_sources.json'
)
local(
cmd_dumpdata + 'experiments | '
'tee ../db_dump/experiments.json'
)
local(
cmd_dumpdata + 'analyses.GenomeReference | '
'tee ../db_dump/genome_reference.json'
)
@task
def reborn():
with lcd(src_p.as_posix()):
db_dump_dir = Path(env.cwd, '../db_dump')
if not (
db_dump_dir.joinpath('user_sources.json').exists() and
db_dump_dir.joinpath('genome_reference.json').exists() and
db_dump_dir.joinpath('experiments.json').exists()
):
abort('Backup the import database content first!')
confirm('Destory and re-create the current database?', False)
local('dropdb biocloud')
local('createdb biocloud')
local('python manage.py migrate')
local('python manage.py loaddata ../db_dump/user_sources.json')
local('python manage.py loaddata ../db_dump/genome_reference.json')
local('python manage.py loaddata ../db_dump/experiments.json')
// ... rest of the code ... |
56e6ab84025f071c04701d3dc736b68e82361139 | apitestcase/testcase.py | apitestcase/testcase.py | import types
import unittest
import requests
class TestCase(object):
"""
Add assetion methods for HTTP Requests to TestCase
"""
hosts = []
def assertGet(self, endpoint="", status_code=200, body=""):
"""
Asserts GET requests on a given endpoint
"""
for host in self.hosts:
response = requests.get(host+endpoint)
self.assertEqual(response.status_code, status_code)
if isinstance(body, types.StringType):
self.assertIn(body, response.content)
elif isinstance(body, list):
for content in body:
self.assertIn(content, response.content)
| import unittest
import requests
class TestCase(object):
"""
Add assetion methods for HTTP Requests to TestCase
"""
hosts = []
def assertGet(self, endpoint="", status_code=200, body=""):
"""
Asserts GET requests on a given endpoint
"""
for host in self.hosts:
response = requests.get(host+endpoint)
self.assertEqual(response.status_code, status_code)
if isinstance(body, str):
self.assertIn(body, response.content)
elif isinstance(body, list):
for content in body:
self.assertIn(content, response.content)
| Change assertGet body check from StringType to str | Change assertGet body check from StringType to str
| Python | mit | bramwelt/apitestcase | - import types
import unittest
import requests
class TestCase(object):
"""
Add assetion methods for HTTP Requests to TestCase
"""
hosts = []
def assertGet(self, endpoint="", status_code=200, body=""):
"""
Asserts GET requests on a given endpoint
"""
for host in self.hosts:
response = requests.get(host+endpoint)
self.assertEqual(response.status_code, status_code)
- if isinstance(body, types.StringType):
+ if isinstance(body, str):
self.assertIn(body, response.content)
elif isinstance(body, list):
for content in body:
self.assertIn(content, response.content)
| Change assertGet body check from StringType to str | ## Code Before:
import types
import unittest
import requests
class TestCase(object):
"""
Add assetion methods for HTTP Requests to TestCase
"""
hosts = []
def assertGet(self, endpoint="", status_code=200, body=""):
"""
Asserts GET requests on a given endpoint
"""
for host in self.hosts:
response = requests.get(host+endpoint)
self.assertEqual(response.status_code, status_code)
if isinstance(body, types.StringType):
self.assertIn(body, response.content)
elif isinstance(body, list):
for content in body:
self.assertIn(content, response.content)
## Instruction:
Change assertGet body check from StringType to str
## Code After:
import unittest
import requests
class TestCase(object):
"""
Add assetion methods for HTTP Requests to TestCase
"""
hosts = []
def assertGet(self, endpoint="", status_code=200, body=""):
"""
Asserts GET requests on a given endpoint
"""
for host in self.hosts:
response = requests.get(host+endpoint)
self.assertEqual(response.status_code, status_code)
if isinstance(body, str):
self.assertIn(body, response.content)
elif isinstance(body, list):
for content in body:
self.assertIn(content, response.content)
| ...
import unittest
...
self.assertEqual(response.status_code, status_code)
if isinstance(body, str):
self.assertIn(body, response.content)
... |
153c832f083e8ec801ecb8dbddd2f8e79b735eed | utilities.py | utilities.py |
import pkg_resources
pkg_resources.require('aphla')
import aphla as ap
def get_pv_names(mode):
''' Given a certain ring mode as a string, return all available pvs '''
ap.machines.load(mode)
result = set()
elements = ap.getElements('*')
for element in elements:
pvs = element.pv()
if(len(pvs) > 0):
pv_name = pvs[0].split(':')[0]
result.add(pv_name)
return result
def get_pvs_from_file(filepath):
''' Return a list of pvs from a given file '''
with open(filepath) as f:
contents = f.read().splitlines()
return contents
|
import pkg_resources
pkg_resources.require('aphla')
import aphla as ap
def get_pv_names(mode):
''' Given a certain ring mode as a string, return all available pvs '''
ap.machines.load(mode)
result = set()
elements = ap.getElements('*')
for element in elements:
pvs = element.pv()
if(len(pvs) > 0):
pv_name = pvs[0].split(':')[0]
result.add(pv_name)
return result
def get_pvs_from_file(filepath):
''' Return a list of pvs from a given file '''
with open(filepath) as f:
contents = f.read().splitlines()
return contents
def write_pvs_to_file(filename, data):
''' Write given pvs to file '''
f = open(filename, 'w')
for element in data:
f.write(element, '\n')
f.close()
| Add utility function to write pvs to file | Add utility function to write pvs to file
| Python | apache-2.0 | razvanvasile/Work-Mini-Projects,razvanvasile/Work-Mini-Projects,razvanvasile/Work-Mini-Projects |
import pkg_resources
pkg_resources.require('aphla')
import aphla as ap
+
def get_pv_names(mode):
''' Given a certain ring mode as a string, return all available pvs '''
ap.machines.load(mode)
result = set()
elements = ap.getElements('*')
for element in elements:
pvs = element.pv()
if(len(pvs) > 0):
pv_name = pvs[0].split(':')[0]
result.add(pv_name)
return result
def get_pvs_from_file(filepath):
''' Return a list of pvs from a given file '''
with open(filepath) as f:
contents = f.read().splitlines()
return contents
+ def write_pvs_to_file(filename, data):
+ ''' Write given pvs to file '''
+ f = open(filename, 'w')
+ for element in data:
+ f.write(element, '\n')
+ f.close()
+ | Add utility function to write pvs to file | ## Code Before:
import pkg_resources
pkg_resources.require('aphla')
import aphla as ap
def get_pv_names(mode):
''' Given a certain ring mode as a string, return all available pvs '''
ap.machines.load(mode)
result = set()
elements = ap.getElements('*')
for element in elements:
pvs = element.pv()
if(len(pvs) > 0):
pv_name = pvs[0].split(':')[0]
result.add(pv_name)
return result
def get_pvs_from_file(filepath):
''' Return a list of pvs from a given file '''
with open(filepath) as f:
contents = f.read().splitlines()
return contents
## Instruction:
Add utility function to write pvs to file
## Code After:
import pkg_resources
pkg_resources.require('aphla')
import aphla as ap
def get_pv_names(mode):
''' Given a certain ring mode as a string, return all available pvs '''
ap.machines.load(mode)
result = set()
elements = ap.getElements('*')
for element in elements:
pvs = element.pv()
if(len(pvs) > 0):
pv_name = pvs[0].split(':')[0]
result.add(pv_name)
return result
def get_pvs_from_file(filepath):
''' Return a list of pvs from a given file '''
with open(filepath) as f:
contents = f.read().splitlines()
return contents
def write_pvs_to_file(filename, data):
''' Write given pvs to file '''
f = open(filename, 'w')
for element in data:
f.write(element, '\n')
f.close()
| ...
import aphla as ap
...
def write_pvs_to_file(filename, data):
''' Write given pvs to file '''
f = open(filename, 'w')
for element in data:
f.write(element, '\n')
f.close()
... |
82b45c3ec1344bed87ac7d572d82f43a4320492c | craigomatic/wsgi.py | craigomatic/wsgi.py | import os
from os.path import abspath, dirname
from sys import path
from django.core.wsgi import get_wsgi_application
SITE_ROOT = dirname(dirname(abspath(__file__)))
path.append(SITE_ROOT)
os.environ.setdefault("DJANGO_SETTINGS_MODULE", "craigomatic.settings")
application = get_wsgi_application()
| import os
from os.path import abspath, dirname
from sys import path
from django.core.wsgi import get_wsgi_application
from whitenoise.django import DjangoWhiteNoise
SITE_ROOT = dirname(dirname(abspath(__file__)))
path.append(SITE_ROOT)
os.environ.setdefault("DJANGO_SETTINGS_MODULE", "craigomatic.settings")
application = get_wsgi_application()
application = DjangoWhiteNoise(application)
| Integrate whitenoise with the Django application | Integrate whitenoise with the Django application
This allows Django to serve static files in production.
| Python | mit | rgreinho/craigomatic,rgreinho/craigomatic,rgreinho/craigomatic,rgreinho/craigomatic | import os
from os.path import abspath, dirname
from sys import path
from django.core.wsgi import get_wsgi_application
+ from whitenoise.django import DjangoWhiteNoise
+
SITE_ROOT = dirname(dirname(abspath(__file__)))
path.append(SITE_ROOT)
os.environ.setdefault("DJANGO_SETTINGS_MODULE", "craigomatic.settings")
application = get_wsgi_application()
+ application = DjangoWhiteNoise(application)
| Integrate whitenoise with the Django application | ## Code Before:
import os
from os.path import abspath, dirname
from sys import path
from django.core.wsgi import get_wsgi_application
SITE_ROOT = dirname(dirname(abspath(__file__)))
path.append(SITE_ROOT)
os.environ.setdefault("DJANGO_SETTINGS_MODULE", "craigomatic.settings")
application = get_wsgi_application()
## Instruction:
Integrate whitenoise with the Django application
## Code After:
import os
from os.path import abspath, dirname
from sys import path
from django.core.wsgi import get_wsgi_application
from whitenoise.django import DjangoWhiteNoise
SITE_ROOT = dirname(dirname(abspath(__file__)))
path.append(SITE_ROOT)
os.environ.setdefault("DJANGO_SETTINGS_MODULE", "craigomatic.settings")
application = get_wsgi_application()
application = DjangoWhiteNoise(application)
| ...
from whitenoise.django import DjangoWhiteNoise
SITE_ROOT = dirname(dirname(abspath(__file__)))
...
application = get_wsgi_application()
application = DjangoWhiteNoise(application)
... |
f0bd7658b961daceaace56e4ada415c5c9410d54 | UM/Operations/ScaleToBoundsOperation.py | UM/Operations/ScaleToBoundsOperation.py |
from UM.Operations.Operation import Operation
from UM.Math.Vector import Vector
## Operation subclass that will scale a node to fit within the bounds provided.
class ScaleToBoundsOperation(Operation):
def __init__(self, node, bounds):
super().__init__()
self._node = node
self._old_scale = node.getScale()
bbox = self._node.getBoundingBox()
largest_dimension = max(bbox.width, bbox.height, bbox.depth)
scale_factor = 1.0
if largest_dimension == bbox.width:
scale_factor = self._old_scale.x * (bounds.width / bbox.width)
elif largest_dimension == bbox.height:
scale_factor = self._old_scale.y * (bounds.height / bbox.height)
else:
scale_factor = self._old_scale.z * (bounds.depth / bbox.depth)
self._new_scale = Vector(scale_factor, scale_factor, scale_factor)
def undo(self):
self._node.setScale(self._old_scale)
def redo(self):
self._node.setPosition(Vector(0, 0, 0))
self._node.setScale(self._new_scale)
|
from UM.Operations.Operation import Operation
from UM.Math.Vector import Vector
## Operation subclass that will scale a node to fit within the bounds provided.
class ScaleToBoundsOperation(Operation):
def __init__(self, node, bounds):
super().__init__()
self._node = node
self._old_scale = node.getScale()
bbox = self._node.getBoundingBox()
largest_dimension = max(bbox.width, bbox.height, bbox.depth)
scale_factor = 1.0
if largest_dimension == bbox.depth:
scale_factor = self._old_scale.z * (bounds.depth / bbox.depth)
elif largest_dimension == bbox.width:
scale_factor = self._old_scale.x * (bounds.width / bbox.width)
elif largest_dimension == bbox.height:
scale_factor = self._old_scale.y * (bounds.height / bbox.height)
self._new_scale = Vector(scale_factor, scale_factor, scale_factor)
def undo(self):
self._node.setScale(self._old_scale)
def redo(self):
self._node.setPosition(Vector(0, 0, 0))
self._node.setScale(self._new_scale)
| Check depth before width since that is more likely to be the smaller dimension | Check depth before width since that is more likely to be the smaller dimension
Contributes to Asana issue 37107676459484
| Python | agpl-3.0 | onitake/Uranium,onitake/Uranium |
from UM.Operations.Operation import Operation
from UM.Math.Vector import Vector
## Operation subclass that will scale a node to fit within the bounds provided.
class ScaleToBoundsOperation(Operation):
def __init__(self, node, bounds):
super().__init__()
self._node = node
self._old_scale = node.getScale()
bbox = self._node.getBoundingBox()
largest_dimension = max(bbox.width, bbox.height, bbox.depth)
scale_factor = 1.0
+ if largest_dimension == bbox.depth:
+ scale_factor = self._old_scale.z * (bounds.depth / bbox.depth)
- if largest_dimension == bbox.width:
+ elif largest_dimension == bbox.width:
scale_factor = self._old_scale.x * (bounds.width / bbox.width)
elif largest_dimension == bbox.height:
scale_factor = self._old_scale.y * (bounds.height / bbox.height)
- else:
- scale_factor = self._old_scale.z * (bounds.depth / bbox.depth)
self._new_scale = Vector(scale_factor, scale_factor, scale_factor)
def undo(self):
self._node.setScale(self._old_scale)
def redo(self):
self._node.setPosition(Vector(0, 0, 0))
self._node.setScale(self._new_scale)
| Check depth before width since that is more likely to be the smaller dimension | ## Code Before:
from UM.Operations.Operation import Operation
from UM.Math.Vector import Vector
## Operation subclass that will scale a node to fit within the bounds provided.
class ScaleToBoundsOperation(Operation):
def __init__(self, node, bounds):
super().__init__()
self._node = node
self._old_scale = node.getScale()
bbox = self._node.getBoundingBox()
largest_dimension = max(bbox.width, bbox.height, bbox.depth)
scale_factor = 1.0
if largest_dimension == bbox.width:
scale_factor = self._old_scale.x * (bounds.width / bbox.width)
elif largest_dimension == bbox.height:
scale_factor = self._old_scale.y * (bounds.height / bbox.height)
else:
scale_factor = self._old_scale.z * (bounds.depth / bbox.depth)
self._new_scale = Vector(scale_factor, scale_factor, scale_factor)
def undo(self):
self._node.setScale(self._old_scale)
def redo(self):
self._node.setPosition(Vector(0, 0, 0))
self._node.setScale(self._new_scale)
## Instruction:
Check depth before width since that is more likely to be the smaller dimension
## Code After:
from UM.Operations.Operation import Operation
from UM.Math.Vector import Vector
## Operation subclass that will scale a node to fit within the bounds provided.
class ScaleToBoundsOperation(Operation):
def __init__(self, node, bounds):
super().__init__()
self._node = node
self._old_scale = node.getScale()
bbox = self._node.getBoundingBox()
largest_dimension = max(bbox.width, bbox.height, bbox.depth)
scale_factor = 1.0
if largest_dimension == bbox.depth:
scale_factor = self._old_scale.z * (bounds.depth / bbox.depth)
elif largest_dimension == bbox.width:
scale_factor = self._old_scale.x * (bounds.width / bbox.width)
elif largest_dimension == bbox.height:
scale_factor = self._old_scale.y * (bounds.height / bbox.height)
self._new_scale = Vector(scale_factor, scale_factor, scale_factor)
def undo(self):
self._node.setScale(self._old_scale)
def redo(self):
self._node.setPosition(Vector(0, 0, 0))
self._node.setScale(self._new_scale)
| // ... existing code ...
scale_factor = 1.0
if largest_dimension == bbox.depth:
scale_factor = self._old_scale.z * (bounds.depth / bbox.depth)
elif largest_dimension == bbox.width:
scale_factor = self._old_scale.x * (bounds.width / bbox.width)
// ... modified code ...
scale_factor = self._old_scale.y * (bounds.height / bbox.height)
// ... rest of the code ... |
57d3f5a78385b07fb4d7f91ac97edb6e9dc850aa | waterbutler/providers/osfstorage/metadata.py | waterbutler/providers/osfstorage/metadata.py | from waterbutler.core import metadata
class BaseOsfStorageMetadata:
@property
def provider(self):
return 'osfstorage'
class OsfStorageFileMetadata(BaseOsfStorageMetadata, metadata.BaseFileMetadata):
@property
def name(self):
return self.raw['name']
@property
def path(self):
if self.raw['path'][0] != '/':
return '/' + self.raw['path']
return self.raw['path']
@property
def modified(self):
return self.raw.get('modified')
@property
def size(self):
return self.raw.get('size')
@property
def content_type(self):
return None
@property
def extra(self):
return {
'downloads': self.raw['downloads']
}
class OsfStorageFolderMetadata(BaseOsfStorageMetadata, metadata.BaseFolderMetadata):
@property
def name(self):
return self.raw['name']
@property
def path(self):
return self.raw['path']
| from waterbutler.core import metadata
class BaseOsfStorageMetadata:
@property
def provider(self):
return 'osfstorage'
class OsfStorageFileMetadata(BaseOsfStorageMetadata, metadata.BaseFileMetadata):
@property
def name(self):
return self.raw['name']
@property
def path(self):
if self.raw['path'][0].startswith('/'):
return self.raw['path']
return '/' + self.raw['path']
@property
def modified(self):
return self.raw.get('modified')
@property
def size(self):
return self.raw.get('size')
@property
def content_type(self):
return None
@property
def extra(self):
return {
'downloads': self.raw['downloads']
}
class OsfStorageFolderMetadata(BaseOsfStorageMetadata, metadata.BaseFolderMetadata):
@property
def name(self):
return self.raw['name']
@property
def path(self):
return self.raw['path']
| Use startswith in favor of indexing | Use startswith in favor of indexing
| Python | apache-2.0 | CenterForOpenScience/waterbutler,cosenal/waterbutler,icereval/waterbutler,chrisseto/waterbutler,RCOSDP/waterbutler,rafaeldelucena/waterbutler,Johnetordoff/waterbutler,TomBaxter/waterbutler,felliott/waterbutler,rdhyee/waterbutler,kwierman/waterbutler,Ghalko/waterbutler,hmoco/waterbutler | from waterbutler.core import metadata
class BaseOsfStorageMetadata:
@property
def provider(self):
return 'osfstorage'
class OsfStorageFileMetadata(BaseOsfStorageMetadata, metadata.BaseFileMetadata):
@property
def name(self):
return self.raw['name']
@property
def path(self):
- if self.raw['path'][0] != '/':
+ if self.raw['path'][0].startswith('/'):
+ return self.raw['path']
- return '/' + self.raw['path']
+ return '/' + self.raw['path']
- return self.raw['path']
@property
def modified(self):
return self.raw.get('modified')
@property
def size(self):
return self.raw.get('size')
@property
def content_type(self):
return None
@property
def extra(self):
return {
'downloads': self.raw['downloads']
}
class OsfStorageFolderMetadata(BaseOsfStorageMetadata, metadata.BaseFolderMetadata):
@property
def name(self):
return self.raw['name']
@property
def path(self):
return self.raw['path']
| Use startswith in favor of indexing | ## Code Before:
from waterbutler.core import metadata
class BaseOsfStorageMetadata:
@property
def provider(self):
return 'osfstorage'
class OsfStorageFileMetadata(BaseOsfStorageMetadata, metadata.BaseFileMetadata):
@property
def name(self):
return self.raw['name']
@property
def path(self):
if self.raw['path'][0] != '/':
return '/' + self.raw['path']
return self.raw['path']
@property
def modified(self):
return self.raw.get('modified')
@property
def size(self):
return self.raw.get('size')
@property
def content_type(self):
return None
@property
def extra(self):
return {
'downloads': self.raw['downloads']
}
class OsfStorageFolderMetadata(BaseOsfStorageMetadata, metadata.BaseFolderMetadata):
@property
def name(self):
return self.raw['name']
@property
def path(self):
return self.raw['path']
## Instruction:
Use startswith in favor of indexing
## Code After:
from waterbutler.core import metadata
class BaseOsfStorageMetadata:
@property
def provider(self):
return 'osfstorage'
class OsfStorageFileMetadata(BaseOsfStorageMetadata, metadata.BaseFileMetadata):
@property
def name(self):
return self.raw['name']
@property
def path(self):
if self.raw['path'][0].startswith('/'):
return self.raw['path']
return '/' + self.raw['path']
@property
def modified(self):
return self.raw.get('modified')
@property
def size(self):
return self.raw.get('size')
@property
def content_type(self):
return None
@property
def extra(self):
return {
'downloads': self.raw['downloads']
}
class OsfStorageFolderMetadata(BaseOsfStorageMetadata, metadata.BaseFolderMetadata):
@property
def name(self):
return self.raw['name']
@property
def path(self):
return self.raw['path']
| # ... existing code ...
def path(self):
if self.raw['path'][0].startswith('/'):
return self.raw['path']
return '/' + self.raw['path']
# ... rest of the code ... |
a6061ef140e371101c3d04c5b85562586293eee8 | scrappyr/scraps/tests/test_models.py | scrappyr/scraps/tests/test_models.py | from test_plus.test import TestCase
from ..models import Scrap
class TestScrap(TestCase):
def test__str__(self):
scrap = Scrap(raw_title='hello')
assert str(scrap) == 'hello'
def test_html_title(self):
scrap = Scrap(raw_title='hello')
assert scrap.html_title == 'hello'
def test_html_title_bold(self):
scrap = Scrap(raw_title='**hello**')
assert scrap.html_title == '<strong>hello</strong>'
| from test_plus.test import TestCase
from ..models import Scrap
class TestScrap(TestCase):
def test__str__(self):
scrap = Scrap(raw_title='hello')
assert str(scrap) == 'hello'
def test_html_title(self):
scrap = Scrap(raw_title='hello')
assert scrap.html_title == 'hello'
def test_html_title_bold(self):
scrap = Scrap(raw_title='**hello**')
assert scrap.html_title == '<strong>hello</strong>'
def test_html_title_with_block_element_gets_escaped(self):
scrap = Scrap(raw_title='<div>hello</div>')
assert scrap.html_title == '<div>hello</div>'
| Add test of block-elements in scrap title | Add test of block-elements in scrap title
| Python | mit | tonysyu/scrappyr-app,tonysyu/scrappyr-app,tonysyu/scrappyr-app,tonysyu/scrappyr-app | from test_plus.test import TestCase
from ..models import Scrap
class TestScrap(TestCase):
def test__str__(self):
scrap = Scrap(raw_title='hello')
assert str(scrap) == 'hello'
def test_html_title(self):
scrap = Scrap(raw_title='hello')
assert scrap.html_title == 'hello'
def test_html_title_bold(self):
scrap = Scrap(raw_title='**hello**')
assert scrap.html_title == '<strong>hello</strong>'
+ def test_html_title_with_block_element_gets_escaped(self):
+ scrap = Scrap(raw_title='<div>hello</div>')
+ assert scrap.html_title == '<div>hello</div>'
+ | Add test of block-elements in scrap title | ## Code Before:
from test_plus.test import TestCase
from ..models import Scrap
class TestScrap(TestCase):
def test__str__(self):
scrap = Scrap(raw_title='hello')
assert str(scrap) == 'hello'
def test_html_title(self):
scrap = Scrap(raw_title='hello')
assert scrap.html_title == 'hello'
def test_html_title_bold(self):
scrap = Scrap(raw_title='**hello**')
assert scrap.html_title == '<strong>hello</strong>'
## Instruction:
Add test of block-elements in scrap title
## Code After:
from test_plus.test import TestCase
from ..models import Scrap
class TestScrap(TestCase):
def test__str__(self):
scrap = Scrap(raw_title='hello')
assert str(scrap) == 'hello'
def test_html_title(self):
scrap = Scrap(raw_title='hello')
assert scrap.html_title == 'hello'
def test_html_title_bold(self):
scrap = Scrap(raw_title='**hello**')
assert scrap.html_title == '<strong>hello</strong>'
def test_html_title_with_block_element_gets_escaped(self):
scrap = Scrap(raw_title='<div>hello</div>')
assert scrap.html_title == '<div>hello</div>'
| ...
assert scrap.html_title == '<strong>hello</strong>'
def test_html_title_with_block_element_gets_escaped(self):
scrap = Scrap(raw_title='<div>hello</div>')
assert scrap.html_title == '<div>hello</div>'
... |
2a0b1d070996bfb3d950d4fae70b264ddabc7d2f | sheldon/config.py | sheldon/config.py |
import os
class Config:
def __init__(self, prefix='SHELDON_'):
"""
Load config from environment variables.
:param prefix: string, all needed environment variables
starts from it.
Default - 'SHELDON_'. So, environment
variables will be looking like:
'SHELDON_BOT_NAME', 'SHELDON_TWITTER_KEY'
:return:
"""
# Bot config variables
self.variables = {}
for variable in os.environ:
if variable.startswith(prefix):
self.variables[variable] = os.environ[variable]
def get(self, variable, default_value):
"""
:param variable: string, needed variable
:param default_value: string, value that returns if
variable is not set
:return:
"""
if variable not in self.variables:
return default_value
return self.variables[variable]
|
import os
class Config:
def __init__(self, prefix='SHELDON_'):
"""
Load config from environment variables.
:param prefix: string, all needed environment variables
starts from it.
Default - 'SHELDON_'. So, environment
variables will be looking like:
'SHELDON_BOT_NAME', 'SHELDON_TWITTER_KEY'
:return:
"""
# Bot config variables
self.variables = {}
for variable in os.environ:
if variable.startswith(prefix):
self.variables[variable] = os.environ[variable]
def get(self, variable, default_value):
"""
Get variable value from environment
:param variable: string, needed variable
:param default_value: string, value that returns if
variable is not set
:return: variable value
"""
if variable not in self.variables:
return default_value
return self.variables[variable]
def get_installed_plugins(self):
"""
Return list of installed plugins from installed_plugins.txt
:return: list of strings with names of plugins
"""
plugins_file = open('installed_plugins.txt')
return plugins_file.readlines()
| Add function for getting installed plugins | Add function for getting installed plugins
| Python | mit | lises/sheldon |
import os
class Config:
def __init__(self, prefix='SHELDON_'):
"""
Load config from environment variables.
:param prefix: string, all needed environment variables
starts from it.
Default - 'SHELDON_'. So, environment
variables will be looking like:
'SHELDON_BOT_NAME', 'SHELDON_TWITTER_KEY'
:return:
"""
# Bot config variables
self.variables = {}
for variable in os.environ:
if variable.startswith(prefix):
self.variables[variable] = os.environ[variable]
def get(self, variable, default_value):
"""
+ Get variable value from environment
:param variable: string, needed variable
:param default_value: string, value that returns if
variable is not set
- :return:
+ :return: variable value
"""
if variable not in self.variables:
return default_value
return self.variables[variable]
+ def get_installed_plugins(self):
+ """
+ Return list of installed plugins from installed_plugins.txt
+ :return: list of strings with names of plugins
+ """
+ plugins_file = open('installed_plugins.txt')
+ return plugins_file.readlines()
+ | Add function for getting installed plugins | ## Code Before:
import os
class Config:
def __init__(self, prefix='SHELDON_'):
"""
Load config from environment variables.
:param prefix: string, all needed environment variables
starts from it.
Default - 'SHELDON_'. So, environment
variables will be looking like:
'SHELDON_BOT_NAME', 'SHELDON_TWITTER_KEY'
:return:
"""
# Bot config variables
self.variables = {}
for variable in os.environ:
if variable.startswith(prefix):
self.variables[variable] = os.environ[variable]
def get(self, variable, default_value):
"""
:param variable: string, needed variable
:param default_value: string, value that returns if
variable is not set
:return:
"""
if variable not in self.variables:
return default_value
return self.variables[variable]
## Instruction:
Add function for getting installed plugins
## Code After:
import os
class Config:
def __init__(self, prefix='SHELDON_'):
"""
Load config from environment variables.
:param prefix: string, all needed environment variables
starts from it.
Default - 'SHELDON_'. So, environment
variables will be looking like:
'SHELDON_BOT_NAME', 'SHELDON_TWITTER_KEY'
:return:
"""
# Bot config variables
self.variables = {}
for variable in os.environ:
if variable.startswith(prefix):
self.variables[variable] = os.environ[variable]
def get(self, variable, default_value):
"""
Get variable value from environment
:param variable: string, needed variable
:param default_value: string, value that returns if
variable is not set
:return: variable value
"""
if variable not in self.variables:
return default_value
return self.variables[variable]
def get_installed_plugins(self):
"""
Return list of installed plugins from installed_plugins.txt
:return: list of strings with names of plugins
"""
plugins_file = open('installed_plugins.txt')
return plugins_file.readlines()
| ...
"""
Get variable value from environment
...
variable is not set
:return: variable value
"""
...
def get_installed_plugins(self):
"""
Return list of installed plugins from installed_plugins.txt
:return: list of strings with names of plugins
"""
plugins_file = open('installed_plugins.txt')
return plugins_file.readlines()
... |
e51d35545d038b5cb7035cc74f39e4a5c2b0756a | thinglang/execution/classes.py | thinglang/execution/classes.py | from thinglang.lexer.symbols.base import LexicalIdentifier
class ThingInstance(object):
def __init__(self, cls):
self.cls = cls
self.methods = {
x.name: x for x in self.cls.children
}
self.members = {}
def __contains__(self, item):
return item in self.members or item in self.methods
def __getitem__(self, item):
return self.members.get(item) or self.methods.get(item)
def __str__(self):
return f'Thing<{self.cls}>(members={self.members}, methods={self.methods})' | from thinglang.lexer.symbols.base import LexicalIdentifier
class ThingInstance(object):
def __init__(self, cls):
self.cls = cls
self.methods = {
x.name: x for x in self.cls.children
}
self.members = {}
def __contains__(self, item):
return item in self.members or item in self.methods
def __getitem__(self, item):
return self.members.get(item) if item in self.members else self.methods[item]
def __str__(self):
return f'Thing<{self.cls}>(members={self.members}, methods={self.methods})' | Fix bug in ThingInstace __setitem__ | Fix bug in ThingInstace __setitem__
| Python | mit | ytanay/thinglang,ytanay/thinglang,ytanay/thinglang,ytanay/thinglang | from thinglang.lexer.symbols.base import LexicalIdentifier
class ThingInstance(object):
def __init__(self, cls):
self.cls = cls
self.methods = {
x.name: x for x in self.cls.children
}
self.members = {}
def __contains__(self, item):
return item in self.members or item in self.methods
def __getitem__(self, item):
- return self.members.get(item) or self.methods.get(item)
+ return self.members.get(item) if item in self.members else self.methods[item]
def __str__(self):
return f'Thing<{self.cls}>(members={self.members}, methods={self.methods})' | Fix bug in ThingInstace __setitem__ | ## Code Before:
from thinglang.lexer.symbols.base import LexicalIdentifier
class ThingInstance(object):
def __init__(self, cls):
self.cls = cls
self.methods = {
x.name: x for x in self.cls.children
}
self.members = {}
def __contains__(self, item):
return item in self.members or item in self.methods
def __getitem__(self, item):
return self.members.get(item) or self.methods.get(item)
def __str__(self):
return f'Thing<{self.cls}>(members={self.members}, methods={self.methods})'
## Instruction:
Fix bug in ThingInstace __setitem__
## Code After:
from thinglang.lexer.symbols.base import LexicalIdentifier
class ThingInstance(object):
def __init__(self, cls):
self.cls = cls
self.methods = {
x.name: x for x in self.cls.children
}
self.members = {}
def __contains__(self, item):
return item in self.members or item in self.methods
def __getitem__(self, item):
return self.members.get(item) if item in self.members else self.methods[item]
def __str__(self):
return f'Thing<{self.cls}>(members={self.members}, methods={self.methods})' | # ... existing code ...
def __getitem__(self, item):
return self.members.get(item) if item in self.members else self.methods[item]
# ... rest of the code ... |
a10e21a8fe811e896998ba510255592a966f0782 | infra/recipes/build_windows.py | infra/recipes/build_windows.py |
from recipe_engine.post_process import Filter
PYTHON_VERSION_COMPATIBILITY = "PY3"
DEPS = [
"crosvm",
"recipe_engine/buildbucket",
"recipe_engine/context",
"recipe_engine/properties",
"recipe_engine/step",
]
def RunSteps(api):
# Note: The recipe does work on linux as well, if the required dependencies have been installed
# on the host via ./tools/install-deps.
# This allows the build to be tested via `./recipe.py run build_windows`
with api.crosvm.host_build_context():
api.step(
"Build crosvm tests",
[
"vpython3",
"./tools/run_tests",
"--verbose",
"--build-only",
],
)
api.step(
"Run crosvm tests",
[
"vpython3",
"./tools/run_tests",
"--verbose",
],
)
def GenTests(api):
filter_steps = Filter("Build crosvm tests", "Run crosvm tests")
yield (
api.test(
"build",
api.buildbucket.ci_build(project="crosvm/crosvm"),
)
+ api.post_process(filter_steps)
)
|
from recipe_engine.post_process import Filter
PYTHON_VERSION_COMPATIBILITY = "PY3"
DEPS = [
"crosvm",
"recipe_engine/buildbucket",
"recipe_engine/context",
"recipe_engine/properties",
"recipe_engine/step",
]
def RunSteps(api):
# Note: The recipe does work on linux as well, if the required dependencies have been installed
# on the host via ./tools/install-deps.
# This allows the build to be tested via `./recipe.py run build_windows`
with api.crosvm.host_build_context():
api.step(
"Build crosvm tests",
[
"vpython3",
"./tools/run_tests",
"--verbose",
"--build-only",
],
)
api.step(
"Run crosvm tests",
[
"vpython3",
"./tools/run_tests",
"--verbose",
],
)
api.step(
"Clippy windows crosvm",
[
"vpython3",
"./tools/clippy",
],
)
def GenTests(api):
filter_steps = Filter("Build crosvm tests", "Run crosvm tests")
yield (
api.test(
"build",
api.buildbucket.ci_build(project="crosvm/crosvm"),
)
+ api.post_process(filter_steps)
)
| Enable clippy in windows LUCI | crosvm: Enable clippy in windows LUCI
For linux based systems, clippy continues to run in health_check
BUG=b:257249038
TEST=CQ
Change-Id: I39d3d45a0db72c61e79fd2c51b195b82c067a244
Reviewed-on: https://chromium-review.googlesource.com/c/crosvm/crosvm/+/3993934
Reviewed-by: Dennis Kempin <[email protected]>
Commit-Queue: Vikram Auradkar <[email protected]>
| Python | bsd-3-clause | google/crosvm,google/crosvm,google/crosvm,google/crosvm |
from recipe_engine.post_process import Filter
PYTHON_VERSION_COMPATIBILITY = "PY3"
DEPS = [
"crosvm",
"recipe_engine/buildbucket",
"recipe_engine/context",
"recipe_engine/properties",
"recipe_engine/step",
]
def RunSteps(api):
# Note: The recipe does work on linux as well, if the required dependencies have been installed
# on the host via ./tools/install-deps.
# This allows the build to be tested via `./recipe.py run build_windows`
with api.crosvm.host_build_context():
api.step(
"Build crosvm tests",
[
"vpython3",
"./tools/run_tests",
"--verbose",
"--build-only",
],
)
api.step(
"Run crosvm tests",
[
"vpython3",
"./tools/run_tests",
"--verbose",
],
)
+ api.step(
+ "Clippy windows crosvm",
+ [
+ "vpython3",
+ "./tools/clippy",
+ ],
+ )
def GenTests(api):
filter_steps = Filter("Build crosvm tests", "Run crosvm tests")
yield (
api.test(
"build",
api.buildbucket.ci_build(project="crosvm/crosvm"),
)
+ api.post_process(filter_steps)
)
| Enable clippy in windows LUCI | ## Code Before:
from recipe_engine.post_process import Filter
PYTHON_VERSION_COMPATIBILITY = "PY3"
DEPS = [
"crosvm",
"recipe_engine/buildbucket",
"recipe_engine/context",
"recipe_engine/properties",
"recipe_engine/step",
]
def RunSteps(api):
# Note: The recipe does work on linux as well, if the required dependencies have been installed
# on the host via ./tools/install-deps.
# This allows the build to be tested via `./recipe.py run build_windows`
with api.crosvm.host_build_context():
api.step(
"Build crosvm tests",
[
"vpython3",
"./tools/run_tests",
"--verbose",
"--build-only",
],
)
api.step(
"Run crosvm tests",
[
"vpython3",
"./tools/run_tests",
"--verbose",
],
)
def GenTests(api):
filter_steps = Filter("Build crosvm tests", "Run crosvm tests")
yield (
api.test(
"build",
api.buildbucket.ci_build(project="crosvm/crosvm"),
)
+ api.post_process(filter_steps)
)
## Instruction:
Enable clippy in windows LUCI
## Code After:
from recipe_engine.post_process import Filter
PYTHON_VERSION_COMPATIBILITY = "PY3"
DEPS = [
"crosvm",
"recipe_engine/buildbucket",
"recipe_engine/context",
"recipe_engine/properties",
"recipe_engine/step",
]
def RunSteps(api):
# Note: The recipe does work on linux as well, if the required dependencies have been installed
# on the host via ./tools/install-deps.
# This allows the build to be tested via `./recipe.py run build_windows`
with api.crosvm.host_build_context():
api.step(
"Build crosvm tests",
[
"vpython3",
"./tools/run_tests",
"--verbose",
"--build-only",
],
)
api.step(
"Run crosvm tests",
[
"vpython3",
"./tools/run_tests",
"--verbose",
],
)
api.step(
"Clippy windows crosvm",
[
"vpython3",
"./tools/clippy",
],
)
def GenTests(api):
filter_steps = Filter("Build crosvm tests", "Run crosvm tests")
yield (
api.test(
"build",
api.buildbucket.ci_build(project="crosvm/crosvm"),
)
+ api.post_process(filter_steps)
)
| # ... existing code ...
)
api.step(
"Clippy windows crosvm",
[
"vpython3",
"./tools/clippy",
],
)
# ... rest of the code ... |
fc39c6afa49a312413468dfdffcc2de94bb7d78e | tests/test_runner.py | tests/test_runner.py | import unittest
from mo.runner import Variable
class TestVariable(unittest.TestCase):
def test_default(self):
v = Variable('name', {'default': 'default'})
self.assertEqual(v.value, 'default')
self.assertEqual(str(v), 'default')
def test_value(self):
v = Variable('name', {'default': 'default'}, 'value')
self.assertEqual(v.value, 'value')
self.assertEqual(str(v), 'value')
def test_str(self):
v = Variable('name', {'default': 'abc'})
self.assertEqual(str(v), v.value)
| import unittest
from mo.runner import Task, Variable
class TestVariable(unittest.TestCase):
def test_default(self):
v = Variable('name', {'default': 'default'})
self.assertEqual(v.value, 'default')
self.assertEqual(str(v), 'default')
def test_value(self):
v = Variable('name', {'default': 'default'}, 'value')
self.assertEqual(v.value, 'value')
self.assertEqual(str(v), 'value')
def test_str(self):
v = Variable('name', {'default': 'abc'})
self.assertEqual(str(v), v.value)
class TestTask(unittest.TestCase):
def test_variables(self):
t = Task('name', {'description': '', 'command': '{{ v }}'},
{'v': 'variable'})
self.assertEqual(t.commands[0], 'variable')
| Add some more tests for tasks | Add some more tests for tasks
| Python | mit | thomasleese/mo | import unittest
- from mo.runner import Variable
+ from mo.runner import Task, Variable
class TestVariable(unittest.TestCase):
def test_default(self):
v = Variable('name', {'default': 'default'})
self.assertEqual(v.value, 'default')
self.assertEqual(str(v), 'default')
def test_value(self):
v = Variable('name', {'default': 'default'}, 'value')
self.assertEqual(v.value, 'value')
self.assertEqual(str(v), 'value')
def test_str(self):
v = Variable('name', {'default': 'abc'})
self.assertEqual(str(v), v.value)
+
+ class TestTask(unittest.TestCase):
+ def test_variables(self):
+ t = Task('name', {'description': '', 'command': '{{ v }}'},
+ {'v': 'variable'})
+ self.assertEqual(t.commands[0], 'variable')
+ | Add some more tests for tasks | ## Code Before:
import unittest
from mo.runner import Variable
class TestVariable(unittest.TestCase):
def test_default(self):
v = Variable('name', {'default': 'default'})
self.assertEqual(v.value, 'default')
self.assertEqual(str(v), 'default')
def test_value(self):
v = Variable('name', {'default': 'default'}, 'value')
self.assertEqual(v.value, 'value')
self.assertEqual(str(v), 'value')
def test_str(self):
v = Variable('name', {'default': 'abc'})
self.assertEqual(str(v), v.value)
## Instruction:
Add some more tests for tasks
## Code After:
import unittest
from mo.runner import Task, Variable
class TestVariable(unittest.TestCase):
def test_default(self):
v = Variable('name', {'default': 'default'})
self.assertEqual(v.value, 'default')
self.assertEqual(str(v), 'default')
def test_value(self):
v = Variable('name', {'default': 'default'}, 'value')
self.assertEqual(v.value, 'value')
self.assertEqual(str(v), 'value')
def test_str(self):
v = Variable('name', {'default': 'abc'})
self.assertEqual(str(v), v.value)
class TestTask(unittest.TestCase):
def test_variables(self):
t = Task('name', {'description': '', 'command': '{{ v }}'},
{'v': 'variable'})
self.assertEqual(t.commands[0], 'variable')
| // ... existing code ...
from mo.runner import Task, Variable
// ... modified code ...
self.assertEqual(str(v), v.value)
class TestTask(unittest.TestCase):
def test_variables(self):
t = Task('name', {'description': '', 'command': '{{ v }}'},
{'v': 'variable'})
self.assertEqual(t.commands[0], 'variable')
// ... rest of the code ... |
7643635278fc1c92289e8fdd456614ce85a2c2f3 | addons/osfstorage/models.py | addons/osfstorage/models.py | import logging
from addons.base.models import BaseNodeSettings, BaseStorageAddon
logger = logging.getLogger(__name__)
class NodeSettings(BaseStorageAddon, BaseNodeSettings):
pass
| import logging
from addons.base.models import BaseNodeSettings, BaseStorageAddon
logger = logging.getLogger(__name__)
class NodeSettings(BaseStorageAddon, BaseNodeSettings):
# Required overrides
complete = True
has_auth = True
| Add required overrides to osfstorage.NodeSettings | Add required overrides to osfstorage.NodeSettings
| Python | apache-2.0 | felliott/osf.io,CenterForOpenScience/osf.io,laurenrevere/osf.io,cslzchen/osf.io,monikagrabowska/osf.io,caneruguz/osf.io,Nesiehr/osf.io,alexschiller/osf.io,aaxelb/osf.io,caseyrollins/osf.io,Johnetordoff/osf.io,caneruguz/osf.io,mluo613/osf.io,chrisseto/osf.io,monikagrabowska/osf.io,cslzchen/osf.io,monikagrabowska/osf.io,monikagrabowska/osf.io,saradbowman/osf.io,leb2dg/osf.io,cslzchen/osf.io,alexschiller/osf.io,hmoco/osf.io,mfraezz/osf.io,saradbowman/osf.io,CenterForOpenScience/osf.io,mluo613/osf.io,brianjgeiger/osf.io,acshi/osf.io,acshi/osf.io,leb2dg/osf.io,Nesiehr/osf.io,erinspace/osf.io,erinspace/osf.io,chrisseto/osf.io,alexschiller/osf.io,binoculars/osf.io,icereval/osf.io,brianjgeiger/osf.io,alexschiller/osf.io,Nesiehr/osf.io,crcresearch/osf.io,adlius/osf.io,adlius/osf.io,chrisseto/osf.io,sloria/osf.io,sloria/osf.io,chennan47/osf.io,laurenrevere/osf.io,HalcyonChimera/osf.io,laurenrevere/osf.io,chennan47/osf.io,leb2dg/osf.io,icereval/osf.io,felliott/osf.io,mattclark/osf.io,HalcyonChimera/osf.io,acshi/osf.io,erinspace/osf.io,baylee-d/osf.io,TomBaxter/osf.io,cwisecarver/osf.io,TomBaxter/osf.io,mfraezz/osf.io,leb2dg/osf.io,baylee-d/osf.io,acshi/osf.io,aaxelb/osf.io,icereval/osf.io,monikagrabowska/osf.io,Johnetordoff/osf.io,caneruguz/osf.io,pattisdr/osf.io,hmoco/osf.io,caseyrollins/osf.io,chennan47/osf.io,TomBaxter/osf.io,caseyrollins/osf.io,felliott/osf.io,Johnetordoff/osf.io,mluo613/osf.io,binoculars/osf.io,mluo613/osf.io,acshi/osf.io,adlius/osf.io,HalcyonChimera/osf.io,baylee-d/osf.io,Johnetordoff/osf.io,cslzchen/osf.io,aaxelb/osf.io,sloria/osf.io,cwisecarver/osf.io,mluo613/osf.io,adlius/osf.io,alexschiller/osf.io,cwisecarver/osf.io,CenterForOpenScience/osf.io,cwisecarver/osf.io,crcresearch/osf.io,mfraezz/osf.io,caneruguz/osf.io,hmoco/osf.io,brianjgeiger/osf.io,brianjgeiger/osf.io,mfraezz/osf.io,mattclark/osf.io,HalcyonChimera/osf.io,CenterForOpenScience/osf.io,Nesiehr/osf.io,mattclark/osf.io,chrisseto/osf.io,crcresearch/osf.io,pattisdr/osf.io,binoculars/osf.io,pattisdr/osf.io,felliott/osf.io,aaxelb/osf.io,hmoco/osf.io | import logging
from addons.base.models import BaseNodeSettings, BaseStorageAddon
logger = logging.getLogger(__name__)
class NodeSettings(BaseStorageAddon, BaseNodeSettings):
- pass
+ # Required overrides
+ complete = True
+ has_auth = True
| Add required overrides to osfstorage.NodeSettings | ## Code Before:
import logging
from addons.base.models import BaseNodeSettings, BaseStorageAddon
logger = logging.getLogger(__name__)
class NodeSettings(BaseStorageAddon, BaseNodeSettings):
pass
## Instruction:
Add required overrides to osfstorage.NodeSettings
## Code After:
import logging
from addons.base.models import BaseNodeSettings, BaseStorageAddon
logger = logging.getLogger(__name__)
class NodeSettings(BaseStorageAddon, BaseNodeSettings):
# Required overrides
complete = True
has_auth = True
| # ... existing code ...
class NodeSettings(BaseStorageAddon, BaseNodeSettings):
# Required overrides
complete = True
has_auth = True
# ... rest of the code ... |
28ecf02c3d08eae725512e1563cf74f1831bd02d | gears/engines/base.py | gears/engines/base.py | import subprocess
from functools import wraps
class EngineProcessFailed(Exception):
pass
class BaseEngine(object):
result_mimetype = None
@classmethod
def as_engine(cls, **initkwargs):
@wraps(cls, updated=())
def engine(asset):
instance = engine.engine_class(**initkwargs)
return instance.process(asset)
engine.engine_class = cls
engine.result_mimetype = cls.result_mimetype
return engine
def process(self, asset):
raise NotImplementedError()
class ExecEngine(BaseEngine):
executable = None
params = []
def __init__(self, executable=None):
if executable is not None:
self.executable = executable
def process(self, asset):
self.asset = asset
p = subprocess.Popen(
args=self.get_args(),
stdin=subprocess.PIPE,
stdout=subprocess.PIPE,
stderr=subprocess.PIPE)
output, errors = p.communicate(input=asset.processed_source)
if p.returncode != 0:
raise EngineProcessFailed(errors)
asset.processed_source = output
def get_args(self):
return [self.executable] + self.params
| import subprocess
from functools import wraps
class EngineProcessFailed(Exception):
pass
class BaseEngine(object):
result_mimetype = None
@classmethod
def as_engine(cls, **initkwargs):
@wraps(cls, updated=())
def engine(asset):
instance = engine.engine_class(**initkwargs)
return instance.process(asset)
engine.engine_class = cls
engine.result_mimetype = cls.result_mimetype
return engine
def process(self, asset):
raise NotImplementedError()
class ExecEngine(BaseEngine):
executable = None
params = []
def __init__(self, executable=None):
if executable is not None:
self.executable = executable
def process(self, asset):
self.asset = asset
p = subprocess.Popen(
args=self.get_args(),
stdin=subprocess.PIPE,
stdout=subprocess.PIPE,
stderr=subprocess.PIPE)
output, errors = p.communicate(input=asset.processed_source.encode('utf-8'))
if p.returncode != 0:
raise EngineProcessFailed(errors)
asset.processed_source = output.decode('utf-8')
def get_args(self):
return [self.executable] + self.params
| Fix unicode support in ExecEngine | Fix unicode support in ExecEngine
| Python | isc | gears/gears,gears/gears,gears/gears | import subprocess
from functools import wraps
class EngineProcessFailed(Exception):
pass
class BaseEngine(object):
result_mimetype = None
@classmethod
def as_engine(cls, **initkwargs):
@wraps(cls, updated=())
def engine(asset):
instance = engine.engine_class(**initkwargs)
return instance.process(asset)
engine.engine_class = cls
engine.result_mimetype = cls.result_mimetype
return engine
def process(self, asset):
raise NotImplementedError()
class ExecEngine(BaseEngine):
executable = None
params = []
def __init__(self, executable=None):
if executable is not None:
self.executable = executable
def process(self, asset):
self.asset = asset
p = subprocess.Popen(
args=self.get_args(),
stdin=subprocess.PIPE,
stdout=subprocess.PIPE,
stderr=subprocess.PIPE)
- output, errors = p.communicate(input=asset.processed_source)
+ output, errors = p.communicate(input=asset.processed_source.encode('utf-8'))
if p.returncode != 0:
raise EngineProcessFailed(errors)
- asset.processed_source = output
+ asset.processed_source = output.decode('utf-8')
def get_args(self):
return [self.executable] + self.params
| Fix unicode support in ExecEngine | ## Code Before:
import subprocess
from functools import wraps
class EngineProcessFailed(Exception):
pass
class BaseEngine(object):
result_mimetype = None
@classmethod
def as_engine(cls, **initkwargs):
@wraps(cls, updated=())
def engine(asset):
instance = engine.engine_class(**initkwargs)
return instance.process(asset)
engine.engine_class = cls
engine.result_mimetype = cls.result_mimetype
return engine
def process(self, asset):
raise NotImplementedError()
class ExecEngine(BaseEngine):
executable = None
params = []
def __init__(self, executable=None):
if executable is not None:
self.executable = executable
def process(self, asset):
self.asset = asset
p = subprocess.Popen(
args=self.get_args(),
stdin=subprocess.PIPE,
stdout=subprocess.PIPE,
stderr=subprocess.PIPE)
output, errors = p.communicate(input=asset.processed_source)
if p.returncode != 0:
raise EngineProcessFailed(errors)
asset.processed_source = output
def get_args(self):
return [self.executable] + self.params
## Instruction:
Fix unicode support in ExecEngine
## Code After:
import subprocess
from functools import wraps
class EngineProcessFailed(Exception):
pass
class BaseEngine(object):
result_mimetype = None
@classmethod
def as_engine(cls, **initkwargs):
@wraps(cls, updated=())
def engine(asset):
instance = engine.engine_class(**initkwargs)
return instance.process(asset)
engine.engine_class = cls
engine.result_mimetype = cls.result_mimetype
return engine
def process(self, asset):
raise NotImplementedError()
class ExecEngine(BaseEngine):
executable = None
params = []
def __init__(self, executable=None):
if executable is not None:
self.executable = executable
def process(self, asset):
self.asset = asset
p = subprocess.Popen(
args=self.get_args(),
stdin=subprocess.PIPE,
stdout=subprocess.PIPE,
stderr=subprocess.PIPE)
output, errors = p.communicate(input=asset.processed_source.encode('utf-8'))
if p.returncode != 0:
raise EngineProcessFailed(errors)
asset.processed_source = output.decode('utf-8')
def get_args(self):
return [self.executable] + self.params
| # ... existing code ...
stderr=subprocess.PIPE)
output, errors = p.communicate(input=asset.processed_source.encode('utf-8'))
if p.returncode != 0:
# ... modified code ...
raise EngineProcessFailed(errors)
asset.processed_source = output.decode('utf-8')
# ... rest of the code ... |
03a54ea1920a3716e9c8d326f5c4c408f45b7d08 | apps/api/urls.py | apps/api/urls.py | from django.conf.urls import url, include
from rest_framework.routers import DefaultRouter
from rest_framework_jwt import views as jwt_views
from . import views
router = DefaultRouter()
router.register(r'quotes', views.QuoteViewSet)
router.register(r'authors', views.AuthorViewSet)
router.register(r'categories', views.CategoryViewSet)
router.register(r'tags', views.TagViewSet)
urlpatterns = [
url(r'^docs/$', views.schema_view),
url(r'^', include(router.urls)),
url(r'^token/new/$', jwt_views.obtain_jwt_token),
url(r'^token/refresh/$', jwt_views.refresh_jwt_token),
url(r'^filters/$', views.FiltersOptionsView.as_view()),
url(r'^templates/(?P<page>[-\w]+.html)/$', views.AngularTemplateView.as_view()),
]
| from django.conf.urls import url, include
from rest_framework.routers import DefaultRouter
from rest_framework_jwt import views as jwt_views
from . import views
router = DefaultRouter()
router.register(r'quotes', views.QuoteViewSet)
router.register(r'authors', views.AuthorViewSet)
router.register(r'categories', views.CategoryViewSet)
router.register(r'tags', views.TagViewSet)
urlpatterns = [
url(r'^docs/$', views.schema_view),
url(r'^', include(router.urls)),
url(r'^token/new/$', jwt_views.ObtainJSONWebToken.as_view()),
url(r'^token/refresh/$', jwt_views.RefreshJSONWebToken.as_view()),
url(r'^token/verify/$', jwt_views.VerifyJSONWebToken.as_view()),
url(r'^filters/$', views.FiltersOptionsView.as_view()),
url(r'^templates/(?P<page>[-\w]+.html)/$', views.AngularTemplateView.as_view()),
]
| Add option to verify jwt token | Add option to verify jwt token
| Python | bsd-3-clause | lucifurtun/myquotes,lucifurtun/myquotes,lucifurtun/myquotes,lucifurtun/myquotes | from django.conf.urls import url, include
from rest_framework.routers import DefaultRouter
from rest_framework_jwt import views as jwt_views
from . import views
router = DefaultRouter()
router.register(r'quotes', views.QuoteViewSet)
router.register(r'authors', views.AuthorViewSet)
router.register(r'categories', views.CategoryViewSet)
router.register(r'tags', views.TagViewSet)
urlpatterns = [
url(r'^docs/$', views.schema_view),
url(r'^', include(router.urls)),
- url(r'^token/new/$', jwt_views.obtain_jwt_token),
+ url(r'^token/new/$', jwt_views.ObtainJSONWebToken.as_view()),
- url(r'^token/refresh/$', jwt_views.refresh_jwt_token),
+ url(r'^token/refresh/$', jwt_views.RefreshJSONWebToken.as_view()),
+ url(r'^token/verify/$', jwt_views.VerifyJSONWebToken.as_view()),
url(r'^filters/$', views.FiltersOptionsView.as_view()),
url(r'^templates/(?P<page>[-\w]+.html)/$', views.AngularTemplateView.as_view()),
]
| Add option to verify jwt token | ## Code Before:
from django.conf.urls import url, include
from rest_framework.routers import DefaultRouter
from rest_framework_jwt import views as jwt_views
from . import views
router = DefaultRouter()
router.register(r'quotes', views.QuoteViewSet)
router.register(r'authors', views.AuthorViewSet)
router.register(r'categories', views.CategoryViewSet)
router.register(r'tags', views.TagViewSet)
urlpatterns = [
url(r'^docs/$', views.schema_view),
url(r'^', include(router.urls)),
url(r'^token/new/$', jwt_views.obtain_jwt_token),
url(r'^token/refresh/$', jwt_views.refresh_jwt_token),
url(r'^filters/$', views.FiltersOptionsView.as_view()),
url(r'^templates/(?P<page>[-\w]+.html)/$', views.AngularTemplateView.as_view()),
]
## Instruction:
Add option to verify jwt token
## Code After:
from django.conf.urls import url, include
from rest_framework.routers import DefaultRouter
from rest_framework_jwt import views as jwt_views
from . import views
router = DefaultRouter()
router.register(r'quotes', views.QuoteViewSet)
router.register(r'authors', views.AuthorViewSet)
router.register(r'categories', views.CategoryViewSet)
router.register(r'tags', views.TagViewSet)
urlpatterns = [
url(r'^docs/$', views.schema_view),
url(r'^', include(router.urls)),
url(r'^token/new/$', jwt_views.ObtainJSONWebToken.as_view()),
url(r'^token/refresh/$', jwt_views.RefreshJSONWebToken.as_view()),
url(r'^token/verify/$', jwt_views.VerifyJSONWebToken.as_view()),
url(r'^filters/$', views.FiltersOptionsView.as_view()),
url(r'^templates/(?P<page>[-\w]+.html)/$', views.AngularTemplateView.as_view()),
]
| # ... existing code ...
url(r'^', include(router.urls)),
url(r'^token/new/$', jwt_views.ObtainJSONWebToken.as_view()),
url(r'^token/refresh/$', jwt_views.RefreshJSONWebToken.as_view()),
url(r'^token/verify/$', jwt_views.VerifyJSONWebToken.as_view()),
url(r'^filters/$', views.FiltersOptionsView.as_view()),
# ... rest of the code ... |
6f6f6e183b574f8505b53ddb7651c8766992b953 | pywikibot/families/lingualibre_family.py | pywikibot/families/lingualibre_family.py | """Family module for Lingua Libre."""
#
# (C) Pywikibot team, 2021
#
# Distributed under the terms of the MIT license.
#
from pywikibot import family
# The Lingua Libre family
class Family(family.WikimediaFamily):
"""Family class for Lingua Libre.
*New in version 6.5.*
"""
name = 'lingualibre'
langs = {
'lingualibre': 'lingualibre.org'
}
interwiki_forward = 'wikipedia'
def scriptpath(self, code):
"""Return the script path for this family."""
return ''
| """Family module for Lingua Libre."""
#
# (C) Pywikibot team, 2021
#
# Distributed under the terms of the MIT license.
#
from pywikibot import family
# The Lingua Libre family
class Family(family.WikimediaFamily):
"""Family class for Lingua Libre.
*New in version 6.5.*
"""
name = 'lingualibre'
langs = {
'lingualibre': 'lingualibre.org'
}
interwiki_forward = 'wikipedia'
def scriptpath(self, code):
"""Return the script path for this family."""
return ''
def interface(self, code):
"""Return 'DataSite'."""
return 'DataSite'
| Allow to request for item on Lingua Libre | Allow to request for item on Lingua Libre
Bug: T286303
Change-Id: Ic0d8824d1bf326f2182fbb64d0cc2ed77f82fd4c
| Python | mit | wikimedia/pywikibot-core,wikimedia/pywikibot-core | """Family module for Lingua Libre."""
#
# (C) Pywikibot team, 2021
#
# Distributed under the terms of the MIT license.
#
from pywikibot import family
# The Lingua Libre family
class Family(family.WikimediaFamily):
"""Family class for Lingua Libre.
*New in version 6.5.*
"""
name = 'lingualibre'
langs = {
'lingualibre': 'lingualibre.org'
}
interwiki_forward = 'wikipedia'
def scriptpath(self, code):
"""Return the script path for this family."""
return ''
+ def interface(self, code):
+ """Return 'DataSite'."""
+ return 'DataSite'
+ | Allow to request for item on Lingua Libre | ## Code Before:
"""Family module for Lingua Libre."""
#
# (C) Pywikibot team, 2021
#
# Distributed under the terms of the MIT license.
#
from pywikibot import family
# The Lingua Libre family
class Family(family.WikimediaFamily):
"""Family class for Lingua Libre.
*New in version 6.5.*
"""
name = 'lingualibre'
langs = {
'lingualibre': 'lingualibre.org'
}
interwiki_forward = 'wikipedia'
def scriptpath(self, code):
"""Return the script path for this family."""
return ''
## Instruction:
Allow to request for item on Lingua Libre
## Code After:
"""Family module for Lingua Libre."""
#
# (C) Pywikibot team, 2021
#
# Distributed under the terms of the MIT license.
#
from pywikibot import family
# The Lingua Libre family
class Family(family.WikimediaFamily):
"""Family class for Lingua Libre.
*New in version 6.5.*
"""
name = 'lingualibre'
langs = {
'lingualibre': 'lingualibre.org'
}
interwiki_forward = 'wikipedia'
def scriptpath(self, code):
"""Return the script path for this family."""
return ''
def interface(self, code):
"""Return 'DataSite'."""
return 'DataSite'
| // ... existing code ...
return ''
def interface(self, code):
"""Return 'DataSite'."""
return 'DataSite'
// ... rest of the code ... |
faf35a814d045ce3d71921ed0d4ac268d5a9811c | app/notify_client/provider_client.py | app/notify_client/provider_client.py |
from app.notify_client import _attach_current_user, NotifyAdminAPIClient
class ProviderClient(NotifyAdminAPIClient):
def __init__(self):
super().__init__("a", "b", "c")
def init_app(self, app):
self.base_url = app.config['API_HOST_NAME']
self.service_id = app.config['ADMIN_CLIENT_USER_NAME']
self.api_key = app.config['ADMIN_CLIENT_SECRET']
def get_all_providers(self):
return self.get(
url='/provider-details'
)
def get_provider_by_id(self, provider_id):
return self.get(
url='/provider-details/{}'.format(provider_id)
)
def update_provider(self, provider_id, priority):
data = {
"priority": priority
}
data = _attach_current_user(data)
return self.post(url='/provider-details/{}'.format(provider_id), data=data)
|
from app.notify_client import _attach_current_user, NotifyAdminAPIClient
class ProviderClient(NotifyAdminAPIClient):
def __init__(self):
super().__init__("a", "b", "c")
def init_app(self, app):
self.base_url = app.config['API_HOST_NAME']
self.service_id = app.config['ADMIN_CLIENT_USER_NAME']
self.api_key = app.config['ADMIN_CLIENT_SECRET']
def get_all_providers(self):
return self.get(
url='/provider-details'
)
def get_provider_by_id(self, provider_id):
return self.get(
url='/provider-details/{}'.format(provider_id)
)
def get_provider_versions(self, provider_id):
return self.get(
url='/provider-details/{}/versions'.format(provider_id)
)
def update_provider(self, provider_id, priority):
data = {
"priority": priority
}
data = _attach_current_user(data)
return self.post(url='/provider-details/{}'.format(provider_id), data=data)
| Add provider client method to get provider version history | Add provider client method to get provider version history
| Python | mit | gov-cjwaszczuk/notifications-admin,alphagov/notifications-admin,alphagov/notifications-admin,alphagov/notifications-admin,alphagov/notifications-admin,gov-cjwaszczuk/notifications-admin,gov-cjwaszczuk/notifications-admin,gov-cjwaszczuk/notifications-admin |
from app.notify_client import _attach_current_user, NotifyAdminAPIClient
class ProviderClient(NotifyAdminAPIClient):
def __init__(self):
super().__init__("a", "b", "c")
def init_app(self, app):
self.base_url = app.config['API_HOST_NAME']
self.service_id = app.config['ADMIN_CLIENT_USER_NAME']
self.api_key = app.config['ADMIN_CLIENT_SECRET']
def get_all_providers(self):
return self.get(
url='/provider-details'
)
def get_provider_by_id(self, provider_id):
return self.get(
url='/provider-details/{}'.format(provider_id)
)
+ def get_provider_versions(self, provider_id):
+ return self.get(
+ url='/provider-details/{}/versions'.format(provider_id)
+ )
+
def update_provider(self, provider_id, priority):
data = {
"priority": priority
}
data = _attach_current_user(data)
return self.post(url='/provider-details/{}'.format(provider_id), data=data)
| Add provider client method to get provider version history | ## Code Before:
from app.notify_client import _attach_current_user, NotifyAdminAPIClient
class ProviderClient(NotifyAdminAPIClient):
def __init__(self):
super().__init__("a", "b", "c")
def init_app(self, app):
self.base_url = app.config['API_HOST_NAME']
self.service_id = app.config['ADMIN_CLIENT_USER_NAME']
self.api_key = app.config['ADMIN_CLIENT_SECRET']
def get_all_providers(self):
return self.get(
url='/provider-details'
)
def get_provider_by_id(self, provider_id):
return self.get(
url='/provider-details/{}'.format(provider_id)
)
def update_provider(self, provider_id, priority):
data = {
"priority": priority
}
data = _attach_current_user(data)
return self.post(url='/provider-details/{}'.format(provider_id), data=data)
## Instruction:
Add provider client method to get provider version history
## Code After:
from app.notify_client import _attach_current_user, NotifyAdminAPIClient
class ProviderClient(NotifyAdminAPIClient):
def __init__(self):
super().__init__("a", "b", "c")
def init_app(self, app):
self.base_url = app.config['API_HOST_NAME']
self.service_id = app.config['ADMIN_CLIENT_USER_NAME']
self.api_key = app.config['ADMIN_CLIENT_SECRET']
def get_all_providers(self):
return self.get(
url='/provider-details'
)
def get_provider_by_id(self, provider_id):
return self.get(
url='/provider-details/{}'.format(provider_id)
)
def get_provider_versions(self, provider_id):
return self.get(
url='/provider-details/{}/versions'.format(provider_id)
)
def update_provider(self, provider_id, priority):
data = {
"priority": priority
}
data = _attach_current_user(data)
return self.post(url='/provider-details/{}'.format(provider_id), data=data)
| // ... existing code ...
def get_provider_versions(self, provider_id):
return self.get(
url='/provider-details/{}/versions'.format(provider_id)
)
def update_provider(self, provider_id, priority):
// ... rest of the code ... |
b42dddaa45a8915a653f4b145f2a58eb6996f28a | home/openbox/lib/helpers.py | home/openbox/lib/helpers.py | import os.path, os
def run_as_browser(fn):
def wrapped(*args, **kwargs):
rv = fn(*args, **kwargs)
return 'sudo -Hiu browser %s' % rv
return wrapped
class Helpers:
@property
@run_as_browser
def default_firefox_bin(self):
candidates = [
'/usr/local/lib/firefox/firefox-bin',
'/usr/local/lib/firefox3/firefox-bin',
'/usr/bin/iceweasel',
]
return self._pick(candidates, os.path.exists)
@property
@run_as_browser
def default_firefox_wrapper(self):
candidates = [
'firefox', 'firefox3'
]
return self._pick(candidates, self._wrapper_tester)
default_firefox = default_firefox_wrapper
@property
def as_browser(self):
return 'sudo -Hiu browser'
@property
def opera(self):
return 'sudo -Hiu browser opera'
@property
def chrome(self):
return 'sudo -Hiu browser chrome'
def have_bin(self, basename):
return self._wrapper_tester(basename)
def _wrapper_tester(self, candidate):
dirs = os.environ['PATH'].split(':')
for dir in dirs:
path = os.path.join(dir, candidate)
if os.path.exists(path):
return True
return False
def _pick(self, candidates, tester):
for candidate in candidates:
if tester(candidate):
return candidate
# consider raising here
return None
| import os.path, os
def run_as_browser(fn):
def wrapped(*args, **kwargs):
rv = fn(*args, **kwargs)
return 'sudo -Hiu browser env XAUTHORITY=/home/browser/.Xauthority %s' % rv
return wrapped
class Helpers:
@property
@run_as_browser
def default_firefox_bin(self):
candidates = [
'/usr/local/lib/firefox/firefox-bin',
'/usr/local/lib/firefox3/firefox-bin',
'/usr/bin/iceweasel',
]
return self._pick(candidates, os.path.exists)
@property
@run_as_browser
def default_firefox_wrapper(self):
candidates = [
'firefox', 'firefox3'
]
return self._pick(candidates, self._wrapper_tester)
default_firefox = default_firefox_wrapper
@property
def as_browser(self):
return 'sudo -Hiu browser'
@property
def opera(self):
return 'sudo -Hiu browser opera'
@property
def chrome(self):
return 'sudo -Hiu browser chrome'
def have_bin(self, basename):
return self._wrapper_tester(basename)
def _wrapper_tester(self, candidate):
dirs = os.environ['PATH'].split(':')
for dir in dirs:
path = os.path.join(dir, candidate)
if os.path.exists(path):
return True
return False
def _pick(self, candidates, tester):
for candidate in candidates:
if tester(candidate):
return candidate
# consider raising here
return None
| Fix firefox invocation as browser | Fix firefox invocation as browser
| Python | bsd-2-clause | p/pubfiles,p/pubfiles,p/pubfiles,p/pubfiles,p/pubfiles | import os.path, os
def run_as_browser(fn):
def wrapped(*args, **kwargs):
rv = fn(*args, **kwargs)
- return 'sudo -Hiu browser %s' % rv
+ return 'sudo -Hiu browser env XAUTHORITY=/home/browser/.Xauthority %s' % rv
return wrapped
class Helpers:
@property
@run_as_browser
def default_firefox_bin(self):
candidates = [
'/usr/local/lib/firefox/firefox-bin',
'/usr/local/lib/firefox3/firefox-bin',
'/usr/bin/iceweasel',
]
return self._pick(candidates, os.path.exists)
@property
@run_as_browser
def default_firefox_wrapper(self):
candidates = [
'firefox', 'firefox3'
]
return self._pick(candidates, self._wrapper_tester)
default_firefox = default_firefox_wrapper
@property
def as_browser(self):
return 'sudo -Hiu browser'
@property
def opera(self):
return 'sudo -Hiu browser opera'
@property
def chrome(self):
return 'sudo -Hiu browser chrome'
def have_bin(self, basename):
return self._wrapper_tester(basename)
def _wrapper_tester(self, candidate):
dirs = os.environ['PATH'].split(':')
for dir in dirs:
path = os.path.join(dir, candidate)
if os.path.exists(path):
return True
return False
def _pick(self, candidates, tester):
for candidate in candidates:
if tester(candidate):
return candidate
# consider raising here
return None
| Fix firefox invocation as browser | ## Code Before:
import os.path, os
def run_as_browser(fn):
def wrapped(*args, **kwargs):
rv = fn(*args, **kwargs)
return 'sudo -Hiu browser %s' % rv
return wrapped
class Helpers:
@property
@run_as_browser
def default_firefox_bin(self):
candidates = [
'/usr/local/lib/firefox/firefox-bin',
'/usr/local/lib/firefox3/firefox-bin',
'/usr/bin/iceweasel',
]
return self._pick(candidates, os.path.exists)
@property
@run_as_browser
def default_firefox_wrapper(self):
candidates = [
'firefox', 'firefox3'
]
return self._pick(candidates, self._wrapper_tester)
default_firefox = default_firefox_wrapper
@property
def as_browser(self):
return 'sudo -Hiu browser'
@property
def opera(self):
return 'sudo -Hiu browser opera'
@property
def chrome(self):
return 'sudo -Hiu browser chrome'
def have_bin(self, basename):
return self._wrapper_tester(basename)
def _wrapper_tester(self, candidate):
dirs = os.environ['PATH'].split(':')
for dir in dirs:
path = os.path.join(dir, candidate)
if os.path.exists(path):
return True
return False
def _pick(self, candidates, tester):
for candidate in candidates:
if tester(candidate):
return candidate
# consider raising here
return None
## Instruction:
Fix firefox invocation as browser
## Code After:
import os.path, os
def run_as_browser(fn):
def wrapped(*args, **kwargs):
rv = fn(*args, **kwargs)
return 'sudo -Hiu browser env XAUTHORITY=/home/browser/.Xauthority %s' % rv
return wrapped
class Helpers:
@property
@run_as_browser
def default_firefox_bin(self):
candidates = [
'/usr/local/lib/firefox/firefox-bin',
'/usr/local/lib/firefox3/firefox-bin',
'/usr/bin/iceweasel',
]
return self._pick(candidates, os.path.exists)
@property
@run_as_browser
def default_firefox_wrapper(self):
candidates = [
'firefox', 'firefox3'
]
return self._pick(candidates, self._wrapper_tester)
default_firefox = default_firefox_wrapper
@property
def as_browser(self):
return 'sudo -Hiu browser'
@property
def opera(self):
return 'sudo -Hiu browser opera'
@property
def chrome(self):
return 'sudo -Hiu browser chrome'
def have_bin(self, basename):
return self._wrapper_tester(basename)
def _wrapper_tester(self, candidate):
dirs = os.environ['PATH'].split(':')
for dir in dirs:
path = os.path.join(dir, candidate)
if os.path.exists(path):
return True
return False
def _pick(self, candidates, tester):
for candidate in candidates:
if tester(candidate):
return candidate
# consider raising here
return None
| # ... existing code ...
rv = fn(*args, **kwargs)
return 'sudo -Hiu browser env XAUTHORITY=/home/browser/.Xauthority %s' % rv
return wrapped
# ... rest of the code ... |
fc7ba9019b42f056713b81bfee70f9e780b4aab5 | models/rasmachine/twitter_client.py | models/rasmachine/twitter_client.py | from __future__ import absolute_import, print_function, unicode_literals
from builtins import dict, str
import tweepy
def get_oauth(auth_file):
try:
fh = open(auth_file, 'rt')
except IOError:
print('Could not get Twitter credentials.')
return None
lines = [l.strip() for l in fh.readlines()]
oauth = tweepy.OAuthHandler(lines[0], lines[1])
oauth.set_access_token(lines[2], lines[3])
fh.close()
return oauth
def update_status(msg, auth_file='twitter_cred.txt'):
twitter_auth = get_oauth(auth_file)
if twitter_auth is None:
return
twitter_api = tweepy.API(twitter_auth)
twitter_api.update_status(msg)
| from __future__ import absolute_import, print_function, unicode_literals
from builtins import dict, str
import tweepy
def get_oauth_file(auth_file):
try:
fh = open(auth_file, 'rt')
except IOError:
print('Could not get Twitter credentials.')
return None
lines = [l.strip() for l in fh.readlines()]
oauth = tweepy.OAuthHandler(lines[0], lines[1])
oauth.set_access_token(lines[2], lines[3])
fh.close()
return oauth
def get_oauth_dict(auth_dict):
oauth = tweepy.OAuthHandler(auth_dict.get('consumer_token'),
auth_dict.get('consumer_secred'))
oauth.set_access_token(auth_dict.get('access_token'),
auth_dict.get('access_secret'))
return oauth
def update_status(msg, twitter_cred):
twitter_auth = get_oauth_dict(twitter_cred)
if twitter_auth is None:
return
twitter_api = tweepy.API(twitter_auth)
twitter_api.update_status(msg)
| Implement dict credentials in Twitter client | Implement dict credentials in Twitter client
| Python | bsd-2-clause | pvtodorov/indra,johnbachman/belpy,jmuhlich/indra,jmuhlich/indra,bgyori/indra,johnbachman/belpy,pvtodorov/indra,sorgerlab/belpy,johnbachman/belpy,pvtodorov/indra,johnbachman/indra,bgyori/indra,pvtodorov/indra,johnbachman/indra,sorgerlab/belpy,johnbachman/indra,sorgerlab/indra,bgyori/indra,sorgerlab/belpy,sorgerlab/indra,jmuhlich/indra,sorgerlab/indra | from __future__ import absolute_import, print_function, unicode_literals
from builtins import dict, str
import tweepy
- def get_oauth(auth_file):
+ def get_oauth_file(auth_file):
try:
fh = open(auth_file, 'rt')
except IOError:
print('Could not get Twitter credentials.')
return None
lines = [l.strip() for l in fh.readlines()]
oauth = tweepy.OAuthHandler(lines[0], lines[1])
oauth.set_access_token(lines[2], lines[3])
fh.close()
return oauth
+ def get_oauth_dict(auth_dict):
+ oauth = tweepy.OAuthHandler(auth_dict.get('consumer_token'),
+ auth_dict.get('consumer_secred'))
+ oauth.set_access_token(auth_dict.get('access_token'),
+ auth_dict.get('access_secret'))
+ return oauth
+
- def update_status(msg, auth_file='twitter_cred.txt'):
+ def update_status(msg, twitter_cred):
- twitter_auth = get_oauth(auth_file)
+ twitter_auth = get_oauth_dict(twitter_cred)
if twitter_auth is None:
return
twitter_api = tweepy.API(twitter_auth)
twitter_api.update_status(msg)
| Implement dict credentials in Twitter client | ## Code Before:
from __future__ import absolute_import, print_function, unicode_literals
from builtins import dict, str
import tweepy
def get_oauth(auth_file):
try:
fh = open(auth_file, 'rt')
except IOError:
print('Could not get Twitter credentials.')
return None
lines = [l.strip() for l in fh.readlines()]
oauth = tweepy.OAuthHandler(lines[0], lines[1])
oauth.set_access_token(lines[2], lines[3])
fh.close()
return oauth
def update_status(msg, auth_file='twitter_cred.txt'):
twitter_auth = get_oauth(auth_file)
if twitter_auth is None:
return
twitter_api = tweepy.API(twitter_auth)
twitter_api.update_status(msg)
## Instruction:
Implement dict credentials in Twitter client
## Code After:
from __future__ import absolute_import, print_function, unicode_literals
from builtins import dict, str
import tweepy
def get_oauth_file(auth_file):
try:
fh = open(auth_file, 'rt')
except IOError:
print('Could not get Twitter credentials.')
return None
lines = [l.strip() for l in fh.readlines()]
oauth = tweepy.OAuthHandler(lines[0], lines[1])
oauth.set_access_token(lines[2], lines[3])
fh.close()
return oauth
def get_oauth_dict(auth_dict):
oauth = tweepy.OAuthHandler(auth_dict.get('consumer_token'),
auth_dict.get('consumer_secred'))
oauth.set_access_token(auth_dict.get('access_token'),
auth_dict.get('access_secret'))
return oauth
def update_status(msg, twitter_cred):
twitter_auth = get_oauth_dict(twitter_cred)
if twitter_auth is None:
return
twitter_api = tweepy.API(twitter_auth)
twitter_api.update_status(msg)
| ...
def get_oauth_file(auth_file):
try:
...
def get_oauth_dict(auth_dict):
oauth = tweepy.OAuthHandler(auth_dict.get('consumer_token'),
auth_dict.get('consumer_secred'))
oauth.set_access_token(auth_dict.get('access_token'),
auth_dict.get('access_secret'))
return oauth
def update_status(msg, twitter_cred):
twitter_auth = get_oauth_dict(twitter_cred)
if twitter_auth is None:
... |
6a3fbb7280c1078b574736eae3c6a3e4e42d3f46 | seaborn/__init__.py | seaborn/__init__.py | import matplotlib as mpl
_orig_rc_params = mpl.rcParams.copy()
# Import seaborn objects
from .rcmod import *
from .utils import *
from .palettes import *
from .relational import *
from .regression import *
from .categorical import *
from .distributions import *
from .timeseries import *
from .matrix import *
from .miscplot import *
from .axisgrid import *
from .widgets import *
from .colors import xkcd_rgb, crayons
from . import cm
__version__ = "0.9.1.dev0"
| import matplotlib as mpl
_orig_rc_params = mpl.rcParams.copy()
# Import seaborn objects
from .rcmod import *
from .utils import *
from .palettes import *
from .relational import *
from .regression import *
from .categorical import *
from .distributions import *
from .matrix import *
from .miscplot import *
from .axisgrid import *
from .widgets import *
from .colors import xkcd_rgb, crayons
from . import cm
__version__ = "0.9.1.dev0"
| Remove top-level import of timeseries module | Remove top-level import of timeseries module
| Python | bsd-3-clause | arokem/seaborn,mwaskom/seaborn,mwaskom/seaborn,arokem/seaborn,anntzer/seaborn,anntzer/seaborn | import matplotlib as mpl
_orig_rc_params = mpl.rcParams.copy()
# Import seaborn objects
from .rcmod import *
from .utils import *
from .palettes import *
from .relational import *
from .regression import *
from .categorical import *
from .distributions import *
- from .timeseries import *
from .matrix import *
from .miscplot import *
from .axisgrid import *
from .widgets import *
from .colors import xkcd_rgb, crayons
from . import cm
__version__ = "0.9.1.dev0"
| Remove top-level import of timeseries module | ## Code Before:
import matplotlib as mpl
_orig_rc_params = mpl.rcParams.copy()
# Import seaborn objects
from .rcmod import *
from .utils import *
from .palettes import *
from .relational import *
from .regression import *
from .categorical import *
from .distributions import *
from .timeseries import *
from .matrix import *
from .miscplot import *
from .axisgrid import *
from .widgets import *
from .colors import xkcd_rgb, crayons
from . import cm
__version__ = "0.9.1.dev0"
## Instruction:
Remove top-level import of timeseries module
## Code After:
import matplotlib as mpl
_orig_rc_params = mpl.rcParams.copy()
# Import seaborn objects
from .rcmod import *
from .utils import *
from .palettes import *
from .relational import *
from .regression import *
from .categorical import *
from .distributions import *
from .matrix import *
from .miscplot import *
from .axisgrid import *
from .widgets import *
from .colors import xkcd_rgb, crayons
from . import cm
__version__ = "0.9.1.dev0"
| ...
from .distributions import *
from .matrix import *
... |
592c6550255793772add694cb941a0db0883713b | kamboo/core.py | kamboo/core.py |
import logging
import botocore
from kotocore.session import Session
log = logging.getLogger(__name__)
class KambooConnection(object):
"""
Kamboo connection with botocore session initialized
"""
session = botocore.session.get_session()
def __init__(self, service_name="ec2", region_name="us-east-1",
account_id=None,
credentials=None):
self.region = region_name
self.account_id = account_id
self.credentials = credentials
if self.credentials:
self.session.set_credentials(**self.credentials)
Connection = Session(session=self.session).get_connection(service_name)
self.conn = Connection(region_name=self.region)
|
import logging
import botocore
from kotocore.session import Session
log = logging.getLogger(__name__)
class KambooConnection(object):
"""
Kamboo connection with botocore session initialized
"""
def __init__(self, service_name="ec2", region_name="us-east-1",
account_id=None,
credentials=None):
self.session = botocore.session.get_session()
self.service = service_name
self.region = region_name
self.account_id = account_id
self.credentials = credentials
if self.credentials:
self.session.set_credentials(**self.credentials)
Connection = Session(session=self.session).get_connection(service_name)
self.conn = Connection(region_name=self.region)
def __repr__(self):
return "KambooConnection: [%s, %s, %s]" % (self.account_id,
self.region,
self.service)
| Fix the issue: "session" shared in different connections | Fix the issue: "session" shared in different connections
| Python | apache-2.0 | henrysher/kamboo,henrysher/kamboo |
import logging
import botocore
from kotocore.session import Session
log = logging.getLogger(__name__)
class KambooConnection(object):
"""
Kamboo connection with botocore session initialized
"""
- session = botocore.session.get_session()
def __init__(self, service_name="ec2", region_name="us-east-1",
account_id=None,
credentials=None):
+ self.session = botocore.session.get_session()
+ self.service = service_name
self.region = region_name
self.account_id = account_id
self.credentials = credentials
if self.credentials:
self.session.set_credentials(**self.credentials)
Connection = Session(session=self.session).get_connection(service_name)
self.conn = Connection(region_name=self.region)
+ def __repr__(self):
+ return "KambooConnection: [%s, %s, %s]" % (self.account_id,
+ self.region,
+ self.service)
+ | Fix the issue: "session" shared in different connections | ## Code Before:
import logging
import botocore
from kotocore.session import Session
log = logging.getLogger(__name__)
class KambooConnection(object):
"""
Kamboo connection with botocore session initialized
"""
session = botocore.session.get_session()
def __init__(self, service_name="ec2", region_name="us-east-1",
account_id=None,
credentials=None):
self.region = region_name
self.account_id = account_id
self.credentials = credentials
if self.credentials:
self.session.set_credentials(**self.credentials)
Connection = Session(session=self.session).get_connection(service_name)
self.conn = Connection(region_name=self.region)
## Instruction:
Fix the issue: "session" shared in different connections
## Code After:
import logging
import botocore
from kotocore.session import Session
log = logging.getLogger(__name__)
class KambooConnection(object):
"""
Kamboo connection with botocore session initialized
"""
def __init__(self, service_name="ec2", region_name="us-east-1",
account_id=None,
credentials=None):
self.session = botocore.session.get_session()
self.service = service_name
self.region = region_name
self.account_id = account_id
self.credentials = credentials
if self.credentials:
self.session.set_credentials(**self.credentials)
Connection = Session(session=self.session).get_connection(service_name)
self.conn = Connection(region_name=self.region)
def __repr__(self):
return "KambooConnection: [%s, %s, %s]" % (self.account_id,
self.region,
self.service)
| # ... existing code ...
"""
# ... modified code ...
credentials=None):
self.session = botocore.session.get_session()
self.service = service_name
self.region = region_name
...
self.conn = Connection(region_name=self.region)
def __repr__(self):
return "KambooConnection: [%s, %s, %s]" % (self.account_id,
self.region,
self.service)
# ... rest of the code ... |
3e40f0bd3941a48e3b792573c0058f1cc339d5db | fabfile.py | fabfile.py |
from fabric.api import *
import slackbot_settings as settings
from urllib import request, parse
env.hosts = settings.DEPLOY_HOSTS
def deploy():
slack("Deploy Started")
with cd("/var/bot/slack-shogi"):
run("git pull")
run("supervisorctl reload")
slack("Deploy Finished")
def slack(text):
if settings.WEBHOOK_URL:
payload = ("payload={\"text\": \"" + parse.quote(text) +
"\", \"username\": \"Mr.deploy\"}").encode("utf-8")
request.urlopen(url=settings.WEBHOOK_URL, data=payload)
|
from fabric.api import *
import slackbot_settings as settings
from urllib import request, parse
env.hosts = settings.DEPLOY_HOSTS
def deploy():
slack("Deploy Started")
try:
with cd("/var/bot/slack-shogi"):
run("git pull")
run("supervisorctl reload")
slack("Deploy Finished")
except:
slack("Deploy Failed")
def slack(text):
if settings.WEBHOOK_URL:
payload = ("payload={\"text\": \"" + parse.quote(text) +
"\", \"username\": \"Mr.deploy\"}").encode("utf-8")
request.urlopen(url=settings.WEBHOOK_URL, data=payload)
| Add fallback slack message for deploy script | Add fallback slack message for deploy script
| Python | mit | setokinto/slack-shogi |
from fabric.api import *
import slackbot_settings as settings
from urllib import request, parse
env.hosts = settings.DEPLOY_HOSTS
def deploy():
slack("Deploy Started")
+ try:
- with cd("/var/bot/slack-shogi"):
+ with cd("/var/bot/slack-shogi"):
- run("git pull")
+ run("git pull")
- run("supervisorctl reload")
+ run("supervisorctl reload")
- slack("Deploy Finished")
+ slack("Deploy Finished")
+ except:
+ slack("Deploy Failed")
def slack(text):
if settings.WEBHOOK_URL:
payload = ("payload={\"text\": \"" + parse.quote(text) +
"\", \"username\": \"Mr.deploy\"}").encode("utf-8")
request.urlopen(url=settings.WEBHOOK_URL, data=payload)
+ | Add fallback slack message for deploy script | ## Code Before:
from fabric.api import *
import slackbot_settings as settings
from urllib import request, parse
env.hosts = settings.DEPLOY_HOSTS
def deploy():
slack("Deploy Started")
with cd("/var/bot/slack-shogi"):
run("git pull")
run("supervisorctl reload")
slack("Deploy Finished")
def slack(text):
if settings.WEBHOOK_URL:
payload = ("payload={\"text\": \"" + parse.quote(text) +
"\", \"username\": \"Mr.deploy\"}").encode("utf-8")
request.urlopen(url=settings.WEBHOOK_URL, data=payload)
## Instruction:
Add fallback slack message for deploy script
## Code After:
from fabric.api import *
import slackbot_settings as settings
from urllib import request, parse
env.hosts = settings.DEPLOY_HOSTS
def deploy():
slack("Deploy Started")
try:
with cd("/var/bot/slack-shogi"):
run("git pull")
run("supervisorctl reload")
slack("Deploy Finished")
except:
slack("Deploy Failed")
def slack(text):
if settings.WEBHOOK_URL:
payload = ("payload={\"text\": \"" + parse.quote(text) +
"\", \"username\": \"Mr.deploy\"}").encode("utf-8")
request.urlopen(url=settings.WEBHOOK_URL, data=payload)
| # ... existing code ...
slack("Deploy Started")
try:
with cd("/var/bot/slack-shogi"):
run("git pull")
run("supervisorctl reload")
slack("Deploy Finished")
except:
slack("Deploy Failed")
# ... modified code ...
request.urlopen(url=settings.WEBHOOK_URL, data=payload)
# ... rest of the code ... |
fa3ec9a764ca0d646588e908395367ce553981e1 | tca/chat/views.py | tca/chat/views.py | from django.shortcuts import render
from rest_framework import viewsets
from chat.models import Member
from chat.models import ChatRoom
from chat.serializers import MemberSerializer
from chat.serializers import ChatRoomSerializer
class MemberViewSet(viewsets.ModelViewSet):
model = Member
serializer_class = MemberSerializer
class ChatRoomViewSet(viewsets.ModelViewSet):
model = ChatRoom
serializer_class = ChatRoomSerializer
| from django.shortcuts import render
from django.shortcuts import get_object_or_404
from rest_framework import viewsets
from rest_framework import status
from rest_framework.decorators import action
from rest_framework.response import Response
from chat.models import Member
from chat.models import ChatRoom
from chat.serializers import MemberSerializer
from chat.serializers import ChatRoomSerializer
class MemberViewSet(viewsets.ModelViewSet):
model = Member
serializer_class = MemberSerializer
class ChatRoomViewSet(viewsets.ModelViewSet):
model = ChatRoom
serializer_class = ChatRoomSerializer
@action()
def add_member(self, request, pk=None):
chat_room = self.get_object()
if 'lrz_id' not in request.DATA:
# Invalid request
return Response(status=status.HTTP_400_BAD_REQUEST)
member = get_object_or_404(Member, lrz_id=request.DATA['lrz_id'])
chat_room.members.add(member)
return Response({
'status': 'success',
})
| Add an action for adding members to a chat room | Add an action for adding members to a chat room
Even though django-rest-framework supports a Ruby-on-Rails style
of updating existing resources by issuing a PATCH or PUT request,
such updates are unsafe and can cause race-conditions to lose
some state. The implementation of this action isn't fully RESTful,
but neither is the PATCH-style used here, so it is still a good
choice for solving a possible race condition situation.
| Python | bsd-3-clause | mlalic/TumCampusAppBackend,mlalic/TumCampusAppBackend | from django.shortcuts import render
+ from django.shortcuts import get_object_or_404
from rest_framework import viewsets
+ from rest_framework import status
+ from rest_framework.decorators import action
+ from rest_framework.response import Response
from chat.models import Member
from chat.models import ChatRoom
from chat.serializers import MemberSerializer
from chat.serializers import ChatRoomSerializer
class MemberViewSet(viewsets.ModelViewSet):
model = Member
serializer_class = MemberSerializer
class ChatRoomViewSet(viewsets.ModelViewSet):
model = ChatRoom
serializer_class = ChatRoomSerializer
+ @action()
+ def add_member(self, request, pk=None):
+ chat_room = self.get_object()
+ if 'lrz_id' not in request.DATA:
+ # Invalid request
+ return Response(status=status.HTTP_400_BAD_REQUEST)
+
+ member = get_object_or_404(Member, lrz_id=request.DATA['lrz_id'])
+ chat_room.members.add(member)
+
+ return Response({
+ 'status': 'success',
+ })
+ | Add an action for adding members to a chat room | ## Code Before:
from django.shortcuts import render
from rest_framework import viewsets
from chat.models import Member
from chat.models import ChatRoom
from chat.serializers import MemberSerializer
from chat.serializers import ChatRoomSerializer
class MemberViewSet(viewsets.ModelViewSet):
model = Member
serializer_class = MemberSerializer
class ChatRoomViewSet(viewsets.ModelViewSet):
model = ChatRoom
serializer_class = ChatRoomSerializer
## Instruction:
Add an action for adding members to a chat room
## Code After:
from django.shortcuts import render
from django.shortcuts import get_object_or_404
from rest_framework import viewsets
from rest_framework import status
from rest_framework.decorators import action
from rest_framework.response import Response
from chat.models import Member
from chat.models import ChatRoom
from chat.serializers import MemberSerializer
from chat.serializers import ChatRoomSerializer
class MemberViewSet(viewsets.ModelViewSet):
model = Member
serializer_class = MemberSerializer
class ChatRoomViewSet(viewsets.ModelViewSet):
model = ChatRoom
serializer_class = ChatRoomSerializer
@action()
def add_member(self, request, pk=None):
chat_room = self.get_object()
if 'lrz_id' not in request.DATA:
# Invalid request
return Response(status=status.HTTP_400_BAD_REQUEST)
member = get_object_or_404(Member, lrz_id=request.DATA['lrz_id'])
chat_room.members.add(member)
return Response({
'status': 'success',
})
| // ... existing code ...
from django.shortcuts import render
from django.shortcuts import get_object_or_404
// ... modified code ...
from rest_framework import viewsets
from rest_framework import status
from rest_framework.decorators import action
from rest_framework.response import Response
...
serializer_class = ChatRoomSerializer
@action()
def add_member(self, request, pk=None):
chat_room = self.get_object()
if 'lrz_id' not in request.DATA:
# Invalid request
return Response(status=status.HTTP_400_BAD_REQUEST)
member = get_object_or_404(Member, lrz_id=request.DATA['lrz_id'])
chat_room.members.add(member)
return Response({
'status': 'success',
})
// ... rest of the code ... |
ff85fc05e179e451dabb1f20781dfc5a90314d71 | scripts/adb-wrapper.py | scripts/adb-wrapper.py | import subprocess
import sys
import re
# Note: no output will be printed until the entire test suite has finished
result = subprocess.run(sys.argv[1], shell=True, stdout=subprocess.PIPE, stderr=subprocess.PIPE, universal_newlines=True)
successRegex = re.compile('OK \(\d+ tests\)')
print(result.stderr)
print(result.stdout)
if successRegex.search(result.stderr + result.stdout):
sys.exit(0)
else:
sys.exit(1)
| import subprocess
import sys
import re
# Note: no output will be printed until the entire test suite has finished
p = subprocess.Popen(sys.argv[1], shell=True, stdout=subprocess.PIPE, stderr=subprocess.PIPE, universal_newlines=True)
stdoutResult, stderrResult = p.communicate()
successRegex = re.compile('OK \(\d+ tests\)')
print(stdoutResult)
print(stderrResult)
if successRegex.search(stderrResult + stdoutResult):
sys.exit(0)
else:
sys.exit(1)
| Refactor the python wrapper script because apparently apt-get doesn't install 3.5, and subprocess.run() is only in 3.5 | Refactor the python wrapper script because apparently apt-get doesn't install 3.5, and subprocess.run() is only in 3.5
| Python | apache-2.0 | sbosley/squidb,yahoo/squidb,yahoo/squidb,sbosley/squidb,sbosley/squidb,sbosley/squidb,sbosley/squidb,yahoo/squidb,yahoo/squidb,yahoo/squidb | import subprocess
import sys
import re
# Note: no output will be printed until the entire test suite has finished
- result = subprocess.run(sys.argv[1], shell=True, stdout=subprocess.PIPE, stderr=subprocess.PIPE, universal_newlines=True)
+ p = subprocess.Popen(sys.argv[1], shell=True, stdout=subprocess.PIPE, stderr=subprocess.PIPE, universal_newlines=True)
+ stdoutResult, stderrResult = p.communicate()
successRegex = re.compile('OK \(\d+ tests\)')
- print(result.stderr)
- print(result.stdout)
+ print(stdoutResult)
+ print(stderrResult)
- if successRegex.search(result.stderr + result.stdout):
+ if successRegex.search(stderrResult + stdoutResult):
sys.exit(0)
else:
sys.exit(1)
| Refactor the python wrapper script because apparently apt-get doesn't install 3.5, and subprocess.run() is only in 3.5 | ## Code Before:
import subprocess
import sys
import re
# Note: no output will be printed until the entire test suite has finished
result = subprocess.run(sys.argv[1], shell=True, stdout=subprocess.PIPE, stderr=subprocess.PIPE, universal_newlines=True)
successRegex = re.compile('OK \(\d+ tests\)')
print(result.stderr)
print(result.stdout)
if successRegex.search(result.stderr + result.stdout):
sys.exit(0)
else:
sys.exit(1)
## Instruction:
Refactor the python wrapper script because apparently apt-get doesn't install 3.5, and subprocess.run() is only in 3.5
## Code After:
import subprocess
import sys
import re
# Note: no output will be printed until the entire test suite has finished
p = subprocess.Popen(sys.argv[1], shell=True, stdout=subprocess.PIPE, stderr=subprocess.PIPE, universal_newlines=True)
stdoutResult, stderrResult = p.communicate()
successRegex = re.compile('OK \(\d+ tests\)')
print(stdoutResult)
print(stderrResult)
if successRegex.search(stderrResult + stdoutResult):
sys.exit(0)
else:
sys.exit(1)
| ...
# Note: no output will be printed until the entire test suite has finished
p = subprocess.Popen(sys.argv[1], shell=True, stdout=subprocess.PIPE, stderr=subprocess.PIPE, universal_newlines=True)
stdoutResult, stderrResult = p.communicate()
...
print(stdoutResult)
print(stderrResult)
if successRegex.search(stderrResult + stdoutResult):
sys.exit(0)
... |
c3a6554ce24781a3eaa9229f9609a09e4e018069 | lcd_restful/__main__.py | lcd_restful/__main__.py | import sys
from getopt import getopt, GetoptError
from .api import Server
from .fake import FakeLcdApi
USAGE = """\
Usage %s [-h|--help]
\t-h or --help\tThis help message
"""
def get_args(args):
try:
opts, args = getopt(args[1:], 'hf', ['help', 'fake'])
except GetoptError as e:
print('GetoptError %s' % e)
sys.exit(2)
ret_args = {}
ret_args['fake'] = False
for opt, arg in opts:
if opt in ['-h', '--help']:
print(USAGE % args[0])
sys.exit(0)
elif opt in ['-f', '--fake']:
ret_args['fake'] = True
else:
print(USAGE % args[0])
sys.exit(1)
return ret_args
def main_serv(clargs=sys.argv):
opts = get_args(clargs)
lcd = None
if opts['fake']:
lcd = FakeLcdApi()
s = Server(lcd=lcd)
s.run()
return 0
if __name__ == "__main__":
sys.exit(main_serv())
| import sys
from getopt import getopt, GetoptError
from .api import Server
from .lcd import Lcd
USAGE = """\
Usage %s [-h|--help] [-f|--fake]
\t-h or --help\tThis help message
\t-f or --fake\tIf on RPi, use FakeHw
"""
def get_args(args):
arg0 = args[0]
try:
opts, args = getopt(args[1:], 'hf', ['help', 'fake'])
except GetoptError as e:
print('GetoptError %s' % e)
sys.exit(2)
ret_args = {}
ret_args['fake'] = False
for opt, arg in opts:
if opt in ['-h', '--help']:
print(USAGE % arg0)
sys.exit(0)
elif opt in ['-f', '--fake']:
ret_args['fake'] = True
else:
print(USAGE % arg0)
sys.exit(1)
return ret_args
def main_serv(clargs=sys.argv):
opts = get_args(clargs)
s = Server(lcd=Lcd(opts['fake']))
s.run()
return 0
if __name__ == "__main__":
sys.exit(main_serv())
| Update main Lcd import and init, and fix help msg | Update main Lcd import and init, and fix help msg
| Python | mit | rfarley3/lcd-restful,rfarley3/lcd-restful | import sys
from getopt import getopt, GetoptError
from .api import Server
- from .fake import FakeLcdApi
+ from .lcd import Lcd
USAGE = """\
- Usage %s [-h|--help]
+ Usage %s [-h|--help] [-f|--fake]
\t-h or --help\tThis help message
+ \t-f or --fake\tIf on RPi, use FakeHw
"""
def get_args(args):
+ arg0 = args[0]
try:
opts, args = getopt(args[1:], 'hf', ['help', 'fake'])
except GetoptError as e:
print('GetoptError %s' % e)
sys.exit(2)
ret_args = {}
ret_args['fake'] = False
for opt, arg in opts:
if opt in ['-h', '--help']:
- print(USAGE % args[0])
+ print(USAGE % arg0)
sys.exit(0)
elif opt in ['-f', '--fake']:
ret_args['fake'] = True
else:
- print(USAGE % args[0])
+ print(USAGE % arg0)
sys.exit(1)
return ret_args
def main_serv(clargs=sys.argv):
opts = get_args(clargs)
+ s = Server(lcd=Lcd(opts['fake']))
- lcd = None
- if opts['fake']:
- lcd = FakeLcdApi()
- s = Server(lcd=lcd)
s.run()
return 0
if __name__ == "__main__":
sys.exit(main_serv())
+ | Update main Lcd import and init, and fix help msg | ## Code Before:
import sys
from getopt import getopt, GetoptError
from .api import Server
from .fake import FakeLcdApi
USAGE = """\
Usage %s [-h|--help]
\t-h or --help\tThis help message
"""
def get_args(args):
try:
opts, args = getopt(args[1:], 'hf', ['help', 'fake'])
except GetoptError as e:
print('GetoptError %s' % e)
sys.exit(2)
ret_args = {}
ret_args['fake'] = False
for opt, arg in opts:
if opt in ['-h', '--help']:
print(USAGE % args[0])
sys.exit(0)
elif opt in ['-f', '--fake']:
ret_args['fake'] = True
else:
print(USAGE % args[0])
sys.exit(1)
return ret_args
def main_serv(clargs=sys.argv):
opts = get_args(clargs)
lcd = None
if opts['fake']:
lcd = FakeLcdApi()
s = Server(lcd=lcd)
s.run()
return 0
if __name__ == "__main__":
sys.exit(main_serv())
## Instruction:
Update main Lcd import and init, and fix help msg
## Code After:
import sys
from getopt import getopt, GetoptError
from .api import Server
from .lcd import Lcd
USAGE = """\
Usage %s [-h|--help] [-f|--fake]
\t-h or --help\tThis help message
\t-f or --fake\tIf on RPi, use FakeHw
"""
def get_args(args):
arg0 = args[0]
try:
opts, args = getopt(args[1:], 'hf', ['help', 'fake'])
except GetoptError as e:
print('GetoptError %s' % e)
sys.exit(2)
ret_args = {}
ret_args['fake'] = False
for opt, arg in opts:
if opt in ['-h', '--help']:
print(USAGE % arg0)
sys.exit(0)
elif opt in ['-f', '--fake']:
ret_args['fake'] = True
else:
print(USAGE % arg0)
sys.exit(1)
return ret_args
def main_serv(clargs=sys.argv):
opts = get_args(clargs)
s = Server(lcd=Lcd(opts['fake']))
s.run()
return 0
if __name__ == "__main__":
sys.exit(main_serv())
| // ... existing code ...
from .api import Server
from .lcd import Lcd
// ... modified code ...
USAGE = """\
Usage %s [-h|--help] [-f|--fake]
\t-h or --help\tThis help message
\t-f or --fake\tIf on RPi, use FakeHw
...
def get_args(args):
arg0 = args[0]
try:
...
if opt in ['-h', '--help']:
print(USAGE % arg0)
sys.exit(0)
...
else:
print(USAGE % arg0)
sys.exit(1)
...
opts = get_args(clargs)
s = Server(lcd=Lcd(opts['fake']))
s.run()
...
sys.exit(main_serv())
// ... rest of the code ... |
bc5fa08e84cd11349dc44c3065b7b5380d60ebd9 | raven/contrib/django/handlers.py | raven/contrib/django/handlers.py |
from __future__ import absolute_import
import logging
from raven.handlers.logging import SentryHandler as BaseSentryHandler
class SentryHandler(BaseSentryHandler):
def __init__(self):
logging.Handler.__init__(self)
def _get_client(self):
from raven.contrib.django.models import client
return client
client = property(_get_client)
def _emit(self, record):
request = getattr(record, 'request', None)
return super(SentryHandler, self)._emit(record, request=request)
|
from __future__ import absolute_import
import logging
from raven.handlers.logging import SentryHandler as BaseSentryHandler
class SentryHandler(BaseSentryHandler):
def __init__(self, level=logging.NOTSET):
logging.Handler.__init__(self, level=level)
def _get_client(self):
from raven.contrib.django.models import client
return client
client = property(_get_client)
def _emit(self, record):
request = getattr(record, 'request', None)
return super(SentryHandler, self)._emit(record, request=request)
| Allow level param in Django SentryHandler.__init__ | Allow level param in Django SentryHandler.__init__
For consistency with superclass and with logging.Handler
| Python | bsd-3-clause | arthurlogilab/raven-python,inspirehep/raven-python,ewdurbin/raven-python,jbarbuto/raven-python,getsentry/raven-python,dbravender/raven-python,jbarbuto/raven-python,Photonomie/raven-python,percipient/raven-python,akheron/raven-python,icereval/raven-python,nikolas/raven-python,akalipetis/raven-python,someonehan/raven-python,nikolas/raven-python,ewdurbin/raven-python,inspirehep/raven-python,icereval/raven-python,danriti/raven-python,danriti/raven-python,Goldmund-Wyldebeast-Wunderliebe/raven-python,Goldmund-Wyldebeast-Wunderliebe/raven-python,someonehan/raven-python,danriti/raven-python,Goldmund-Wyldebeast-Wunderliebe/raven-python,akheron/raven-python,akheron/raven-python,arthurlogilab/raven-python,hzy/raven-python,nikolas/raven-python,arthurlogilab/raven-python,jbarbuto/raven-python,inspirehep/raven-python,smarkets/raven-python,arthurlogilab/raven-python,smarkets/raven-python,johansteffner/raven-python,akalipetis/raven-python,lepture/raven-python,getsentry/raven-python,akalipetis/raven-python,ronaldevers/raven-python,johansteffner/raven-python,percipient/raven-python,icereval/raven-python,jmp0xf/raven-python,jmagnusson/raven-python,lepture/raven-python,smarkets/raven-python,johansteffner/raven-python,Photonomie/raven-python,someonehan/raven-python,ronaldevers/raven-python,dbravender/raven-python,ronaldevers/raven-python,lepture/raven-python,percipient/raven-python,jbarbuto/raven-python,jmagnusson/raven-python,jmp0xf/raven-python,smarkets/raven-python,recht/raven-python,jmp0xf/raven-python,icereval/raven-python,jmagnusson/raven-python,Photonomie/raven-python,Goldmund-Wyldebeast-Wunderliebe/raven-python,dbravender/raven-python,recht/raven-python,nikolas/raven-python,getsentry/raven-python,hzy/raven-python,inspirehep/raven-python,recht/raven-python,ewdurbin/raven-python,hzy/raven-python |
from __future__ import absolute_import
import logging
from raven.handlers.logging import SentryHandler as BaseSentryHandler
class SentryHandler(BaseSentryHandler):
- def __init__(self):
+ def __init__(self, level=logging.NOTSET):
- logging.Handler.__init__(self)
+ logging.Handler.__init__(self, level=level)
def _get_client(self):
from raven.contrib.django.models import client
return client
client = property(_get_client)
def _emit(self, record):
request = getattr(record, 'request', None)
return super(SentryHandler, self)._emit(record, request=request)
| Allow level param in Django SentryHandler.__init__ | ## Code Before:
from __future__ import absolute_import
import logging
from raven.handlers.logging import SentryHandler as BaseSentryHandler
class SentryHandler(BaseSentryHandler):
def __init__(self):
logging.Handler.__init__(self)
def _get_client(self):
from raven.contrib.django.models import client
return client
client = property(_get_client)
def _emit(self, record):
request = getattr(record, 'request', None)
return super(SentryHandler, self)._emit(record, request=request)
## Instruction:
Allow level param in Django SentryHandler.__init__
## Code After:
from __future__ import absolute_import
import logging
from raven.handlers.logging import SentryHandler as BaseSentryHandler
class SentryHandler(BaseSentryHandler):
def __init__(self, level=logging.NOTSET):
logging.Handler.__init__(self, level=level)
def _get_client(self):
from raven.contrib.django.models import client
return client
client = property(_get_client)
def _emit(self, record):
request = getattr(record, 'request', None)
return super(SentryHandler, self)._emit(record, request=request)
| // ... existing code ...
class SentryHandler(BaseSentryHandler):
def __init__(self, level=logging.NOTSET):
logging.Handler.__init__(self, level=level)
// ... rest of the code ... |
d32e1d8349c115027e3095d61f8fa882fca1ab52 | functions/test_lambda.py | functions/test_lambda.py | """ Explore how python works with lambda expressions. """
import unittest
import string
import random
class TestGetWebsites(unittest.TestCase):
def test_closure(self, m):
""" See that python supports closures similar to JavaScript """
def gibberish():
""" Some random string """
return ''.join([random.choice(string.ascii_letters) for i in range(20)])
value1 = gibberish()
value2 = gibberish()
result = (lambda x: lambda y: x)(value1)(value2)
self.assertEqual(result, value1)
if __name__ == "__main__":
unittest.main() | """ Explore how python works with lambda expressions. """
import unittest
import string
import random
class TestGetWebsites(unittest.TestCase):
def test_closure(self):
""" See that python supports closures similar to JavaScript """
def gibberish():
""" Some random string """
return ''.join([random.choice(string.ascii_letters) for i in range(20)])
value1 = gibberish()
result = (lambda x: lambda: x)(value1)()
self.assertEqual(result, value1)
def test_closure_ids(self):
""" Show how a variable passed to a function remains in scope
even after it returns """
def make_lambdas(num):
""" Build a lambda generator """
for i in range(num):
func = lambda x: lambda: x
yield func(i)
functions = list(make_lambdas(random.randint(1, 10)))
random_index = random.randint(0, len(functions)-1)
random_function = functions[random_index]()
print("{0} equals {1}".format(random_index, random_function))
self.assertEqual(random_function, random_index)
if __name__ == "__main__":
unittest.main() | Add a test to generate a range of functions with closures. Check if it returns the expected value | Add a test to generate a range of functions with closures. Check if it returns the expected value
| Python | mit | b-ritter/python-notes,b-ritter/python-notes | """ Explore how python works with lambda expressions. """
import unittest
import string
import random
class TestGetWebsites(unittest.TestCase):
- def test_closure(self, m):
+ def test_closure(self):
""" See that python supports closures similar to JavaScript """
def gibberish():
""" Some random string """
return ''.join([random.choice(string.ascii_letters) for i in range(20)])
value1 = gibberish()
- value2 = gibberish()
- result = (lambda x: lambda y: x)(value1)(value2)
+ result = (lambda x: lambda: x)(value1)()
self.assertEqual(result, value1)
+
+ def test_closure_ids(self):
+ """ Show how a variable passed to a function remains in scope
+ even after it returns """
+ def make_lambdas(num):
+ """ Build a lambda generator """
+ for i in range(num):
+ func = lambda x: lambda: x
+ yield func(i)
+ functions = list(make_lambdas(random.randint(1, 10)))
+ random_index = random.randint(0, len(functions)-1)
+ random_function = functions[random_index]()
+ print("{0} equals {1}".format(random_index, random_function))
+ self.assertEqual(random_function, random_index)
if __name__ == "__main__":
unittest.main() | Add a test to generate a range of functions with closures. Check if it returns the expected value | ## Code Before:
""" Explore how python works with lambda expressions. """
import unittest
import string
import random
class TestGetWebsites(unittest.TestCase):
def test_closure(self, m):
""" See that python supports closures similar to JavaScript """
def gibberish():
""" Some random string """
return ''.join([random.choice(string.ascii_letters) for i in range(20)])
value1 = gibberish()
value2 = gibberish()
result = (lambda x: lambda y: x)(value1)(value2)
self.assertEqual(result, value1)
if __name__ == "__main__":
unittest.main()
## Instruction:
Add a test to generate a range of functions with closures. Check if it returns the expected value
## Code After:
""" Explore how python works with lambda expressions. """
import unittest
import string
import random
class TestGetWebsites(unittest.TestCase):
def test_closure(self):
""" See that python supports closures similar to JavaScript """
def gibberish():
""" Some random string """
return ''.join([random.choice(string.ascii_letters) for i in range(20)])
value1 = gibberish()
result = (lambda x: lambda: x)(value1)()
self.assertEqual(result, value1)
def test_closure_ids(self):
""" Show how a variable passed to a function remains in scope
even after it returns """
def make_lambdas(num):
""" Build a lambda generator """
for i in range(num):
func = lambda x: lambda: x
yield func(i)
functions = list(make_lambdas(random.randint(1, 10)))
random_index = random.randint(0, len(functions)-1)
random_function = functions[random_index]()
print("{0} equals {1}".format(random_index, random_function))
self.assertEqual(random_function, random_index)
if __name__ == "__main__":
unittest.main() | // ... existing code ...
def test_closure(self):
""" See that python supports closures similar to JavaScript """
// ... modified code ...
value1 = gibberish()
result = (lambda x: lambda: x)(value1)()
self.assertEqual(result, value1)
def test_closure_ids(self):
""" Show how a variable passed to a function remains in scope
even after it returns """
def make_lambdas(num):
""" Build a lambda generator """
for i in range(num):
func = lambda x: lambda: x
yield func(i)
functions = list(make_lambdas(random.randint(1, 10)))
random_index = random.randint(0, len(functions)-1)
random_function = functions[random_index]()
print("{0} equals {1}".format(random_index, random_function))
self.assertEqual(random_function, random_index)
// ... rest of the code ... |
42e16bf376a64995a8b70a91829a82d7b0f3e1a1 | gameanalysis/__main__.py | gameanalysis/__main__.py | """Command line module"""
import argparse
import pkgutil
import sys
import gameanalysis
from gameanalysis import script
def create_parser():
"""Create the default parser"""
modules = [imp.find_module(name).load_module(name) for imp, name, _
in pkgutil.iter_modules(script.__path__)]
parser = argparse.ArgumentParser(
description="""Command line access to the game analysis toolkit.""")
parser.add_argument('-V', '--version', action='version',
version='%(prog)s {}'.format(gameanalysis.__version__))
subparsers = parser.add_subparsers(
title='commands', dest='command', metavar='<command>', help="""The
commands to execute. Available commands are:""")
for module in modules:
subparser = module.add_parser(subparsers)
subparser.main = module.main
return parser, subparsers.choices
def amain(*argv):
"""Entry point for game analysis"""
parser, commands = create_parser()
args = parser.parse_args(argv)
if args.command is None:
parser.print_help()
sys.exit(1)
else:
commands[args.command].main(args)
def main():
"""Entry point for game analysis"""
amain(*sys.argv[1:])
if __name__ == '__main__':
main()
| """Command line module"""
import argparse
import logging
import pkgutil
import sys
import gameanalysis
from gameanalysis import script
def create_parser():
"""Create the default parser"""
modules = [imp.find_module(name).load_module(name) for imp, name, _
in pkgutil.iter_modules(script.__path__)]
parser = argparse.ArgumentParser(
description="""Command line access to the game analysis toolkit.""")
parser.add_argument('-V', '--version', action='version',
version='%(prog)s {}'.format(gameanalysis.__version__))
parser.add_argument(
'-v', '--verbose', action='count', default=0, help="""Set the verbosity
level depending on the number of times specified, up to a maximum of
three.""")
subparsers = parser.add_subparsers(
title='commands', dest='command', metavar='<command>', help="""The
commands to execute. Available commands are:""")
for module in modules:
subparser = module.add_parser(subparsers)
subparser.main = module.main
return parser, subparsers.choices
def amain(*argv):
"""Entry point for game analysis"""
parser, commands = create_parser()
args = parser.parse_args(argv)
if args.command is None:
parser.print_help()
sys.exit(1)
else:
logging.basicConfig(level=40 - 10 * min(args.verbose, 3))
commands[args.command].main(args)
def main():
"""Entry point for game analysis"""
amain(*sys.argv[1:])
if __name__ == '__main__':
main()
| Add logging verbosity to game analysis | Add logging verbosity to game analysis
| Python | apache-2.0 | egtaonline/GameAnalysis | """Command line module"""
import argparse
+ import logging
import pkgutil
import sys
import gameanalysis
from gameanalysis import script
def create_parser():
"""Create the default parser"""
modules = [imp.find_module(name).load_module(name) for imp, name, _
in pkgutil.iter_modules(script.__path__)]
parser = argparse.ArgumentParser(
description="""Command line access to the game analysis toolkit.""")
parser.add_argument('-V', '--version', action='version',
version='%(prog)s {}'.format(gameanalysis.__version__))
+ parser.add_argument(
+ '-v', '--verbose', action='count', default=0, help="""Set the verbosity
+ level depending on the number of times specified, up to a maximum of
+ three.""")
subparsers = parser.add_subparsers(
title='commands', dest='command', metavar='<command>', help="""The
commands to execute. Available commands are:""")
for module in modules:
subparser = module.add_parser(subparsers)
subparser.main = module.main
return parser, subparsers.choices
def amain(*argv):
"""Entry point for game analysis"""
parser, commands = create_parser()
args = parser.parse_args(argv)
if args.command is None:
parser.print_help()
sys.exit(1)
else:
+ logging.basicConfig(level=40 - 10 * min(args.verbose, 3))
commands[args.command].main(args)
def main():
"""Entry point for game analysis"""
amain(*sys.argv[1:])
if __name__ == '__main__':
main()
| Add logging verbosity to game analysis | ## Code Before:
"""Command line module"""
import argparse
import pkgutil
import sys
import gameanalysis
from gameanalysis import script
def create_parser():
"""Create the default parser"""
modules = [imp.find_module(name).load_module(name) for imp, name, _
in pkgutil.iter_modules(script.__path__)]
parser = argparse.ArgumentParser(
description="""Command line access to the game analysis toolkit.""")
parser.add_argument('-V', '--version', action='version',
version='%(prog)s {}'.format(gameanalysis.__version__))
subparsers = parser.add_subparsers(
title='commands', dest='command', metavar='<command>', help="""The
commands to execute. Available commands are:""")
for module in modules:
subparser = module.add_parser(subparsers)
subparser.main = module.main
return parser, subparsers.choices
def amain(*argv):
"""Entry point for game analysis"""
parser, commands = create_parser()
args = parser.parse_args(argv)
if args.command is None:
parser.print_help()
sys.exit(1)
else:
commands[args.command].main(args)
def main():
"""Entry point for game analysis"""
amain(*sys.argv[1:])
if __name__ == '__main__':
main()
## Instruction:
Add logging verbosity to game analysis
## Code After:
"""Command line module"""
import argparse
import logging
import pkgutil
import sys
import gameanalysis
from gameanalysis import script
def create_parser():
"""Create the default parser"""
modules = [imp.find_module(name).load_module(name) for imp, name, _
in pkgutil.iter_modules(script.__path__)]
parser = argparse.ArgumentParser(
description="""Command line access to the game analysis toolkit.""")
parser.add_argument('-V', '--version', action='version',
version='%(prog)s {}'.format(gameanalysis.__version__))
parser.add_argument(
'-v', '--verbose', action='count', default=0, help="""Set the verbosity
level depending on the number of times specified, up to a maximum of
three.""")
subparsers = parser.add_subparsers(
title='commands', dest='command', metavar='<command>', help="""The
commands to execute. Available commands are:""")
for module in modules:
subparser = module.add_parser(subparsers)
subparser.main = module.main
return parser, subparsers.choices
def amain(*argv):
"""Entry point for game analysis"""
parser, commands = create_parser()
args = parser.parse_args(argv)
if args.command is None:
parser.print_help()
sys.exit(1)
else:
logging.basicConfig(level=40 - 10 * min(args.verbose, 3))
commands[args.command].main(args)
def main():
"""Entry point for game analysis"""
amain(*sys.argv[1:])
if __name__ == '__main__':
main()
| # ... existing code ...
import argparse
import logging
import pkgutil
# ... modified code ...
version='%(prog)s {}'.format(gameanalysis.__version__))
parser.add_argument(
'-v', '--verbose', action='count', default=0, help="""Set the verbosity
level depending on the number of times specified, up to a maximum of
three.""")
subparsers = parser.add_subparsers(
...
else:
logging.basicConfig(level=40 - 10 * min(args.verbose, 3))
commands[args.command].main(args)
# ... rest of the code ... |
d902045e991cc778dabe31e34a6dcd119e19ccd0 | attributes/license/main.py | attributes/license/main.py | from utilities import url_to_json
def run(project_id, repo_path, cursor, **options):
query = 'SELECT url FROM projects WHERE id = ' + str(project_id)
cursor.execute(query)
record = cursor.fetchone()
full_url = record[0].rstrip()
json_response = url_to_json(full_url, headers={
'Accept': 'application/vnd.github.drax-preview+json'
}
)
if 'license' in json_response:
result = 1
else:
result = 0
return result
if __name__ == '__main__':
print("Attribute plugins are not meant to be executed directly.")
| from core import tokenize
from utilities import url_to_json
def run(project_id, repo_path, cursor, **options):
query = 'SELECT url FROM projects WHERE id = ' + str(project_id)
cursor.execute(query)
record = cursor.fetchone()
full_url = tokenize(record[0].rstrip())
json_response = url_to_json(full_url, headers={
'Accept': 'application/vnd.github.drax-preview+json'
}
)
result = 'license' in json_response
return result, int(result)
if __name__ == '__main__':
print("Attribute plugins are not meant to be executed directly.")
| Update license attribute to return binary and raw result | Update license attribute to return binary and raw result
| Python | apache-2.0 | RepoReapers/reaper,RepoReapers/reaper,RepoReapers/reaper,RepoReapers/reaper | + from core import tokenize
from utilities import url_to_json
def run(project_id, repo_path, cursor, **options):
query = 'SELECT url FROM projects WHERE id = ' + str(project_id)
cursor.execute(query)
record = cursor.fetchone()
- full_url = record[0].rstrip()
+ full_url = tokenize(record[0].rstrip())
json_response = url_to_json(full_url, headers={
'Accept': 'application/vnd.github.drax-preview+json'
}
)
- if 'license' in json_response:
+ result = 'license' in json_response
+ return result, int(result)
- result = 1
- else:
- result = 0
-
- return result
if __name__ == '__main__':
print("Attribute plugins are not meant to be executed directly.")
| Update license attribute to return binary and raw result | ## Code Before:
from utilities import url_to_json
def run(project_id, repo_path, cursor, **options):
query = 'SELECT url FROM projects WHERE id = ' + str(project_id)
cursor.execute(query)
record = cursor.fetchone()
full_url = record[0].rstrip()
json_response = url_to_json(full_url, headers={
'Accept': 'application/vnd.github.drax-preview+json'
}
)
if 'license' in json_response:
result = 1
else:
result = 0
return result
if __name__ == '__main__':
print("Attribute plugins are not meant to be executed directly.")
## Instruction:
Update license attribute to return binary and raw result
## Code After:
from core import tokenize
from utilities import url_to_json
def run(project_id, repo_path, cursor, **options):
query = 'SELECT url FROM projects WHERE id = ' + str(project_id)
cursor.execute(query)
record = cursor.fetchone()
full_url = tokenize(record[0].rstrip())
json_response = url_to_json(full_url, headers={
'Accept': 'application/vnd.github.drax-preview+json'
}
)
result = 'license' in json_response
return result, int(result)
if __name__ == '__main__':
print("Attribute plugins are not meant to be executed directly.")
| ...
from core import tokenize
from utilities import url_to_json
...
full_url = tokenize(record[0].rstrip())
json_response = url_to_json(full_url, headers={
...
result = 'license' in json_response
return result, int(result)
... |
5977eb82f2614efe8cde843913db62a93c7978f5 | navigation_extensions.py | navigation_extensions.py | from django.utils.text import capfirst
from django.utils.translation import ugettext_lazy as _
from feincms.module.page.extensions.navigation import NavigationExtension, PagePretender
class ZivinetzNavigationExtension(NavigationExtension):
name = _('Zivinetz navigation extension')
def children(self, page, **kwargs):
request = kwargs.get('request')
if request.user.is_authenticated() and request.user.is_staff:
urls = [
(_('scheduling'), 'admin/scheduling/'),
(_('waitlist'), 'admin/waitlist/'),
(_('drudges'), 'admin/drudges/'),
(_('assignments'), 'admin/assignments/'),
(_('job references'), 'admin/jobreferences/'),
(_('expense reports'), 'admin/expense_reports/'),
(_('regional offices'), 'admin/regional_offices/'),
(_('scope statements'), 'admin/scope_statements/'),
(_('specifications'), 'admin/specifications/'),
]
else:
urls = [
(_('dashboard'), 'dashboard/'),
(_('profile'), 'profile/'),
]
return [PagePretender(
title=capfirst(title),
url='%s%s' % (page.get_navigation_url(), url),
level=page.level+1,
tree_id=page.tree_id,
) for title, url in urls]
| from django.utils.text import capfirst
from django.utils.translation import ugettext_lazy as _
from feincms.module.page.extensions.navigation import NavigationExtension, PagePretender
class ZivinetzNavigationExtension(NavigationExtension):
name = _('Zivinetz navigation extension')
def children(self, page, **kwargs):
request = kwargs.get('request')
if request.user.is_authenticated() and request.user.is_staff:
urls = [
(_('scheduling'), 'admin/scheduling/'),
(_('waitlist'), 'admin/waitlist/'),
(_('drudges'), 'admin/drudges/'),
(_('assignments'), 'admin/assignments/'),
(_('expense reports'), 'admin/expense_reports/'),
(_('regional offices'), 'admin/regional_offices/'),
(_('scope statements'), 'admin/scope_statements/'),
(_('specifications'), 'admin/specifications/'),
]
else:
urls = [
(_('dashboard'), 'dashboard/'),
(_('profile'), 'profile/'),
]
return [PagePretender(
title=capfirst(title),
url='%s%s' % (page.get_navigation_url(), url),
level=page.level+1,
tree_id=page.tree_id,
) for title, url in urls]
| Remove job references from navigation | Remove job references from navigation
| Python | mit | matthiask/zivinetz,matthiask/zivinetz,matthiask/zivinetz,matthiask/zivinetz | from django.utils.text import capfirst
from django.utils.translation import ugettext_lazy as _
from feincms.module.page.extensions.navigation import NavigationExtension, PagePretender
class ZivinetzNavigationExtension(NavigationExtension):
name = _('Zivinetz navigation extension')
def children(self, page, **kwargs):
request = kwargs.get('request')
if request.user.is_authenticated() and request.user.is_staff:
urls = [
(_('scheduling'), 'admin/scheduling/'),
(_('waitlist'), 'admin/waitlist/'),
(_('drudges'), 'admin/drudges/'),
(_('assignments'), 'admin/assignments/'),
- (_('job references'), 'admin/jobreferences/'),
(_('expense reports'), 'admin/expense_reports/'),
(_('regional offices'), 'admin/regional_offices/'),
(_('scope statements'), 'admin/scope_statements/'),
(_('specifications'), 'admin/specifications/'),
]
else:
urls = [
(_('dashboard'), 'dashboard/'),
(_('profile'), 'profile/'),
]
return [PagePretender(
title=capfirst(title),
url='%s%s' % (page.get_navigation_url(), url),
level=page.level+1,
tree_id=page.tree_id,
) for title, url in urls]
| Remove job references from navigation | ## Code Before:
from django.utils.text import capfirst
from django.utils.translation import ugettext_lazy as _
from feincms.module.page.extensions.navigation import NavigationExtension, PagePretender
class ZivinetzNavigationExtension(NavigationExtension):
name = _('Zivinetz navigation extension')
def children(self, page, **kwargs):
request = kwargs.get('request')
if request.user.is_authenticated() and request.user.is_staff:
urls = [
(_('scheduling'), 'admin/scheduling/'),
(_('waitlist'), 'admin/waitlist/'),
(_('drudges'), 'admin/drudges/'),
(_('assignments'), 'admin/assignments/'),
(_('job references'), 'admin/jobreferences/'),
(_('expense reports'), 'admin/expense_reports/'),
(_('regional offices'), 'admin/regional_offices/'),
(_('scope statements'), 'admin/scope_statements/'),
(_('specifications'), 'admin/specifications/'),
]
else:
urls = [
(_('dashboard'), 'dashboard/'),
(_('profile'), 'profile/'),
]
return [PagePretender(
title=capfirst(title),
url='%s%s' % (page.get_navigation_url(), url),
level=page.level+1,
tree_id=page.tree_id,
) for title, url in urls]
## Instruction:
Remove job references from navigation
## Code After:
from django.utils.text import capfirst
from django.utils.translation import ugettext_lazy as _
from feincms.module.page.extensions.navigation import NavigationExtension, PagePretender
class ZivinetzNavigationExtension(NavigationExtension):
name = _('Zivinetz navigation extension')
def children(self, page, **kwargs):
request = kwargs.get('request')
if request.user.is_authenticated() and request.user.is_staff:
urls = [
(_('scheduling'), 'admin/scheduling/'),
(_('waitlist'), 'admin/waitlist/'),
(_('drudges'), 'admin/drudges/'),
(_('assignments'), 'admin/assignments/'),
(_('expense reports'), 'admin/expense_reports/'),
(_('regional offices'), 'admin/regional_offices/'),
(_('scope statements'), 'admin/scope_statements/'),
(_('specifications'), 'admin/specifications/'),
]
else:
urls = [
(_('dashboard'), 'dashboard/'),
(_('profile'), 'profile/'),
]
return [PagePretender(
title=capfirst(title),
url='%s%s' % (page.get_navigation_url(), url),
level=page.level+1,
tree_id=page.tree_id,
) for title, url in urls]
| ...
(_('assignments'), 'admin/assignments/'),
(_('expense reports'), 'admin/expense_reports/'),
... |
267c6591ef7ab3354b0965902585203fbfe04dee | pybot/http_client.py | pybot/http_client.py | import requests, json
from resources.urls import FACEBOOK_MESSAGES_POST_URL
class HttpClient():
"""
Client which excutes the call to
facebook's messenger api
"""
def submit_request(self, path, method, payload, completion):
assert len(path) > 0
path = self.get_api_url(path)
headers = self.get_headers()
if method == 'GET':
request = requests.get(path, headers=headers)
if request.status_code is not 200:
error = self.get_error_from_request(request)
completion(None, error)
else:
json_ = self.get_json(request.text)
completion(json_, None)
elif method == 'POST':
raise NotImplementedError
def get_error_from_request(self, request):
return {
'error': self.get_json(request.text)
}
def get_json(self, string):
return json.loads(string)
def get_api_url(self, path):
return FACEBOOK_MESSAGES_POST_URL + path
def get_headers(self):
return {}
| import requests, json
from resources.urls import FACEBOOK_MESSAGES_POST_URL
class HttpClient():
"""
Client which excutes the call to
facebook's messenger api
"""
def submit_request(self, path, method, payload, completion):
assert len(path) > 0
path = self.get_api_url(path)
headers = self.get_headers()
if method == 'GET':
response = requests.get(path, headers=headers)
if response.status_code is not 200:
error = self.get_error_from_request(response)
completion(None, error)
else:
json_ = self.get_json(response.text)
completion(json_, None)
elif method == 'POST':
re
def get_error_from_response(self, response):
return {
'error': self.get_json(response.text)
}
def get_json(self, string):
return json.loads(string)
def get_api_url(self, path):
return FACEBOOK_MESSAGES_POST_URL + path
def get_headers(self):
return {}
| Change wording to use response | Change wording to use response
| Python | mit | ben-cunningham/pybot,ben-cunningham/python-messenger-bot | import requests, json
from resources.urls import FACEBOOK_MESSAGES_POST_URL
class HttpClient():
"""
Client which excutes the call to
facebook's messenger api
"""
def submit_request(self, path, method, payload, completion):
assert len(path) > 0
path = self.get_api_url(path)
headers = self.get_headers()
if method == 'GET':
- request = requests.get(path, headers=headers)
+ response = requests.get(path, headers=headers)
- if request.status_code is not 200:
+ if response.status_code is not 200:
- error = self.get_error_from_request(request)
+ error = self.get_error_from_request(response)
completion(None, error)
else:
- json_ = self.get_json(request.text)
+ json_ = self.get_json(response.text)
completion(json_, None)
elif method == 'POST':
- raise NotImplementedError
+ re
- def get_error_from_request(self, request):
+ def get_error_from_response(self, response):
return {
- 'error': self.get_json(request.text)
+ 'error': self.get_json(response.text)
}
def get_json(self, string):
return json.loads(string)
def get_api_url(self, path):
return FACEBOOK_MESSAGES_POST_URL + path
def get_headers(self):
return {}
| Change wording to use response | ## Code Before:
import requests, json
from resources.urls import FACEBOOK_MESSAGES_POST_URL
class HttpClient():
"""
Client which excutes the call to
facebook's messenger api
"""
def submit_request(self, path, method, payload, completion):
assert len(path) > 0
path = self.get_api_url(path)
headers = self.get_headers()
if method == 'GET':
request = requests.get(path, headers=headers)
if request.status_code is not 200:
error = self.get_error_from_request(request)
completion(None, error)
else:
json_ = self.get_json(request.text)
completion(json_, None)
elif method == 'POST':
raise NotImplementedError
def get_error_from_request(self, request):
return {
'error': self.get_json(request.text)
}
def get_json(self, string):
return json.loads(string)
def get_api_url(self, path):
return FACEBOOK_MESSAGES_POST_URL + path
def get_headers(self):
return {}
## Instruction:
Change wording to use response
## Code After:
import requests, json
from resources.urls import FACEBOOK_MESSAGES_POST_URL
class HttpClient():
"""
Client which excutes the call to
facebook's messenger api
"""
def submit_request(self, path, method, payload, completion):
assert len(path) > 0
path = self.get_api_url(path)
headers = self.get_headers()
if method == 'GET':
response = requests.get(path, headers=headers)
if response.status_code is not 200:
error = self.get_error_from_request(response)
completion(None, error)
else:
json_ = self.get_json(response.text)
completion(json_, None)
elif method == 'POST':
re
def get_error_from_response(self, response):
return {
'error': self.get_json(response.text)
}
def get_json(self, string):
return json.loads(string)
def get_api_url(self, path):
return FACEBOOK_MESSAGES_POST_URL + path
def get_headers(self):
return {}
| ...
if method == 'GET':
response = requests.get(path, headers=headers)
if response.status_code is not 200:
error = self.get_error_from_request(response)
completion(None, error)
...
else:
json_ = self.get_json(response.text)
completion(json_, None)
...
elif method == 'POST':
re
def get_error_from_response(self, response):
...
return {
'error': self.get_json(response.text)
}
... |
5786942c88420be913705790489676780dcd9fc0 | nlppln/utils.py | nlppln/utils.py | """NLP pipeline utility functionality"""
import os
def remove_ext(fname):
"""Removes the extension from a filename
"""
bn = os.path.basename(fname)
return os.path.splitext(bn)[0]
def create_dirs(fname):
"""Create (output) directories if they don't exist
"""
fname = os.path.dirname(fname)
if not os.path.exists(fname):
os.makedirs(fname)
def out_file_name(out_dir, fname, ext):
fname = remove_ext(fname)
return os.path.join(out_dir, '{}.{}'.format(fname, ext))
| """NLP pipeline utility functionality"""
import os
def remove_ext(fname):
"""Removes the extension from a filename
"""
bn = os.path.basename(fname)
return os.path.splitext(bn)[0]
def create_dirs(fname):
"""Create (output) directories if they don't exist
"""
fname = os.path.dirname(os.path.abspath(fname))
if not os.path.exists(fname):
os.makedirs(fname)
def out_file_name(out_dir, fname, ext):
fname = remove_ext(fname)
return os.path.join(out_dir, '{}.{}'.format(fname, ext))
| Update createdirs to determine the absolute path of files | Update createdirs to determine the absolute path of files
| Python | apache-2.0 | WhatWorksWhenForWhom/nlppln,WhatWorksWhenForWhom/nlppln,WhatWorksWhenForWhom/nlppln | """NLP pipeline utility functionality"""
import os
def remove_ext(fname):
"""Removes the extension from a filename
"""
bn = os.path.basename(fname)
return os.path.splitext(bn)[0]
def create_dirs(fname):
"""Create (output) directories if they don't exist
"""
- fname = os.path.dirname(fname)
+ fname = os.path.dirname(os.path.abspath(fname))
if not os.path.exists(fname):
os.makedirs(fname)
def out_file_name(out_dir, fname, ext):
fname = remove_ext(fname)
return os.path.join(out_dir, '{}.{}'.format(fname, ext))
| Update createdirs to determine the absolute path of files | ## Code Before:
"""NLP pipeline utility functionality"""
import os
def remove_ext(fname):
"""Removes the extension from a filename
"""
bn = os.path.basename(fname)
return os.path.splitext(bn)[0]
def create_dirs(fname):
"""Create (output) directories if they don't exist
"""
fname = os.path.dirname(fname)
if not os.path.exists(fname):
os.makedirs(fname)
def out_file_name(out_dir, fname, ext):
fname = remove_ext(fname)
return os.path.join(out_dir, '{}.{}'.format(fname, ext))
## Instruction:
Update createdirs to determine the absolute path of files
## Code After:
"""NLP pipeline utility functionality"""
import os
def remove_ext(fname):
"""Removes the extension from a filename
"""
bn = os.path.basename(fname)
return os.path.splitext(bn)[0]
def create_dirs(fname):
"""Create (output) directories if they don't exist
"""
fname = os.path.dirname(os.path.abspath(fname))
if not os.path.exists(fname):
os.makedirs(fname)
def out_file_name(out_dir, fname, ext):
fname = remove_ext(fname)
return os.path.join(out_dir, '{}.{}'.format(fname, ext))
| // ... existing code ...
"""
fname = os.path.dirname(os.path.abspath(fname))
// ... rest of the code ... |
29aa1a440a9ff225d3f9a4773f9097a5efcbd0de | test/integration/test_output.py | test/integration/test_output.py | from ..helpers import *
def test_honcho_start_joins_stderr_into_stdout():
ret, out, err = get_honcho_output(['-f', 'Procfile.output', 'start'])
assert_equal(ret, 0)
assert_in('some normal output', out)
assert_in('and then write to stderr', out)
assert_equal(err, '')
def test_honcho_run_keeps_stderr_and_stdout_separate():
ret, out, err = get_honcho_output(['run', 'python', 'output.py'])
assert_equal(ret, 0)
assert_equal(out, 'some normal output\n')
assert_equal(err, 'and then write to stderr\n')
| from ..helpers import *
def test_honcho_start_joins_stderr_into_stdout():
ret, out, err = get_honcho_output(['-f', 'Procfile.output', 'start'])
assert_equal(ret, 0)
assert_regexp_matches(out, r'some normal output')
assert_regexp_matches(out, r'and then write to stderr')
assert_equal(err, '')
def test_honcho_run_keeps_stderr_and_stdout_separate():
ret, out, err = get_honcho_output(['run', 'python', 'output.py'])
assert_equal(ret, 0)
assert_equal(out, 'some normal output\n')
assert_equal(err, 'and then write to stderr\n')
| Rewrite assertions for Python 2.6 compatibility | Rewrite assertions for Python 2.6 compatibility
| Python | mit | janusnic/honcho,xarisd/honcho,myyk/honcho,gratipay/honcho,nickstenning/honcho,nickstenning/honcho,gratipay/honcho | from ..helpers import *
def test_honcho_start_joins_stderr_into_stdout():
ret, out, err = get_honcho_output(['-f', 'Procfile.output', 'start'])
assert_equal(ret, 0)
- assert_in('some normal output', out)
- assert_in('and then write to stderr', out)
+ assert_regexp_matches(out, r'some normal output')
+ assert_regexp_matches(out, r'and then write to stderr')
assert_equal(err, '')
def test_honcho_run_keeps_stderr_and_stdout_separate():
ret, out, err = get_honcho_output(['run', 'python', 'output.py'])
assert_equal(ret, 0)
assert_equal(out, 'some normal output\n')
assert_equal(err, 'and then write to stderr\n')
| Rewrite assertions for Python 2.6 compatibility | ## Code Before:
from ..helpers import *
def test_honcho_start_joins_stderr_into_stdout():
ret, out, err = get_honcho_output(['-f', 'Procfile.output', 'start'])
assert_equal(ret, 0)
assert_in('some normal output', out)
assert_in('and then write to stderr', out)
assert_equal(err, '')
def test_honcho_run_keeps_stderr_and_stdout_separate():
ret, out, err = get_honcho_output(['run', 'python', 'output.py'])
assert_equal(ret, 0)
assert_equal(out, 'some normal output\n')
assert_equal(err, 'and then write to stderr\n')
## Instruction:
Rewrite assertions for Python 2.6 compatibility
## Code After:
from ..helpers import *
def test_honcho_start_joins_stderr_into_stdout():
ret, out, err = get_honcho_output(['-f', 'Procfile.output', 'start'])
assert_equal(ret, 0)
assert_regexp_matches(out, r'some normal output')
assert_regexp_matches(out, r'and then write to stderr')
assert_equal(err, '')
def test_honcho_run_keeps_stderr_and_stdout_separate():
ret, out, err = get_honcho_output(['run', 'python', 'output.py'])
assert_equal(ret, 0)
assert_equal(out, 'some normal output\n')
assert_equal(err, 'and then write to stderr\n')
| ...
assert_regexp_matches(out, r'some normal output')
assert_regexp_matches(out, r'and then write to stderr')
assert_equal(err, '')
... |
ee9df63aeaabb4111cea3698a4f0e30b4502e519 | test/disable_captcha.py | test/disable_captcha.py | import owebunit
import urlparse
from wolis_test_case import WolisTestCase
class AcpLoginTestCase(WolisTestCase):
def test_disable_captcha(self):
self.login('morpheus', 'morpheus')
self.acp_login('morpheus', 'morpheus')
start_url = '/adm/index.php'
self.get_with_sid(start_url)
self.assert_status(200)
assert 'Board statistics' in self.response.body
url = self.link_href_by_text('Spambot countermeasures')
# already has sid
self.get(urlparse.urljoin(start_url, url))
self.assert_status(200)
assert 'Enable spambot countermeasures' in self.response.body
assert len(self.response.forms) == 1
form = self.response.forms[0]
params = {
'enable_confirm': '0',
}
params = owebunit.extend_params(form.params.list, params)
self.post(form.computed_action, body=params)
self.assert_status(200)
assert 'Configuration updated successfully' in self.response.body
if __name__ == '__main__':
import unittest
unittest.main()
| import owebunit
import urlparse
from wolis_test_case import WolisTestCase
class AcpLoginTestCase(WolisTestCase):
def test_disable_captcha(self):
self.login('morpheus', 'morpheus')
self.acp_login('morpheus', 'morpheus')
self.change_acp_knob(
link_text='Spambot countermeasures',
check_page_text='Enable spambot countermeasures',
name='enable_confirm',
value='0',
)
def change_acp_knob(self, link_text, check_page_text, name, value):
start_url = '/adm/index.php'
self.get_with_sid(start_url)
self.assert_status(200)
assert 'Board statistics' in self.response.body
url = self.link_href_by_text(link_text)
# already has sid
self.get(urlparse.urljoin(start_url, url))
self.assert_status(200)
assert check_page_text in self.response.body
assert len(self.response.forms) == 1
form = self.response.forms[0]
params = {
name: value,
}
params = owebunit.extend_params(form.params.list, params)
self.post(form.computed_action, body=params)
self.assert_status(200)
assert 'Configuration updated successfully' in self.response.body
if __name__ == '__main__':
import unittest
unittest.main()
| Generalize to allow editing other configuration | Generalize to allow editing other configuration
| Python | bsd-2-clause | p/wolis-phpbb,p/wolis-phpbb | import owebunit
import urlparse
from wolis_test_case import WolisTestCase
class AcpLoginTestCase(WolisTestCase):
def test_disable_captcha(self):
self.login('morpheus', 'morpheus')
self.acp_login('morpheus', 'morpheus')
+ self.change_acp_knob(
+ link_text='Spambot countermeasures',
+ check_page_text='Enable spambot countermeasures',
+ name='enable_confirm',
+ value='0',
+ )
+
+ def change_acp_knob(self, link_text, check_page_text, name, value):
start_url = '/adm/index.php'
self.get_with_sid(start_url)
self.assert_status(200)
assert 'Board statistics' in self.response.body
- url = self.link_href_by_text('Spambot countermeasures')
+ url = self.link_href_by_text(link_text)
# already has sid
self.get(urlparse.urljoin(start_url, url))
self.assert_status(200)
- assert 'Enable spambot countermeasures' in self.response.body
+ assert check_page_text in self.response.body
assert len(self.response.forms) == 1
form = self.response.forms[0]
params = {
- 'enable_confirm': '0',
+ name: value,
}
params = owebunit.extend_params(form.params.list, params)
self.post(form.computed_action, body=params)
self.assert_status(200)
assert 'Configuration updated successfully' in self.response.body
if __name__ == '__main__':
import unittest
unittest.main()
| Generalize to allow editing other configuration | ## Code Before:
import owebunit
import urlparse
from wolis_test_case import WolisTestCase
class AcpLoginTestCase(WolisTestCase):
def test_disable_captcha(self):
self.login('morpheus', 'morpheus')
self.acp_login('morpheus', 'morpheus')
start_url = '/adm/index.php'
self.get_with_sid(start_url)
self.assert_status(200)
assert 'Board statistics' in self.response.body
url = self.link_href_by_text('Spambot countermeasures')
# already has sid
self.get(urlparse.urljoin(start_url, url))
self.assert_status(200)
assert 'Enable spambot countermeasures' in self.response.body
assert len(self.response.forms) == 1
form = self.response.forms[0]
params = {
'enable_confirm': '0',
}
params = owebunit.extend_params(form.params.list, params)
self.post(form.computed_action, body=params)
self.assert_status(200)
assert 'Configuration updated successfully' in self.response.body
if __name__ == '__main__':
import unittest
unittest.main()
## Instruction:
Generalize to allow editing other configuration
## Code After:
import owebunit
import urlparse
from wolis_test_case import WolisTestCase
class AcpLoginTestCase(WolisTestCase):
def test_disable_captcha(self):
self.login('morpheus', 'morpheus')
self.acp_login('morpheus', 'morpheus')
self.change_acp_knob(
link_text='Spambot countermeasures',
check_page_text='Enable spambot countermeasures',
name='enable_confirm',
value='0',
)
def change_acp_knob(self, link_text, check_page_text, name, value):
start_url = '/adm/index.php'
self.get_with_sid(start_url)
self.assert_status(200)
assert 'Board statistics' in self.response.body
url = self.link_href_by_text(link_text)
# already has sid
self.get(urlparse.urljoin(start_url, url))
self.assert_status(200)
assert check_page_text in self.response.body
assert len(self.response.forms) == 1
form = self.response.forms[0]
params = {
name: value,
}
params = owebunit.extend_params(form.params.list, params)
self.post(form.computed_action, body=params)
self.assert_status(200)
assert 'Configuration updated successfully' in self.response.body
if __name__ == '__main__':
import unittest
unittest.main()
| ...
self.change_acp_knob(
link_text='Spambot countermeasures',
check_page_text='Enable spambot countermeasures',
name='enable_confirm',
value='0',
)
def change_acp_knob(self, link_text, check_page_text, name, value):
start_url = '/adm/index.php'
...
url = self.link_href_by_text(link_text)
...
assert check_page_text in self.response.body
...
params = {
name: value,
}
... |
73e5fe29074f52e0b769fd2a6c40669040bef330 | app/notify_client/invite_api_client.py | app/notify_client/invite_api_client.py |
from notifications_python_client.base import BaseAPIClient
from app.notify_client.models import User
class InviteApiClient(BaseAPIClient):
def __init__(self, base_url=None, client_id=None, secret=None):
super(self.__class__, self).__init__(base_url=base_url or 'base_url',
client_id=client_id or 'client_id',
secret=secret or 'secret')
def init_app(self, app):
self.base_url = app.config['API_HOST_NAME']
self.client_id = app.config['ADMIN_CLIENT_USER_NAME']
self.secret = app.config['ADMIN_CLIENT_SECRET']
def create_invite(self, invite_from_id, service_id, email_address, permissions):
data = {
'service': str(service_id),
'email_address': email_address,
'from_user': invite_from_id,
'permissions': permissions
}
resp = self.post(url='/service/{}/invite'.format(service_id), data=data)
return resp['data']
def get_invites_for_service(self, service_id):
endpoint = '/service/{}/invite'.format(service_id)
resp = self.get(endpoint)
return [User(data) for data in resp['data']]
def cancel_invited_user(self, service_id, invited_user_id):
data = {'status': 'cancelled'}
resp = self.post(url='/service/{0}/invite/{1}'.format(service_id, invited_user_id),
data=data)
return resp['data']
|
from notifications_python_client.base import BaseAPIClient
from app.notify_client.models import User
class InviteApiClient(BaseAPIClient):
def __init__(self, base_url=None, client_id=None, secret=None):
super(self.__class__, self).__init__(base_url=base_url or 'base_url',
client_id=client_id or 'client_id',
secret=secret or 'secret')
def init_app(self, app):
self.base_url = app.config['API_HOST_NAME']
self.client_id = app.config['ADMIN_CLIENT_USER_NAME']
self.secret = app.config['ADMIN_CLIENT_SECRET']
def create_invite(self, invite_from_id, service_id, email_address, permissions):
data = {
'service': str(service_id),
'email_address': email_address,
'from_user': invite_from_id,
'permissions': permissions
}
resp = self.post(url='/service/{}/invite'.format(service_id), data=data)
return resp['data']
def get_invites_for_service(self, service_id):
endpoint = '/service/{}/invite'.format(service_id)
resp = self.get(endpoint)
return [User(data) for data in resp['data']]
def cancel_invited_user(self, service_id, invited_user_id):
data = {'status': 'cancelled'}
self.post(url='/service/{0}/invite/{1}'.format(service_id, invited_user_id),
data=data)
| Change cancel_invited_user client to not return anything. | Change cancel_invited_user client to not return anything.
| Python | mit | alphagov/notifications-admin,gov-cjwaszczuk/notifications-admin,alphagov/notifications-admin,gov-cjwaszczuk/notifications-admin,gov-cjwaszczuk/notifications-admin,alphagov/notifications-admin,alphagov/notifications-admin,gov-cjwaszczuk/notifications-admin |
from notifications_python_client.base import BaseAPIClient
from app.notify_client.models import User
class InviteApiClient(BaseAPIClient):
def __init__(self, base_url=None, client_id=None, secret=None):
super(self.__class__, self).__init__(base_url=base_url or 'base_url',
client_id=client_id or 'client_id',
secret=secret or 'secret')
def init_app(self, app):
self.base_url = app.config['API_HOST_NAME']
self.client_id = app.config['ADMIN_CLIENT_USER_NAME']
self.secret = app.config['ADMIN_CLIENT_SECRET']
def create_invite(self, invite_from_id, service_id, email_address, permissions):
data = {
'service': str(service_id),
'email_address': email_address,
'from_user': invite_from_id,
'permissions': permissions
}
resp = self.post(url='/service/{}/invite'.format(service_id), data=data)
return resp['data']
def get_invites_for_service(self, service_id):
endpoint = '/service/{}/invite'.format(service_id)
resp = self.get(endpoint)
return [User(data) for data in resp['data']]
def cancel_invited_user(self, service_id, invited_user_id):
data = {'status': 'cancelled'}
- resp = self.post(url='/service/{0}/invite/{1}'.format(service_id, invited_user_id),
+ self.post(url='/service/{0}/invite/{1}'.format(service_id, invited_user_id),
- data=data)
+ data=data)
- return resp['data']
| Change cancel_invited_user client to not return anything. | ## Code Before:
from notifications_python_client.base import BaseAPIClient
from app.notify_client.models import User
class InviteApiClient(BaseAPIClient):
def __init__(self, base_url=None, client_id=None, secret=None):
super(self.__class__, self).__init__(base_url=base_url or 'base_url',
client_id=client_id or 'client_id',
secret=secret or 'secret')
def init_app(self, app):
self.base_url = app.config['API_HOST_NAME']
self.client_id = app.config['ADMIN_CLIENT_USER_NAME']
self.secret = app.config['ADMIN_CLIENT_SECRET']
def create_invite(self, invite_from_id, service_id, email_address, permissions):
data = {
'service': str(service_id),
'email_address': email_address,
'from_user': invite_from_id,
'permissions': permissions
}
resp = self.post(url='/service/{}/invite'.format(service_id), data=data)
return resp['data']
def get_invites_for_service(self, service_id):
endpoint = '/service/{}/invite'.format(service_id)
resp = self.get(endpoint)
return [User(data) for data in resp['data']]
def cancel_invited_user(self, service_id, invited_user_id):
data = {'status': 'cancelled'}
resp = self.post(url='/service/{0}/invite/{1}'.format(service_id, invited_user_id),
data=data)
return resp['data']
## Instruction:
Change cancel_invited_user client to not return anything.
## Code After:
from notifications_python_client.base import BaseAPIClient
from app.notify_client.models import User
class InviteApiClient(BaseAPIClient):
def __init__(self, base_url=None, client_id=None, secret=None):
super(self.__class__, self).__init__(base_url=base_url or 'base_url',
client_id=client_id or 'client_id',
secret=secret or 'secret')
def init_app(self, app):
self.base_url = app.config['API_HOST_NAME']
self.client_id = app.config['ADMIN_CLIENT_USER_NAME']
self.secret = app.config['ADMIN_CLIENT_SECRET']
def create_invite(self, invite_from_id, service_id, email_address, permissions):
data = {
'service': str(service_id),
'email_address': email_address,
'from_user': invite_from_id,
'permissions': permissions
}
resp = self.post(url='/service/{}/invite'.format(service_id), data=data)
return resp['data']
def get_invites_for_service(self, service_id):
endpoint = '/service/{}/invite'.format(service_id)
resp = self.get(endpoint)
return [User(data) for data in resp['data']]
def cancel_invited_user(self, service_id, invited_user_id):
data = {'status': 'cancelled'}
self.post(url='/service/{0}/invite/{1}'.format(service_id, invited_user_id),
data=data)
| ...
data = {'status': 'cancelled'}
self.post(url='/service/{0}/invite/{1}'.format(service_id, invited_user_id),
data=data)
... |
60625877a23e26e66c2c97cbeb4f139ede717eda | B.py | B.py |
from collections import namedtuple
import matplotlib.pyplot as plt
BCand = namedtuple('BCand', ['m', 'merr', 'pt', 'p'])
bs = []
with open('B.txt') as f:
for line in f.readlines()[1:]:
bs.append(BCand(*[float(v) for v in line.strip().split(',')]))
masses = [b.m for b in bs]
plt.hist(masses, 60, histtype='stepfilled')
plt.xlabel(r'$m_B / \mathrm{GeV}$')
plt.savefig('mass.pdf')
|
from collections import namedtuple
import matplotlib.pyplot as plt
import numpy as np
BCand = namedtuple('BCand', ['m', 'merr', 'pt', 'p'])
bs = [BCand(*b) for b in np.genfromtxt('B.txt', skip_header=1, delimiter=',')]
masses = [b.m for b in bs]
ns, bins, _ = plt.hist(masses, 60, histtype='stepfilled', facecolor='r',
edgecolor='none')
centers = bins[:-1] + (bins[1:] - bins[:-1]) / 2
merr = np.sqrt(ns)
plt.errorbar(centers, ns, yerr=merr, fmt='b+')
plt.xlabel(r'$m_B / \mathrm{GeV}$')
plt.savefig('mass.pdf')
| Use numpy for readin and add errorbars. | Use numpy for readin and add errorbars.
| Python | mit | bixel/python-introduction |
from collections import namedtuple
import matplotlib.pyplot as plt
+ import numpy as np
BCand = namedtuple('BCand', ['m', 'merr', 'pt', 'p'])
+ bs = [BCand(*b) for b in np.genfromtxt('B.txt', skip_header=1, delimiter=',')]
- bs = []
-
- with open('B.txt') as f:
- for line in f.readlines()[1:]:
- bs.append(BCand(*[float(v) for v in line.strip().split(',')]))
masses = [b.m for b in bs]
- plt.hist(masses, 60, histtype='stepfilled')
+ ns, bins, _ = plt.hist(masses, 60, histtype='stepfilled', facecolor='r',
+ edgecolor='none')
+ centers = bins[:-1] + (bins[1:] - bins[:-1]) / 2
+ merr = np.sqrt(ns)
+ plt.errorbar(centers, ns, yerr=merr, fmt='b+')
plt.xlabel(r'$m_B / \mathrm{GeV}$')
plt.savefig('mass.pdf')
| Use numpy for readin and add errorbars. | ## Code Before:
from collections import namedtuple
import matplotlib.pyplot as plt
BCand = namedtuple('BCand', ['m', 'merr', 'pt', 'p'])
bs = []
with open('B.txt') as f:
for line in f.readlines()[1:]:
bs.append(BCand(*[float(v) for v in line.strip().split(',')]))
masses = [b.m for b in bs]
plt.hist(masses, 60, histtype='stepfilled')
plt.xlabel(r'$m_B / \mathrm{GeV}$')
plt.savefig('mass.pdf')
## Instruction:
Use numpy for readin and add errorbars.
## Code After:
from collections import namedtuple
import matplotlib.pyplot as plt
import numpy as np
BCand = namedtuple('BCand', ['m', 'merr', 'pt', 'p'])
bs = [BCand(*b) for b in np.genfromtxt('B.txt', skip_header=1, delimiter=',')]
masses = [b.m for b in bs]
ns, bins, _ = plt.hist(masses, 60, histtype='stepfilled', facecolor='r',
edgecolor='none')
centers = bins[:-1] + (bins[1:] - bins[:-1]) / 2
merr = np.sqrt(ns)
plt.errorbar(centers, ns, yerr=merr, fmt='b+')
plt.xlabel(r'$m_B / \mathrm{GeV}$')
plt.savefig('mass.pdf')
| ...
import matplotlib.pyplot as plt
import numpy as np
...
bs = [BCand(*b) for b in np.genfromtxt('B.txt', skip_header=1, delimiter=',')]
...
ns, bins, _ = plt.hist(masses, 60, histtype='stepfilled', facecolor='r',
edgecolor='none')
centers = bins[:-1] + (bins[1:] - bins[:-1]) / 2
merr = np.sqrt(ns)
plt.errorbar(centers, ns, yerr=merr, fmt='b+')
plt.xlabel(r'$m_B / \mathrm{GeV}$')
... |
f814e945d3e62c87c5f86ef5ac37c5feb733b83d | tests/test_ext.py | tests/test_ext.py | from __future__ import absolute_import, unicode_literals
import unittest
from mopidy import config, ext
class ExtensionTest(unittest.TestCase):
def setUp(self): # noqa: N802
self.ext = ext.Extension()
def test_dist_name_is_none(self):
self.assertIsNone(self.ext.dist_name)
def test_ext_name_is_none(self):
self.assertIsNone(self.ext.ext_name)
def test_version_is_none(self):
self.assertIsNone(self.ext.version)
def test_get_default_config_raises_not_implemented(self):
with self.assertRaises(NotImplementedError):
self.ext.get_default_config()
def test_get_config_schema_returns_extension_schema(self):
schema = self.ext.get_config_schema()
self.assertIsInstance(schema['enabled'], config.Boolean)
def test_validate_environment_does_nothing_by_default(self):
self.assertIsNone(self.ext.validate_environment())
def test_setup_raises_not_implemented(self):
with self.assertRaises(NotImplementedError):
self.ext.setup(None)
| from __future__ import absolute_import, unicode_literals
import pytest
from mopidy import config, ext
@pytest.fixture
def extension():
return ext.Extension()
def test_dist_name_is_none(extension):
assert extension.dist_name is None
def test_ext_name_is_none(extension):
assert extension.ext_name is None
def test_version_is_none(extension):
assert extension.version is None
def test_get_default_config_raises_not_implemented(extension):
with pytest.raises(NotImplementedError):
extension.get_default_config()
def test_get_config_schema_returns_extension_schema(extension):
schema = extension.get_config_schema()
assert isinstance(schema['enabled'], config.Boolean)
def test_validate_environment_does_nothing_by_default(extension):
assert extension.validate_environment() is None
def test_setup_raises_not_implemented(extension):
with pytest.raises(NotImplementedError):
extension.setup(None)
| Convert ext test to pytests | tests: Convert ext test to pytests
| Python | apache-2.0 | mokieyue/mopidy,bencevans/mopidy,ZenithDK/mopidy,jodal/mopidy,quartz55/mopidy,pacificIT/mopidy,pacificIT/mopidy,quartz55/mopidy,ali/mopidy,swak/mopidy,tkem/mopidy,bencevans/mopidy,mopidy/mopidy,SuperStarPL/mopidy,dbrgn/mopidy,hkariti/mopidy,glogiotatidis/mopidy,mokieyue/mopidy,ali/mopidy,bacontext/mopidy,glogiotatidis/mopidy,diandiankan/mopidy,dbrgn/mopidy,mopidy/mopidy,jmarsik/mopidy,glogiotatidis/mopidy,bacontext/mopidy,vrs01/mopidy,kingosticks/mopidy,SuperStarPL/mopidy,diandiankan/mopidy,vrs01/mopidy,dbrgn/mopidy,jcass77/mopidy,glogiotatidis/mopidy,swak/mopidy,bencevans/mopidy,mokieyue/mopidy,adamcik/mopidy,tkem/mopidy,rawdlite/mopidy,bacontext/mopidy,pacificIT/mopidy,ali/mopidy,jcass77/mopidy,rawdlite/mopidy,mopidy/mopidy,vrs01/mopidy,tkem/mopidy,dbrgn/mopidy,SuperStarPL/mopidy,diandiankan/mopidy,tkem/mopidy,jodal/mopidy,diandiankan/mopidy,jmarsik/mopidy,quartz55/mopidy,ZenithDK/mopidy,jodal/mopidy,ZenithDK/mopidy,jmarsik/mopidy,swak/mopidy,ZenithDK/mopidy,bacontext/mopidy,kingosticks/mopidy,quartz55/mopidy,pacificIT/mopidy,SuperStarPL/mopidy,hkariti/mopidy,ali/mopidy,hkariti/mopidy,adamcik/mopidy,vrs01/mopidy,rawdlite/mopidy,hkariti/mopidy,rawdlite/mopidy,mokieyue/mopidy,adamcik/mopidy,swak/mopidy,kingosticks/mopidy,bencevans/mopidy,jmarsik/mopidy,jcass77/mopidy | from __future__ import absolute_import, unicode_literals
- import unittest
+ import pytest
from mopidy import config, ext
- class ExtensionTest(unittest.TestCase):
+ @pytest.fixture
+ def extension():
+ return ext.Extension()
- def setUp(self): # noqa: N802
- self.ext = ext.Extension()
- def test_dist_name_is_none(self):
+ def test_dist_name_is_none(extension):
- self.assertIsNone(self.ext.dist_name)
+ assert extension.dist_name is None
- def test_ext_name_is_none(self):
- self.assertIsNone(self.ext.ext_name)
- def test_version_is_none(self):
- self.assertIsNone(self.ext.version)
+ def test_ext_name_is_none(extension):
+ assert extension.ext_name is None
- def test_get_default_config_raises_not_implemented(self):
- with self.assertRaises(NotImplementedError):
- self.ext.get_default_config()
+ def test_version_is_none(extension):
+ assert extension.version is None
- def test_get_config_schema_returns_extension_schema(self):
- schema = self.ext.get_config_schema()
- self.assertIsInstance(schema['enabled'], config.Boolean)
- def test_validate_environment_does_nothing_by_default(self):
- self.assertIsNone(self.ext.validate_environment())
- def test_setup_raises_not_implemented(self):
+ def test_get_default_config_raises_not_implemented(extension):
- with self.assertRaises(NotImplementedError):
+ with pytest.raises(NotImplementedError):
- self.ext.setup(None)
+ extension.get_default_config()
+
+ def test_get_config_schema_returns_extension_schema(extension):
+ schema = extension.get_config_schema()
+ assert isinstance(schema['enabled'], config.Boolean)
+
+
+ def test_validate_environment_does_nothing_by_default(extension):
+ assert extension.validate_environment() is None
+
+
+ def test_setup_raises_not_implemented(extension):
+ with pytest.raises(NotImplementedError):
+ extension.setup(None)
+ | Convert ext test to pytests | ## Code Before:
from __future__ import absolute_import, unicode_literals
import unittest
from mopidy import config, ext
class ExtensionTest(unittest.TestCase):
def setUp(self): # noqa: N802
self.ext = ext.Extension()
def test_dist_name_is_none(self):
self.assertIsNone(self.ext.dist_name)
def test_ext_name_is_none(self):
self.assertIsNone(self.ext.ext_name)
def test_version_is_none(self):
self.assertIsNone(self.ext.version)
def test_get_default_config_raises_not_implemented(self):
with self.assertRaises(NotImplementedError):
self.ext.get_default_config()
def test_get_config_schema_returns_extension_schema(self):
schema = self.ext.get_config_schema()
self.assertIsInstance(schema['enabled'], config.Boolean)
def test_validate_environment_does_nothing_by_default(self):
self.assertIsNone(self.ext.validate_environment())
def test_setup_raises_not_implemented(self):
with self.assertRaises(NotImplementedError):
self.ext.setup(None)
## Instruction:
Convert ext test to pytests
## Code After:
from __future__ import absolute_import, unicode_literals
import pytest
from mopidy import config, ext
@pytest.fixture
def extension():
return ext.Extension()
def test_dist_name_is_none(extension):
assert extension.dist_name is None
def test_ext_name_is_none(extension):
assert extension.ext_name is None
def test_version_is_none(extension):
assert extension.version is None
def test_get_default_config_raises_not_implemented(extension):
with pytest.raises(NotImplementedError):
extension.get_default_config()
def test_get_config_schema_returns_extension_schema(extension):
schema = extension.get_config_schema()
assert isinstance(schema['enabled'], config.Boolean)
def test_validate_environment_does_nothing_by_default(extension):
assert extension.validate_environment() is None
def test_setup_raises_not_implemented(extension):
with pytest.raises(NotImplementedError):
extension.setup(None)
| // ... existing code ...
import pytest
// ... modified code ...
@pytest.fixture
def extension():
return ext.Extension()
def test_dist_name_is_none(extension):
assert extension.dist_name is None
def test_ext_name_is_none(extension):
assert extension.ext_name is None
def test_version_is_none(extension):
assert extension.version is None
def test_get_default_config_raises_not_implemented(extension):
with pytest.raises(NotImplementedError):
extension.get_default_config()
def test_get_config_schema_returns_extension_schema(extension):
schema = extension.get_config_schema()
assert isinstance(schema['enabled'], config.Boolean)
def test_validate_environment_does_nothing_by_default(extension):
assert extension.validate_environment() is None
def test_setup_raises_not_implemented(extension):
with pytest.raises(NotImplementedError):
extension.setup(None)
// ... rest of the code ... |
8c551fe51ed142305945c0cef530ac84ed3e7eb9 | nodeconductor/logging/perms.py | nodeconductor/logging/perms.py | from nodeconductor.core.permissions import StaffPermissionLogic
PERMISSION_LOGICS = (
('logging.Alert', StaffPermissionLogic(any_permission=True)),
('logging.SystemNotification', StaffPermissionLogic(any_permission=True)),
)
| from nodeconductor.core.permissions import StaffPermissionLogic
PERMISSION_LOGICS = (
('logging.Alert', StaffPermissionLogic(any_permission=True)),
('logging.WebHook', StaffPermissionLogic(any_permission=True)),
('logging.PushHook', StaffPermissionLogic(any_permission=True)),
('logging.EmailHook', StaffPermissionLogic(any_permission=True)),
('logging.SystemNotification', StaffPermissionLogic(any_permission=True)),
)
| Allow staff user to manage hooks. | Allow staff user to manage hooks.
| Python | mit | opennode/nodeconductor,opennode/nodeconductor,opennode/nodeconductor | from nodeconductor.core.permissions import StaffPermissionLogic
PERMISSION_LOGICS = (
('logging.Alert', StaffPermissionLogic(any_permission=True)),
+ ('logging.WebHook', StaffPermissionLogic(any_permission=True)),
+ ('logging.PushHook', StaffPermissionLogic(any_permission=True)),
+ ('logging.EmailHook', StaffPermissionLogic(any_permission=True)),
('logging.SystemNotification', StaffPermissionLogic(any_permission=True)),
)
| Allow staff user to manage hooks. | ## Code Before:
from nodeconductor.core.permissions import StaffPermissionLogic
PERMISSION_LOGICS = (
('logging.Alert', StaffPermissionLogic(any_permission=True)),
('logging.SystemNotification', StaffPermissionLogic(any_permission=True)),
)
## Instruction:
Allow staff user to manage hooks.
## Code After:
from nodeconductor.core.permissions import StaffPermissionLogic
PERMISSION_LOGICS = (
('logging.Alert', StaffPermissionLogic(any_permission=True)),
('logging.WebHook', StaffPermissionLogic(any_permission=True)),
('logging.PushHook', StaffPermissionLogic(any_permission=True)),
('logging.EmailHook', StaffPermissionLogic(any_permission=True)),
('logging.SystemNotification', StaffPermissionLogic(any_permission=True)),
)
| // ... existing code ...
('logging.Alert', StaffPermissionLogic(any_permission=True)),
('logging.WebHook', StaffPermissionLogic(any_permission=True)),
('logging.PushHook', StaffPermissionLogic(any_permission=True)),
('logging.EmailHook', StaffPermissionLogic(any_permission=True)),
('logging.SystemNotification', StaffPermissionLogic(any_permission=True)),
// ... rest of the code ... |
9c16b71ecbb38115f107c7baba56304fb9630ec5 | ocds/export/__init__.py | ocds/export/__init__.py | from .models import (
Release,
ReleasePackage,
Record,
RecordPackage
)
from .schema import Tender
from .helpers import (
mode_test,
get_ocid
)
def release_tender(tender, prefix):
""" returns Release object created from `tender` with ocid `prefix` """
date = tender.get('dateModified', '')
ocid = get_ocid(prefix, tender['tenderID'])
return Release(dict(tender=tender, ocid=ocid, date=date))
def release_tenders(tenders, prefix):
""" returns list of Release object created from `tenders` with amendment info and ocid `prefix` """
prev_tender = next(tenders)
for tender in tenders:
yield Tender.with_diff(prev_tender, tender)
prev_tender = tender
def package_tenders(tenders, params):
data = {}
for field in ReleasePackage._fields:
if field in params:
data[field] = params.get(field, '')
data['releases'] = [release_tender(tender, params.get('prefix')) for tender in tenders]
return ReleasePackage(dict(**data)).serialize()
| from .models import (
Release,
ReleasePackage,
Record,
RecordPackage
)
from .schema import Tender
from .helpers import (
mode_test,
get_ocid
)
def release_tender(tender, prefix):
""" returns Release object created from `tender` with ocid `prefix` """
date = tender.get('dateModified', '')
ocid = get_ocid(prefix, tender['tenderID'])
return Release(dict(tender=tender, ocid=ocid, date=date))
def release_tenders(tenders, prefix):
""" returns list of Release object created from `tenders` with amendment info and ocid `prefix` """
prev_tender = next(tenders)
for tender in tenders:
data = {}
for field in ['tender', 'awards', 'contracts']:
model = getattr(Release, field).model_class
if field in tender:
data[field] = model.fromDiff(prev_tender.get(field, ''), new_tender.get(field, ''))
elif field == 'tender':
data['tender'] = model.fromDiff(prev_tender, tender)
yield Release(data)
prev_tender = tender
def package_tenders(tenders, params):
data = {}
for field in ReleasePackage._fields:
if field in params:
data[field] = params.get(field, '')
data['releases'] = [release_tender(tender, params.get('prefix')) for tender in tenders]
return ReleasePackage(dict(**data)).serialize()
| Update helpers for generating releases | Update helpers for generating releases
| Python | apache-2.0 | yshalenyk/openprocurement.ocds.export,yshalenyk/openprocurement.ocds.export,yshalenyk/ocds.export | from .models import (
Release,
ReleasePackage,
Record,
RecordPackage
)
from .schema import Tender
from .helpers import (
mode_test,
get_ocid
)
def release_tender(tender, prefix):
""" returns Release object created from `tender` with ocid `prefix` """
date = tender.get('dateModified', '')
ocid = get_ocid(prefix, tender['tenderID'])
return Release(dict(tender=tender, ocid=ocid, date=date))
def release_tenders(tenders, prefix):
""" returns list of Release object created from `tenders` with amendment info and ocid `prefix` """
prev_tender = next(tenders)
for tender in tenders:
- yield Tender.with_diff(prev_tender, tender)
+ data = {}
+ for field in ['tender', 'awards', 'contracts']:
+ model = getattr(Release, field).model_class
+ if field in tender:
+ data[field] = model.fromDiff(prev_tender.get(field, ''), new_tender.get(field, ''))
+ elif field == 'tender':
+ data['tender'] = model.fromDiff(prev_tender, tender)
+ yield Release(data)
prev_tender = tender
+
def package_tenders(tenders, params):
data = {}
for field in ReleasePackage._fields:
if field in params:
data[field] = params.get(field, '')
data['releases'] = [release_tender(tender, params.get('prefix')) for tender in tenders]
return ReleasePackage(dict(**data)).serialize()
| Update helpers for generating releases | ## Code Before:
from .models import (
Release,
ReleasePackage,
Record,
RecordPackage
)
from .schema import Tender
from .helpers import (
mode_test,
get_ocid
)
def release_tender(tender, prefix):
""" returns Release object created from `tender` with ocid `prefix` """
date = tender.get('dateModified', '')
ocid = get_ocid(prefix, tender['tenderID'])
return Release(dict(tender=tender, ocid=ocid, date=date))
def release_tenders(tenders, prefix):
""" returns list of Release object created from `tenders` with amendment info and ocid `prefix` """
prev_tender = next(tenders)
for tender in tenders:
yield Tender.with_diff(prev_tender, tender)
prev_tender = tender
def package_tenders(tenders, params):
data = {}
for field in ReleasePackage._fields:
if field in params:
data[field] = params.get(field, '')
data['releases'] = [release_tender(tender, params.get('prefix')) for tender in tenders]
return ReleasePackage(dict(**data)).serialize()
## Instruction:
Update helpers for generating releases
## Code After:
from .models import (
Release,
ReleasePackage,
Record,
RecordPackage
)
from .schema import Tender
from .helpers import (
mode_test,
get_ocid
)
def release_tender(tender, prefix):
""" returns Release object created from `tender` with ocid `prefix` """
date = tender.get('dateModified', '')
ocid = get_ocid(prefix, tender['tenderID'])
return Release(dict(tender=tender, ocid=ocid, date=date))
def release_tenders(tenders, prefix):
""" returns list of Release object created from `tenders` with amendment info and ocid `prefix` """
prev_tender = next(tenders)
for tender in tenders:
data = {}
for field in ['tender', 'awards', 'contracts']:
model = getattr(Release, field).model_class
if field in tender:
data[field] = model.fromDiff(prev_tender.get(field, ''), new_tender.get(field, ''))
elif field == 'tender':
data['tender'] = model.fromDiff(prev_tender, tender)
yield Release(data)
prev_tender = tender
def package_tenders(tenders, params):
data = {}
for field in ReleasePackage._fields:
if field in params:
data[field] = params.get(field, '')
data['releases'] = [release_tender(tender, params.get('prefix')) for tender in tenders]
return ReleasePackage(dict(**data)).serialize()
| # ... existing code ...
for tender in tenders:
data = {}
for field in ['tender', 'awards', 'contracts']:
model = getattr(Release, field).model_class
if field in tender:
data[field] = model.fromDiff(prev_tender.get(field, ''), new_tender.get(field, ''))
elif field == 'tender':
data['tender'] = model.fromDiff(prev_tender, tender)
yield Release(data)
prev_tender = tender
# ... rest of the code ... |
a7328bd229070126ca5b09bb1c9fe4c5e319bb04 | members/urls.py | members/urls.py | from django.conf.urls import patterns, url
from django.contrib import auth
urlpatterns = patterns('members.views',
url(r'^login/$', 'login', name='login'),
url(r'^logout/$', 'logout', name='logout'),
url(r'^search/(?P<name>.*)/$', 'search', name='search'),
url(r'^archive/$', 'archive_student_council', name='archive_student_council'),
)
| from django.conf.urls import patterns, url
from django.contrib import auth
urlpatterns = patterns('members.views',
url(r'^login/$', 'login', name='login'),
url(r'^logout/$', 'logout', name='logout'),
url(r'^search/(?P<name>.*)/$', 'search', name='search'),
url(r'^archive/$', 'archive_student_council', name='archive_student_council'),
url(r'^profile/$', 'user_projects', name='user-projects'),
)
| Add url for user's profile | Add url for user's profile
| Python | mit | Hackfmi/Diaphanum,Hackfmi/Diaphanum | from django.conf.urls import patterns, url
from django.contrib import auth
urlpatterns = patterns('members.views',
url(r'^login/$', 'login', name='login'),
url(r'^logout/$', 'logout', name='logout'),
url(r'^search/(?P<name>.*)/$', 'search', name='search'),
url(r'^archive/$', 'archive_student_council', name='archive_student_council'),
+ url(r'^profile/$', 'user_projects', name='user-projects'),
)
| Add url for user's profile | ## Code Before:
from django.conf.urls import patterns, url
from django.contrib import auth
urlpatterns = patterns('members.views',
url(r'^login/$', 'login', name='login'),
url(r'^logout/$', 'logout', name='logout'),
url(r'^search/(?P<name>.*)/$', 'search', name='search'),
url(r'^archive/$', 'archive_student_council', name='archive_student_council'),
)
## Instruction:
Add url for user's profile
## Code After:
from django.conf.urls import patterns, url
from django.contrib import auth
urlpatterns = patterns('members.views',
url(r'^login/$', 'login', name='login'),
url(r'^logout/$', 'logout', name='logout'),
url(r'^search/(?P<name>.*)/$', 'search', name='search'),
url(r'^archive/$', 'archive_student_council', name='archive_student_council'),
url(r'^profile/$', 'user_projects', name='user-projects'),
)
| # ... existing code ...
url(r'^archive/$', 'archive_student_council', name='archive_student_council'),
url(r'^profile/$', 'user_projects', name='user-projects'),
)
# ... rest of the code ... |
019d33092226d1ff8fe36897c03d25ddd48e34b1 | serve.py | serve.py |
import datetime as dt
import sys
import flask
import sqlalchemy as sa
import coils
import tables
import mapping
app = flask.Flask(__name__)
# Load configuration file.
CONFIG = sys.argv[1] if len(sys.argv)>=2 else 'wabbit.cfg'
config = coils.Config(CONFIG)
@app.route('/')
def index():
"""Render the index page."""
return flask.render_template('index.html')
@app.route('/info')
def info():
"""Return JSON of server info."""
# Connect to database engine.
engine = sa.create_engine(
'mysql://{}:{}@{}/{}'.format(
config['username'], config['password'],
config['host'], config['db_name']))
Session = sa.orm.sessionmaker(bind=engine)
session = Session()
now = dt.datetime.now()
datum = session.query(mapping.Datum).\
filter(mapping.Datum.name=='size')[0]
return flask.jsonify(server_time=now, db_size=datum.value)
if __name__ == '__main__':
app.run()
|
import datetime as dt
import sys
import flask
from flask.ext.sqlalchemy import SQLAlchemy
import coils
import mapping
# Load configuration file.
CONFIG = sys.argv[1] if len(sys.argv)>=2 else 'wabbit.cfg'
config = coils.Config(CONFIG)
# Initialize Flask and SQLAlchemy.
app = flask.Flask(__name__)
app.config['SQLALCHEMY_DATABASE_URI'] = 'mysql://{}:{}@{}/{}'.format(
config['username'], config['password'],
config['host'], config['db_name'])
db = SQLAlchemy(app)
@app.route('/')
def index():
"""Render the index page."""
return flask.render_template('index.html')
@app.route('/info')
def info():
"""Return JSON of server info."""
now = dt.datetime.now()
datum = db.session.query(mapping.Datum).\
filter(mapping.Datum.name=='size')[0]
return flask.jsonify(server_time=now, db_size=datum.value)
if __name__ == '__main__':
app.run()
| Use SQLAlchemy extension in Flask app. | Use SQLAlchemy extension in Flask app.
| Python | mit | vmlaker/wabbit,vmlaker/wabbit,vmlaker/wabbit,vmlaker/wabbit |
import datetime as dt
import sys
import flask
- import sqlalchemy as sa
+ from flask.ext.sqlalchemy import SQLAlchemy
import coils
- import tables
import mapping
-
- app = flask.Flask(__name__)
# Load configuration file.
CONFIG = sys.argv[1] if len(sys.argv)>=2 else 'wabbit.cfg'
config = coils.Config(CONFIG)
+
+ # Initialize Flask and SQLAlchemy.
+ app = flask.Flask(__name__)
+ app.config['SQLALCHEMY_DATABASE_URI'] = 'mysql://{}:{}@{}/{}'.format(
+ config['username'], config['password'],
+ config['host'], config['db_name'])
+ db = SQLAlchemy(app)
@app.route('/')
def index():
"""Render the index page."""
return flask.render_template('index.html')
@app.route('/info')
def info():
"""Return JSON of server info."""
- # Connect to database engine.
- engine = sa.create_engine(
- 'mysql://{}:{}@{}/{}'.format(
- config['username'], config['password'],
- config['host'], config['db_name']))
- Session = sa.orm.sessionmaker(bind=engine)
- session = Session()
now = dt.datetime.now()
- datum = session.query(mapping.Datum).\
+ datum = db.session.query(mapping.Datum).\
filter(mapping.Datum.name=='size')[0]
return flask.jsonify(server_time=now, db_size=datum.value)
if __name__ == '__main__':
app.run()
| Use SQLAlchemy extension in Flask app. | ## Code Before:
import datetime as dt
import sys
import flask
import sqlalchemy as sa
import coils
import tables
import mapping
app = flask.Flask(__name__)
# Load configuration file.
CONFIG = sys.argv[1] if len(sys.argv)>=2 else 'wabbit.cfg'
config = coils.Config(CONFIG)
@app.route('/')
def index():
"""Render the index page."""
return flask.render_template('index.html')
@app.route('/info')
def info():
"""Return JSON of server info."""
# Connect to database engine.
engine = sa.create_engine(
'mysql://{}:{}@{}/{}'.format(
config['username'], config['password'],
config['host'], config['db_name']))
Session = sa.orm.sessionmaker(bind=engine)
session = Session()
now = dt.datetime.now()
datum = session.query(mapping.Datum).\
filter(mapping.Datum.name=='size')[0]
return flask.jsonify(server_time=now, db_size=datum.value)
if __name__ == '__main__':
app.run()
## Instruction:
Use SQLAlchemy extension in Flask app.
## Code After:
import datetime as dt
import sys
import flask
from flask.ext.sqlalchemy import SQLAlchemy
import coils
import mapping
# Load configuration file.
CONFIG = sys.argv[1] if len(sys.argv)>=2 else 'wabbit.cfg'
config = coils.Config(CONFIG)
# Initialize Flask and SQLAlchemy.
app = flask.Flask(__name__)
app.config['SQLALCHEMY_DATABASE_URI'] = 'mysql://{}:{}@{}/{}'.format(
config['username'], config['password'],
config['host'], config['db_name'])
db = SQLAlchemy(app)
@app.route('/')
def index():
"""Render the index page."""
return flask.render_template('index.html')
@app.route('/info')
def info():
"""Return JSON of server info."""
now = dt.datetime.now()
datum = db.session.query(mapping.Datum).\
filter(mapping.Datum.name=='size')[0]
return flask.jsonify(server_time=now, db_size=datum.value)
if __name__ == '__main__':
app.run()
| ...
import flask
from flask.ext.sqlalchemy import SQLAlchemy
import coils
import mapping
...
config = coils.Config(CONFIG)
# Initialize Flask and SQLAlchemy.
app = flask.Flask(__name__)
app.config['SQLALCHEMY_DATABASE_URI'] = 'mysql://{}:{}@{}/{}'.format(
config['username'], config['password'],
config['host'], config['db_name'])
db = SQLAlchemy(app)
...
"""Return JSON of server info."""
now = dt.datetime.now()
datum = db.session.query(mapping.Datum).\
filter(mapping.Datum.name=='size')[0]
... |
8ed94e1fb93252eed47239d8c6a5f28796802a36 | src/cclib/__init__.py | src/cclib/__init__.py |
__version__ = "1.3"
from . import parser
from . import progress
from . import method
from . import bridge
# The test module can be imported if it was installed with cclib.
try:
from . import test
except ImportError:
pass
|
# This file is part of cclib (http://cclib.sf.net), a library for parsing
# and interpreting the results of computational chemistry packages.
#
# Copyright (C) 2006-2014 the cclib development team
#
# The library is free software, distributed under the terms of
# the GNU Lesser General Public version 2.1 or later. You should have
# received a copy of the license along with cclib. You can also access
# the full license online at http://www.gnu.org/copyleft/lgpl.html.
__version__ = "1.3"
from . import parser
from . import progress
from . import method
from . import bridge
# The test module can be imported if it was installed with cclib.
try:
from . import test
except ImportError:
pass
| Add a descriptive docstring to main cclib module | Add a descriptive docstring to main cclib module
| Python | bsd-3-clause | berquist/cclib,jchodera/cclib,ghutchis/cclib,ben-albrecht/cclib,andersx/cclib,gaursagar/cclib,Clyde-fare/cclib,ghutchis/cclib,langner/cclib,andersx/cclib,cclib/cclib,Schamnad/cclib,ATenderholt/cclib,berquist/cclib,cclib/cclib,ATenderholt/cclib,langner/cclib,berquist/cclib,cclib/cclib,langner/cclib,gaursagar/cclib,jchodera/cclib,ben-albrecht/cclib,Schamnad/cclib,Clyde-fare/cclib | +
+ # This file is part of cclib (http://cclib.sf.net), a library for parsing
+ # and interpreting the results of computational chemistry packages.
+ #
+ # Copyright (C) 2006-2014 the cclib development team
+ #
+ # The library is free software, distributed under the terms of
+ # the GNU Lesser General Public version 2.1 or later. You should have
+ # received a copy of the license along with cclib. You can also access
+ # the full license online at http://www.gnu.org/copyleft/lgpl.html.
__version__ = "1.3"
from . import parser
from . import progress
from . import method
from . import bridge
# The test module can be imported if it was installed with cclib.
try:
from . import test
except ImportError:
pass
| Add a descriptive docstring to main cclib module | ## Code Before:
__version__ = "1.3"
from . import parser
from . import progress
from . import method
from . import bridge
# The test module can be imported if it was installed with cclib.
try:
from . import test
except ImportError:
pass
## Instruction:
Add a descriptive docstring to main cclib module
## Code After:
# This file is part of cclib (http://cclib.sf.net), a library for parsing
# and interpreting the results of computational chemistry packages.
#
# Copyright (C) 2006-2014 the cclib development team
#
# The library is free software, distributed under the terms of
# the GNU Lesser General Public version 2.1 or later. You should have
# received a copy of the license along with cclib. You can also access
# the full license online at http://www.gnu.org/copyleft/lgpl.html.
__version__ = "1.3"
from . import parser
from . import progress
from . import method
from . import bridge
# The test module can be imported if it was installed with cclib.
try:
from . import test
except ImportError:
pass
| # ... existing code ...
# This file is part of cclib (http://cclib.sf.net), a library for parsing
# and interpreting the results of computational chemistry packages.
#
# Copyright (C) 2006-2014 the cclib development team
#
# The library is free software, distributed under the terms of
# the GNU Lesser General Public version 2.1 or later. You should have
# received a copy of the license along with cclib. You can also access
# the full license online at http://www.gnu.org/copyleft/lgpl.html.
# ... rest of the code ... |
838012c457d6c963707bb16259cd72d28c231672 | cellcounter/accounts/decorators.py | cellcounter/accounts/decorators.py | __author__ = 'jvc26'
| from functools import wraps
from ratelimit.exceptions import Ratelimited
from ratelimit.helpers import is_ratelimited
def registration_ratelimit(ip=True, block=False, method=['POST'], field=None, rate='1/h',
skip_if=None, keys=None):
def decorator(fn):
@wraps(fn)
def _wrapped(request, *args, **kw):
request.limited = getattr(request, 'limited', False)
if skip_if is None or not skip_if(request):
ratelimited = is_ratelimited(request=request, increment=False,
ip=ip, method=method, field=field,
rate=rate, keys=keys)
if ratelimited and block:
raise Ratelimited()
return_val, success = fn(request, *args, **kw)
if success:
is_ratelimited(request=request, increment=True, ip=ip,
method=method, field=field, rate=rate, keys=keys)
return return_val
return _wrapped
return decorator | Use custom decorator to allow ratelimiting only on successful POST - prevents blocking form errors | Use custom decorator to allow ratelimiting only on successful POST - prevents blocking form errors
| Python | mit | haematologic/cellcounter,cellcounter/cellcounter,cellcounter/cellcounter,haematologic/cellcounter,cellcounter/cellcounter,cellcounter/cellcounter,haematologic/cellcounter | - __author__ = 'jvc26'
+ from functools import wraps
+ from ratelimit.exceptions import Ratelimited
+ from ratelimit.helpers import is_ratelimited
+
+
+ def registration_ratelimit(ip=True, block=False, method=['POST'], field=None, rate='1/h',
+ skip_if=None, keys=None):
+ def decorator(fn):
+ @wraps(fn)
+ def _wrapped(request, *args, **kw):
+ request.limited = getattr(request, 'limited', False)
+ if skip_if is None or not skip_if(request):
+ ratelimited = is_ratelimited(request=request, increment=False,
+ ip=ip, method=method, field=field,
+ rate=rate, keys=keys)
+ if ratelimited and block:
+ raise Ratelimited()
+ return_val, success = fn(request, *args, **kw)
+ if success:
+ is_ratelimited(request=request, increment=True, ip=ip,
+ method=method, field=field, rate=rate, keys=keys)
+ return return_val
+ return _wrapped
+ return decorator | Use custom decorator to allow ratelimiting only on successful POST - prevents blocking form errors | ## Code Before:
__author__ = 'jvc26'
## Instruction:
Use custom decorator to allow ratelimiting only on successful POST - prevents blocking form errors
## Code After:
from functools import wraps
from ratelimit.exceptions import Ratelimited
from ratelimit.helpers import is_ratelimited
def registration_ratelimit(ip=True, block=False, method=['POST'], field=None, rate='1/h',
skip_if=None, keys=None):
def decorator(fn):
@wraps(fn)
def _wrapped(request, *args, **kw):
request.limited = getattr(request, 'limited', False)
if skip_if is None or not skip_if(request):
ratelimited = is_ratelimited(request=request, increment=False,
ip=ip, method=method, field=field,
rate=rate, keys=keys)
if ratelimited and block:
raise Ratelimited()
return_val, success = fn(request, *args, **kw)
if success:
is_ratelimited(request=request, increment=True, ip=ip,
method=method, field=field, rate=rate, keys=keys)
return return_val
return _wrapped
return decorator | # ... existing code ...
from functools import wraps
from ratelimit.exceptions import Ratelimited
from ratelimit.helpers import is_ratelimited
def registration_ratelimit(ip=True, block=False, method=['POST'], field=None, rate='1/h',
skip_if=None, keys=None):
def decorator(fn):
@wraps(fn)
def _wrapped(request, *args, **kw):
request.limited = getattr(request, 'limited', False)
if skip_if is None or not skip_if(request):
ratelimited = is_ratelimited(request=request, increment=False,
ip=ip, method=method, field=field,
rate=rate, keys=keys)
if ratelimited and block:
raise Ratelimited()
return_val, success = fn(request, *args, **kw)
if success:
is_ratelimited(request=request, increment=True, ip=ip,
method=method, field=field, rate=rate, keys=keys)
return return_val
return _wrapped
return decorator
# ... rest of the code ... |
a90889b773010d2fe2ed1dff133f951c0b5baea4 | demo/__init__.py | demo/__init__.py | """Package for PythonTemplateDemo."""
__project__ = 'PythonTemplateDemo'
__version__ = '0.0.0'
VERSION = __project__ + '-' + __version__
PYTHON_VERSION = 2, 7
import sys
if not sys.version_info >= PYTHON_VERSION: # pragma: no cover (manual test)
exit("Python {}.{}+ is required.".format(*PYTHON_VERSION))
| """Package for PythonTemplateDemo."""
import sys
__project__ = 'PythonTemplateDemo'
__version__ = '0.0.0'
VERSION = __project__ + '-' + __version__
PYTHON_VERSION = 2, 7
if not sys.version_info >= PYTHON_VERSION: # pragma: no cover (manual test)
exit("Python {}.{}+ is required.".format(*PYTHON_VERSION))
| Deploy Travis CI build 387 to GitHub | Deploy Travis CI build 387 to GitHub
| Python | mit | jacebrowning/template-python-demo | """Package for PythonTemplateDemo."""
+
+ import sys
__project__ = 'PythonTemplateDemo'
__version__ = '0.0.0'
VERSION = __project__ + '-' + __version__
PYTHON_VERSION = 2, 7
- import sys
if not sys.version_info >= PYTHON_VERSION: # pragma: no cover (manual test)
exit("Python {}.{}+ is required.".format(*PYTHON_VERSION))
| Deploy Travis CI build 387 to GitHub | ## Code Before:
"""Package for PythonTemplateDemo."""
__project__ = 'PythonTemplateDemo'
__version__ = '0.0.0'
VERSION = __project__ + '-' + __version__
PYTHON_VERSION = 2, 7
import sys
if not sys.version_info >= PYTHON_VERSION: # pragma: no cover (manual test)
exit("Python {}.{}+ is required.".format(*PYTHON_VERSION))
## Instruction:
Deploy Travis CI build 387 to GitHub
## Code After:
"""Package for PythonTemplateDemo."""
import sys
__project__ = 'PythonTemplateDemo'
__version__ = '0.0.0'
VERSION = __project__ + '-' + __version__
PYTHON_VERSION = 2, 7
if not sys.version_info >= PYTHON_VERSION: # pragma: no cover (manual test)
exit("Python {}.{}+ is required.".format(*PYTHON_VERSION))
| ...
"""Package for PythonTemplateDemo."""
import sys
...
if not sys.version_info >= PYTHON_VERSION: # pragma: no cover (manual test)
... |
8eb66d72452d69d683a576c75cdf2be72b2370fa | tests/test_utils.py | tests/test_utils.py | import pytest
from mangacork import utils
@pytest.fixture
def sample_page_bad_format():
sample_page = {'chapter': "chapter1", 'page': 3}
return sample_page
@pytest.fixture
def sample_page_good_format():
sample_page = {'chapter':'manga_ch1', 'page':'x_v001-001'}
return sample_page
def test_build_img_path(sample_page_bad_format):
chapter = sample_page_bad_format["chapter"]
page = sample_page_bad_format["page"]
expected_output = "/chapter1/3"
assert utils.build_img_path(chapter,page) == expected_output
def test_increment_page_number_bad_formate(sample_page_bad_format):
with pytest.raises(ValueError):
current_page = utils.build_img_path(sample_page_bad_format["chapter"],
sample_page_bad_format["page"])
utils.increment_page_number(current_page)
| import pytest
from mangacork import utils
@pytest.fixture
def sample_page_bad_format():
sample_page = {'chapter': "chapter1", 'page': 3}
return sample_page
@pytest.fixture
def sample_page_good_format():
sample_page = {'chapter':'manga_ch1', 'page':'x_v001-001'}
return sample_page
def test_build_img_path(sample_page_bad_format):
chapter = sample_page_bad_format["chapter"]
page = sample_page_bad_format["page"]
expected_output = "/chapter1/3"
assert utils.build_img_path(chapter,page) == expected_output
def test_increment_page_number_bad_format(sample_page_bad_format):
with pytest.raises(ValueError):
current_page = utils.build_img_path(sample_page_bad_format["chapter"],
sample_page_bad_format["page"])
utils.increment_page_number(current_page)
def test_increment_page_number_good_format(sample_page_good_format):
chapter = sample_page_good_format["chapter"]
page = sample_page_good_format["page"]
current_page = utils.build_img_path(chapter, page)
next_page = utils.increment_page_number(current_page)
expected_output = '/manga_ch1/x_v001-002'
assert next_page == expected_output
| Add test for inc page num, good format | Add test for inc page num, good format
| Python | mit | ma3lstrom/manga-cork,ma3lstrom/manga-cork,ma3lstrom/manga-cork | import pytest
from mangacork import utils
@pytest.fixture
def sample_page_bad_format():
sample_page = {'chapter': "chapter1", 'page': 3}
return sample_page
@pytest.fixture
def sample_page_good_format():
sample_page = {'chapter':'manga_ch1', 'page':'x_v001-001'}
return sample_page
-
def test_build_img_path(sample_page_bad_format):
chapter = sample_page_bad_format["chapter"]
page = sample_page_bad_format["page"]
expected_output = "/chapter1/3"
assert utils.build_img_path(chapter,page) == expected_output
- def test_increment_page_number_bad_formate(sample_page_bad_format):
+ def test_increment_page_number_bad_format(sample_page_bad_format):
with pytest.raises(ValueError):
current_page = utils.build_img_path(sample_page_bad_format["chapter"],
sample_page_bad_format["page"])
utils.increment_page_number(current_page)
+
+ def test_increment_page_number_good_format(sample_page_good_format):
+ chapter = sample_page_good_format["chapter"]
+ page = sample_page_good_format["page"]
+ current_page = utils.build_img_path(chapter, page)
+ next_page = utils.increment_page_number(current_page)
+ expected_output = '/manga_ch1/x_v001-002'
+ assert next_page == expected_output
+ | Add test for inc page num, good format | ## Code Before:
import pytest
from mangacork import utils
@pytest.fixture
def sample_page_bad_format():
sample_page = {'chapter': "chapter1", 'page': 3}
return sample_page
@pytest.fixture
def sample_page_good_format():
sample_page = {'chapter':'manga_ch1', 'page':'x_v001-001'}
return sample_page
def test_build_img_path(sample_page_bad_format):
chapter = sample_page_bad_format["chapter"]
page = sample_page_bad_format["page"]
expected_output = "/chapter1/3"
assert utils.build_img_path(chapter,page) == expected_output
def test_increment_page_number_bad_formate(sample_page_bad_format):
with pytest.raises(ValueError):
current_page = utils.build_img_path(sample_page_bad_format["chapter"],
sample_page_bad_format["page"])
utils.increment_page_number(current_page)
## Instruction:
Add test for inc page num, good format
## Code After:
import pytest
from mangacork import utils
@pytest.fixture
def sample_page_bad_format():
sample_page = {'chapter': "chapter1", 'page': 3}
return sample_page
@pytest.fixture
def sample_page_good_format():
sample_page = {'chapter':'manga_ch1', 'page':'x_v001-001'}
return sample_page
def test_build_img_path(sample_page_bad_format):
chapter = sample_page_bad_format["chapter"]
page = sample_page_bad_format["page"]
expected_output = "/chapter1/3"
assert utils.build_img_path(chapter,page) == expected_output
def test_increment_page_number_bad_format(sample_page_bad_format):
with pytest.raises(ValueError):
current_page = utils.build_img_path(sample_page_bad_format["chapter"],
sample_page_bad_format["page"])
utils.increment_page_number(current_page)
def test_increment_page_number_good_format(sample_page_good_format):
chapter = sample_page_good_format["chapter"]
page = sample_page_good_format["page"]
current_page = utils.build_img_path(chapter, page)
next_page = utils.increment_page_number(current_page)
expected_output = '/manga_ch1/x_v001-002'
assert next_page == expected_output
| # ... existing code ...
def test_build_img_path(sample_page_bad_format):
# ... modified code ...
def test_increment_page_number_bad_format(sample_page_bad_format):
with pytest.raises(ValueError):
...
utils.increment_page_number(current_page)
def test_increment_page_number_good_format(sample_page_good_format):
chapter = sample_page_good_format["chapter"]
page = sample_page_good_format["page"]
current_page = utils.build_img_path(chapter, page)
next_page = utils.increment_page_number(current_page)
expected_output = '/manga_ch1/x_v001-002'
assert next_page == expected_output
# ... rest of the code ... |
66edf9f04c1b23681fae4234a8b297868e66b7aa | osmaxx-py/osmaxx/excerptexport/models/excerpt.py | osmaxx-py/osmaxx/excerptexport/models/excerpt.py | from django.db import models
from django.contrib.auth.models import User
from django.utils.translation import gettext_lazy as _
class Excerpt(models.Model):
name = models.CharField(max_length=128, verbose_name=_('name'), blank=False)
is_public = models.BooleanField(default=False, verbose_name=_('is public'))
is_active = models.BooleanField(default=True, verbose_name=_('is active'))
owner = models.ForeignKey(User, related_name='excerpts', verbose_name=_('owner'))
bounding_geometry = models.OneToOneField('BoundingGeometry', verbose_name=_('bounding geometry'))
@property
def type_of_geometry(self):
return self.bounding_geometry.type_of_geometry
@property
def extent(self):
return self.bounding_geometry.extent
def __str__(self):
return self.name
def _active_excerpts():
return Excerpt.objects.filter(is_active=True).filter(
bounding_geometry__bboxboundinggeometry__isnull=False
)
def private_user_excerpts(user):
return _active_excerpts().filter(is_public=False, owner=user)
def public_user_excerpts(user):
return _active_excerpts().filter(is_public=True, owner=user)
def other_users_public_excerpts(user):
return _active_excerpts().filter(is_public=True).exclude(owner=user)
| from django.db import models
from django.contrib.auth.models import User
from django.utils.translation import gettext_lazy as _
class Excerpt(models.Model):
name = models.CharField(max_length=128, verbose_name=_('name'))
is_public = models.BooleanField(default=False, verbose_name=_('is public'))
is_active = models.BooleanField(default=True, verbose_name=_('is active'))
owner = models.ForeignKey(User, related_name='excerpts', verbose_name=_('owner'))
bounding_geometry = models.OneToOneField('BoundingGeometry', verbose_name=_('bounding geometry'))
@property
def type_of_geometry(self):
return self.bounding_geometry.type_of_geometry
@property
def extent(self):
return self.bounding_geometry.extent
def __str__(self):
return self.name
def _active_excerpts():
return Excerpt.objects.filter(is_active=True).filter(
bounding_geometry__bboxboundinggeometry__isnull=False
)
def private_user_excerpts(user):
return _active_excerpts().filter(is_public=False, owner=user)
def public_user_excerpts(user):
return _active_excerpts().filter(is_public=True, owner=user)
def other_users_public_excerpts(user):
return _active_excerpts().filter(is_public=True).exclude(owner=user)
| Remove value which is already default | Remove value which is already default
| Python | mit | geometalab/osmaxx,geometalab/drf-utm-zone-info,geometalab/osmaxx,geometalab/osmaxx-frontend,geometalab/osmaxx-frontend,geometalab/osmaxx-frontend,geometalab/osmaxx,geometalab/osmaxx-frontend,geometalab/drf-utm-zone-info,geometalab/osmaxx | from django.db import models
from django.contrib.auth.models import User
from django.utils.translation import gettext_lazy as _
class Excerpt(models.Model):
- name = models.CharField(max_length=128, verbose_name=_('name'), blank=False)
+ name = models.CharField(max_length=128, verbose_name=_('name'))
is_public = models.BooleanField(default=False, verbose_name=_('is public'))
is_active = models.BooleanField(default=True, verbose_name=_('is active'))
owner = models.ForeignKey(User, related_name='excerpts', verbose_name=_('owner'))
bounding_geometry = models.OneToOneField('BoundingGeometry', verbose_name=_('bounding geometry'))
@property
def type_of_geometry(self):
return self.bounding_geometry.type_of_geometry
@property
def extent(self):
return self.bounding_geometry.extent
def __str__(self):
return self.name
def _active_excerpts():
return Excerpt.objects.filter(is_active=True).filter(
bounding_geometry__bboxboundinggeometry__isnull=False
)
def private_user_excerpts(user):
return _active_excerpts().filter(is_public=False, owner=user)
def public_user_excerpts(user):
return _active_excerpts().filter(is_public=True, owner=user)
def other_users_public_excerpts(user):
return _active_excerpts().filter(is_public=True).exclude(owner=user)
| Remove value which is already default | ## Code Before:
from django.db import models
from django.contrib.auth.models import User
from django.utils.translation import gettext_lazy as _
class Excerpt(models.Model):
name = models.CharField(max_length=128, verbose_name=_('name'), blank=False)
is_public = models.BooleanField(default=False, verbose_name=_('is public'))
is_active = models.BooleanField(default=True, verbose_name=_('is active'))
owner = models.ForeignKey(User, related_name='excerpts', verbose_name=_('owner'))
bounding_geometry = models.OneToOneField('BoundingGeometry', verbose_name=_('bounding geometry'))
@property
def type_of_geometry(self):
return self.bounding_geometry.type_of_geometry
@property
def extent(self):
return self.bounding_geometry.extent
def __str__(self):
return self.name
def _active_excerpts():
return Excerpt.objects.filter(is_active=True).filter(
bounding_geometry__bboxboundinggeometry__isnull=False
)
def private_user_excerpts(user):
return _active_excerpts().filter(is_public=False, owner=user)
def public_user_excerpts(user):
return _active_excerpts().filter(is_public=True, owner=user)
def other_users_public_excerpts(user):
return _active_excerpts().filter(is_public=True).exclude(owner=user)
## Instruction:
Remove value which is already default
## Code After:
from django.db import models
from django.contrib.auth.models import User
from django.utils.translation import gettext_lazy as _
class Excerpt(models.Model):
name = models.CharField(max_length=128, verbose_name=_('name'))
is_public = models.BooleanField(default=False, verbose_name=_('is public'))
is_active = models.BooleanField(default=True, verbose_name=_('is active'))
owner = models.ForeignKey(User, related_name='excerpts', verbose_name=_('owner'))
bounding_geometry = models.OneToOneField('BoundingGeometry', verbose_name=_('bounding geometry'))
@property
def type_of_geometry(self):
return self.bounding_geometry.type_of_geometry
@property
def extent(self):
return self.bounding_geometry.extent
def __str__(self):
return self.name
def _active_excerpts():
return Excerpt.objects.filter(is_active=True).filter(
bounding_geometry__bboxboundinggeometry__isnull=False
)
def private_user_excerpts(user):
return _active_excerpts().filter(is_public=False, owner=user)
def public_user_excerpts(user):
return _active_excerpts().filter(is_public=True, owner=user)
def other_users_public_excerpts(user):
return _active_excerpts().filter(is_public=True).exclude(owner=user)
| ...
class Excerpt(models.Model):
name = models.CharField(max_length=128, verbose_name=_('name'))
is_public = models.BooleanField(default=False, verbose_name=_('is public'))
... |
03ee406800fb59ff3e7565397107fa9aad0d54d0 | website/notifications/listeners.py | website/notifications/listeners.py | import logging
from website.notifications.exceptions import InvalidSubscriptionError
from website.notifications.utils import subscribe_user_to_notifications, subscribe_user_to_global_notifications
from website.project.signals import contributor_added, project_created
from framework.auth.signals import user_confirmed
logger = logging.getLogger(__name__)
@project_created.connect
def subscribe_creator(node):
if node.is_collection or node.is_deleted:
return None
try:
subscribe_user_to_notifications(node, node.creator)
except InvalidSubscriptionError as err:
user = node.creator._id if node.creator else 'None'
logger.warn('Skipping subscription of user {} to node {}'.format(user, node._id))
logger.warn('Reason: {}'.format(str(err)))
@contributor_added.connect
def subscribe_contributor(node, contributor, auth=None, *args, **kwargs):
try:
subscribe_user_to_notifications(node, contributor)
except InvalidSubscriptionError as err:
logger.warn('Skipping subscription of user {} to node {}'.format(contributor, node._id))
logger.warn('Reason: {}'.format(str(err)))
@user_confirmed.connect
def subscribe_confirmed_user(user):
try:
subscribe_user_to_global_notifications(user)
except InvalidSubscriptionError as err:
logger.warn('Skipping subscription of user {} to global subscriptions'.format(user))
logger.warn('Reason: {}'.format(str(err)))
| import logging
from website.notifications.exceptions import InvalidSubscriptionError
from website.notifications.utils import subscribe_user_to_notifications, subscribe_user_to_global_notifications
from website.project.signals import contributor_added, project_created
from framework.auth.signals import user_confirmed
logger = logging.getLogger(__name__)
@project_created.connect
def subscribe_creator(node):
if node.institution_id or node.is_collection or node.is_deleted:
return None
try:
subscribe_user_to_notifications(node, node.creator)
except InvalidSubscriptionError as err:
user = node.creator._id if node.creator else 'None'
logger.warn('Skipping subscription of user {} to node {}'.format(user, node._id))
logger.warn('Reason: {}'.format(str(err)))
@contributor_added.connect
def subscribe_contributor(node, contributor, auth=None, *args, **kwargs):
try:
subscribe_user_to_notifications(node, contributor)
except InvalidSubscriptionError as err:
logger.warn('Skipping subscription of user {} to node {}'.format(contributor, node._id))
logger.warn('Reason: {}'.format(str(err)))
@user_confirmed.connect
def subscribe_confirmed_user(user):
try:
subscribe_user_to_global_notifications(user)
except InvalidSubscriptionError as err:
logger.warn('Skipping subscription of user {} to global subscriptions'.format(user))
logger.warn('Reason: {}'.format(str(err)))
| Revert "Remove incorrect check for institution_id" | Revert "Remove incorrect check for institution_id"
This reverts commit 617df13670573b858b6c23249f4287786807d8b6.
| Python | apache-2.0 | hmoco/osf.io,cslzchen/osf.io,monikagrabowska/osf.io,cslzchen/osf.io,cslzchen/osf.io,Nesiehr/osf.io,aaxelb/osf.io,CenterForOpenScience/osf.io,chrisseto/osf.io,chennan47/osf.io,crcresearch/osf.io,Nesiehr/osf.io,felliott/osf.io,Johnetordoff/osf.io,acshi/osf.io,baylee-d/osf.io,HalcyonChimera/osf.io,brianjgeiger/osf.io,aaxelb/osf.io,binoculars/osf.io,cwisecarver/osf.io,CenterForOpenScience/osf.io,mattclark/osf.io,sloria/osf.io,TomBaxter/osf.io,caneruguz/osf.io,adlius/osf.io,hmoco/osf.io,caneruguz/osf.io,chennan47/osf.io,chrisseto/osf.io,mattclark/osf.io,saradbowman/osf.io,aaxelb/osf.io,mfraezz/osf.io,felliott/osf.io,monikagrabowska/osf.io,sloria/osf.io,brianjgeiger/osf.io,CenterForOpenScience/osf.io,TomBaxter/osf.io,erinspace/osf.io,erinspace/osf.io,leb2dg/osf.io,adlius/osf.io,pattisdr/osf.io,laurenrevere/osf.io,caneruguz/osf.io,mfraezz/osf.io,CenterForOpenScience/osf.io,crcresearch/osf.io,adlius/osf.io,adlius/osf.io,cwisecarver/osf.io,leb2dg/osf.io,baylee-d/osf.io,mfraezz/osf.io,caseyrollins/osf.io,mattclark/osf.io,icereval/osf.io,brianjgeiger/osf.io,Nesiehr/osf.io,leb2dg/osf.io,crcresearch/osf.io,monikagrabowska/osf.io,felliott/osf.io,erinspace/osf.io,laurenrevere/osf.io,leb2dg/osf.io,binoculars/osf.io,HalcyonChimera/osf.io,Johnetordoff/osf.io,pattisdr/osf.io,chennan47/osf.io,hmoco/osf.io,acshi/osf.io,cslzchen/osf.io,monikagrabowska/osf.io,Nesiehr/osf.io,pattisdr/osf.io,icereval/osf.io,HalcyonChimera/osf.io,caseyrollins/osf.io,caseyrollins/osf.io,aaxelb/osf.io,laurenrevere/osf.io,TomBaxter/osf.io,monikagrabowska/osf.io,acshi/osf.io,Johnetordoff/osf.io,Johnetordoff/osf.io,cwisecarver/osf.io,chrisseto/osf.io,baylee-d/osf.io,icereval/osf.io,binoculars/osf.io,felliott/osf.io,mfraezz/osf.io,HalcyonChimera/osf.io,brianjgeiger/osf.io,hmoco/osf.io,sloria/osf.io,saradbowman/osf.io,acshi/osf.io,caneruguz/osf.io,acshi/osf.io,chrisseto/osf.io,cwisecarver/osf.io | import logging
from website.notifications.exceptions import InvalidSubscriptionError
from website.notifications.utils import subscribe_user_to_notifications, subscribe_user_to_global_notifications
from website.project.signals import contributor_added, project_created
from framework.auth.signals import user_confirmed
logger = logging.getLogger(__name__)
@project_created.connect
def subscribe_creator(node):
- if node.is_collection or node.is_deleted:
+ if node.institution_id or node.is_collection or node.is_deleted:
return None
try:
subscribe_user_to_notifications(node, node.creator)
except InvalidSubscriptionError as err:
user = node.creator._id if node.creator else 'None'
logger.warn('Skipping subscription of user {} to node {}'.format(user, node._id))
logger.warn('Reason: {}'.format(str(err)))
@contributor_added.connect
def subscribe_contributor(node, contributor, auth=None, *args, **kwargs):
try:
subscribe_user_to_notifications(node, contributor)
except InvalidSubscriptionError as err:
logger.warn('Skipping subscription of user {} to node {}'.format(contributor, node._id))
logger.warn('Reason: {}'.format(str(err)))
@user_confirmed.connect
def subscribe_confirmed_user(user):
try:
subscribe_user_to_global_notifications(user)
except InvalidSubscriptionError as err:
logger.warn('Skipping subscription of user {} to global subscriptions'.format(user))
logger.warn('Reason: {}'.format(str(err)))
| Revert "Remove incorrect check for institution_id" | ## Code Before:
import logging
from website.notifications.exceptions import InvalidSubscriptionError
from website.notifications.utils import subscribe_user_to_notifications, subscribe_user_to_global_notifications
from website.project.signals import contributor_added, project_created
from framework.auth.signals import user_confirmed
logger = logging.getLogger(__name__)
@project_created.connect
def subscribe_creator(node):
if node.is_collection or node.is_deleted:
return None
try:
subscribe_user_to_notifications(node, node.creator)
except InvalidSubscriptionError as err:
user = node.creator._id if node.creator else 'None'
logger.warn('Skipping subscription of user {} to node {}'.format(user, node._id))
logger.warn('Reason: {}'.format(str(err)))
@contributor_added.connect
def subscribe_contributor(node, contributor, auth=None, *args, **kwargs):
try:
subscribe_user_to_notifications(node, contributor)
except InvalidSubscriptionError as err:
logger.warn('Skipping subscription of user {} to node {}'.format(contributor, node._id))
logger.warn('Reason: {}'.format(str(err)))
@user_confirmed.connect
def subscribe_confirmed_user(user):
try:
subscribe_user_to_global_notifications(user)
except InvalidSubscriptionError as err:
logger.warn('Skipping subscription of user {} to global subscriptions'.format(user))
logger.warn('Reason: {}'.format(str(err)))
## Instruction:
Revert "Remove incorrect check for institution_id"
## Code After:
import logging
from website.notifications.exceptions import InvalidSubscriptionError
from website.notifications.utils import subscribe_user_to_notifications, subscribe_user_to_global_notifications
from website.project.signals import contributor_added, project_created
from framework.auth.signals import user_confirmed
logger = logging.getLogger(__name__)
@project_created.connect
def subscribe_creator(node):
if node.institution_id or node.is_collection or node.is_deleted:
return None
try:
subscribe_user_to_notifications(node, node.creator)
except InvalidSubscriptionError as err:
user = node.creator._id if node.creator else 'None'
logger.warn('Skipping subscription of user {} to node {}'.format(user, node._id))
logger.warn('Reason: {}'.format(str(err)))
@contributor_added.connect
def subscribe_contributor(node, contributor, auth=None, *args, **kwargs):
try:
subscribe_user_to_notifications(node, contributor)
except InvalidSubscriptionError as err:
logger.warn('Skipping subscription of user {} to node {}'.format(contributor, node._id))
logger.warn('Reason: {}'.format(str(err)))
@user_confirmed.connect
def subscribe_confirmed_user(user):
try:
subscribe_user_to_global_notifications(user)
except InvalidSubscriptionError as err:
logger.warn('Skipping subscription of user {} to global subscriptions'.format(user))
logger.warn('Reason: {}'.format(str(err)))
| // ... existing code ...
def subscribe_creator(node):
if node.institution_id or node.is_collection or node.is_deleted:
return None
// ... rest of the code ... |
85814828d2caedd8612db6ce0ecec92025a34330 | tests/test_main.py | tests/test_main.py | from cookiecutter.main import is_repo_url
def test_is_repo_url():
"""Verify is_repo_url works."""
assert is_repo_url('gitolite@server:team/repo') is True
assert is_repo_url('[email protected]:audreyr/cookiecutter.git') is True
assert is_repo_url('https://github.com/audreyr/cookiecutter.git') is True
assert is_repo_url('gh:audreyr/cookiecutter-pypackage') is True
assert is_repo_url('/audreyr/cookiecutter.git') is False
assert is_repo_url('/home/audreyr/cookiecutter') is False
| from cookiecutter.main import is_repo_url
def test_is_repo_url():
"""Verify is_repo_url works."""
assert is_repo_url('gitolite@server:team/repo') is True
assert is_repo_url('[email protected]:audreyr/cookiecutter.git') is True
assert is_repo_url('https://github.com/audreyr/cookiecutter.git') is True
assert is_repo_url('gh:audreyr/cookiecutter-pypackage') is True
assert is_repo_url('https://bitbucket.org/pokoli/cookiecutter.hg') is True
assert is_repo_url('/audreyr/cookiecutter.git') is False
assert is_repo_url('/home/audreyr/cookiecutter') is False
| Add test for bitbucket domain | Add test for bitbucket domain
| Python | bsd-3-clause | michaeljoseph/cookiecutter,Springerle/cookiecutter,Springerle/cookiecutter,venumech/cookiecutter,cguardia/cookiecutter,luzfcb/cookiecutter,pjbull/cookiecutter,agconti/cookiecutter,willingc/cookiecutter,audreyr/cookiecutter,audreyr/cookiecutter,venumech/cookiecutter,takeflight/cookiecutter,dajose/cookiecutter,takeflight/cookiecutter,agconti/cookiecutter,terryjbates/cookiecutter,cguardia/cookiecutter,stevepiercy/cookiecutter,pjbull/cookiecutter,hackebrot/cookiecutter,luzfcb/cookiecutter,christabor/cookiecutter,ramiroluz/cookiecutter,hackebrot/cookiecutter,christabor/cookiecutter,benthomasson/cookiecutter,benthomasson/cookiecutter,michaeljoseph/cookiecutter,ramiroluz/cookiecutter,moi65/cookiecutter,stevepiercy/cookiecutter,dajose/cookiecutter,terryjbates/cookiecutter,moi65/cookiecutter,willingc/cookiecutter | from cookiecutter.main import is_repo_url
def test_is_repo_url():
"""Verify is_repo_url works."""
assert is_repo_url('gitolite@server:team/repo') is True
assert is_repo_url('[email protected]:audreyr/cookiecutter.git') is True
assert is_repo_url('https://github.com/audreyr/cookiecutter.git') is True
assert is_repo_url('gh:audreyr/cookiecutter-pypackage') is True
+ assert is_repo_url('https://bitbucket.org/pokoli/cookiecutter.hg') is True
assert is_repo_url('/audreyr/cookiecutter.git') is False
assert is_repo_url('/home/audreyr/cookiecutter') is False
| Add test for bitbucket domain | ## Code Before:
from cookiecutter.main import is_repo_url
def test_is_repo_url():
"""Verify is_repo_url works."""
assert is_repo_url('gitolite@server:team/repo') is True
assert is_repo_url('[email protected]:audreyr/cookiecutter.git') is True
assert is_repo_url('https://github.com/audreyr/cookiecutter.git') is True
assert is_repo_url('gh:audreyr/cookiecutter-pypackage') is True
assert is_repo_url('/audreyr/cookiecutter.git') is False
assert is_repo_url('/home/audreyr/cookiecutter') is False
## Instruction:
Add test for bitbucket domain
## Code After:
from cookiecutter.main import is_repo_url
def test_is_repo_url():
"""Verify is_repo_url works."""
assert is_repo_url('gitolite@server:team/repo') is True
assert is_repo_url('[email protected]:audreyr/cookiecutter.git') is True
assert is_repo_url('https://github.com/audreyr/cookiecutter.git') is True
assert is_repo_url('gh:audreyr/cookiecutter-pypackage') is True
assert is_repo_url('https://bitbucket.org/pokoli/cookiecutter.hg') is True
assert is_repo_url('/audreyr/cookiecutter.git') is False
assert is_repo_url('/home/audreyr/cookiecutter') is False
| // ... existing code ...
assert is_repo_url('gh:audreyr/cookiecutter-pypackage') is True
assert is_repo_url('https://bitbucket.org/pokoli/cookiecutter.hg') is True
// ... rest of the code ... |
461019099c41ca4ef2fc7ccfec0141ed5b7e3bd6 | tests/test_unicode.py | tests/test_unicode.py | import sys
import pytest
import jupytext
from .utils import list_all_notebooks
@pytest.mark.parametrize('nb_file', list_all_notebooks('.ipynb') +
list_all_notebooks('.Rmd'))
def test_notebook_contents_is_unicode(nb_file):
nb = jupytext.readf(nb_file)
for cell in nb.cells:
if sys.version_info < (3, 0):
assert cell.source == '' or isinstance(cell.source, unicode)
else:
assert isinstance(cell.source, str)
def test_write_non_ascii(tmpdir):
nb = jupytext.reads(u'Non-ascii contênt', ext='.Rmd')
jupytext.writef(nb, str(tmpdir.join('notebook.Rmd')))
jupytext.writef(nb, str(tmpdir.join('notebook.ipynb')))
| import sys
import pytest
import jupytext
from .utils import list_all_notebooks
try:
unicode # Python 2
except NameError:
unicode = str # Python 3
@pytest.mark.parametrize('nb_file', list_all_notebooks('.ipynb') +
list_all_notebooks('.Rmd'))
def test_notebook_contents_is_unicode(nb_file):
nb = jupytext.readf(nb_file)
for cell in nb.cells:
if sys.version_info < (3, 0):
assert cell.source == '' or isinstance(cell.source, unicode)
else:
assert isinstance(cell.source, str)
def test_write_non_ascii(tmpdir):
nb = jupytext.reads(u'Non-ascii contênt', ext='.Rmd')
jupytext.writef(nb, str(tmpdir.join('notebook.Rmd')))
jupytext.writef(nb, str(tmpdir.join('notebook.ipynb')))
| Define unicode in Python 3 | Define unicode in Python 3
__unicode__ was removed in Python 3 because all __str__ are Unicode.
[flake8](http://flake8.pycqa.org) testing of https://github.com/mwouts/jupytext on Python 3.7.0
$ __flake8 . --count --select=E901,E999,F821,F822,F823 --show-source --statistics__
```
./.jupyter/jupyter_notebook_config.py:1:1: F821 undefined name 'c'
c.NotebookApp.contents_manager_class = 'jupytext.TextFileContentsManager'
^
./tests/test_unicode.py:15:65: F821 undefined name 'unicode'
assert cell.source == '' or isinstance(cell.source, unicode)
^
./tests/mirror/jupyter_again.py:32:1: E999 SyntaxError: invalid syntax
?next
^
1 E999 SyntaxError: invalid syntax
2 F821 undefined name 'c'
3
``` | Python | mit | mwouts/jupytext,mwouts/jupytext,mwouts/jupytext,mwouts/jupytext,mwouts/jupytext,mwouts/jupytext,mwouts/jupytext,mwouts/jupytext,mwouts/jupytext,mwouts/jupytext | import sys
import pytest
import jupytext
from .utils import list_all_notebooks
+
+ try:
+ unicode # Python 2
+ except NameError:
+ unicode = str # Python 3
@pytest.mark.parametrize('nb_file', list_all_notebooks('.ipynb') +
list_all_notebooks('.Rmd'))
def test_notebook_contents_is_unicode(nb_file):
nb = jupytext.readf(nb_file)
for cell in nb.cells:
if sys.version_info < (3, 0):
assert cell.source == '' or isinstance(cell.source, unicode)
else:
assert isinstance(cell.source, str)
def test_write_non_ascii(tmpdir):
nb = jupytext.reads(u'Non-ascii contênt', ext='.Rmd')
jupytext.writef(nb, str(tmpdir.join('notebook.Rmd')))
jupytext.writef(nb, str(tmpdir.join('notebook.ipynb')))
| Define unicode in Python 3 | ## Code Before:
import sys
import pytest
import jupytext
from .utils import list_all_notebooks
@pytest.mark.parametrize('nb_file', list_all_notebooks('.ipynb') +
list_all_notebooks('.Rmd'))
def test_notebook_contents_is_unicode(nb_file):
nb = jupytext.readf(nb_file)
for cell in nb.cells:
if sys.version_info < (3, 0):
assert cell.source == '' or isinstance(cell.source, unicode)
else:
assert isinstance(cell.source, str)
def test_write_non_ascii(tmpdir):
nb = jupytext.reads(u'Non-ascii contênt', ext='.Rmd')
jupytext.writef(nb, str(tmpdir.join('notebook.Rmd')))
jupytext.writef(nb, str(tmpdir.join('notebook.ipynb')))
## Instruction:
Define unicode in Python 3
## Code After:
import sys
import pytest
import jupytext
from .utils import list_all_notebooks
try:
unicode # Python 2
except NameError:
unicode = str # Python 3
@pytest.mark.parametrize('nb_file', list_all_notebooks('.ipynb') +
list_all_notebooks('.Rmd'))
def test_notebook_contents_is_unicode(nb_file):
nb = jupytext.readf(nb_file)
for cell in nb.cells:
if sys.version_info < (3, 0):
assert cell.source == '' or isinstance(cell.source, unicode)
else:
assert isinstance(cell.source, str)
def test_write_non_ascii(tmpdir):
nb = jupytext.reads(u'Non-ascii contênt', ext='.Rmd')
jupytext.writef(nb, str(tmpdir.join('notebook.Rmd')))
jupytext.writef(nb, str(tmpdir.join('notebook.ipynb')))
| // ... existing code ...
from .utils import list_all_notebooks
try:
unicode # Python 2
except NameError:
unicode = str # Python 3
// ... rest of the code ... |
3290d532f3dd9c1e24921c4b80aeb6e860bc86a8 | spock/plugins/__init__.py | spock/plugins/__init__.py | from spock.plugins.core import auth, event, net, ticker, timer
from spock.plugins.helpers import clientinfo, entities, interact, inventory,\
keepalive, movement, physics, respawn, start, world
from spock.plugins.base import PluginBase # noqa
core_plugins = [
('auth', auth.AuthPlugin),
('event', event.EventPlugin),
('net', net.NetPlugin),
('ticker', ticker.TickerPlugin),
('timers', timer.TimerPlugin),
]
helper_plugins = [
('clientinfo', clientinfo.ClientInfoPlugin),
('entities', entities.EntitiesPlugin),
('interact', interact.InteractPlugin),
('inventory', inventory.InventoryPlugin),
('keepalive', keepalive.KeepalivePlugin),
('movement', movement.MovementPlugin),
('physics', physics.PhysicsPlugin),
('respawn', respawn.RespawnPlugin),
('start', start.StartPlugin),
('world', world.WorldPlugin),
]
default_plugins = core_plugins + helper_plugins
| from spock.plugins.core import auth, event, net, ticker, timer
from spock.plugins.helpers import chat, clientinfo, entities, interact, \
inventory, keepalive, movement, physics, respawn, start, world
from spock.plugins.base import PluginBase # noqa
core_plugins = [
('auth', auth.AuthPlugin),
('event', event.EventPlugin),
('net', net.NetPlugin),
('ticker', ticker.TickerPlugin),
('timers', timer.TimerPlugin),
]
helper_plugins = [
('chat', chat.ChatPlugin),
('clientinfo', clientinfo.ClientInfoPlugin),
('entities', entities.EntitiesPlugin),
('interact', interact.InteractPlugin),
('inventory', inventory.InventoryPlugin),
('keepalive', keepalive.KeepalivePlugin),
('movement', movement.MovementPlugin),
('physics', physics.PhysicsPlugin),
('respawn', respawn.RespawnPlugin),
('start', start.StartPlugin),
('world', world.WorldPlugin),
]
default_plugins = core_plugins + helper_plugins
| Fix removal of chat plugin | Fix removal of chat plugin
| Python | mit | nickelpro/SpockBot,MrSwiss/SpockBot,Gjum/SpockBot,SpockBotMC/SpockBot,gamingrobot/SpockBot,luken/SpockBot | from spock.plugins.core import auth, event, net, ticker, timer
- from spock.plugins.helpers import clientinfo, entities, interact, inventory,\
+ from spock.plugins.helpers import chat, clientinfo, entities, interact, \
- keepalive, movement, physics, respawn, start, world
+ inventory, keepalive, movement, physics, respawn, start, world
-
from spock.plugins.base import PluginBase # noqa
core_plugins = [
('auth', auth.AuthPlugin),
('event', event.EventPlugin),
('net', net.NetPlugin),
('ticker', ticker.TickerPlugin),
('timers', timer.TimerPlugin),
]
helper_plugins = [
+ ('chat', chat.ChatPlugin),
('clientinfo', clientinfo.ClientInfoPlugin),
('entities', entities.EntitiesPlugin),
('interact', interact.InteractPlugin),
('inventory', inventory.InventoryPlugin),
('keepalive', keepalive.KeepalivePlugin),
('movement', movement.MovementPlugin),
('physics', physics.PhysicsPlugin),
('respawn', respawn.RespawnPlugin),
('start', start.StartPlugin),
('world', world.WorldPlugin),
]
default_plugins = core_plugins + helper_plugins
| Fix removal of chat plugin | ## Code Before:
from spock.plugins.core import auth, event, net, ticker, timer
from spock.plugins.helpers import clientinfo, entities, interact, inventory,\
keepalive, movement, physics, respawn, start, world
from spock.plugins.base import PluginBase # noqa
core_plugins = [
('auth', auth.AuthPlugin),
('event', event.EventPlugin),
('net', net.NetPlugin),
('ticker', ticker.TickerPlugin),
('timers', timer.TimerPlugin),
]
helper_plugins = [
('clientinfo', clientinfo.ClientInfoPlugin),
('entities', entities.EntitiesPlugin),
('interact', interact.InteractPlugin),
('inventory', inventory.InventoryPlugin),
('keepalive', keepalive.KeepalivePlugin),
('movement', movement.MovementPlugin),
('physics', physics.PhysicsPlugin),
('respawn', respawn.RespawnPlugin),
('start', start.StartPlugin),
('world', world.WorldPlugin),
]
default_plugins = core_plugins + helper_plugins
## Instruction:
Fix removal of chat plugin
## Code After:
from spock.plugins.core import auth, event, net, ticker, timer
from spock.plugins.helpers import chat, clientinfo, entities, interact, \
inventory, keepalive, movement, physics, respawn, start, world
from spock.plugins.base import PluginBase # noqa
core_plugins = [
('auth', auth.AuthPlugin),
('event', event.EventPlugin),
('net', net.NetPlugin),
('ticker', ticker.TickerPlugin),
('timers', timer.TimerPlugin),
]
helper_plugins = [
('chat', chat.ChatPlugin),
('clientinfo', clientinfo.ClientInfoPlugin),
('entities', entities.EntitiesPlugin),
('interact', interact.InteractPlugin),
('inventory', inventory.InventoryPlugin),
('keepalive', keepalive.KeepalivePlugin),
('movement', movement.MovementPlugin),
('physics', physics.PhysicsPlugin),
('respawn', respawn.RespawnPlugin),
('start', start.StartPlugin),
('world', world.WorldPlugin),
]
default_plugins = core_plugins + helper_plugins
| ...
from spock.plugins.core import auth, event, net, ticker, timer
from spock.plugins.helpers import chat, clientinfo, entities, interact, \
inventory, keepalive, movement, physics, respawn, start, world
from spock.plugins.base import PluginBase # noqa
...
helper_plugins = [
('chat', chat.ChatPlugin),
('clientinfo', clientinfo.ClientInfoPlugin),
... |
d5a3285b05d96ffc99049867256cdba87a5b420a | packages/mono_crypto.py | packages/mono_crypto.py | from mono_master import MonoMasterPackage
from bockbuild.util.util import *
class MonoMasterEncryptedPackage (MonoMasterPackage):
def __init__(self):
MonoMasterPackage.__init__ (self)
self.configure_flags.extend(['--enable-extension-module=crypto --enable-native-types'])
def prep(self):
MonoMasterPackage.prep(self)
retry (self.checkout_mono_extensions)
def checkout_mono_extensions(self):
ext = '[email protected]:xamarin/mono-extensions.git'
dirname = os.path.join(self.profile.build_root, "mono-extensions")
if not os.path.exists(dirname):
self.sh('%' + '{git} clone --local --shared "%s" "%s"' % (ext, dirname))
self.pushd(dirname)
try:
self.sh('%{git} clean -xffd')
self.sh('%{git} fetch --all --prune')
if "pr/" not in self.git_branch:
self.sh('%' + '{git} checkout origin/%s' % self.git_branch)
else:
self.sh('%{git} checkout origin/master')
self.sh ('%{git} reset --hard')
except Exception as e:
self.popd ()
self.rm_if_exists (dirname)
raise
finally:
info ('Mono crypto extensions (rev. %s)' % git_get_revision (self))
self.popd ()
MonoMasterEncryptedPackage() | from mono_master import MonoMasterPackage
from bockbuild.util.util import *
class MonoMasterEncryptedPackage (MonoMasterPackage):
def __init__(self):
MonoMasterPackage.__init__ (self)
self.configure_flags.extend(['--enable-extension-module=crypto --enable-native-types'])
def prep(self):
MonoMasterPackage.prep(self)
retry (self.checkout_mono_extensions)
def checkout_mono_extensions(self):
ext = '[email protected]:xamarin/mono-extensions.git'
dirname = os.path.join(self.profile.build_root, "mono-extensions")
if not os.path.exists(dirname):
self.sh('%' + '{git} clone --local --shared "%s" "%s"' % (ext, dirname))
self.pushd(dirname)
try:
self.sh('%{git} clean -xffd')
self.sh('%{git} fetch --all --prune')
if 'pull/' in self.git_branch: # pull request
self.sh('%{git} checkout origin/master')
else:
self.sh('%' + '{git} checkout origin/%s' % self.git_branch)
self.sh ('%{git} reset --hard')
except Exception as e:
self.popd ()
self.rm_if_exists (dirname)
raise
finally:
info ('Mono crypto extensions (rev. %s)' % git_get_revision (self))
self.popd ()
MonoMasterEncryptedPackage() | Fix mono-extensions checkout for PR branches ('origin/pull/N/merge') | Fix mono-extensions checkout for PR branches ('origin/pull/N/merge')
| Python | mit | mono/bockbuild,mono/bockbuild | from mono_master import MonoMasterPackage
from bockbuild.util.util import *
class MonoMasterEncryptedPackage (MonoMasterPackage):
def __init__(self):
MonoMasterPackage.__init__ (self)
self.configure_flags.extend(['--enable-extension-module=crypto --enable-native-types'])
def prep(self):
MonoMasterPackage.prep(self)
retry (self.checkout_mono_extensions)
def checkout_mono_extensions(self):
ext = '[email protected]:xamarin/mono-extensions.git'
dirname = os.path.join(self.profile.build_root, "mono-extensions")
if not os.path.exists(dirname):
self.sh('%' + '{git} clone --local --shared "%s" "%s"' % (ext, dirname))
self.pushd(dirname)
try:
self.sh('%{git} clean -xffd')
self.sh('%{git} fetch --all --prune')
- if "pr/" not in self.git_branch:
+ if 'pull/' in self.git_branch: # pull request
+ self.sh('%{git} checkout origin/master')
+ else:
self.sh('%' + '{git} checkout origin/%s' % self.git_branch)
+
- else:
- self.sh('%{git} checkout origin/master')
self.sh ('%{git} reset --hard')
except Exception as e:
self.popd ()
self.rm_if_exists (dirname)
raise
finally:
info ('Mono crypto extensions (rev. %s)' % git_get_revision (self))
self.popd ()
MonoMasterEncryptedPackage() | Fix mono-extensions checkout for PR branches ('origin/pull/N/merge') | ## Code Before:
from mono_master import MonoMasterPackage
from bockbuild.util.util import *
class MonoMasterEncryptedPackage (MonoMasterPackage):
def __init__(self):
MonoMasterPackage.__init__ (self)
self.configure_flags.extend(['--enable-extension-module=crypto --enable-native-types'])
def prep(self):
MonoMasterPackage.prep(self)
retry (self.checkout_mono_extensions)
def checkout_mono_extensions(self):
ext = '[email protected]:xamarin/mono-extensions.git'
dirname = os.path.join(self.profile.build_root, "mono-extensions")
if not os.path.exists(dirname):
self.sh('%' + '{git} clone --local --shared "%s" "%s"' % (ext, dirname))
self.pushd(dirname)
try:
self.sh('%{git} clean -xffd')
self.sh('%{git} fetch --all --prune')
if "pr/" not in self.git_branch:
self.sh('%' + '{git} checkout origin/%s' % self.git_branch)
else:
self.sh('%{git} checkout origin/master')
self.sh ('%{git} reset --hard')
except Exception as e:
self.popd ()
self.rm_if_exists (dirname)
raise
finally:
info ('Mono crypto extensions (rev. %s)' % git_get_revision (self))
self.popd ()
MonoMasterEncryptedPackage()
## Instruction:
Fix mono-extensions checkout for PR branches ('origin/pull/N/merge')
## Code After:
from mono_master import MonoMasterPackage
from bockbuild.util.util import *
class MonoMasterEncryptedPackage (MonoMasterPackage):
def __init__(self):
MonoMasterPackage.__init__ (self)
self.configure_flags.extend(['--enable-extension-module=crypto --enable-native-types'])
def prep(self):
MonoMasterPackage.prep(self)
retry (self.checkout_mono_extensions)
def checkout_mono_extensions(self):
ext = '[email protected]:xamarin/mono-extensions.git'
dirname = os.path.join(self.profile.build_root, "mono-extensions")
if not os.path.exists(dirname):
self.sh('%' + '{git} clone --local --shared "%s" "%s"' % (ext, dirname))
self.pushd(dirname)
try:
self.sh('%{git} clean -xffd')
self.sh('%{git} fetch --all --prune')
if 'pull/' in self.git_branch: # pull request
self.sh('%{git} checkout origin/master')
else:
self.sh('%' + '{git} checkout origin/%s' % self.git_branch)
self.sh ('%{git} reset --hard')
except Exception as e:
self.popd ()
self.rm_if_exists (dirname)
raise
finally:
info ('Mono crypto extensions (rev. %s)' % git_get_revision (self))
self.popd ()
MonoMasterEncryptedPackage() | ...
self.sh('%{git} fetch --all --prune')
if 'pull/' in self.git_branch: # pull request
self.sh('%{git} checkout origin/master')
else:
self.sh('%' + '{git} checkout origin/%s' % self.git_branch)
self.sh ('%{git} reset --hard')
... |
a92121cfdbb94d36d021fb8d1386031829ee86a2 | patterns/solid.py | patterns/solid.py | import blinkypattern
class Solid(blinkypattern.BlinkyPattern):
def __init__(self, blinkytape, solid_color):
super(Solid, self).__init__(blinkytape)
self._pixels = [solid_color] * self._blinkytape.pixel_count
def setup(self):
super(Solid, self).setup()
self._blinkytape.set_pixels(self._pixels)
self._blinkytape.update()
| class Solid(object):
def __init__(self, pixel_count, color):
self._pixels = [color] * pixel_count
@property
def pixels(self):
return self._pixels
| Update Solid pattern for refactor | Update Solid pattern for refactor
| Python | mit | jonspeicher/blinkyfun | - import blinkypattern
+ class Solid(object):
+ def __init__(self, pixel_count, color):
+ self._pixels = [color] * pixel_count
+ @property
+ def pixels(self):
+ return self._pixels
- class Solid(blinkypattern.BlinkyPattern):
- def __init__(self, blinkytape, solid_color):
- super(Solid, self).__init__(blinkytape)
- self._pixels = [solid_color] * self._blinkytape.pixel_count
- def setup(self):
- super(Solid, self).setup()
- self._blinkytape.set_pixels(self._pixels)
- self._blinkytape.update()
- | Update Solid pattern for refactor | ## Code Before:
import blinkypattern
class Solid(blinkypattern.BlinkyPattern):
def __init__(self, blinkytape, solid_color):
super(Solid, self).__init__(blinkytape)
self._pixels = [solid_color] * self._blinkytape.pixel_count
def setup(self):
super(Solid, self).setup()
self._blinkytape.set_pixels(self._pixels)
self._blinkytape.update()
## Instruction:
Update Solid pattern for refactor
## Code After:
class Solid(object):
def __init__(self, pixel_count, color):
self._pixels = [color] * pixel_count
@property
def pixels(self):
return self._pixels
| // ... existing code ...
class Solid(object):
def __init__(self, pixel_count, color):
self._pixels = [color] * pixel_count
@property
def pixels(self):
return self._pixels
// ... rest of the code ... |
b39518482da1d3e064cdbc34490e4a9924f6d5f1 | quantecon/tests/test_ecdf.py | quantecon/tests/test_ecdf.py | import unittest
import numpy as np
from quantecon import ECDF
class TestECDF(unittest.TestCase):
@classmethod
def setUpClass(cls):
cls.obs = np.random.rand(40) # observations defining dist
cls.ecdf = ECDF(cls.obs)
def test_call_high(self):
"ecdf: x above all obs give 1.0"
# all of self.obs <= 1 so ecdf(1.1) should be 1
self.assertAlmostEqual(self.ecdf(1.1), 1.0)
def test_call_low(self):
"ecdf: x below all obs give 0.0"
# all of self.obs <= 1 so ecdf(1.1) should be 1
self.assertAlmostEqual(self.ecdf(-0.1), 0.0)
def test_ascending(self):
"ecdf: larger values should return F(x) at least as big"
x = np.random.rand()
F_1 = self.ecdf(x)
F_2 = self.ecdf(1.1 * x)
self.assertGreaterEqual(F_2, F_1)
| import unittest
import numpy as np
from quantecon import ECDF
class TestECDF(unittest.TestCase):
@classmethod
def setUpClass(cls):
cls.obs = np.random.rand(40) # observations defining dist
cls.ecdf = ECDF(cls.obs)
def test_call_high(self):
"ecdf: x above all obs give 1.0"
# all of self.obs <= 1 so ecdf(1.1) should be 1
self.assertAlmostEqual(self.ecdf(1.1), 1.0)
def test_call_low(self):
"ecdf: x below all obs give 0.0"
# all of self.obs <= 1 so ecdf(1.1) should be 1
self.assertAlmostEqual(self.ecdf(-0.1), 0.0)
def test_ascending(self):
"ecdf: larger values should return F(x) at least as big"
x = np.random.rand()
F_1 = self.ecdf(x)
F_2 = self.ecdf(1.1 * x)
self.assertGreaterEqual(F_2, F_1)
def test_vectorized(self):
"ecdf: testing vectorized __call__ method"
t = np.linspace(-1, 1, 100)
self.assertEqual(t.shape, self.ecdf(t).shape)
t = np.linspace(-1, 1, 100).reshape(2, 2, 25)
self.assertEqual(t.shape, self.ecdf(t).shape)
| Add a test for vectorized call | TST: Add a test for vectorized call
| Python | bsd-3-clause | oyamad/QuantEcon.py,QuantEcon/QuantEcon.py,oyamad/QuantEcon.py,QuantEcon/QuantEcon.py | import unittest
import numpy as np
from quantecon import ECDF
class TestECDF(unittest.TestCase):
@classmethod
def setUpClass(cls):
cls.obs = np.random.rand(40) # observations defining dist
cls.ecdf = ECDF(cls.obs)
def test_call_high(self):
"ecdf: x above all obs give 1.0"
# all of self.obs <= 1 so ecdf(1.1) should be 1
self.assertAlmostEqual(self.ecdf(1.1), 1.0)
def test_call_low(self):
"ecdf: x below all obs give 0.0"
# all of self.obs <= 1 so ecdf(1.1) should be 1
self.assertAlmostEqual(self.ecdf(-0.1), 0.0)
def test_ascending(self):
"ecdf: larger values should return F(x) at least as big"
x = np.random.rand()
F_1 = self.ecdf(x)
F_2 = self.ecdf(1.1 * x)
self.assertGreaterEqual(F_2, F_1)
+ def test_vectorized(self):
+ "ecdf: testing vectorized __call__ method"
+ t = np.linspace(-1, 1, 100)
+ self.assertEqual(t.shape, self.ecdf(t).shape)
+ t = np.linspace(-1, 1, 100).reshape(2, 2, 25)
+ self.assertEqual(t.shape, self.ecdf(t).shape)
+ | Add a test for vectorized call | ## Code Before:
import unittest
import numpy as np
from quantecon import ECDF
class TestECDF(unittest.TestCase):
@classmethod
def setUpClass(cls):
cls.obs = np.random.rand(40) # observations defining dist
cls.ecdf = ECDF(cls.obs)
def test_call_high(self):
"ecdf: x above all obs give 1.0"
# all of self.obs <= 1 so ecdf(1.1) should be 1
self.assertAlmostEqual(self.ecdf(1.1), 1.0)
def test_call_low(self):
"ecdf: x below all obs give 0.0"
# all of self.obs <= 1 so ecdf(1.1) should be 1
self.assertAlmostEqual(self.ecdf(-0.1), 0.0)
def test_ascending(self):
"ecdf: larger values should return F(x) at least as big"
x = np.random.rand()
F_1 = self.ecdf(x)
F_2 = self.ecdf(1.1 * x)
self.assertGreaterEqual(F_2, F_1)
## Instruction:
Add a test for vectorized call
## Code After:
import unittest
import numpy as np
from quantecon import ECDF
class TestECDF(unittest.TestCase):
@classmethod
def setUpClass(cls):
cls.obs = np.random.rand(40) # observations defining dist
cls.ecdf = ECDF(cls.obs)
def test_call_high(self):
"ecdf: x above all obs give 1.0"
# all of self.obs <= 1 so ecdf(1.1) should be 1
self.assertAlmostEqual(self.ecdf(1.1), 1.0)
def test_call_low(self):
"ecdf: x below all obs give 0.0"
# all of self.obs <= 1 so ecdf(1.1) should be 1
self.assertAlmostEqual(self.ecdf(-0.1), 0.0)
def test_ascending(self):
"ecdf: larger values should return F(x) at least as big"
x = np.random.rand()
F_1 = self.ecdf(x)
F_2 = self.ecdf(1.1 * x)
self.assertGreaterEqual(F_2, F_1)
def test_vectorized(self):
"ecdf: testing vectorized __call__ method"
t = np.linspace(-1, 1, 100)
self.assertEqual(t.shape, self.ecdf(t).shape)
t = np.linspace(-1, 1, 100).reshape(2, 2, 25)
self.assertEqual(t.shape, self.ecdf(t).shape)
| // ... existing code ...
self.assertGreaterEqual(F_2, F_1)
def test_vectorized(self):
"ecdf: testing vectorized __call__ method"
t = np.linspace(-1, 1, 100)
self.assertEqual(t.shape, self.ecdf(t).shape)
t = np.linspace(-1, 1, 100).reshape(2, 2, 25)
self.assertEqual(t.shape, self.ecdf(t).shape)
// ... rest of the code ... |
2cdb6a5eeb1730627cea2a812d590efed82d03fb | acceptance_tests/test_course_learners.py | acceptance_tests/test_course_learners.py | from unittest import skipUnless
from bok_choy.web_app_test import WebAppTest
from acceptance_tests import ENABLE_LEARNER_ANALYTICS
from acceptance_tests.mixins import CoursePageTestsMixin
from acceptance_tests.pages import CourseLearnersPage
@skipUnless(ENABLE_LEARNER_ANALYTICS, 'Learner Analytics must be enabled to run CourseLearnersTests')
class CourseLearnersTests(CoursePageTestsMixin, WebAppTest):
def setUp(self):
super(CourseLearnersTests, self).setUp()
self.page = CourseLearnersPage(self.browser)
def _test_data_update_message(self):
# Don't test the update message for now, since it won't exist
# until the SPA adds it to the page in AN-6205.
pass
def _get_data_update_message(self):
# Don't test the update message for now, since it won't exist
# until the SPA adds it to the page in AN-6205.
return ''
| from unittest import skipUnless
from bok_choy.web_app_test import WebAppTest
from acceptance_tests import ENABLE_LEARNER_ANALYTICS
from acceptance_tests.mixins import CoursePageTestsMixin
from acceptance_tests.pages import CourseLearnersPage
@skipUnless(ENABLE_LEARNER_ANALYTICS, 'Learner Analytics must be enabled to run CourseLearnersTests')
class CourseLearnersTests(CoursePageTestsMixin, WebAppTest):
help_path = 'engagement/learners.html'
def setUp(self):
super(CourseLearnersTests, self).setUp()
self.page = CourseLearnersPage(self.browser)
def _test_data_update_message(self):
# Don't test the update message for now, since it won't exist
# until the SPA adds it to the page in AN-6205.
pass
def _get_data_update_message(self):
# Don't test the update message for now, since it won't exist
# until the SPA adds it to the page in AN-6205.
return ''
| Add test for learners help link | Add test for learners help link
| Python | agpl-3.0 | Stanford-Online/edx-analytics-dashboard,Stanford-Online/edx-analytics-dashboard,edx/edx-analytics-dashboard,Stanford-Online/edx-analytics-dashboard,edx/edx-analytics-dashboard,edx/edx-analytics-dashboard,Stanford-Online/edx-analytics-dashboard,edx/edx-analytics-dashboard | from unittest import skipUnless
from bok_choy.web_app_test import WebAppTest
from acceptance_tests import ENABLE_LEARNER_ANALYTICS
from acceptance_tests.mixins import CoursePageTestsMixin
from acceptance_tests.pages import CourseLearnersPage
@skipUnless(ENABLE_LEARNER_ANALYTICS, 'Learner Analytics must be enabled to run CourseLearnersTests')
class CourseLearnersTests(CoursePageTestsMixin, WebAppTest):
+ help_path = 'engagement/learners.html'
+
def setUp(self):
super(CourseLearnersTests, self).setUp()
self.page = CourseLearnersPage(self.browser)
def _test_data_update_message(self):
# Don't test the update message for now, since it won't exist
# until the SPA adds it to the page in AN-6205.
pass
def _get_data_update_message(self):
# Don't test the update message for now, since it won't exist
# until the SPA adds it to the page in AN-6205.
return ''
| Add test for learners help link | ## Code Before:
from unittest import skipUnless
from bok_choy.web_app_test import WebAppTest
from acceptance_tests import ENABLE_LEARNER_ANALYTICS
from acceptance_tests.mixins import CoursePageTestsMixin
from acceptance_tests.pages import CourseLearnersPage
@skipUnless(ENABLE_LEARNER_ANALYTICS, 'Learner Analytics must be enabled to run CourseLearnersTests')
class CourseLearnersTests(CoursePageTestsMixin, WebAppTest):
def setUp(self):
super(CourseLearnersTests, self).setUp()
self.page = CourseLearnersPage(self.browser)
def _test_data_update_message(self):
# Don't test the update message for now, since it won't exist
# until the SPA adds it to the page in AN-6205.
pass
def _get_data_update_message(self):
# Don't test the update message for now, since it won't exist
# until the SPA adds it to the page in AN-6205.
return ''
## Instruction:
Add test for learners help link
## Code After:
from unittest import skipUnless
from bok_choy.web_app_test import WebAppTest
from acceptance_tests import ENABLE_LEARNER_ANALYTICS
from acceptance_tests.mixins import CoursePageTestsMixin
from acceptance_tests.pages import CourseLearnersPage
@skipUnless(ENABLE_LEARNER_ANALYTICS, 'Learner Analytics must be enabled to run CourseLearnersTests')
class CourseLearnersTests(CoursePageTestsMixin, WebAppTest):
help_path = 'engagement/learners.html'
def setUp(self):
super(CourseLearnersTests, self).setUp()
self.page = CourseLearnersPage(self.browser)
def _test_data_update_message(self):
# Don't test the update message for now, since it won't exist
# until the SPA adds it to the page in AN-6205.
pass
def _get_data_update_message(self):
# Don't test the update message for now, since it won't exist
# until the SPA adds it to the page in AN-6205.
return ''
| // ... existing code ...
class CourseLearnersTests(CoursePageTestsMixin, WebAppTest):
help_path = 'engagement/learners.html'
def setUp(self):
// ... rest of the code ... |
3ecc978421e1bcceb30635e875333e52272e07a3 | tests/providers/test_ovh.py | tests/providers/test_ovh.py | from unittest import TestCase
from lexicon.providers.ovh import Provider
from lexicon.common.options_handler import env_auth_options
from integration_tests import IntegrationTests
# Hook into testing framework by inheriting unittest.TestCase and reuse
# the tests which *each and every* implementation of the interface must
# pass, by inheritance from integration_tests.IntegrationTests
class OvhProviderTests(TestCase, IntegrationTests):
Provider = Provider
provider_name = 'ovh'
domain = 'elogium.net'
def _filter_headers(self):
return ['X-Ovh-Application', 'X-Ovh-Consumer', 'X-Ovh-Signature']
# Override _test_options to call env_auth_options and then import auth config from env variables
def _test_options(self):
cmd_options = env_auth_options(self.provider_name)
cmd_options['domain'] = self.domain
return cmd_options
| from unittest import TestCase
from lexicon.providers.ovh import Provider
from lexicon.common.options_handler import env_auth_options
from integration_tests import IntegrationTests
# Hook into testing framework by inheriting unittest.TestCase and reuse
# the tests which *each and every* implementation of the interface must
# pass, by inheritance from integration_tests.IntegrationTests
class OvhProviderTests(TestCase, IntegrationTests):
Provider = Provider
provider_name = 'ovh'
domain = 'elogium.net'
def _filter_headers(self):
return ['X-Ovh-Application', 'X-Ovh-Consumer', 'X-Ovh-Signature']
# Override _test_options to call env_auth_options and then import auth config from env variables
def _test_options(self):
cmd_options = env_auth_options(self.provider_name)
cmd_options['auth_entrypoint'] = 'ovh-eu'
cmd_options['domain'] = self.domain
return cmd_options
| Select ovh-eu entrypoint for test integration | Select ovh-eu entrypoint for test integration
| Python | mit | tnwhitwell/lexicon,AnalogJ/lexicon,AnalogJ/lexicon,tnwhitwell/lexicon | from unittest import TestCase
from lexicon.providers.ovh import Provider
from lexicon.common.options_handler import env_auth_options
from integration_tests import IntegrationTests
# Hook into testing framework by inheriting unittest.TestCase and reuse
# the tests which *each and every* implementation of the interface must
# pass, by inheritance from integration_tests.IntegrationTests
class OvhProviderTests(TestCase, IntegrationTests):
Provider = Provider
provider_name = 'ovh'
domain = 'elogium.net'
def _filter_headers(self):
return ['X-Ovh-Application', 'X-Ovh-Consumer', 'X-Ovh-Signature']
# Override _test_options to call env_auth_options and then import auth config from env variables
def _test_options(self):
cmd_options = env_auth_options(self.provider_name)
+ cmd_options['auth_entrypoint'] = 'ovh-eu'
cmd_options['domain'] = self.domain
return cmd_options
| Select ovh-eu entrypoint for test integration | ## Code Before:
from unittest import TestCase
from lexicon.providers.ovh import Provider
from lexicon.common.options_handler import env_auth_options
from integration_tests import IntegrationTests
# Hook into testing framework by inheriting unittest.TestCase and reuse
# the tests which *each and every* implementation of the interface must
# pass, by inheritance from integration_tests.IntegrationTests
class OvhProviderTests(TestCase, IntegrationTests):
Provider = Provider
provider_name = 'ovh'
domain = 'elogium.net'
def _filter_headers(self):
return ['X-Ovh-Application', 'X-Ovh-Consumer', 'X-Ovh-Signature']
# Override _test_options to call env_auth_options and then import auth config from env variables
def _test_options(self):
cmd_options = env_auth_options(self.provider_name)
cmd_options['domain'] = self.domain
return cmd_options
## Instruction:
Select ovh-eu entrypoint for test integration
## Code After:
from unittest import TestCase
from lexicon.providers.ovh import Provider
from lexicon.common.options_handler import env_auth_options
from integration_tests import IntegrationTests
# Hook into testing framework by inheriting unittest.TestCase and reuse
# the tests which *each and every* implementation of the interface must
# pass, by inheritance from integration_tests.IntegrationTests
class OvhProviderTests(TestCase, IntegrationTests):
Provider = Provider
provider_name = 'ovh'
domain = 'elogium.net'
def _filter_headers(self):
return ['X-Ovh-Application', 'X-Ovh-Consumer', 'X-Ovh-Signature']
# Override _test_options to call env_auth_options and then import auth config from env variables
def _test_options(self):
cmd_options = env_auth_options(self.provider_name)
cmd_options['auth_entrypoint'] = 'ovh-eu'
cmd_options['domain'] = self.domain
return cmd_options
| # ... existing code ...
cmd_options = env_auth_options(self.provider_name)
cmd_options['auth_entrypoint'] = 'ovh-eu'
cmd_options['domain'] = self.domain
# ... rest of the code ... |
9121c8c074a31fd3668f8281c7f093360ed72988 | salad/cli.py | salad/cli.py | import sys
import argparse
from lettuce.bin import main as lettuce_main
from lettuce import world
from selenium.webdriver.common.desired_capabilities import DesiredCapabilities
BROWSER_CHOICES = [browser.lower()
for browser in DesiredCapabilities.__dict__.keys()
if not browser.startswith('_')]
BROWSER_CHOICES.append('zope.testbrowser')
BROWSER_CHOICES.sort()
DEFAULT_BROWSER = 'firefox'
def main(args=sys.argv[1:]):
parser = argparse.ArgumentParser(prog="Salad", description='BDD browswer-automation made tasty.')
parser.add_argument('--browser', default=DEFAULT_BROWSER,
metavar='BROWSER', choices=BROWSER_CHOICES,
help=('Browser to use. Options: %s Default is %s.' %
(BROWSER_CHOICES, DEFAULT_BROWSER)))
parser.add_argument('--remote-url',
help='Selenium server url for remote browsers')
parser.add_argument('args', nargs=argparse.REMAINDER)
parsed_args = parser.parse_args()
world.drivers = [parsed_args.browser]
world.remote_url = parsed_args.remote_url
lettuce_main(args=parsed_args.args)
if __name__ == '__main__':
main()
| import sys
import argparse
from lettuce.bin import main as lettuce_main
from lettuce import world
from selenium.webdriver.common.desired_capabilities import DesiredCapabilities
BROWSER_CHOICES = [browser.lower()
for browser in DesiredCapabilities.__dict__.keys()
if not browser.startswith('_')]
BROWSER_CHOICES.append('zope.testbrowser')
BROWSER_CHOICES.sort()
DEFAULT_BROWSER = 'firefox'
def main(args=sys.argv[1:]):
parser = argparse.ArgumentParser(prog="Salad", description='BDD browswer-automation made tasty.')
parser.add_argument('--browser', default=DEFAULT_BROWSER,
metavar='BROWSER', choices=BROWSER_CHOICES,
help=('Browser to use. Options: %s Default is %s.' %
(BROWSER_CHOICES, DEFAULT_BROWSER)))
parser.add_argument('--remote-url',
help='Selenium server url for remote browsers')
(parsed_args, leftovers) = parser.parse_known_args()
world.drivers = [parsed_args.browser]
world.remote_url = parsed_args.remote_url
lettuce_main(args=leftovers)
if __name__ == '__main__':
main()
| Use parse_known_args and pass leftovers to lettuce | Use parse_known_args and pass leftovers to lettuce
Seems to be more reliable at handling weird argument ordering than
REMAINDER was
| Python | bsd-3-clause | salad/salad,salad/salad,beanqueen/salad,beanqueen/salad | import sys
import argparse
from lettuce.bin import main as lettuce_main
from lettuce import world
from selenium.webdriver.common.desired_capabilities import DesiredCapabilities
BROWSER_CHOICES = [browser.lower()
for browser in DesiredCapabilities.__dict__.keys()
if not browser.startswith('_')]
BROWSER_CHOICES.append('zope.testbrowser')
BROWSER_CHOICES.sort()
DEFAULT_BROWSER = 'firefox'
def main(args=sys.argv[1:]):
parser = argparse.ArgumentParser(prog="Salad", description='BDD browswer-automation made tasty.')
parser.add_argument('--browser', default=DEFAULT_BROWSER,
metavar='BROWSER', choices=BROWSER_CHOICES,
help=('Browser to use. Options: %s Default is %s.' %
(BROWSER_CHOICES, DEFAULT_BROWSER)))
parser.add_argument('--remote-url',
help='Selenium server url for remote browsers')
- parser.add_argument('args', nargs=argparse.REMAINDER)
- parsed_args = parser.parse_args()
+ (parsed_args, leftovers) = parser.parse_known_args()
world.drivers = [parsed_args.browser]
world.remote_url = parsed_args.remote_url
- lettuce_main(args=parsed_args.args)
+ lettuce_main(args=leftovers)
if __name__ == '__main__':
main()
| Use parse_known_args and pass leftovers to lettuce | ## Code Before:
import sys
import argparse
from lettuce.bin import main as lettuce_main
from lettuce import world
from selenium.webdriver.common.desired_capabilities import DesiredCapabilities
BROWSER_CHOICES = [browser.lower()
for browser in DesiredCapabilities.__dict__.keys()
if not browser.startswith('_')]
BROWSER_CHOICES.append('zope.testbrowser')
BROWSER_CHOICES.sort()
DEFAULT_BROWSER = 'firefox'
def main(args=sys.argv[1:]):
parser = argparse.ArgumentParser(prog="Salad", description='BDD browswer-automation made tasty.')
parser.add_argument('--browser', default=DEFAULT_BROWSER,
metavar='BROWSER', choices=BROWSER_CHOICES,
help=('Browser to use. Options: %s Default is %s.' %
(BROWSER_CHOICES, DEFAULT_BROWSER)))
parser.add_argument('--remote-url',
help='Selenium server url for remote browsers')
parser.add_argument('args', nargs=argparse.REMAINDER)
parsed_args = parser.parse_args()
world.drivers = [parsed_args.browser]
world.remote_url = parsed_args.remote_url
lettuce_main(args=parsed_args.args)
if __name__ == '__main__':
main()
## Instruction:
Use parse_known_args and pass leftovers to lettuce
## Code After:
import sys
import argparse
from lettuce.bin import main as lettuce_main
from lettuce import world
from selenium.webdriver.common.desired_capabilities import DesiredCapabilities
BROWSER_CHOICES = [browser.lower()
for browser in DesiredCapabilities.__dict__.keys()
if not browser.startswith('_')]
BROWSER_CHOICES.append('zope.testbrowser')
BROWSER_CHOICES.sort()
DEFAULT_BROWSER = 'firefox'
def main(args=sys.argv[1:]):
parser = argparse.ArgumentParser(prog="Salad", description='BDD browswer-automation made tasty.')
parser.add_argument('--browser', default=DEFAULT_BROWSER,
metavar='BROWSER', choices=BROWSER_CHOICES,
help=('Browser to use. Options: %s Default is %s.' %
(BROWSER_CHOICES, DEFAULT_BROWSER)))
parser.add_argument('--remote-url',
help='Selenium server url for remote browsers')
(parsed_args, leftovers) = parser.parse_known_args()
world.drivers = [parsed_args.browser]
world.remote_url = parsed_args.remote_url
lettuce_main(args=leftovers)
if __name__ == '__main__':
main()
| # ... existing code ...
help='Selenium server url for remote browsers')
(parsed_args, leftovers) = parser.parse_known_args()
world.drivers = [parsed_args.browser]
# ... modified code ...
world.remote_url = parsed_args.remote_url
lettuce_main(args=leftovers)
# ... rest of the code ... |
fc4fa5d06ea0ca557d69112d1c8d0f10c8e594e0 | diet_gtfs.py | diet_gtfs.py | import csv
import sys
# agency.txt done
# feed_info.txt nothing to change
# calendar_dates.txt depends on service_id.
# routes.txt depends on agency.txt
# shapes.txt depends on trips.txt
# stops.txt depends on stop_times.txt
# stop_times.txt depends on trip_id.
# transfers.txt depends on stop_id from and to, routes.
# trips.txt contains shape_id, also route_id to trip_id.
def clean_agency_file(*agencies):
with open('agency.txt', 'r') as f:
reader = csv.reader(f)
next(f)
for row in reader:
if row[0] in agencies:
print(row)
def main():
agencies = sys.argv[1:]
clean_agency_file(*agencies)
if __name__ == '__main__':
main()
| import csv
import sys
# agency.txt done
# feed_info.txt nothing to change
# calendar_dates.txt depends on service_id.
# routes.txt depends on agency.txt
# shapes.txt depends on trips.txt
# stops.txt depends on stop_times.txt
# stop_times.txt depends on trip_id.
# transfers.txt depends on stop_id from and to, routes.
# trips.txt contains shape_id, also route_id to trip_id.
def clean_agency_file(*agencies):
with open('agency.txt', 'r') as f:
reader = csv.reader(f)
filtered_rows = []
filtered_rows.append(next(reader))
for row in reader:
if row[0] in agencies:
filtered_rows.append(row)
with open('cleaned/agency.txt', 'w') as f:
writer = csv.writer(f)
writer.writerows(filtered_rows)
def main():
agencies = sys.argv[1:]
clean_agency_file(*agencies)
if __name__ == '__main__':
main()
| Create a complete filtered output agency.txt | Create a complete filtered output agency.txt
Filter based on arguments passed from shell.
| Python | bsd-2-clause | sensiblecodeio/diet-gtfs | import csv
import sys
# agency.txt done
# feed_info.txt nothing to change
# calendar_dates.txt depends on service_id.
# routes.txt depends on agency.txt
# shapes.txt depends on trips.txt
# stops.txt depends on stop_times.txt
# stop_times.txt depends on trip_id.
# transfers.txt depends on stop_id from and to, routes.
# trips.txt contains shape_id, also route_id to trip_id.
+
def clean_agency_file(*agencies):
with open('agency.txt', 'r') as f:
reader = csv.reader(f)
- next(f)
+ filtered_rows = []
+ filtered_rows.append(next(reader))
+
for row in reader:
if row[0] in agencies:
- print(row)
+ filtered_rows.append(row)
+
+ with open('cleaned/agency.txt', 'w') as f:
+ writer = csv.writer(f)
+ writer.writerows(filtered_rows)
def main():
agencies = sys.argv[1:]
clean_agency_file(*agencies)
if __name__ == '__main__':
main()
| Create a complete filtered output agency.txt | ## Code Before:
import csv
import sys
# agency.txt done
# feed_info.txt nothing to change
# calendar_dates.txt depends on service_id.
# routes.txt depends on agency.txt
# shapes.txt depends on trips.txt
# stops.txt depends on stop_times.txt
# stop_times.txt depends on trip_id.
# transfers.txt depends on stop_id from and to, routes.
# trips.txt contains shape_id, also route_id to trip_id.
def clean_agency_file(*agencies):
with open('agency.txt', 'r') as f:
reader = csv.reader(f)
next(f)
for row in reader:
if row[0] in agencies:
print(row)
def main():
agencies = sys.argv[1:]
clean_agency_file(*agencies)
if __name__ == '__main__':
main()
## Instruction:
Create a complete filtered output agency.txt
## Code After:
import csv
import sys
# agency.txt done
# feed_info.txt nothing to change
# calendar_dates.txt depends on service_id.
# routes.txt depends on agency.txt
# shapes.txt depends on trips.txt
# stops.txt depends on stop_times.txt
# stop_times.txt depends on trip_id.
# transfers.txt depends on stop_id from and to, routes.
# trips.txt contains shape_id, also route_id to trip_id.
def clean_agency_file(*agencies):
with open('agency.txt', 'r') as f:
reader = csv.reader(f)
filtered_rows = []
filtered_rows.append(next(reader))
for row in reader:
if row[0] in agencies:
filtered_rows.append(row)
with open('cleaned/agency.txt', 'w') as f:
writer = csv.writer(f)
writer.writerows(filtered_rows)
def main():
agencies = sys.argv[1:]
clean_agency_file(*agencies)
if __name__ == '__main__':
main()
| // ... existing code ...
def clean_agency_file(*agencies):
// ... modified code ...
reader = csv.reader(f)
filtered_rows = []
filtered_rows.append(next(reader))
for row in reader:
...
if row[0] in agencies:
filtered_rows.append(row)
with open('cleaned/agency.txt', 'w') as f:
writer = csv.writer(f)
writer.writerows(filtered_rows)
// ... rest of the code ... |
ee24b8b57bc73947cd5140aca15389861b33ab00 | gui/qt.py | gui/qt.py | from lib.version import AMON_VERSION
from lib.keybase import KeybaseUser
from lib.gmail import GmailUser
from lib.addresses import AddressBook
import lib.gpg as gpg
import sys
import logging
import json
from PyQt4 import QtGui
class Amon(QtGui.QMainWindow):
def __init__(self):
super(Amon, self).__init__()
self.keybase_user = KeybaseUser()
self.gmail = GmailUser()
self.address_book = AddressBook()
| from lib.version import AMON_VERSION
from lib.keybase import KeybaseUser
from lib.gmail import GmailUser
from lib.addresses import AddressBook
import lib.gpg as gpg
import sys
import logging
import json
from PyQt4 import QtGui
class Amon(QtGui.QMainWindow):
def __init__(self):
super(Amon, self).__init__()
self.keybase_user = KeybaseUser()
self.gmail = GmailUser()
self.address_book = AddressBook()
self.initUI()
def initUI(self):
exitAction = QtGui.QAction(QtGui.QIcon('exit.png'), '&Exit', self)
exitAction.setShortcut('Ctrl+Q')
exitAction.setStatusTip('Exit application')
exitAction.triggered.connect(QtGui.qApp.quit)
menubar = self.menuBar()
menubar.setNativeMenuBar(False)
fileMenu = menubar.addMenu('&File')
fileMenu.addAction(exitAction)
self.statusBar().showMessage('Ready')
self.setGeometry(300, 300, 300, 200)
self.setWindowTitle('Amon ' + AMON_VERSION)
self.show()
def main():
app = QtGui.QApplication(sys.argv)
amon = Amon()
sys.exit(app.exec_())
if __name__ == '__main__':
main()
| Update Qt gui to have status bar and menu bar | Update Qt gui to have status bar and menu bar
| Python | unlicense | CodingAnarchy/Amon | from lib.version import AMON_VERSION
from lib.keybase import KeybaseUser
from lib.gmail import GmailUser
from lib.addresses import AddressBook
import lib.gpg as gpg
import sys
import logging
import json
from PyQt4 import QtGui
class Amon(QtGui.QMainWindow):
def __init__(self):
super(Amon, self).__init__()
self.keybase_user = KeybaseUser()
self.gmail = GmailUser()
self.address_book = AddressBook()
+ self.initUI()
+
+ def initUI(self):
+ exitAction = QtGui.QAction(QtGui.QIcon('exit.png'), '&Exit', self)
+ exitAction.setShortcut('Ctrl+Q')
+ exitAction.setStatusTip('Exit application')
+ exitAction.triggered.connect(QtGui.qApp.quit)
+
+ menubar = self.menuBar()
+ menubar.setNativeMenuBar(False)
+ fileMenu = menubar.addMenu('&File')
+ fileMenu.addAction(exitAction)
+
+ self.statusBar().showMessage('Ready')
+ self.setGeometry(300, 300, 300, 200)
+ self.setWindowTitle('Amon ' + AMON_VERSION)
+ self.show()
+
+
+ def main():
+ app = QtGui.QApplication(sys.argv)
+ amon = Amon()
+ sys.exit(app.exec_())
+
+ if __name__ == '__main__':
+ main()
+ | Update Qt gui to have status bar and menu bar | ## Code Before:
from lib.version import AMON_VERSION
from lib.keybase import KeybaseUser
from lib.gmail import GmailUser
from lib.addresses import AddressBook
import lib.gpg as gpg
import sys
import logging
import json
from PyQt4 import QtGui
class Amon(QtGui.QMainWindow):
def __init__(self):
super(Amon, self).__init__()
self.keybase_user = KeybaseUser()
self.gmail = GmailUser()
self.address_book = AddressBook()
## Instruction:
Update Qt gui to have status bar and menu bar
## Code After:
from lib.version import AMON_VERSION
from lib.keybase import KeybaseUser
from lib.gmail import GmailUser
from lib.addresses import AddressBook
import lib.gpg as gpg
import sys
import logging
import json
from PyQt4 import QtGui
class Amon(QtGui.QMainWindow):
def __init__(self):
super(Amon, self).__init__()
self.keybase_user = KeybaseUser()
self.gmail = GmailUser()
self.address_book = AddressBook()
self.initUI()
def initUI(self):
exitAction = QtGui.QAction(QtGui.QIcon('exit.png'), '&Exit', self)
exitAction.setShortcut('Ctrl+Q')
exitAction.setStatusTip('Exit application')
exitAction.triggered.connect(QtGui.qApp.quit)
menubar = self.menuBar()
menubar.setNativeMenuBar(False)
fileMenu = menubar.addMenu('&File')
fileMenu.addAction(exitAction)
self.statusBar().showMessage('Ready')
self.setGeometry(300, 300, 300, 200)
self.setWindowTitle('Amon ' + AMON_VERSION)
self.show()
def main():
app = QtGui.QApplication(sys.argv)
amon = Amon()
sys.exit(app.exec_())
if __name__ == '__main__':
main()
| # ... existing code ...
self.address_book = AddressBook()
self.initUI()
def initUI(self):
exitAction = QtGui.QAction(QtGui.QIcon('exit.png'), '&Exit', self)
exitAction.setShortcut('Ctrl+Q')
exitAction.setStatusTip('Exit application')
exitAction.triggered.connect(QtGui.qApp.quit)
menubar = self.menuBar()
menubar.setNativeMenuBar(False)
fileMenu = menubar.addMenu('&File')
fileMenu.addAction(exitAction)
self.statusBar().showMessage('Ready')
self.setGeometry(300, 300, 300, 200)
self.setWindowTitle('Amon ' + AMON_VERSION)
self.show()
def main():
app = QtGui.QApplication(sys.argv)
amon = Amon()
sys.exit(app.exec_())
if __name__ == '__main__':
main()
# ... rest of the code ... |
92d9e9885e241e0bb7df64d3cd696db09cdfc74d | utils.py | utils.py | def fix_str(value):
try:
return unicode(value)
except UnicodeDecodeError:
return unicode(value.decode('latin1'))
def pandas_to_dict(df):
return [{colname: (fix_str(row[i]) if type(row[i]) is str else row[i])
for i, colname in enumerate(df.columns)
if colname not in ['pedido_data', 'cliente_data']}
for row in df.values]
| def fix_str(value):
try:
return unicode(value)
except UnicodeDecodeError:
return unicode(value.decode('latin1'))
def pandas_to_dict(df):
return [{colname: (fix_str(row[i]) if type(row[i]) is str else row[i])
for i, colname in enumerate(df.columns)}
for row in df.values]
| Remove columns filter on pandas_to_dict | Remove columns filter on pandas_to_dict
| Python | mit | mlgruby/mining,mlgruby/mining,mining/mining,chrisdamba/mining,jgabriellima/mining,chrisdamba/mining,AndrzejR/mining,seagoat/mining,jgabriellima/mining,mining/mining,avelino/mining,AndrzejR/mining,seagoat/mining,mlgruby/mining,avelino/mining | def fix_str(value):
try:
return unicode(value)
except UnicodeDecodeError:
return unicode(value.decode('latin1'))
def pandas_to_dict(df):
return [{colname: (fix_str(row[i]) if type(row[i]) is str else row[i])
- for i, colname in enumerate(df.columns)
+ for i, colname in enumerate(df.columns)}
- if colname not in ['pedido_data', 'cliente_data']}
for row in df.values]
| Remove columns filter on pandas_to_dict | ## Code Before:
def fix_str(value):
try:
return unicode(value)
except UnicodeDecodeError:
return unicode(value.decode('latin1'))
def pandas_to_dict(df):
return [{colname: (fix_str(row[i]) if type(row[i]) is str else row[i])
for i, colname in enumerate(df.columns)
if colname not in ['pedido_data', 'cliente_data']}
for row in df.values]
## Instruction:
Remove columns filter on pandas_to_dict
## Code After:
def fix_str(value):
try:
return unicode(value)
except UnicodeDecodeError:
return unicode(value.decode('latin1'))
def pandas_to_dict(df):
return [{colname: (fix_str(row[i]) if type(row[i]) is str else row[i])
for i, colname in enumerate(df.columns)}
for row in df.values]
| # ... existing code ...
return [{colname: (fix_str(row[i]) if type(row[i]) is str else row[i])
for i, colname in enumerate(df.columns)}
for row in df.values]
# ... rest of the code ... |
22f9fc8a56882f0595d051cb8c5d20fd97091e8c | custom/opm/tests/test_snapshot.py | custom/opm/tests/test_snapshot.py | from datetime import date
from unittest import TestCase
from couchforms.models import XFormInstance
from ..constants import *
from ..reports import get_report, BeneficiaryPaymentReport, MetReport
from .case_reports import Report, OPMCase, MockCaseRow, MockDataProvider
class TestGetReportUtil(TestCase):
def get_report_class(self, report_class):
obj_dict = {
'get_rows': lambda slf, datespan: [
OPMCase(
forms=[],
edd=date(2014, 11, 10),
),
OPMCase(
forms=[],
dod=date(2014, 1, 12),
),
OPMCase(
forms=[],
dod=date(2014, 3, 12),
),
],
'data_provider': MockDataProvider(),
}
return type(report_class.__name__, (Report, report_class), obj_dict)
def test_basic_BPR(self):
report_class = self.get_report_class(BeneficiaryPaymentReport)
report = get_report(report_class, month=6, year=2014, block="Atri")
report.rows
def test_basic_CMR(self):
report_class = self.get_report_class(MetReport)
report = get_report(report_class, month=6, year=2014, block="Atri")
report.rows
| from datetime import date
from unittest import TestCase
from mock import patch
from corehq.apps.users.models import CommCareUser
from couchforms.models import XFormInstance
from ..constants import *
from ..reports import get_report, BeneficiaryPaymentReport, MetReport
from .case_reports import Report, OPMCase, MockCaseRow, MockDataProvider
class TestGetReportUtil(TestCase):
def get_report_class(self, report_class):
obj_dict = {
'get_rows': lambda slf, datespan: [
OPMCase(
forms=[],
edd=date(2014, 11, 10),
),
OPMCase(
forms=[],
dod=date(2014, 1, 12),
),
OPMCase(
forms=[],
dod=date(2014, 3, 12),
),
],
'data_provider': MockDataProvider(),
}
return type(report_class.__name__, (Report, report_class), obj_dict)
def test_basic_BPR(self):
report_class = self.get_report_class(BeneficiaryPaymentReport)
report = get_report(report_class, month=6, year=2014, block="Atri")
report.rows
@patch.object(CommCareUser, 'by_domain', return_value=[])
def test_basic_CMR(self, user_mock):
report_class = self.get_report_class(MetReport)
report = get_report(report_class, month=6, year=2014, block="Atri")
report.rows
| Fix for test (add mock for CommCareUser) | Fix for test (add mock for CommCareUser)
| Python | bsd-3-clause | puttarajubr/commcare-hq,puttarajubr/commcare-hq,qedsoftware/commcare-hq,qedsoftware/commcare-hq,qedsoftware/commcare-hq,qedsoftware/commcare-hq,dimagi/commcare-hq,puttarajubr/commcare-hq,dimagi/commcare-hq,qedsoftware/commcare-hq,dimagi/commcare-hq,puttarajubr/commcare-hq,dimagi/commcare-hq,dimagi/commcare-hq | from datetime import date
from unittest import TestCase
-
+ from mock import patch
+ from corehq.apps.users.models import CommCareUser
from couchforms.models import XFormInstance
from ..constants import *
from ..reports import get_report, BeneficiaryPaymentReport, MetReport
from .case_reports import Report, OPMCase, MockCaseRow, MockDataProvider
class TestGetReportUtil(TestCase):
def get_report_class(self, report_class):
obj_dict = {
'get_rows': lambda slf, datespan: [
OPMCase(
forms=[],
edd=date(2014, 11, 10),
),
OPMCase(
forms=[],
dod=date(2014, 1, 12),
),
OPMCase(
forms=[],
dod=date(2014, 3, 12),
),
],
'data_provider': MockDataProvider(),
}
return type(report_class.__name__, (Report, report_class), obj_dict)
def test_basic_BPR(self):
report_class = self.get_report_class(BeneficiaryPaymentReport)
report = get_report(report_class, month=6, year=2014, block="Atri")
report.rows
+ @patch.object(CommCareUser, 'by_domain', return_value=[])
- def test_basic_CMR(self):
+ def test_basic_CMR(self, user_mock):
report_class = self.get_report_class(MetReport)
report = get_report(report_class, month=6, year=2014, block="Atri")
report.rows
| Fix for test (add mock for CommCareUser) | ## Code Before:
from datetime import date
from unittest import TestCase
from couchforms.models import XFormInstance
from ..constants import *
from ..reports import get_report, BeneficiaryPaymentReport, MetReport
from .case_reports import Report, OPMCase, MockCaseRow, MockDataProvider
class TestGetReportUtil(TestCase):
def get_report_class(self, report_class):
obj_dict = {
'get_rows': lambda slf, datespan: [
OPMCase(
forms=[],
edd=date(2014, 11, 10),
),
OPMCase(
forms=[],
dod=date(2014, 1, 12),
),
OPMCase(
forms=[],
dod=date(2014, 3, 12),
),
],
'data_provider': MockDataProvider(),
}
return type(report_class.__name__, (Report, report_class), obj_dict)
def test_basic_BPR(self):
report_class = self.get_report_class(BeneficiaryPaymentReport)
report = get_report(report_class, month=6, year=2014, block="Atri")
report.rows
def test_basic_CMR(self):
report_class = self.get_report_class(MetReport)
report = get_report(report_class, month=6, year=2014, block="Atri")
report.rows
## Instruction:
Fix for test (add mock for CommCareUser)
## Code After:
from datetime import date
from unittest import TestCase
from mock import patch
from corehq.apps.users.models import CommCareUser
from couchforms.models import XFormInstance
from ..constants import *
from ..reports import get_report, BeneficiaryPaymentReport, MetReport
from .case_reports import Report, OPMCase, MockCaseRow, MockDataProvider
class TestGetReportUtil(TestCase):
def get_report_class(self, report_class):
obj_dict = {
'get_rows': lambda slf, datespan: [
OPMCase(
forms=[],
edd=date(2014, 11, 10),
),
OPMCase(
forms=[],
dod=date(2014, 1, 12),
),
OPMCase(
forms=[],
dod=date(2014, 3, 12),
),
],
'data_provider': MockDataProvider(),
}
return type(report_class.__name__, (Report, report_class), obj_dict)
def test_basic_BPR(self):
report_class = self.get_report_class(BeneficiaryPaymentReport)
report = get_report(report_class, month=6, year=2014, block="Atri")
report.rows
@patch.object(CommCareUser, 'by_domain', return_value=[])
def test_basic_CMR(self, user_mock):
report_class = self.get_report_class(MetReport)
report = get_report(report_class, month=6, year=2014, block="Atri")
report.rows
| # ... existing code ...
from unittest import TestCase
from mock import patch
from corehq.apps.users.models import CommCareUser
from couchforms.models import XFormInstance
# ... modified code ...
@patch.object(CommCareUser, 'by_domain', return_value=[])
def test_basic_CMR(self, user_mock):
report_class = self.get_report_class(MetReport)
# ... rest of the code ... |
06f10e09f5b1c5766815b6e7eb219b4e33082709 | check_urls.py | check_urls.py |
import re, sys, markdown, requests, bs4 as BeautifulSoup
reload(sys)
sys.setdefaultencoding('utf8')
def check_url(url):
try:
return bool(requests.head(url, allow_redirects=True))
except Exception as e:
print 'Error checking URL %s: %s' % (url, e)
return False
def retrieve_urls(filename):
with open(filename) as fd:
mdtext = fd.read()
html_text = markdown.markdown(mdtext)
soup = BeautifulSoup.BeautifulSoup(html_text, "html.parser")
return [a['href'] for a in soup.findAll('a')]
def check_urls(filename):
print 'checking URLs for %s' % (filename,)
ok = True
for url in retrieve_urls(filename):
r = "(?:http[s]?://[^)]+)"
u = re.findall(r, url)
if not u: continue
msg = 'Checking %s => ' % (u[0],)
if check_url(u[0]):
print msg, 'OK'
else:
print msg, 'FAILED'
ok = False
return ok
def main():
ok = True
for filename in sys.argv[1:]:
try:
ok &= check_urls(filename)
except IOError as e:
print e
ok = False
exit (0 if ok else 1)
if __name__ == '__main__':
main()
|
from __future__ import print_function
import re, sys, markdown, requests, bs4 as BeautifulSoup
try: # Python 2
reload
except NameError: # Python 3
from importlib import reload
reload(sys)
sys.setdefaultencoding('utf8')
def check_url(url):
try:
return bool(requests.head(url, allow_redirects=True))
except Exception as e:
print('Error checking URL %s: %s' % (url, e))
return False
def retrieve_urls(filename):
with open(filename) as fd:
mdtext = fd.read()
html_text = markdown.markdown(mdtext)
soup = BeautifulSoup.BeautifulSoup(html_text, "html.parser")
return [a['href'] for a in soup.findAll('a')]
def check_urls(filename):
print('checking URLs for %s' % filename)
ok = True
for url in retrieve_urls(filename):
r = "(?:http[s]?://[^)]+)"
u = re.findall(r, url)
if not u: continue
msg = 'Checking %s => ' % (u[0],)
if check_url(u[0]):
print(msg, 'OK')
else:
print(msg, 'FAILED')
ok = False
return ok
def main():
ok = True
for filename in sys.argv[1:]:
try:
ok &= check_urls(filename)
except IOError as e:
print(e)
ok = False
exit(0 if ok else 1)
if __name__ == '__main__':
main()
| Add Python 3 compatibility and flake8 testing | Add Python 3 compatibility and flake8 testing | Python | unlicense | ligurio/free-software-testing-books |
+ from __future__ import print_function
import re, sys, markdown, requests, bs4 as BeautifulSoup
+
+ try: # Python 2
+ reload
+ except NameError: # Python 3
+ from importlib import reload
reload(sys)
sys.setdefaultencoding('utf8')
def check_url(url):
try:
return bool(requests.head(url, allow_redirects=True))
except Exception as e:
- print 'Error checking URL %s: %s' % (url, e)
+ print('Error checking URL %s: %s' % (url, e))
return False
def retrieve_urls(filename):
with open(filename) as fd:
mdtext = fd.read()
html_text = markdown.markdown(mdtext)
soup = BeautifulSoup.BeautifulSoup(html_text, "html.parser")
return [a['href'] for a in soup.findAll('a')]
def check_urls(filename):
- print 'checking URLs for %s' % (filename,)
+ print('checking URLs for %s' % filename)
ok = True
for url in retrieve_urls(filename):
r = "(?:http[s]?://[^)]+)"
u = re.findall(r, url)
if not u: continue
msg = 'Checking %s => ' % (u[0],)
if check_url(u[0]):
- print msg, 'OK'
+ print(msg, 'OK')
else:
- print msg, 'FAILED'
+ print(msg, 'FAILED')
ok = False
return ok
def main():
ok = True
for filename in sys.argv[1:]:
try:
ok &= check_urls(filename)
except IOError as e:
- print e
+ print(e)
ok = False
- exit (0 if ok else 1)
+ exit(0 if ok else 1)
if __name__ == '__main__':
main()
| Add Python 3 compatibility and flake8 testing | ## Code Before:
import re, sys, markdown, requests, bs4 as BeautifulSoup
reload(sys)
sys.setdefaultencoding('utf8')
def check_url(url):
try:
return bool(requests.head(url, allow_redirects=True))
except Exception as e:
print 'Error checking URL %s: %s' % (url, e)
return False
def retrieve_urls(filename):
with open(filename) as fd:
mdtext = fd.read()
html_text = markdown.markdown(mdtext)
soup = BeautifulSoup.BeautifulSoup(html_text, "html.parser")
return [a['href'] for a in soup.findAll('a')]
def check_urls(filename):
print 'checking URLs for %s' % (filename,)
ok = True
for url in retrieve_urls(filename):
r = "(?:http[s]?://[^)]+)"
u = re.findall(r, url)
if not u: continue
msg = 'Checking %s => ' % (u[0],)
if check_url(u[0]):
print msg, 'OK'
else:
print msg, 'FAILED'
ok = False
return ok
def main():
ok = True
for filename in sys.argv[1:]:
try:
ok &= check_urls(filename)
except IOError as e:
print e
ok = False
exit (0 if ok else 1)
if __name__ == '__main__':
main()
## Instruction:
Add Python 3 compatibility and flake8 testing
## Code After:
from __future__ import print_function
import re, sys, markdown, requests, bs4 as BeautifulSoup
try: # Python 2
reload
except NameError: # Python 3
from importlib import reload
reload(sys)
sys.setdefaultencoding('utf8')
def check_url(url):
try:
return bool(requests.head(url, allow_redirects=True))
except Exception as e:
print('Error checking URL %s: %s' % (url, e))
return False
def retrieve_urls(filename):
with open(filename) as fd:
mdtext = fd.read()
html_text = markdown.markdown(mdtext)
soup = BeautifulSoup.BeautifulSoup(html_text, "html.parser")
return [a['href'] for a in soup.findAll('a')]
def check_urls(filename):
print('checking URLs for %s' % filename)
ok = True
for url in retrieve_urls(filename):
r = "(?:http[s]?://[^)]+)"
u = re.findall(r, url)
if not u: continue
msg = 'Checking %s => ' % (u[0],)
if check_url(u[0]):
print(msg, 'OK')
else:
print(msg, 'FAILED')
ok = False
return ok
def main():
ok = True
for filename in sys.argv[1:]:
try:
ok &= check_urls(filename)
except IOError as e:
print(e)
ok = False
exit(0 if ok else 1)
if __name__ == '__main__':
main()
| // ... existing code ...
from __future__ import print_function
import re, sys, markdown, requests, bs4 as BeautifulSoup
try: # Python 2
reload
except NameError: # Python 3
from importlib import reload
// ... modified code ...
except Exception as e:
print('Error checking URL %s: %s' % (url, e))
return False
...
def check_urls(filename):
print('checking URLs for %s' % filename)
ok = True
...
if check_url(u[0]):
print(msg, 'OK')
else:
print(msg, 'FAILED')
ok = False
...
except IOError as e:
print(e)
ok = False
exit(0 if ok else 1)
// ... rest of the code ... |
6d8b99b5e4dab49c5a2e90b07f02072c116a7367 | robots/models.py | robots/models.py | from django.db import models
from django.contrib.sites.models import Site
from django.utils.translation import ugettext_lazy as _
class File(models.Model):
site = models.OneToOneField(Site, verbose_name=_(u'site'))
content = models.TextField(_(u'file content'))
objects = models.Manager()
class Meta:
verbose_name = _(u'robots.txt file')
verbose_name_plural = _(u'robots.txt files')
def __unicode__(self):
return u'/'.join([self.site.domain, u'robots.txt'])
| from django.db import models
from django.contrib.sites.models import Site
from django.utils.translation import ugettext_lazy as _
class File(models.Model):
site = models.OneToOneField(Site, verbose_name=_(u'site'))
content = models.TextField(_(u'file content'))
class Meta:
verbose_name = _(u'robots.txt file')
verbose_name_plural = _(u'robots.txt files')
def __unicode__(self):
return u'/'.join([self.site.domain, u'robots.txt'])
| Remove unnecessary manager declaration from File model | Remove unnecessary manager declaration from File model
| Python | isc | trilan/lemon-robots,trilan/lemon-robots | from django.db import models
from django.contrib.sites.models import Site
from django.utils.translation import ugettext_lazy as _
class File(models.Model):
site = models.OneToOneField(Site, verbose_name=_(u'site'))
content = models.TextField(_(u'file content'))
- objects = models.Manager()
-
class Meta:
verbose_name = _(u'robots.txt file')
verbose_name_plural = _(u'robots.txt files')
def __unicode__(self):
return u'/'.join([self.site.domain, u'robots.txt'])
| Remove unnecessary manager declaration from File model | ## Code Before:
from django.db import models
from django.contrib.sites.models import Site
from django.utils.translation import ugettext_lazy as _
class File(models.Model):
site = models.OneToOneField(Site, verbose_name=_(u'site'))
content = models.TextField(_(u'file content'))
objects = models.Manager()
class Meta:
verbose_name = _(u'robots.txt file')
verbose_name_plural = _(u'robots.txt files')
def __unicode__(self):
return u'/'.join([self.site.domain, u'robots.txt'])
## Instruction:
Remove unnecessary manager declaration from File model
## Code After:
from django.db import models
from django.contrib.sites.models import Site
from django.utils.translation import ugettext_lazy as _
class File(models.Model):
site = models.OneToOneField(Site, verbose_name=_(u'site'))
content = models.TextField(_(u'file content'))
class Meta:
verbose_name = _(u'robots.txt file')
verbose_name_plural = _(u'robots.txt files')
def __unicode__(self):
return u'/'.join([self.site.domain, u'robots.txt'])
| ...
class Meta:
... |
638ea1b12b71f74b357d60b09f1284625db73b2d | migrations/versions/0040_adjust_mmg_provider_rate.py | migrations/versions/0040_adjust_mmg_provider_rate.py |
# revision identifiers, used by Alembic.
revision = '0040_adjust_mmg_provider_rate'
down_revision = '0039_fix_notifications'
import uuid
from datetime import datetime
from alembic import op
import sqlalchemy as sa
def upgrade():
### commands auto generated by Alembic - please adjust! ###
conn = op.get_bind()
conn.execute(
sa.sql.text(("INSERT INTO provider_rates (id, valid_from, rate, provider_id) "
"VALUES (:id, :valid_from, :rate, (SELECT id FROM provider_details WHERE identifier = 'mmg'))")),
id=uuid.uuid4(),
valid_from=datetime.utcnow(),
rate=1.65
)
### end Alembic commands ###
def downgrade():
### commands auto generated by Alembic - please adjust! ###
conn = op.get_bind()
conn.execute(("DELETE FROM provider_rates "
"WHERE provider_id = (SELECT id FROM provider_details WHERE identifier = 'mmg') "
"AND rate = 1.65"))
### end Alembic commands ###
|
# revision identifiers, used by Alembic.
revision = '0040_adjust_mmg_provider_rate'
down_revision = '0039_fix_notifications'
import uuid
from datetime import datetime
from alembic import op
import sqlalchemy as sa
def upgrade():
### commands auto generated by Alembic - please adjust! ###
conn = op.get_bind()
conn.execute(
sa.sql.text(("INSERT INTO provider_rates (id, valid_from, rate, provider_id) "
"VALUES (:id, :valid_from, :rate, (SELECT id FROM provider_details WHERE identifier = 'mmg'))")),
id=uuid.uuid4(),
valid_from=datetime(2016, 7, 1),
rate=1.65
)
### end Alembic commands ###
def downgrade():
### commands auto generated by Alembic - please adjust! ###
conn = op.get_bind()
conn.execute(("DELETE FROM provider_rates "
"WHERE provider_id = (SELECT id FROM provider_details WHERE identifier = 'mmg') "
"AND rate = 1.65"))
### end Alembic commands ###
| Set the start date for the new rate as July 1 | Set the start date for the new rate as July 1
| Python | mit | alphagov/notifications-api,alphagov/notifications-api |
# revision identifiers, used by Alembic.
revision = '0040_adjust_mmg_provider_rate'
down_revision = '0039_fix_notifications'
import uuid
from datetime import datetime
from alembic import op
import sqlalchemy as sa
def upgrade():
### commands auto generated by Alembic - please adjust! ###
conn = op.get_bind()
conn.execute(
sa.sql.text(("INSERT INTO provider_rates (id, valid_from, rate, provider_id) "
"VALUES (:id, :valid_from, :rate, (SELECT id FROM provider_details WHERE identifier = 'mmg'))")),
id=uuid.uuid4(),
- valid_from=datetime.utcnow(),
+ valid_from=datetime(2016, 7, 1),
rate=1.65
)
### end Alembic commands ###
def downgrade():
### commands auto generated by Alembic - please adjust! ###
conn = op.get_bind()
conn.execute(("DELETE FROM provider_rates "
"WHERE provider_id = (SELECT id FROM provider_details WHERE identifier = 'mmg') "
"AND rate = 1.65"))
### end Alembic commands ###
| Set the start date for the new rate as July 1 | ## Code Before:
# revision identifiers, used by Alembic.
revision = '0040_adjust_mmg_provider_rate'
down_revision = '0039_fix_notifications'
import uuid
from datetime import datetime
from alembic import op
import sqlalchemy as sa
def upgrade():
### commands auto generated by Alembic - please adjust! ###
conn = op.get_bind()
conn.execute(
sa.sql.text(("INSERT INTO provider_rates (id, valid_from, rate, provider_id) "
"VALUES (:id, :valid_from, :rate, (SELECT id FROM provider_details WHERE identifier = 'mmg'))")),
id=uuid.uuid4(),
valid_from=datetime.utcnow(),
rate=1.65
)
### end Alembic commands ###
def downgrade():
### commands auto generated by Alembic - please adjust! ###
conn = op.get_bind()
conn.execute(("DELETE FROM provider_rates "
"WHERE provider_id = (SELECT id FROM provider_details WHERE identifier = 'mmg') "
"AND rate = 1.65"))
### end Alembic commands ###
## Instruction:
Set the start date for the new rate as July 1
## Code After:
# revision identifiers, used by Alembic.
revision = '0040_adjust_mmg_provider_rate'
down_revision = '0039_fix_notifications'
import uuid
from datetime import datetime
from alembic import op
import sqlalchemy as sa
def upgrade():
### commands auto generated by Alembic - please adjust! ###
conn = op.get_bind()
conn.execute(
sa.sql.text(("INSERT INTO provider_rates (id, valid_from, rate, provider_id) "
"VALUES (:id, :valid_from, :rate, (SELECT id FROM provider_details WHERE identifier = 'mmg'))")),
id=uuid.uuid4(),
valid_from=datetime(2016, 7, 1),
rate=1.65
)
### end Alembic commands ###
def downgrade():
### commands auto generated by Alembic - please adjust! ###
conn = op.get_bind()
conn.execute(("DELETE FROM provider_rates "
"WHERE provider_id = (SELECT id FROM provider_details WHERE identifier = 'mmg') "
"AND rate = 1.65"))
### end Alembic commands ###
| # ... existing code ...
id=uuid.uuid4(),
valid_from=datetime(2016, 7, 1),
rate=1.65
# ... rest of the code ... |
783766b4f4d65dfb4b41e6386edd8ea2df32d727 | tests/test_creation.py | tests/test_creation.py | import generic as g
class CreationTest(g.unittest.TestCase):
def test_soup(self):
count = 100
mesh = g.trimesh.creation.random_soup(face_count=count)
self.assertTrue(len(mesh.faces) == count)
self.assertTrue(len(mesh.face_adjacency) == 0)
self.assertTrue(len(mesh.split(only_watertight=True)) == 0)
self.assertTrue(len(mesh.split(only_watertight=False)) == count)
def test_uv(self):
sphere = g.trimesh.creation.uv_sphere()
self.assertTrue(sphere.is_watertight)
self.assertTrue(sphere.is_winding_consistent)
if __name__ == '__main__':
g.trimesh.util.attach_to_log()
g.unittest.main()
| import generic as g
class CreationTest(g.unittest.TestCase):
def test_soup(self):
count = 100
mesh = g.trimesh.creation.random_soup(face_count=count)
self.assertTrue(len(mesh.faces) == count)
self.assertTrue(len(mesh.face_adjacency) == 0)
self.assertTrue(len(mesh.split(only_watertight=True)) == 0)
self.assertTrue(len(mesh.split(only_watertight=False)) == count)
def test_uv(self):
sphere = g.trimesh.creation.uv_sphere()
self.assertTrue(sphere.is_watertight)
self.assertTrue(sphere.is_winding_consistent)
def test_path_extrude(self):
# Create base polygon
vec = g.np.array([0,1])*0.2
n_comps = 100
angle = g.np.pi * 2.0 / n_comps
rotmat = g.np.array([
[g.np.cos(angle), -g.np.sin(angle)],
[g.np.sin(angle), g.np.cos(angle)]
])
perim = []
for i in range(n_comps):
perim.append(vec)
vec = g.np.dot(rotmat, vec)
poly = g.Polygon(perim)
# Create 3D path
angles = g.np.linspace(0, 8*g.np.pi, 1000)
x = angles / 10.0
y = g.np.cos(angles)
z = g.np.sin(angles)
path = g.np.c_[x,y,z]
# Extrude
mesh = g.trimesh.creation.extrude_polygon_along_path(poly, path)
self.assertTrue(mesh.is_volume)
if __name__ == '__main__':
g.trimesh.util.attach_to_log()
g.unittest.main()
| Add integration test for extrusion | Add integration test for extrusion
| Python | mit | mikedh/trimesh,mikedh/trimesh,mikedh/trimesh,dajusc/trimesh,mikedh/trimesh,dajusc/trimesh | import generic as g
class CreationTest(g.unittest.TestCase):
def test_soup(self):
count = 100
mesh = g.trimesh.creation.random_soup(face_count=count)
self.assertTrue(len(mesh.faces) == count)
self.assertTrue(len(mesh.face_adjacency) == 0)
self.assertTrue(len(mesh.split(only_watertight=True)) == 0)
self.assertTrue(len(mesh.split(only_watertight=False)) == count)
def test_uv(self):
sphere = g.trimesh.creation.uv_sphere()
self.assertTrue(sphere.is_watertight)
self.assertTrue(sphere.is_winding_consistent)
+ def test_path_extrude(self):
+ # Create base polygon
+ vec = g.np.array([0,1])*0.2
+ n_comps = 100
+ angle = g.np.pi * 2.0 / n_comps
+ rotmat = g.np.array([
+ [g.np.cos(angle), -g.np.sin(angle)],
+ [g.np.sin(angle), g.np.cos(angle)]
+ ])
+ perim = []
+ for i in range(n_comps):
+ perim.append(vec)
+ vec = g.np.dot(rotmat, vec)
+ poly = g.Polygon(perim)
+
+ # Create 3D path
+ angles = g.np.linspace(0, 8*g.np.pi, 1000)
+ x = angles / 10.0
+ y = g.np.cos(angles)
+ z = g.np.sin(angles)
+ path = g.np.c_[x,y,z]
+
+ # Extrude
+ mesh = g.trimesh.creation.extrude_polygon_along_path(poly, path)
+ self.assertTrue(mesh.is_volume)
+
if __name__ == '__main__':
g.trimesh.util.attach_to_log()
g.unittest.main()
| Add integration test for extrusion | ## Code Before:
import generic as g
class CreationTest(g.unittest.TestCase):
def test_soup(self):
count = 100
mesh = g.trimesh.creation.random_soup(face_count=count)
self.assertTrue(len(mesh.faces) == count)
self.assertTrue(len(mesh.face_adjacency) == 0)
self.assertTrue(len(mesh.split(only_watertight=True)) == 0)
self.assertTrue(len(mesh.split(only_watertight=False)) == count)
def test_uv(self):
sphere = g.trimesh.creation.uv_sphere()
self.assertTrue(sphere.is_watertight)
self.assertTrue(sphere.is_winding_consistent)
if __name__ == '__main__':
g.trimesh.util.attach_to_log()
g.unittest.main()
## Instruction:
Add integration test for extrusion
## Code After:
import generic as g
class CreationTest(g.unittest.TestCase):
def test_soup(self):
count = 100
mesh = g.trimesh.creation.random_soup(face_count=count)
self.assertTrue(len(mesh.faces) == count)
self.assertTrue(len(mesh.face_adjacency) == 0)
self.assertTrue(len(mesh.split(only_watertight=True)) == 0)
self.assertTrue(len(mesh.split(only_watertight=False)) == count)
def test_uv(self):
sphere = g.trimesh.creation.uv_sphere()
self.assertTrue(sphere.is_watertight)
self.assertTrue(sphere.is_winding_consistent)
def test_path_extrude(self):
# Create base polygon
vec = g.np.array([0,1])*0.2
n_comps = 100
angle = g.np.pi * 2.0 / n_comps
rotmat = g.np.array([
[g.np.cos(angle), -g.np.sin(angle)],
[g.np.sin(angle), g.np.cos(angle)]
])
perim = []
for i in range(n_comps):
perim.append(vec)
vec = g.np.dot(rotmat, vec)
poly = g.Polygon(perim)
# Create 3D path
angles = g.np.linspace(0, 8*g.np.pi, 1000)
x = angles / 10.0
y = g.np.cos(angles)
z = g.np.sin(angles)
path = g.np.c_[x,y,z]
# Extrude
mesh = g.trimesh.creation.extrude_polygon_along_path(poly, path)
self.assertTrue(mesh.is_volume)
if __name__ == '__main__':
g.trimesh.util.attach_to_log()
g.unittest.main()
| // ... existing code ...
def test_path_extrude(self):
# Create base polygon
vec = g.np.array([0,1])*0.2
n_comps = 100
angle = g.np.pi * 2.0 / n_comps
rotmat = g.np.array([
[g.np.cos(angle), -g.np.sin(angle)],
[g.np.sin(angle), g.np.cos(angle)]
])
perim = []
for i in range(n_comps):
perim.append(vec)
vec = g.np.dot(rotmat, vec)
poly = g.Polygon(perim)
# Create 3D path
angles = g.np.linspace(0, 8*g.np.pi, 1000)
x = angles / 10.0
y = g.np.cos(angles)
z = g.np.sin(angles)
path = g.np.c_[x,y,z]
# Extrude
mesh = g.trimesh.creation.extrude_polygon_along_path(poly, path)
self.assertTrue(mesh.is_volume)
// ... rest of the code ... |
1f914a04adb4ad7d39ca7104e2ea36acc76b18bd | pvextractor/tests/test_gui.py | pvextractor/tests/test_gui.py | import numpy as np
from numpy.testing import assert_allclose
import pytest
from astropy.io import fits
from ..pvextractor import extract_pv_slice
from ..geometry.path import Path
from ..gui import PVSlicer
from .test_slicer import make_test_hdu
try:
import PyQt5
PYQT5OK = True
except ImportError:
PYQT5OK = False
import matplotlib as mpl
if mpl.__version__[0] == '2':
MPLOK = False
else:
MPLOK = True
@pytest.mark.skipif('not PYQT5OK or not MPLOK')
def test_gui():
hdu = make_test_hdu()
pv = PVSlicer(hdu, clim=(-0.02, 2))
pv.show(block=False)
x = [100,200,220,330,340]
y = [100,200,300,420,430]
for i in range(len(x)):
pv.fig.canvas.motion_notify_event(x[i],y[i])
pv.fig.canvas.button_press_event(x[i],y[i],1)
pv.fig.canvas.key_press_event('enter')
pv.fig.canvas.motion_notify_event(310,420)
pv.fig.canvas.button_press_event(410,420,1)
pv.fig.canvas.draw()
assert pv.pv_slice.data.shape == (5,2)
| import pytest
from distutils.version import LooseVersion
import matplotlib as mpl
from ..gui import PVSlicer
from .test_slicer import make_test_hdu
try:
import PyQt5
PYQT5OK = True
except ImportError:
PYQT5OK = False
if LooseVersion(mpl.__version__) < LooseVersion('2'):
MPLOK = True
else:
MPLOK = False
@pytest.mark.skipif('not PYQT5OK or not MPLOK')
def test_gui():
hdu = make_test_hdu()
pv = PVSlicer(hdu, clim=(-0.02, 2))
pv.show(block=False)
x = [100, 200, 220, 330, 340]
y = [100, 200, 300, 420, 430]
for i in range(len(x)):
pv.fig.canvas.motion_notify_event(x[i], y[i])
pv.fig.canvas.button_press_event(x[i], y[i], 1)
pv.fig.canvas.key_press_event('enter')
pv.fig.canvas.motion_notify_event(310, 420)
pv.fig.canvas.button_press_event(410, 420, 1)
pv.fig.canvas.draw()
assert pv.pv_slice.data.shape == (5, 2)
| Use LooseVersion to compare version numbers | Use LooseVersion to compare version numbers
| Python | bsd-3-clause | radio-astro-tools/pvextractor,keflavich/pvextractor | - import numpy as np
- from numpy.testing import assert_allclose
import pytest
+ from distutils.version import LooseVersion
+ import matplotlib as mpl
- from astropy.io import fits
- from ..pvextractor import extract_pv_slice
- from ..geometry.path import Path
from ..gui import PVSlicer
from .test_slicer import make_test_hdu
try:
import PyQt5
PYQT5OK = True
except ImportError:
PYQT5OK = False
- import matplotlib as mpl
- if mpl.__version__[0] == '2':
+
+ if LooseVersion(mpl.__version__) < LooseVersion('2'):
+ MPLOK = True
+ else:
MPLOK = False
+
- else:
- MPLOK = True
@pytest.mark.skipif('not PYQT5OK or not MPLOK')
def test_gui():
hdu = make_test_hdu()
pv = PVSlicer(hdu, clim=(-0.02, 2))
pv.show(block=False)
- x = [100,200,220,330,340]
+ x = [100, 200, 220, 330, 340]
- y = [100,200,300,420,430]
+ y = [100, 200, 300, 420, 430]
for i in range(len(x)):
- pv.fig.canvas.motion_notify_event(x[i],y[i])
+ pv.fig.canvas.motion_notify_event(x[i], y[i])
- pv.fig.canvas.button_press_event(x[i],y[i],1)
+ pv.fig.canvas.button_press_event(x[i], y[i], 1)
pv.fig.canvas.key_press_event('enter')
- pv.fig.canvas.motion_notify_event(310,420)
+ pv.fig.canvas.motion_notify_event(310, 420)
- pv.fig.canvas.button_press_event(410,420,1)
+ pv.fig.canvas.button_press_event(410, 420, 1)
pv.fig.canvas.draw()
- assert pv.pv_slice.data.shape == (5,2)
+ assert pv.pv_slice.data.shape == (5, 2)
| Use LooseVersion to compare version numbers | ## Code Before:
import numpy as np
from numpy.testing import assert_allclose
import pytest
from astropy.io import fits
from ..pvextractor import extract_pv_slice
from ..geometry.path import Path
from ..gui import PVSlicer
from .test_slicer import make_test_hdu
try:
import PyQt5
PYQT5OK = True
except ImportError:
PYQT5OK = False
import matplotlib as mpl
if mpl.__version__[0] == '2':
MPLOK = False
else:
MPLOK = True
@pytest.mark.skipif('not PYQT5OK or not MPLOK')
def test_gui():
hdu = make_test_hdu()
pv = PVSlicer(hdu, clim=(-0.02, 2))
pv.show(block=False)
x = [100,200,220,330,340]
y = [100,200,300,420,430]
for i in range(len(x)):
pv.fig.canvas.motion_notify_event(x[i],y[i])
pv.fig.canvas.button_press_event(x[i],y[i],1)
pv.fig.canvas.key_press_event('enter')
pv.fig.canvas.motion_notify_event(310,420)
pv.fig.canvas.button_press_event(410,420,1)
pv.fig.canvas.draw()
assert pv.pv_slice.data.shape == (5,2)
## Instruction:
Use LooseVersion to compare version numbers
## Code After:
import pytest
from distutils.version import LooseVersion
import matplotlib as mpl
from ..gui import PVSlicer
from .test_slicer import make_test_hdu
try:
import PyQt5
PYQT5OK = True
except ImportError:
PYQT5OK = False
if LooseVersion(mpl.__version__) < LooseVersion('2'):
MPLOK = True
else:
MPLOK = False
@pytest.mark.skipif('not PYQT5OK or not MPLOK')
def test_gui():
hdu = make_test_hdu()
pv = PVSlicer(hdu, clim=(-0.02, 2))
pv.show(block=False)
x = [100, 200, 220, 330, 340]
y = [100, 200, 300, 420, 430]
for i in range(len(x)):
pv.fig.canvas.motion_notify_event(x[i], y[i])
pv.fig.canvas.button_press_event(x[i], y[i], 1)
pv.fig.canvas.key_press_event('enter')
pv.fig.canvas.motion_notify_event(310, 420)
pv.fig.canvas.button_press_event(410, 420, 1)
pv.fig.canvas.draw()
assert pv.pv_slice.data.shape == (5, 2)
| ...
import pytest
from distutils.version import LooseVersion
import matplotlib as mpl
from ..gui import PVSlicer
...
if LooseVersion(mpl.__version__) < LooseVersion('2'):
MPLOK = True
else:
MPLOK = False
...
x = [100, 200, 220, 330, 340]
y = [100, 200, 300, 420, 430]
...
for i in range(len(x)):
pv.fig.canvas.motion_notify_event(x[i], y[i])
pv.fig.canvas.button_press_event(x[i], y[i], 1)
...
pv.fig.canvas.key_press_event('enter')
pv.fig.canvas.motion_notify_event(310, 420)
pv.fig.canvas.button_press_event(410, 420, 1)
...
assert pv.pv_slice.data.shape == (5, 2)
... |
1bd344a3ccda43f4ac1d4b94b1a18fc816c9b6ae | slurmscale/jobs/jobs.py | slurmscale/jobs/jobs.py | """Get info about jobs running on this cluster."""
import pyslurm
from job import Job
class Jobs(object):
"""A service object to inspect jobs."""
@property
def _jobs(self):
"""Fetch fresh data."""
return pyslurm.job().get()
def list(self):
"""List the current jobs on the cluster."""
current_jobs = self._jobs
return [Job(current_jobs[j]) for j in current_jobs]
| """Get info about jobs running on this cluster."""
import pyslurm
from job import Job
class Jobs(object):
"""A service object to inspect jobs."""
@property
def _jobs(self):
"""Fetch fresh data."""
return pyslurm.job().get()
def list(self, states=None):
"""
List the current jobs on the cluster.
:type states: List of ``str``
:param states: Filter jobs in the given states. Available states are
``PENDING``, ``RUNNING``, ``CANCELLED``, ``CONFIGURING``,
``COMPLETING``, ``COMPLETED``, ``FAILED``, ``TIMEOUT``,
``PREEMPTED``, ``NODE_FAIL`` and ``SPECIAL_EXIT``.
:rtype: List of ``Job``
:return: A list of current cluster jobs, possibly filtered by supplied
states.
"""
current_jobs = self._jobs
jobs = []
if states:
for i in current_jobs:
if current_jobs[i].get('job_state', '') in states:
jobs.append(Job(current_jobs[i]))
else:
jobs = [Job(current_jobs[j]) for j in current_jobs]
return jobs
| Add ability to filter job list by job state | Add ability to filter job list by job state
| Python | mit | afgane/slurmscale,afgane/slurmscale | """Get info about jobs running on this cluster."""
import pyslurm
from job import Job
class Jobs(object):
"""A service object to inspect jobs."""
@property
def _jobs(self):
"""Fetch fresh data."""
return pyslurm.job().get()
- def list(self):
+ def list(self, states=None):
+ """
- """List the current jobs on the cluster."""
+ List the current jobs on the cluster.
+
+ :type states: List of ``str``
+ :param states: Filter jobs in the given states. Available states are
+ ``PENDING``, ``RUNNING``, ``CANCELLED``, ``CONFIGURING``,
+ ``COMPLETING``, ``COMPLETED``, ``FAILED``, ``TIMEOUT``,
+ ``PREEMPTED``, ``NODE_FAIL`` and ``SPECIAL_EXIT``.
+
+ :rtype: List of ``Job``
+ :return: A list of current cluster jobs, possibly filtered by supplied
+ states.
+ """
current_jobs = self._jobs
+ jobs = []
+ if states:
+ for i in current_jobs:
+ if current_jobs[i].get('job_state', '') in states:
+ jobs.append(Job(current_jobs[i]))
+ else:
- return [Job(current_jobs[j]) for j in current_jobs]
+ jobs = [Job(current_jobs[j]) for j in current_jobs]
+ return jobs
| Add ability to filter job list by job state | ## Code Before:
"""Get info about jobs running on this cluster."""
import pyslurm
from job import Job
class Jobs(object):
"""A service object to inspect jobs."""
@property
def _jobs(self):
"""Fetch fresh data."""
return pyslurm.job().get()
def list(self):
"""List the current jobs on the cluster."""
current_jobs = self._jobs
return [Job(current_jobs[j]) for j in current_jobs]
## Instruction:
Add ability to filter job list by job state
## Code After:
"""Get info about jobs running on this cluster."""
import pyslurm
from job import Job
class Jobs(object):
"""A service object to inspect jobs."""
@property
def _jobs(self):
"""Fetch fresh data."""
return pyslurm.job().get()
def list(self, states=None):
"""
List the current jobs on the cluster.
:type states: List of ``str``
:param states: Filter jobs in the given states. Available states are
``PENDING``, ``RUNNING``, ``CANCELLED``, ``CONFIGURING``,
``COMPLETING``, ``COMPLETED``, ``FAILED``, ``TIMEOUT``,
``PREEMPTED``, ``NODE_FAIL`` and ``SPECIAL_EXIT``.
:rtype: List of ``Job``
:return: A list of current cluster jobs, possibly filtered by supplied
states.
"""
current_jobs = self._jobs
jobs = []
if states:
for i in current_jobs:
if current_jobs[i].get('job_state', '') in states:
jobs.append(Job(current_jobs[i]))
else:
jobs = [Job(current_jobs[j]) for j in current_jobs]
return jobs
| # ... existing code ...
def list(self, states=None):
"""
List the current jobs on the cluster.
:type states: List of ``str``
:param states: Filter jobs in the given states. Available states are
``PENDING``, ``RUNNING``, ``CANCELLED``, ``CONFIGURING``,
``COMPLETING``, ``COMPLETED``, ``FAILED``, ``TIMEOUT``,
``PREEMPTED``, ``NODE_FAIL`` and ``SPECIAL_EXIT``.
:rtype: List of ``Job``
:return: A list of current cluster jobs, possibly filtered by supplied
states.
"""
current_jobs = self._jobs
jobs = []
if states:
for i in current_jobs:
if current_jobs[i].get('job_state', '') in states:
jobs.append(Job(current_jobs[i]))
else:
jobs = [Job(current_jobs[j]) for j in current_jobs]
return jobs
# ... rest of the code ... |
0ac3ab3278e81aebe4717e0f599f752b4fda06d3 | examples/swat-s1/tests.py | examples/swat-s1/tests.py |
# from mininet.cli import CLI
from mininet.net import Mininet
from nose.plugins.skip import SkipTest
from utils import STATE, RWT_INIT_LEVEL
from utils import TANK_SECTION
from topo import SwatTopo
from physical_process import RawWaterTank
# import subprocess
# import sys
@SkipTest
def test_init():
pass
def test_topo():
topo = SwatTopo()
net = Mininet(topo=topo)
net.start()
net.pingAll()
net.stop()
def test_raw_water_tank():
RawWaterTank(
name='test_rwt',
state=STATE,
protocol=None,
section=TANK_SECTION,
level=RWT_INIT_LEVEL
)
|
# from mininet.cli import CLI
from mininet.net import Mininet
from utils import STATE, RWT_INIT_LEVEL
from utils import TANK_SECTION
from topo import SwatTopo
from physical_process import RawWaterTank
# import subprocess
# import sys
def test_init():
pass
def test_topo():
topo = SwatTopo()
net = Mininet(topo=topo)
net.start()
net.pingAll()
net.stop()
def test_raw_water_tank():
RawWaterTank(
name='test_rwt',
state=STATE,
protocol=None,
section=TANK_SECTION,
level=RWT_INIT_LEVEL
)
| Remove examples dep from nose | Remove examples dep from nose
| Python | mit | remmihsorp/minicps,scy-phy/minicps,remmihsorp/minicps,scy-phy/minicps |
# from mininet.cli import CLI
from mininet.net import Mininet
-
- from nose.plugins.skip import SkipTest
from utils import STATE, RWT_INIT_LEVEL
from utils import TANK_SECTION
from topo import SwatTopo
from physical_process import RawWaterTank
# import subprocess
# import sys
- @SkipTest
def test_init():
pass
def test_topo():
topo = SwatTopo()
net = Mininet(topo=topo)
net.start()
net.pingAll()
net.stop()
def test_raw_water_tank():
RawWaterTank(
name='test_rwt',
state=STATE,
protocol=None,
section=TANK_SECTION,
level=RWT_INIT_LEVEL
)
| Remove examples dep from nose | ## Code Before:
# from mininet.cli import CLI
from mininet.net import Mininet
from nose.plugins.skip import SkipTest
from utils import STATE, RWT_INIT_LEVEL
from utils import TANK_SECTION
from topo import SwatTopo
from physical_process import RawWaterTank
# import subprocess
# import sys
@SkipTest
def test_init():
pass
def test_topo():
topo = SwatTopo()
net = Mininet(topo=topo)
net.start()
net.pingAll()
net.stop()
def test_raw_water_tank():
RawWaterTank(
name='test_rwt',
state=STATE,
protocol=None,
section=TANK_SECTION,
level=RWT_INIT_LEVEL
)
## Instruction:
Remove examples dep from nose
## Code After:
# from mininet.cli import CLI
from mininet.net import Mininet
from utils import STATE, RWT_INIT_LEVEL
from utils import TANK_SECTION
from topo import SwatTopo
from physical_process import RawWaterTank
# import subprocess
# import sys
def test_init():
pass
def test_topo():
topo = SwatTopo()
net = Mininet(topo=topo)
net.start()
net.pingAll()
net.stop()
def test_raw_water_tank():
RawWaterTank(
name='test_rwt',
state=STATE,
protocol=None,
section=TANK_SECTION,
level=RWT_INIT_LEVEL
)
| ...
from mininet.net import Mininet
...
def test_init():
... |
e3312c773e9e3ac9b939bc3e0ca6a872dae5cdef | pre_commit_hooks/trailing_whitespace_fixer.py | pre_commit_hooks/trailing_whitespace_fixer.py | from __future__ import print_function
import argparse
import sys
from plumbum import local
from pre_commit_hooks.util import entry
@entry
def fix_trailing_whitespace(argv):
parser = argparse.ArgumentParser()
parser.add_argument('filenames', nargs='*', help='Filenames to fix')
args = parser.parse_args(argv)
bad_whitespace_files = local['grep'][
('-l', '[[:space:]]$') + tuple(args.filenames)
](retcode=None).strip().splitlines()
if bad_whitespace_files:
for bad_whitespace_file in bad_whitespace_files:
print('Fixing {0}'.format(bad_whitespace_file))
local['sed']['-i', '-e', 's/[[:space:]]*$//', bad_whitespace_file]()
return 1
else:
return 0
if __name__ == '__main__':
sys.exit(fix_trailing_whitespace())
| from __future__ import print_function
import argparse
import fileinput
import sys
from plumbum import local
from pre_commit_hooks.util import entry
def _fix_file(filename):
for line in fileinput.input([filename], inplace=True):
print(line.rstrip())
@entry
def fix_trailing_whitespace(argv):
parser = argparse.ArgumentParser()
parser.add_argument('filenames', nargs='*', help='Filenames to fix')
args = parser.parse_args(argv)
bad_whitespace_files = local['grep'][
('-l', '[[:space:]]$') + tuple(args.filenames)
](retcode=None).strip().splitlines()
if bad_whitespace_files:
for bad_whitespace_file in bad_whitespace_files:
print('Fixing {0}'.format(bad_whitespace_file))
_fix_file(bad_whitespace_file)
return 1
else:
return 0
if __name__ == '__main__':
sys.exit(fix_trailing_whitespace())
| Use fileinput instead of sed. | Use fileinput instead of sed.
| Python | mit | Coverfox/pre-commit-hooks,Harwood/pre-commit-hooks,bgschiller/pre-commit-hooks,pre-commit/pre-commit-hooks,jordant/pre-commit-hooks,jordant/pre-commit-hooks,chriskuehl/pre-commit-hooks,dupuy/pre-commit-hooks,arahayrabedian/pre-commit-hooks | from __future__ import print_function
import argparse
+ import fileinput
import sys
from plumbum import local
from pre_commit_hooks.util import entry
+
+
+ def _fix_file(filename):
+ for line in fileinput.input([filename], inplace=True):
+ print(line.rstrip())
@entry
def fix_trailing_whitespace(argv):
parser = argparse.ArgumentParser()
parser.add_argument('filenames', nargs='*', help='Filenames to fix')
args = parser.parse_args(argv)
bad_whitespace_files = local['grep'][
('-l', '[[:space:]]$') + tuple(args.filenames)
](retcode=None).strip().splitlines()
if bad_whitespace_files:
for bad_whitespace_file in bad_whitespace_files:
print('Fixing {0}'.format(bad_whitespace_file))
- local['sed']['-i', '-e', 's/[[:space:]]*$//', bad_whitespace_file]()
+ _fix_file(bad_whitespace_file)
return 1
else:
return 0
if __name__ == '__main__':
sys.exit(fix_trailing_whitespace())
| Use fileinput instead of sed. | ## Code Before:
from __future__ import print_function
import argparse
import sys
from plumbum import local
from pre_commit_hooks.util import entry
@entry
def fix_trailing_whitespace(argv):
parser = argparse.ArgumentParser()
parser.add_argument('filenames', nargs='*', help='Filenames to fix')
args = parser.parse_args(argv)
bad_whitespace_files = local['grep'][
('-l', '[[:space:]]$') + tuple(args.filenames)
](retcode=None).strip().splitlines()
if bad_whitespace_files:
for bad_whitespace_file in bad_whitespace_files:
print('Fixing {0}'.format(bad_whitespace_file))
local['sed']['-i', '-e', 's/[[:space:]]*$//', bad_whitespace_file]()
return 1
else:
return 0
if __name__ == '__main__':
sys.exit(fix_trailing_whitespace())
## Instruction:
Use fileinput instead of sed.
## Code After:
from __future__ import print_function
import argparse
import fileinput
import sys
from plumbum import local
from pre_commit_hooks.util import entry
def _fix_file(filename):
for line in fileinput.input([filename], inplace=True):
print(line.rstrip())
@entry
def fix_trailing_whitespace(argv):
parser = argparse.ArgumentParser()
parser.add_argument('filenames', nargs='*', help='Filenames to fix')
args = parser.parse_args(argv)
bad_whitespace_files = local['grep'][
('-l', '[[:space:]]$') + tuple(args.filenames)
](retcode=None).strip().splitlines()
if bad_whitespace_files:
for bad_whitespace_file in bad_whitespace_files:
print('Fixing {0}'.format(bad_whitespace_file))
_fix_file(bad_whitespace_file)
return 1
else:
return 0
if __name__ == '__main__':
sys.exit(fix_trailing_whitespace())
| # ... existing code ...
import argparse
import fileinput
import sys
# ... modified code ...
from pre_commit_hooks.util import entry
def _fix_file(filename):
for line in fileinput.input([filename], inplace=True):
print(line.rstrip())
...
print('Fixing {0}'.format(bad_whitespace_file))
_fix_file(bad_whitespace_file)
return 1
# ... rest of the code ... |
eb368c344075ce78606d4656ebfb19c7e7ccdf50 | src/054.py | src/054.py | from path import dirpath
def ans():
lines = open(dirpath() + '054.txt').readlines()
cards = [line.strip().split() for line in lines]
return None
if __name__ == '__main__':
print(ans())
| from collections import (
defaultdict,
namedtuple,
)
from path import dirpath
def _value(rank):
try:
return int(rank)
except ValueError:
return 10 + 'TJQKA'.index(rank)
def _sort_by_rank(hand):
return list(reversed(sorted(
hand,
key=lambda card: _value(card[0]),
)))
def _of_a_kind(hand, count):
counts = defaultdict(list)
for card in hand:
counts[card[0]].append(card)
filtered = {
rank: cards for
rank, cards in counts.items() if
count <= len(cards)
}
if len(filtered) < 1:
return None
return max(
filtered.values(),
key=lambda cards: _value(cards[0][0])
)
def high_card(hand):
return _of_a_kind(hand, 1)
def two_of_a_kind(hand):
return _of_a_kind(hand, 2)
def three_of_a_kind(hand):
return _of_a_kind(hand, 3)
def four_of_a_kind(hand):
return _of_a_kind(hand, 4)
def full_house(hand):
three = three_of_a_kind(hand)
if not three:
return None
pair = two_of_a_kind([card for card in hand if card not in three])
if not pair:
return None
return three + pair
def straight(hand):
sorted_ = sorted([_value(card[0]) for card in hand])
if sorted_ == list(range(sorted_[0], sorted_[-1] + 1)):
return _sort_by_rank(hand)
return None
def flush(hand):
counts = defaultdict(list)
for card in hand:
counts[card[1]].append(card)
for cards in counts.values():
if len(cards) == 5:
return _sort_by_rank(cards)
return None
def straight_flush(hand):
return flush(hand) if straight(hand) else None
def ans():
lines = open(dirpath() + '054.txt').readlines()
turns = [line.strip().split() for line in lines]
num_wins = 0
for cards in turns:
one = cards[:5]
two = cards[5:]
return None
if __name__ == '__main__':
print(ans())
| Write some logic for 54 | Write some logic for 54
| Python | mit | mackorone/euler | + from collections import (
+ defaultdict,
+ namedtuple,
+ )
from path import dirpath
+
+
+ def _value(rank):
+ try:
+ return int(rank)
+ except ValueError:
+ return 10 + 'TJQKA'.index(rank)
+
+
+ def _sort_by_rank(hand):
+ return list(reversed(sorted(
+ hand,
+ key=lambda card: _value(card[0]),
+ )))
+
+
+ def _of_a_kind(hand, count):
+ counts = defaultdict(list)
+ for card in hand:
+ counts[card[0]].append(card)
+ filtered = {
+ rank: cards for
+ rank, cards in counts.items() if
+ count <= len(cards)
+ }
+ if len(filtered) < 1:
+ return None
+ return max(
+ filtered.values(),
+ key=lambda cards: _value(cards[0][0])
+ )
+
+
+ def high_card(hand):
+ return _of_a_kind(hand, 1)
+
+
+ def two_of_a_kind(hand):
+ return _of_a_kind(hand, 2)
+
+
+ def three_of_a_kind(hand):
+ return _of_a_kind(hand, 3)
+
+
+ def four_of_a_kind(hand):
+ return _of_a_kind(hand, 4)
+
+
+ def full_house(hand):
+ three = three_of_a_kind(hand)
+ if not three:
+ return None
+ pair = two_of_a_kind([card for card in hand if card not in three])
+ if not pair:
+ return None
+ return three + pair
+
+
+ def straight(hand):
+ sorted_ = sorted([_value(card[0]) for card in hand])
+ if sorted_ == list(range(sorted_[0], sorted_[-1] + 1)):
+ return _sort_by_rank(hand)
+ return None
+
+
+ def flush(hand):
+ counts = defaultdict(list)
+ for card in hand:
+ counts[card[1]].append(card)
+ for cards in counts.values():
+ if len(cards) == 5:
+ return _sort_by_rank(cards)
+ return None
+
+
+ def straight_flush(hand):
+ return flush(hand) if straight(hand) else None
def ans():
lines = open(dirpath() + '054.txt').readlines()
- cards = [line.strip().split() for line in lines]
+ turns = [line.strip().split() for line in lines]
+ num_wins = 0
+ for cards in turns:
+ one = cards[:5]
+ two = cards[5:]
return None
if __name__ == '__main__':
print(ans())
| Write some logic for 54 | ## Code Before:
from path import dirpath
def ans():
lines = open(dirpath() + '054.txt').readlines()
cards = [line.strip().split() for line in lines]
return None
if __name__ == '__main__':
print(ans())
## Instruction:
Write some logic for 54
## Code After:
from collections import (
defaultdict,
namedtuple,
)
from path import dirpath
def _value(rank):
try:
return int(rank)
except ValueError:
return 10 + 'TJQKA'.index(rank)
def _sort_by_rank(hand):
return list(reversed(sorted(
hand,
key=lambda card: _value(card[0]),
)))
def _of_a_kind(hand, count):
counts = defaultdict(list)
for card in hand:
counts[card[0]].append(card)
filtered = {
rank: cards for
rank, cards in counts.items() if
count <= len(cards)
}
if len(filtered) < 1:
return None
return max(
filtered.values(),
key=lambda cards: _value(cards[0][0])
)
def high_card(hand):
return _of_a_kind(hand, 1)
def two_of_a_kind(hand):
return _of_a_kind(hand, 2)
def three_of_a_kind(hand):
return _of_a_kind(hand, 3)
def four_of_a_kind(hand):
return _of_a_kind(hand, 4)
def full_house(hand):
three = three_of_a_kind(hand)
if not three:
return None
pair = two_of_a_kind([card for card in hand if card not in three])
if not pair:
return None
return three + pair
def straight(hand):
sorted_ = sorted([_value(card[0]) for card in hand])
if sorted_ == list(range(sorted_[0], sorted_[-1] + 1)):
return _sort_by_rank(hand)
return None
def flush(hand):
counts = defaultdict(list)
for card in hand:
counts[card[1]].append(card)
for cards in counts.values():
if len(cards) == 5:
return _sort_by_rank(cards)
return None
def straight_flush(hand):
return flush(hand) if straight(hand) else None
def ans():
lines = open(dirpath() + '054.txt').readlines()
turns = [line.strip().split() for line in lines]
num_wins = 0
for cards in turns:
one = cards[:5]
two = cards[5:]
return None
if __name__ == '__main__':
print(ans())
| // ... existing code ...
from collections import (
defaultdict,
namedtuple,
)
from path import dirpath
def _value(rank):
try:
return int(rank)
except ValueError:
return 10 + 'TJQKA'.index(rank)
def _sort_by_rank(hand):
return list(reversed(sorted(
hand,
key=lambda card: _value(card[0]),
)))
def _of_a_kind(hand, count):
counts = defaultdict(list)
for card in hand:
counts[card[0]].append(card)
filtered = {
rank: cards for
rank, cards in counts.items() if
count <= len(cards)
}
if len(filtered) < 1:
return None
return max(
filtered.values(),
key=lambda cards: _value(cards[0][0])
)
def high_card(hand):
return _of_a_kind(hand, 1)
def two_of_a_kind(hand):
return _of_a_kind(hand, 2)
def three_of_a_kind(hand):
return _of_a_kind(hand, 3)
def four_of_a_kind(hand):
return _of_a_kind(hand, 4)
def full_house(hand):
three = three_of_a_kind(hand)
if not three:
return None
pair = two_of_a_kind([card for card in hand if card not in three])
if not pair:
return None
return three + pair
def straight(hand):
sorted_ = sorted([_value(card[0]) for card in hand])
if sorted_ == list(range(sorted_[0], sorted_[-1] + 1)):
return _sort_by_rank(hand)
return None
def flush(hand):
counts = defaultdict(list)
for card in hand:
counts[card[1]].append(card)
for cards in counts.values():
if len(cards) == 5:
return _sort_by_rank(cards)
return None
def straight_flush(hand):
return flush(hand) if straight(hand) else None
// ... modified code ...
lines = open(dirpath() + '054.txt').readlines()
turns = [line.strip().split() for line in lines]
num_wins = 0
for cards in turns:
one = cards[:5]
two = cards[5:]
// ... rest of the code ... |
4cdf5be2a3c01e1b16a5e49bdf770f9d8573e16e | icekit/utils/testing.py | icekit/utils/testing.py | import glob
import os
import uuid
from django.core.files.base import ContentFile
from PIL import Image
from StringIO import StringIO
def new_test_image():
"""
Creates an automatically generated test image.
In your testing `tearDown` method make sure to delete the test
image with the helper function `delete_test_image`.
The recommended way of using this helper function is as follows:
object_1.image_property.save(*new_test_image())
:return: Image name and image content file.
"""
image_name = 'test-{}.png'.format(uuid.uuid4())
image_buf = StringIO()
image = Image.new('RGBA', size=(50, 50), color=(256, 0, 0))
image.save(image_buf, 'png')
image_buf.seek(0)
return image_name, ContentFile(image_buf.read(), image_name)
def delete_test_image(image_field):
"""
Deletes test image generated as well as thumbnails if created.
The recommended way of using this helper function is as follows:
delete_test_image(object_1.image_property)
:param image_field: The image field on an object.
:return: None.
"""
# ensure all thumbs are deleted
for filename in glob.glob(
os.path.join('public', 'media', 'thumbs', image_field.name) + '*'):
os.unlink(filename)
# delete the saved file
image_field.delete()
| import glob
import os
import uuid
from PIL import Image
from django.core.files.base import ContentFile
from django.utils import six
def new_test_image():
"""
Creates an automatically generated test image.
In your testing `tearDown` method make sure to delete the test
image with the helper function `delete_test_image`.
The recommended way of using this helper function is as follows:
object_1.image_property.save(*new_test_image())
:return: Image name and image content file.
"""
image_name = 'test-{}.png'.format(uuid.uuid4())
image_buf = six.StringIO()
image = Image.new('RGBA', size=(50, 50), color=(256, 0, 0))
image.save(image_buf, 'png')
image_buf.seek(0)
return image_name, ContentFile(image_buf.read(), image_name)
def delete_test_image(image_field):
"""
Deletes test image generated as well as thumbnails if created.
The recommended way of using this helper function is as follows:
delete_test_image(object_1.image_property)
:param image_field: The image field on an object.
:return: None.
"""
# ensure all thumbs are deleted
for filename in glob.glob(
os.path.join('public', 'media', 'thumbs', image_field.name) + '*'):
os.unlink(filename)
# delete the saved file
image_field.delete()
| Update StringIO import for Python3 compat | Update StringIO import for Python3 compat
| Python | mit | ic-labs/django-icekit,ic-labs/django-icekit,ic-labs/django-icekit,ic-labs/django-icekit | import glob
import os
import uuid
+
+ from PIL import Image
+
from django.core.files.base import ContentFile
+ from django.utils import six
- from PIL import Image
- from StringIO import StringIO
def new_test_image():
"""
Creates an automatically generated test image.
In your testing `tearDown` method make sure to delete the test
image with the helper function `delete_test_image`.
The recommended way of using this helper function is as follows:
object_1.image_property.save(*new_test_image())
:return: Image name and image content file.
"""
image_name = 'test-{}.png'.format(uuid.uuid4())
- image_buf = StringIO()
+ image_buf = six.StringIO()
image = Image.new('RGBA', size=(50, 50), color=(256, 0, 0))
image.save(image_buf, 'png')
image_buf.seek(0)
return image_name, ContentFile(image_buf.read(), image_name)
def delete_test_image(image_field):
"""
Deletes test image generated as well as thumbnails if created.
The recommended way of using this helper function is as follows:
delete_test_image(object_1.image_property)
:param image_field: The image field on an object.
:return: None.
"""
# ensure all thumbs are deleted
for filename in glob.glob(
os.path.join('public', 'media', 'thumbs', image_field.name) + '*'):
os.unlink(filename)
# delete the saved file
image_field.delete()
| Update StringIO import for Python3 compat | ## Code Before:
import glob
import os
import uuid
from django.core.files.base import ContentFile
from PIL import Image
from StringIO import StringIO
def new_test_image():
"""
Creates an automatically generated test image.
In your testing `tearDown` method make sure to delete the test
image with the helper function `delete_test_image`.
The recommended way of using this helper function is as follows:
object_1.image_property.save(*new_test_image())
:return: Image name and image content file.
"""
image_name = 'test-{}.png'.format(uuid.uuid4())
image_buf = StringIO()
image = Image.new('RGBA', size=(50, 50), color=(256, 0, 0))
image.save(image_buf, 'png')
image_buf.seek(0)
return image_name, ContentFile(image_buf.read(), image_name)
def delete_test_image(image_field):
"""
Deletes test image generated as well as thumbnails if created.
The recommended way of using this helper function is as follows:
delete_test_image(object_1.image_property)
:param image_field: The image field on an object.
:return: None.
"""
# ensure all thumbs are deleted
for filename in glob.glob(
os.path.join('public', 'media', 'thumbs', image_field.name) + '*'):
os.unlink(filename)
# delete the saved file
image_field.delete()
## Instruction:
Update StringIO import for Python3 compat
## Code After:
import glob
import os
import uuid
from PIL import Image
from django.core.files.base import ContentFile
from django.utils import six
def new_test_image():
"""
Creates an automatically generated test image.
In your testing `tearDown` method make sure to delete the test
image with the helper function `delete_test_image`.
The recommended way of using this helper function is as follows:
object_1.image_property.save(*new_test_image())
:return: Image name and image content file.
"""
image_name = 'test-{}.png'.format(uuid.uuid4())
image_buf = six.StringIO()
image = Image.new('RGBA', size=(50, 50), color=(256, 0, 0))
image.save(image_buf, 'png')
image_buf.seek(0)
return image_name, ContentFile(image_buf.read(), image_name)
def delete_test_image(image_field):
"""
Deletes test image generated as well as thumbnails if created.
The recommended way of using this helper function is as follows:
delete_test_image(object_1.image_property)
:param image_field: The image field on an object.
:return: None.
"""
# ensure all thumbs are deleted
for filename in glob.glob(
os.path.join('public', 'media', 'thumbs', image_field.name) + '*'):
os.unlink(filename)
# delete the saved file
image_field.delete()
| ...
import uuid
from PIL import Image
from django.core.files.base import ContentFile
from django.utils import six
...
image_name = 'test-{}.png'.format(uuid.uuid4())
image_buf = six.StringIO()
image = Image.new('RGBA', size=(50, 50), color=(256, 0, 0))
... |
c224fdecf174077f3b7a15f056e65b10282fed38 | tasks.py | tasks.py | from invoke import Collection
from invocations import docs, testing
# TODO: let from_module specify new name
api = Collection.from_module(docs)
# TODO: maybe allow rolling configuration into it too heh
api.configure({
'sphinx.source': 'sites/docs',
'sphinx.target': 'sites/docs/_build',
})
api.name = 'docs'
main = Collection.from_module(docs)
main.name = 'main'
main.configure({
'sphinx.source': 'sites/main',
'sphinx.target': 'sites/main/_build',
})
ns = Collection(testing.test, docs=api, main=main)
| from invoke import Collection
from invocations import docs, testing
# Usage doc/API site
api = Collection.from_module(docs, name='docs', config={
'sphinx.source': 'sites/docs',
'sphinx.target': 'sites/docs/_build',
})
# Main/about/changelog site
main = Collection.from_module(docs, name='main', config={
'sphinx.source': 'sites/main',
'sphinx.target': 'sites/main/_build',
})
ns = Collection(testing.test, docs=api, main=main)
| Use new behavior from newer Invoke | Use new behavior from newer Invoke
| Python | lgpl-2.1 | mirrorcoder/paramiko,thisch/paramiko,paramiko/paramiko,remram44/paramiko,zpzgone/paramiko,dlitz/paramiko,Automatic/paramiko,varunarya10/paramiko,ameily/paramiko,mhdaimi/paramiko,redixin/paramiko,rcorrieri/paramiko,thusoy/paramiko,torkil/paramiko,SebastianDeiss/paramiko,zarr12steven/paramiko,dorianpula/paramiko,toby82/paramiko,esc/paramiko,digitalquacks/paramiko,jorik041/paramiko,CptLemming/paramiko,davidbistolas/paramiko,selboo/paramiko,reaperhulk/paramiko,jaraco/paramiko,anadigi/paramiko,fvicente/paramiko | from invoke import Collection
from invocations import docs, testing
+ # Usage doc/API site
+ api = Collection.from_module(docs, name='docs', config={
- # TODO: let from_module specify new name
- api = Collection.from_module(docs)
- # TODO: maybe allow rolling configuration into it too heh
- api.configure({
'sphinx.source': 'sites/docs',
'sphinx.target': 'sites/docs/_build',
})
+ # Main/about/changelog site
+ main = Collection.from_module(docs, name='main', config={
- api.name = 'docs'
- main = Collection.from_module(docs)
- main.name = 'main'
- main.configure({
'sphinx.source': 'sites/main',
'sphinx.target': 'sites/main/_build',
})
ns = Collection(testing.test, docs=api, main=main)
| Use new behavior from newer Invoke | ## Code Before:
from invoke import Collection
from invocations import docs, testing
# TODO: let from_module specify new name
api = Collection.from_module(docs)
# TODO: maybe allow rolling configuration into it too heh
api.configure({
'sphinx.source': 'sites/docs',
'sphinx.target': 'sites/docs/_build',
})
api.name = 'docs'
main = Collection.from_module(docs)
main.name = 'main'
main.configure({
'sphinx.source': 'sites/main',
'sphinx.target': 'sites/main/_build',
})
ns = Collection(testing.test, docs=api, main=main)
## Instruction:
Use new behavior from newer Invoke
## Code After:
from invoke import Collection
from invocations import docs, testing
# Usage doc/API site
api = Collection.from_module(docs, name='docs', config={
'sphinx.source': 'sites/docs',
'sphinx.target': 'sites/docs/_build',
})
# Main/about/changelog site
main = Collection.from_module(docs, name='main', config={
'sphinx.source': 'sites/main',
'sphinx.target': 'sites/main/_build',
})
ns = Collection(testing.test, docs=api, main=main)
| ...
# Usage doc/API site
api = Collection.from_module(docs, name='docs', config={
'sphinx.source': 'sites/docs',
...
})
# Main/about/changelog site
main = Collection.from_module(docs, name='main', config={
'sphinx.source': 'sites/main',
... |
96b06d80f6108997fab44ac1e6042fcae93cc82a | server.py | server.py | import json
import tornado.ioloop
import tornado.web
import Adafruit_BMP.BMP085 as BMP085
class SensorAccess(tornado.web.RequestHandler):
@tornado.web.asynchronous
def get(self):
self.write(json.dumps(self.read_sensor()))
self.finish()
def read_sensor(self):
pass
class TempSensorAccess(SensorAccess):
def read_sensor(self):
sensor = BMP085.BMP085(mode=BMP085.BMP085_ULTRAHIGHRES)
return {
'temperature': sensor.read_temperature(),
'pressure': sensor.read_pressure(),
}
class IndexHandler(tornado.web.RequestHandler):
@tornado.web.asynchronous
def get(self):
self.write(json.dumps({
'inde': 'pitools service'
}))
def start_server():
application = tornado.web.Application([
(r"/", IndexHandler),
(r"/sensors/env", TempSensorAccess),
])
application.listen(9876)
tornado.ioloop.IOLoop.instance().start()
if __name__ == "__main__":
start_server()
| import json
import tornado.ioloop
import tornado.web
import Adafruit_BMP.BMP085 as BMP085
class SensorAccess(tornado.web.RequestHandler):
@tornado.web.asynchronous
def get(self):
self.write(json.dumps(self.read_sensor()))
self.finish()
def read_sensor(self):
pass
class TempSensorAccess(SensorAccess):
def read_sensor(self):
sensor = BMP085.BMP085(mode=BMP085.BMP085_ULTRAHIGHRES)
return {
'temperature': sensor.read_temperature(),
'pressure': sensor.read_pressure(),
}
class IndexHandler(tornado.web.RequestHandler):
@tornado.web.asynchronous
def get(self):
self.write(json.dumps({
'index': 'pitools service'
}))
self.finish()
def start_server():
application = tornado.web.Application([
(r"/", IndexHandler),
(r"/sensors/env", TempSensorAccess),
])
application.listen(9876)
tornado.ioloop.IOLoop.instance().start()
if __name__ == "__main__":
start_server()
| Fix typo; Fix request never finish | Fix typo; Fix request never finish
| Python | bsd-2-clause | JokerQyou/pitools | import json
import tornado.ioloop
import tornado.web
import Adafruit_BMP.BMP085 as BMP085
class SensorAccess(tornado.web.RequestHandler):
@tornado.web.asynchronous
def get(self):
self.write(json.dumps(self.read_sensor()))
self.finish()
def read_sensor(self):
pass
class TempSensorAccess(SensorAccess):
def read_sensor(self):
sensor = BMP085.BMP085(mode=BMP085.BMP085_ULTRAHIGHRES)
return {
'temperature': sensor.read_temperature(),
'pressure': sensor.read_pressure(),
}
class IndexHandler(tornado.web.RequestHandler):
@tornado.web.asynchronous
def get(self):
self.write(json.dumps({
- 'inde': 'pitools service'
+ 'index': 'pitools service'
}))
+ self.finish()
def start_server():
application = tornado.web.Application([
(r"/", IndexHandler),
(r"/sensors/env", TempSensorAccess),
])
application.listen(9876)
tornado.ioloop.IOLoop.instance().start()
if __name__ == "__main__":
start_server()
| Fix typo; Fix request never finish | ## Code Before:
import json
import tornado.ioloop
import tornado.web
import Adafruit_BMP.BMP085 as BMP085
class SensorAccess(tornado.web.RequestHandler):
@tornado.web.asynchronous
def get(self):
self.write(json.dumps(self.read_sensor()))
self.finish()
def read_sensor(self):
pass
class TempSensorAccess(SensorAccess):
def read_sensor(self):
sensor = BMP085.BMP085(mode=BMP085.BMP085_ULTRAHIGHRES)
return {
'temperature': sensor.read_temperature(),
'pressure': sensor.read_pressure(),
}
class IndexHandler(tornado.web.RequestHandler):
@tornado.web.asynchronous
def get(self):
self.write(json.dumps({
'inde': 'pitools service'
}))
def start_server():
application = tornado.web.Application([
(r"/", IndexHandler),
(r"/sensors/env", TempSensorAccess),
])
application.listen(9876)
tornado.ioloop.IOLoop.instance().start()
if __name__ == "__main__":
start_server()
## Instruction:
Fix typo; Fix request never finish
## Code After:
import json
import tornado.ioloop
import tornado.web
import Adafruit_BMP.BMP085 as BMP085
class SensorAccess(tornado.web.RequestHandler):
@tornado.web.asynchronous
def get(self):
self.write(json.dumps(self.read_sensor()))
self.finish()
def read_sensor(self):
pass
class TempSensorAccess(SensorAccess):
def read_sensor(self):
sensor = BMP085.BMP085(mode=BMP085.BMP085_ULTRAHIGHRES)
return {
'temperature': sensor.read_temperature(),
'pressure': sensor.read_pressure(),
}
class IndexHandler(tornado.web.RequestHandler):
@tornado.web.asynchronous
def get(self):
self.write(json.dumps({
'index': 'pitools service'
}))
self.finish()
def start_server():
application = tornado.web.Application([
(r"/", IndexHandler),
(r"/sensors/env", TempSensorAccess),
])
application.listen(9876)
tornado.ioloop.IOLoop.instance().start()
if __name__ == "__main__":
start_server()
| # ... existing code ...
self.write(json.dumps({
'index': 'pitools service'
}))
self.finish()
# ... rest of the code ... |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.